Re: 2 issues after upgrade 9.4 -> 10.1

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Nikhil,

there are only 2 replicate 3 volumes, and yes, i did the
online-upgrade - a couple of times now, all worked well :-)

I should've been more precise: online upgrade on all 6 servers, on 5
servers the ports stayed the same, and only on one server the ports
changed - without reboot. First i thought that smth might have gone
wrong, rebooted the one server, but ports stayed different. So
directly after the online upgrade on that one server the ports were
different. But as ports don't matter anymore: issue solved :-)

And the mount issue solved as well, excellent! Thank you!


Best regards,
Hubert

Am Di., 1. März 2022 um 11:40 Uhr schrieb Nikhil Ladha <nladha@xxxxxxxxxx>:
>
> Hi Hu Bert,
>
> Do you have a distributed-replicate volume and you followed an online upgrade procedure? If so, then that is the reason that the ports are different on only 1 server as you mentioned you did a reboot on it.
> Secondly, the glusterfs mount issue is already fixed (check patch #3211) and will be available in the next release.
>
> --
> Thanks and Regards,
> NiKHIL LADHA
>
>
> On Tue, Mar 1, 2022 at 3:10 PM Hu Bert <revirii@xxxxxxxxxxxxxx> wrote:
>>
>> Hey,
>>
>> ok, i think i found the reason for the port issue:
>>
>> https://docs.gluster.org/en/latest/release-notes/10.0/
>> https://github.com/gluster/glusterfs/issues/786
>>
>> Should've look closer... mea culpa. But, quite interesting, that
>> happened only on one server, i upgraded 6 servers in total.
>>
>> So only the issue with the glusterfs mount and backup-volfile-servers remains.
>>
>>
>> Thx,
>> Hubert
>>
>> Am Di., 1. März 2022 um 06:19 Uhr schrieb Hu Bert <revirii@xxxxxxxxxxxxxx>:
>> >
>> > Good morning,
>> >
>> > just did an upgrade of 3 gluster volumes and x clients from 9.4 to
>> > 10.1. In principle the upgrade went fine, just 2 smaller issues
>> > appeared.
>> >
>> > 1) on one of the servers the ports are screwed up.
>> >
>> > gluster volume status
>> > Status of volume: workdata
>> > Gluster process                             TCP Port  RDMA Port  Online  Pid
>> > ------------------------------------------------------------------------------
>> > Brick glusterpub1:/gluster/md3/workdata     49152     0          Y       1452
>> > Brick glusterpub2:/gluster/md3/workdata     49152     0          Y       1839
>> > Brick glusterpub3:/gluster/md3/workdata     54105     0          Y       1974
>> > Brick glusterpub1:/gluster/md4/workdata     49153     0          Y       1459
>> > Brick glusterpub2:/gluster/md4/workdata     49153     0          Y       1849
>> > Brick glusterpub3:/gluster/md4/workdata     58177     0          Y       1997
>> > Brick glusterpub1:/gluster/md5/workdata     49154     0          Y       1468
>> > Brick glusterpub2:/gluster/md5/workdata     49154     0          Y       1857
>> > Brick glusterpub3:/gluster/md5/workdata     59071     0          Y       2003
>> > Brick glusterpub1:/gluster/md6/workdata     49155     0          Y       1481
>> > Brick glusterpub2:/gluster/md6/workdata     49155     0          Y       1868
>> > Brick glusterpub3:/gluster/md6/workdata     53309     0          Y       2008
>> > Brick glusterpub1:/gluster/md7/workdata     49156     0          Y       1490
>> > Brick glusterpub2:/gluster/md7/workdata     49156     0          Y       1878
>> > Brick glusterpub3:/gluster/md7/workdata     54310     0          Y       2027
>> > Self-heal Daemon on localhost               N/A       N/A        Y       2108
>> > Self-heal Daemon on glusterpub1             N/A       N/A        Y       1210749
>> > Self-heal Daemon on glusterpub2             N/A       N/A        Y       950871
>> >
>> > Task Status of Volume workdata
>> > ------------------------------------------------------------------------------
>> > There are no active volume tasks
>> >
>> > glusterpub3 has different ports. I know, this is no problem, the
>> > volume is good, but even after a reboot the ports stay like this.
>> >
>> > glustershd.log:
>> > [2022-03-01 04:58:13.993349 +0000] I
>> > [rpc-clnt.c:1969:rpc_clnt_reconfig] 0-workdata-client-0: changing port
>> > to 49152 (from 0)
>> > [2022-03-01 04:58:13.993410 +0000] I [socket.c:834:__socket_shutdown]
>> > 0-workdata-client-0: intentional socket shutdown(13)
>> > [............]
>> > [2022-03-01 04:58:14.008111 +0000] I
>> > [rpc-clnt.c:1969:rpc_clnt_reconfig] 0-workdata-client-1: changing port
>> > to 49152 (from 0)
>> > [2022-03-01 04:58:14.008148 +0000] I [socket.c:834:__socket_shutdown]
>> > 0-workdata-client-1: intentional socket shutdown(14)
>> > [............]
>> > [2022-03-01 04:58:14.011416 +0000] I
>> > [rpc-clnt.c:1969:rpc_clnt_reconfig] 0-workdata-client-2: changing port
>> > to 54105 (from 0)
>> > [2022-03-01 04:58:14.011469 +0000] I [socket.c:834:__socket_shutdown]
>> > 0-workdata-client-2: intentional socket shutdown(13)
>> >
>> > same for the other 4 bricks. Probably some more related message,
>> > unsure which ones to c+p. And some error messages like these (appear
>> > on all servers):
>> >
>> > [2022-03-01 04:58:14.012523 +0000] E
>> > [rpc-clnt.c:331:saved_frames_unwind] (-->
>> > /lib/x86_64-linux-gnu/libglusterfs.so.0(_gf_log_callingfn+0x195)[0x7f4cec48c2a5]
>> > (--> /lib/x86_64-linux-gnu/libgfrpc.so.0(+0x729c)[0x7f4cec42529c] (-->
>> > /lib/
>> > x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x10f)[0x7f4cec42d20f]
>> > (--> /lib/x86_64-linux-gnu/libgfrpc.so.0(+0x10118)[0x7f4cec42e118]
>> > (--> /lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_transport_notify+0x26)[0x7f4cec429646]
>> > )))
>> > )) 0-workdata-client-5: forced unwinding frame type(GF-DUMP)
>> > op(DUMP(1)) called at 2022-03-01 04:58:14.011943 +0000 (xid=0x5)
>> >
>> > very strange.
>> >
>> > 2) when mounting on the clients (after upgrade):
>> >
>> > /sbin/mount.glusterfs: 90: [: glusterpub2 glusterpub3 SyntaxOK:
>> > unexpected operator
>> > /sbin/mount.glusterfs: 366: [: SyntaxOK: unexpected operator
>> >
>> > Syntax ok, but unexpected operator? Has the mount syntax changed?
>> >
>> > glusterpub1:/workdata /data/repository/shared/public glusterfs
>> > defaults,_netdev,attribute-timeout=0,entry-timeout=0,backup-volfile-servers=glusterpub2:glusterpub3
>> > 0 0
>> >
>> >
>> > thx,
>> > Hubert
>> ________
>>
>>
>>
>> Community Meeting Calendar:
>>
>> Schedule -
>> Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
>> Bridge: https://meet.google.com/cpu-eiue-hvk
>> Gluster-users mailing list
>> Gluster-users@xxxxxxxxxxx
>> https://lists.gluster.org/mailman/listinfo/gluster-users
________



Community Meeting Calendar:

Schedule -
Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
Bridge: https://meet.google.com/cpu-eiue-hvk
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
https://lists.gluster.org/mailman/listinfo/gluster-users




[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux