----- Original Message ----- From: "Atin Mukherjee" <atin.mukherjee83@xxxxxxxxx> To: "Rafi Kavungal Chundattu Parambil" <rkavunga@xxxxxxxxxx>, "Riccardo Murri" <riccardo.murri@xxxxxxxxx> Cc: gluster-users@xxxxxxxxxxx Sent: Wednesday, March 27, 2019 4:07:42 PM Subject: Re: cannot add server back to cluster after reinstallation On Wed, 27 Mar 2019 at 16:02, Riccardo Murri <riccardo.murri@xxxxxxxxx> wrote: > Hello Atin, > > > Check cluster.op-version, peer status, volume status output. If they are > all fine you’re good. > > Both `op-version` and `peer status` look fine: > ``` > # gluster volume get all cluster.max-op-version > Option Value > ------ ----- > cluster.max-op-version 31202 > > # gluster peer status > Number of Peers: 4 > > Hostname: glusterfs-server-004 > Uuid: 9a5763d2-1941-4e5d-8d33-8d6756f7f318 > State: Peer in Cluster (Connected) > > Hostname: glusterfs-server-005 > Uuid: d53398f6-19d4-4633-8bc3-e493dac41789 > State: Peer in Cluster (Connected) > > Hostname: glusterfs-server-003 > Uuid: 3c74d2b4-a4f3-42d4-9511-f6174b0a641d > State: Peer in Cluster (Connected) > > Hostname: glusterfs-server-001 > Uuid: 60bcc47e-ccbe-493e-b4ea-d45d63123977 > State: Peer in Cluster (Connected) > ``` > > However, `volume status` shows a missing snapshotd on the reinstalled > server (the 002 one). I believe you ran this command on 002? And in that case its showing as localhost. > We're not using snapshots so I guess this is fine too? Is features.uss enabled for this volume? Otherwise we don’t show snapd information in status output. Rafi - am I correct? Yes. We don't show snapd information unless uss is enabled. So please check whether uss is enabled or not. You can use gluster v get glusterfs features.uss . If you are not using any snapshot then it doesn't make sense to use uss. You can disable it using gluster v set glusterfs features.uss disable Please note that if you are doing the rolling upgrade, it is not recommended to do any configuration changes. In that case you can disable it after completing the upgrade. Rafi KC > > ``` > # gluster volume status > Status of volume: glusterfs > Gluster process TCP Port RDMA Port Online > Pid > > ------------------------------------------------------------------------------ > Brick glusterfs-server-005:/s > rv/glusterfs 49152 0 Y > 1410 > Brick glusterfs-server-004:/s > rv/glusterfs 49152 0 Y > 1416 > Brick glusterfs-server-003:/s > rv/glusterfs 49152 0 Y > 1520 > Brick glusterfs-server-001:/s > rv/glusterfs 49152 0 Y > 1266 > Brick glusterfs-server-002:/s > rv/glusterfs 49152 0 Y > 3011 > Snapshot Daemon on localhost N/A N/A Y > 3029 > Snapshot Daemon on glusterfs- > server-001 49153 0 Y > 1361 > Snapshot Daemon on glusterfs- > server-005 49153 0 Y > 1478 > Snapshot Daemon on glusterfs- > server-004 49153 0 Y > 1490 > Snapshot Daemon on glusterfs- > server-003 49153 0 Y > 1563 > > Task Status of Volume glusterfs > > ------------------------------------------------------------------------------ > Task : Rebalance > ID : 0eaf6ad1-df95-48f4-b941-17488010ddcc > Status : failed > ``` > > Thanks, > Riccardo > -- --Atin _______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx https://lists.gluster.org/mailman/listinfo/gluster-users