Thank you Mauro. Rémi -----Message d'origine----- De : Mauro Mozzarelli [mailto:mauro@xxxxxxxxxxxx] Envoyé : vendredi 30 octobre 2015 12:58 À : Atin Mukherjee <atin.mukherjee83@xxxxxxxxx> Cc : Remi Serrano <rserrano@xxxxxxxx>; gluster-users@xxxxxxxxxxx Objet : Re: How-to start gluster when only one node is up ? Hi, Atin keeps giving the same answer: "it is by design" I keep saying "the design is wrong and it should be changed to cater for standby servers" In the meantime this is the workaround I am using: When the single node starts I stop and start the volume, and then it becomes mountable. On CentOS 6 and CentOS 7 it works with release up to 3.7.4. Release 3.7.5 is broken so I reverted back to 3.7.4. In my experience glusterfs releases are a bit of a hit and miss. Often something stops working with newer releases, then after a few more releases it works again or there is a workaround ... Not quite the stability one would want for commercial use, and thus at the moment I can risk using it only for my home servers, hence the cluster with a node always ON and the second as STANDBY. MOUNT=/home LABEL="GlusterFS:" if grep -qs $MOUNT /proc/mounts; then echo "$LABEL $MOUNT is mounted"; gluster volume start gv_home 2>/dev/null else echo "$LABEL $MOUNT is NOT mounted"; echo "$LABEL Restarting gluster volume ..." yes|gluster volume stop gv_home > /dev/null gluster volume start gv_home mount -t glusterfs sirius-ib:/gv_home $MOUNT; if grep -qs $MOUNT /proc/mounts; then echo "$LABEL $MOUNT is mounted"; gluster volume start gv_home 2>/dev/null else echo "$LABEL failure to mount $MOUNT"; fi fi I hope this helps. Mauro On Fri, October 30, 2015 11:48, Atin Mukherjee wrote: > -Atin > Sent from one plus one > On Oct 30, 2015 4:35 PM, "Remi Serrano" <rserrano@xxxxxxxx> wrote: >> >> Hello, >> >> >> >> I setup a gluster file cluster with 2 nodes. It works fine. >> >> But, when I shut down the 2 nodes, and startup only one node, I >> cannot > mount the share : >> >> >> >> [root@xxx ~]# mount -t glusterfs 10.32.0.11:/gv0 /glusterLocalShare >> >> Mount failed. Please check the log file for more details. >> >> >> >> Log says : >> >> [2015-10-30 10:33:26.147003] I [MSGID: 100030] >> [glusterfsd.c:2318:main] > 0-/usr/sbin/glusterfs: Started running /usr/sbin/glusterfs version > 3.7.5 > (args: /usr/sbin/glusterfs -127.0.0.1 --volfile-id=/gv0 > /glusterLocalShare) >> >> [2015-10-30 10:33:26.171964] I [MSGID: 101190] > [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started > thread with index 1 >> >> [2015-10-30 10:33:26.185685] I [MSGID: 101190] > [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started > thread with index 2 >> >> [2015-10-30 10:33:26.186972] I [MSGID: 114020] [client.c:2118:notify] > 0-gv0-client-0: parent translators are ready, attempting connect on > transport >> >> [2015-10-30 10:33:26.191823] I [MSGID: 114020] [client.c:2118:notify] > 0-gv0-client-1: parent translators are ready, attempting connect on > transport >> >> [2015-10-30 10:33:26.192209] E [MSGID: 114058] > [client-handshake.c:1524:client_query_portmap_cbk] 0-gv0-client-0: > failed to get the port number for remote subvolume. Please ume status' > on server to see if brick process is running. >> >> [2015-10-30 10:33:26.192339] I [MSGID: 114018] > [client.c:2042:client_rpc_notify] 0-gv0-client-0: disconnected from > gv0-client-0. Client process will keep trying to connect t brick's > port is available >> >> >> >> And when I check the volumes I get: >> >> [root@xxx ~]# gluster volume status >> >> Status of volume: gv0 >> >> Gluster process TCP Port RDMA Port Online > Pid >> >> > ---------------------------------------------------------------------- > -------- >> >> Brick 10.32.0.11:/glusterBrick1/gv0 N/A N/A N > N/A >> >> NFS Server on localhost N/A N/A N > N/A >> >> NFS Server on localhost N/A N/A N > N/A >> >> >> >> Task Status of Volume gv0 >> >> > ---------------------------------------------------------------------- > -------- >> >> There are no active volume tasks >> >> >> >> If I start th second node, all is OK. >> >> >> >> Is this normal ? > This behaviour is by design. In a multi node cluster when GlusterD > comes up it doesn't start the bricks until it receives the > configuration from its one of the friends to ensure that stale > information is not been referred. > In your case since the other node is down bricks are not started and > hence mount fails. > As a workaround, we recommend to add a dummy node to the cluster to > avoid this issue. >> >> >> >> Regards, >> >> >> >> Rémi >> >> >> >> >> _______________________________________________ >> Gluster-users mailing list >> Gluster-users@xxxxxxxxxxx >> http://cp.mcafee.com/d/1jWVIi4zqb33ytT6kn3tPqdQnC1Oab8USzt5Vwsyes76Qr >> ELfLcCzAQsCzASztBZxVOXUUT6DNzIawFqdqsKrmzmDbCZs8fIIKfZvC7zhPtMtRXBQSn >> zhPNEVsp7cLYJt6OaaJNOfaxVZicHs3jq9JcTvC4n3hOUehssodTdw0y9Xt9_7w0e2qKM >> M-l9OwXn29Xt9_4KTjUQdGWT6EaYkmfQ6Y3jrxEVupdwLQzh0qmMAuTivNbJQ-d3iWq80 >> LkMnDkQgkDIpCy0yuTivNd41wDY3h1m1VoQglxVEwxl3Ph1axEwrmd418SUedXoJO > _______________________________________________ > Gluster-users mailing list > Gluster-users@xxxxxxxxxxx > http://cp.mcafee.com/d/2DRPowcy0w76Qm674XKcEK6XCQrELc3AkmhNJ6WbP0V4sUe > dEThuvupd79EVd79J6XbX3PBTNNKdfz7ol1iQqQVsSJ6JendWUgvppsvW_cf6zCXwXHTbF > IL6zDzhOUOepvVqWdAklrzAul3PWApmU6CQPqpK_c8K6zBMsyUUMrKr014jSWj-f00s4Rt > xxYGjB1SK4jSWj-9tKDNErlRKdglUEIvEdU6CT3hOYOr1vF6y0QJx8ZKA_ynrFYq6BQQg1 > uFwLeFEwFfoPd414ZKA_yq831fU6y2I3ONEwH3Ph12G7Cy2l3h0SIq82hJMsrYI-l -- Mauro Mozzarelli Phone: +44 7941 727378 eMail: mauro@xxxxxxxxxxxx _______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-users