Hi Strahil, problem found: the driver for the network card was buggy or not working properly, which caused the network interface going up and down. I'll check if a systemd mount unit is an option. Thx! Best regards, Hubert Am Sa., 21. März 2020 um 21:06 Uhr schrieb Strahil Nikolov <hunter86_bg@xxxxxxxxx>: > > On March 20, 2020 8:13:28 AM GMT+02:00, Hu Bert <revirii@xxxxxxxxxxxxxx> wrote: > >Hello, > > > >i just reinstall a server (debian buster). I added 2 entries to > >/etc/fstab: > > > >gluster1:/persistent /data/repository/shared/private glusterfs > >defaults,_netdev,attribute-timeout=0,entry-timeout=0,backup-volfile-servers=gluster2:gluster3 > >0 0 > >gluster1:/workdata /data/repository/shared/public glusterfs > >defaults,_netdev,attribute-timeout=0,entry-timeout=0,backup-volfile-servers=gluster2:gluster3 > >0 0 > > > >Entries in /etc/hosts for gluster1+2+3 are there. But after a reboot > >the mount of the 2 gluster volumes fails. There are additional servers > >with exact the same entries, they don't have a problem with mounting. > >But only this server does. The log entries for one of the volumes > >shows this: > > > >[2020-03-20 05:32:13.089703] I [MSGID: 100030] > >[glusterfsd.c:2725:main] 0-/usr/sbin/glusterfs: Started running > >/usr/sbin/glusterfs version 5.11 (args: /usr/sbin/glusterfs > >--attribute-timeout=0 --entry-timeout=0 --process-name fuse > >--volfile-server=gluster1 --volfile-server=gluster2 > >--volfile-server=gluster3 --volfile-id=/persistent > >/data/repository/shared/private) > >[2020-03-20 05:32:13.120904] I [MSGID: 101190] > >[event-epoll.c:621:event_dispatch_epoll_worker] 0-epoll: Started > >thread with index 1 > >[2020-03-20 05:32:16.196568] I > >[glusterfsd-mgmt.c:2424:mgmt_rpc_notify] 0-glusterfsd-mgmt: > >disconnected from remote-host: gluster1 > >[2020-03-20 05:32:16.196614] I > >[glusterfsd-mgmt.c:2464:mgmt_rpc_notify] 0-glusterfsd-mgmt: connecting > >to next volfile server gluster2 > >[2020-03-20 05:32:20.164538] I > >[glusterfsd-mgmt.c:2464:mgmt_rpc_notify] 0-glusterfsd-mgmt: connecting > >to next volfile server gluster3 > >[2020-03-20 05:32:26.180546] I > >[glusterfsd-mgmt.c:2444:mgmt_rpc_notify] 0-glusterfsd-mgmt: Exhausted > >all volfile servers > >[2020-03-20 05:32:26.181618] W [glusterfsd.c:1500:cleanup_and_exit] > >(-->/lib/x86_64-linux-gnu/libgfrpc.so.0(+0xee13) [0x7fbf38a98e13] > >-->/usr/sbin/glusterfs(+0x127d7) [0x55bac75517d7] > >-->/usr/sbin/glusterfs(cleanup_and_exit+0x54) [0x55bac7549f54] ) 0-: > >received signum (1), shutting down > >[2020-03-20 05:32:26.181744] I [fuse-bridge.c:5914:fini] 0-fuse: > >Unmounting '/data/repository/shared/private'. > >[2020-03-20 05:32:26.200708] I [fuse-bridge.c:5919:fini] 0-fuse: > >Closing fuse connection to '/data/repository/shared/private'. > >[2020-03-20 05:32:26.200885] W [glusterfsd.c:1500:cleanup_and_exit] > >(-->/lib/x86_64-linux-gnu/libpthread.so.0(+0x7fa3) [0x7fbf38661fa3] > >-->/usr/sbin/glusterfs(glusterfs_sigwaiter+0xfd) [0x55bac754a0fd] > >-->/usr/sbin/glusterfs(cleanup_and_exit+0x54) [0x55bac7549f54] ) 0-: > >received signum (15), shutting down > > > >The messages for the other volume are identical. If i do a 'mount -a', > >the volumes get mounted. > > > >Did i miss anything? > > > > > >Regards, > >Hubert > >________ > > > > > > > >Community Meeting Calendar: > > > >Schedule - > >Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC > >Bridge: https://bluejeans.com/441850968 > > > >Gluster-users mailing list > >Gluster-users@xxxxxxxxxxx > >https://lists.gluster.org/mailman/listinfo/gluster-users > > Why don't you set it up as a systemd '.mount' unit ? > You can define dependencies. > > Currently, after a reboot you can check the following: > 'systemctl status persistent-data-repository-shared-private.mount' > > > Best Regards, > Strahil Nikolov ________ Community Meeting Calendar: Schedule - Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC Bridge: https://bluejeans.com/441850968 Gluster-users mailing list Gluster-users@xxxxxxxxxxx https://lists.gluster.org/mailman/listinfo/gluster-users