Working 172.17.2.30, [root@centclust2 glusterfs]# telnet 172.17.2.30 49152 Trying 172.17.2.30... Connected to 172.17.2.30. Escape character is '^]'. Working 192.168.135.36, [root@centclust1 ssl]# telnet 192.168.135.36 49152 Trying 192.168.135.36... Connected to centclust1 (192.168.135.36). Escape character is '^]'. Working 172.17.2.31, [root@centclust1 ~]# telnet 172.17.2.31 49152 Trying 172.17.2.31... Connected to centclust2 (172.17.2.31). Escape character is '^]'. Working 192.168.135.46, [root@centclust1 ~]# telnet 192.168.135.46 49152 Trying 192.168.135.46... Connected to centclust2 (192.168.135.46). Escape character is '^]'. This did work even before! Greetings Daniel EDV Daniel Müller Leitung EDV Tropenklinik Paul-Lechler-Krankenhaus Paul-Lechler-Str. 24 72076 Tübingen Tel.: 07071/206-463, Fax: 07071/206-499 eMail: mueller@xxxxxxxxxxxxxxx Internet: www.tropenklinik.de -----Ursprüngliche Nachricht----- Von: Vijaikumar M [mailto:vmallika@xxxxxxxxxx] Gesendet: Donnerstag, 31. Juli 2014 13:33 An: mueller@xxxxxxxxxxxxxxx Cc: 'Krishnan Parthasarathi'; gluster-devel-bounces@xxxxxxxxxxx; gluster-users@xxxxxxxxxxx Betreff: Re: WG: Strange issu concerning glusterfs 3.5.1 on centos 6.5 Hi Daniel, Check if telnet works on brick port from both the interface. telnet 172.17.2.30 <brick-port1> telnet 192.168.135.36 <brick-port1> telnet 172.17.2.31 <brick-port2> telnet 192.168.135.46 <brick-port2> Thanks, Vijay On Thursday 31 July 2014 04:37 PM, Daniel Müller wrote: > So, > > [root@centclust1 ~]# ifconfig > eth0 Link encap:Ethernet Hardware Adresse 00:25:90:80:D9:E8 > inet Adresse:172.17.2.30 Bcast:172.17.2.255 Maske:255.255.255.0 > inet6 Adresse: fe80::225:90ff:fe80:d9e8/64 Gültigkeitsbereich:Verbindung > UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 > RX packets:3506528 errors:0 dropped:0 overruns:0 frame:0 > TX packets:169905 errors:0 dropped:0 overruns:0 carrier:0 > Kollisionen:0 Sendewarteschlangenlänge:1000 > RX bytes:476128477 (454.0 MiB) TX bytes:18788266 (17.9 MiB) > Speicher:fe860000-fe880000 > > eth1 Link encap:Ethernet Hardware Adresse 00:25:90:80:D9:E9 > inet Adresse:192.168.135.36 Bcast:192.168.135.255 Maske:255.255.255.0 > inet6 Adresse: fe80::225:90ff:fe80:d9e9/64 Gültigkeitsbereich:Verbindung > UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 > RX packets:381664693 errors:0 dropped:0 overruns:0 frame:0 > TX packets:380924973 errors:0 dropped:0 overruns:0 carrier:0 > Kollisionen:0 Sendewarteschlangenlänge:1000 > RX bytes:477454156923 (444.6 GiB) TX bytes:476729269342 (443.9 GiB) > Speicher:fe8e0000-fe900000 > > lo Link encap:Lokale Schleife > inet Adresse:127.0.0.1 Maske:255.0.0.0 > inet6 Adresse: ::1/128 Gültigkeitsbereich:Maschine > UP LOOPBACK RUNNING MTU:16436 Metric:1 > RX packets:93922879 errors:0 dropped:0 overruns:0 frame:0 > TX packets:93922879 errors:0 dropped:0 overruns:0 carrier:0 > Kollisionen:0 Sendewarteschlangenlänge:0 > RX bytes:462579764180 (430.8 GiB) TX bytes:462579764180 > (430.8 GiB) > > > [root@centclust2 ~]# ifconfig > eth0 Link encap:Ethernet Hardware Adresse 00:25:90:80:EF:00 > inet Adresse:172.17.2.31 Bcast:172.17.2.255 Maske:255.255.255.0 > inet6 Adresse: fe80::225:90ff:fe80:ef00/64 Gültigkeitsbereich:Verbindung > UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 > RX packets:1383117 errors:0 dropped:0 overruns:0 frame:0 > TX packets:45828 errors:0 dropped:0 overruns:0 carrier:0 > Kollisionen:0 Sendewarteschlangenlänge:1000 > RX bytes:185634714 (177.0 MiB) TX bytes:5357926 (5.1 MiB) > Speicher:fe860000-fe880000 > > eth1 Link encap:Ethernet Hardware Adresse 00:25:90:80:EF:01 > inet Adresse:192.168.135.46 Bcast:192.168.135.255 Maske:255.255.255.0 > inet6 Adresse: fe80::225:90ff:fe80:ef01/64 Gültigkeitsbereich:Verbindung > UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 > RX packets:340364283 errors:0 dropped:0 overruns:0 frame:0 > TX packets:59930672 errors:0 dropped:0 overruns:0 carrier:0 > Kollisionen:0 Sendewarteschlangenlänge:1000 > RX bytes:473823738544 (441.2 GiB) TX bytes:9973035418 (9.2 GiB) > Speicher:fe8e0000-fe900000 > > lo Link encap:Lokale Schleife > inet Adresse:127.0.0.1 Maske:255.0.0.0 > inet6 Adresse: ::1/128 Gültigkeitsbereich:Maschine > UP LOOPBACK RUNNING MTU:16436 Metric:1 > RX packets:1102979 errors:0 dropped:0 overruns:0 frame:0 > TX packets:1102979 errors:0 dropped:0 overruns:0 carrier:0 > Kollisionen:0 Sendewarteschlangenlänge:0 > RX bytes:126066547 (120.2 MiB) TX bytes:126066547 (120.2 > MiB) > > > [root@centclust1 ~]# route > Kernel IP Routentabelle > Ziel Router Genmask Flags Metric Ref Use Iface > 192.168.135.0 * 255.255.255.0 U 1 0 0 eth1 > 172.17.2.0 * 255.255.255.0 U 1 0 0 eth0 > default s4master 0.0.0.0 UG 0 0 0 eth1 > > > [root@centclust2 ~]# route > Kernel IP Routentabelle > Ziel Router Genmask Flags Metric Ref Use Iface > 192.168.135.0 * 255.255.255.0 U 0 0 0 eth1 > 172.17.2.0 * 255.255.255.0 U 0 0 0 eth0 > link-local * 255.255.0.0 U 1002 0 0 eth0 > link-local * 255.255.0.0 U 1003 0 0 eth1 > default s4master 0.0.0.0 UG 0 0 0 eth1 > > [root@centclust1 ~]# route -n > Kernel IP Routentabelle > Ziel Router Genmask Flags Metric Ref Use Iface > 192.168.135.0 0.0.0.0 255.255.255.0 U 1 0 0 eth1 > 172.17.2.0 0.0.0.0 255.255.255.0 U 1 0 0 eth0 > 0.0.0.0 192.168.135.230 0.0.0.0 UG 0 0 0 eth1 > > [root@centclust2 ~]# route -n > Kernel IP Routentabelle > Ziel Router Genmask Flags Metric Ref Use Iface > 192.168.135.0 0.0.0.0 255.255.255.0 U 0 0 0 eth1 > 172.17.2.0 0.0.0.0 255.255.255.0 U 0 0 0 eth0 > 169.254.0.0 0.0.0.0 255.255.0.0 U 1002 0 0 eth0 > 169.254.0.0 0.0.0.0 255.255.0.0 U 1003 0 0 eth1 > 0.0.0.0 192.168.135.230 0.0.0.0 UG 0 0 0 eth1 > > > > EDV Daniel Müller > > Leitung EDV > Tropenklinik Paul-Lechler-Krankenhaus > Paul-Lechler-Str. 24 > 72076 Tübingen > Tel.: 07071/206-463, Fax: 07071/206-499 > eMail: mueller@xxxxxxxxxxxxxxx > Internet: www.tropenklinik.de > > > > > > -----Ursprüngliche Nachricht----- > Von: Krishnan Parthasarathi [mailto:kparthas@xxxxxxxxxx] > Gesendet: Donnerstag, 31. Juli 2014 12:55 > An: mueller@xxxxxxxxxxxxxxx > Cc: gluster-devel-bounces@xxxxxxxxxxx; gluster-users@xxxxxxxxxxx > Betreff: Re: WG: Strange issu concerning glusterfs > 3.5.1 on centos 6.5 > > Daniel, > > Could you provide the following details from your original two NIC setup probed using hostname? > 1) output of ifconfig of the two NICs on both the nodes . > 2) output of route from both the nodes. > > ~KP > ----- Original Message ----- >> Hello and thank you so far, >> What I have recognized is, having more than one nic running this is >> confusing glusterfs 3.5. I never saw this on my glusterfs 3.4 and 3.2 >> systems still working. >> So I set up just clean erased gluster with yum glusterfs* erase and did: >> Logged in to my both nodes in the 135 subnet,ex: >> Ssh 192.168.135.36 (centclust1) (172.17.2.30 is the 2nd nic) Ssh >> 192.168.135.46 (centclust2) (172.17.2.31 is the 2nd nic) Started >> gluster on both nodes , service glusterd start. >> Did the peer probe on 192.168.135.36/centclust1: >> Gluster peer probe 192.168.135.46 //Former I did gluster peer probe >> centclust2 >> This result in: >> [root@centclust1 ~]# gluster peer status Number of Peers: 1 >> >> Hostname: 192.168.135.46 >> Uuid: c395c15d-5187-4e5b-b680-57afcb88b881 >> State: Peer in Cluster (Connected) >> >> [root@centclust2 backup]# gluster peer status Number of Peers: 1 >> >> Hostname: 192.168.135.36 >> Uuid: 94d5903b-ebe9-40d6-93bf-c2f2e92909a0 >> State: Peer in Cluster (Connected) >> The signifent difference gluster now shows the ip of both nodes >> >> Now I did the create the replicating vol: >> gluster volume create smbcluster replica 2 transport tcp >> 192.168.135.36:/sbu/glusterfs/export >> 192.168.135.46:/sbu/glusterfs/export >> started the volume >> gluster volume status >> Status of volume: smbcluster >> Gluster process Port Online Pid >> ------------------------------------------------------------------------------ >> Brick 192.168.135.36:/sbu/glusterfs/export 49152 Y 27421 >> Brick 192.168.135.46:/sbu/glusterfs/export 49152 Y 12186 >> NFS Server on localhost 2049 Y 27435 >> Self-heal Daemon on localhost N/A Y 27439 >> NFS Server on 192.168.135.46 2049 Y 12200 >> Self-heal Daemon on 192.168.135.46 N/A Y 12204 >> >> Task Status of Volume smbcluster >> --------------------------------------------------------------------- >> - >> -------- >> There are no active volume tasks >> >> Mounted the volumes: >> >> Centclust1:mount -t glusterfs 192.168.135.36:/smbcluster /mntgluster >> -o acl Centclust2:mount -t glusterfs 192.168.135.46:/smbcluster >> /mntgluster -o acl >> >> And BINGO up and running!!!!!!! >> >> >> EDV Daniel Müller >> >> Leitung EDV >> Tropenklinik Paul-Lechler-Krankenhaus Paul-Lechler-Str. 24 >> 72076 Tübingen >> Tel.: 07071/206-463, Fax: 07071/206-499 >> eMail: mueller@xxxxxxxxxxxxxxx >> Internet: www.tropenklinik.de >> >> >> >> >> -----Ursprüngliche Nachricht----- >> Von: Krishnan Parthasarathi [mailto:kparthas@xxxxxxxxxx] >> Gesendet: Mittwoch, 30. Juli 2014 16:52 >> An: mueller@xxxxxxxxxxxxxxx >> Cc: gluster-devel-bounces@xxxxxxxxxxx; gluster-users@xxxxxxxxxxx >> Betreff: Re: WG: Strange issu concerning glusterfs >> 3.5.1 on centos 6.5 >> >> Daniel, >> >> I didn't get a chance to follow up with debugging this issue. I will >> look into this and get back to you. I suspect that there is something >> different about the network layer behaviour in your setup. >> >> ~KP >> >> ----- Original Message ----- >>> Just another other test: >>> [root@centclust1 sicherung]# getfattr -d -e hex -m . /sicherung/bu >>> getfattr: Entferne führenden '/' von absoluten Pfadnamen # file: >>> sicherung/bu >>> security.selinux=0x756e636f6e66696e65645f753a6f626a6563745f723a66696 >>> c6 >>> 55f743a733000 >>> trusted.afr.smbbackup-client-0=0x000000000000000000000000 >>> trusted.afr.smbbackup-client-1=0x000000000000000200000001 >>> trusted.gfid=0x00000000000000000000000000000001 >>> trusted.glusterfs.dht=0x000000010000000000000000ffffffff >>> trusted.glusterfs.volume-id=0x6f51d002e634437db58d9b952693f1df >>> >>> [root@centclust2 glusterfs]# getfattr -d -e hex -m . /sicherung/bu >>> getfattr: Entferne führenden '/' von absoluten Pfadnamen # file: >>> sicherung/bu >>> security.selinux=0x756e636f6e66696e65645f753a6f626a6563745f723a66696 >>> c6 >>> 55f743a733000 >>> trusted.afr.smbbackup-client-0=0x000000000000000200000001 >>> trusted.afr.smbbackup-client-1=0x000000000000000000000000 >>> trusted.gfid=0x00000000000000000000000000000001 >>> trusted.glusterfs.dht=0x000000010000000000000000ffffffff >>> trusted.glusterfs.volume-id=0x6f51d002e634437db58d9b952693f1df >>> >>> Is this ok? >>> >>> After long testing and doing a /etc/init.d/network restart the >>> replication started once/a short time then ended up!? >>> Any idea??????? >>> >>> >>> EDV Daniel Müller >>> >>> Leitung EDV >>> Tropenklinik Paul-Lechler-Krankenhaus Paul-Lechler-Str. 24 >>> 72076 Tübingen >>> Tel.: 07071/206-463, Fax: 07071/206-499 >>> eMail: mueller@xxxxxxxxxxxxxxx >>> Internet: www.tropenklinik.de >>> >>> "Der Mensch ist die Medizin des Menschen" >>> >>> >>> >>> >>> -----Ursprüngliche Nachricht----- >>> Von: Krishnan Parthasarathi [mailto:kparthas@xxxxxxxxxx] >>> Gesendet: Mittwoch, 30. Juli 2014 11:09 >>> An: mueller@xxxxxxxxxxxxxxx >>> Cc: gluster-devel-bounces@xxxxxxxxxxx; gluster-users@xxxxxxxxxxx >>> Betreff: Re: WG: Strange issu concerning glusterfs >>> 3.5.1 on centos 6.5 >>> >>> Could you provide the output of the following command? >>> >>> netstat -ntap | grep gluster >>> >>> This should tell us if glusterfsd processes (bricks) are listening >>> on all interfaces. >>> >>> ~KP >>> >>> ----- Original Message ----- >>>> Just one idea >>>> I add a second NIC with a 172.2.17... adress on both machines. >>>> Could this cause the trouble!? >>>> >>>> EDV Daniel Müller >>>> >>>> Leitung EDV >>>> Tropenklinik Paul-Lechler-Krankenhaus Paul-Lechler-Str. 24 >>>> 72076 Tübingen >>>> Tel.: 07071/206-463, Fax: 07071/206-499 >>>> eMail: mueller@xxxxxxxxxxxxxxx >>>> Internet: www.tropenklinik.de >>>> >>>> >>>> >>>> >>>> -----Ursprüngliche Nachricht----- >>>> Von: Krishnan Parthasarathi [mailto:kparthas@xxxxxxxxxx] >>>> Gesendet: Mittwoch, 30. Juli 2014 09:29 >>>> An: mueller@xxxxxxxxxxxxxxx >>>> Cc: gluster-devel-bounces@xxxxxxxxxxx; gluster-users@xxxxxxxxxxx >>>> Betreff: Re: WG: Strange issu concerning glusterfs >>>> 3.5.1 on centos 6.5 >>>> >>>> Daniel, >>>> >>>> From a quick look, I see that glustershd and the nfs client is >>>> unable to connect to one of the bricks. This is resulting in data >>>> from mounts being written to local bricks only. >>>> I should have asked this before, could you provide the bricks logs >>>> as well? >>>> >>>> Could you also try to connect to the bricks using telnet? >>>> For eg, from centclust1, telnet centclust2 <brick-port>. >>>> >>>> ~KP >>>> >>>> ----- Original Message ----- >>>>> So my logs. I disable ssl meanwhile but it is the same situation. >>>>> No replication!? >>>>> >>>>> >>>>> >>>>> EDV Daniel Müller >>>>> >>>>> Leitung EDV >>>>> Tropenklinik Paul-Lechler-Krankenhaus Paul-Lechler-Str. 24 >>>>> 72076 Tübingen >>>>> Tel.: 07071/206-463, Fax: 07071/206-499 >>>>> eMail: mueller@xxxxxxxxxxxxxxx >>>>> Internet: www.tropenklinik.de >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> -----Ursprüngliche Nachricht----- >>>>> Von: Krishnan Parthasarathi [mailto:kparthas@xxxxxxxxxx] >>>>> Gesendet: Mittwoch, 30. Juli 2014 08:56 >>>>> An: mueller@xxxxxxxxxxxxxxx >>>>> Cc: gluster-users@xxxxxxxxxxx; gluster-devel-bounces@xxxxxxxxxxx >>>>> Betreff: Re: WG: Strange issu concerning >>>>> glusterfs >>>>> 3.5.1 on centos 6.5 >>>>> >>>>> Could you attach the entire mount and glustershd log files to >>>>> this thread? >>>>> >>>>> ~KP >>>>> >>>>> ----- Original Message ----- >>>>>> NO ONE!?? >>>>>> This is an entry of my glustershd.log: >>>>>> [2014-07-30 06:40:59.294334] W >>>>>> [client-handshake.c:1846:client_dump_version_cbk] >>>>>> 0-smbbackup-client-1: >>>>>> received RPC status error >>>>>> [2014-07-30 06:40:59.294352] I >>>>>> [client.c:2229:client_rpc_notify] >>>>>> 0-smbbackup-client-1: disconnected from 172.17.2.31:49152. >>>>>> Client process will keep trying to connect to glusterd until >>>>>> brick's port is available >>>>>> >>>>>> >>>>>> This is from mnt-sicherung.log: >>>>>> [2014-07-30 06:40:38.259850] E [socket.c:2820:socket_connect] >>>>>> 1-smbbackup-client-0: connection attempt on 172.17.2.30:24007 >>>>>> failed, (Connection timed out) [2014-07-30 06:40:41.275120] I >>>>>> [rpc-clnt.c:1729:rpc_clnt_reconfig] >>>>>> 1-smbbackup-client-0: changing port to 49152 (from 0) >>>>>> >>>>>> [root@centclust1 sicherung]# gluster --remote-host=centclust1 >>>>>> peer status Number of Peers: 1 >>>>>> >>>>>> Hostname: centclust2 >>>>>> Uuid: 4f15e9bd-9b5a-435b-83d2-4ed202c66b11 >>>>>> State: Peer in Cluster (Connected) >>>>>> >>>>>> [root@centclust1 sicherung]# gluster --remote-host=centclust2 >>>>>> peer status Number of Peers: 1 >>>>>> >>>>>> Hostname: 172.17.2.30 >>>>>> Uuid: 99fe6a2c-df7e-4475-a7bc-a35abba620fb >>>>>> State: Peer in Cluster (Connected) >>>>>> >>>>>> [root@centclust1 ssl]# ps aux | grep gluster >>>>>> root 13655 0.0 0.0 413848 16872 ? Ssl 08:10 0:00 >>>>>> /usr/sbin/glusterd --pid-file=/var/run/glusterd.pid >>>>>> root 13958 0.0 0.0 12139920 44812 ? Ssl 08:11 0:00 >>>>>> /usr/sbin/glusterfsd -s centclust1.tplk.loc --volfile-id >>>>>> smbbackup.centclust1.tplk.loc.sicherung-bu -p >>>>>> /var/lib/glusterd/vols/smbbackup/run/centclust1.tplk.loc-sicherung-bu. >>>>>> pid -S /var/run/4c65260e12e2d3a9a5549446f491f383.socket >>>>>> --brick-name /sicherung/bu -l >>>>>> /var/log/glusterfs/bricks/sicherung-bu.log >>>>>> --xlator-option >>>>>> *-posix.glusterd-uuid=99fe6a2c-df7e-4475-a7bc-a35abba620fb >>>>>> --brick-port >>>>>> 49152 --xlator-option smbbackup-server.listen-port=49152 >>>>>> root 13972 0.0 0.0 815748 58252 ? Ssl 08:11 0:00 >>>>>> /usr/sbin/glusterfs -s localhost --volfile-id gluster/nfs -p >>>>>> /var/lib/glusterd/nfs/run/nfs.pid -l >>>>>> /var/log/glusterfs/nfs.log -S /var/run/ee6f37fc79b9cb1968eca387930b39fb.socket >>>>>> root 13976 0.0 0.0 831160 29492 ? Ssl 08:11 0:00 >>>>>> /usr/sbin/glusterfs -s localhost --volfile-id >>>>>> gluster/glustershd -p >>>>>> /var/lib/glusterd/glustershd/run/glustershd.pid -l >>>>>> /var/log/glusterfs/glustershd.log -S >>>>>> /var/run/aa970d146eb23ba7124e6c4511879850.socket --xlator-option *replicate*.node-uuid=99fe6a2c-df7e-4475-a7bc-a35abba620fb >>>>>> root 15781 0.0 0.0 105308 932 pts/1 S+ 08:47 0:00 grep >>>>>> gluster >>>>>> root 29283 0.0 0.0 451116 56812 ? Ssl Jul29 0:21 >>>>>> /usr/sbin/glusterfs -s localhost --volfile-id gluster/nfs -p >>>>>> /var/lib/glusterd/nfs/run/nfs.pid -l >>>>>> /var/log/glusterfs/nfs.log -S /var/run/a7fcb1d1d3a769d28df80b85ae5d13c4.socket >>>>>> root 29287 0.0 0.0 335432 25848 ? Ssl Jul29 0:21 >>>>>> /usr/sbin/glusterfs -s localhost --volfile-id >>>>>> gluster/glustershd -p >>>>>> /var/lib/glusterd/glustershd/run/glustershd.pid -l >>>>>> /var/log/glusterfs/glustershd.log -S >>>>>> /var/run/833e60f976365c2a307f92fb233942a2.socket --xlator-option *replicate*.node-uuid=64b1a7eb-2df3-47bd-9379-39c29e5a001a >>>>>> root 31698 0.0 0.0 1438392 57952 ? Ssl Jul29 0:12 >>>>>> /usr/sbin/glusterfs --acl --volfile-server=centclust1.tplk.loc >>>>>> --volfile-id=/smbbackup /mnt/sicherung >>>>>> >>>>>> [root@centclust2 glusterfs]# ps aux | grep gluster >>>>>> root 1561 0.0 0.0 1481492 60152 ? Ssl Jul29 0:12 >>>>>> /usr/sbin/glusterfs --acl --volfile-server=centclust2.tplk.loc >>>>>> --volfile-id=/smbbackup /mnt/sicherung >>>>>> root 15656 0.0 0.0 413848 16832 ? Ssl 08:11 0:01 >>>>>> /usr/sbin/glusterd --pid-file=/var/run/glusterd.pid >>>>>> root 15942 0.0 0.0 12508704 43860 ? Ssl 08:11 0:00 >>>>>> /usr/sbin/glusterfsd -s centclust2.tplk.loc --volfile-id >>>>>> smbbackup.centclust2.tplk.loc.sicherung-bu -p >>>>>> /var/lib/glusterd/vols/smbbackup/run/centclust2.tplk.loc-sicherung-bu. >>>>>> pid -S /var/run/40a554af3860eddd5794b524576d0520.socket >>>>>> --brick-name /sicherung/bu -l >>>>>> /var/log/glusterfs/bricks/sicherung-bu.log >>>>>> --xlator-option >>>>>> *-posix.glusterd-uuid=4f15e9bd-9b5a-435b-83d2-4ed202c66b11 >>>>>> --brick-port >>>>>> 49152 --xlator-option smbbackup-server.listen-port=49152 >>>>>> root 15956 0.0 0.0 825992 57496 ? Ssl 08:11 0:00 >>>>>> /usr/sbin/glusterfs -s localhost --volfile-id gluster/nfs -p >>>>>> /var/lib/glusterd/nfs/run/nfs.pid -l >>>>>> /var/log/glusterfs/nfs.log -S /var/run/602d1d8ba7b80ded2b70305ed7417cf5.socket >>>>>> root 15960 0.0 0.0 841404 26760 ? Ssl 08:11 0:00 >>>>>> /usr/sbin/glusterfs -s localhost --volfile-id >>>>>> gluster/glustershd -p >>>>>> /var/lib/glusterd/glustershd/run/glustershd.pid -l >>>>>> /var/log/glusterfs/glustershd.log -S >>>>>> /var/run/504d01c7f7df8b8306951cc2aaeaf52c.socket >>>>>> --xlator-option >>>>>> *replicate*.node-uuid=4f15e9bd-9b5a-435b-83d2-4ed202c66b11 >>>>>> root 17728 0.0 0.0 105312 936 pts/0 S+ 08:48 0:00 grep >>>>>> gluster >>>>>> root 32363 0.0 0.0 451100 55584 ? Ssl Jul29 0:21 >>>>>> /usr/sbin/glusterfs -s localhost --volfile-id gluster/nfs -p >>>>>> /var/lib/glusterd/nfs/run/nfs.pid -l >>>>>> /var/log/glusterfs/nfs.log -S /var/run/73054288d1cadfb87b4b9827bd205c7b.socket >>>>>> root 32370 0.0 0.0 335432 26220 ? Ssl Jul29 0:21 >>>>>> /usr/sbin/glusterfs -s localhost --volfile-id >>>>>> gluster/glustershd -p >>>>>> /var/lib/glusterd/glustershd/run/glustershd.pid -l >>>>>> /var/log/glusterfs/glustershd.log -S >>>>>> /var/run/de1427ce373c792c76c38b12c106f029.socket >>>>>> --xlator-option >>>>>> *replicate*.node-uuid=83e6d78c-0119-4537-8922-b3e731718864 >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> Leitung EDV >>>>>> Tropenklinik Paul-Lechler-Krankenhaus Paul-Lechler-Str. 24 >>>>>> 72076 Tübingen >>>>>> Tel.: 07071/206-463, Fax: 07071/206-499 >>>>>> eMail: mueller@xxxxxxxxxxxxxxx >>>>>> Internet: www.tropenklinik.de >>>>>> >>>>>> >>>>>> >>>>>> -----Ursprüngliche Nachricht----- >>>>>> Von: Daniel Müller [mailto:mueller@xxxxxxxxxxxxxxx] >>>>>> Gesendet: Dienstag, 29. Juli 2014 16:02 >>>>>> An: 'gluster-users@xxxxxxxxxxx' >>>>>> Betreff: Strange issu concerning glusterfs 3.5.1 on centos 6.5 >>>>>> >>>>>> Dear all, >>>>>> >>>>>> there is a strange issue centos6.5 and glusterfs 3.5.1: >>>>>> >>>>>> glusterd -V >>>>>> glusterfs 3.5.1 built on Jun 24 2014 15:09:41 Repository revision: >>>>>> git://git.gluster.com/glusterfs.git >>>>>> Copyright (c) 2006-2013 Red Hat, Inc. <http://www.redhat.com/> >>>>>> GlusterFS comes with ABSOLUTELY NO WARRANTY. >>>>>> It is licensed to you under your choice of the GNU Lesser >>>>>> General Public License, version 3 or any later version (LGPLv3 >>>>>> or later), or the GNU General Public License, version 2 >>>>>> (GPLv2), in all cases as published by the Free Software >>>>>> Foundation >>>>>> >>>>>> I try to set up a replicated 2 brick vol on two centos 6.5 server. >>>>>> I can probe well and my nodes are reporting no errors: >>>>>> >>>>>> [root@centclust1 mnt]# gluster peer status Number of Peers: 1 >>>>>> >>>>>> Hostname: centclust2 >>>>>> Uuid: 4f15e9bd-9b5a-435b-83d2-4ed202c66b11 >>>>>> State: Peer in Cluster (Connected) >>>>>> >>>>>> [root@centclust2 sicherung]# gluster peer status Number of Peers: >>>>>> 1 >>>>>> >>>>>> Hostname: 172.17.2.30 >>>>>> Uuid: 99fe6a2c-df7e-4475-a7bc-a35abba620fb >>>>>> State: Peer in Cluster (Connected) >>>>>> >>>>>> Now I set up a replicating VOl on an XFS-Disk: /dev/sdb1 on >>>>>> /sicherung type xfs (rw) >>>>>> >>>>>> gluster volume create smbbackup replica 2 transport tcp >>>>>> centclust1.tplk.loc:/sicherung/bu >>>>>> centclust2.tplk.loc:/sicherung/bu >>>>>> >>>>>> gluster volume smbbackup status reports ok: >>>>>> >>>>>> [root@centclust1 mnt]# gluster volume status smbbackup Status >>>>>> of >>>>>> volume: smbbackup >>>>>> Gluster process Port >>>>>> Online >>>>>> Pid >>>>>> -------------------------------------------------------------- >>>>>> -- >>>>>> -- >>>>>> -- >>>>>> -- >>>>>> ------ >>>>>> -- >>>>>> Brick centclust1.tplk.loc:/sicherung/bu 49152 Y >>>>>> 31969 >>>>>> Brick centclust2.tplk.loc:/sicherung/bu 49152 Y >>>>>> 2124 >>>>>> NFS Server on localhost 2049 Y >>>>>> 31983 >>>>>> Self-heal Daemon on localhost N/A Y >>>>>> 31987 >>>>>> NFS Server on centclust2 2049 Y >>>>>> 2138 >>>>>> Self-heal Daemon on centclust2 N/A Y >>>>>> 2142 >>>>>> >>>>>> Task Status of Volume smbbackup >>>>>> -------------------------------------------------------------- >>>>>> -- >>>>>> -- >>>>>> -- >>>>>> -- >>>>>> ------ >>>>>> -- >>>>>> There are no active volume tasks >>>>>> >>>>>> [root@centclust2 sicherung]# gluster volume status smbbackup >>>>>> Status of >>>>>> volume: smbbackup >>>>>> Gluster process Port >>>>>> Online >>>>>> Pid >>>>>> -------------------------------------------------------------- >>>>>> -- >>>>>> -- >>>>>> -- >>>>>> -- >>>>>> ------ >>>>>> -- >>>>>> Brick centclust1.tplk.loc:/sicherung/bu 49152 Y >>>>>> 31969 >>>>>> Brick centclust2.tplk.loc:/sicherung/bu 49152 Y >>>>>> 2124 >>>>>> NFS Server on localhost 2049 Y >>>>>> 2138 >>>>>> Self-heal Daemon on localhost N/A Y >>>>>> 2142 >>>>>> NFS Server on 172.17.2.30 2049 Y >>>>>> 31983 >>>>>> Self-heal Daemon on 172.17.2.30 N/A Y >>>>>> 31987 >>>>>> >>>>>> Task Status of Volume smbbackup >>>>>> -------------------------------------------------------------- >>>>>> -- >>>>>> -- >>>>>> -- >>>>>> -- >>>>>> ------ >>>>>> -- >>>>>> There are no active volume tasks >>>>>> >>>>>> I mounted the vol on both servers with: >>>>>> >>>>>> mount -t glusterfs centclust1.tplk.loc:/smbbackup >>>>>> /mnt/sicherung -o acl mount -t glusterfs >>>>>> centclust2.tplk.loc:/smbbackup /mnt/sicherung -o acl >>>>>> >>>>>> But when I write in /mnt/sicherung the files are not >>>>>> replicated to the other node in anyway!?? >>>>>> >>>>>> They rest on the local server in /mnt/sicherung and >>>>>> /sicherung/bu On each node separate:# >>>>>> [root@centclust1 sicherung]# pwd /mnt/sicherung >>>>>> >>>>>> [root@centclust1 sicherung]# touch test.txt >>>>>> [root@centclust1 sicherung]# ls test.txt >>>>>> [root@centclust2 sicherung]# pwd /mnt/sicherung >>>>>> [root@centclust2 sicherung]# ls more.txt >>>>>> [root@centclust1 sicherung]# ls -la /sicherung/bu insgesamt 0 >>>>>> drwxr-xr-x. 3 root root 38 29. Jul 15:56 . >>>>>> drwxr-xr-x. 3 root root 15 29. Jul 14:31 .. >>>>>> drw-------. 15 root root 142 29. Jul 15:56 .glusterfs >>>>>> -rw-r--r--. 2 root root 0 29. Jul 15:56 test.txt >>>>>> [root@centclust2 sicherung]# ls -la /sicherung/bu insgesamt 0 >>>>>> drwxr-xr-x. 3 root root 38 29. Jul 15:32 . >>>>>> drwxr-xr-x. 3 root root 15 29. Jul 14:31 .. >>>>>> drw-------. 7 root root 70 29. Jul 15:32 .glusterfs -rw-r--r--. >>>>>> 2 root root 0 29. Jul 15:32 more.txt >>>>>> >>>>>> >>>>>> >>>>>> Greetings >>>>>> Daniel >>>>>> >>>>>> >>>>>> >>>>>> EDV Daniel Müller >>>>>> >>>>>> Leitung EDV >>>>>> Tropenklinik Paul-Lechler-Krankenhaus Paul-Lechler-Str. 24 >>>>>> 72076 Tübingen >>>>>> Tel.: 07071/206-463, Fax: 07071/206-499 >>>>>> eMail: mueller@xxxxxxxxxxxxxxx >>>>>> Internet: www.tropenklinik.de >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> _______________________________________________ >>>>>> Gluster-users mailing list >>>>>> Gluster-users@xxxxxxxxxxx >>>>>> http://supercolony.gluster.org/mailman/listinfo/gluster-users >>>>>> >>>>> _______________________________________________ >>>>> Gluster-users mailing list >>>>> Gluster-users@xxxxxxxxxxx >>>>> http://supercolony.gluster.org/mailman/listinfo/gluster-users >>>> _______________________________________________ >>>> Gluster-users mailing list >>>> Gluster-users@xxxxxxxxxxx >>>> http://supercolony.gluster.org/mailman/listinfo/gluster-users >>> _______________________________________________ >>> Gluster-users mailing list >>> Gluster-users@xxxxxxxxxxx >>> http://supercolony.gluster.org/mailman/listinfo/gluster-users >> _______________________________________________ >> Gluster-users mailing list >> Gluster-users@xxxxxxxxxxx >> http://supercolony.gluster.org/mailman/listinfo/gluster-users > _______________________________________________ > Gluster-users mailing list > Gluster-users@xxxxxxxxxxx > http://supercolony.gluster.org/mailman/listinfo/gluster-users _______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://supercolony.gluster.org/mailman/listinfo/gluster-users