Hi Daniel,
Check if telnet works on brick port from both the interface.
telnet 172.17.2.30 <brick-port1>
telnet 192.168.135.36 <brick-port1>
telnet 172.17.2.31 <brick-port2>
telnet 192.168.135.46 <brick-port2>
Thanks,
Vijay
On Thursday 31 July 2014 04:37 PM, Daniel Müller wrote:
So,
[root@centclust1 ~]# ifconfig
eth0 Link encap:Ethernet Hardware Adresse 00:25:90:80:D9:E8
inet Adresse:172.17.2.30 Bcast:172.17.2.255 Maske:255.255.255.0
inet6 Adresse: fe80::225:90ff:fe80:d9e8/64 Gültigkeitsbereich:Verbindung
UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
RX packets:3506528 errors:0 dropped:0 overruns:0 frame:0
TX packets:169905 errors:0 dropped:0 overruns:0 carrier:0
Kollisionen:0 Sendewarteschlangenlänge:1000
RX bytes:476128477 (454.0 MiB) TX bytes:18788266 (17.9 MiB)
Speicher:fe860000-fe880000
eth1 Link encap:Ethernet Hardware Adresse 00:25:90:80:D9:E9
inet Adresse:192.168.135.36 Bcast:192.168.135.255 Maske:255.255.255.0
inet6 Adresse: fe80::225:90ff:fe80:d9e9/64 Gültigkeitsbereich:Verbindung
UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
RX packets:381664693 errors:0 dropped:0 overruns:0 frame:0
TX packets:380924973 errors:0 dropped:0 overruns:0 carrier:0
Kollisionen:0 Sendewarteschlangenlänge:1000
RX bytes:477454156923 (444.6 GiB) TX bytes:476729269342 (443.9 GiB)
Speicher:fe8e0000-fe900000
lo Link encap:Lokale Schleife
inet Adresse:127.0.0.1 Maske:255.0.0.0
inet6 Adresse: ::1/128 Gültigkeitsbereich:Maschine
UP LOOPBACK RUNNING MTU:16436 Metric:1
RX packets:93922879 errors:0 dropped:0 overruns:0 frame:0
TX packets:93922879 errors:0 dropped:0 overruns:0 carrier:0
Kollisionen:0 Sendewarteschlangenlänge:0
RX bytes:462579764180 (430.8 GiB) TX bytes:462579764180 (430.8 GiB)
[root@centclust2 ~]# ifconfig
eth0 Link encap:Ethernet Hardware Adresse 00:25:90:80:EF:00
inet Adresse:172.17.2.31 Bcast:172.17.2.255 Maske:255.255.255.0
inet6 Adresse: fe80::225:90ff:fe80:ef00/64 Gültigkeitsbereich:Verbindung
UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
RX packets:1383117 errors:0 dropped:0 overruns:0 frame:0
TX packets:45828 errors:0 dropped:0 overruns:0 carrier:0
Kollisionen:0 Sendewarteschlangenlänge:1000
RX bytes:185634714 (177.0 MiB) TX bytes:5357926 (5.1 MiB)
Speicher:fe860000-fe880000
eth1 Link encap:Ethernet Hardware Adresse 00:25:90:80:EF:01
inet Adresse:192.168.135.46 Bcast:192.168.135.255 Maske:255.255.255.0
inet6 Adresse: fe80::225:90ff:fe80:ef01/64 Gültigkeitsbereich:Verbindung
UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
RX packets:340364283 errors:0 dropped:0 overruns:0 frame:0
TX packets:59930672 errors:0 dropped:0 overruns:0 carrier:0
Kollisionen:0 Sendewarteschlangenlänge:1000
RX bytes:473823738544 (441.2 GiB) TX bytes:9973035418 (9.2 GiB)
Speicher:fe8e0000-fe900000
lo Link encap:Lokale Schleife
inet Adresse:127.0.0.1 Maske:255.0.0.0
inet6 Adresse: ::1/128 Gültigkeitsbereich:Maschine
UP LOOPBACK RUNNING MTU:16436 Metric:1
RX packets:1102979 errors:0 dropped:0 overruns:0 frame:0
TX packets:1102979 errors:0 dropped:0 overruns:0 carrier:0
Kollisionen:0 Sendewarteschlangenlänge:0
RX bytes:126066547 (120.2 MiB) TX bytes:126066547 (120.2 MiB)
[root@centclust1 ~]# route
Kernel IP Routentabelle
Ziel Router Genmask Flags Metric Ref Use Iface
192.168.135.0 * 255.255.255.0 U 1 0 0 eth1
172.17.2.0 * 255.255.255.0 U 1 0 0 eth0
default s4master 0.0.0.0 UG 0 0 0 eth1
[root@centclust2 ~]# route
Kernel IP Routentabelle
Ziel Router Genmask Flags Metric Ref Use Iface
192.168.135.0 * 255.255.255.0 U 0 0 0 eth1
172.17.2.0 * 255.255.255.0 U 0 0 0 eth0
link-local * 255.255.0.0 U 1002 0 0 eth0
link-local * 255.255.0.0 U 1003 0 0 eth1
default s4master 0.0.0.0 UG 0 0 0 eth1
[root@centclust1 ~]# route -n
Kernel IP Routentabelle
Ziel Router Genmask Flags Metric Ref Use Iface
192.168.135.0 0.0.0.0 255.255.255.0 U 1 0 0 eth1
172.17.2.0 0.0.0.0 255.255.255.0 U 1 0 0 eth0
0.0.0.0 192.168.135.230 0.0.0.0 UG 0 0 0 eth1
[root@centclust2 ~]# route -n
Kernel IP Routentabelle
Ziel Router Genmask Flags Metric Ref Use Iface
192.168.135.0 0.0.0.0 255.255.255.0 U 0 0 0 eth1
172.17.2.0 0.0.0.0 255.255.255.0 U 0 0 0 eth0
169.254.0.0 0.0.0.0 255.255.0.0 U 1002 0 0 eth0
169.254.0.0 0.0.0.0 255.255.0.0 U 1003 0 0 eth1
0.0.0.0 192.168.135.230 0.0.0.0 UG 0 0 0 eth1
EDV Daniel Müller
Leitung EDV
Tropenklinik Paul-Lechler-Krankenhaus
Paul-Lechler-Str. 24
72076 Tübingen
Tel.: 07071/206-463, Fax: 07071/206-499
eMail: mueller@xxxxxxxxxxxxxxx
Internet: www.tropenklinik.de
-----Ursprüngliche Nachricht-----
Von: Krishnan Parthasarathi [mailto:kparthas@xxxxxxxxxx]
Gesendet: Donnerstag, 31. Juli 2014 12:55
An: mueller@xxxxxxxxxxxxxxx
Cc: gluster-devel-bounces@xxxxxxxxxxx; gluster-users@xxxxxxxxxxx
Betreff: Re: WG: Strange issu concerning glusterfs 3.5.1 on centos 6.5
Daniel,
Could you provide the following details from your original two NIC setup probed using hostname?
1) output of ifconfig of the two NICs on both the nodes .
2) output of route from both the nodes.
~KP
----- Original Message -----
Hello and thank you so far,
What I have recognized is, having more than one nic running this is
confusing glusterfs 3.5. I never saw this on my glusterfs 3.4 and 3.2
systems still working.
So I set up just clean erased gluster with yum glusterfs* erase and did:
Logged in to my both nodes in the 135 subnet,ex:
Ssh 192.168.135.36 (centclust1) (172.17.2.30 is the 2nd nic) Ssh
192.168.135.46 (centclust2) (172.17.2.31 is the 2nd nic) Started
gluster on both nodes , service glusterd start.
Did the peer probe on 192.168.135.36/centclust1:
Gluster peer probe 192.168.135.46 //Former I did gluster peer probe
centclust2
This result in:
[root@centclust1 ~]# gluster peer status Number of Peers: 1
Hostname: 192.168.135.46
Uuid: c395c15d-5187-4e5b-b680-57afcb88b881
State: Peer in Cluster (Connected)
[root@centclust2 backup]# gluster peer status Number of Peers: 1
Hostname: 192.168.135.36
Uuid: 94d5903b-ebe9-40d6-93bf-c2f2e92909a0
State: Peer in Cluster (Connected)
The signifent difference gluster now shows the ip of both nodes
Now I did the create the replicating vol:
gluster volume create smbcluster replica 2 transport tcp
192.168.135.36:/sbu/glusterfs/export
192.168.135.46:/sbu/glusterfs/export
started the volume
gluster volume status
Status of volume: smbcluster
Gluster process Port Online Pid
------------------------------------------------------------------------------
Brick 192.168.135.36:/sbu/glusterfs/export 49152 Y 27421
Brick 192.168.135.46:/sbu/glusterfs/export 49152 Y 12186
NFS Server on localhost 2049 Y 27435
Self-heal Daemon on localhost N/A Y 27439
NFS Server on 192.168.135.46 2049 Y 12200
Self-heal Daemon on 192.168.135.46 N/A Y 12204
Task Status of Volume smbcluster
----------------------------------------------------------------------
--------
There are no active volume tasks
Mounted the volumes:
Centclust1:mount -t glusterfs 192.168.135.36:/smbcluster /mntgluster
-o acl Centclust2:mount -t glusterfs 192.168.135.46:/smbcluster
/mntgluster -o acl
And BINGO up and running!!!!!!!
EDV Daniel Müller
Leitung EDV
Tropenklinik Paul-Lechler-Krankenhaus
Paul-Lechler-Str. 24
72076 Tübingen
Tel.: 07071/206-463, Fax: 07071/206-499
eMail: mueller@xxxxxxxxxxxxxxx
Internet: www.tropenklinik.de
-----Ursprüngliche Nachricht-----
Von: Krishnan Parthasarathi [mailto:kparthas@xxxxxxxxxx]
Gesendet: Mittwoch, 30. Juli 2014 16:52
An: mueller@xxxxxxxxxxxxxxx
Cc: gluster-devel-bounces@xxxxxxxxxxx; gluster-users@xxxxxxxxxxx
Betreff: Re: WG: Strange issu concerning glusterfs
3.5.1 on centos 6.5
Daniel,
I didn't get a chance to follow up with debugging this issue. I will
look into this and get back to you. I suspect that there is something
different about the network layer behaviour in your setup.
~KP
----- Original Message -----
Just another other test:
[root@centclust1 sicherung]# getfattr -d -e hex -m . /sicherung/bu
getfattr: Entferne führenden '/' von absoluten Pfadnamen # file:
sicherung/bu
security.selinux=0x756e636f6e66696e65645f753a6f626a6563745f723a66696
c6
55f743a733000
trusted.afr.smbbackup-client-0=0x000000000000000000000000
trusted.afr.smbbackup-client-1=0x000000000000000200000001
trusted.gfid=0x00000000000000000000000000000001
trusted.glusterfs.dht=0x000000010000000000000000ffffffff
trusted.glusterfs.volume-id=0x6f51d002e634437db58d9b952693f1df
[root@centclust2 glusterfs]# getfattr -d -e hex -m . /sicherung/bu
getfattr: Entferne führenden '/' von absoluten Pfadnamen # file:
sicherung/bu
security.selinux=0x756e636f6e66696e65645f753a6f626a6563745f723a66696
c6
55f743a733000
trusted.afr.smbbackup-client-0=0x000000000000000200000001
trusted.afr.smbbackup-client-1=0x000000000000000000000000
trusted.gfid=0x00000000000000000000000000000001
trusted.glusterfs.dht=0x000000010000000000000000ffffffff
trusted.glusterfs.volume-id=0x6f51d002e634437db58d9b952693f1df
Is this ok?
After long testing and doing a /etc/init.d/network restart the
replication started once/a short time then ended up!?
Any idea???????
EDV Daniel Müller
Leitung EDV
Tropenklinik Paul-Lechler-Krankenhaus Paul-Lechler-Str. 24
72076 Tübingen
Tel.: 07071/206-463, Fax: 07071/206-499
eMail: mueller@xxxxxxxxxxxxxxx
Internet: www.tropenklinik.de
"Der Mensch ist die Medizin des Menschen"
-----Ursprüngliche Nachricht-----
Von: Krishnan Parthasarathi [mailto:kparthas@xxxxxxxxxx]
Gesendet: Mittwoch, 30. Juli 2014 11:09
An: mueller@xxxxxxxxxxxxxxx
Cc: gluster-devel-bounces@xxxxxxxxxxx; gluster-users@xxxxxxxxxxx
Betreff: Re: WG: Strange issu concerning glusterfs
3.5.1 on centos 6.5
Could you provide the output of the following command?
netstat -ntap | grep gluster
This should tell us if glusterfsd processes (bricks) are listening
on all interfaces.
~KP
----- Original Message -----
Just one idea
I add a second NIC with a 172.2.17... adress on both machines.
Could this cause the trouble!?
EDV Daniel Müller
Leitung EDV
Tropenklinik Paul-Lechler-Krankenhaus Paul-Lechler-Str. 24
72076 Tübingen
Tel.: 07071/206-463, Fax: 07071/206-499
eMail: mueller@xxxxxxxxxxxxxxx
Internet: www.tropenklinik.de
-----Ursprüngliche Nachricht-----
Von: Krishnan Parthasarathi [mailto:kparthas@xxxxxxxxxx]
Gesendet: Mittwoch, 30. Juli 2014 09:29
An: mueller@xxxxxxxxxxxxxxx
Cc: gluster-devel-bounces@xxxxxxxxxxx; gluster-users@xxxxxxxxxxx
Betreff: Re: WG: Strange issu concerning glusterfs
3.5.1 on centos 6.5
Daniel,
From a quick look, I see that glustershd and the nfs client is
unable to connect to one of the bricks. This is resulting in data
from mounts being written to local bricks only.
I should have asked this before, could you provide the bricks logs
as well?
Could you also try to connect to the bricks using telnet?
For eg, from centclust1, telnet centclust2 <brick-port>.
~KP
----- Original Message -----
So my logs. I disable ssl meanwhile but it is the same situation.
No replication!?
EDV Daniel Müller
Leitung EDV
Tropenklinik Paul-Lechler-Krankenhaus Paul-Lechler-Str. 24
72076 Tübingen
Tel.: 07071/206-463, Fax: 07071/206-499
eMail: mueller@xxxxxxxxxxxxxxx
Internet: www.tropenklinik.de
-----Ursprüngliche Nachricht-----
Von: Krishnan Parthasarathi [mailto:kparthas@xxxxxxxxxx]
Gesendet: Mittwoch, 30. Juli 2014 08:56
An: mueller@xxxxxxxxxxxxxxx
Cc: gluster-users@xxxxxxxxxxx; gluster-devel-bounces@xxxxxxxxxxx
Betreff: Re: WG: Strange issu concerning
glusterfs
3.5.1 on centos 6.5
Could you attach the entire mount and glustershd log files to
this thread?
~KP
----- Original Message -----
NO ONE!??
This is an entry of my glustershd.log:
[2014-07-30 06:40:59.294334] W
[client-handshake.c:1846:client_dump_version_cbk]
0-smbbackup-client-1:
received RPC status error
[2014-07-30 06:40:59.294352] I
[client.c:2229:client_rpc_notify]
0-smbbackup-client-1: disconnected from 172.17.2.31:49152.
Client process will keep trying to connect to glusterd until
brick's port is available
This is from mnt-sicherung.log:
[2014-07-30 06:40:38.259850] E [socket.c:2820:socket_connect]
1-smbbackup-client-0: connection attempt on 172.17.2.30:24007
failed, (Connection timed out) [2014-07-30 06:40:41.275120] I
[rpc-clnt.c:1729:rpc_clnt_reconfig]
1-smbbackup-client-0: changing port to 49152 (from 0)
[root@centclust1 sicherung]# gluster --remote-host=centclust1
peer status Number of Peers: 1
Hostname: centclust2
Uuid: 4f15e9bd-9b5a-435b-83d2-4ed202c66b11
State: Peer in Cluster (Connected)
[root@centclust1 sicherung]# gluster --remote-host=centclust2
peer status Number of Peers: 1
Hostname: 172.17.2.30
Uuid: 99fe6a2c-df7e-4475-a7bc-a35abba620fb
State: Peer in Cluster (Connected)
[root@centclust1 ssl]# ps aux | grep gluster
root 13655 0.0 0.0 413848 16872 ? Ssl 08:10 0:00
/usr/sbin/glusterd --pid-file=/var/run/glusterd.pid
root 13958 0.0 0.0 12139920 44812 ? Ssl 08:11 0:00
/usr/sbin/glusterfsd -s centclust1.tplk.loc --volfile-id
smbbackup.centclust1.tplk.loc.sicherung-bu -p
/var/lib/glusterd/vols/smbbackup/run/centclust1.tplk.loc-sicherung-bu.
pid -S /var/run/4c65260e12e2d3a9a5549446f491f383.socket
--brick-name /sicherung/bu -l
/var/log/glusterfs/bricks/sicherung-bu.log
--xlator-option
*-posix.glusterd-uuid=99fe6a2c-df7e-4475-a7bc-a35abba620fb
--brick-port
49152 --xlator-option smbbackup-server.listen-port=49152
root 13972 0.0 0.0 815748 58252 ? Ssl 08:11 0:00
/usr/sbin/glusterfs -s localhost --volfile-id gluster/nfs -p
/var/lib/glusterd/nfs/run/nfs.pid -l
/var/log/glusterfs/nfs.log -S /var/run/ee6f37fc79b9cb1968eca387930b39fb.socket
root 13976 0.0 0.0 831160 29492 ? Ssl 08:11 0:00
/usr/sbin/glusterfs -s localhost --volfile-id
gluster/glustershd -p
/var/lib/glusterd/glustershd/run/glustershd.pid -l
/var/log/glusterfs/glustershd.log -S
/var/run/aa970d146eb23ba7124e6c4511879850.socket --xlator-option *replicate*.node-uuid=99fe6a2c-df7e-4475-a7bc-a35abba620fb
root 15781 0.0 0.0 105308 932 pts/1 S+ 08:47 0:00 grep
gluster
root 29283 0.0 0.0 451116 56812 ? Ssl Jul29 0:21
/usr/sbin/glusterfs -s localhost --volfile-id gluster/nfs -p
/var/lib/glusterd/nfs/run/nfs.pid -l
/var/log/glusterfs/nfs.log -S /var/run/a7fcb1d1d3a769d28df80b85ae5d13c4.socket
root 29287 0.0 0.0 335432 25848 ? Ssl Jul29 0:21
/usr/sbin/glusterfs -s localhost --volfile-id
gluster/glustershd -p
/var/lib/glusterd/glustershd/run/glustershd.pid -l
/var/log/glusterfs/glustershd.log -S
/var/run/833e60f976365c2a307f92fb233942a2.socket --xlator-option *replicate*.node-uuid=64b1a7eb-2df3-47bd-9379-39c29e5a001a
root 31698 0.0 0.0 1438392 57952 ? Ssl Jul29 0:12
/usr/sbin/glusterfs --acl --volfile-server=centclust1.tplk.loc
--volfile-id=/smbbackup /mnt/sicherung
[root@centclust2 glusterfs]# ps aux | grep gluster
root 1561 0.0 0.0 1481492 60152 ? Ssl Jul29 0:12
/usr/sbin/glusterfs --acl --volfile-server=centclust2.tplk.loc
--volfile-id=/smbbackup /mnt/sicherung
root 15656 0.0 0.0 413848 16832 ? Ssl 08:11 0:01
/usr/sbin/glusterd --pid-file=/var/run/glusterd.pid
root 15942 0.0 0.0 12508704 43860 ? Ssl 08:11 0:00
/usr/sbin/glusterfsd -s centclust2.tplk.loc --volfile-id
smbbackup.centclust2.tplk.loc.sicherung-bu -p
/var/lib/glusterd/vols/smbbackup/run/centclust2.tplk.loc-sicherung-bu.
pid -S /var/run/40a554af3860eddd5794b524576d0520.socket
--brick-name /sicherung/bu -l
/var/log/glusterfs/bricks/sicherung-bu.log
--xlator-option
*-posix.glusterd-uuid=4f15e9bd-9b5a-435b-83d2-4ed202c66b11
--brick-port
49152 --xlator-option smbbackup-server.listen-port=49152
root 15956 0.0 0.0 825992 57496 ? Ssl 08:11 0:00
/usr/sbin/glusterfs -s localhost --volfile-id gluster/nfs -p
/var/lib/glusterd/nfs/run/nfs.pid -l
/var/log/glusterfs/nfs.log -S /var/run/602d1d8ba7b80ded2b70305ed7417cf5.socket
root 15960 0.0 0.0 841404 26760 ? Ssl 08:11 0:00
/usr/sbin/glusterfs -s localhost --volfile-id
gluster/glustershd -p
/var/lib/glusterd/glustershd/run/glustershd.pid -l
/var/log/glusterfs/glustershd.log -S
/var/run/504d01c7f7df8b8306951cc2aaeaf52c.socket
--xlator-option
*replicate*.node-uuid=4f15e9bd-9b5a-435b-83d2-4ed202c66b11
root 17728 0.0 0.0 105312 936 pts/0 S+ 08:48 0:00 grep
gluster
root 32363 0.0 0.0 451100 55584 ? Ssl Jul29 0:21
/usr/sbin/glusterfs -s localhost --volfile-id gluster/nfs -p
/var/lib/glusterd/nfs/run/nfs.pid -l
/var/log/glusterfs/nfs.log -S /var/run/73054288d1cadfb87b4b9827bd205c7b.socket
root 32370 0.0 0.0 335432 26220 ? Ssl Jul29 0:21
/usr/sbin/glusterfs -s localhost --volfile-id
gluster/glustershd -p
/var/lib/glusterd/glustershd/run/glustershd.pid -l
/var/log/glusterfs/glustershd.log -S
/var/run/de1427ce373c792c76c38b12c106f029.socket
--xlator-option
*replicate*.node-uuid=83e6d78c-0119-4537-8922-b3e731718864
Leitung EDV
Tropenklinik Paul-Lechler-Krankenhaus Paul-Lechler-Str. 24
72076 Tübingen
Tel.: 07071/206-463, Fax: 07071/206-499
eMail: mueller@xxxxxxxxxxxxxxx
Internet: www.tropenklinik.de
-----Ursprüngliche Nachricht-----
Von: Daniel Müller [mailto:mueller@xxxxxxxxxxxxxxx]
Gesendet: Dienstag, 29. Juli 2014 16:02
An: 'gluster-users@xxxxxxxxxxx'
Betreff: Strange issu concerning glusterfs 3.5.1 on centos 6.5
Dear all,
there is a strange issue centos6.5 and glusterfs 3.5.1:
glusterd -V
glusterfs 3.5.1 built on Jun 24 2014 15:09:41 Repository revision:
git://git.gluster.com/glusterfs.git
Copyright (c) 2006-2013 Red Hat, Inc. <http://www.redhat.com/>
GlusterFS comes with ABSOLUTELY NO WARRANTY.
It is licensed to you under your choice of the GNU Lesser
General Public License, version 3 or any later version (LGPLv3
or later), or the GNU General Public License, version 2
(GPLv2), in all cases as published by the Free Software
Foundation
I try to set up a replicated 2 brick vol on two centos 6.5 server.
I can probe well and my nodes are reporting no errors:
[root@centclust1 mnt]# gluster peer status Number of Peers: 1
Hostname: centclust2
Uuid: 4f15e9bd-9b5a-435b-83d2-4ed202c66b11
State: Peer in Cluster (Connected)
[root@centclust2 sicherung]# gluster peer status Number of Peers:
1
Hostname: 172.17.2.30
Uuid: 99fe6a2c-df7e-4475-a7bc-a35abba620fb
State: Peer in Cluster (Connected)
Now I set up a replicating VOl on an XFS-Disk: /dev/sdb1 on
/sicherung type xfs (rw)
gluster volume create smbbackup replica 2 transport tcp
centclust1.tplk.loc:/sicherung/bu
centclust2.tplk.loc:/sicherung/bu
gluster volume smbbackup status reports ok:
[root@centclust1 mnt]# gluster volume status smbbackup Status
of
volume: smbbackup
Gluster process Port
Online
Pid
--------------------------------------------------------------
--
--
--
--
------
--
Brick centclust1.tplk.loc:/sicherung/bu 49152 Y
31969
Brick centclust2.tplk.loc:/sicherung/bu 49152 Y
2124
NFS Server on localhost 2049 Y
31983
Self-heal Daemon on localhost N/A Y
31987
NFS Server on centclust2 2049 Y
2138
Self-heal Daemon on centclust2 N/A Y
2142
Task Status of Volume smbbackup
--------------------------------------------------------------
--
--
--
--
------
--
There are no active volume tasks
[root@centclust2 sicherung]# gluster volume status smbbackup
Status of
volume: smbbackup
Gluster process Port
Online
Pid
--------------------------------------------------------------
--
--
--
--
------
--
Brick centclust1.tplk.loc:/sicherung/bu 49152 Y
31969
Brick centclust2.tplk.loc:/sicherung/bu 49152 Y
2124
NFS Server on localhost 2049 Y
2138
Self-heal Daemon on localhost N/A Y
2142
NFS Server on 172.17.2.30 2049 Y
31983
Self-heal Daemon on 172.17.2.30 N/A Y
31987
Task Status of Volume smbbackup
--------------------------------------------------------------
--
--
--
--
------
--
There are no active volume tasks
I mounted the vol on both servers with:
mount -t glusterfs centclust1.tplk.loc:/smbbackup
/mnt/sicherung -o acl mount -t glusterfs
centclust2.tplk.loc:/smbbackup /mnt/sicherung -o acl
But when I write in /mnt/sicherung the files are not
replicated to the other node in anyway!??
They rest on the local server in /mnt/sicherung and
/sicherung/bu On each node separate:#
[root@centclust1 sicherung]# pwd /mnt/sicherung
[root@centclust1 sicherung]# touch test.txt
[root@centclust1 sicherung]# ls test.txt
[root@centclust2 sicherung]# pwd /mnt/sicherung
[root@centclust2 sicherung]# ls more.txt
[root@centclust1 sicherung]# ls -la /sicherung/bu insgesamt 0
drwxr-xr-x. 3 root root 38 29. Jul 15:56 .
drwxr-xr-x. 3 root root 15 29. Jul 14:31 ..
drw-------. 15 root root 142 29. Jul 15:56 .glusterfs
-rw-r--r--. 2 root root 0 29. Jul 15:56 test.txt
[root@centclust2 sicherung]# ls -la /sicherung/bu insgesamt 0
drwxr-xr-x. 3 root root 38 29. Jul 15:32 .
drwxr-xr-x. 3 root root 15 29. Jul 14:31 ..
drw-------. 7 root root 70 29. Jul 15:32 .glusterfs -rw-r--r--.
2 root root 0 29. Jul 15:32 more.txt
Greetings
Daniel
EDV Daniel Müller
Leitung EDV
Tropenklinik Paul-Lechler-Krankenhaus Paul-Lechler-Str. 24
72076 Tübingen
Tel.: 07071/206-463, Fax: 07071/206-499
eMail: mueller@xxxxxxxxxxxxxxx
Internet: www.tropenklinik.de
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://supercolony.gluster.org/mailman/listinfo/gluster-users
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://supercolony.gluster.org/mailman/listinfo/gluster-users
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://supercolony.gluster.org/mailman/listinfo/gluster-users
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://supercolony.gluster.org/mailman/listinfo/gluster-users
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://supercolony.gluster.org/mailman/listinfo/gluster-users
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://supercolony.gluster.org/mailman/listinfo/gluster-users
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://supercolony.gluster.org/mailman/listinfo/gluster-users