Hi, I solved the problem by deleting the volume and create it again. It seems that the problem is due to something else and resetting gluster might be the way to solve the problem. Does anybody know a better way to completely reset gluster? (Deleting and recreating the volume seems to be not a good way in a production system.) Thanks. On Sat, Mar 22, 2014 at 2:14 PM, Peng Yu <pengyu.ut@xxxxxxxxx> wrote: > Hi, > > Here are the respective IP addresses of both servers. Why should I > remove "auth.allow: 172.17.*.*"? (And how to remove it?) > > pengy@rigel:~$ ifconfig |grep -A 7 '^br1' > br1 Link encap:Ethernet HWaddr c8:1f:66:e2:90:45 > inet addr:172.17.1.1 Bcast:172.17.255.255 Mask:255.255.0.0 > inet6 addr: fe80::ca1f:66ff:fee2:9045/64 Scope:Link > UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 > RX packets:312191 errors:0 dropped:0 overruns:0 frame:0 > TX packets:210807 errors:0 dropped:0 overruns:0 carrier:0 > collisions:0 txqueuelen:0 > RX bytes:3741197826 (3.7 GB) TX bytes:25954291 (25.9 MB) > pengy@betelgeuse:~$ ifconfig |grep -A 7 '^br1' > br1 Link encap:Ethernet HWaddr c8:1f:66:df:01:0b > inet addr:172.17.2.1 Bcast:172.17.255.255 Mask:255.255.0.0 > inet6 addr: fe80::ca1f:66ff:fedf:10b/64 Scope:Link > UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 > RX packets:197382 errors:0 dropped:0 overruns:0 frame:0 > TX packets:90443 errors:0 dropped:0 overruns:0 carrier:0 > collisions:0 txqueuelen:0 > RX bytes:11914450 (11.9 MB) TX bytes:10016451 (10.0 MB) > > > Here are are the firewall information. I don't see anything is wrong. > Do you see anything wrong? Thanks. > > pengy@rigel:~$ sudo ufw app list > Available applications: > OpenSSH > pengy@rigel:~$ sudo ufw status > Status: inactive > pengy@rigel:~$ sudo iptables -L > Chain INPUT (policy ACCEPT) > target prot opt source destination > ACCEPT udp -- anywhere anywhere udp dpt:domain > ACCEPT tcp -- anywhere anywhere tcp dpt:domain > ACCEPT udp -- anywhere anywhere udp dpt:bootps > ACCEPT tcp -- anywhere anywhere tcp dpt:bootps > > Chain FORWARD (policy ACCEPT) > target prot opt source destination > ACCEPT tcp -- anywhere 192.168.122.200 state > NEW,RELATED,ESTABLISHED tcp dpt:ssh > ACCEPT all -- anywhere 192.168.122.0/24 ctstate > RELATED,ESTABLISHED > ACCEPT all -- 192.168.122.0/24 anywhere > ACCEPT all -- anywhere anywhere > REJECT all -- anywhere anywhere > reject-with icmp-port-unreachable > REJECT all -- anywhere anywhere > reject-with icmp-port-unreachable > > Chain OUTPUT (policy ACCEPT) > target prot opt source destination > > pengy@betelgeuse:~$ sudo ufw app list > Available applications: > OpenSSH > pengy@betelgeuse:~$ sudo ufw status > Status: inactive > pengy@betelgeuse:~$ sudo iptables -L > Chain INPUT (policy ACCEPT) > target prot opt source destination > ACCEPT udp -- anywhere anywhere udp dpt:domain > ACCEPT tcp -- anywhere anywhere tcp dpt:domain > ACCEPT udp -- anywhere anywhere udp dpt:bootps > ACCEPT tcp -- anywhere anywhere tcp dpt:bootps > > Chain FORWARD (policy ACCEPT) > target prot opt source destination > ACCEPT all -- anywhere 192.168.122.0/24 ctstate > RELATED,ESTABLISHED > ACCEPT all -- 192.168.122.0/24 anywhere > ACCEPT all -- anywhere anywhere > REJECT all -- anywhere anywhere > reject-with icmp-port-unreachable > REJECT all -- anywhere anywhere > reject-with icmp-port-unreachable > > Chain OUTPUT (policy ACCEPT) > target prot opt source destination > > > On Sat, Mar 22, 2014 at 2:01 PM, Carlos Capriotti > <capriotti.carlos@xxxxxxxxx> wrote: >> One thing that caught my eyes: >> >> auth.allow: 172.17.*.* >> >> Can you remove that, restart glusterd/the nodes and try again ? >> >> Also, do you have firewall/iptables rules enabled ? If yes, consider testing >> with iptables/firewall disabled. >> >> >> >> >> On Sat, Mar 22, 2014 at 7:09 PM, Peng Yu <pengyu.ut@xxxxxxxxx> wrote: >>> >>> Hi, >>> >>> There should be two bricks in the volume "gv". But `sudo gluster >>> volume status` does not show `betelgeuse:/mnt/raid6/glusterfs_export`. >>> Does anybody know what is wrong with this? Thanks. >>> >>> pengy@rigel:~$ sudo gluster volume status >>> Status of volume: gv >>> Gluster process Port Online Pid >>> >>> ------------------------------------------------------------------------------ >>> Brick rigel:/mnt/raid6/glusterfs_export 49152 Y 38971 >>> NFS Server on localhost N/A N N/A >>> Self-heal Daemon on localhost N/A N N/A >>> >>> There are no active volume tasks >>> pengy@rigel:~$ sudo gluster volume info >>> >>> Volume Name: gv >>> Type: Replicate >>> Volume ID: 64754d6c-3736-41d8-afb5-d8071a6a6a07 >>> Status: Started >>> Number of Bricks: 1 x 2 = 2 >>> Transport-type: tcp >>> Bricks: >>> Brick1: rigel:/mnt/raid6/glusterfs_export >>> Brick2: betelgeuse:/mnt/raid6/glusterfs_export >>> Options Reconfigured: >>> auth.allow: 172.17.*.* >>> >>> -- >>> Regards, >>> Peng >>> _______________________________________________ >>> Gluster-users mailing list >>> Gluster-users@xxxxxxxxxxx >>> http://supercolony.gluster.org/mailman/listinfo/gluster-users >> >> > > > > -- > Regards, > Peng -- Regards, Peng _______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://supercolony.gluster.org/mailman/listinfo/gluster-users