-Atin
Sent from one plus one
On Aug 10, 2015 9:37 PM, "Kingsley" <gluster@xxxxxxxxxxxxxxxxxxx> wrote:
>
> On Mon, 2015-08-10 at 21:34 +0530, Atin Mukherjee wrote:
> > -Atin
> > Sent from one plus one
> > On Aug 10, 2015 7:19 PM, "Kingsley" <gluster@xxxxxxxxxxxxxxxxxxx>
> > wrote:
> > >
> > > Further to this, the volume doesn't seem overly healthy. Any idea
> > how I
> > > can get it back into a working state?
> > >
> > > Trying to access one particular directory on the clients just hangs.
> > If
> > > I query heal info, that directory appears in the output as possibly
> > > undergoing heal (actual directory name changed as it's private
> > info):
> > Can you execute strace and see which call is stuck? That would help us
> > to get to the exact component which we would need to look at.
>
> Hi,
>
> I've never used strace before. Could you give me the command line to
> type?
Just type strace followed by the command
Is this what you meant (I renamed the broken directory so that I could create another and let the system continue to work with a freshly created one). It ran very quickly and returned be back to the command prompt, but I then "cd"d into that directory and did a plain "ls" which then hung, ie:
--8<--
[root@voicemail1b-1 14391.broken]# ls
^Z
fg
--8<--
Anyway, the strace:
[root@voicemail1b-1 834723]# strace ls 14391.broken
execve("/usr/bin/ls", ["ls", "14391.broken"], [/* 27 vars */]) = 0
brk(0) = 0x158c000
mmap(NULL, 4096, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_ANONYMOUS, -1, 0) = 0x7f7d2494c000
access("/etc/ld.so.preload", R_OK) = -1 ENOENT (No such file or directory)
open("/etc/ld.so.cache", O_RDONLY|O_CLOEXEC) = 3
fstat(3, {st_mode=S_IFREG|0644, st_size=31874, ...}) = 0
mmap(NULL, 31874, PROT_READ, MAP_PRIVATE, 3, 0) = 0x7f7d24944000
close(3) = 0
open("/lib64/libselinux.so.1", O_RDONLY|O_CLOEXEC) = 3
read(3, "\177ELF\2\1\1\0\0\0\0\0\0\0\0\0\3\0>\0\1\0\0\0\240d\0\0\0\0\0\0"..., 832) = 832
fstat(3, {st_mode=S_IFREG|0755, st_size=147120, ...}) = 0
mmap(NULL, 2246784, PROT_READ|PROT_EXEC, MAP_PRIVATE|MAP_DENYWRITE, 3, 0) = 0x7f7d24509000
mprotect(0x7f7d2452a000, 2097152, PROT_NONE) = 0
mmap(0x7f7d2472a000, 8192, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x21000) = 0x7f7d2472a000
mmap(0x7f7d2472c000, 6272, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_ANONYMOUS, -1, 0) = 0x7f7d2472c000
close(3) = 0
open("/lib64/libcap.so.2", O_RDONLY|O_CLOEXEC) = 3
read(3, "\177ELF\2\1\1\0\0\0\0\0\0\0\0\0\3\0>\0\1\0\0\0 \26\0\0\0\0\0\0"..., 832) = 832
fstat(3, {st_mode=S_IFREG|0755, st_size=20024, ...}) = 0
mmap(NULL, 2114112, PROT_READ|PROT_EXEC, MAP_PRIVATE|MAP_DENYWRITE, 3, 0) = 0x7f7d24304000
mprotect(0x7f7d24308000, 2093056, PROT_NONE) = 0
mmap(0x7f7d24507000, 8192, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x3000) = 0x7f7d24507000
close(3) = 0
open("/lib64/libacl.so.1", O_RDONLY|O_CLOEXEC) = 3
read(3, "\177ELF\2\1\1\0\0\0\0\0\0\0\0\0\3\0>\0\1\0\0\0\200\37\0\0\0\0\0\0"..., 832) = 832
fstat(3, {st_mode=S_IFREG|0755, st_size=37056, ...}) = 0
mmap(NULL, 4096, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_ANONYMOUS, -1, 0) = 0x7f7d24943000
mmap(NULL, 2130560, PROT_READ|PROT_EXEC, MAP_PRIVATE|MAP_DENYWRITE, 3, 0) = 0x7f7d240fb000
mprotect(0x7f7d24102000, 2097152, PROT_NONE) = 0
mmap(0x7f7d24302000, 8192, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x7000) = 0x7f7d24302000
close(3) = 0
open("/lib64/libc.so.6", O_RDONLY|O_CLOEXEC) = 3
read(3, "\177ELF\2\1\1\3\0\0\0\0\0\0\0\0\3\0>\0\1\0\0\0\0\34\2\0\0\0\0\0"..., 832) = 832
fstat(3, {st_mode=S_IFREG|0755, st_size=2107760, ...}) = 0
mmap(NULL, 3932736, PROT_READ|PROT_EXEC, MAP_PRIVATE|MAP_DENYWRITE, 3, 0) = 0x7f7d23d3a000
mprotect(0x7f7d23ef0000, 2097152, PROT_NONE) = 0
mmap(0x7f7d240f0000, 24576, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x1b6000) = 0x7f7d240f0000
mmap(0x7f7d240f6000, 16960, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_ANONYMOUS, -1, 0) = 0x7f7d240f6000
close(3) = 0
open("/lib64/libpcre.so.1", O_RDONLY|O_CLOEXEC) = 3
read(3, "\177ELF\2\1\1\0\0\0\0\0\0\0\0\0\3\0>\0\1\0\0\0\360\25\0\0\0\0\0\0"..., 832) = 832
fstat(3, {st_mode=S_IFREG|0755, st_size=398272, ...}) = 0
mmap(NULL, 2490888, PROT_READ|PROT_EXEC, MAP_PRIVATE|MAP_DENYWRITE, 3, 0) = 0x7f7d23ad9000
mprotect(0x7f7d23b38000, 2097152, PROT_NONE) = 0
mmap(0x7f7d23d38000, 8192, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x5f000) = 0x7f7d23d38000
close(3) = 0
open("/lib64/liblzma.so.5", O_RDONLY|O_CLOEXEC) = 3
read(3, "\177ELF\2\1\1\0\0\0\0\0\0\0\0\0\3\0>\0\1\0\0\0000/\0\0\0\0\0\0"..., 832) = 832
fstat(3, {st_mode=S_IFREG|0755, st_size=153184, ...}) = 0
mmap(NULL, 4096, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_ANONYMOUS, -1, 0) = 0x7f7d24942000
mmap(NULL, 2245240, PROT_READ|PROT_EXEC, MAP_PRIVATE|MAP_DENYWRITE, 3, 0) = 0x7f7d238b4000
mprotect(0x7f7d238d8000, 2093056, PROT_NONE) = 0
mmap(0x7f7d23ad7000, 8192, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x23000) = 0x7f7d23ad7000
close(3) = 0
open("/lib64/libdl.so.2", O_RDONLY|O_CLOEXEC) = 3
read(3, "\177ELF\2\1\1\0\0\0\0\0\0\0\0\0\3\0>\0\1\0\0\0\320\16\0\0\0\0\0\0"..., 832) = 832
fstat(3, {st_mode=S_IFREG|0755, st_size=19512, ...}) = 0
mmap(NULL, 2109744, PROT_READ|PROT_EXEC, MAP_PRIVATE|MAP_DENYWRITE, 3, 0) = 0x7f7d236b0000
mprotect(0x7f7d236b3000, 2093056, PROT_NONE) = 0
mmap(0x7f7d238b2000, 8192, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x2000) = 0x7f7d238b2000
close(3) = 0
open("/lib64/libattr.so.1", O_RDONLY|O_CLOEXEC) = 3
read(3, "\177ELF\2\1\1\0\0\0\0\0\0\0\0\0\3\0>\0\1\0\0\0\320\23\0\0\0\0\0\0"..., 832) = 832
fstat(3, {st_mode=S_IFREG|0755, st_size=19888, ...}) = 0
mmap(NULL, 2113904, PROT_READ|PROT_EXEC, MAP_PRIVATE|MAP_DENYWRITE, 3, 0) = 0x7f7d234ab000
mprotect(0x7f7d234af000, 2093056, PROT_NONE) = 0
mmap(0x7f7d236ae000, 8192, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x3000) = 0x7f7d236ae000
close(3) = 0
open("/lib64/libpthread.so.0", O_RDONLY|O_CLOEXEC) = 3
read(3, "\177ELF\2\1\1\0\0\0\0\0\0\0\0\0\3\0>\0\1\0\0\0\240l\0\0\0\0\0\0"..., 832) = 832
fstat(3, {st_mode=S_IFREG|0755, st_size=141616, ...}) = 0
mmap(NULL, 4096, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_ANONYMOUS, -1, 0) = 0x7f7d24941000
mmap(NULL, 2208864, PROT_READ|PROT_EXEC, MAP_PRIVATE|MAP_DENYWRITE, 3, 0) = 0x7f7d2328f000
mprotect(0x7f7d232a5000, 2097152, PROT_NONE) = 0
mmap(0x7f7d234a5000, 8192, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x16000) = 0x7f7d234a5000
mmap(0x7f7d234a7000, 13408, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_ANONYMOUS, -1, 0) = 0x7f7d234a7000
close(3) = 0
mmap(NULL, 4096, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_ANONYMOUS, -1, 0) = 0x7f7d24940000
mmap(NULL, 8192, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_ANONYMOUS, -1, 0) = 0x7f7d2493e000
arch_prctl(ARCH_SET_FS, 0x7f7d2493e800) = 0
mprotect(0x7f7d240f0000, 16384, PROT_READ) = 0
mprotect(0x7f7d234a5000, 4096, PROT_READ) = 0
mprotect(0x7f7d236ae000, 4096, PROT_READ) = 0
mprotect(0x7f7d238b2000, 4096, PROT_READ) = 0
mprotect(0x7f7d23ad7000, 4096, PROT_READ) = 0
mprotect(0x7f7d23d38000, 4096, PROT_READ) = 0
mprotect(0x7f7d24302000, 4096, PROT_READ) = 0
mprotect(0x7f7d24507000, 4096, PROT_READ) = 0
mprotect(0x7f7d2472a000, 4096, PROT_READ) = 0
mprotect(0x61a000, 4096, PROT_READ) = 0
mprotect(0x7f7d2494f000, 4096, PROT_READ) = 0
munmap(0x7f7d24944000, 31874) = 0
set_tid_address(0x7f7d2493ead0) = 17906
set_robust_list(0x7f7d2493eae0, 24) = 0
rt_sigaction(SIGRTMIN, {0x7f7d23295780, [], SA_RESTORER|SA_SIGINFO, 0x7f7d2329e130}, NULL, 8) = 0
rt_sigaction(SIGRT_1, {0x7f7d23295810, [], SA_RESTORER|SA_RESTART|SA_SIGINFO, 0x7f7d2329e130}, NULL, 8) = 0
rt_sigprocmask(SIG_UNBLOCK, [RTMIN RT_1], NULL, 8) = 0
getrlimit(RLIMIT_STACK, {rlim_cur=8192*1024, rlim_max=RLIM64_INFINITY}) = 0
statfs("/sys/fs/selinux", {f_type=0xf97cff8c, f_bsize=4096, f_blocks=0, f_bfree=0, f_bavail=0, f_files=0, f_ffree=0, f_fsid={0, 0}, f_namelen=255, f_frsize=4096}) = 0
statfs("/sys/fs/selinux", {f_type=0xf97cff8c, f_bsize=4096, f_blocks=0, f_bfree=0, f_bavail=0, f_files=0, f_ffree=0, f_fsid={0, 0}, f_namelen=255, f_frsize=4096}) = 0
stat("/sys/fs/selinux", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
brk(0) = 0x158c000
brk(0x15ad000) = 0x15ad000
ioctl(1, SNDCTL_TMR_TIMEBASE or SNDRV_TIMER_IOCTL_NEXT_DEVICE or TCGETS, {B38400 opost isig icanon echo ...}) = 0
ioctl(1, TIOCGWINSZ, {ws_row=41, ws_col=202, ws_xpixel=0, ws_ypixel=0}) = 0
stat("14391.broken", {st_mode=S_IFDIR|0755, st_size=8192, ...}) = 0
openat(AT_FDCWD, "14391.broken", O_RDONLY|O_NONBLOCK|O_DIRECTORY|O_CLOEXEC) = 3
getdents(3, /* 23 entries */, 32768) = 552
getdents(3, /* 23 entries */, 32768) = 552
getdents(3, /* 23 entries */, 32768) = 552
getdents(3, /* 23 entries */, 32768) = 552
getdents(3, /* 23 entries */, 32768) = 552
getdents(3, /* 23 entries */, 32768) = 552
getdents(3, /* 23 entries */, 32768) = 552
getdents(3, /* 23 entries */, 32768) = 552
getdents(3, /* 23 entries */, 32768) = 552
getdents(3, /* 23 entries */, 32768) = 552
getdents(3, /* 23 entries */, 32768) = 552
getdents(3, /* 23 entries */, 32768) = 552
getdents(3, /* 23 entries */, 32768) = 552
getdents(3, /* 19 entries */, 32768) = 456
getdents(3, /* 0 entries */, 32768) = 0
close(3) = 0
fstat(1, {st_mode=S_IFCHR|0620, st_rdev=makedev(136, 3), ...}) = 0
mmap(NULL, 4096, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_ANONYMOUS, -1, 0) = 0x7f7d2494b000
write(1, "012 033 046 063 076 087 09"..., 195012 033 046 063 076 087 096 104 112 120 128 136 144 152 160 172 180 195 209 225 235 246 258 279 298 313 343 389 628 900 908 918 926 934 942 950 958 968 980 994
) = 195
write(1, "013 034 049 065 079 088 09"..., 195013 034 049 065 079 088 097 105 113 121 129 137 145 153 161 173 184 196 212 226 236 247 266 281 299 314 348 394 843 901 910 919 927 935 943 951 959 970 981 996
) = 195
write(1, "014 035 050 066 080 089 09"..., 195014 035 050 066 080 089 098 106 114 122 130 138 146 154 162 174 185 197 215 227 237 248 267 288 301 317 349 396 869 902 911 920 928 936 944 952 960 972 982 997
) = 195
write(1, "016 039 052 071 081 090 09"..., 195016 039 052 071 081 090 099 107 115 123 131 139 147 155 163 175 186 198 219 229 238 250 269 291 305 321 350 405 882 903 912 921 929 937 945 953 961 973 984 998
) = 195
write(1, "018 041 055 072 082 091 10"..., 190018 041 055 072 082 091 100 108 116 124 132 140 148 156 164 176 187 203 221 230 239 251 270 292 306 328 354 407 890 904 914 922 930 938 946 954 962 974 985
) = 190
write(1, "019 042 057 073 084 092 10"..., 190019 042 057 073 084 092 101 109 117 125 133 141 149 157 165 177 190 204 222 231 240 253 272 293 308 336 357 413 892 905 915 923 931 939 947 955 965 976 988
) = 190
write(1, "024 043 059 074 085 093 10"..., 190024 043 059 074 085 093 102 110 118 126 134 142 150 158 166 178 193 206 223 232 241 255 274 294 309 339 370 470 895 906 916 924 932 940 948 956 966 977 989
) = 190
write(1, "031 044 060 075 086 095 10"..., 190031 044 060 075 086 095 103 111 119 127 135 143 151 159 167 179 194 207 224 234 243 257 275 296 310 342 386 517 899 907 917 925 933 941 949 957 967 978 993
) = 190
close(1) = 0
munmap(0x7f7d2494b000, 4096) = 0
close(2) = 0
exit_group(0) = ?
+++ exited with 0 +++
>
> Then ... do I need to run something on one of the bricks while strace is
> running?
>
> Cheers,
> Kingsley.
>
>
> > >
> > > [root@gluster1b-1 ~]# gluster volume heal callrec info
> > > Brick gluster1a-1.dns99.co.uk:/data/brick/callrec/
> > > <gfid:164f888f-2049-49e6-ad26-c758ee091863>
> > > /recordings/834723/14391 - Possibly undergoing heal
> > >
> > > <gfid:e280b40c-d8b7-43c5-9da7-4737054d7a7f>
> > > <gfid:b1fbda4a-732f-4f5d-b5a1-8355d786073e>
> > > <gfid:edb74524-b4b7-4190-85e7-4aad002f6e7c>
> > > <gfid:9b8b8446-1e27-4113-93c2-6727b1f457eb>
> > > <gfid:650efeca-b45c-413b-acc3-f0a5853ccebd>
> > > Number of entries: 7
> > >
> > > Brick gluster1b-1.dns99.co.uk:/data/brick/callrec/
> > > Number of entries: 0
> > >
> > > Brick gluster2a-1.dns99.co.uk:/data/brick/callrec/
> > > <gfid:e280b40c-d8b7-43c5-9da7-4737054d7a7f>
> > > <gfid:164f888f-2049-49e6-ad26-c758ee091863>
> > > <gfid:650efeca-b45c-413b-acc3-f0a5853ccebd>
> > > <gfid:b1fbda4a-732f-4f5d-b5a1-8355d786073e>
> > > /recordings/834723/14391 - Possibly undergoing heal
> > >
> > > <gfid:edb74524-b4b7-4190-85e7-4aad002f6e7c>
> > > <gfid:9b8b8446-1e27-4113-93c2-6727b1f457eb>
> > > Number of entries: 7
> > >
> > > Brick gluster2b-1.dns99.co.uk:/data/brick/callrec/
> > > Number of entries: 0
> > >
> > >
> > > If I query each brick directly for the number of files/directories
> > > within that, I get 1731 on gluster1a-1 and gluster2a-1, but 1737 on
> > the
> > > other two, using this command:
> > >
> > > # find /data/brick/callrec/recordings/834723/14391 -print | wc -l
> > >
> > > Cheers,
> > > Kingsley.
> > >
> > > On Mon, 2015-08-10 at 11:05 +0100, Kingsley wrote:
> > > > Sorry for the blind panic - restarting the volume seems to have
> > fixed
> > > > it.
> > > >
> > > > But then my next question - why is this necessary? Surely it
> > undermines
> > > > the whole point of a high availability system?
> > > >
> > > > Cheers,
> > > > Kingsley.
> > > >
> > > > On Mon, 2015-08-10 at 10:53 +0100, Kingsley wrote:
> > > > > Hi,
> > > > >
> > > > > We have a 4 way replicated volume using gluster 3.6.3 on CentOS
> > 7.
> > > > >
> > > > > Over the weekend I did a yum update on each of the bricks in
> > turn, but
> > > > > now when clients (using fuse mounts) try to access the volume,
> > it hangs.
> > > > > Gluster itself wasn't updated (we've disabled that repo so that
> > we keep
> > > > > to 3.6.3 for now).
> > > > >
> > > > > This was what I did:
> > > > >
> > > > > * on first brick, "yum update"
> > > > > * reboot brick
> > > > > * watch "gluster volume status" on another brick and wait
> > for it
> > > > > to say all 4 bricks are online before proceeding to
> > update the
> > > > > next brick
> > > > >
> > > > > I was expecting the clients might pause 30 seconds while they
> > notice a
> > > > > brick is offline, but then recover.
> > > > >
> > > > > I've tried re-mounting clients, but that hasn't helped.
> > > > >
> > > > > I can't see much data in any of the log files.
> > > > >
> > > > > I've tried "gluster volume heal callrec" but it doesn't seem to
> > have
> > > > > helped.
> > > > >
> > > > > What shall I do next?
> > > > >
> > > > > I've pasted some stuff below in case any of it helps.
> > > > >
> > > > > Cheers,
> > > > > Kingsley.
> > > > >
> > > > > [root@gluster1b-1 ~]# gluster volume info callrec
> > > > >
> > > > > Volume Name: callrec
> > > > > Type: Replicate
> > > > > Volume ID: a39830b7-eddb-4061-b381-39411274131a
> > > > > Status: Started
> > > > > Number of Bricks: 1 x 4 = 4
> > > > > Transport-type: tcp
> > > > > Bricks:
> > > > > Brick1: gluster1a-1:/data/brick/callrec
> > > > > Brick2: gluster1b-1:/data/brick/callrec
> > > > > Brick3: gluster2a-1:/data/brick/callrec
> > > > > Brick4: gluster2b-1:/data/brick/callrec
> > > > > Options Reconfigured:
> > > > > performance.flush-behind: off
> > > > > [root@gluster1b-1 ~]#
> > > > >
> > > > >
> > > > > [root@gluster1b-1 ~]# gluster volume status callrec
> > > > > Status of volume: callrec
> > > > > Gluster process Port
> > Online Pid
> > > > >
> > ------------------------------------------------------------------------------
> > > > > Brick gluster1a-1:/data/brick/callrec 49153
> > Y 6803
> > > > > Brick gluster1b-1:/data/brick/callrec 49153
> > Y 2614
> > > > > Brick gluster2a-1:/data/brick/callrec 49153
> > Y 2645
> > > > > Brick gluster2b-1:/data/brick/callrec 49153
> > Y 4325
> > > > > NFS Server on localhost 2049
> > Y 2769
> > > > > Self-heal Daemon on localhost N/A
> > Y 2789
> > > > > NFS Server on gluster2a-1 2049
> > Y 2857
> > > > > Self-heal Daemon on gluster2a-1 N/A
> > Y 2814
> > > > > NFS Server on 88.151.41.100 2049
> > Y 6833
> > > > > Self-heal Daemon on 88.151.41.100 N/A
> > Y 6824
> > > > > NFS Server on gluster2b-1 2049
> > Y 4428
> > > > > Self-heal Daemon on gluster2b-1 N/A
> > Y 4387
> > > > >
> > > > > Task Status of Volume callrec
> > > > >
> > ------------------------------------------------------------------------------
> > > > > There are no active volume tasks
> > > > >
> > > > > [root@gluster1b-1 ~]#
> > > > >
> > > > >
> > > > > [root@gluster1b-1 ~]# gluster volume heal callrec info
> > > > > Brick gluster1a-1.dns99.co.uk:/data/brick/callrec/
> > > > > /to_process - Possibly undergoing heal
> > > > >
> > > > > Number of entries: 1
> > > > >
> > > > > Brick gluster1b-1.dns99.co.uk:/data/brick/callrec/
> > > > > Number of entries: 0
> > > > >
> > > > > Brick gluster2a-1.dns99.co.uk:/data/brick/callrec/
> > > > > /to_process - Possibly undergoing heal
> > > > >
> > > > > Number of entries: 1
> > > > >
> > > > > Brick gluster2b-1.dns99.co.uk:/data/brick/callrec/
> > > > > Number of entries: 0
> > > > >
> > > > > [root@gluster1b-1 ~]#
> > > > >
> > > > >
> > > > > _______________________________________________
> > > > > Gluster-users mailing list
> > > > > Gluster-users@xxxxxxxxxxx
> > > > > http://www.gluster.org/mailman/listinfo/gluster-users
> > > > >
> > > >
> > > > _______________________________________________
> > > > Gluster-users mailing list
> > > > Gluster-users@xxxxxxxxxxx
> > > > http://www.gluster.org/mailman/listinfo/gluster-users
> > > >
> > >
> > > _______________________________________________
> > > Gluster-users mailing list
> > > Gluster-users@xxxxxxxxxxx
> > > http://www.gluster.org/mailman/listinfo/gluster-users
> >
> >
>
_______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-users