On Fri, 2012-11-30 at 12:51 -0500, Jeff Darcy wrote: > On 11/28/2012 10:27 AM, At?f CEYLAN wrote: > > My first question: if GlusterFS was start before than the imap/pop3 server > > can't be map 993 and 995 ports by imap/pop3 server. Because GlusterFS use them. > > I didn't understand why it use these ports? > > Like many other system programs, GlusterFS tries to use ports below 1024 which > are supposed to be privileged, hunting downward until it finds one that's > available. If this is a problem for you, I suggest looking into the > "portreserve" command. > > > Second, one of two debian was crash and boot up again. When it was start, > > GlusterFS heal process was start. But a few minutes later written below records > > into the log and GlusterFS native client (or FUSE) was crash. > > > > [2012-11-28 12:11:33.763486] E > > [afr-self-heal-data.c:763:afr_sh_data_fxattrop_fstat_done] 0-m3-replicate-0: > > Unable to self-heal contents of > > '/domains/1/abc.com/info/Maildir/dovecot.index.log' (possible split-brain). > > Please delete the file from all but the preferred subvolume. > > [2012-11-28 12:11:33.763659] E > > [afr-self-heal-common.c:2160:afr_self_heal_completion_cbk] 0-m3-replicate-0: > > background meta-data data self-heal failed on > > /domains/1/O/abc.com/info/Maildir/dovecot.index.log > > [2012-11-28 12:11:33.763927] W [afr-open.c:213:afr_open] 0-m3-replicate-0: > > failed to open as split brain seen, returning EIO > > [2012-11-28 12:11:33.763958] W [fuse-bridge.c:1948:fuse_readv_cbk] > > 0-glusterfs-fuse: 432877: READ =-1 (Input/output error) > > [2012-11-28 12:11:33.764039] W [afr-open.c:213:afr_open] 0-m3-replicate-0: > > failed to open as split brain seen, returning EIO > > [2012-11-28 12:11:33.764062] W [fuse-bridge.c:1948:fuse_readv_cbk] > > 0-glusterfs-fuse: 432878: READ =-1 (Input/output error) > > [2012-11-28 12:11:33.764062] W [fuse-bridge.c:1948:fuse_readv_cbk] > > 0-glusterfs-fuse: 432878: READ =-1 (Input/output error) > > [2012-11-28 12:11:36.274580] E > > [afr-self-heal-data.c:763:afr_sh_data_fxattrop_fstat_done] 0-m3-replicate-0: > > Unable to self-heal contents of > > '/domains/xxx.com/info/Maildir/dovecot.index.log' (possible split-brain). > > Please delete the file from all but the preferred subvolume. > > [2012-11-28 12:11:36.274781] E > > [afr-self-heal-common.c:2160:afr_self_heal_completion_cbk] 0-m3-replicate-0: > > background meta-data data self-heal failed on > > /domains/xxx.com/info/Maildir/dovecot.index.log > > The phrase "split brain" means that we detected changes to both replicas, and > it would be unsafe to let one override the other (i.e. might lost data) so we > keep our hands off until the user has a chance to intervene. This can happen > in two distinct ways: > > * Network partition: client A can only reach replica X, client B can only reach > replica Y, both make changes which end up causing split brain. > > * Multiple failures over time. X goes down, changes occur only on Y, then Y > goes down and X comes up (or X comes up and Y goes down before self-heal is > finished) so changes only occur at X. > > The quorum feature should address both of these, at the expense of returning > errors if an insufficient number of replicas are available (so it works best > with replica count >= 3). > > It's also usually worth figuring out why such problems happened in the first > place. Do you have a lot of network problems or server failures? Are these > servers widely separated? Either is likely to cau se problems not only with > GlusterFS but with any distributed filesystem, so it's a good idea to address > such issues or at least mention them when reporting problems. One of my servers is currently disabled. Because when I run glusterfs service that client(fuse) give "/ mnt/s2/mail/mail1-2: Transport endpoint is not connected" error message. I want to run both servers without shutdown my http and other services. So I can't start the glusterfs service of the crashed server. How I make debugging and fixing the errors without stopping worked services? There are lots of file consistency errors at gluster.log looks like below. Shall I move below files to outside of cluster directories and run heal command and I move files again to old directories over client? [2012-11-28 12:11:33.763486] E [afr-self-heal-data.c:763:afr_sh_data_fxattrop_fstat_done] 0-m3-replicate-0: Unable to self-heal contents of '/domains/1/abc.com/info/Maildir/dovecot.index.log' (possible split-brain). Please delete the file from all but the preferred subvolume. [2012-11-28 12:11:33.763659] E [afr-self-heal-common.c:2160:afr_self_heal_completion_cbk] 0-m3-replicate-0: background meta-data data self-heal failed on /domains/1/O/abc.com/info/Maildir/dovecot.index.log [2012-11-28 12:11:33.763927] W [afr-open.c:213:afr_open] 0-m3-replicate-0: failed to open as split brain seen, returning EIO [2012-11-28 12:11:33.763958] W [fuse-bridge.c:1948:fuse_readv_cbk] 0-glusterfs-fuse: 432877: READ =-1 (Input/output error) [2012-11-28 12:11:33.764039] W [afr-open.c:213:afr_open] 0-m3-replicate-0: failed to open as split brain seen, returning EIO [2012-11-28 12:11:33.764062] W [fuse-bridge.c:1948:fuse_readv_cbk] 0-glusterfs-fuse: 432878: READ =-1 (Input/output error) [2012-11-28 12:11:33.764062] W [fuse-bridge.c:1948:fuse_readv_cbk] 0-glusterfs-fuse: 432878: READ =-1 (Input/output error) [2012-11-28 12:11:36.274580] E [afr-self-heal-data.c:763:afr_sh_data_fxattrop_fstat_done] 0-m3-replicate-0: Unable to self-heal contents of '/domains/xxx.com/info/Maildir/dovecot.index.log' (possible split-brain). Please delete the file from all but the preferred subvolume. [2012-11-28 12:11:36.274781] E [afr-self-heal-common.c:2160:afr_self_heal_completion_cbk] 0-m3-replicate-0: background meta-data data self-heal failed on /domains/xxx.com/info/Maildir/dovecot.index.log -- M.At?f CEYLAN Yurdum Yaz?l?m -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20121201/8b5b6f58/attachment.html>