<!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.0 TRANSITIONAL//EN">
<HTML>
<HEAD>
<META HTTP-EQUIV="Content-Type" CONTENT="text/html; CHARSET=UTF-8">
<META NAME="GENERATOR" CONTENT="GtkHTML/3.30.3">
</HEAD>
<BODY>
On Fri, 2012-11-30 at 12:51 -0500, Jeff Darcy wrote:
<BLOCKQUOTE TYPE=CITE>
<PRE>
On 11/28/2012 10:27 AM, Atıf CEYLAN wrote:
> My first question: if GlusterFS was start before than the imap/pop3 server
> can't be map 993 and 995 ports by imap/pop3 server. Because GlusterFS use them.
> I didn't understand why it use these ports?
Like many other system programs, GlusterFS tries to use ports below 1024 which
are supposed to be privileged, hunting downward until it finds one that's
available. If this is a problem for you, I suggest looking into the
"portreserve" command.
> Second, one of two debian was crash and boot up again. When it was start,
> GlusterFS heal process was start. But a few minutes later written below records
> into the log and GlusterFS native client (or FUSE) was crash.
>
> [2012-11-28 12:11:33.763486] E
> [afr-self-heal-data.c:763:afr_sh_data_fxattrop_fstat_done] 0-m3-replicate-0:
> Unable to self-heal contents of
> '/domains/1/abc.com/info/Maildir/dovecot.index.log' (possible split-brain).
> Please delete the file from all but the preferred subvolume.
> [2012-11-28 12:11:33.763659] E
> [afr-self-heal-common.c:2160:afr_self_heal_completion_cbk] 0-m3-replicate-0:
> background meta-data data self-heal failed on
> /domains/1/O/abc.com/info/Maildir/dovecot.index.log
> [2012-11-28 12:11:33.763927] W [afr-open.c:213:afr_open] 0-m3-replicate-0:
> failed to open as split brain seen, returning EIO
> [2012-11-28 12:11:33.763958] W [fuse-bridge.c:1948:fuse_readv_cbk]
> 0-glusterfs-fuse: 432877: READ =-1 (Input/output error)
> [2012-11-28 12:11:33.764039] W [afr-open.c:213:afr_open] 0-m3-replicate-0:
> failed to open as split brain seen, returning EIO
> [2012-11-28 12:11:33.764062] W [fuse-bridge.c:1948:fuse_readv_cbk]
> 0-glusterfs-fuse: 432878: READ =-1 (Input/output error)
> [2012-11-28 12:11:33.764062] W [fuse-bridge.c:1948:fuse_readv_cbk]
> 0-glusterfs-fuse: 432878: READ =-1 (Input/output error)
> [2012-11-28 12:11:36.274580] E
> [afr-self-heal-data.c:763:afr_sh_data_fxattrop_fstat_done] 0-m3-replicate-0:
> Unable to self-heal contents of
> '/domains/xxx.com/info/Maildir/dovecot.index.log' (possible split-brain).
> Please delete the file from all but the preferred subvolume.
> [2012-11-28 12:11:36.274781] E
> [afr-self-heal-common.c:2160:afr_self_heal_completion_cbk] 0-m3-replicate-0:
> background meta-data data self-heal failed on
> /domains/xxx.com/info/Maildir/dovecot.index.log
The phrase "split brain" means that we detected changes to both replicas, and
it would be unsafe to let one override the other (i.e. might lost data) so we
keep our hands off until the user has a chance to intervene. This can happen
in two distinct ways:
* Network partition: client A can only reach replica X, client B can only reach
replica Y, both make changes which end up causing split brain.
* Multiple failures over time. X goes down, changes occur only on Y, then Y
goes down and X comes up (or X comes up and Y goes down before self-heal is
finished) so changes only occur at X.
The quorum feature should address both of these, at the expense of returning
errors if an insufficient number of replicas are available (so it works best
with replica count >= 3).
It's also usually worth figuring out why such problems happened in the first
place. Do you have a lot of network problems or server failures? Are these
servers widely separated? Either is likely to cau        se problems not only with
GlusterFS but with any distributed filesystem, so it's a good idea to address
such issues or at least mention them when reporting problems.
</PRE>
</BLOCKQUOTE>
One of my servers is currently disabled. Because when I run glusterfs service that client(fuse) give "/ mnt/s2/mail/mail1-2: Transport endpoint is not connected" error message. I want to run both servers without shutdown my http and other services. So I can't start the glusterfs service of the crashed server. How I make debugging and fixing the errors without stopping worked services? <BR>
<BR>
There are lots of file consistency errors at gluster.log looks like below.<BR>
<BR>
Shall I move below files to outside of cluster directories and run heal command and I move files again to old directories over client?<BR>
<BR>
[2012-11-28 12:11:33.763486] E [afr-self-heal-data.c:763:afr_sh_data_fxattrop_fstat_done] 0-m3-replicate-0: Unable to self-heal contents of '/domains/1/abc.com/info/Maildir/dovecot.index.log' (possible split-brain). Please delete the file from all but the preferred subvolume.<BR>
[2012-11-28 12:11:33.763659] E [afr-self-heal-common.c:2160:afr_self_heal_completion_cbk] 0-m3-replicate-0: background meta-data data self-heal failed on /domains/1/O/abc.com/info/Maildir/dovecot.index.log<BR>
[2012-11-28 12:11:33.763927] W [afr-open.c:213:afr_open] 0-m3-replicate-0: failed to open as split brain seen, returning EIO<BR>
[2012-11-28 12:11:33.763958] W [fuse-bridge.c:1948:fuse_readv_cbk] 0-glusterfs-fuse: 432877: READ =-1 (Input/output error)<BR>
[2012-11-28 12:11:33.764039] W [afr-open.c:213:afr_open] 0-m3-replicate-0: failed to open as split brain seen, returning EIO<BR>
[2012-11-28 12:11:33.764062] W [fuse-bridge.c:1948:fuse_readv_cbk] 0-glusterfs-fuse: 432878: READ =-1 (Input/output error)<BR>
[2012-11-28 12:11:33.764062] W [fuse-bridge.c:1948:fuse_readv_cbk] 0-glusterfs-fuse: 432878: READ =-1 (Input/output error)<BR>
[2012-11-28 12:11:36.274580] E [afr-self-heal-data.c:763:afr_sh_data_fxattrop_fstat_done] 0-m3-replicate-0: Unable to self-heal contents of '/domains/xxx.com/info/Maildir/dovecot.index.log' (possible split-brain). Please delete the file from all but the preferred subvolume.<BR>
[2012-11-28 12:11:36.274781] E [afr-self-heal-common.c:2160:afr_self_heal_completion_cbk] 0-m3-replicate-0: background meta-data data self-heal failed on /domains/xxx.com/info/Maildir/dovecot.index.log<BR>
<BR>
<TABLE CELLSPACING="0" CELLPADDING="0" WIDTH="100%">
<TR>
<TD>
-- <BR>
<B>M.Atıf CEYLAN</B><BR>
Yurdum Yazılım
</TD>
</TR>
</TABLE>
</BODY>
</HTML>