<div dir="ltr">Thanks! I turned off drc as suggested and will have to wait and see how that works. Here are the packages I have installed via yum:<div># rpm -qa |grep -i gluster</div><div>glusterfs-cli-3.5.0-2.el6.x86_64</div>
<div>glusterfs-libs-3.5.0-2.el6.x86_64</div><div>glusterfs-fuse-3.5.0-2.el6.x86_64</div><div>glusterfs-server-3.5.0-2.el6.x86_64</div><div>glusterfs-3.5.0-2.el6.x86_64</div><div>glusterfs-geo-replication-3.5.0-2.el6.x86_64</div>
<div><br></div><div>The nfs server service was showing to be running even when stuff wasn't working. This is from while it was broken:<br></div><div><br></div><div><div># gluster volume status</div><div>Status of volume: gv0</div>
<div>Gluster process Port Online Pid</div><div>------------------------------------------------------------------------------------------------------------</div><div>
Brick eapps-gluster01.my.domain:/export/sdb1/gv0 49152 Y 39593</div><div>Brick eapps-gluster02.my.domain:/export/sdb1/gv0 49152 Y 2472</div><div>Brick eapps-gluster03.my.domain:/export/sdb1/gv0 49152 Y 1866</div>
<div>NFS Server on localhost 2049 Y 39603</div><div>Self-heal Daemon on localhost N/A Y 39610</div><div>NFS Server on eapps-gluster03.my.domain 2049 Y 35125</div>
<div>Self-heal Daemon on eapps-gluster03.my.domain N/A Y 35132</div><div>NFS Server on eapps-gluster02.my.domain 2049 Y 37103</div><div>Self-heal Daemon on eapps-gluster02.my.domain N/A Y 37110</div>
<div><br></div><div>Task Status of Volume gv0</div><div>---------------------------------------------------------------------------------------------------------------</div></div><div><br></div><div><br></div><div><div>Running 'service glusterd restart' on the NFS server made things start working again after this.</div>
</div><div><br></div><div class="gmail_extra"><br></div><div class="gmail_extra">-- Gene<br><div><div dir="ltr"><div style="font-family:tahoma,sans-serif"><br></div></div></div>
<br><br><div class="gmail_quote">On Tue, Jun 10, 2014 at 12:10 PM, Niels de Vos <span dir="ltr"><<a href="mailto:ndevos@redhat.com" target="_blank">ndevos@redhat.com</a>></span> wrote:<br><blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex">
<div class="">On Tue, Jun 10, 2014 at 11:32:50AM -0400, Gene Liverman wrote:<br>
> Twice now I have had my nfs connection to a replicated gluster volume stop<br>
> responding. On both servers that connect to the system I have the following<br>
> symptoms:<br>
><br>
</div>> 1. Accessing the mount with the native client is still working fine (the<br>
<div class="">> volume is mounted both that way and via nfs. One app requires the nfs<br>
> version)<br>
</div>> 2. The logs have messages stating the following: "kernel: nfs: server<br>
<div class="">> my-servers-name not responding, still trying"<br>
><br>
> How can I fix this?<br>
<br>
</div>You should check if the NFS-server (a glusterfs process) is still<br>
running:<br>
<br>
# gluster volume status<br>
<br>
If the NFS-server is not running anymore, you can start it with:<br>
<br>
# gluster volume start $VOLUME force<br>
(you only need to do that for one volume)<br>
<br>
<br>
In case this is with GlusterFS 3.5, you may be hitting a memory leak in<br>
the DRC (Duplicate Request Cache) implementation of the NFS-server. You<br>
can disable DRC with this:<br>
<br>
# gluster volume set $VOLUME nfs.drc off<br>
<br>
In glusterfs-3.5.1 DRC will be disabled by default, there have been too<br>
many issues with DRC to enable it for everyone. We need to do more tests<br>
and fix DRC in the current development (master) branch.<br>
<br>
HTH,<br>
Niels<br>
</blockquote></div><br></div></div>