<div dir="ltr"><div class="" lang="x-western">
  
    
  
  
    <div class="">Hi Shwetha,</div><div class=""><br></div><div class="">







<p class="">[root@nas-01 ~]# getfattr -d -e hex -m . /datastore</p>
<p class="">getfattr: Removing leading &#39;/&#39; from absolute path names</p>
<p class=""># file: datastore</p>
<p class="">trusted.afr.datastore1-client-0=0x000000000000000000000000</p>
<p class="">trusted.afr.datastore1-client-1=0x000000000000000000000000</p>
<p class="">trusted.gfid=0x00000000000000000000000000000001</p>
<p class="">trusted.glusterfs.dht=0x000000010000000000000000ffffffff</p>
<p class="">trusted.glusterfs.volume-id=0xfdff519085ef4cba9056a6bbbd8d6863</p><p class=""><br></p><p class=""><br></p><p class="">[root@nas-02 ~]# getfattr -d -e hex -m . /datastore</p><p class="">getfattr: Removing leading &#39;/&#39; from absolute path names</p>
<p class=""># file: datastore</p><p class="">










</p><p class="">security.selinux=0x73797374656d5f753a6f626a6563745f723a6574635f72756e74696d655f743a733000</p><p class=""><br></p><p class="">I don&#39;t understand what happened .</p><p class=""><br></p><p class="">gr</p>
<p class="">Patrick</p></div><div class=""><br></div><div class="">Hi Gandalf,<br>
      <br>
      can you run the following command on the brick path? <br>
      <br>
      &quot;getfattr -d -e hex -m . /datastore&quot; on both &quot;nas-01-data&quot; and
      &quot;nas-02-data&quot; nodes. <br>
      <br>
      This will let us know whether there is
      &quot;trusted.glusterfs.volume-id&quot; set. <br>
      <br>
      -Shwetha<br>
      <br>
      On 11/26/2013 07:36 PM, gandalf istari wrote:<br>
    </div>
    <blockquote cite="mid:CAFMZTixON=7OkQxyU+QV-jspvLx35BnHSP9mg7BZ_4zzq3pKbQ@mail.gmail.com" type="cite">
      <div dir="ltr">
        <div class="" lang="x-western">
          <div class="">hi thanks for the quick answer.</div>
          <div class=""><br>
          </div>
          <div class="">I&#39;m running glusterfs 3.4.1</div>
          <div class=""><br>
          </div>
          <div class="">[root@nas-02 datastore]# gluster volume start
            datastore1 force<br>
          </div>
          <div class="">
            <p class="">volume start: datastore1: failed: Failed to get
              extended attribute trusted.glusterfs.volume-id for brick
              dir /datastore. Reason : No data available</p>
            <p class="">It seems that the .gluster directory is missing
              for some reason.</p>
            <p class=""><br>
            </p>
            <pre style="white-space:pre-wrap">volume replace-brick datastore1 nas-01-data:/datastore <span style="font-family:arial">nas-02-data:/datastore</span>
commit force</pre>
            <pre style="white-space:pre-wrap"></pre>
            <pre style="white-space:pre-wrap">To rebuild/replace the missing brick ?</pre>
            <pre style="white-space:pre-wrap">I&#39;m quite new with glusterfs</pre>
            <pre style="white-space:pre-wrap"></pre>
            <pre style="white-space:pre-wrap">Thanks </pre>
            <p class=""><br>
            </p>
            <p class=""><br>
            </p>
          </div>
          <div class=""><br>
          </div>
          <div class=""><br>
          </div>
          <div class=""><br>
          </div>
          <div class="">On 26/11/13 12:47, gandalf istari wrote:<br>
          </div>
          <blockquote cite="mid:CAFMZTiwYXv3V69+hryNwRfo=-xmsxwJeHC2XRhXRDgSiVcY7fA@mail.gmail.com" type="cite">
            <div dir="ltr">Hi have setup a two node replication
              glusterfs. After the initial installation the &quot;master&quot;
              node was put into the datacenter and after two week we
              moved the second one also to the datacenter.
              <div><br>
              </div>
              <div>But the sync has not started yet.</div>
              <div><br>
              </div>
              <div>On the &quot;master&quot;</div>
              <div>
                <p class="">gluster&gt; volume info all </p>
                <p class="">Volume Name: datastore1</p>
                <p class="">Type: Replicate</p>
                <p class="">Volume ID:
                  fdff5190-85ef-4cba-9056-a6bbbd8d6863</p>
                <p class="">Status: Started</p>
                <p class="">Number of Bricks: 1 x 2 = 2</p>
                <p class="">Transport-type: tcp</p>
                <p class="">Bricks:</p>
                <p class="">Brick1: nas-01-data:/datastore</p>
                <p class="">Brick2: nas-02-data:/datastore</p>
                <p class="">gluster&gt; peer status</p>
                <p class="">Number of Peers: 1</p>
                <p class=""><br>
                </p>
                <p class="">Hostname: nas-02-data</p>
                <p class="">Uuid: 71df9f86-a87b-481d-896c-c0d4ab679cfa</p>
                <p class=""> </p>
                <p class="">State: Peer in Cluster (Connected)</p>
                <p class=""><br>
                </p>
                <p class="">On the &quot;slave&quot;</p>
                <p class="">gluster&gt; peer status</p>
                <p class="">Number of Peers: 1</p>
                <p class="">Hostname: 192.168.70.6<br>
                </p>
                <p class="">Uuid: 97ef0154-ad7b-402a-b0cb-22be09134a3c</p>
                <p class=""> </p>
                <p class="">State: Peer in Cluster (Connected)</p>
                <p class=""><br>
                </p>
                <p class="">gluster&gt; volume status all</p>
                <p class="">Status of volume: datastore1</p>
                <p class="">Gluster process<span class=""> </span><span class=""> </span><span class=""> </span><span class=""> </span><span class=""> </span><span class=""> </span>Port<span class=""> </span>Online<span class=""> </span>Pid</p>

                <p class="">------------------------------------------------------------------------------</p>
                <p class="">Brick nas-01-data:/datastore<span class="">
                  </span><span class=""> </span><span class=""> </span><span class=""> </span>49152<span class=""> </span>Y<span class=""> </span>2130</p>
                <p class="">Brick nas-02-data:/datastore<span class="">
                  </span><span class=""> </span><span class=""> </span><span class=""> </span>N/A<span class=""> </span>N<span class=""> </span>N/A</p>
                <p class="">NFS Server on localhost<span class=""> </span><span class=""> </span><span class=""> </span><span class=""> </span><span class=""> </span>2049<span class=""> </span>Y<span class=""> </span>8064</p>

                <p class="">Self-heal Daemon on localhost<span class="">
                  </span><span class=""> </span><span class=""> </span><span class=""> </span>N/A<span class=""> </span>Y<span class=""> </span>8073</p>
                <p class="">NFS Server on 192.168.70.6<span class=""> </span><span class=""> </span><span class=""> </span><span class=""> </span>2049<span class=""> </span>Y<span class=""> </span>3379</p>
                <p class="">Self-heal Daemon on 192.168.70.6<span class=""> </span><span class=""> </span><span class=""> </span>N/A<span class=""> </span>Y<span class=""> </span>3384</p>
              </div>
            </div>
          </blockquote>
          Which version of glusterfs are you running?<br>
          <br>
          volume status suggests that the second brick
          (nas-02-data:/datastore) is not running. <br>
          <br>
          Can you run &quot;gluster volume start &lt;volname&gt; force&quot; in
          any of these two nodes and try again? <br>
          Then you would also required to run `find . | xargs stat` on
          the mountpoint of the volume. That should trigger the self
          heal.<br>
          <blockquote cite="mid:CAFMZTiwYXv3V69+hryNwRfo=-xmsxwJeHC2XRhXRDgSiVcY7fA@mail.gmail.com" type="cite">
            <div dir="ltr">
              <div>
                <p class=""> </p>
                <p class=""> </p>
                <p class="">There are no active volume tasks</p>
                <p class=""><br>
                </p>
                <p class="">I would like to run on the &quot;slave&quot; gluster
                  volume sync nas-01-data datastore1</p>
              </div>
            </div>
          </blockquote>
          BTW, There is no concept of &quot;master&quot; and &quot;slave&quot; in afr
          (replication). However there is concept of &quot;master volume&quot; and
          &quot;slave volume&quot; in gluster geo-replication.<br>
          <blockquote cite="mid:CAFMZTiwYXv3V69+hryNwRfo=-xmsxwJeHC2XRhXRDgSiVcY7fA@mail.gmail.com" type="cite">
            <div dir="ltr">
              <div>
                <p class="">But then the virtual machines hosted will be
                  unavailible is there another way to start the
                  replication ?</p>
                <p class=""><br>
                </p>
                <p class="">Thanks</p>
                <p class=""><br>
                </p>
                <p class=""><br>
                </p>
                <p class=""><br>
                </p>
              </div>
            </div>
            <br>
            <fieldset class=""></fieldset>
            <br>
            <pre>_______________________________________________
Gluster-users mailing list
<a class="" href="mailto:Gluster-users@gluster.org">Gluster-users@gluster.org</a>
<a class="" href="http://supercolony.gluster.org/mailman/listinfo/gluster-users">http://supercolony.gluster.org/mailman/listinfo/gluster-users</a></pre>
          </blockquote>
          <br>
        </div>
      </div>
      <br>
      <fieldset class=""></fieldset>
      <br>
      <pre>_______________________________________________
Gluster-users mailing list
<a class="" href="mailto:Gluster-users@gluster.org">Gluster-users@gluster.org</a>
<a class="" href="http://supercolony.gluster.org/mailman/listinfo/gluster-users">http://supercolony.gluster.org/mailman/listinfo/gluster-users</a></pre>
    </blockquote>
    <br>
    <br>
  


</div><div class="" lang="x-western"><br></div></div>