<html>
<head>
<meta content="text/html; charset=UTF-8" http-equiv="Content-Type">
</head>
<body bgcolor="#FFFFFF" text="#000000">
<div class="moz-cite-prefix">To start self-healing data execute :
"gluster volume heal datastore1 full" . <br>
<br>
To monitor self-heal completion status exeucte : "gluster volume
heal datastore1 info". The number of entries under each brick
should be 0. When the number of entries count becomes 0 the
self-heal is completed. <br>
<br>
Other way to trigger self-heal is to create a new fuse mount and
execute "find . | xargs stat" from the mount point. <br>
<br>
-Shwetha <br>
On 11/27/2013 01:07 PM, gandalf istari wrote:<br>
</div>
<blockquote
cite="mid:CAFMZTiyq3Bmr9y6DUWMGo0F2JMGK4a8v8t87BtXt6H+X1TAFAw@mail.gmail.com"
type="cite">
<div dir="ltr">
<div class="" lang="x-unicode">
<div class="">Thank you so much,</div>
<div class=""><br>
</div>
<div class="">It seems to be working the directories are now
created but still empty. I suppose this will take a while to
sync 44GB</div>
<div class=""><br>
</div>
<div class="">The only change I have done was: leave out
nas-02-data for your command.</div>
<div class=""><br>
</div>
<div class="">whats the best way now to monitor the sync
process ?</div>
<div class=""><br>
</div>
<div class="">
<br>
</div>
<div class=""><br>
</div>
<div class="">You couldn't force start the volume because the
brick "nas-02-data:/datastore" doesn't have the
"trusted.glusterfs.volume-id" . <br>
<br>
From nas-02 node execute : <br>
<br>
1. setfattr -n trusted.glusterfs.volume-id -v
0xfdff519085ef4cba9056a6bbbd8d6863 nas-02-data:/datastore <br>
<br>
2. gluster volume start datastore1 force. <br>
<br>
<br>
-Shwetha<br>
On 11/27/2013 12:23 PM, gandalf istari wrote:<br>
</div>
<blockquote
cite="mid:CAFMZTiy9kk2KW3mE0LSx=gNexre2Lb5nd1EBaQYdCqDcEMHGEw@mail.gmail.com"
type="cite">
<div dir="ltr">
<div class="" lang="x-western">
<div class="">Hi Shwetha,</div>
<div class=""><br>
</div>
<div class="">
<p class="">[root@nas-01 ~]# getfattr -d -e hex -m .
/datastore</p>
<p class="">getfattr: Removing leading '/' from
absolute path names</p>
<p class=""># file: datastore</p>
<p class="">trusted.afr.datastore1-client-0=0x000000000000000000000000</p>
<p class="">trusted.afr.datastore1-client-1=0x000000000000000000000000</p>
<p class="">trusted.gfid=0x00000000000000000000000000000001</p>
<p class="">trusted.glusterfs.dht=0x000000010000000000000000ffffffff</p>
<p class="">trusted.glusterfs.volume-id=0xfdff519085ef4cba9056a6bbbd8d6863</p>
<p class=""><br>
</p>
<p class=""><br>
</p>
<p class="">[root@nas-02 ~]# getfattr -d -e hex -m .
/datastore</p>
<p class="">getfattr: Removing leading '/' from
absolute path names</p>
<p class=""># file: datastore</p>
<p class=""> </p>
<p class="">security.selinux=0x73797374656d5f753a6f626a6563745f723a6574635f72756e74696d655f743a733000</p>
</div>
</div>
</div>
</blockquote>
<blockquote
cite="mid:CAFMZTiy9kk2KW3mE0LSx=gNexre2Lb5nd1EBaQYdCqDcEMHGEw@mail.gmail.com"
type="cite">
<div dir="ltr">
<div class="" lang="x-western">
<div class="">
<p class="">I don't understand what happened .</p>
<p class=""><br>
</p>
<p class="">gr</p>
<p class="">Patrick</p>
</div>
<div class=""><br>
</div>
<div class="">Hi Gandalf,<br>
<br>
can you run the following command on the brick path? <br>
<br>
"getfattr -d -e hex -m . /datastore" on both
"nas-01-data" and "nas-02-data" nodes. <br>
<br>
This will let us know whether there is
"trusted.glusterfs.volume-id" set. <br>
<br>
-Shwetha<br>
<br>
On 11/26/2013 07:36 PM, gandalf istari wrote:<br>
</div>
<blockquote
cite="mid:CAFMZTixON=7OkQxyU+QV-jspvLx35BnHSP9mg7BZ_4zzq3pKbQ@mail.gmail.com"
type="cite">
<div dir="ltr">
<div class="" lang="x-western">
<div class="">hi thanks for the quick answer.</div>
<div class=""><br>
</div>
<div class="">I'm running glusterfs 3.4.1</div>
<div class=""><br>
</div>
<div class="">[root@nas-02 datastore]# gluster
volume start datastore1 force<br>
</div>
<div class="">
<p class="">volume start: datastore1: failed:
Failed to get extended attribute
trusted.glusterfs.volume-id for brick dir
/datastore. Reason : No data available</p>
<p class="">It seems that the .gluster directory
is missing for some reason.</p>
<p class=""><br>
</p>
<pre style="white-space:pre-wrap">volume replace-brick datastore1 nas-01-data:/datastore <span style="font-family:arial">nas-02-data:/datastore</span>
commit force</pre>
<pre style="white-space:pre-wrap">To rebuild/replace the missing brick ?</pre>
<pre style="white-space:pre-wrap">I'm quite new with glusterfs</pre>
<pre style="white-space:pre-wrap">Thanks </pre>
<p class=""><br>
</p>
<p class=""><br>
</p>
</div>
<div class=""><br>
</div>
<div class=""><br>
</div>
<div class=""><br>
</div>
<div class="">On 26/11/13 12:47, gandalf istari
wrote:<br>
</div>
<blockquote
cite="mid:CAFMZTiwYXv3V69+hryNwRfo=-xmsxwJeHC2XRhXRDgSiVcY7fA@mail.gmail.com"
type="cite">
<div dir="ltr">Hi have setup a two node
replication glusterfs. After the initial
installation the "master" node was put into
the datacenter and after two week we moved the
second one also to the datacenter.
<div><br>
</div>
<div>But the sync has not started yet.</div>
<div><br>
</div>
<div>On the "master"</div>
<div>
<p class="">gluster> volume info all </p>
<p class="">Volume Name: datastore1</p>
<p class="">Type: Replicate</p>
<p class="">Volume ID:
fdff5190-85ef-4cba-9056-a6bbbd8d6863</p>
<p class="">Status: Started</p>
<p class="">Number of Bricks: 1 x 2 = 2</p>
<p class="">Transport-type: tcp</p>
<p class="">Bricks:</p>
<p class="">Brick1: nas-01-data:/datastore</p>
<p class="">Brick2: nas-02-data:/datastore</p>
<p class="">gluster> peer status</p>
<p class="">Number of Peers: 1</p>
<p class=""><br>
</p>
<p class="">Hostname: nas-02-data</p>
<p class="">Uuid:
71df9f86-a87b-481d-896c-c0d4ab679cfa</p>
<p class=""> </p>
<p class="">State: Peer in Cluster
(Connected)</p>
<p class=""><br>
</p>
<p class="">On the "slave"</p>
<p class="">gluster> peer status</p>
<p class="">Number of Peers: 1</p>
<p class="">Hostname: 192.168.70.6<br>
</p>
<p class="">Uuid:
97ef0154-ad7b-402a-b0cb-22be09134a3c</p>
<p class=""> </p>
<p class="">State: Peer in Cluster
(Connected)</p>
<p class=""><br>
</p>
<p class="">gluster> volume status all</p>
<p class="">Status of volume: datastore1</p>
<p class="">Gluster process<span class=""> </span><span
class=""> </span><span class=""> </span><span
class=""> </span><span class=""> </span><span
class=""> </span>Port<span class=""> </span>Online<span
class=""> </span>Pid</p>
<p class="">------------------------------------------------------------------------------</p>
<p class="">Brick nas-01-data:/datastore<span
class=""> </span><span class=""> </span><span
class=""> </span><span class=""> </span>49152<span
class=""> </span>Y<span class=""> </span>2130</p>
<p class="">Brick nas-02-data:/datastore<span
class=""> </span><span class=""> </span><span
class=""> </span><span class=""> </span>N/A<span
class=""> </span>N<span class=""> </span>N/A</p>
<p class="">NFS Server on localhost<span
class=""> </span><span class=""> </span><span
class=""> </span><span class=""> </span><span
class=""> </span>2049<span class=""> </span>Y<span
class=""> </span>8064</p>
<p class="">Self-heal Daemon on localhost<span
class=""> </span><span class=""> </span><span
class=""> </span><span class=""> </span>N/A<span
class=""> </span>Y<span class=""> </span>8073</p>
<p class="">NFS Server on 192.168.70.6<span
class=""> </span><span class=""> </span><span
class=""> </span><span class=""> </span>2049<span
class=""> </span>Y<span class=""> </span>3379</p>
<p class="">Self-heal Daemon on 192.168.70.6<span
class=""> </span><span class=""> </span><span
class=""> </span>N/A<span class=""> </span>Y<span
class=""> </span>3384</p>
</div>
</div>
</blockquote>
Which version of glusterfs are you running?<br>
<br>
volume status suggests that the second brick
(nas-02-data:/datastore) is not running. <br>
<br>
Can you run "gluster volume start <volname>
force" in any of these two nodes and try again? <br>
Then you would also required to run `find . |
xargs stat` on the mountpoint of the volume. That
should trigger the self heal.<br>
<blockquote
cite="mid:CAFMZTiwYXv3V69+hryNwRfo=-xmsxwJeHC2XRhXRDgSiVcY7fA@mail.gmail.com"
type="cite">
<div dir="ltr">
<div>
<p class=""> </p>
<p class=""> </p>
<p class="">There are no active volume tasks</p>
<p class=""><br>
</p>
<p class="">I would like to run on the
"slave" gluster volume sync nas-01-data
datastore1</p>
</div>
</div>
</blockquote>
BTW, There is no concept of "master" and "slave"
in afr (replication). However there is concept of
"master volume" and "slave volume" in gluster
geo-replication.<br>
<blockquote
cite="mid:CAFMZTiwYXv3V69+hryNwRfo=-xmsxwJeHC2XRhXRDgSiVcY7fA@mail.gmail.com"
type="cite">
<div dir="ltr">
<div>
<p class="">But then the virtual machines
hosted will be unavailible is there
another way to start the replication ?</p>
<p class=""><br>
</p>
<p class="">Thanks</p>
<p class=""><br>
</p>
<p class=""><br>
</p>
<p class=""><br>
</p>
</div>
</div>
<br>
<fieldset class=""></fieldset>
<br>
<pre>_______________________________________________
Gluster-users mailing list
<a moz-do-not-send="true" class="" href="mailto:Gluster-users@gluster.org">Gluster-users@gluster.org</a>
<a moz-do-not-send="true" class="" href="http://supercolony.gluster.org/mailman/listinfo/gluster-users">http://supercolony.gluster.org/mailman/listinfo/gluster-users</a></pre>
</blockquote>
<br>
</div>
</div>
<br>
<fieldset class=""></fieldset>
<br>
<pre>_______________________________________________
Gluster-users mailing list
<a moz-do-not-send="true" class="" href="mailto:Gluster-users@gluster.org">Gluster-users@gluster.org</a>
<a moz-do-not-send="true" class="" href="http://supercolony.gluster.org/mailman/listinfo/gluster-users">http://supercolony.gluster.org/mailman/listinfo/gluster-users</a></pre>
</blockquote>
<br>
<br>
</div>
<div class="" lang="x-western"><br>
</div>
</div>
</blockquote>
<br>
<br>
</div>
</div>
</blockquote>
<br>
<br>
</body>
</html>