<html>
<head>
<meta content="text/html; charset=UTF-8" http-equiv="Content-Type">
</head>
<body text="#000000" bgcolor="#FFFFFF">
<br>
<div class="moz-cite-prefix">On 11/04/2014 03:20 PM, Roman wrote:<br>
</div>
<blockquote
cite="mid:CAFR=TBr=+L6xuL8w0kQ0Z1vh+FR9ng1222fNEC10DwJGWsRQHw@mail.gmail.com"
type="cite">
<div dir="ltr">Hello,
<div><br>
</div>
<div>some news on this?</div>
</div>
</blockquote>
I thought it is working fine for you, after setting the option? The
delay is happening because fsync takes too long to complete on the
brick.<br>
<br>
Pranith<br>
<blockquote
cite="mid:CAFR=TBr=+L6xuL8w0kQ0Z1vh+FR9ng1222fNEC10DwJGWsRQHw@mail.gmail.com"
type="cite">
<div class="gmail_extra"><br>
<div class="gmail_quote">2014-10-21 12:38 GMT+03:00 Roman <span
dir="ltr"><<a moz-do-not-send="true"
href="mailto:romeo.r@gmail.com" target="_blank">romeo.r@gmail.com</a>></span>:<br>
<blockquote class="gmail_quote" style="margin:0 0 0
.8ex;border-left:1px #ccc solid;padding-left:1ex">
<div dir="ltr">
<div>root@stor2:~# gluster volume info HA-testvol-1T</div>
<div><br>
</div>
<div>Volume Name: HA-testvol-1T</div>
<div>Type: Replicate</div>
<div>Volume ID: 224a294c-d22a-4b63-8441-2b293527cbab</div>
<div>Status: Started</div>
<div>Number of Bricks: 1 x 2 = 2</div>
<div>Transport-type: tcp</div>
<div>Bricks:</div>
<div>Brick1: stor1:/exports/testvol/1T</div>
<div>Brick2: stor2:/exports/testvol/1T</div>
<div>Options Reconfigured:</div>
<div>diagnostics.count-fop-hits: on</div>
<div>diagnostics.latency-measurement: on</div>
<div>cluster.ensure-durability: off</div>
<div>nfs.disable: 1</div>
<div>network.ping-timeout: 10</div>
<div><br>
</div>
<div><br>
</div>
<div>and this of course.</div>
</div>
<div class="gmail_extra"><br>
<div class="gmail_quote">2014-10-21 12:37 GMT+03:00 Roman
<span dir="ltr"><<a moz-do-not-send="true"
href="mailto:romeo.r@gmail.com" target="_blank">romeo.r@gmail.com</a>></span>:<br>
<blockquote class="gmail_quote" style="margin:0 0 0
.8ex;border-left:1px #ccc solid;padding-left:1ex">
<div dir="ltr">Hi,
<div>well, this time it was a lot faster and no
warnings from VM-s</div>
<div><br>
</div>
<div>
<div>root@glstor-cli:/srv/gfs/HA-testvol-1T# dd
if=/dev/zero of=900G-disk bs=2G count=450
iflag=fullblock</div>
<div>450+0 records in</div>
<div>450+0 records out</div>
<div>966367641600 bytes (966 GB) copied, 5292.48
s, 183 MB/s</div>
</div>
<div><br>
</div>
<div>
<div>root@stor2:~# gluster volume profile
HA-testvol-1T info</div>
<div>Brick: stor2:/exports/testvol/1T</div>
<div>--------------------------------</div>
<div>Cumulative Stats:</div>
<div> Block Size: 4b+
4096b+ 65536b+</div>
<div> No. of Reads: 0
0 0</div>
<div>No. of Writes: 1
4 5</div>
<div><br>
</div>
<div> Block Size: 131072b+</div>
<div> No. of Reads: 0</div>
<div>No. of Writes: 7454715</div>
<div> %-latency Avg-latency Min-Latency
Max-Latency No. of calls Fop</div>
<div> --------- ----------- -----------
----------- ------------ ----</div>
<div> 0.00 0.00 us 0.00 us
0.00 us 10 FORGET</div>
<div> 0.00 0.00 us 0.00 us
0.00 us 29 RELEASE</div>
<div> 0.00 0.00 us 0.00 us
0.00 us 70 RELEASEDIR</div>
<div> 0.00 115.00 us 115.00 us
115.00 us 1 SETXATTR</div>
<div> 0.00 81.50 us 62.00 us
101.00 us 2 SETATTR</div>
<div> 0.00 57.44 us 27.00 us
108.00 us 9 STATFS</div>
<div> 0.00 32.22 us 10.00 us
76.00 us 18 FSTAT</div>
<div> 0.00 40.73 us 13.00 us
64.00 us 15 FLUSH</div>
<div> 0.00 48.22 us 19.00 us
166.00 us 18 OPEN</div>
<div> 0.00 38.24 us 18.00 us
87.00 us 29 GETXATTR</div>
<div> 0.00 43.29 us 22.00 us
146.00 us 42 ENTRYLK</div>
<div> 0.00 46.10 us 22.00 us
81.00 us 40 READDIR</div>
<div> 0.00 226.36 us 184.00 us
282.00 us 11 CREATE</div>
<div> 0.00 68.77 us 27.00 us
119.00 us 70 OPENDIR</div>
<div> 0.00 6378.00 us 6378.00 us
6378.00 us 1 FSYNC</div>
<div> 0.00 149.32 us 22.00 us
375.00 us 84 READDIRP</div>
<div> 0.00 32.07 us 8.00 us
218.00 us 1094 FINODELK</div>
<div> 0.03 7272.05 us 12.00 us
699192.00 us 97 INODELK</div>
<div> 0.06 4326.87 us 20.00 us
834845.00 us 294 LOOKUP</div>
<div> 0.07 150574.00 us 74.00 us
1504668.00 us 10 UNLINK</div>
<div> 1.70 51392.90 us 20.00 us
5756079.00 us 704 FXATTROP</div>
<div> 98.13 280.28 us 71.00 us
2507482.00 us 7454725 WRITE</div>
<div><br>
</div>
<div> Duration: 11974 seconds</div>
<div> Data Read: 0 bytes</div>
<div>Data Written: 977105055751 bytes</div>
<div><br>
</div>
<div>Interval 1 Stats:</div>
<div> Block Size: 4b+
4096b+ 65536b+</div>
<div> No. of Reads: 0
0 0</div>
<div>No. of Writes: 1
4 5</div>
<div><br>
</div>
<div> Block Size: 131072b+</div>
<div> No. of Reads: 0</div>
<div>No. of Writes: 7454715</div>
<div> %-latency Avg-latency Min-Latency
Max-Latency No. of calls Fop</div>
<div> --------- ----------- -----------
----------- ------------ ----</div>
<div> 0.00 0.00 us 0.00 us
0.00 us 10 FORGET</div>
<div> 0.00 0.00 us 0.00 us
0.00 us 29 RELEASE</div>
<div> 0.00 0.00 us 0.00 us
0.00 us 70 RELEASEDIR</div>
<div> 0.00 115.00 us 115.00 us
115.00 us 1 SETXATTR</div>
<div> 0.00 81.50 us 62.00 us
101.00 us 2 SETATTR</div>
<div> 0.00 57.44 us 27.00 us
108.00 us 9 STATFS</div>
<div> 0.00 32.22 us 10.00 us
76.00 us 18 FSTAT</div>
<div> 0.00 40.73 us 13.00 us
64.00 us 15 FLUSH</div>
<div> 0.00 48.22 us 19.00 us
166.00 us 18 OPEN</div>
<div> 0.00 38.24 us 18.00 us
87.00 us 29 GETXATTR</div>
<div> 0.00 43.29 us 22.00 us
146.00 us 42 ENTRYLK</div>
<div> 0.00 46.10 us 22.00 us
81.00 us 40 READDIR</div>
<div> 0.00 226.36 us 184.00 us
282.00 us 11 CREATE</div>
<div> 0.00 68.77 us 27.00 us
119.00 us 70 OPENDIR</div>
<div> 0.00 6378.00 us 6378.00 us
6378.00 us 1 FSYNC</div>
<div> 0.00 149.32 us 22.00 us
375.00 us 84 READDIRP</div>
<div> 0.00 32.07 us 8.00 us
218.00 us 1094 FINODELK</div>
<div> 0.03 7272.05 us 12.00 us
699192.00 us 97 INODELK</div>
<div> 0.06 4326.87 us 20.00 us
834845.00 us 294 LOOKUP</div>
<div> 0.07 150574.00 us 74.00 us
1504668.00 us 10 UNLINK</div>
<div> 1.70 51392.90 us 20.00 us
5756079.00 us 704 FXATTROP</div>
<div> 98.13 280.28 us 71.00 us
2507482.00 us 7454725 WRITE</div>
<div><br>
</div>
<div> Duration: 11948 seconds</div>
<div> Data Read: 0 bytes</div>
<div>Data Written: 977105055751 bytes</div>
<div><br>
</div>
<div>Brick: stor1:/exports/testvol/1T</div>
<div>--------------------------------</div>
<div>Cumulative Stats:</div>
<div> Block Size: 4b+
4096b+ 65536b+</div>
<div> No. of Reads: 1
0 0</div>
<div>No. of Writes: 1
4 5</div>
<div><br>
</div>
<div> Block Size: 131072b+</div>
<div> No. of Reads: 0</div>
<div>No. of Writes: 7454715</div>
<div> %-latency Avg-latency Min-Latency
Max-Latency No. of calls Fop</div>
<div> --------- ----------- -----------
----------- ------------ ----</div>
<div> 0.00 0.00 us 0.00 us
0.00 us 10 FORGET</div>
<div> 0.00 0.00 us 0.00 us
0.00 us 29 RELEASE</div>
<div> 0.00 0.00 us 0.00 us
0.00 us 70 RELEASEDIR</div>
<div> 0.00 37.00 us 37.00 us
37.00 us 1 STAT</div>
<div> 0.00 140.00 us 140.00 us
140.00 us 1 READ</div>
<div> 0.00 151.00 us 151.00 us
151.00 us 1 SETXATTR</div>
<div> 0.00 87.00 us 76.00 us
98.00 us 2 SETATTR</div>
<div> 0.00 33.13 us 13.00 us
65.00 us 15 FLUSH</div>
<div> 0.00 28.83 us 12.00 us
97.00 us 18 FSTAT</div>
<div> 0.00 66.00 us 26.00 us
104.00 us 9 STATFS</div>
<div> 0.00 55.33 us 19.00 us
105.00 us 18 OPEN</div>
<div> 0.00 41.55 us 23.00 us
91.00 us 29 GETXATTR</div>
<div> 0.00 50.52 us 22.00 us
80.00 us 40 READDIR</div>
<div> 0.00 53.12 us 22.00 us
149.00 us 42 ENTRYLK</div>
<div> 0.00 238.00 us 187.00 us
283.00 us 11 CREATE</div>
<div> 0.00 74.44 us 22.00 us
121.00 us 70 OPENDIR</div>
<div> 0.00 6660.00 us 6660.00 us
6660.00 us 1 FSYNC</div>
<div> 0.00 98.08 us 20.00 us
281.00 us 294 LOOKUP</div>
<div> 0.06 143484.30 us 80.00 us
1433636.00 us 10 UNLINK</div>
<div> 0.46 117012.34 us 14.00 us
4468461.00 us 97 INODELK</div>
<div> 1.98 74942.61 us 20.00 us
5196144.00 us 646 FXATTROP</div>
<div> 3.68 83834.87 us 10.00 us
4469758.00 us 1072 FINODELK</div>
<div> 93.82 307.76 us 62.00 us
2507005.00 us 7454725 WRITE</div>
<div><br>
</div>
<div> Duration: 11972 seconds</div>
<div> Data Read: 7 bytes</div>
<div>Data Written: 977105055751 bytes</div>
<div><br>
</div>
<div>Interval 1 Stats:</div>
<div> Block Size: 4b+
4096b+ 65536b+</div>
<div> No. of Reads: 1
0 0</div>
<div>No. of Writes: 1
4 5</div>
<div><br>
</div>
<div> Block Size: 131072b+</div>
<div> No. of Reads: 0</div>
<div>No. of Writes: 7454715</div>
<div> %-latency Avg-latency Min-Latency
Max-Latency No. of calls Fop</div>
<div> --------- ----------- -----------
----------- ------------ ----</div>
<div> 0.00 0.00 us 0.00 us
0.00 us 10 FORGET</div>
<div> 0.00 0.00 us 0.00 us
0.00 us 29 RELEASE</div>
<div> 0.00 0.00 us 0.00 us
0.00 us 70 RELEASEDIR</div>
<div> 0.00 37.00 us 37.00 us
37.00 us 1 STAT</div>
<div> 0.00 140.00 us 140.00 us
140.00 us 1 READ</div>
<div> 0.00 151.00 us 151.00 us
151.00 us 1 SETXATTR</div>
<div> 0.00 87.00 us 76.00 us
98.00 us 2 SETATTR</div>
<div> 0.00 33.13 us 13.00 us
65.00 us 15 FLUSH</div>
<div> 0.00 28.83 us 12.00 us
97.00 us 18 FSTAT</div>
<div> 0.00 66.00 us 26.00 us
104.00 us 9 STATFS</div>
<div> 0.00 55.33 us 19.00 us
105.00 us 18 OPEN</div>
<div> 0.00 41.55 us 23.00 us
91.00 us 29 GETXATTR</div>
<div> 0.00 50.52 us 22.00 us
80.00 us 40 READDIR</div>
<div> 0.00 53.12 us 22.00 us
149.00 us 42 ENTRYLK</div>
<div> 0.00 238.00 us 187.00 us
283.00 us 11 CREATE</div>
<div> 0.00 74.44 us 22.00 us
121.00 us 70 OPENDIR</div>
<div> 0.00 6660.00 us 6660.00 us
6660.00 us 1 FSYNC</div>
<div> 0.00 98.08 us 20.00 us
281.00 us 294 LOOKUP</div>
<div> 0.06 143484.30 us 80.00 us
1433636.00 us 10 UNLINK</div>
<div> 0.46 117012.34 us 14.00 us
4468461.00 us 97 INODELK</div>
<div> 1.98 74942.61 us 20.00 us
5196144.00 us 646 FXATTROP</div>
<div> 3.68 83834.87 us 10.00 us
4469758.00 us 1072 FINODELK</div>
<div> 93.82 307.76 us 62.00 us
2507005.00 us 7454725 WRITE</div>
<div><br>
</div>
<div> Duration: 11948 seconds</div>
<div> Data Read: 7 bytes</div>
<div>Data Written: 977105055751 bytes</div>
</div>
<div><br>
</div>
<div>got something interesting? :)</div>
</div>
<div class="gmail_extra"><br>
<div class="gmail_quote">2014-10-21 9:21 GMT+03:00
Roman <span dir="ltr"><<a
moz-do-not-send="true"
href="mailto:romeo.r@gmail.com"
target="_blank">romeo.r@gmail.com</a>></span>:<br>
<blockquote class="gmail_quote" style="margin:0 0
0 .8ex;border-left:1px #ccc
solid;padding-left:1ex">
<div dir="ltr">Hi,
<div><br>
</div>
<div>I'm sorry, I had no time to make tests
yesterday. Am starting to do them now. So
soon I'll reply.</div>
</div>
<div class="gmail_extra"><br>
<div class="gmail_quote">2014-10-18 19:29
GMT+03:00 Roman <span dir="ltr"><<a
moz-do-not-send="true"
href="mailto:romeo.r@gmail.com"
target="_blank">romeo.r@gmail.com</a>></span>:<br>
<blockquote class="gmail_quote"
style="margin:0 0 0 .8ex;border-left:1px
#ccc solid;padding-left:1ex">
<div dir="ltr">This volume is now being
tested by my collegue for windows
purposes.
<div>I will create new one on monday and
will test with parameters you've sent
me.</div>
</div>
<div class="gmail_extra"><br>
<div class="gmail_quote">2014-10-17
17:36 GMT+03:00 Pranith Kumar
Karampuri <span dir="ltr"><<a
moz-do-not-send="true"
href="mailto:pkarampu@redhat.com"
target="_blank">pkarampu@redhat.com</a>></span>:<br>
<blockquote class="gmail_quote"
style="margin:0 0 0
.8ex;border-left:1px #ccc
solid;padding-left:1ex">
<div text="#000000"
bgcolor="#FFFFFF"> Roman,<br>
Everything in the logs look okay
to me, except the following
profile number:<br>
3.91 1255944.81 us
127.00 us 23397532.00 us
189 FSYNC<br>
<br>
It seems that at least one of the
fsyncs is taking almost 23 seconds
to complete. According to all the
data you gave till now, I feel
this is the only thing I feel
could have done it. To test this
bit, could you turn off the
following option using and try
again?<br>
<br>
gluster volume set <volname>
cluster.ensure-durability off<br>
<br>
Let me know what happened. I am
extremely curious to here about
it.<br>
<br>
Pranith<br>
<br>
<div>On 10/17/2014 12:04 PM, Roman
wrote:<br>
</div>
<blockquote type="cite">
<div dir="ltr">mount
<div><br>
</div>
<div>
<div>[2014-10-13
17:36:56.758654] I
[glusterfsd.c:1959:main]
0-/usr/sbin/glusterfs:
Started running
/usr/sbin/glusterfs
version 3.5.2
(/usr/sbin/glusterfs
--direct-io-mode=enable
--fuse-mountopts=default_permissions,allow_other,max_read=131072
--volfile-server=stor1
--volfile-server=stor2
--volfile-id=HA-WIN-TT-1T
--fuse-mountopts=default_permissions,allow_other,max_read=131072
/srv/nfs/HA-WIN-TT-1T)</div>
<div>[2014-10-13
17:36:56.762162] I
[socket.c:3561:socket_init]
0-glusterfs: SSL support
is NOT enabled</div>
<div>[2014-10-13
17:36:56.762223] I
[socket.c:3576:socket_init]
0-glusterfs: using system
polling thread</div>
<div>[2014-10-13
17:36:56.766686] I
[dht-shared.c:311:dht_init_regex]
0-HA-WIN-TT-1T-dht: using
regex rsync-hash-regex =
^\.(.+)\.[^.]+$</div>
<div>[2014-10-13
17:36:56.768887] I
[socket.c:3561:socket_init]
0-HA-WIN-TT-1T-client-1:
SSL support is NOT enabled</div>
<div>[2014-10-13
17:36:56.768939] I
[socket.c:3576:socket_init]
0-HA-WIN-TT-1T-client-1:
using system polling
thread</div>
<div>[2014-10-13
17:36:56.769280] I
[socket.c:3561:socket_init]
0-HA-WIN-TT-1T-client-0:
SSL support is NOT enabled</div>
<div>[2014-10-13
17:36:56.769294] I
[socket.c:3576:socket_init]
0-HA-WIN-TT-1T-client-0:
using system polling
thread</div>
<div>[2014-10-13
17:36:56.769336] I
[client.c:2294:notify]
0-HA-WIN-TT-1T-client-0:
parent translators are
ready, attempting connect
on transport</div>
<div>[2014-10-13
17:36:56.769829] I
[client.c:2294:notify]
0-HA-WIN-TT-1T-client-1:
parent translators are
ready, attempting connect
on transport</div>
<div>Final graph:</div>
<div>+------------------------------------------------------------------------------+</div>
<div> 1: volume
HA-WIN-TT-1T-client-0</div>
<div> 2: type
protocol/client</div>
<div> 3: option
remote-host stor1</div>
<div> 4: option
remote-subvolume
/exports/NFS-WIN/1T</div>
<div> 5: option
transport-type socket</div>
<div> 6: option
ping-timeout 10</div>
<div> 7: option
send-gids true</div>
<div> 8: end-volume</div>
<div> 9:</div>
<div> 10: volume
HA-WIN-TT-1T-client-1</div>
<div> 11: type
protocol/client</div>
<div> 12: option
remote-host stor2</div>
<div> 13: option
remote-subvolume
/exports/NFS-WIN/1T</div>
<div> 14: option
transport-type socket</div>
<div> 15: option
ping-timeout 10</div>
<div> 16: option
send-gids true</div>
<div> 17: end-volume</div>
<div> 18:</div>
<div> 19: volume
HA-WIN-TT-1T-replicate-0</div>
<div> 20: type
cluster/replicate</div>
<div> 21: subvolumes
HA-WIN-TT-1T-client-0
HA-WIN-TT-1T-client-1</div>
<div> 22: end-volume</div>
<div> 23:</div>
<div> 24: volume
HA-WIN-TT-1T-dht</div>
<div> 25: type
cluster/distribute</div>
<div> 26: subvolumes
HA-WIN-TT-1T-replicate-0</div>
<div> 27: end-volume</div>
<div> 28:</div>
<div> 29: volume
HA-WIN-TT-1T-write-behind</div>
<div> 30: type
performance/write-behind</div>
<div> 31: subvolumes
HA-WIN-TT-1T-dht</div>
<div> 32: end-volume</div>
<div> 33:</div>
<div> 34: volume
HA-WIN-TT-1T-read-ahead</div>
<div> 35: type
performance/read-ahead</div>
<div> 36: subvolumes
HA-WIN-TT-1T-write-behind</div>
<div> 37: end-volume</div>
<div> 38:</div>
<div> 39: volume
HA-WIN-TT-1T-io-cache</div>
<div> 40: type
performance/io-cache</div>
<div> 41: subvolumes
HA-WIN-TT-1T-read-ahead</div>
<div> 42: end-volume</div>
<div> 43:</div>
<div> 44: volume
HA-WIN-TT-1T-quick-read</div>
<div> 45: type
performance/quick-read</div>
<div> 46: subvolumes
HA-WIN-TT-1T-io-cache</div>
<div> 47: end-volume</div>
<div> 48:</div>
<div> 49: volume
HA-WIN-TT-1T-open-behind</div>
<div> 50: type
performance/open-behind</div>
<div> 51: subvolumes
HA-WIN-TT-1T-quick-read</div>
<div> 52: end-volume</div>
<div> 53:</div>
<div> 54: volume
HA-WIN-TT-1T-md-cache</div>
<div> 55: type
performance/md-cache</div>
<div> 56: subvolumes
HA-WIN-TT-1T-open-behind</div>
<div> 57: end-volume</div>
<div> 58:</div>
<div> 59: volume
HA-WIN-TT-1T</div>
<div> 60: type
debug/io-stats</div>
<div> 61: option
latency-measurement off</div>
<div> 62: option
count-fop-hits off</div>
<div> 63: subvolumes
HA-WIN-TT-1T-md-cache</div>
<div> 64: end-volume</div>
<div> 65:</div>
<div>+------------------------------------------------------------------------------+</div>
<div>[2014-10-13
17:36:56.770718] I
[rpc-clnt.c:1729:rpc_clnt_reconfig]
0-HA-WIN-TT-1T-client-1:
changing port to 49160
(from 0)</div>
<div>[2014-10-13
17:36:56.771378] I
[rpc-clnt.c:1729:rpc_clnt_reconfig]
0-HA-WIN-TT-1T-client-0:
changing port to 49160
(from 0)</div>
<div>[2014-10-13
17:36:56.772008] I
[client-handshake.c:1677:select_server_supported_programs]
0-HA-WIN-TT-1T-client-1:
Using Program GlusterFS
3.3, Num (1298437),
Version (330)</div>
<div>[2014-10-13
17:36:56.772083] I
[client-handshake.c:1677:select_server_supported_programs]
0-HA-WIN-TT-1T-client-0:
Using Program GlusterFS
3.3, Num (1298437),
Version (330)</div>
<div>[2014-10-13
17:36:56.772338] I
[client-handshake.c:1462:client_setvolume_cbk]
0-HA-WIN-TT-1T-client-1:
Connected to <a
moz-do-not-send="true"
href="http://10.250.0.2:49160"
target="_blank">10.250.0.2:49160</a>,
attached to remote volume
'/exports/NFS-WIN/1T'.</div>
<div>[2014-10-13
17:36:56.772361] I
[client-handshake.c:1474:client_setvolume_cbk]
0-HA-WIN-TT-1T-client-1:
Server and Client
lk-version numbers are not
same, reopening the fds</div>
<div>[2014-10-13
17:36:56.772424] I
[afr-common.c:4131:afr_notify]
0-HA-WIN-TT-1T-replicate-0:
Subvolume
'HA-WIN-TT-1T-client-1'
came back up; going
online.</div>
<div>[2014-10-13
17:36:56.772463] I
[client-handshake.c:1462:client_setvolume_cbk]
0-HA-WIN-TT-1T-client-0:
Connected to <a
moz-do-not-send="true"
href="http://10.250.0.1:49160"
target="_blank">10.250.0.1:49160</a>,
attached to remote volume
'/exports/NFS-WIN/1T'.</div>
<div>[2014-10-13
17:36:56.772477] I
[client-handshake.c:1474:client_setvolume_cbk]
0-HA-WIN-TT-1T-client-0:
Server and Client
lk-version numbers are not
same, reopening the fds</div>
<div>[2014-10-13
17:36:56.779099] I
[fuse-bridge.c:4977:fuse_graph_setup]
0-fuse: switched to graph
0</div>
<div>[2014-10-13
17:36:56.779338] I
[client-handshake.c:450:client_set_lk_version_cbk]
0-HA-WIN-TT-1T-client-0:
Server lk version = 1</div>
<div>[2014-10-13
17:36:56.779367] I
[fuse-bridge.c:3914:fuse_init]
0-glusterfs-fuse: FUSE
inited with protocol
versions: glusterfs 7.22
kernel 7.17</div>
<div>[2014-10-13
17:36:56.779438] I
[client-handshake.c:450:client_set_lk_version_cbk]
0-HA-WIN-TT-1T-client-1:
Server lk version = 1</div>
<div>[2014-10-13
17:37:02.010942] I
[fuse-bridge.c:4818:fuse_thread_proc]
0-fuse: unmounting
/srv/nfs/HA-WIN-TT-1T</div>
<div>[2014-10-13
17:37:02.011296] W
[glusterfsd.c:1095:cleanup_and_exit]
(-->/lib/x86_64-linux-gnu/libc.so.6(clone+0x6d)
[0x7fc7b7672e6d]
(-->/lib/x86_64-linux-gnu/libpthread.so.0(+0x6b50)
[0x7fc7b7d20b50]
(-->/usr/sbin/glusterfs(glusterfs_sigwaiter+0xd5)
[0x7fc7b95add55]))) 0-:
received signum (15),
shutting down</div>
<div>[2014-10-13
17:37:02.011316] I
[fuse-bridge.c:5475:fini]
0-fuse: Unmounting
'/srv/nfs/HA-WIN-TT-1T'.</div>
<div>[2014-10-13
17:37:31.133036] W
[socket.c:522:__socket_rwv]
0-HA-WIN-TT-1T-client-0:
readv on <a
moz-do-not-send="true"
href="http://10.250.0.1:49160"
target="_blank">10.250.0.1:49160</a>
failed (No data available)</div>
<div>[2014-10-13
17:37:31.133110] I
[client.c:2229:client_rpc_notify]
0-HA-WIN-TT-1T-client-0:
disconnected from <a
moz-do-not-send="true"
href="http://10.250.0.1:49160"
target="_blank">10.250.0.1:49160</a>.
Client process will keep
trying to connect to
glusterd until brick's
port is available</div>
<div>[2014-10-13
17:37:33.317437] W
[socket.c:522:__socket_rwv]
0-HA-WIN-TT-1T-client-1:
readv on <a
moz-do-not-send="true"
href="http://10.250.0.2:49160"
target="_blank">10.250.0.2:49160</a>
failed (No data available)</div>
<div>[2014-10-13
17:37:33.317478] I
[client.c:2229:client_rpc_notify]
0-HA-WIN-TT-1T-client-1:
disconnected from <a
moz-do-not-send="true"
href="http://10.250.0.2:49160"
target="_blank">10.250.0.2:49160</a>.
Client process will keep
trying to connect to
glusterd until brick's
port is available</div>
<div>[2014-10-13
17:37:33.317496] E
[afr-common.c:4168:afr_notify]
0-HA-WIN-TT-1T-replicate-0:
All subvolumes are down.
Going offline until
atleast one of them comes
back up.</div>
<div>[2014-10-13
17:37:42.045604] I
[rpc-clnt.c:1729:rpc_clnt_reconfig]
0-HA-WIN-TT-1T-client-0:
changing port to 49160
(from 0)</div>
<div>[2014-10-13
17:37:42.046177] I
[client-handshake.c:1677:select_server_supported_programs]
0-HA-WIN-TT-1T-client-0:
Using Program GlusterFS
3.3, Num (1298437),
Version (330)</div>
<div>[2014-10-13
17:37:42.048863] I
[client-handshake.c:1462:client_setvolume_cbk]
0-HA-WIN-TT-1T-client-0:
Connected to <a
moz-do-not-send="true"
href="http://10.250.0.1:49160"
target="_blank">10.250.0.1:49160</a>,
attached to remote volume
'/exports/NFS-WIN/1T'.</div>
<div>[2014-10-13
17:37:42.048883] I
[client-handshake.c:1474:client_setvolume_cbk]
0-HA-WIN-TT-1T-client-0:
Server and Client
lk-version numbers are not
same, reopening the fds</div>
<div>[2014-10-13
17:37:42.048897] I
[client-handshake.c:1314:client_post_handshake]
0-HA-WIN-TT-1T-client-0: 1
fds open - Delaying
child_up until they are
re-opened</div>
<div>[2014-10-13
17:37:42.049299] W
[client-handshake.c:980:client3_3_reopen_cbk]
0-HA-WIN-TT-1T-client-0:
reopen on
<gfid:b00e322a-7bae-479f-91e0-1fd77c73692b>
failed (Stale NFS file
handle)</div>
<div>[2014-10-13
17:37:42.049328] I
[client-handshake.c:936:client_child_up_reopen_done]
0-HA-WIN-TT-1T-client-0:
last fd
open'd/lock-self-heal'd -
notifying CHILD-UP</div>
<div>[2014-10-13
17:37:42.049360] I
[afr-common.c:4131:afr_notify]
0-HA-WIN-TT-1T-replicate-0:
Subvolume
'HA-WIN-TT-1T-client-0'
came back up; going
online.</div>
<div>[2014-10-13
17:37:42.049446] I
[client-handshake.c:450:client_set_lk_version_cbk]
0-HA-WIN-TT-1T-client-0:
Server lk version = 1</div>
<div>[2014-10-13
17:37:45.087592] I
[rpc-clnt.c:1729:rpc_clnt_reconfig]
0-HA-WIN-TT-1T-client-1:
changing port to 49160
(from 0)</div>
<div>[2014-10-13
17:37:45.088132] I
[client-handshake.c:1677:select_server_supported_programs]
0-HA-WIN-TT-1T-client-1:
Using Program GlusterFS
3.3, Num (1298437),
Version (330)</div>
<div>[2014-10-13
17:37:45.088343] I
[client-handshake.c:1462:client_setvolume_cbk]
0-HA-WIN-TT-1T-client-1:
Connected to <a
moz-do-not-send="true"
href="http://10.250.0.2:49160"
target="_blank">10.250.0.2:49160</a>,
attached to remote volume
'/exports/NFS-WIN/1T'.</div>
<div>[2014-10-13
17:37:45.088360] I
[client-handshake.c:1474:client_setvolume_cbk]
0-HA-WIN-TT-1T-client-1:
Server and Client
lk-version numbers are not
same, reopening the fds</div>
<div>[2014-10-13
17:37:45.088373] I
[client-handshake.c:1314:client_post_handshake]
0-HA-WIN-TT-1T-client-1: 1
fds open - Delaying
child_up until they are
re-opened</div>
<div>[2014-10-13
17:37:45.088681] W
[client-handshake.c:980:client3_3_reopen_cbk]
0-HA-WIN-TT-1T-client-1:
reopen on
<gfid:b00e322a-7bae-479f-91e0-1fd77c73692b>
failed (Stale NFS file
handle)</div>
<div>[2014-10-13
17:37:45.088697] I
[client-handshake.c:936:client_child_up_reopen_done]
0-HA-WIN-TT-1T-client-1:
last fd
open'd/lock-self-heal'd -
notifying CHILD-UP</div>
<div>[2014-10-13
17:37:45.088819] I
[client-handshake.c:450:client_set_lk_version_cbk]
0-HA-WIN-TT-1T-client-1:
Server lk version = 1</div>
<div>[2014-10-13
17:37:54.601822] I
[glusterfsd.c:1959:main]
0-/usr/sbin/glusterfs:
Started running
/usr/sbin/glusterfs
version 3.5.2
(/usr/sbin/glusterfs
--direct-io-mode=enable
--fuse-mountopts=default_permissions,allow_other,max_read=131072
--volfile-server=stor1
--volfile-server=stor2
--volfile-id=HA-WIN-TT-1T
--fuse-mountopts=default_permissions,allow_other,max_read=131072
/srv/nfs/HA-WIN-TT-1T)</div>
<div>[2014-10-13
17:37:54.604972] I
[socket.c:3561:socket_init]
0-glusterfs: SSL support
is NOT enabled</div>
<div>[2014-10-13
17:37:54.605034] I
[socket.c:3576:socket_init]
0-glusterfs: using system
polling thread</div>
<div>[2014-10-13
17:37:54.609219] I
[dht-shared.c:311:dht_init_regex]
0-HA-WIN-TT-1T-dht: using
regex rsync-hash-regex =
^\.(.+)\.[^.]+$</div>
<div>[2014-10-13
17:37:54.611421] I
[socket.c:3561:socket_init]
0-HA-WIN-TT-1T-client-1:
SSL support is NOT enabled</div>
<div>[2014-10-13
17:37:54.611466] I
[socket.c:3576:socket_init]
0-HA-WIN-TT-1T-client-1:
using system polling
thread</div>
<div>[2014-10-13
17:37:54.611808] I
[socket.c:3561:socket_init]
0-HA-WIN-TT-1T-client-0:
SSL support is NOT enabled</div>
<div>[2014-10-13
17:37:54.611821] I
[socket.c:3576:socket_init]
0-HA-WIN-TT-1T-client-0:
using system polling
thread</div>
<div>[2014-10-13
17:37:54.611862] I
[client.c:2294:notify]
0-HA-WIN-TT-1T-client-0:
parent translators are
ready, attempting connect
on transport</div>
<div>[2014-10-13
17:37:54.612354] I
[client.c:2294:notify]
0-HA-WIN-TT-1T-client-1:
parent translators are
ready, attempting connect
on transport</div>
<div>Final graph:</div>
<div>+------------------------------------------------------------------------------+</div>
<div> 1: volume
HA-WIN-TT-1T-client-0</div>
<div> 2: type
protocol/client</div>
<div> 3: option
remote-host stor1</div>
<div> 4: option
remote-subvolume
/exports/NFS-WIN/1T</div>
<div> 5: option
transport-type socket</div>
<div> 6: option
ping-timeout 10</div>
<div> 7: option
send-gids true</div>
<div> 8: end-volume</div>
<div> 9:</div>
<div> 10: volume
HA-WIN-TT-1T-client-1</div>
<div> 11: type
protocol/client</div>
<div> 12: option
remote-host stor2</div>
<div> 13: option
remote-subvolume
/exports/NFS-WIN/1T</div>
<div> 14: option
transport-type socket</div>
<div> 15: option
ping-timeout 10</div>
<div> 16: option
send-gids true</div>
<div> 17: end-volume</div>
<div> 18:</div>
<div> 19: volume
HA-WIN-TT-1T-replicate-0</div>
<div> 20: type
cluster/replicate</div>
<div> 21: subvolumes
HA-WIN-TT-1T-client-0
HA-WIN-TT-1T-client-1</div>
<div> 22: end-volume</div>
<div> 23:</div>
<div> 24: volume
HA-WIN-TT-1T-dht</div>
<div> 25: type
cluster/distribute</div>
<div> 26: subvolumes
HA-WIN-TT-1T-replicate-0</div>
<div> 27: end-volume</div>
<div> 28:</div>
<div> 29: volume
HA-WIN-TT-1T-write-behind</div>
<div> 30: type
performance/write-behind</div>
<div> 31: subvolumes
HA-WIN-TT-1T-dht</div>
<div> 32: end-volume</div>
<div> 33:</div>
<div> 34: volume
HA-WIN-TT-1T-read-ahead</div>
<div> 35: type
performance/read-ahead</div>
<div> 36: subvolumes
HA-WIN-TT-1T-write-behind</div>
<div> 37: end-volume</div>
<div> 38:</div>
<div> 39: volume
HA-WIN-TT-1T-io-cache</div>
<div> 40: type
performance/io-cache</div>
<div> 41: subvolumes
HA-WIN-TT-1T-read-ahead</div>
<div> 42: end-volume</div>
<div> 43:</div>
<div> 44: volume
HA-WIN-TT-1T-quick-read</div>
<div> 45: type
performance/quick-read</div>
<div> 46: subvolumes
HA-WIN-TT-1T-io-cache</div>
<div> 47: end-volume</div>
<div> 48:</div>
<div> 49: volume
HA-WIN-TT-1T-open-behind</div>
<div> 50: type
performance/open-behind</div>
<div> 51: subvolumes
HA-WIN-TT-1T-quick-read</div>
<div> 52: end-volume</div>
<div> 53:</div>
<div> 54: volume
HA-WIN-TT-1T-md-cache</div>
<div> 55: type
performance/md-cache</div>
<div> 56: subvolumes
HA-WIN-TT-1T-open-behind</div>
<div> 57: end-volume</div>
<div> 58:</div>
<div> 59: volume
HA-WIN-TT-1T</div>
<div> 60: type
debug/io-stats</div>
<div> 61: option
latency-measurement off</div>
<div> 62: option
count-fop-hits off</div>
<div> 63: subvolumes
HA-WIN-TT-1T-md-cache</div>
<div> 64: end-volume</div>
<div> 65:</div>
<div>+------------------------------------------------------------------------------+</div>
<div>[2014-10-13
17:37:54.613137] I
[rpc-clnt.c:1729:rpc_clnt_reconfig]
0-HA-WIN-TT-1T-client-0:
changing port to 49160
(from 0)</div>
<div>[2014-10-13
17:37:54.613521] I
[rpc-clnt.c:1729:rpc_clnt_reconfig]
0-HA-WIN-TT-1T-client-1:
changing port to 49160
(from 0)</div>
<div>[2014-10-13
17:37:54.614228] I
[client-handshake.c:1677:select_server_supported_programs]
0-HA-WIN-TT-1T-client-0:
Using Program GlusterFS
3.3, Num (1298437),
Version (330)</div>
<div>[2014-10-13
17:37:54.614399] I
[client-handshake.c:1677:select_server_supported_programs]
0-HA-WIN-TT-1T-client-1:
Using Program GlusterFS
3.3, Num (1298437),
Version (330)</div>
<div>[2014-10-13
17:37:54.614483] I
[client-handshake.c:1462:client_setvolume_cbk]
0-HA-WIN-TT-1T-client-0:
Connected to <a
moz-do-not-send="true"
href="http://10.250.0.1:49160"
target="_blank">10.250.0.1:49160</a>,
attached to remote volume
'/exports/NFS-WIN/1T'.</div>
<div>[2014-10-13
17:37:54.614499] I
[client-handshake.c:1474:client_setvolume_cbk]
0-HA-WIN-TT-1T-client-0:
Server and Client
lk-version numbers are not
same, reopening the fds</div>
<div>[2014-10-13
17:37:54.614557] I
[afr-common.c:4131:afr_notify]
0-HA-WIN-TT-1T-replicate-0:
Subvolume
'HA-WIN-TT-1T-client-0'
came back up; going
online.</div>
<div>[2014-10-13
17:37:54.614625] I
[client-handshake.c:450:client_set_lk_version_cbk]
0-HA-WIN-TT-1T-client-0:
Server lk version = 1</div>
<div>[2014-10-13
17:37:54.614709] I
[client-handshake.c:1462:client_setvolume_cbk]
0-HA-WIN-TT-1T-client-1:
Connected to <a
moz-do-not-send="true"
href="http://10.250.0.2:49160"
target="_blank">10.250.0.2:49160</a>,
attached to remote volume
'/exports/NFS-WIN/1T'.</div>
<div>[2014-10-13
17:37:54.614724] I
[client-handshake.c:1474:client_setvolume_cbk]
0-HA-WIN-TT-1T-client-1:
Server and Client
lk-version numbers are not
same, reopening the fds</div>
<div>[2014-10-13
17:37:54.621318] I
[fuse-bridge.c:4977:fuse_graph_setup]
0-fuse: switched to graph
0</div>
<div>[2014-10-13
17:37:54.621545] I
[client-handshake.c:450:client_set_lk_version_cbk]
0-HA-WIN-TT-1T-client-1:
Server lk version = 1</div>
<div>[2014-10-13
17:37:54.621617] I
[fuse-bridge.c:3914:fuse_init]
0-glusterfs-fuse: FUSE
inited with protocol
versions: glusterfs 7.22
kernel 7.17</div>
<div>[2014-10-13
17:38:25.951778] W
[client-rpc-fops.c:4235:client3_3_flush]
0-HA-WIN-TT-1T-client-0:
(b00e322a-7bae-479f-91e0-1fd77c73692b)
remote_fd is -1. EBADFD</div>
<div>[2014-10-13
17:38:25.951827] W
[client-rpc-fops.c:4235:client3_3_flush]
0-HA-WIN-TT-1T-client-1:
(b00e322a-7bae-479f-91e0-1fd77c73692b)
remote_fd is -1. EBADFD</div>
<div>[2014-10-13
17:38:25.966963] I
[fuse-bridge.c:4818:fuse_thread_proc]
0-fuse: unmounting
/srv/nfs/HA-WIN-TT-1T</div>
<div>[2014-10-13
17:38:25.967174] W
[glusterfsd.c:1095:cleanup_and_exit]
(-->/lib/x86_64-linux-gnu/libc.so.6(clone+0x6d)
[0x7ffec893de6d]
(-->/lib/x86_64-linux-gnu/libpthread.so.0(+0x6b50)
[0x7ffec8febb50]
(-->/usr/sbin/glusterfs(glusterfs_sigwaiter+0xd5)
[0x7ffeca878d55]))) 0-:
received signum (15),
shutting down</div>
<div>[2014-10-13
17:38:25.967194] I
[fuse-bridge.c:5475:fini]
0-fuse: Unmounting
'/srv/nfs/HA-WIN-TT-1T'.</div>
<div>[2014-10-13
17:40:21.500514] I
[glusterfsd-mgmt.c:56:mgmt_cbk_spec]
0-mgmt: Volume file
changed</div>
<div>[2014-10-13
17:40:21.517782] I
[glusterfsd-mgmt.c:56:mgmt_cbk_spec]
0-mgmt: Volume file
changed</div>
<div>[2014-10-13
17:40:21.524056] I
[dht-shared.c:311:dht_init_regex]
0-HA-WIN-TT-1T-dht: using
regex rsync-hash-regex =
^\.(.+)\.[^.]+$</div>
<div>[2014-10-13
17:40:21.528430] I
[glusterfsd-mgmt.c:1307:mgmt_getspec_cbk]
0-glusterfs: No change in
volfile, continuing</div>
</div>
<div><br>
</div>
<div>glusterfshd stor1</div>
<div><br>
</div>
<div>
<div>2014-10-13
17:38:17.203360] I
[glusterfsd.c:1959:main]
0-/usr/sbin/glusterfs:
Started running
/usr/sbin/glusterfs
version 3.5.2
(/usr/sbin/glusterfs -s
localhost --volfile-id
gluster/glustershd -p
/var/lib/glusterd/glustershd/run/glustershd.pid
-l
/var/log/glusterfs/glustershd.log
-S
/var/run/75bbc77a676bde0d0afe20f40dc9e3e1.socket
--xlator-option
*replicate*.node-uuid=e09cbbc2-08a3-4e5b-83b8-48eb11a1c7b3)</div>
<div>[2014-10-13
17:38:17.204958] I
[socket.c:3561:socket_init]
0-socket.glusterfsd: SSL
support is NOT enabled</div>
<div>[2014-10-13
17:38:17.205016] I
[socket.c:3576:socket_init]
0-socket.glusterfsd: using
system polling thread</div>
<div>[2014-10-13
17:38:17.205188] I
[socket.c:3561:socket_init]
0-glusterfs: SSL support
is NOT enabled</div>
<div>[2014-10-13
17:38:17.205209] I
[socket.c:3576:socket_init]
0-glusterfs: using system
polling thread</div>
<div>[2014-10-13
17:38:17.207840] I
[graph.c:254:gf_add_cmdline_options]
0-HA-2TB-TT-Proxmox-cluster-replicate-0:
adding option 'node-uuid'
for volume
'HA-2TB-TT-Proxmox-cluster-replicate-0'
with value
'e09cbbc2-08a3-4e5b-83b8-48eb11a1c7b3'</div>
<div>[2014-10-13
17:38:17.209433] I
[socket.c:3561:socket_init]
0-HA-2TB-TT-Proxmox-cluster-client-1:
SSL support is NOT enabled</div>
<div>[2014-10-13
17:38:17.209448] I
[socket.c:3576:socket_init]
0-HA-2TB-TT-Proxmox-cluster-client-1:
using system polling
thread</div>
<div>[2014-10-13
17:38:17.209625] I
[socket.c:3561:socket_init]
0-HA-2TB-TT-Proxmox-cluster-client-0:
SSL support is NOT enabled</div>
<div>[2014-10-13
17:38:17.209634] I
[socket.c:3576:socket_init]
0-HA-2TB-TT-Proxmox-cluster-client-0:
using system polling
thread</div>
<div>[2014-10-13
17:38:17.209652] I
[client.c:2294:notify]
0-HA-2TB-TT-Proxmox-cluster-client-0:
parent translators are
ready, attempting connect
on transport</div>
<div>[2014-10-13
17:38:17.210241] I
[client.c:2294:notify]
0-HA-2TB-TT-Proxmox-cluster-client-1:
parent translators are
ready, attempting connect
on transport</div>
<div>Final graph:</div>
<div>+------------------------------------------------------------------------------+</div>
<div> 1: volume
HA-2TB-TT-Proxmox-cluster-client-0</div>
<div> 2: type
protocol/client</div>
<div> 3: option
remote-host stor1</div>
<div> 4: option
remote-subvolume
/exports/HA-2TB-TT-Proxmox-cluster/2TB</div>
<div> 5: option
transport-type socket</div>
<div> 6: option
username
59c66122-55c1-4c28-956e-6189fcb1aff5</div>
<div> 7: option
password
34b79afb-a93c-431b-900a-b688e67cdbc9</div>
<div> 8: option
ping-timeout 10</div>
<div> 9: end-volume</div>
<div> 10:</div>
<div> 11: volume
HA-2TB-TT-Proxmox-cluster-client-1</div>
<div> 12: type
protocol/client</div>
<div> 13: option
remote-host stor2</div>
<div> 14: option
remote-subvolume
/exports/HA-2TB-TT-Proxmox-cluster/2TB</div>
<div> 15: option
transport-type socket</div>
<div> 16: option
username
59c66122-55c1-4c28-956e-6189fcb1aff5</div>
<div> 17: option
password
34b79afb-a93c-431b-900a-b688e67cdbc9</div>
<div> 18: option
ping-timeout 10</div>
<div> 19: end-volume</div>
<div> 20:</div>
<div> 21: volume
HA-2TB-TT-Proxmox-cluster-replicate-0</div>
<div> 22: type
cluster/replicate</div>
<div> 23: option
node-uuid
e09cbbc2-08a3-4e5b-83b8-48eb11a1c7b3</div>
<div> 24: option
background-self-heal-count
0</div>
<div> 25: option
metadata-self-heal on</div>
<div> 26: option
data-self-heal on</div>
<div> 27: option
entry-self-heal on</div>
<div> 28: option
self-heal-daemon on</div>
<div> 29: option
iam-self-heal-daemon yes</div>
<div> 30: subvolumes
HA-2TB-TT-Proxmox-cluster-client-0
HA-2TB-TT-Proxmox-cluster-client-1</div>
<div> 31: end-volume</div>
<div> 32:</div>
<div> 33: volume glustershd</div>
<div> 34: type
debug/io-stats</div>
<div> 35: subvolumes
HA-2TB-TT-Proxmox-cluster-replicate-0</div>
<div> 36: end-volume</div>
<div> 37:</div>
<div>+------------------------------------------------------------------------------+</div>
<div>[2014-10-13
17:38:17.210709] I
[rpc-clnt.c:1729:rpc_clnt_reconfig]
0-HA-2TB-TT-Proxmox-cluster-client-0:
changing port to 49159
(from 0)</div>
<div>[2014-10-13
17:38:17.211008] I
[client-handshake.c:1677:select_server_supported_programs]
0-HA-2TB-TT-Proxmox-cluster-client-0:
Using Program GlusterFS
3.3, Num (1298437),
Version (330)</div>
<div>[2014-10-13
17:38:17.211170] I
[client-handshake.c:1462:client_setvolume_cbk]
0-HA-2TB-TT-Proxmox-cluster-client-0:
Connected to <a
moz-do-not-send="true"
href="http://10.250.0.1:49159"
target="_blank">10.250.0.1:49159</a>,
attached to remote volume
'/exports/HA-2TB-TT-Proxmox-cluster/2TB'.</div>
<div>[2014-10-13
17:38:17.211195] I
[client-handshake.c:1474:client_setvolume_cbk]
0-HA-2TB-TT-Proxmox-cluster-client-0:
Server and Client
lk-version numbers are not
same, reopening the fds</div>
<div>[2014-10-13
17:38:17.211250] I
[afr-common.c:4131:afr_notify]
0-HA-2TB-TT-Proxmox-cluster-replicate-0:
Subvolume
'HA-2TB-TT-Proxmox-cluster-client-0'
came back up; going
online.</div>
<div>[2014-10-13
17:38:17.211297] I
[client-handshake.c:450:client_set_lk_version_cbk]
0-HA-2TB-TT-Proxmox-cluster-client-0:
Server lk version = 1</div>
<div>[2014-10-13
17:38:17.211656] I
[afr-self-heald.c:1690:afr_dir_exclusive_crawl]
0-HA-2TB-TT-Proxmox-cluster-replicate-0:
Another crawl is in
progress for
HA-2TB-TT-Proxmox-cluster-client-0</div>
<div>[2014-10-13
17:38:17.211661] E
[afr-self-heald.c:1479:afr_find_child_position]
0-HA-2TB-TT-Proxmox-cluster-replicate-0:
getxattr failed on
HA-2TB-TT-Proxmox-cluster-client-1
- (Transport endpoint is
not connected)</div>
<div>[2014-10-13
17:38:17.216327] E
[afr-self-heal-data.c:1611:afr_sh_data_open_cbk]
0-HA-2TB-TT-Proxmox-cluster-replicate-0:
open of
<gfid:65381af4-8e0b-4721-8214-71d29dcf5237>
failed on child
HA-2TB-TT-Proxmox-cluster-client-1
(Transport endpoint is not
connected)</div>
<div>[2014-10-13
17:38:17.217372] E
[afr-self-heal-data.c:1611:afr_sh_data_open_cbk]
0-HA-2TB-TT-Proxmox-cluster-replicate-0:
open of
<gfid:65381af4-8e0b-4721-8214-71d29dcf5237>
failed on child
HA-2TB-TT-Proxmox-cluster-client-1
(Transport endpoint is not
connected)</div>
<div>[2014-10-13
17:38:19.226057] I
[rpc-clnt.c:1729:rpc_clnt_reconfig]
0-HA-2TB-TT-Proxmox-cluster-client-1:
changing port to 49159
(from 0)</div>
<div>[2014-10-13
17:38:19.226704] I
[client-handshake.c:1677:select_server_supported_programs]
0-HA-2TB-TT-Proxmox-cluster-client-1:
Using Program GlusterFS
3.3, Num (1298437),
Version (330)</div>
<div>[2014-10-13
17:38:19.226896] I
[client-handshake.c:1462:client_setvolume_cbk]
0-HA-2TB-TT-Proxmox-cluster-client-1:
Connected to <a
moz-do-not-send="true"
href="http://10.250.0.2:49159"
target="_blank">10.250.0.2:49159</a>,
attached to remote volume
'/exports/HA-2TB-TT-Proxmox-cluster/2TB'.</div>
<div>[2014-10-13
17:38:19.226916] I
[client-handshake.c:1474:client_setvolume_cbk]
0-HA-2TB-TT-Proxmox-cluster-client-1:
Server and Client
lk-version numbers are not
same, reopening the fds</div>
<div>[2014-10-13
17:38:19.227031] I
[client-handshake.c:450:client_set_lk_version_cbk]
0-HA-2TB-TT-Proxmox-cluster-client-1:
Server lk version = 1</div>
<div>[2014-10-13
17:38:25.933950] W
[glusterfsd.c:1095:cleanup_and_exit]
(-->/lib/x86_64-linux-gnu/libc.so.6(clone+0x6d)
[0x7f1a7c03ce6d]
(-->/lib/x86_64-linux-gnu/libpthread.so.0(+0x6b50)
[0x7f1a7c6eab50]
(-->/usr/sbin/glusterfs(glusterfs_sigwaiter+0xd5)
[0x7f1a7df77d55]))) 0-:
received signum (15),
shutting down</div>
<div>[2014-10-13
17:38:26.942918] I
[glusterfsd.c:1959:main]
0-/usr/sbin/glusterfs:
Started running
/usr/sbin/glusterfs
version 3.5.2
(/usr/sbin/glusterfs -s
localhost --volfile-id
gluster/glustershd -p
/var/lib/glusterd/glustershd/run/glustershd.pid
-l
/var/log/glusterfs/glustershd.log
-S
/var/run/75bbc77a676bde0d0afe20f40dc9e3e1.socket
--xlator-option
*replicate*.node-uuid=e09cbbc2-08a3-4e5b-83b8-48eb11a1c7b3)</div>
<div>[2014-10-13
17:38:26.944548] I
[socket.c:3561:socket_init]
0-socket.glusterfsd: SSL
support is NOT enabled</div>
<div>[2014-10-13
17:38:26.944584] I
[socket.c:3576:socket_init]
0-socket.glusterfsd: using
system polling thread</div>
<div>[2014-10-13
17:38:26.944689] I
[socket.c:3561:socket_init]
0-glusterfs: SSL support
is NOT enabled</div>
<div>[2014-10-13
17:38:26.944701] I
[socket.c:3576:socket_init]
0-glusterfs: using system
polling thread</div>
<div>[2014-10-13
17:38:26.946667] I
[graph.c:254:gf_add_cmdline_options]
0-HA-2TB-TT-Proxmox-cluster-replicate-0:
adding option 'node-uuid'
for volume
'HA-2TB-TT-Proxmox-cluster-replicate-0'
with value
'e09cbbc2-08a3-4e5b-83b8-48eb11a1c7b3'</div>
<div>[2014-10-13
17:38:26.946684] I
[graph.c:254:gf_add_cmdline_options]
0-HA-WIN-TT-1T-replicate-0:
adding option 'node-uuid'
for volume
'HA-WIN-TT-1T-replicate-0'
with value
'e09cbbc2-08a3-4e5b-83b8-48eb11a1c7b3'</div>
<div>[2014-10-13
17:38:26.948783] I
[socket.c:3561:socket_init]
0-HA-2TB-TT-Proxmox-cluster-client-1:
SSL support is NOT enabled</div>
<div>[2014-10-13
17:38:26.948809] I
[socket.c:3576:socket_init]
0-HA-2TB-TT-Proxmox-cluster-client-1:
using system polling
thread</div>
<div>[2014-10-13
17:38:26.949118] I
[socket.c:3561:socket_init]
0-HA-2TB-TT-Proxmox-cluster-client-0:
SSL support is NOT enabled</div>
<div>[2014-10-13
17:38:26.949134] I
[socket.c:3576:socket_init]
0-HA-2TB-TT-Proxmox-cluster-client-0:
using system polling
thread</div>
<div>[2014-10-13
17:38:26.951698] I
[socket.c:3561:socket_init]
0-HA-WIN-TT-1T-client-1:
SSL support is NOT enabled</div>
<div>[2014-10-13
17:38:26.951715] I
[socket.c:3576:socket_init]
0-HA-WIN-TT-1T-client-1:
using system polling
thread</div>
<div>[2014-10-13
17:38:26.951921] I
[socket.c:3561:socket_init]
0-HA-WIN-TT-1T-client-0:
SSL support is NOT enabled</div>
<div>[2014-10-13
17:38:26.951932] I
[socket.c:3576:socket_init]
0-HA-WIN-TT-1T-client-0:
using system polling
thread</div>
<div>[2014-10-13
17:38:26.951959] I
[client.c:2294:notify]
0-HA-2TB-TT-Proxmox-cluster-client-0:
parent translators are
ready, attempting connect
on transport</div>
<div>[2014-10-13
17:38:26.952612] I
[client.c:2294:notify]
0-HA-2TB-TT-Proxmox-cluster-client-1:
parent translators are
ready, attempting connect
on transport</div>
<div>[2014-10-13
17:38:26.952862] I
[client.c:2294:notify]
0-HA-WIN-TT-1T-client-0:
parent translators are
ready, attempting connect
on transport</div>
<div>[2014-10-13
17:38:26.953447] I
[client.c:2294:notify]
0-HA-WIN-TT-1T-client-1:
parent translators are
ready, attempting connect
on transport</div>
<div>Final graph:</div>
<div>+------------------------------------------------------------------------------+</div>
<div> 1: volume
HA-2TB-TT-Proxmox-cluster-client-0</div>
<div> 2: type
protocol/client</div>
<div> 3: option
remote-host stor1</div>
<div> 4: option
remote-subvolume
/exports/HA-2TB-TT-Proxmox-cluster/2TB</div>
<div> 5: option
transport-type socket</div>
<div> 6: option
username
59c66122-55c1-4c28-956e-6189fcb1aff5</div>
<div> 7: option
password
34b79afb-a93c-431b-900a-b688e67cdbc9</div>
<div> 8: option
ping-timeout 10</div>
<div> 9: end-volume</div>
<div> 10:</div>
<div> 11: volume
HA-2TB-TT-Proxmox-cluster-client-1</div>
<div> 12: type
protocol/client</div>
<div> 13: option
remote-host stor2</div>
<div> 14: option
remote-subvolume
/exports/HA-2TB-TT-Proxmox-cluster/2TB</div>
<div> 15: option
transport-type socket</div>
<div> 16: option
username
59c66122-55c1-4c28-956e-6189fcb1aff5</div>
<div> 17: option
password
34b79afb-a93c-431b-900a-b688e67cdbc9</div>
<div> 18: option
ping-timeout 10</div>
<div> 19: end-volume</div>
<div> 20:</div>
<div> 21: volume
HA-2TB-TT-Proxmox-cluster-replicate-0</div>
<div> 22: type
cluster/replicate</div>
<div> 23: option
node-uuid
e09cbbc2-08a3-4e5b-83b8-48eb11a1c7b3</div>
<div> 24: option
background-self-heal-count
0</div>
<div> 25: option
metadata-self-heal on</div>
<div> 26: option
data-self-heal on</div>
<div> 27: option
entry-self-heal on</div>
<div> 28: option
self-heal-daemon on</div>
<div> 29: option
iam-self-heal-daemon yes</div>
<div> 30: subvolumes
HA-2TB-TT-Proxmox-cluster-client-0
HA-2TB-TT-Proxmox-cluster-client-1</div>
<div> 31: end-volume</div>
<div> 32:</div>
<div> 33: volume
HA-WIN-TT-1T-client-0</div>
<div> 34: type
protocol/client</div>
<div> 35: option
remote-host stor1</div>
<div> 36: option
remote-subvolume
/exports/NFS-WIN/1T</div>
<div> 37: option
transport-type socket</div>
<div> 38: option
username
101b907c-ff21-47da-8ba6-37e2920691ce</div>
<div> 39: option
password
f4f29094-891f-4241-8736-5e3302ed8bc8</div>
<div> 40: option
ping-timeout 10</div>
<div> 41: end-volume</div>
<div> 42:</div>
<div> 43: volume
HA-WIN-TT-1T-client-1</div>
<div> 44: type
protocol/client</div>
<div> 45: option
remote-host stor2</div>
<div> 46: option
remote-subvolume
/exports/NFS-WIN/1T</div>
<div> 47: option
transport-type socket</div>
<div> 48: option
username
101b907c-ff21-47da-8ba6-37e2920691ce</div>
<div> 49: option
password
f4f29094-891f-4241-8736-5e3302ed8bc8</div>
<div> 50: option
ping-timeout 10</div>
<div> 51: end-volume</div>
<div> 52:</div>
<div> 53: volume
HA-WIN-TT-1T-replicate-0</div>
<div> 54: type
cluster/replicate</div>
<div> 55: option
node-uuid
e09cbbc2-08a3-4e5b-83b8-48eb11a1c7b3</div>
<div> 56: option
background-self-heal-count
0</div>
<div> 57: option
metadata-self-heal on</div>
<div> 58: option
data-self-heal on</div>
<div> 59: option
entry-self-heal on</div>
<div> 60: option
self-heal-daemon on</div>
<div> 61: option
iam-self-heal-daemon yes</div>
<div> 62: subvolumes
HA-WIN-TT-1T-client-0
HA-WIN-TT-1T-client-1</div>
<div> 63: end-volume</div>
<div> 64:</div>
<div> 65: volume glustershd</div>
<div> 66: type
debug/io-stats</div>
<div> 67: subvolumes
HA-2TB-TT-Proxmox-cluster-replicate-0
HA-WIN-TT-1T-replicate-0</div>
<div> 68: end-volume</div>
<div> 69:</div>
<div>+------------------------------------------------------------------------------+</div>
<div>[2014-10-13
17:38:26.954036] I
[rpc-clnt.c:1729:rpc_clnt_reconfig]
0-HA-2TB-TT-Proxmox-cluster-client-0:
changing port to 49159
(from 0)</div>
<div>[2014-10-13
17:38:26.954308] I
[rpc-clnt.c:1729:rpc_clnt_reconfig]
0-HA-WIN-TT-1T-client-0:
changing port to 49160
(from 0)</div>
<div>[2014-10-13
17:38:26.954741] I
[client-handshake.c:1677:select_server_supported_programs]
0-HA-2TB-TT-Proxmox-cluster-client-0:
Using Program GlusterFS
3.3, Num (1298437),
Version (330)</div>
<div>[2014-10-13
17:38:26.954815] I
[client-handshake.c:1677:select_server_supported_programs]
0-HA-WIN-TT-1T-client-0:
Using Program GlusterFS
3.3, Num (1298437),
Version (330)</div>
<div>[2014-10-13
17:38:26.954999] I
[client-handshake.c:1462:client_setvolume_cbk]
0-HA-2TB-TT-Proxmox-cluster-client-0:
Connected to <a
moz-do-not-send="true"
href="http://10.250.0.1:49159"
target="_blank">10.250.0.1:49159</a>,
attached to remote volume
'/exports/HA-2TB-TT-Proxmox-cluster/2TB'.</div>
<div>[2014-10-13
17:38:26.955017] I
[client-handshake.c:1474:client_setvolume_cbk]
0-HA-2TB-TT-Proxmox-cluster-client-0:
Server and Client
lk-version numbers are not
same, reopening the fds</div>
<div>[2014-10-13
17:38:26.955073] I
[afr-common.c:4131:afr_notify]
0-HA-2TB-TT-Proxmox-cluster-replicate-0:
Subvolume
'HA-2TB-TT-Proxmox-cluster-client-0'
came back up; going
online.</div>
<div>[2014-10-13
17:38:26.955127] I
[client-handshake.c:450:client_set_lk_version_cbk]
0-HA-2TB-TT-Proxmox-cluster-client-0:
Server lk version = 1</div>
<div>[2014-10-13
17:38:26.955151] I
[client-handshake.c:1462:client_setvolume_cbk]
0-HA-WIN-TT-1T-client-0:
Connected to <a
moz-do-not-send="true"
href="http://10.250.0.1:49160"
target="_blank">10.250.0.1:49160</a>,
attached to remote volume
'/exports/NFS-WIN/1T'.</div>
<div>[2014-10-13
17:38:26.955161] I
[client-handshake.c:1474:client_setvolume_cbk]
0-HA-WIN-TT-1T-client-0:
Server and Client
lk-version numbers are not
same, reopening the fds</div>
<div>[2014-10-13
17:38:26.955226] I
[afr-common.c:4131:afr_notify]
0-HA-WIN-TT-1T-replicate-0:
Subvolume
'HA-WIN-TT-1T-client-0'
came back up; going
online.</div>
<div>[2014-10-13
17:38:26.955297] I
[client-handshake.c:450:client_set_lk_version_cbk]
0-HA-WIN-TT-1T-client-0:
Server lk version = 1</div>
<div>[2014-10-13
17:38:26.955583] I
[afr-self-heald.c:1690:afr_dir_exclusive_crawl]
0-HA-2TB-TT-Proxmox-cluster-replicate-0:
Another crawl is in
progress for
HA-2TB-TT-Proxmox-cluster-client-0</div>
<div>[2014-10-13
17:38:26.955589] E
[afr-self-heald.c:1479:afr_find_child_position]
0-HA-2TB-TT-Proxmox-cluster-replicate-0:
getxattr failed on
HA-2TB-TT-Proxmox-cluster-client-1
- (Transport endpoint is
not connected)</div>
<div>[2014-10-13
17:38:26.955832] I
[afr-self-heald.c:1690:afr_dir_exclusive_crawl]
0-HA-WIN-TT-1T-replicate-0:
Another crawl is in
progress for
HA-WIN-TT-1T-client-0</div>
<div>[2014-10-13
17:38:26.955858] E
[afr-self-heald.c:1479:afr_find_child_position]
0-HA-WIN-TT-1T-replicate-0:
getxattr failed on
HA-WIN-TT-1T-client-1 -
(Transport endpoint is not
connected)</div>
<div>[2014-10-13
17:38:26.964913] I
[rpc-clnt.c:1729:rpc_clnt_reconfig]
0-HA-2TB-TT-Proxmox-cluster-client-1:
changing port to 49159
(from 0)</div>
<div>[2014-10-13
17:38:26.965553] I
[client-handshake.c:1677:select_server_supported_programs]
0-HA-2TB-TT-Proxmox-cluster-client-1:
Using Program GlusterFS
3.3, Num (1298437),
Version (330)</div>
<div>[2014-10-13
17:38:26.965794] I
[client-handshake.c:1462:client_setvolume_cbk]
0-HA-2TB-TT-Proxmox-cluster-client-1:
Connected to <a
moz-do-not-send="true"
href="http://10.250.0.2:49159"
target="_blank">10.250.0.2:49159</a>,
attached to remote volume
'/exports/HA-2TB-TT-Proxmox-cluster/2TB'.</div>
<div>[2014-10-13
17:38:26.965815] I
[client-handshake.c:1474:client_setvolume_cbk]
0-HA-2TB-TT-Proxmox-cluster-client-1:
Server and Client
lk-version numbers are not
same, reopening the fds</div>
<div>[2014-10-13
17:38:26.965968] I
[client-handshake.c:450:client_set_lk_version_cbk]
0-HA-2TB-TT-Proxmox-cluster-client-1:
Server lk version = 1</div>
<div>[2014-10-13
17:38:26.967510] I
[afr-self-heald.c:1690:afr_dir_exclusive_crawl]
0-HA-2TB-TT-Proxmox-cluster-replicate-0:
Another crawl is in
progress for
HA-2TB-TT-Proxmox-cluster-client-0</div>
<div>[2014-10-13
17:38:27.971374] I
[rpc-clnt.c:1729:rpc_clnt_reconfig]
0-HA-WIN-TT-1T-client-1:
changing port to 49160
(from 0)</div>
<div>[2014-10-13
17:38:27.971940] I
[client-handshake.c:1677:select_server_supported_programs]
0-HA-WIN-TT-1T-client-1:
Using Program GlusterFS
3.3, Num (1298437),
Version (330)</div>
<div>[2014-10-13
17:38:27.975460] I
[client-handshake.c:1462:client_setvolume_cbk]
0-HA-WIN-TT-1T-client-1:
Connected to <a
moz-do-not-send="true"
href="http://10.250.0.2:49160"
target="_blank">10.250.0.2:49160</a>,
attached to remote volume
'/exports/NFS-WIN/1T'.</div>
<div>[2014-10-13
17:38:27.975481] I
[client-handshake.c:1474:client_setvolume_cbk]
0-HA-WIN-TT-1T-client-1:
Server and Client
lk-version numbers are not
same, reopening the fds</div>
<div>[2014-10-13
17:38:27.976656] I
[client-handshake.c:450:client_set_lk_version_cbk]
0-HA-WIN-TT-1T-client-1:
Server lk version = 1</div>
<div>[2014-10-13
17:41:05.390992] I
[glusterfsd-mgmt.c:56:mgmt_cbk_spec]
0-mgmt: Volume file
changed</div>
<div>[2014-10-13
17:41:05.408292] I
[glusterfsd-mgmt.c:56:mgmt_cbk_spec]
0-mgmt: Volume file
changed</div>
<div>[2014-10-13
17:41:05.412221] I
[glusterfsd-mgmt.c:1307:mgmt_getspec_cbk]
0-glusterfs: No change in
volfile, continuing</div>
<div>[2014-10-13
17:41:05.417388] I
[glusterfsd-mgmt.c:1307:mgmt_getspec_cbk]
0-glusterfs: No change in
volfile, continuing</div>
<div>root@stor1:~#</div>
</div>
<div><br>
</div>
<div>glusterfshd stor2</div>
<div><br>
</div>
<div>
<div>[2014-10-13
17:38:28.992891] I
[glusterfsd.c:1959:main]
0-/usr/sbin/glusterfs:
Started running
/usr/sbin/glusterfs
version 3.5.2
(/usr/sbin/glusterfs -s
localhost --volfile-id
gluster/glustershd -p
/var/lib/glusterd/glustershd/run/glustershd.pid
-l
/var/log/glusterfs/glustershd.log
-S
/var/run/b1494ca4d047df6e8590d7080131908f.socket
--xlator-option
*replicate*.node-uuid=abf9e3a7-eb91-4273-acdf-876cd6ba1fe3)</div>
<div>[2014-10-13
17:38:28.994439] I
[socket.c:3561:socket_init]
0-socket.glusterfsd: SSL
support is NOT enabled</div>
<div>[2014-10-13
17:38:28.994476] I
[socket.c:3576:socket_init]
0-socket.glusterfsd: using
system polling thread</div>
<div>[2014-10-13
17:38:28.994581] I
[socket.c:3561:socket_init]
0-glusterfs: SSL support
is NOT enabled</div>
<div>[2014-10-13
17:38:28.994594] I
[socket.c:3576:socket_init]
0-glusterfs: using system
polling thread</div>
<div>[2014-10-13
17:38:28.996569] I
[graph.c:254:gf_add_cmdline_options]
0-HA-2TB-TT-Proxmox-cluster-replicate-0:
adding option 'node-uuid'
for volume
'HA-2TB-TT-Proxmox-cluster-replicate-0'
with value
'abf9e3a7-eb91-4273-acdf-876cd6ba1fe3'</div>
<div>[2014-10-13
17:38:28.996585] I
[graph.c:254:gf_add_cmdline_options]
0-HA-WIN-TT-1T-replicate-0:
adding option 'node-uuid'
for volume
'HA-WIN-TT-1T-replicate-0'
with value
'abf9e3a7-eb91-4273-acdf-876cd6ba1fe3'</div>
<div>[2014-10-13
17:38:28.998463] I
[socket.c:3561:socket_init]
0-HA-2TB-TT-Proxmox-cluster-client-1:
SSL support is NOT enabled</div>
<div>[2014-10-13
17:38:28.998483] I
[socket.c:3576:socket_init]
0-HA-2TB-TT-Proxmox-cluster-client-1:
using system polling
thread</div>
<div>[2014-10-13
17:38:28.998695] I
[socket.c:3561:socket_init]
0-HA-2TB-TT-Proxmox-cluster-client-0:
SSL support is NOT enabled</div>
<div>[2014-10-13
17:38:28.998707] I
[socket.c:3576:socket_init]
0-HA-2TB-TT-Proxmox-cluster-client-0:
using system polling
thread</div>
<div>[2014-10-13
17:38:29.000506] I
[socket.c:3561:socket_init]
0-HA-WIN-TT-1T-client-1:
SSL support is NOT enabled</div>
<div>[2014-10-13
17:38:29.000520] I
[socket.c:3576:socket_init]
0-HA-WIN-TT-1T-client-1:
using system polling
thread</div>
<div>[2014-10-13
17:38:29.000723] I
[socket.c:3561:socket_init]
0-HA-WIN-TT-1T-client-0:
SSL support is NOT enabled</div>
<div>[2014-10-13
17:38:29.000734] I
[socket.c:3576:socket_init]
0-HA-WIN-TT-1T-client-0:
using system polling
thread</div>
<div>[2014-10-13
17:38:29.000762] I
[client.c:2294:notify]
0-HA-2TB-TT-Proxmox-cluster-client-0:
parent translators are
ready, attempting connect
on transport</div>
<div>[2014-10-13
17:38:29.001064] I
[client.c:2294:notify]
0-HA-2TB-TT-Proxmox-cluster-client-1:
parent translators are
ready, attempting connect
on transport</div>
<div>[2014-10-13
17:38:29.001639] I
[client.c:2294:notify]
0-HA-WIN-TT-1T-client-0:
parent translators are
ready, attempting connect
on transport</div>
<div>[2014-10-13
17:38:29.001877] I
[client.c:2294:notify]
0-HA-WIN-TT-1T-client-1:
parent translators are
ready, attempting connect
on transport</div>
<div>Final graph:</div>
<div>+------------------------------------------------------------------------------+</div>
<div> 1: volume
HA-2TB-TT-Proxmox-cluster-client-0</div>
<div> 2: type
protocol/client</div>
<div> 3: option
remote-host stor1</div>
<div> 4: option
remote-subvolume
/exports/HA-2TB-TT-Proxmox-cluster/2TB</div>
<div> 5: option
transport-type socket</div>
<div> 6: option
username
59c66122-55c1-4c28-956e-6189fcb1aff5</div>
<div> 7: option
password
34b79afb-a93c-431b-900a-b688e67cdbc9</div>
<div> 8: option
ping-timeout 10</div>
<div> 9: end-volume</div>
<div> 10:</div>
<div> 11: volume
HA-2TB-TT-Proxmox-cluster-client-1</div>
<div> 12: type
protocol/client</div>
<div> 13: option
remote-host stor2</div>
<div> 14: option
remote-subvolume
/exports/HA-2TB-TT-Proxmox-cluster/2TB</div>
<div> 15: option
transport-type socket</div>
<div> 16: option
username
59c66122-55c1-4c28-956e-6189fcb1aff5</div>
<div> 17: option
password
34b79afb-a93c-431b-900a-b688e67cdbc9</div>
<div> 18: option
ping-timeout 10</div>
<div> 19: end-volume</div>
<div> 20:</div>
<div> 21: volume
HA-2TB-TT-Proxmox-cluster-replicate-0</div>
<div> 22: type
cluster/replicate</div>
<div> 23: option
node-uuid
abf9e3a7-eb91-4273-acdf-876cd6ba1fe3</div>
<div> 24: option
background-self-heal-count
0</div>
<div> 25: option
metadata-self-heal on</div>
<div> 26: option
data-self-heal on</div>
<div> 27: option
entry-self-heal on</div>
<div> 28: option
self-heal-daemon on</div>
<div> 29: option
iam-self-heal-daemon yes</div>
<div> 30: subvolumes
HA-2TB-TT-Proxmox-cluster-client-0
HA-2TB-TT-Proxmox-cluster-client-1</div>
<div> 31: end-volume</div>
<div> 32:</div>
<div> 33: volume
HA-WIN-TT-1T-client-0</div>
<div> 34: type
protocol/client</div>
<div> 35: option
remote-host stor1</div>
<div> 36: option
remote-subvolume
/exports/NFS-WIN/1T</div>
<div> 37: option
transport-type socket</div>
<div> 38: option
username
101b907c-ff21-47da-8ba6-37e2920691ce</div>
<div> 39: option
password
f4f29094-891f-4241-8736-5e3302ed8bc8</div>
<div> 40: option
ping-timeout 10</div>
<div> 41: end-volume</div>
<div> 42:</div>
<div> 43: volume
HA-WIN-TT-1T-client-1</div>
<div> 44: type
protocol/client</div>
<div> 45: option
remote-host stor2</div>
<div> 46: option
remote-subvolume
/exports/NFS-WIN/1T</div>
<div> 47: option
transport-type socket</div>
<div> 48: option
username
101b907c-ff21-47da-8ba6-37e2920691ce</div>
<div> 49: option
password
f4f29094-891f-4241-8736-5e3302ed8bc8</div>
<div> 50: option
ping-timeout 10</div>
<div> 51: end-volume</div>
<div> 52:</div>
<div> 53: volume
HA-WIN-TT-1T-replicate-0</div>
<div> 54: type
cluster/replicate</div>
<div> 55: option
node-uuid
abf9e3a7-eb91-4273-acdf-876cd6ba1fe3</div>
<div> 56: option
background-self-heal-count
0</div>
<div> 57: option
metadata-self-heal on</div>
<div> 58: option
data-self-heal on</div>
<div> 59: option
entry-self-heal on</div>
<div> 60: option
self-heal-daemon on</div>
<div> 61: option
iam-self-heal-daemon yes</div>
<div> 62: subvolumes
HA-WIN-TT-1T-client-0
HA-WIN-TT-1T-client-1</div>
<div> 63: end-volume</div>
<div> 64:</div>
<div> 65: volume glustershd</div>
<div> 66: type
debug/io-stats</div>
<div> 67: subvolumes
HA-2TB-TT-Proxmox-cluster-replicate-0
HA-WIN-TT-1T-replicate-0</div>
<div> 68: end-volume</div>
<div> 69:</div>
<div>+------------------------------------------------------------------------------+</div>
<div>[2014-10-13
17:38:29.002743] I
[rpc-clnt.c:1729:rpc_clnt_reconfig]
0-HA-2TB-TT-Proxmox-cluster-client-1:
changing port to 49159
(from 0)</div>
<div>[2014-10-13
17:38:29.003027] I
[rpc-clnt.c:1729:rpc_clnt_reconfig]
0-HA-WIN-TT-1T-client-1:
changing port to 49160
(from 0)</div>
<div>[2014-10-13
17:38:29.003290] I
[rpc-clnt.c:1729:rpc_clnt_reconfig]
0-HA-2TB-TT-Proxmox-cluster-client-0:
changing port to 49159
(from 0)</div>
<div>[2014-10-13
17:38:29.003334] I
[rpc-clnt.c:1729:rpc_clnt_reconfig]
0-HA-WIN-TT-1T-client-0:
changing port to 49160
(from 0)</div>
<div>[2014-10-13
17:38:29.003922] I
[client-handshake.c:1677:select_server_supported_programs]
0-HA-2TB-TT-Proxmox-cluster-client-1:
Using Program GlusterFS
3.3, Num (1298437),
Version (330)</div>
<div>[2014-10-13
17:38:29.004023] I
[client-handshake.c:1677:select_server_supported_programs]
0-HA-WIN-TT-1T-client-1:
Using Program GlusterFS
3.3, Num (1298437),
Version (330)</div>
<div>[2014-10-13
17:38:29.004139] I
[client-handshake.c:1677:select_server_supported_programs]
0-HA-2TB-TT-Proxmox-cluster-client-0:
Using Program GlusterFS
3.3, Num (1298437),
Version (330)</div>
<div>[2014-10-13
17:38:29.004202] I
[client-handshake.c:1462:client_setvolume_cbk]
0-HA-2TB-TT-Proxmox-cluster-client-1:
Connected to <a
moz-do-not-send="true"
href="http://10.250.0.2:49159"
target="_blank">10.250.0.2:49159</a>,
attached to remote volume
'/exports/HA-2TB-TT-Proxmox-cluster/2TB'.</div>
<div>[2014-10-13
17:38:29.004217] I
[client-handshake.c:1474:client_setvolume_cbk]
0-HA-2TB-TT-Proxmox-cluster-client-1:
Server and Client
lk-version numbers are not
same, reopening the fds</div>
<div>[2014-10-13
17:38:29.004266] I
[afr-common.c:4131:afr_notify]
0-HA-2TB-TT-Proxmox-cluster-replicate-0:
Subvolume
'HA-2TB-TT-Proxmox-cluster-client-1'
came back up; going
online.</div>
<div>[2014-10-13
17:38:29.004318] I
[client-handshake.c:1677:select_server_supported_programs]
0-HA-WIN-TT-1T-client-0:
Using Program GlusterFS
3.3, Num (1298437),
Version (330)</div>
<div>[2014-10-13
17:38:29.004368] I
[client-handshake.c:1462:client_setvolume_cbk]
0-HA-WIN-TT-1T-client-1:
Connected to <a
moz-do-not-send="true"
href="http://10.250.0.2:49160"
target="_blank">10.250.0.2:49160</a>,
attached to remote volume
'/exports/NFS-WIN/1T'.</div>
<div>[2014-10-13
17:38:29.004383] I
[client-handshake.c:1474:client_setvolume_cbk]
0-HA-WIN-TT-1T-client-1:
Server and Client
lk-version numbers are not
same, reopening the fds</div>
<div>[2014-10-13
17:38:29.004429] I
[afr-common.c:4131:afr_notify]
0-HA-WIN-TT-1T-replicate-0:
Subvolume
'HA-WIN-TT-1T-client-1'
came back up; going
online.</div>
<div>[2014-10-13
17:38:29.004483] I
[client-handshake.c:450:client_set_lk_version_cbk]
0-HA-2TB-TT-Proxmox-cluster-client-1:
Server lk version = 1</div>
<div>[2014-10-13
17:38:29.004506] I
[client-handshake.c:450:client_set_lk_version_cbk]
0-HA-WIN-TT-1T-client-1:
Server lk version = 1</div>
<div>[2014-10-13
17:38:29.004526] I
[client-handshake.c:1462:client_setvolume_cbk]
0-HA-2TB-TT-Proxmox-cluster-client-0:
Connected to <a
moz-do-not-send="true"
href="http://10.250.0.1:49159"
target="_blank">10.250.0.1:49159</a>,
attached to remote volume
'/exports/HA-2TB-TT-Proxmox-cluster/2TB'.</div>
<div>[2014-10-13
17:38:29.004535] I
[client-handshake.c:1474:client_setvolume_cbk]
0-HA-2TB-TT-Proxmox-cluster-client-0:
Server and Client
lk-version numbers are not
same, reopening the fds</div>
<div>[2014-10-13
17:38:29.004613] I
[client-handshake.c:1462:client_setvolume_cbk]
0-HA-WIN-TT-1T-client-0:
Connected to <a
moz-do-not-send="true"
href="http://10.250.0.1:49160"
target="_blank">10.250.0.1:49160</a>,
attached to remote volume
'/exports/NFS-WIN/1T'.</div>
<div>[2014-10-13
17:38:29.004626] I
[client-handshake.c:1474:client_setvolume_cbk]
0-HA-WIN-TT-1T-client-0:
Server and Client
lk-version numbers are not
same, reopening the fds</div>
<div>[2014-10-13
17:38:29.004731] I
[client-handshake.c:450:client_set_lk_version_cbk]
0-HA-2TB-TT-Proxmox-cluster-client-0:
Server lk version = 1</div>
<div>[2014-10-13
17:38:29.004796] I
[client-handshake.c:450:client_set_lk_version_cbk]
0-HA-WIN-TT-1T-client-0:
Server lk version = 1</div>
<div>[2014-10-13
17:38:29.005291] I
[afr-self-heald.c:1690:afr_dir_exclusive_crawl]
0-HA-WIN-TT-1T-replicate-0:
Another crawl is in
progress for
HA-WIN-TT-1T-client-1</div>
<div>[2014-10-13
17:38:29.005303] I
[afr-self-heald.c:1690:afr_dir_exclusive_crawl]
0-HA-2TB-TT-Proxmox-cluster-replicate-0:
Another crawl is in
progress for
HA-2TB-TT-Proxmox-cluster-client-1</div>
<div>[2014-10-13
17:38:29.005443] I
[afr-self-heald.c:1690:afr_dir_exclusive_crawl]
0-HA-2TB-TT-Proxmox-cluster-replicate-0:
Another crawl is in
progress for
HA-2TB-TT-Proxmox-cluster-client-1</div>
<div>[2014-10-13
17:41:05.427867] I
[glusterfsd-mgmt.c:56:mgmt_cbk_spec]
0-mgmt: Volume file
changed</div>
<div>[2014-10-13
17:41:05.443271] I
[glusterfsd-mgmt.c:56:mgmt_cbk_spec]
0-mgmt: Volume file
changed</div>
<div>[2014-10-13
17:41:05.444111] I
[glusterfsd-mgmt.c:1307:mgmt_getspec_cbk]
0-glusterfs: No change in
volfile, continuing</div>
<div>[2014-10-13
17:41:05.444807] I
[glusterfsd-mgmt.c:1307:mgmt_getspec_cbk]
0-glusterfs: No change in
volfile, continuing</div>
</div>
<div><br>
</div>
<div>brick stor2</div>
<div><br>
</div>
<div>
<div>[2014-10-13
17:38:17.213386] W
[glusterfsd.c:1095:cleanup_and_exit]
(-->/lib/x86_64-linux-gnu/libc.so.6(+0x462a0)
[0x7f343271f2a0]
(-->/usr/lib/x86_64-linux-gnu/libglusterfs.so.0(synctask_wrap+0x12)
[0x7f343371db12]
(-->/usr/sbin/glusterfsd(glusterfs_handle_terminate+0x15)
[0x7f3434790dd5]))) 0-:
received signum (15),
shutting down</div>
<div>[2014-10-13
17:38:26.957312] I
[glusterfsd.c:1959:main]
0-/usr/sbin/glusterfsd:
Started running
/usr/sbin/glusterfsd
version 3.5.2
(/usr/sbin/glusterfsd -s
stor2 --volfile-id
HA-WIN-TT-1T.stor2.exports-NFS-WIN-1T
-p
/var/lib/glusterd/vols/HA-WIN-TT-1T/run/stor2-exports-NFS-WIN-1T.pid
-S
/var/run/91514691033d00e666bb151f9c771a26.socket
--brick-name
/exports/NFS-WIN/1T -l
/var/log/glusterfs/bricks/exports-NFS-WIN-1T.log
--xlator-option
*-posix.glusterd-uuid=abf9e3a7-eb91-4273-acdf-876cd6ba1fe3
--brick-port 49160
--xlator-option
HA-WIN-TT-1T-server.listen-port=49160)</div>
<div>[2014-10-13
17:38:26.958864] I
[socket.c:3561:socket_init]
0-socket.glusterfsd: SSL
support is NOT enabled</div>
<div>[2014-10-13
17:38:26.958899] I
[socket.c:3576:socket_init]
0-socket.glusterfsd: using
system polling thread</div>
<div>[2014-10-13
17:38:26.959003] I
[socket.c:3561:socket_init]
0-glusterfs: SSL support
is NOT enabled</div>
<div>[2014-10-13
17:38:26.959015] I
[socket.c:3576:socket_init]
0-glusterfs: using system
polling thread</div>
<div>[2014-10-13
17:38:26.961860] I
[graph.c:254:gf_add_cmdline_options]
0-HA-WIN-TT-1T-server:
adding option
'listen-port' for volume
'HA-WIN-TT-1T-server' with
value '49160'</div>
<div>[2014-10-13
17:38:26.961878] I
[graph.c:254:gf_add_cmdline_options]
0-HA-WIN-TT-1T-posix:
adding option
'glusterd-uuid' for volume
'HA-WIN-TT-1T-posix' with
value
'abf9e3a7-eb91-4273-acdf-876cd6ba1fe3'</div>
<div>[2014-10-13
17:38:26.965032] I
[rpcsvc.c:2127:rpcsvc_set_outstanding_rpc_limit]
0-rpc-service: Configured
rpc.outstanding-rpc-limit
with value 64</div>
<div>[2014-10-13
17:38:26.965075] W
[options.c:888:xl_opt_validate]
0-HA-WIN-TT-1T-server:
option 'listen-port' is
deprecated, preferred is
'transport.socket.listen-port',
continuing with correction</div>
<div>[2014-10-13
17:38:26.965097] I
[socket.c:3561:socket_init]
0-tcp.HA-WIN-TT-1T-server:
SSL support is NOT enabled</div>
<div>[2014-10-13
17:38:26.965105] I
[socket.c:3576:socket_init]
0-tcp.HA-WIN-TT-1T-server:
using system polling
thread</div>
<div>[2014-10-13
17:38:26.965602] W
[graph.c:329:_log_if_unknown_option]
0-HA-WIN-TT-1T-quota:
option 'timeout' is not
recognized</div>
<div>Final graph:</div>
<div>+------------------------------------------------------------------------------+</div>
<div> 1: volume
HA-WIN-TT-1T-posix</div>
<div> 2: type
storage/posix</div>
<div> 3: option
glusterd-uuid
abf9e3a7-eb91-4273-acdf-876cd6ba1fe3</div>
<div> 4: option
directory
/exports/NFS-WIN/1T</div>
<div> 5: option
volume-id
2937ac01-4cba-44a8-8ff8-0161b67f8ee4</div>
<div> 6: end-volume</div>
<div> 7:</div>
<div> 8: volume
HA-WIN-TT-1T-changelog</div>
<div> 9: type
features/changelog</div>
<div> 10: option
changelog-brick
/exports/NFS-WIN/1T</div>
<div> 11: option
changelog-dir
/exports/NFS-WIN/1T/.glusterfs/changelogs</div>
<div> 12: subvolumes
HA-WIN-TT-1T-posix</div>
<div> 13: end-volume</div>
<div> 14:</div>
<div> 15: volume
HA-WIN-TT-1T-access-control</div>
<div> 16: type
features/access-control</div>
<div> 17: subvolumes
HA-WIN-TT-1T-changelog</div>
<div> 18: end-volume</div>
<div> 19:</div>
<div> 20: volume
HA-WIN-TT-1T-locks</div>
<div> 21: type
features/locks</div>
<div> 22: subvolumes
HA-WIN-TT-1T-access-control</div>
<div> 23: end-volume</div>
<div> 24:</div>
<div> 25: volume
HA-WIN-TT-1T-io-threads</div>
<div> 26: type
performance/io-threads</div>
<div> 27: subvolumes
HA-WIN-TT-1T-locks</div>
<div> 28: end-volume</div>
<div> 29:</div>
<div> 30: volume
HA-WIN-TT-1T-index</div>
<div> 31: type
features/index</div>
<div> 32: option
index-base
/exports/NFS-WIN/1T/.glusterfs/indices</div>
<div> 33: subvolumes
HA-WIN-TT-1T-io-threads</div>
<div> 34: end-volume</div>
<div> 35:</div>
<div> 36: volume
HA-WIN-TT-1T-marker</div>
<div> 37: type
features/marker</div>
<div> 38: option
volume-uuid
2937ac01-4cba-44a8-8ff8-0161b67f8ee4</div>
<div> 39: option
timestamp-file
/var/lib/glusterd/vols/HA-WIN-TT-1T/marker.tstamp</div>
<div> 40: option xtime
off</div>
<div> 41: option
gsync-force-xtime off</div>
<div> 42: option quota
off</div>
<div> 43: subvolumes
HA-WIN-TT-1T-index</div>
<div> 44: end-volume</div>
<div> 45:</div>
<div> 46: volume
HA-WIN-TT-1T-quota</div>
<div> 47: type
features/quota</div>
<div> 48: option
volume-uuid HA-WIN-TT-1T</div>
<div> 49: option
server-quota off</div>
<div> 50: option timeout
0</div>
<div> 51: option
deem-statfs off</div>
<div> 52: subvolumes
HA-WIN-TT-1T-marker</div>
<div> 53: end-volume</div>
<div> 54:</div>
<div> 55: volume
/exports/NFS-WIN/1T</div>
<div> 56: type
debug/io-stats</div>
<div> 57: option
latency-measurement off</div>
<div> 58: option
count-fop-hits off</div>
<div> 59: subvolumes
HA-WIN-TT-1T-quota</div>
<div> 60: end-volume</div>
<div> 61:</div>
<div> 62: volume
HA-WIN-TT-1T-server</div>
<div> 63: type
protocol/server</div>
<div> 64: option
transport.socket.listen-port
49160</div>
<div> 65: option
rpc-auth.auth-glusterfs on</div>
<div> 66: option
rpc-auth.auth-unix on</div>
<div> 67: option
rpc-auth.auth-null on</div>
<div> 68: option
transport-type tcp</div>
<div> 69: option
auth.login./exports/NFS-WIN/1T.allow
101b907c-ff21-47da-8ba6-37e2920691ce</div>
<div> 70: option
auth.login.101b907c-ff21-47da-8ba6-37e2920691ce.password
f4f29094-891f-4241-8736-5e3302ed8bc8</div>
<div> 71: option
auth.addr./exports/NFS-WIN/1T.allow
*</div>
<div> 72: subvolumes
/exports/NFS-WIN/1T</div>
<div> 73: end-volume</div>
<div> 74:</div>
<div>+------------------------------------------------------------------------------+</div>
<div>[2014-10-13
17:38:27.985048] I
[server-handshake.c:575:server_setvolume]
0-HA-WIN-TT-1T-server:
accepted client from
stor1-14362-2014/10/13-17:38:26:938194-HA-WIN-TT-1T-client-1-0-0
(version: 3.5.2)</div>
<div>[2014-10-13
17:38:28.988700] I
[server-handshake.c:575:server_setvolume]
0-HA-WIN-TT-1T-server:
accepted client from
glstor-cli-20753-2014/10/13-11:50:40:959211-HA-WIN-TT-1T-client-1-0-1
(version: 3.5.2)</div>
<div>[2014-10-13
17:38:29.004121] I
[server-handshake.c:575:server_setvolume]
0-HA-WIN-TT-1T-server:
accepted client from
stor2-15494-2014/10/13-17:38:28:989227-HA-WIN-TT-1T-client-1-0-0
(version: 3.5.2)</div>
<div>[2014-10-13
17:38:38.515315] I
[server-handshake.c:575:server_setvolume]
0-HA-WIN-TT-1T-server:
accepted client from
glstor-cli-23823-2014/10/13-17:37:54:595571-HA-WIN-TT-1T-client-1-0-0
(version: 3.5.2)</div>
<div>[2014-10-13
17:39:09.872223] I
[server.c:520:server_rpc_notify]
0-HA-WIN-TT-1T-server:
disconnecting
connectionfrom
glstor-cli-20753-2014/10/13-11:50:40:959211-HA-WIN-TT-1T-client-1-0-1</div>
<div>[2014-10-13
17:39:09.872299] I
[client_t.c:417:gf_client_unref]
0-HA-WIN-TT-1T-server:
Shutting down connection
glstor-cli-20753-2014/10/13-11:50:40:959211-HA-WIN-TT-1T-client-1-0-1</div>
<div>[2014-10-13
17:41:05.427810] I
[glusterfsd-mgmt.c:56:mgmt_cbk_spec]
0-mgmt: Volume file
changed</div>
<div>[2014-10-13
17:41:05.443234] I
[glusterfsd-mgmt.c:56:mgmt_cbk_spec]
0-mgmt: Volume file
changed</div>
<div>[2014-10-13
17:41:05.445049] I
[glusterfsd-mgmt.c:1307:mgmt_getspec_cbk]
0-glusterfs: No change in
volfile, continuing</div>
<div>root@stor2:~#</div>
</div>
<div><br>
</div>
<div>brick stor1</div>
<div><br>
</div>
<div>
<div>[2014-10-13
17:38:24.900066] I
[glusterfsd.c:1959:main]
0-/usr/sbin/glusterfsd:
Started running
/usr/sbin/glusterfsd
version 3.5.2
(/usr/sbin/glusterfsd -s
stor1 --volfile-id
HA-WIN-TT-1T.stor1.exports-NFS-WIN-1T
-p
/var/lib/glusterd/vols/HA-WIN-TT-1T/run/stor1-exports-NFS-WIN-1T.pid
-S
/var/run/02580c93278849804f3f34f7ed8314b2.socket
--brick-name
/exports/NFS-WIN/1T -l
/var/log/glusterfs/bricks/exports-NFS-WIN-1T.log
--xlator-option
*-posix.glusterd-uuid=e09cbbc2-08a3-4e5b-83b8-48eb11a1c7b3
--brick-port 49160
--xlator-option
HA-WIN-TT-1T-server.listen-port=49160)</div>
<div>[2014-10-13
17:38:24.902022] I
[socket.c:3561:socket_init]
0-socket.glusterfsd: SSL
support is NOT enabled</div>
<div>[2014-10-13
17:38:24.902077] I
[socket.c:3576:socket_init]
0-socket.glusterfsd: using
system polling thread</div>
<div>[2014-10-13
17:38:24.902214] I
[socket.c:3561:socket_init]
0-glusterfs: SSL support
is NOT enabled</div>
<div>[2014-10-13
17:38:24.902239] I
[socket.c:3576:socket_init]
0-glusterfs: using system
polling thread</div>
<div>[2014-10-13
17:38:24.906698] I
[graph.c:254:gf_add_cmdline_options]
0-HA-WIN-TT-1T-server:
adding option
'listen-port' for volume
'HA-WIN-TT-1T-server' with
value '49160'</div>
<div>[2014-10-13
17:38:24.906731] I
[graph.c:254:gf_add_cmdline_options]
0-HA-WIN-TT-1T-posix:
adding option
'glusterd-uuid' for volume
'HA-WIN-TT-1T-posix' with
value
'e09cbbc2-08a3-4e5b-83b8-48eb11a1c7b3'</div>
<div>[2014-10-13
17:38:24.908378] I
[rpcsvc.c:2127:rpcsvc_set_outstanding_rpc_limit]
0-rpc-service: Configured
rpc.outstanding-rpc-limit
with value 64</div>
<div>[2014-10-13
17:38:24.908435] W
[options.c:888:xl_opt_validate]
0-HA-WIN-TT-1T-server:
option 'listen-port' is
deprecated, preferred is
'transport.socket.listen-port',
continuing with correction</div>
<div>[2014-10-13
17:38:24.908472] I
[socket.c:3561:socket_init]
0-tcp.HA-WIN-TT-1T-server:
SSL support is NOT enabled</div>
<div>[2014-10-13
17:38:24.908485] I
[socket.c:3576:socket_init]
0-tcp.HA-WIN-TT-1T-server:
using system polling
thread</div>
<div>[2014-10-13
17:38:24.909105] W
[graph.c:329:_log_if_unknown_option]
0-HA-WIN-TT-1T-quota:
option 'timeout' is not
recognized</div>
<div>Final graph:</div>
<div>+------------------------------------------------------------------------------+</div>
<div> 1: volume
HA-WIN-TT-1T-posix</div>
<div> 2: type
storage/posix</div>
<div> 3: option
glusterd-uuid
e09cbbc2-08a3-4e5b-83b8-48eb11a1c7b3</div>
<div> 4: option
directory
/exports/NFS-WIN/1T</div>
<div> 5: option
volume-id
2937ac01-4cba-44a8-8ff8-0161b67f8ee4</div>
<div> 6: end-volume</div>
<div> 7:</div>
<div> 8: volume
HA-WIN-TT-1T-changelog</div>
<div> 9: type
features/changelog</div>
<div> 10: option
changelog-brick
/exports/NFS-WIN/1T</div>
<div> 11: option
changelog-dir
/exports/NFS-WIN/1T/.glusterfs/changelogs</div>
<div> 12: subvolumes
HA-WIN-TT-1T-posix</div>
<div> 13: end-volume</div>
<div> 14:</div>
<div> 15: volume
HA-WIN-TT-1T-access-control</div>
<div> 16: type
features/access-control</div>
<div> 17: subvolumes
HA-WIN-TT-1T-changelog</div>
<div> 18: end-volume</div>
<div> 19:</div>
<div> 20: volume
HA-WIN-TT-1T-locks</div>
<div> 21: type
features/locks</div>
<div> 22: subvolumes
HA-WIN-TT-1T-access-control</div>
<div> 23: end-volume</div>
<div> 24:</div>
<div> 25: volume
HA-WIN-TT-1T-io-threads</div>
<div> 26: type
performance/io-threads</div>
<div> 27: subvolumes
HA-WIN-TT-1T-locks</div>
<div> 28: end-volume</div>
<div> 29:</div>
<div> 30: volume
HA-WIN-TT-1T-index</div>
<div> 31: type
features/index</div>
<div> 32: option
index-base
/exports/NFS-WIN/1T/.glusterfs/indices</div>
<div> 33: subvolumes
HA-WIN-TT-1T-io-threads</div>
<div> 34: end-volume</div>
<div> 35:</div>
<div> 36: volume
HA-WIN-TT-1T-marker</div>
<div> 37: type
features/marker</div>
<div> 38: option
volume-uuid
2937ac01-4cba-44a8-8ff8-0161b67f8ee4</div>
<div> 39: option
timestamp-file
/var/lib/glusterd/vols/HA-WIN-TT-1T/marker.tstamp</div>
<div> 40: option xtime
off</div>
<div> 41: option
gsync-force-xtime off</div>
<div> 42: option quota
off</div>
<div> 43: subvolumes
HA-WIN-TT-1T-index</div>
<div> 44: end-volume</div>
<div> 45:</div>
<div> 46: volume
HA-WIN-TT-1T-quota</div>
<div> 47: type
features/quota</div>
<div> 48: option
volume-uuid HA-WIN-TT-1T</div>
<div> 49: option
server-quota off</div>
<div> 50: option timeout
0</div>
<div> 51: option
deem-statfs off</div>
<div> 52: subvolumes
HA-WIN-TT-1T-marker</div>
<div> 53: end-volume</div>
<div> 54:</div>
<div> 55: volume
/exports/NFS-WIN/1T</div>
<div> 56: type
debug/io-stats</div>
<div> 57: option
latency-measurement off</div>
<div> 58: option
count-fop-hits off</div>
<div> 59: subvolumes
HA-WIN-TT-1T-quota</div>
<div> 60: end-volume</div>
<div> 61:</div>
<div> 62: volume
HA-WIN-TT-1T-server</div>
<div> 63: type
protocol/server</div>
<div> 64: option
transport.socket.listen-port
49160</div>
<div> 65: option
rpc-auth.auth-glusterfs on</div>
<div> 66: option
rpc-auth.auth-unix on</div>
<div> 67: option
rpc-auth.auth-null on</div>
<div> 68: option
transport-type tcp</div>
<div> 69: option
auth.login./exports/NFS-WIN/1T.allow
101b907c-ff21-47da-8ba6-37e2920691ce</div>
<div> 70: option
auth.login.101b907c-ff21-47da-8ba6-37e2920691ce.password
f4f29094-891f-4241-8736-5e3302ed8bc8</div>
<div> 71: option
auth.addr./exports/NFS-WIN/1T.allow
*</div>
<div> 72: subvolumes
/exports/NFS-WIN/1T</div>
<div> 73: end-volume</div>
<div> 74:</div>
<div>+------------------------------------------------------------------------------+</div>
<div>[2014-10-13
17:38:25.933796] I
[server-handshake.c:575:server_setvolume]
0-HA-WIN-TT-1T-server:
accepted client from
glstor-cli-20753-2014/10/13-11:50:40:959211-HA-WIN-TT-1T-client-0-0-1
(version: 3.5.2)</div>
<div>[2014-10-13
17:38:26.954924] I
[server-handshake.c:575:server_setvolume]
0-HA-WIN-TT-1T-server:
accepted client from
stor1-14362-2014/10/13-17:38:26:938194-HA-WIN-TT-1T-client-0-0-0
(version: 3.5.2)</div>
<div>[2014-10-13
17:38:28.991488] I
[server-handshake.c:575:server_setvolume]
0-HA-WIN-TT-1T-server:
accepted client from
stor2-15494-2014/10/13-17:38:28:989227-HA-WIN-TT-1T-client-0-0-0
(version: 3.5.2)</div>
<div>[2014-10-13
17:38:38.502056] I
[server-handshake.c:575:server_setvolume]
0-HA-WIN-TT-1T-server:
accepted client from
glstor-cli-23823-2014/10/13-17:37:54:595571-HA-WIN-TT-1T-client-0-0-0
(version: 3.5.2)</div>
<div>[2014-10-13
17:39:09.858784] I
[server.c:520:server_rpc_notify]
0-HA-WIN-TT-1T-server:
disconnecting
connectionfrom
glstor-cli-20753-2014/10/13-11:50:40:959211-HA-WIN-TT-1T-client-0-0-1</div>
<div>[2014-10-13
17:39:09.858863] I
[client_t.c:417:gf_client_unref]
0-HA-WIN-TT-1T-server:
Shutting down connection
glstor-cli-20753-2014/10/13-11:50:40:959211-HA-WIN-TT-1T-client-0-0-1</div>
<div>[2014-10-13
17:41:05.390918] I
[glusterfsd-mgmt.c:56:mgmt_cbk_spec]
0-mgmt: Volume file
changed</div>
<div>[2014-10-13
17:41:05.408236] I
[glusterfsd-mgmt.c:56:mgmt_cbk_spec]
0-mgmt: Volume file
changed</div>
<div>[2014-10-13
17:41:05.414813] I
[glusterfsd-mgmt.c:1307:mgmt_getspec_cbk]
0-glusterfs: No change in
volfile, continuing</div>
</div>
<div><br>
</div>
<div><br>
</div>
<div>seems to be the right
part of logs :)</div>
<div><br>
</div>
</div>
<div class="gmail_extra"><br>
<div class="gmail_quote">2014-10-15
18:24 GMT+03:00 Pranith
Kumar Karampuri <span
dir="ltr"><<a
moz-do-not-send="true"
href="mailto:pkarampu@redhat.com"
target="_blank">pkarampu@redhat.com</a>></span>:<br>
<blockquote
class="gmail_quote"
style="margin:0 0 0
.8ex;border-left:1px #ccc
solid;padding-left:1ex">
<div text="#000000"
bgcolor="#FFFFFF"><span>
<br>
<div>On 10/14/2014
01:20 AM, Roman
wrote:<br>
</div>
<blockquote
type="cite">
<div dir="ltr">ok.
done.
<div>this time
there were no
disconnects, at
least all of vms
are working, but
got some mails
from VM about IO
writes again.</div>
<div><span
style="font-size:11pt;font-family:Calibri,sans-serif"><br>
</span></div>
<div><span
style="font-size:11pt;font-family:Calibri,sans-serif">WARNINGs:
Read IO Wait
time is 1.45
(outside range
[0:1]).</span><br>
</div>
</div>
</blockquote>
</span> This warning
says 'Read IO wait' and
there is not a single
READ operation that came
to gluster. Wondering
why that is :-/. Any
clue? There is at least
one write which took 3
seconds according to the
stats. At least one
synchronization
operation (FINODELK)
took 23 seconds. Could
you give logs of this
run? for mount,
glustershd, bricks.<span><font
color="#888888"><br>
<br>
Pranith</font></span>
<div>
<div><br>
<blockquote
type="cite">
<div dir="ltr">
<div><span
style="font-size:11pt;font-family:Calibri,sans-serif"><br>
</span></div>
<div>here is the
output</div>
<div><br>
</div>
<div>
<div>root@stor1:~#
gluster volume
profile
HA-WIN-TT-1T
info</div>
<div>Brick:
stor1:/exports/NFS-WIN/1T</div>
<div>--------------------------------</div>
<div>Cumulative
Stats:</div>
<div> Block
Size:
131072b+
262144b+</div>
<div> No. of
Reads:
0
0</div>
<div>No. of
Writes:
7372798
1</div>
<div> %-latency
Avg-latency
Min-Latency
Max-Latency
No. of calls
Fop</div>
<div> ---------
-----------
-----------
-----------
------------
----</div>
<div>
0.00
0.00 us
0.00 us
0.00 us
25
RELEASE</div>
<div>
0.00
0.00 us
0.00 us
0.00 us
16
RELEASEDIR</div>
<div>
0.00
64.00 us
52.00 us
76.00 us
2
ENTRYLK</div>
<div>
0.00
73.50 us
51.00 us
96.00 us
2
FLUSH</div>
<div>
0.00
68.43 us
30.00 us
135.00 us
7
STATFS</div>
<div>
0.00
54.31 us
44.00 us
109.00 us
16
OPENDIR</div>
<div>
0.00
50.75 us
16.00 us
74.00 us
24
FSTAT</div>
<div>
0.00
47.77 us
19.00 us
119.00 us
26
GETXATTR</div>
<div>
0.00
59.21 us
21.00 us
89.00 us
24
OPEN</div>
<div>
0.00
59.39 us
22.00 us
296.00 us
28
READDIR</div>
<div>
0.00
4972.00 us
4972.00 us
4972.00 us
1
CREATE</div>
<div>
0.00
97.42 us
19.00 us
184.00 us
62
LOOKUP</div>
<div>
0.00
89.49 us
20.00 us
656.00 us
324
FXATTROP</div>
<div>
3.91
1255944.81 us
127.00 us
23397532.00 us
189
FSYNC</div>
<div>
7.40
3406275.50 us
17.00 us
23398013.00 us
132
INODELK</div>
<div>
34.96
94598.02 us
8.00 us
23398705.00 us
22445
FINODELK</div>
<div>
53.73
442.66 us
79.00 us
3116494.00 us
7372799
WRITE</div>
<div><br>
</div>
<div>
Duration: 7813
seconds</div>
<div> Data
Read: 0 bytes</div>
<div>Data
Written:
966367641600
bytes</div>
<div><br>
</div>
<div>Interval
0 Stats:</div>
<div> Block
Size:
131072b+
262144b+</div>
<div> No. of
Reads:
0
0</div>
<div>No. of
Writes:
7372798
1</div>
<div> %-latency
Avg-latency
Min-Latency
Max-Latency
No. of calls
Fop</div>
<div> ---------
-----------
-----------
-----------
------------
----</div>
<div>
0.00
0.00 us
0.00 us
0.00 us
25
RELEASE</div>
<div>
0.00
0.00 us
0.00 us
0.00 us
16
RELEASEDIR</div>
<div>
0.00
64.00 us
52.00 us
76.00 us
2
ENTRYLK</div>
<div>
0.00
73.50 us
51.00 us
96.00 us
2
FLUSH</div>
<div>
0.00
68.43 us
30.00 us
135.00 us
7
STATFS</div>
<div>
0.00
54.31 us
44.00 us
109.00 us
16
OPENDIR</div>
<div>
0.00
50.75 us
16.00 us
74.00 us
24
FSTAT</div>
<div>
0.00
47.77 us
19.00 us
119.00 us
26
GETXATTR</div>
<div>
0.00
59.21 us
21.00 us
89.00 us
24
OPEN</div>
<div>
0.00
59.39 us
22.00 us
296.00 us
28
READDIR</div>
<div>
0.00
4972.00 us
4972.00 us
4972.00 us
1
CREATE</div>
<div>
0.00
97.42 us
19.00 us
184.00 us
62
LOOKUP</div>
<div>
0.00
89.49 us
20.00 us
656.00 us
324
FXATTROP</div>
<div>
3.91
1255944.81 us
127.00 us
23397532.00 us
189
FSYNC</div>
<div>
7.40
3406275.50 us
17.00 us
23398013.00 us
132
INODELK</div>
<div>
34.96
94598.02 us
8.00 us
23398705.00 us
22445
FINODELK</div>
<div>
53.73
442.66 us
79.00 us
3116494.00 us
7372799
WRITE</div>
<div><br>
</div>
<div>
Duration: 7813
seconds</div>
<div> Data
Read: 0 bytes</div>
<div>Data
Written:
966367641600
bytes</div>
<div><br>
</div>
<div>Brick:
stor2:/exports/NFS-WIN/1T</div>
<div>--------------------------------</div>
<div>Cumulative
Stats:</div>
<div> Block
Size:
131072b+
262144b+</div>
<div> No. of
Reads:
0
0</div>
<div>No. of
Writes:
7372798
1</div>
<div> %-latency
Avg-latency
Min-Latency
Max-Latency
No. of calls
Fop</div>
<div> ---------
-----------
-----------
-----------
------------
----</div>
<div>
0.00
0.00 us
0.00 us
0.00 us
25
RELEASE</div>
<div>
0.00
0.00 us
0.00 us
0.00 us
16
RELEASEDIR</div>
<div>
0.00
61.50 us
46.00 us
77.00 us
2
ENTRYLK</div>
<div>
0.00
82.00 us
67.00 us
97.00 us
2
FLUSH</div>
<div>
0.00
265.00 us
265.00 us
265.00 us
1
CREATE</div>
<div>
0.00
57.43 us
30.00 us
85.00 us
7
STATFS</div>
<div>
0.00
61.12 us
37.00 us
107.00 us
16
OPENDIR</div>
<div>
0.00
44.04 us
12.00 us
86.00 us
24
FSTAT</div>
<div>
0.00
41.42 us
24.00 us
96.00 us
26
GETXATTR</div>
<div>
0.00
45.93 us
24.00 us
133.00 us
28
READDIR</div>
<div>
0.00
57.17 us
25.00 us
147.00 us
24
OPEN</div>
<div>
0.00
145.28 us
31.00 us
288.00 us
32
READDIRP</div>
<div>
0.00
39.50 us
10.00 us
152.00 us
132
INODELK</div>
<div>
0.00
330.97 us
20.00 us
14280.00 us
62
LOOKUP</div>
<div>
0.00
79.06 us
19.00 us
851.00 us
430
FXATTROP</div>
<div>
0.02
29.32 us
7.00 us
28154.00 us
22568
FINODELK</div>
<div>
7.80
1313096.68 us
125.00 us
23281862.00 us
189
FSYNC</div>
<div>
92.18
397.92 us
76.00 us
1838343.00 us
7372799
WRITE</div>
<div><br>
</div>
<div>
Duration: 7811
seconds</div>
<div> Data
Read: 0 bytes</div>
<div>Data
Written:
966367641600
bytes</div>
<div><br>
</div>
<div>Interval
0 Stats:</div>
<div> Block
Size:
131072b+
262144b+</div>
<div> No. of
Reads:
0
0</div>
<div>No. of
Writes:
7372798
1</div>
<div> %-latency
Avg-latency
Min-Latency
Max-Latency
No. of calls
Fop</div>
<div> ---------
-----------
-----------
-----------
------------
----</div>
<div>
0.00
0.00 us
0.00 us
0.00 us
25
RELEASE</div>
<div>
0.00
0.00 us
0.00 us
0.00 us
16
RELEASEDIR</div>
<div>
0.00
61.50 us
46.00 us
77.00 us
2
ENTRYLK</div>
<div>
0.00
82.00 us
67.00 us
97.00 us
2
FLUSH</div>
<div>
0.00
265.00 us
265.00 us
265.00 us
1
CREATE</div>
<div>
0.00
57.43 us
30.00 us
85.00 us
7
STATFS</div>
<div>
0.00
61.12 us
37.00 us
107.00 us
16
OPENDIR</div>
<div>
0.00
44.04 us
12.00 us
86.00 us
24
FSTAT</div>
<div>
0.00
41.42 us
24.00 us
96.00 us
26
GETXATTR</div>
<div>
0.00
45.93 us
24.00 us
133.00 us
28
READDIR</div>
<div>
0.00
57.17 us
25.00 us
147.00 us
24
OPEN</div>
<div>
0.00
145.28 us
31.00 us
288.00 us
32
READDIRP</div>
<div>
0.00
39.50 us
10.00 us
152.00 us
132
INODELK</div>
<div>
0.00
330.97 us
20.00 us
14280.00 us
62
LOOKUP</div>
<div>
0.00
79.06 us
19.00 us
851.00 us
430
FXATTROP</div>
<div>
0.02
29.32 us
7.00 us
28154.00 us
22568
FINODELK</div>
<div>
7.80
1313096.68 us
125.00 us
23281862.00 us
189
FSYNC</div>
<div>
92.18
397.92 us
76.00 us
1838343.00 us
7372799
WRITE</div>
<div><br>
</div>
<div>
Duration: 7811
seconds</div>
<div> Data
Read: 0 bytes</div>
<div>Data
Written:
966367641600
bytes</div>
<div><br>
</div>
</div>
<div>does it
make something
more clear?</div>
</div>
<div
class="gmail_extra"><br>
<div
class="gmail_quote">2014-10-13
20:40
GMT+03:00
Roman <span
dir="ltr"><<a
moz-do-not-send="true" href="mailto:romeo.r@gmail.com" target="_blank">romeo.r@gmail.com</a>></span>:<br>
<blockquote
class="gmail_quote"
style="margin:0
0 0
.8ex;border-left:1px
#ccc
solid;padding-left:1ex">
<div dir="ltr">i
think i may
know what was
an issue.
There was an
iscsitarget
service
runing, that
was exporting
this generated
block device.
so maybe my
collegue
Windows server
picked it up
and mountd :)
I'll if it
will happen
again.</div>
<div
class="gmail_extra">
<div>
<div><br>
<div
class="gmail_quote">2014-10-13
20:27
GMT+03:00
Roman <span
dir="ltr"><<a
moz-do-not-send="true" href="mailto:romeo.r@gmail.com" target="_blank">romeo.r@gmail.com</a>></span>:<br>
<blockquote
class="gmail_quote"
style="margin:0
0 0
.8ex;border-left:1px
#ccc
solid;padding-left:1ex">
<div dir="ltr">So
may I restart
the volume and
start the
test, or you
need something
else from this
issue?</div>
<div
class="gmail_extra">
<div>
<div><br>
<div
class="gmail_quote">2014-10-13
19:49
GMT+03:00
Pranith Kumar
Karampuri <span
dir="ltr"><<a
moz-do-not-send="true" href="mailto:pkarampu@redhat.com" target="_blank">pkarampu@redhat.com</a>></span>:<br>
<blockquote
class="gmail_quote"
style="margin:0
0 0
.8ex;border-left:1px
#ccc
solid;padding-left:1ex">
<div
text="#000000"
bgcolor="#FFFFFF"><span> <br>
<div>On
10/13/2014
10:03 PM,
Roman wrote:<br>
</div>
<blockquote
type="cite">
<div dir="ltr">hmm,
<div>seems
like another
strange issue?
Seen this
before. Had to
restart the
volume to get
my empty space
back.</div>
<div>
<div>root@glstor-cli:/srv/nfs/HA-WIN-TT-1T#
ls -l</div>
<div>total
943718400</div>
<div>-rw-r--r--
1 root root
966367641600
Oct 13 16:55
disk</div>
<div>root@glstor-cli:/srv/nfs/HA-WIN-TT-1T#
rm disk</div>
<div>root@glstor-cli:/srv/nfs/HA-WIN-TT-1T#
df -h</div>
<div>Filesystem
Size Used
Avail Use%
Mounted on</div>
<div>rootfs
282G
1.1G 266G
1% /</div>
<div>udev
10M
0 10M 0%
/dev</div>
<div>tmpfs
1.4G 228K
1.4G 1%
/run</div>
<div>/dev/disk/by-uuid/c62ee3c0-c0e5-44af-b0cd-7cb3fbcc0fba
282G 1.1G
266G 1% /</div>
<div>tmpfs
5.0M 0
5.0M 0%
/run/lock</div>
<div>tmpfs
5.2G 0
5.2G 0%
/run/shm</div>
<div>stor1:HA-WIN-TT-1T
1008G
901G 57G
95%
/srv/nfs/HA-WIN-TT-1T</div>
</div>
<div><br>
</div>
<div>no file,
but size is
still 901G.</div>
<div>Both
servers show
the same.</div>
<div>Do I
really have to
restart the
volume to fix
that?</div>
</div>
</blockquote>
</span> IMO
this can
happen if
there is an fd
leak. open-fd
is the only
variable that
can change
with volume
restart. How
do you
re-create the
bug?<span><font
color="#888888"><br>
<br>
Pranith</font></span>
<div>
<div><br>
<blockquote
type="cite">
<div
class="gmail_extra"><br>
<div
class="gmail_quote">2014-10-13
19:30
GMT+03:00
Roman <span
dir="ltr"><<a
moz-do-not-send="true" href="mailto:romeo.r@gmail.com" target="_blank">romeo.r@gmail.com</a>></span>:<br>
<blockquote
class="gmail_quote"
style="margin:0
0 0
.8ex;border-left:1px
#ccc
solid;padding-left:1ex">
<div dir="ltr">Sure.
<div>I'll let
it to run for
this night .</div>
</div>
<div
class="gmail_extra">
<div>
<div><br>
<div
class="gmail_quote">2014-10-13
19:19
GMT+03:00
Pranith Kumar
Karampuri <span
dir="ltr"><<a
moz-do-not-send="true" href="mailto:pkarampu@redhat.com" target="_blank">pkarampu@redhat.com</a>></span>:<br>
<blockquote
class="gmail_quote"
style="margin:0
0 0
.8ex;border-left:1px
#ccc
solid;padding-left:1ex">
<div
text="#000000"
bgcolor="#FFFFFF"> hi Roman,<br>
Do you
think we can
run this test
again? this
time, could
you enable
'gluster
volume profile
<volname>
start', do the
same test.
Provide output
of 'gluster
volume profile
<volname>
info' and logs
after the
test?<span><font
color="#888888"><br>
<br>
Pranith</font></span>
<div>
<div><br>
<div>On
10/13/2014
09:45 PM,
Roman wrote:<br>
</div>
<blockquote
type="cite">
<div dir="ltr">Sure
!
<div><br>
</div>
<div>
<div>root@stor1:~#
gluster volume
info</div>
<div><br>
</div>
<div>Volume
Name:
HA-2TB-TT-Proxmox-cluster</div>
<div>Type:
Replicate</div>
<div>Volume
ID:
66e38bde-c5fa-4ce2-be6e-6b2adeaa16c2</div>
<div>Status:
Started</div>
<div>Number of
Bricks: 1 x 2
= 2</div>
<div>Transport-type:
tcp</div>
<div>Bricks:</div>
<div>Brick1:
stor1:/exports/HA-2TB-TT-Proxmox-cluster/2TB</div>
<div>Brick2:
stor2:/exports/HA-2TB-TT-Proxmox-cluster/2TB</div>
<div>Options
Reconfigured:</div>
<div>nfs.disable:
0</div>
<div>network.ping-timeout:
10</div>
<div><br>
</div>
<div>Volume
Name:
HA-WIN-TT-1T</div>
<div>Type:
Replicate</div>
<div>Volume
ID:
2937ac01-4cba-44a8-8ff8-0161b67f8ee4</div>
<div>Status:
Started</div>
<div>Number of
Bricks: 1 x 2
= 2</div>
<div>Transport-type:
tcp</div>
<div>Bricks:</div>
<div>Brick1:
stor1:/exports/NFS-WIN/1T</div>
<div>Brick2:
stor2:/exports/NFS-WIN/1T</div>
<div>Options
Reconfigured:</div>
<div>nfs.disable:
1</div>
<div>network.ping-timeout:
10</div>
<div><br>
</div>
<div><br>
</div>
</div>
</div>
<div
class="gmail_extra"><br>
<div
class="gmail_quote">2014-10-13
19:09
GMT+03:00
Pranith Kumar
Karampuri <span
dir="ltr"><<a
moz-do-not-send="true" href="mailto:pkarampu@redhat.com" target="_blank">pkarampu@redhat.com</a>></span>:<br>
<blockquote
class="gmail_quote"
style="margin:0
0 0
.8ex;border-left:1px
#ccc
solid;padding-left:1ex">
<div
text="#000000"
bgcolor="#FFFFFF"> Could you give your 'gluster volume info' output?<br>
<br>
Pranith
<div>
<div><br>
<div>On
10/13/2014
09:36 PM,
Roman wrote:<br>
</div>
</div>
</div>
<blockquote
type="cite">
<div>
<div>
<div dir="ltr">Hi,
<div><br>
</div>
<div>I've got
this kind of
setup (servers
run replica)</div>
<div><br>
</div>
<div><br>
</div>
<div>@ 10G
backend</div>
<div>gluster
storage1</div>
<div>gluster
storage2</div>
<div>gluster
client1</div>
<div><br>
</div>
<div>@1g
backend</div>
<div>other
gluster
clients</div>
<div><br>
</div>
<div>Servers
got HW RAID5
with SAS
disks.</div>
<div><br>
</div>
<div>So today
I've desided
to create a
900GB file for
iscsi target
that will be
located @
glusterfs
separate
volume, using
dd (just a
dummy file
filled with
zeros, bs=1G
count 900)</div>
<div>For the
first of all
the process
took pretty
lots of time,
the writing
speed was 130
MB/sec (client
port was 2
gbps, servers
ports were
running @
1gbps).</div>
<div>Then it
reported
something like
"endpoint is
not connected"
and all of my
VMs on the
other volume
started to
give me IO
errors.</div>
<div>Servers
load was
around 4,6
(total 12
cores)</div>
<div><br>
</div>
<div>Maybe it
was due to
timeout of 2
secs, so I've
made it a big
higher, 10
sec.</div>
<div><br>
</div>
<div>Also
during the dd
image creation
time, VMs very
often reported
me that their
disks are slow
like</div>
<div>
<p>WARNINGs:
Read IO Wait
time is -0.02
(outside range
[0:1]).</p>
<p>Is 130MB
/sec is the
maximum
bandwidth for
all of the
volumes in
total? That
why would we
need 10g
backends?</p>
<p>HW Raid
local speed is
300 MB/sec, so
it should not
be an issue.
any ideas or
mby any
advices?</p>
<p><br>
</p>
<p>Maybe some1
got optimized
sysctl.conf
for 10G
backend?</p>
<p>mine is
pretty simple,
which can be
found from
googling.</p>
<p><br>
</p>
<p>just to
mention: those
VM-s were
connected
using separate
1gbps
intraface,
which means,
they should
not be
affected by
the client
with 10g
backend.</p>
<p><br>
</p>
<p>logs are
pretty
useless, they
just say this
during the
outage</p>
<p><br>
</p>
<p>[2014-10-13
12:09:18.392910]
W
[client-handshake.c:276:client_ping_cbk]
0-HA-2TB-TT-Proxmox-cluster-client-0:
timer must
have expired</p>
<p>[2014-10-13
12:10:08.389708]
C
[client-handshake.c:127:rpc_client_ping_timer_expired]
0-HA-2TB-TT-Proxmox-cluster-client-0:
server <a
moz-do-not-send="true"
href="http://10.250.0.1:49159" target="_blank">10.250.0.1:49159</a> has
not responded
in the last 2
seconds,
disconnecting.</p>
<p>[2014-10-13
12:10:08.390312]
W
[client-handshake.c:276:client_ping_cbk]
0-HA-2TB-TT-Proxmox-cluster-client-0:
timer must
have expired</p>
</div>
<div>so I
decided to set
the timout a
bit higher.</div>
<div>
<div><br>
</div>
<div>So it
seems to me,
that under
high load
GlusterFS is
not useable?
130 MB/s is
not that much
to get some
kind of
timeouts or
makeing the
systme so
slow, that
VM-s feeling
themselves
bad.</div>
<div><br>
</div>
<div>Of
course, after
the
disconnection,
healing
process was
started, but
as VM-s lost
connection to
both of
servers, it
was pretty
useless, they
could not run
anymore. and
BTW, when u
load the
server with
such huge job
(dd of 900GB),
healing
process goes
soooooo slow
:)</div>
<div><br>
</div>
<div><br>
</div>
<div><br>
</div>
-- <br>
Best regards,<br>
Roman. </div>
</div>
<br>
<fieldset></fieldset>
<br>
</div>
</div>
<pre>_______________________________________________
Gluster-users mailing list
<a moz-do-not-send="true" href="mailto:Gluster-users@gluster.org" target="_blank">Gluster-users@gluster.org</a>
<a moz-do-not-send="true" href="http://supercolony.gluster.org/mailman/listinfo/gluster-users" target="_blank">http://supercolony.gluster.org/mailman/listinfo/gluster-users</a></pre>
</blockquote>
<br>
</div>
</blockquote>
</div>
<br>
<br
clear="all">
<div><br>
</div>
-- <br>
Best regards,<br>
Roman. </div>
</blockquote>
<br>
</div>
</div>
</div>
</blockquote>
</div>
<br>
<br
clear="all">
<div><br>
</div>
</div>
</div>
<span><font
color="#888888">--
<br>
Best regards,<br>
Roman. </font></span></div>
</blockquote>
</div>
<br>
<br
clear="all">
<div><br>
</div>
-- <br>
Best regards,<br>
Roman. </div>
</blockquote>
<br>
</div>
</div>
</div>
</blockquote>
</div>
<br>
<br
clear="all">
<div><br>
</div>
</div>
</div>
<span><font
color="#888888">--
<br>
Best regards,<br>
Roman. </font></span></div>
</blockquote>
</div>
<br>
<br
clear="all">
<div><br>
</div>
</div>
</div>
<span><font
color="#888888">--
<br>
Best regards,<br>
Roman. </font></span></div>
</blockquote>
</div>
<br>
<br clear="all">
<span><font
color="#888888">
<div><br>
</div>
-- <br>
Best regards,<br>
Roman. </font></span></div>
<span><font
color="#888888">
</font></span></blockquote>
<span><font
color="#888888">
<br>
</font></span></div>
<span><font
color="#888888"> </font></span></div>
<span><font
color="#888888"> </font></span></div>
<span><font
color="#888888"> </font></span></blockquote>
<span><font color="#888888">
</font></span></div>
<span><font color="#888888"> <br>
<br clear="all">
<span><font
color="#888888">
<div><br>
</div>
-- <br>
Best regards,<br>
Roman. </font></span></font></span></div>
<span><font color="#888888"> </font></span></blockquote>
<span><font color="#888888"> <br>
</font></span></div>
<span><font color="#888888">
</font></span></blockquote>
</div>
<span><font color="#888888"><br>
<br clear="all">
<span><font color="#888888">
<div><br>
</div>
-- <br>
Best regards,<br>
Roman.
</font></span></font></span></div>
<span><font color="#888888">
</font></span></blockquote>
</div>
<span><font color="#888888"><br>
<br clear="all">
<span><font color="#888888">
<div><br>
</div>
-- <br>
Best regards,<br>
Roman.
</font></span></font></span></div>
<span><font color="#888888">
</font></span></blockquote>
</div>
<span><font color="#888888"><br>
<br clear="all">
<span class="HOEnZb"><font color="#888888">
<div><br>
</div>
-- <br>
Best regards,<br>
Roman.
</font></span></font></span></div>
<span class="HOEnZb"><font color="#888888">
</font></span></blockquote>
</div>
<span class="HOEnZb"><font color="#888888"><br>
<br clear="all">
<div><br>
</div>
-- <br>
Best regards,<br>
Roman.
</font></span></div>
</blockquote>
</div>
<br>
<br clear="all">
<div><br>
</div>
-- <br>
<div class="gmail_signature">Best regards,<br>
Roman.</div>
</div>
</blockquote>
<br>
</body>
</html>