This should work without any issues. It is possible that the shard(s)
would get created with different gfids but the ones on the lagging brick
will eventually (by the time heal-info returns all zeroes) get replaced
with shards having the correct gfids.
would get created with different gfids but the ones on the lagging brick
will eventually (by the time heal-info returns all zeroes) get replaced
with shards having the correct gfids.
On Thu, Oct 27, 2016 at 3:48 PM, qingwei wei <tchengwee@xxxxxxxxx> wrote:
Hi,
My final goal of the test is to see the impact of brick replacement
while IO is till running.
One scenario that i think of is as below:
1. random read IO is performed on gluster volume (3 replicas)
2. 1 brick down and IO still ongoing
3. Perform brick replacement and IO still ongoing
4. There will be a full heal on the new brick while IO still ongoing.
Assume we have large number of files, reconstruct (recreate shard
file) could take some time on this new brick. So will it be possible
that some random read IO to yet created shard trigger the similar
error?
Thanks.
Cwtan
On Thu, Oct 27, 2016 at 4:26 PM, Krutika Dhananjay <kdhananj@xxxxxxxxxx> wrote:
> Found the RC. The problem seems to be that sharding translator attempts to
> create
> non-existent shards in read/write codepaths with a newly generated gfid
> attached
> to the create request in case the shard is absent. Replicate translator,
> which sits below
> sharding on the stack takes this request and plays it on all of its
> replicas. On two of them it
> fails with EEXIST, and on the one where the shards were removed from the
> backend, the
> shard path is created but with the newly generated gfid while the other two
> replicas continue to
> hold the original gfid (the one prior to rm -rf). Although this can be
> fixed, it will require one
> additional lookup for each shard for each read/write operation, causing the
> latency of the read/write
> response to the application to increase by a factor of 1 network call.
>
> The test you're doing is partially (but not fully) manipulating and removing
> data from the backend,
> which is not recommended.
>
> My question to you is this - what is the specific failure that you are
> trying to simulate with removal of
> contents of .shard? Normally, the `rm -rf on backend` type of tests are
> performed to simulate disk
> failure and its replacement with a brand new disk, in which case executing
> the replace-brick/reset-brick
> commands should be sufficient to recover all contents from the remaining two
> replicas.
>
> -Krutika
>
> On Thu, Oct 27, 2016 at 12:49 PM, Krutika Dhananjay <kdhananj@xxxxxxxxxx>
> wrote:
>>
>> Now it's reproducible, thanks. :)
>>
>> I think I know the RC. Let me confirm it through tests and report back.
>>
>> -Krutika
>>
>> On Thu, Oct 27, 2016 at 10:42 AM, qingwei wei <tchengwee@xxxxxxxxx> wrote:
>>>
>>> Hi,
>>>
>>> I did few more test runs and it seems that it happens during this
>>> sequence
>>>
>>> 1.populate data using dd
>>> 2. delete away ALL the shard files in one of the brick .shard folder
>>> 3. Trying to access using dd, no error reported
>>> 4. umount and mount.
>>> 5. Trying to access using dd, no error reported
>>> 6. umount and mount.
>>> 7. Trying to access using dd and Input/Output error reported
>>>
>>> during step 3 and 4, no file is created under the .shard directory
>>> For step 7, a shard file is created with same file name but different
>>> gfid compare to other good replicas.
>>>
>>> Below is the client log and brick log with more details in the attached
>>> log.
>>>
>>> Client log
>>>
>>> [2016-10-27 04:34:46.493281] D [MSGID: 0]
>>> [shard.c:3138:shard_common_mknod_cbk] 0-testHeal4-shard: mknod of
>>> shard 1 failed: File exists
>>> [2016-10-27 04:34:46.493351] D [MSGID: 0]
>>> [dht-common.c:2633:dht_lookup] 0-testHeal4-dht: Calling fresh lookup
>>> for /.shard/76bc4b0f-bb18-4736-8327-99098cd0d7ce.1 on
>>> testHeal4-replicate-0
>>> [2016-10-27 04:34:46.494646] W [MSGID: 114031]
>>> [client-rpc-fops.c:2981:client3_3_lookup_cbk] 0-testHeal4-client-0:
>>> remote operation failed. Path: (null)
>>> (00000000-0000-0000-0000-000000000000) [Invalid argument]
>>> [2016-10-27 04:34:46.494673] D [MSGID: 0]
>>> [client-rpc-fops.c:2989:client3_3_lookup_cbk] 0-stack-trace:
>>> stack-address: 0x7f9083edc1c8, testHeal4-client-0 returned -1 error:
>>> Invalid argument [Invalid argument]
>>> [2016-10-27 04:34:46.494705] W [MSGID: 114031]
>>> [client-rpc-fops.c:2981:client3_3_lookup_cbk] 0-testHeal4-client-1:
>>> remote operation failed. Path: (null)
>>> (00000000-0000-0000-0000-000000000000) [Invalid argument]
>>> [2016-10-27 04:34:46.494710] W [MSGID: 114031]
>>> [client-rpc-fops.c:2981:client3_3_lookup_cbk] 0-testHeal4-client-2:
>>> remote operation failed. Path: (null)
>>> (00000000-0000-0000-0000-000000000000) [Invalid argument]
>>> [2016-10-27 04:34:46.494730] D [MSGID: 0]
>>> [client-rpc-fops.c:2989:client3_3_lookup_cbk] 0-stack-trace:
>>> stack-address: 0x7f9083edc1c8, testHeal4-client-1 returned -1 error:
>>> Invalid argument [Invalid argument]
>>> [2016-10-27 04:34:46.494751] D [MSGID: 0]
>>> [client-rpc-fops.c:2989:client3_3_lookup_cbk] 0-stack-trace:
>>> stack-address: 0x7f9083edc1c8, testHeal4-client-2 returned -1 error:
>>> Invalid argument [Invalid argument]
>>> [2016-10-27 04:34:46.495339] D [MSGID: 0]
>>> [afr-common.c:1986:afr_lookup_done] 0-stack-trace: stack-address:
>>> 0x7f9083edbb1c, testHeal4-replicate-0 returned -1 error: Input/output
>>> error [Input/output error]
>>> [2016-10-27 04:34:46.495364] D [MSGID: 0]
>>> [dht-common.c:2220:dht_lookup_cbk] 0-testHeal4-dht: fresh_lookup
>>> returned for /.shard/76bc4b0f-bb18-4736-8327-99098cd0d7ce.1 with
>>> op_ret -1 [Input/output error]
>>> [2016-10-27 04:34:46.495374] D [MSGID: 0]
>>> [dht-common.c:2300:dht_lookup_cbk] 0-testHeal4-dht: Lookup of
>>> /.shard/76bc4b0f-bb18-4736-8327-99098cd0d7ce.1 for subvolume
>>> testHeal4-replicate-0 failed [Input/output error]
>>> [2016-10-27 04:34:46.495384] D [MSGID: 0]
>>> [dht-common.c:2363:dht_lookup_cbk] 0-stack-trace: stack-address:
>>> 0x7f9083edbb1c, testHeal4-dht returned -1 error: Input/output error
>>> [Input/output error]
>>> [2016-10-27 04:34:46.495395] E [MSGID: 133010]
>>> [shard.c:1582:shard_common_lookup_shards_cbk] 0-testHeal4-shard:
>>> Lookup on shard 1 failed. Base file gfid =
>>> 76bc4b0f-bb18-4736-8327-99098cd0d7ce [Input/output error]
>>> [2016-10-27 04:34:46.495406] D [MSGID: 0]
>>> [shard.c:3086:shard_post_lookup_shards_readv_handler] 0-stack-trace:
>>> stack-address: 0x7f9083edbb1c, testHeal4-shard returned -1 error:
>>> Input/output error [Input/output error]
>>> [2016-10-27 04:34:46.495417] D [MSGID: 0]
>>> [defaults.c:1010:default_readv_cbk] 0-stack-trace: stack-address:
>>> 0x7f9083edbb1c, testHeal4-write-behind returned -1 error: Input/output
>>> error [Input/output error]
>>> [2016-10-27 04:34:46.495428] D [MSGID: 0]
>>> [read-ahead.c:462:ra_readv_disabled_cbk] 0-stack-trace: stack-address:
>>> 0x7f9083edbb1c, testHeal4-read-ahead returned -1 error: Input/output
>>> error [Input/output error]
>>>
>>> brick log
>>>
>>> [2016-10-27 04:34:46.492055] D [MSGID: 0]
>>> [io-threads.c:351:iot_schedule] 0-testHeal4-io-threads: STATFS
>>> scheduled as fast fop
>>> [2016-10-27 04:34:46.492157] D [client_t.c:333:gf_client_ref]
>>>
>>> (-->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server3_3_entrylk+0x93)
>>> [0x7efebb37d633]
>>>
>>> -->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( get_frame_from_request+0x257)
>>> [0x7efebb36cfd7] -->/lib64/libglusterfs.so.0(gf_client_ref+0x68)
>>> [0x7efecfadf608] ) 0-client_t:
>>>
>>> fujitsu05.dctopenstack.org-6064-2016/10/27-04:34:44: 217958-testHeal4-client-1-0-0:
>>> ref-count 3
>>> [2016-10-27 04:34:46.492180] D [MSGID: 0]
>>> [io-threads.c:351:iot_schedule] 0-testHeal4-io-threads: ENTRYLK
>>> scheduled as normal fop
>>> [2016-10-27 04:34:46.492239] D [client_t.c:417:gf_client_unref]
>>>
>>> (-->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server_statfs_cbk+0x112)
>>> [0x7efebb36f8e2]
>>>
>>> -->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server_submit_reply+0x123)
>>> [0x7efebb368f13] -->/lib64/libglusterfs.so.0(gf_client_unref+0x77)
>>> [0x7efecfadf787] ) 0-client_t:
>>>
>>> fujitsu05.dctopenstack.org-6064-2016/10/27-04:34:44: 217958-testHeal4-client-1-0-0:
>>> ref-count 2
>>> [2016-10-27 04:34:46.492271] D [client_t.c:417:gf_client_unref]
>>>
>>> (-->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server_entrylk_cbk+0xa6)
>>> [0x7efebb3713a6]
>>>
>>> -->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server_submit_reply+0x123)
>>> [0x7efebb368f13] -->/lib64/libglusterfs.so.0(gf_client_unref+0x77)
>>> [0x7efecfadf787] ) 0-client_t:
>>>
>>> fujitsu05.dctopenstack.org-6064-2016/10/27-04:34:44: 217958-testHeal4-client-1-0-0:
>>> ref-count 1
>>> [2016-10-27 04:34:46.492535] D [client_t.c:333:gf_client_ref]
>>>
>>> (-->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server3_3_mknod+0x80)
>>> [0x7efebb37b690]
>>>
>>> -->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( get_frame_from_request+0x257)
>>> [0x7efebb36cfd7] -->/lib64/libglusterfs.so.0(gf_client_ref+0x68)
>>> [0x7efecfadf608] ) 0-client_t:
>>>
>>> fujitsu05.dctopenstack.org-6064-2016/10/27-04:34:44: 217958-testHeal4-client-1-0-0:
>>> ref-count 2
>>> [2016-10-27 04:34:46.492565] D [MSGID: 0]
>>> [io-threads.c:351:iot_schedule] 0-testHeal4-io-threads: MKNOD
>>> scheduled as normal fop
>>> [2016-10-27 04:34:46.492843] D [client_t.c:417:gf_client_unref]
>>>
>>> (-->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server_mknod_cbk+0x5ad)
>>> [0x7efebb383c9d]
>>>
>>> -->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server_submit_reply+0x123)
>>> [0x7efebb368f13] -->/lib64/libglusterfs.so.0(gf_client_unref+0x77)
>>> [0x7efecfadf787] ) 0-client_t:
>>>
>>> fujitsu05.dctopenstack.org-6064-2016/10/27-04:34:44: 217958-testHeal4-client-1-0-0:
>>> ref-count 1
>>> [2016-10-27 04:34:46.492981] D [client_t.c:333:gf_client_ref]
>>>
>>> (-->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server3_3_xattrop+0x86)
>>> [0x7efebb3789d6]
>>>
>>> -->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( get_frame_from_request+0x257)
>>> [0x7efebb36cfd7] -->/lib64/libglusterfs.so.0(gf_client_ref+0x68)
>>> [0x7efecfadf608] ) 0-client_t:
>>>
>>> fujitsu05.dctopenstack.org-6064-2016/10/27-04:34:44: 217958-testHeal4-client-1-0-0:
>>> ref-count 2
>>> [2016-10-27 04:34:46.493056] D [MSGID: 0]
>>> [io-threads.c:351:iot_schedule] 0-testHeal4-io-threads: XATTROP
>>> scheduled as slow fop
>>> [2016-10-27 04:34:46.493128] D [client_t.c:333:gf_client_ref]
>>>
>>> (-->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server3_3_entrylk+0x93)
>>> [0x7efebb37d633]
>>>
>>> -->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( get_frame_from_request+0x257)
>>> [0x7efebb36cfd7] -->/lib64/libglusterfs.so.0(gf_client_ref+0x68)
>>> [0x7efecfadf608] ) 0-client_t:
>>>
>>> fujitsu05.dctopenstack.org-6064-2016/10/27-04:34:44: 217958-testHeal4-client-1-0-0:
>>> ref-count 3
>>> [2016-10-27 04:34:46.493148] D [MSGID: 0]
>>> [io-threads.c:351:iot_schedule] 0-testHeal4-io-threads: ENTRYLK
>>> scheduled as normal fop
>>> [2016-10-27 04:34:46.493214] D [client_t.c:417:gf_client_unref]
>>>
>>> (-->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server_xattrop_cbk+0xd9)
>>> [0x7efebb370579]
>>>
>>> -->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server_submit_reply+0x123)
>>> [0x7efebb368f13] -->/lib64/libglusterfs.so.0(gf_client_unref+0x77)
>>> [0x7efecfadf787] ) 0-client_t:
>>>
>>> fujitsu05.dctopenstack.org-6064-2016/10/27-04:34:44: 217958-testHeal4-client-1-0-0:
>>> ref-count 2
>>> [2016-10-27 04:34:46.493239] D [client_t.c:417:gf_client_unref]
>>>
>>> (-->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server_entrylk_cbk+0xa6)
>>> [0x7efebb3713a6]
>>>
>>> -->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server_submit_reply+0x123)
>>> [0x7efebb368f13] -->/lib64/libglusterfs.so.0(gf_client_unref+0x77)
>>> [0x7efecfadf787] ) 0-client_t:
>>>
>>> fujitsu05.dctopenstack.org-6064-2016/10/27-04:34:44: 217958-testHeal4-client-1-0-0:
>>> ref-count 1
>>> [2016-10-27 04:34:46.493490] D [client_t.c:333:gf_client_ref]
>>>
>>> (-->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server3_3_lookup+0x8b)
>>> [0x7efebb386beb]
>>>
>>> -->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( get_frame_from_request+0x257)
>>> [0x7efebb36cfd7] -->/lib64/libglusterfs.so.0(gf_client_ref+0x68)
>>> [0x7efecfadf608] ) 0-client_t:
>>>
>>> fujitsu05.dctopenstack.org-6064-2016/10/27-04:34:44: 217958-testHeal4-client-1-0-0:
>>> ref-count 2
>>> [2016-10-27 04:34:46.493514] D [MSGID: 0]
>>> [io-threads.c:351:iot_schedule] 0-testHeal4-io-threads: LOOKUP
>>> scheduled as fast fop
>>> [2016-10-27 04:34:46.493666] D [client_t.c:417:gf_client_unref]
>>>
>>> (-->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server_lookup_cbk+0x548)
>>> [0x7efebb3864c8]
>>>
>>> -->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server_submit_reply+0x123)
>>> [0x7efebb368f13] -->/lib64/libglusterfs.so.0(gf_client_unref+0x77)
>>> [0x7efecfadf787] ) 0-client_t:
>>>
>>> fujitsu05.dctopenstack.org-6064-2016/10/27-04:34:44: 217958-testHeal4-client-1-0-0:
>>> ref-count 1
>>> [2016-10-27 04:34:46.493782] D [client_t.c:333:gf_client_ref]
>>>
>>> (-->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server3_3_lookup+0x8b)
>>> [0x7efebb386beb]
>>>
>>> -->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( get_frame_from_request+0x257)
>>> [0x7efebb36cfd7] -->/lib64/libglusterfs.so.0(gf_client_ref+0x68)
>>> [0x7efecfadf608] ) 0-client_t:
>>>
>>> fujitsu05.dctopenstack.org-6064-2016/10/27-04:34:44: 217958-testHeal4-client-1-0-0:
>>> ref-count 2
>>> [2016-10-27 04:34:46.493986] D [client_t.c:417:gf_client_unref]
>>>
>>> (-->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server_lookup_cbk+0x548)
>>> [0x7efebb3864c8]
>>>
>>> -->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server_submit_reply+0x123)
>>> [0x7efebb368f13] -->/lib64/libglusterfs.so.0(gf_client_unref+0x77)
>>> [0x7efecfadf787] ) 0-client_t:
>>>
>>> fujitsu05.dctopenstack.org-6064-2016/10/27-04:34:44: 217958-testHeal4-client-1-0-0:
>>> ref-count 1
>>> [2016-10-27 04:34:46.494596] D [client_t.c:333:gf_client_ref]
>>>
>>> (-->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server3_3_lookup+0x8b)
>>> [0x7efebb386beb]
>>>
>>> -->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( get_frame_from_request+0x257)
>>> [0x7efebb36cfd7] -->/lib64/libglusterfs.so.0(gf_client_ref+0x68)
>>> [0x7efecfadf608] ) 0-client_t:
>>>
>>> fujitsu05.dctopenstack.org-6064-2016/10/27-04:34:44: 217958-testHeal4-client-1-0-0:
>>> ref-count 2
>>> [2016-10-27 04:34:46.494616] D [logging.c:1954:_gf_msg_internal]
>>> 0-logging-infra: Buffer overflow of a buffer whose size limit is 5.
>>> About to flush least recently used log message to disk
>>> [2016-10-27 04:34:46.493818] D [MSGID: 0]
>>> [io-threads.c:351:iot_schedule] 0-testHeal4-io-threads: LOOKUP
>>> scheduled as fast fop
>>> [2016-10-27 04:34:46.494616] W [MSGID: 115009]
>>> [server-resolve.c:574:server_resolve] 0-testHeal4-server: no
>>> resolution type for (null) (LOOKUP)
>>> [2016-10-27 04:34:46.494650] E [MSGID: 115050]
>>> [server-rpc-fops.c:179:server_lookup_cbk] 0-testHeal4-server: 29:
>>> LOOKUP (null)
>>> (00000000-0000-0000-0000-000000000000/76bc4b0f-bb18- 4736-8327-99098cd0d7ce.1)
>>> ==> (Invalid argument) [Invalid argument]
>>> [2016-10-27 04:34:46.494720] D [client_t.c:417:gf_client_unref]
>>>
>>> (-->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server_lookup_cbk+0x548)
>>> [0x7efebb3864c8]
>>>
>>> -->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server_submit_reply+0x123)
>>> [0x7efebb368f13] -->/lib64/libglusterfs.so.0(gf_client_unref+0x77)
>>> [0x7efecfadf787] ) 0-client_t:
>>>
>>> fujitsu05.dctopenstack.org-6064-2016/10/27-04:34:44: 217958-testHeal4-client-1-0-0:
>>> ref-count 1
>>> [2016-10-27 04:34:46.494936] D [client_t.c:333:gf_client_ref]
>>>
>>> (-->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server3_3_lookup+0x8b)
>>> [0x7efebb386beb]
>>>
>>> -->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( get_frame_from_request+0x257)
>>> [0x7efebb36cfd7] -->/lib64/libglusterfs.so.0(gf_client_ref+0x68)
>>> [0x7efecfadf608] ) 0-client_t:
>>>
>>> fujitsu05.dctopenstack.org-6064-2016/10/27-04:34:44: 217958-testHeal4-client-1-0-0:
>>> ref-count 2
>>> [2016-10-27 04:34:46.494967] D [MSGID: 0]
>>> [io-threads.c:351:iot_schedule] 0-testHeal4-io-threads: LOOKUP
>>> scheduled as fast fop
>>> [2016-10-27 04:34:46.495108] D [client_t.c:417:gf_client_unref]
>>>
>>> (-->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server_lookup_cbk+0x548)
>>> [0x7efebb3864c8]
>>>
>>> -->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server_submit_reply+0x123)
>>> [0x7efebb368f13] -->/lib64/libglusterfs.so.0(gf_client_unref+0x77)
>>> [0x7efecfadf787] ) 0-client_t:
>>>
>>> fujitsu05.dctopenstack.org-6064-2016/10/27-04:34:44: 217958-testHeal4-client-1-0-0:
>>> ref-count 1
>>> [2016-10-27 04:34:46.595813] D [client_t.c:333:gf_client_ref]
>>>
>>> (-->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server3_3_lookup+0x8b)
>>> [0x7efebb386beb]
>>>
>>> -->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( get_frame_from_request+0x257)
>>> [0x7efebb36cfd7] -->/lib64/libglusterfs.so.0(gf_client_ref+0x68)
>>> [0x7efecfadf608] ) 0-client_t:
>>>
>>> fujitsu05.dctopenstack.org-6064-2016/10/27-04:34:44: 217958-testHeal4-client-1-0-0:
>>> ref-count 2
>>> [2016-10-27 04:34:46.595915] D [client_t.c:333:gf_client_ref]
>>>
>>> (-->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server3_3_lookup+0x8b)
>>> [0x7efebb386beb]
>>>
>>> -->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( get_frame_from_request+0x257)
>>> [0x7efebb36cfd7] -->/lib64/libglusterfs.so.0(gf_client_ref+0x68)
>>> [0x7efecfadf608] ) 0-client_t:
>>>
>>> fujitsu05.dctopenstack.org-6064-2016/10/27-04:34:44: 217958-testHeal4-client-1-0-0:
>>> ref-count 3
>>> [2016-10-27 04:34:46.596054] D [client_t.c:417:gf_client_unref]
>>>
>>> (-->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server_lookup_cbk+0x548)
>>> [0x7efebb3864c8]
>>>
>>> -->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server_submit_reply+0x123)
>>> [0x7efebb368f13] -->/lib64/libglusterfs.so.0(gf_client_unref+0x77)
>>> [0x7efecfadf787] ) 0-client_t:
>>>
>>> fujitsu05.dctopenstack.org-6064-2016/10/27-04:34:44: 217958-testHeal4-client-1-0-0:
>>> ref-count 2
>>> [2016-10-27 04:34:46.596162] D [client_t.c:417:gf_client_unref]
>>>
>>> (-->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server_lookup_cbk+0x548)
>>> [0x7efebb3864c8]
>>>
>>> -->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server_submit_reply+0x123)
>>> [0x7efebb368f13] -->/lib64/libglusterfs.so.0(gf_client_unref+0x77)
>>> [0x7efecfadf787] ) 0-client_t:
>>>
>>> fujitsu05.dctopenstack.org-6064-2016/10/27-04:34:44: 217958-testHeal4-client-1-0-0:
>>> ref-count 1
>>> [2016-10-27 04:34:46.596427] D [client_t.c:333:gf_client_ref]
>>>
>>> (-->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server3_3_entrylk+0x93)
>>> [0x7efebb37d633]
>>>
>>> -->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( get_frame_from_request+0x257)
>>> [0x7efebb36cfd7] -->/lib64/libglusterfs.so.0(gf_client_ref+0x68)
>>> [0x7efecfadf608] ) 0-client_t:
>>>
>>> fujitsu05.dctopenstack.org-6064-2016/10/27-04:34:44: 217958-testHeal4-client-1-0-0:
>>> ref-count 2
>>> [2016-10-27 04:34:46.596453] D [logging.c:1954:_gf_msg_internal]
>>> 0-logging-infra: Buffer overflow of a buffer whose size limit is 5.
>>> About to flush least recently used log message to disk
>>> The message "D [MSGID: 0] [io-threads.c:351:iot_schedule]
>>> 0-testHeal4-io-threads: LOOKUP scheduled as fast fop" repeated 2 times
>>> between [2016-10-27 04:34:46.494967] and [2016-10-27 04:34:46.595944]
>>> [2016-10-27 04:34:46.596453] D [MSGID: 0]
>>> [io-threads.c:351:iot_schedule] 0-testHeal4-io-threads: ENTRYLK
>>> scheduled as normal fop
>>> [2016-10-27 04:34:46.596551] D [client_t.c:333:gf_client_ref]
>>>
>>> (-->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server3_3_entrylk+0x93)
>>> [0x7efebb37d633]
>>>
>>> -->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( get_frame_from_request+0x257)
>>> [0x7efebb36cfd7] -->/lib64/libglusterfs.so.0(gf_client_ref+0x68)
>>> [0x7efecfadf608] ) 0-client_t:
>>>
>>> fujitsu05.dctopenstack.org-6064-2016/10/27-04:34:44: 217958-testHeal4-client-1-0-0:
>>> ref-count 3
>>> [2016-10-27 04:34:46.596603] D [logging.c:1954:_gf_msg_internal]
>>> 0-logging-infra: Buffer overflow of a buffer whose size limit is 5.
>>> About to flush least recently used log message to disk
>>> [2016-10-27 04:34:46.596611] D [client_t.c:417:gf_client_unref]
>>>
>>> (-->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server_entrylk_cbk+0xa6)
>>> [0x7efebb3713a6]
>>>
>>> -->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server_submit_reply+0x123)
>>> [0x7efebb368f13] -->/lib64/libglusterfs.so.0(gf_client_unref+0x77)
>>> [0x7efecfadf787] ) 0-client_t:
>>>
>>> fujitsu05.dctopenstack.org-6064-2016/10/27-04:34:44: 217958-testHeal4-client-1-0-0:
>>> ref-count 2
>>> [2016-10-27 04:34:46.596570] D [MSGID: 0]
>>> [io-threads.c:351:iot_schedule] 0-testHeal4-io-threads: ENTRYLK
>>> scheduled as normal fop
>>> [2016-10-27 04:34:46.596602] D [MSGID: 0]
>>> [entrylk.c:701:pl_common_entrylk] 0-stack-trace: stack-address:
>>> 0x7efecd3db738, testHeal4-locks returned -1 error: Resource
>>> temporarily unavailable [Resource temporarily unavailable]
>>> [2016-10-27 04:34:46.596650] D [MSGID: 0]
>>> [defaults.c:1196:default_entrylk_cbk] 0-stack-trace: stack-address:
>>> 0x7efecd3db738, testHeal4-io-threads returned -1 error: Resource
>>> temporarily unavailable [Resource temporarily unavailable]
>>> [2016-10-27 04:34:46.596664] D [MSGID: 0]
>>> [io-stats.c:1811:io_stats_entrylk_cbk] 0-stack-trace: stack-address:
>>> 0x7efecd3db738, /mnt/sdc_mssd/testHeal4 returned -1 error: Resource
>>> temporarily unavailable [Resource temporarily unavailable]
>>> [2016-10-27 04:34:46.596676] D [MSGID: 115054]
>>> [server-rpc-fops.c:350:server_entrylk_cbk] 0-testHeal4-server: 34:
>>> ENTRYLK /.shard (be318638-e8a0-4c6d-977d-7a937aa84806) ==> (Resource
>>> temporarily unavailable) [Resource temporarily unavailable]
>>> [2016-10-27 04:34:46.596764] D [client_t.c:417:gf_client_unref]
>>>
>>> (-->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server_entrylk_cbk+0xa6)
>>> [0x7efebb3713a6]
>>>
>>> -->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server_submit_reply+0x123)
>>> [0x7efebb368f13] -->/lib64/libglusterfs.so.0(gf_client_unref+0x77)
>>> [0x7efecfadf787] ) 0-client_t:
>>>
>>> fujitsu05.dctopenstack.org-6064-2016/10/27-04:34:44: 217958-testHeal4-client-1-0-0:
>>> ref-count 1
>>> [2016-10-27 04:34:46.596791] D [client_t.c:333:gf_client_ref]
>>>
>>> (-->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server3_3_mknod+0x80)
>>> [0x7efebb37b690]
>>>
>>> -->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( get_frame_from_request+0x257)
>>> [0x7efebb36cfd7] -->/lib64/libglusterfs.so.0(gf_client_ref+0x68)
>>> [0x7efecfadf608] ) 0-client_t:
>>>
>>> fujitsu05.dctopenstack.org-6064-2016/10/27-04:34:44: 217958-testHeal4-client-1-0-0:
>>> ref-count 2
>>> [2016-10-27 04:34:46.596813] D [MSGID: 0]
>>> [server-resolve.c:330:resolve_entry_simple] 0-testHeal4-server: inode
>>> (pointer: 0x7efea1aebaa0 gfid:041e3b34-14c2-4bb1-82e2-db352232c3cf
>>> found for path ((null)) while type is RESOLVE_NOT
>>> [2016-10-27 04:34:46.596828] D [MSGID: 115057]
>>> [server-rpc-fops.c:563:server_mknod_cbk] 0-testHeal4-server: 35: MKNOD
>>> (null)
>>> (be318638-e8a0-4c6d-977d-7a937aa84806/76bc4b0f-bb18- 4736-8327-99098cd0d7ce.1)
>>> ==> (File exists) [File exists]
>>> [2016-10-27 04:34:46.596896] D [client_t.c:417:gf_client_unref]
>>>
>>> (-->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so( server_mknod_cbk+0x5ad)
>>> [0x7efebb383c9d]
>>>
>>> -->/usr/lib64/glusterfs/3.7.16/xlator/protocol/server.so(