Re: Lost CLOSE with NFSv4.1 on RHEL7 ( and bejond?)

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Mon, Aug 1, 2016 at 7:08 AM, Mkrtchyan, Tigran
<tigran.mkrtchyan@xxxxxxx> wrote:
> Hi Olga,
>
> we have installed kernel 4.7.0 on one of the nodes and don't see missing
> closes from that node.
>
> Nevertheless, I don't think that the commit you have mentioned is fixing that,
> as it fixes OPEN_DOWNGRADE, but we have a sequence of OPEN->CLOSE->OPEN. The
> OPEN_DOWNGRADE is not expected - file is already closed when a second open
> is sent and both requests using the same session slot.
>
> Have you seen a similar issue on vanilla or rhel kernel?

I had a hard time triggering it consistently. I believe I have seen it
on RHEL7.2 kernel but I think I was more consistently seeing it on
some upstream (Trond's) kernel version (I think it was around 4.2).
The problem was seen by Netapp QA on 4.3-rc7 version.

Thanks for testing on the 4.7 version. I'll see what else went in that
might explain the failure on the older kernel.

>
> Thanks a lot,
>    Tigran.
>
> ----- Original Message -----
>> From: "Olga Kornievskaia" <aglo@xxxxxxxxx>
>> To: "Mkrtchyan, Tigran" <tigran.mkrtchyan@xxxxxxx>
>> Cc: "Andy Adamson" <William.Adamson@xxxxxxxxxx>, "Linux NFS Mailing List" <linux-nfs@xxxxxxxxxxxxxxx>, "Trond Myklebust"
>> <trond.myklebust@xxxxxxxxxxxxxxx>, "Steve Dickson" <steved@xxxxxxxxxx>
>> Sent: Thursday, July 14, 2016 4:52:59 PM
>> Subject: Re: Lost CLOSE with NFSv4.1 on RHEL7 ( and bejond?)
>
>> Hi Tigran,
>>
>> On Wed, Jul 13, 2016 at 12:49 PM, Mkrtchyan, Tigran
>> <tigran.mkrtchyan@xxxxxxx> wrote:
>>>
>>>
>>> Hi Andy,
>>>
>>> I will try to get upstream kernel on one of the nodes. It will take
>>> some time as we need to add a new host into the cluster and get
>>> some traffic go through it.
>>>
>>> In the mean while, with RHEL7 we get it easy reproduced - about 10
>>> such cases per day. Is there any tool that will help us to see where
>>> it happens? Some traces points? Call trace from vfs close to NFS close?
>>
>> There are NFS tracepoints but I don't know think there are VFS
>> tracepoints. Unfortunately, there was a bug in the OPEN tracepoints
>> that caused a kernel crash. I had a bugzilla out for RHEL7.2. It says
>> it's fixed in the later kernel (.381) but it's currently not back
>> ported to RHEL7.2z but hopefully will be soon (just chatted with Steve
>> about getting the fix into zstream). I made no progress in figuring
>> out what could be causing the lack of CLOSE and it was hard for me to
>> reproduce.
>>
>> Just recently Trond fixed a problem where a CLOSE that was suppose to
>> be sent as an OPEN_DOWNGRADE wasn't sent (commit 0979bc2a59) . I
>> wonder if that can be fixing this problem....
>>
>>> There is a one comment in the kernel code, which sounds similar:
>>> (http://git.linux-nfs.org/?p=trondmy/linux-nfs.git;a=blob;f=fs/nfs/nfs4proc.c;h=519368b987622ea23bea210929bebfd0c327e14e;hb=refs/heads/linux-next#l2955)
>>>
>>> nfs4proc.c: 2954
>>> ====
>>>
>>> /*
>>>  * It is possible for data to be read/written from a mem-mapped file
>>>  * after the sys_close call (which hits the vfs layer as a flush).
>>>  * This means that we can't safely call nfsv4 close on a file until
>>>  * the inode is cleared. This in turn means that we are not good
>>>  * NFSv4 citizens - we do not indicate to the server to update the file's
>>>  * share state even when we are done with one of the three share
>>>  * stateid's in the inode.
>>>  *
>>>  * NOTE: Caller must be holding the sp->so_owner semaphore!
>>>  */
>>> int nfs4_do_close(struct nfs4_state *state, gfp_t gfp_mask, int wait)
>>>
>>
>> I'm not sure if the comment means to say that there is a possibility
>> that NFS won't send a CLOSE (or at least I hope not). I thought that
>> because we keep a reference count on the inode and send the CLOSE when
>> it goes down to 0. Basically the last WRITE will trigger the nfs close
>> not the vfs_close.
>>
>>
>>> ====
>>>
>>>
>>> Tigran.
>>>
>>>
>>> ----- Original Message -----
>>>> From: "Andy Adamson" <William.Adamson@xxxxxxxxxx>
>>>> To: "Mkrtchyan, Tigran" <tigran.mkrtchyan@xxxxxxx>
>>>> Cc: "Linux NFS Mailing List" <linux-nfs@xxxxxxxxxxxxxxx>, "Andy Adamson"
>>>> <William.Adamson@xxxxxxxxxx>, "Trond Myklebust"
>>>> <trond.myklebust@xxxxxxxxxxxxxxx>, "Steve Dickson" <steved@xxxxxxxxxx>
>>>> Sent: Tuesday, July 12, 2016 7:16:19 PM
>>>> Subject: Re: Lost CLOSE with NFSv4.1 on RHEL7 ( and bejond?)
>>>
>>>> Hi Tigran
>>>>
>>>> Can you test with an upstream kernel? Olga has seen issues around no CLOSE being
>>>> sent - it is really hard to reproduce….
>>>>
>>>> —>Andy
>>>>
>>>>
>>>>> On Jul 7, 2016, at 6:49 AM, Mkrtchyan, Tigran <tigran.mkrtchyan@xxxxxxx> wrote:
>>>>>
>>>>>
>>>>>
>>>>> Dear NFS folks,
>>>>>
>>>>> we observe orphan open-states on our deployment with nfsv4.1.
>>>>> Our setup - two client nodes, running RHEL-7.2 with kernel
>>>>> 3.10.0-327.22.2.el7.x86_64. Both nodes running ownCloud (like
>>>>> a dropbox) which nfsv4.1 mounts to dCache storage. Some clients
>>>>> connected to node1, others to node2.
>>>>>
>>>>> Time-to-time we see some 'active' transfers on data our DS
>>>>> which do nothing. There is a corresponding state on MDS.
>>>>>
>>>>> I have traced one one such cases:
>>>>>
>>>>>  - node1 uploads the file.
>>>>>  - node2 reads the file couple of times, OPEN+LAYOUTGET+CLOSE
>>>>>  - node2 sends OPEN+LAYOUTGET
>>>>>  - there is no open file on node2 which points to it.
>>>>>  - CLOSE never send to the server.
>>>>>  - node1 eventually removes the removes the file
>>>>>
>>>>> We have many other cases where file is not removed, but this one I was
>>>>> able to trace. The link to capture files:
>>>>>
>>>>> https://desycloud.desy.de/index.php/s/YldowcRzTGJeLbN
>>>>>
>>>>> We had ~ 10^6 transfers in last 2 days and 29 files in such state (~0.0029%).
>>>>>
>>>> > Tigran.
>>> --
>>> To unsubscribe from this list: send the line "unsubscribe linux-nfs" in
>>> the body of a message to majordomo@xxxxxxxxxxxxxxx
>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>> --
>> To unsubscribe from this list: send the line "unsubscribe linux-nfs" in
>> the body of a message to majordomo@xxxxxxxxxxxxxxx
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
--
To unsubscribe from this list: send the line "unsubscribe linux-nfs" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [Linux Filesystem Development]     [Linux USB Development]     [Linux Media Development]     [Video for Linux]     [Linux NILFS]     [Linux Audio Users]     [Yosemite Info]     [Linux SCSI]

  Powered by Linux