On Thu, Dec 14, 2017 at 12:52 AM, Jens-U. Mozdzen <jmozdzen@xxxxxx> wrote: > Hi Yan, > > Zitat von "Yan, Zheng" <ukernel@xxxxxxxxx>: >> >> [...] >> >> It's likely some clients had caps on unlinked inodes, which prevent >> MDS from purging objects. When a file gets deleted, mds notifies all >> clients, clients are supposed to drop corresponding caps if possible. >> You may hit a bug in this area, some clients failed to drop cap for >> unlinked inodes. >> [...] >> There is a reconnect stage during MDS recovers. To reduce reconnect >> message size, clients trim unused inodes from their cache >> aggressively. In your case, most unlinked inodes also got trimmed . >> So mds could purge corresponding objects after it recovered > > > thank you for that detailed explanation. While I've already included the > recent code fix for this issue on a test node, all other mount points > (including the NFS server machine) still run thenon-fixed kernel Ceph > client. So your description makes me believe we've hit exactly what you > describe. > > Seems we'll have to fix the clients :) > > Is there a command I can use to see what caps a client holds, to verify the > proposed patch actually works? > No easy way. 'ceph daemon mds.x session ls' can show how many caps each client holds. 'ceph daemon mds.x dump cache' dump whole mds cache. that information can be extracted from the cache dump. Regards Yan, Zheng > Regards, > Jens > > PS: Is there a command I can use to see what caps a client holds, to verify > the proposed patch actually works? > > _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com