Hi Frank, If you have one active MDS, the stray dir objects in the meta pool are named: 600.00000000 601.00000000 ... 609.00000000 So you can e.g. `rados listomapvals -p con-fs2-meta1 600.00000000` to get an idea about the stray files. Each of those stray dirs hold up to mds_bal_fragment_size_max entries. After they are full you'll get ENOSPACE on rm. It's safe to increase mds_bal_fragment_size_max to 200000 if that starts to happen. Cheers, Dan On Tue, Jan 18, 2022 at 10:53 AM Frank Schilder <frans@xxxxxx> wrote: > > Hi Dan and Patrick, > > this problem seems to develop into a nightmare. I executed a find on the file system and had some initial success. The number of stray files dropped by about 8%. Unfortunately, this is about it. I'm running a find now also on snap dirs, but I don't have much hope. There must be a way to find out what is accumulating in the stray buckets. As I wrote in another reply to this thread, I can't dump the trees: > > > I seem to have a problem. I cannot dump the mds tree: > > > > [root@ceph-08 ~]# ceph daemon mds.ceph-08 dump tree '~mdsdir/stray0' > > root inode is not in cache > > [root@ceph-08 ~]# ceph daemon mds.ceph-08 dump tree '~mds0/stray0' > > root inode is not in cache > > [root@ceph-08 ~]# ceph daemon mds.ceph-08 dump tree '~mds0' 0 > > root inode is not in cache > > [root@ceph-08 ~]# ceph daemon mds.ceph-08 dump tree '~mdsdir' 0 > > root inode is not in cache > > > > [root@ceph-08 ~]# ceph daemon mds.ceph-08 get subtrees | grep path > > "path": "", > > "path": "~mds0", > > > > However, this information is somewhere in rados objects and it should be possible to figure something out similar to > > # rados getxattr --pool=con-fs2-meta1 <OBJ_ID> parent | ceph-dencoder type inode_backtrace_t import - decode dump_json > # rados listomapkeys --pool=con-fs2-meta1 <OBJ_ID> > > What OBJ_IDs am I looking for? How and where can I start to traverse the structure? Version is mimic latest stable. > > Thanks for your help, > ================= > Frank Schilder > AIT Risø Campus > Bygning 109, rum S14 > > ________________________________________ > From: Dan van der Ster <dan@xxxxxxxxxxxxxx> > Sent: 17 January 2022 09:35:02 > To: Patrick Donnelly > Cc: Frank Schilder; ceph-users > Subject: Re: Re: cephfs: [ERR] loaded dup inode > > On Sun, Jan 16, 2022 at 3:54 PM Patrick Donnelly <pdonnell@xxxxxxxxxx> wrote: > > > > Hi Dan, > > > > On Fri, Jan 14, 2022 at 6:32 AM Dan van der Ster <dan@xxxxxxxxxxxxxx> wrote: > > > We had this long ago related to a user generating lots of hard links. > > > Snapshots will have a similar effect. > > > (in these cases, if a user deletes the original file, the file goes > > > into stray until it is "reintegrated"). > > > > > > If you can find the dir where they're working, `ls -lR` will force > > > those to reintegrate (you will see because the num strays will drop > > > back down). > > > You might have to ls -lR in a snap directory, or in the current tree > > > -- you have to browse around and experiment. > > > > > > pacific does this re-integration automatically. > > > > This reintegration is still not automatic (i.e. the MDS does not have > > a mechanism (yet) for hunting for the dentry to do reintegration). > > The next version (planned) of Pacific will have reintegration > > triggered by recursive scrub: > > > > https://github.com/ceph/ceph/pull/44514 > > > > which is significantly less disruptive than `ls -lR` or `find`. > > Oops, sorry, my bad. > I was thinking about https://github.com/ceph/ceph/pull/33479 > > Cheers, Dan > > > > > > -- > > Patrick Donnelly, Ph.D. > > He / Him / His > > Principal Software Engineer > > Red Hat, Inc. > > GPG: 19F28A586F808C2402351B93C3301A3E258DD79D > > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx