The backtraces are written out asynchronously by the MDS to those objects, so there can be a delay between file creation and when they appear. In fact I think backtraces only get written when the inode in question is falling out of the MDS journal, so if you have a relatively small number of flies which are consistently getting updated, they can go arbitrarily long without the backtrace object getting written out to RADOS objects. Side note: I think you should be able to specify your EC pool as the default data pool for a filesystem, which would prevent you from needing a separate replicated pool storing those backtraces. Unless we still have a programmed limit from back when EC pools couldn't handle omap? -Greg On Thu, Mar 17, 2022 at 1:15 PM Vladimir Brik <vladimir.brik@xxxxxxxxxxxxxxxx> wrote: > > Never mind. I don't know what changed, but I am seeing > 0-size objects in the default pool now. > > Vlad > > On 3/16/22 11:02, Vladimir Brik wrote: > > > Are you sure there are no objects? > > Yes. In the 16.2.7 cluster "ceph df" reports no objects in > > the default data pool. I am wondering if I need to something > > special ensure that recovery data is stored in a fast pool > > and not together with data in the EC pool. > > > > In my other cluster that was deployed when Ceph was on > > version 14 (but upgraded to 15 since), there are a lot of > > 0-size objects in the default data pool. > > > > Vlad > > > > > > On 3/16/22 04:16, Frank Schilder wrote: > >> Are you sure there are no objects? Here is what it looks > >> on our FS: > >> > >> NAME ID USED %USED > >> MAX AVAIL OBJECTS > >> con-fs2-meta1 12 474 MiB > >> 0.04 1.0 TiB 35687606 > >> con-fs2-meta2 13 0 B > >> 0 1.0 TiB 300163323 > >> > >> Meta1 is the meta-data pool and meta2 the default data > >> pool. It shows 0 bytes, but contains 10x the objects that > >> sit in the meta data pool. These objects contain only meta > >> data. That's why no actual usage is reported (at least on > >> mimic). > >> > >> The data in this default data pool is a serious challenge > >> for recovery. I put it on fast SSDs, but the large number > >> of objects requires aggressive recovery options. With the > >> default settings recovery of this pool takes longer than > >> the rebuild of data in the EC data pools on HDD. I also > >> allocated lots of PGs to it to reduce the object count per > >> PG. Having this data on fast drives with tuned settings > >> helps a lot with overall recovery and snaptrim. > >> > >> Best regards, > >> ================= > >> Frank Schilder > >> AIT Risø Campus > >> Bygning 109, rum S14 > >> > >> ________________________________________ > >> From: Vladimir Brik <vladimir.brik@xxxxxxxxxxxxxxxx> > >> Sent: 15 March 2022 20:53:25 > >> To: ceph-users > >> Subject: Cephfs default data pool (inode > >> backtrace) no longer a thing? > >> > >> Hello > >> > >> https://docs.ceph.com/en/latest/cephfs/createfs/ mentions a > >> "default data pool" that is used for "inode backtrace > >> information, which is used for hard link management and > >> disaster recovery", and "all CephFS inodes have at least one > >> object in the default data pool". > >> > >> I noticed that when I create a volume using "ceph fs volume > >> create" and then add the EC data pool where my files > >> actually are, the default pool remains empty (no objects). > >> > >> Does this mean that the recommendation from the link above > >> "If erasure-coded pools are planned for file system data, it > >> is best to configure the default as a replicated pool" is no > >> longer applicable, or do I need to configure something to > >> avoid a performance hit when using EC data pools? > >> > >> > >> Thanks > >> > >> Vlad > >> _______________________________________________ > >> ceph-users mailing list -- ceph-users@xxxxxxx > >> To unsubscribe send an email to ceph-users-leave@xxxxxxx > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx