Re: Cephfs default data pool (inode backtrace) no longer a thing?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Never mind. I don't know what changed, but I am seeing 0-size objects in the default pool now.

Vlad

On 3/16/22 11:02, Vladimir Brik wrote:
 > Are you sure there are no objects?
Yes. In the 16.2.7 cluster "ceph df" reports no objects in the default data pool. I am wondering if I need to something special ensure that recovery data is stored in a fast pool and not together with data in the EC pool.

In my other cluster that was deployed when Ceph was on version 14 (but upgraded to 15 since), there are a lot of 0-size objects in the default data pool.

Vlad


On 3/16/22 04:16, Frank Schilder wrote:
Are you sure there are no objects? Here is what it looks on our FS:

     NAME                     ID     USED        %USED MAX AVAIL     OBJECTS      con-fs2-meta1            12     474 MiB 0.04       1.0 TiB      35687606      con-fs2-meta2            13         0 B 0       1.0 TiB     300163323

Meta1 is the meta-data pool and meta2 the default data pool. It shows 0 bytes, but contains 10x the objects that sit in the meta data pool. These objects contain only meta data. That's why no actual usage is reported (at least on mimic).

The data in this default data pool is a serious challenge for recovery. I put it on fast SSDs, but the large number of objects requires aggressive recovery options. With the default settings recovery of this pool takes longer than the rebuild of data in the EC data pools on HDD. I also allocated lots of PGs to it to reduce the object count per PG. Having this data on fast drives with tuned settings helps a lot with overall recovery and snaptrim.

Best regards,
=================
Frank Schilder
AIT Risø Campus
Bygning 109, rum S14

________________________________________
From: Vladimir Brik <vladimir.brik@xxxxxxxxxxxxxxxx>
Sent: 15 March 2022 20:53:25
To: ceph-users
Subject: Cephfs default data pool (inode backtrace) no longer a thing?

Hello

https://docs.ceph.com/en/latest/cephfs/createfs/ mentions a
"default data pool" that is used for "inode backtrace
information, which is used for hard link management and
disaster recovery", and "all CephFS inodes have at least one
object in the default data pool".

I noticed that when I create a volume using "ceph fs volume
create" and then add the EC data pool where my files
actually are, the default pool remains empty (no objects).

Does this mean that the recommendation from the link above
"If erasure-coded pools are planned for file system data, it
is best to configure the default as a replicated pool" is no
longer applicable, or do I need to configure something to
avoid a performance hit when using EC data pools?


Thanks

Vlad
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux