Never mind. I don't know what changed, but I am seeing
0-size objects in the default pool now.
Vlad
On 3/16/22 11:02, Vladimir Brik wrote:
> Are you sure there are no objects?
Yes. In the 16.2.7 cluster "ceph df" reports no objects in
the default data pool. I am wondering if I need to something
special ensure that recovery data is stored in a fast pool
and not together with data in the EC pool.
In my other cluster that was deployed when Ceph was on
version 14 (but upgraded to 15 since), there are a lot of
0-size objects in the default data pool.
Vlad
On 3/16/22 04:16, Frank Schilder wrote:
Are you sure there are no objects? Here is what it looks
on our FS:
NAME ID USED %USED
MAX AVAIL OBJECTS
con-fs2-meta1 12 474 MiB
0.04 1.0 TiB 35687606
con-fs2-meta2 13 0 B
0 1.0 TiB 300163323
Meta1 is the meta-data pool and meta2 the default data
pool. It shows 0 bytes, but contains 10x the objects that
sit in the meta data pool. These objects contain only meta
data. That's why no actual usage is reported (at least on
mimic).
The data in this default data pool is a serious challenge
for recovery. I put it on fast SSDs, but the large number
of objects requires aggressive recovery options. With the
default settings recovery of this pool takes longer than
the rebuild of data in the EC data pools on HDD. I also
allocated lots of PGs to it to reduce the object count per
PG. Having this data on fast drives with tuned settings
helps a lot with overall recovery and snaptrim.
Best regards,
=================
Frank Schilder
AIT Risø Campus
Bygning 109, rum S14
________________________________________
From: Vladimir Brik <vladimir.brik@xxxxxxxxxxxxxxxx>
Sent: 15 March 2022 20:53:25
To: ceph-users
Subject: Cephfs default data pool (inode
backtrace) no longer a thing?
Hello
https://docs.ceph.com/en/latest/cephfs/createfs/ mentions a
"default data pool" that is used for "inode backtrace
information, which is used for hard link management and
disaster recovery", and "all CephFS inodes have at least one
object in the default data pool".
I noticed that when I create a volume using "ceph fs volume
create" and then add the EC data pool where my files
actually are, the default pool remains empty (no objects).
Does this mean that the recommendation from the link above
"If erasure-coded pools are planned for file system data, it
is best to configure the default as a replicated pool" is no
longer applicable, or do I need to configure something to
avoid a performance hit when using EC data pools?
Thanks
Vlad
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx