Re: [ceph 17.2.6] unable to create rbd snapshots for images with erasure code data-pool

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 




Hi,

Am Mi., 19. Apr. 2023 um 11:02 Uhr schrieb Ilya Dryomov <idryomov@xxxxxxxxx>:
On Wed, Apr 19, 2023 at 10:29 AM Reto Gysi <rlgysi@xxxxxxxxx> wrote:
>
> yes, I used the same ecpool_hdd also for cephfs file systems. The new pool ecpool_test I've created for a test, I've also created it with application profile 'cephfs', but there aren't any cephfs filesystem attached to it.

This is not and has never been supported.

Do you mean 1) using the same erasure coded pool for both rbd and cephfs, or  2) multiple cephfs filesystem using the same erasure coded pool via ceph.dir.layout.pool="ecpool_hdd"?

1)
image.png

2)
rgysi cephfs filesystem
rgysi - 5 clients
=====
RANK  STATE           MDS             ACTIVITY     DNS    INOS   DIRS   CAPS   
0    active  rgysi.debian.uhgqen  Reqs:    0 /s   409k   408k  40.8k  16.5k   
      POOL          TYPE     USED  AVAIL   
cephfs.rgysi.meta  metadata  1454M  2114G   
cephfs.rgysi.data    data    4898G  17.6T   
   ecpool_hdd       data    29.3T  29.6T 

root@zephir:~# getfattr -n ceph.dir.layout /home/rgysi/am/ecpool/
getfattr: Removing leading '/' from absolute path names
# file: home/rgysi/am/ecpool/
ceph.dir.layout="stripe_unit=4194304 stripe_count=1 object_size=4194304 pool=ecpool_hdd"

root@zephir:~#

backups cephfs filesystem
backups - 2 clients
=======
RANK  STATE            MDS              ACTIVITY     DNS    INOS   DIRS   CAPS   
0    active  backups.debian.runngh  Reqs:    0 /s   253k   253k  21.3k   899    
       POOL           TYPE     USED  AVAIL   
cephfs.backups.meta  metadata  1364M  2114G   
cephfs.backups.data    data    16.7T  16.4T   
    ecpool_hdd        data    29.3T  29.6T   

root@zephir:~# getfattr -n ceph.dir.layout /mnt/backups/windows/windows-drives/
getfattr: Removing leading '/' from absolute path names
# file: mnt/backups/windows/windows-drives/
ceph.dir.layout="stripe_unit=4194304 stripe_count=1 object_size=4194304 pool=ecpool_hdd"

root@zephir:~#



So I guess I should use a different ec datapool for rbd and for each of the cephfs filesystems in the future, correct?

Thanks & Cheers

Reto

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux