Re: [ceph 17.2.6] unable to create rbd snapshots for images with erasure code data-pool

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Wed, Apr 19, 2023 at 5:57 PM Reto Gysi <rlgysi@xxxxxxxxx> wrote:
>
>
> Hi,
>
> Am Mi., 19. Apr. 2023 um 11:02 Uhr schrieb Ilya Dryomov <idryomov@xxxxxxxxx>:
>>
>> On Wed, Apr 19, 2023 at 10:29 AM Reto Gysi <rlgysi@xxxxxxxxx> wrote:
>> >
>> > yes, I used the same ecpool_hdd also for cephfs file systems. The new pool ecpool_test I've created for a test, I've also created it with application profile 'cephfs', but there aren't any cephfs filesystem attached to it.
>>
>> This is not and has never been supported.
>
>
> Do you mean 1) using the same erasure coded pool for both rbd and cephfs, or  2) multiple cephfs filesystem using the same erasure coded pool via ceph.dir.layout.pool="ecpool_hdd"?

(1), using the same EC pool for both RBD and CephFS.

>
> 1)
>
>
> 2)
> rgysi cephfs filesystem
> rgysi - 5 clients
> =====
> RANK  STATE           MDS             ACTIVITY     DNS    INOS   DIRS   CAPS
> 0    active  rgysi.debian.uhgqen  Reqs:    0 /s   409k   408k  40.8k  16.5k
>       POOL          TYPE     USED  AVAIL
> cephfs.rgysi.meta  metadata  1454M  2114G
> cephfs.rgysi.data    data    4898G  17.6T
>    ecpool_hdd       data    29.3T  29.6T
>
> root@zephir:~# getfattr -n ceph.dir.layout /home/rgysi/am/ecpool/
> getfattr: Removing leading '/' from absolute path names
> # file: home/rgysi/am/ecpool/
> ceph.dir.layout="stripe_unit=4194304 stripe_count=1 object_size=4194304 pool=ecpool_hdd"
>
> root@zephir:~#
>
> backups cephfs filesystem
> backups - 2 clients
> =======
> RANK  STATE            MDS              ACTIVITY     DNS    INOS   DIRS   CAPS
> 0    active  backups.debian.runngh  Reqs:    0 /s   253k   253k  21.3k   899
>        POOL           TYPE     USED  AVAIL
> cephfs.backups.meta  metadata  1364M  2114G
> cephfs.backups.data    data    16.7T  16.4T
>     ecpool_hdd        data    29.3T  29.6T
>
> root@zephir:~# getfattr -n ceph.dir.layout /mnt/backups/windows/windows-drives/
> getfattr: Removing leading '/' from absolute path names
> # file: mnt/backups/windows/windows-drives/
> ceph.dir.layout="stripe_unit=4194304 stripe_count=1 object_size=4194304 pool=ecpool_hdd"
>
> root@zephir:~#
>
>
>
> So I guess I should use a different ec datapool for rbd and for each of the cephfs filesystems in the future, correct?

Definitely a different EC pool for RBD (i.e. don't mix with CephFS).
Not sure about the _each_ of the filesystems bit -- Venky or Patrick can
comment on whether sharing an EC pool between filesystems is OK.

Thanks,

                Ilya
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux