Oh right, I forgot about that, good point! But if that is (still) true
then this should definitely be in the docs as a warning for EC pools
in cephfs!
Zitat von "David C." <david.casier@xxxxxxxx>:
In case the root is EC, it is likely that is not possible to apply the
disaster recovery procedure, (no xattr layout/parent on the data pool).
________________________________________________________
Cordialement,
*David CASIER*
________________________________________________________
Le jeu. 25 janv. 2024 à 13:03, Eugen Block <eblock@xxxxxx> a écrit :
I'm not sure if using EC as default data pool for cephfs is still
discouraged as stated in the output when attempting to do that, the
docs don't mention that (at least not in the link I sent in the last
mail):
ceph:~ # ceph fs new cephfs cephfs_metadata cephfs_data
Error EINVAL: pool 'cephfs_data' (id '8') is an erasure-coded pool.
Use of an EC pool for the default data pool is discouraged; see the
online CephFS documentation for more information. Use --force to
override.
ceph:~ # ceph fs new cephfs cephfs_metadata cephfs_data --force
new fs with metadata pool 6 and data pool 8
CC'ing Zac here to hopefully clear that up.
Zitat von "David C." <david.casier@xxxxxxxx>:
> Albert,
> Never used EC for (root) data pool.
>
> Le jeu. 25 janv. 2024 à 12:08, Albert Shih <Albert.Shih@xxxxxxxx> a
écrit :
>
>> Le 25/01/2024 à 08:42:19+0000, Eugen Block a écrit
>> > Hi,
>> >
>> > it's really as easy as it sounds (fresh test cluster on 18.2.1 without
>> any
>> > pools yet):
>> >
>> > ceph:~ # ceph fs volume create cephfs
>>
>> Yes...I already try that with the label and works fine.
>>
>> But I prefer to use «my» pools. Because I have ssd/hdd and want also try
>> «erasure coding» pool for the data.
>>
>
>> I also need to set the pg_num and pgp_num (I know I can do that after
the
>> creation).
>
>
>> So I manage to do ... half what I want...
>>
>> In fact
>>
>> ceph fs volume create thing
>>
>> will create two pools
>>
>> cephfs.thing.meta
>> cephfs.thing.data
>>
>> and if those pool already existe it will use them.
>>
>> But that's only if the data are replicated no with erasure
coding....(maybe
>> I forget something config on the pool).
>>
>> Well I will currently continue my test with replicated data.
>>
>> > The pools and the daemons are created automatically (you can control
the
>> > placement of the daemons with the --placement option). Note that the
>> > metadata pool needs to be on fast storage, so you might need to change
>> the
>> > ruleset for the metadata pool after creation in case you have HDDs in
>> place.
>> > Changing pools after the creation can be done via ceph fs commands:
>> >
>> > ceph:~ # ceph osd pool create cephfs_data2
>> > pool 'cephfs_data2' created
>> >
>> > ceph:~ # ceph fs add_data_pool cephfs cephfs_data2
>> > Pool 'cephfs_data2' (id '4') has pg autoscale mode 'on' but is not
>> marked
>> > as bulk.
>> > Consider setting the flag by running
>> > # ceph osd pool set cephfs_data2 bulk true
>> > added data pool 4 to fsmap
>> >
>> > ceph:~ # ceph fs status
>> > cephfs - 0 clients
>> > ======
>> > RANK STATE MDS ACTIVITY DNS INOS
DIRS
>> > CAPS
>> > 0 active cephfs.soc9-ceph.uqcybj Reqs: 0 /s 10 13
12
>> > 0
>> > POOL TYPE USED AVAIL
>> > cephfs.cephfs.meta metadata 64.0k 13.8G
>> > cephfs.cephfs.data data 0 13.8G
>> > cephfs_data2 data 0 13.8G
>> >
>> >
>> > You can't remove the default data pool, though (here it's
>> > cephfs.cephfs.data). If you want to control the pool creation you can
>> fall
>> > back to the method you mentioned, create pools as you require them and
>> then
>> > create a new cephfs, and deploy the mds service.
>>
>> Yes, but I'm guessing the
>>
>> ceph fs volume
>>
>> are the «future» so it would be super nice to add (at least) the option
to
>> choose the couple of pool...
>>
>> >
>> > I haven't looked too deep into changing the default pool yet, so there
>> might
>> > be a way to switch that as well.
>>
>> Ok. I will also try but...well...newbie ;-)
>>
>> Anyway thanks.
>>
>> regards
>>
>> --
>> Albert SHIH 🦫 🐸
>> France
>> Heure locale/Local time:
>> jeu. 25 janv. 2024 12:00:08 CET
>> _______________________________________________
>> ceph-users mailing list -- ceph-users@xxxxxxx
>> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>>
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx