In case the root is EC, it is likely that is not possible to apply the disaster recovery procedure, (no xattr layout/parent on the data pool). ________________________________________________________ Cordialement, *David CASIER* ________________________________________________________ Le jeu. 25 janv. 2024 à 13:03, Eugen Block <eblock@xxxxxx> a écrit : > I'm not sure if using EC as default data pool for cephfs is still > discouraged as stated in the output when attempting to do that, the > docs don't mention that (at least not in the link I sent in the last > mail): > > ceph:~ # ceph fs new cephfs cephfs_metadata cephfs_data > Error EINVAL: pool 'cephfs_data' (id '8') is an erasure-coded pool. > Use of an EC pool for the default data pool is discouraged; see the > online CephFS documentation for more information. Use --force to > override. > > ceph:~ # ceph fs new cephfs cephfs_metadata cephfs_data --force > new fs with metadata pool 6 and data pool 8 > > CC'ing Zac here to hopefully clear that up. > > Zitat von "David C." <david.casier@xxxxxxxx>: > > > Albert, > > Never used EC for (root) data pool. > > > > Le jeu. 25 janv. 2024 à 12:08, Albert Shih <Albert.Shih@xxxxxxxx> a > écrit : > > > >> Le 25/01/2024 à 08:42:19+0000, Eugen Block a écrit > >> > Hi, > >> > > >> > it's really as easy as it sounds (fresh test cluster on 18.2.1 without > >> any > >> > pools yet): > >> > > >> > ceph:~ # ceph fs volume create cephfs > >> > >> Yes...I already try that with the label and works fine. > >> > >> But I prefer to use «my» pools. Because I have ssd/hdd and want also try > >> «erasure coding» pool for the data. > >> > > > >> I also need to set the pg_num and pgp_num (I know I can do that after > the > >> creation). > > > > > >> So I manage to do ... half what I want... > >> > >> In fact > >> > >> ceph fs volume create thing > >> > >> will create two pools > >> > >> cephfs.thing.meta > >> cephfs.thing.data > >> > >> and if those pool already existe it will use them. > >> > >> But that's only if the data are replicated no with erasure > coding....(maybe > >> I forget something config on the pool). > >> > >> Well I will currently continue my test with replicated data. > >> > >> > The pools and the daemons are created automatically (you can control > the > >> > placement of the daemons with the --placement option). Note that the > >> > metadata pool needs to be on fast storage, so you might need to change > >> the > >> > ruleset for the metadata pool after creation in case you have HDDs in > >> place. > >> > Changing pools after the creation can be done via ceph fs commands: > >> > > >> > ceph:~ # ceph osd pool create cephfs_data2 > >> > pool 'cephfs_data2' created > >> > > >> > ceph:~ # ceph fs add_data_pool cephfs cephfs_data2 > >> > Pool 'cephfs_data2' (id '4') has pg autoscale mode 'on' but is not > >> marked > >> > as bulk. > >> > Consider setting the flag by running > >> > # ceph osd pool set cephfs_data2 bulk true > >> > added data pool 4 to fsmap > >> > > >> > ceph:~ # ceph fs status > >> > cephfs - 0 clients > >> > ====== > >> > RANK STATE MDS ACTIVITY DNS INOS > DIRS > >> > CAPS > >> > 0 active cephfs.soc9-ceph.uqcybj Reqs: 0 /s 10 13 > 12 > >> > 0 > >> > POOL TYPE USED AVAIL > >> > cephfs.cephfs.meta metadata 64.0k 13.8G > >> > cephfs.cephfs.data data 0 13.8G > >> > cephfs_data2 data 0 13.8G > >> > > >> > > >> > You can't remove the default data pool, though (here it's > >> > cephfs.cephfs.data). If you want to control the pool creation you can > >> fall > >> > back to the method you mentioned, create pools as you require them and > >> then > >> > create a new cephfs, and deploy the mds service. > >> > >> Yes, but I'm guessing the > >> > >> ceph fs volume > >> > >> are the «future» so it would be super nice to add (at least) the option > to > >> choose the couple of pool... > >> > >> > > >> > I haven't looked too deep into changing the default pool yet, so there > >> might > >> > be a way to switch that as well. > >> > >> Ok. I will also try but...well...newbie ;-) > >> > >> Anyway thanks. > >> > >> regards > >> > >> -- > >> Albert SHIH 🦫 🐸 > >> France > >> Heure locale/Local time: > >> jeu. 25 janv. 2024 12:00:08 CET > >> _______________________________________________ > >> ceph-users mailing list -- ceph-users@xxxxxxx > >> To unsubscribe send an email to ceph-users-leave@xxxxxxx > >> > > _______________________________________________ > > ceph-users mailing list -- ceph-users@xxxxxxx > > To unsubscribe send an email to ceph-users-leave@xxxxxxx > > > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx