Hi,
db_slots is still not implemented:
pacific:~ # ceph orch apply -i osd.yml --dry-run
Error EINVAL: Failed to validate Drive Group: Filtering for <db_slots>
is not supported
Question 2: If db_slots still *doesn't* work, is there a coherent
way to divide up a solid state DB drive for use by a bunch of OSDs
when the OSDs may not all be created in one go? At first I thought
it was related to limit, but re-reading the advanced specification
for a 4th time, I don't think that's the case. Of course this
question is moot if db_slots actually works.
You can use the block_db_size filter so ceph-volume would not consume
the entire DB disk (SSD/NVMe), but it would try to deploy all
available data devices if you don't specify more filters. You can use
the --dry-run flag to try out some specs to get a feeling what
ceph-volume would actually do. Just as a short example from my lab
(all-in-one-host with three data devices and one SSD for block_db), this
service_type: osd
service_id: osd_spec_hdd_ssd
service_name: osd.hdd_ssd_mix
placement:
host_pattern: '*'
data_devices:
rotational: 1
limit: 2
db_devices:
rotational: 0
filter_logic: AND
block_db_size: 3G
unmanaged: false
would result in this deployment:
+---------+------------------+---------+----------+----------+-----+
|SERVICE |NAME |HOST |DATA |DB |WAL |
+---------+------------------+---------+----------+----------+-----+
|osd |osd_spec_hdd_ssd |pacific |/dev/vdb |/dev/vde |- |
|osd |osd_spec_hdd_ssd |pacific |/dev/vdc |/dev/vde |- |
+---------+------------------+---------+----------+----------+-----+
These are my available devices:
vdb 1 10G
vdc 1 10G
vdd 1 10G
vde 0 8G
So the limit filter works as expected here. If I don't specify it, I
wouldn't get any OSDs because ceph-volume can't fit three DBs of size
3 GB onto the 8 GB disk.
Does that help?
Regards,
Eugen
Zitat von Edward R Huyer <erhvks@xxxxxxx>:
I recently upgraded my existing cluster to Pacific and cephadm, and
need to reconfigure all the (rotational) OSDs to use NVMe drives for
db storage. I think I have a reasonably good idea how that's going
to work, but the use of db_slots and limit in the OSD service
specification have me scratching my head.
Question 1: Does db_slots actually work in the latest version of
Pacific? It's listed here
https://docs.ceph.com/en/pacific/cephadm/osd/#additional-options but
in the advanced case section
https://docs.ceph.com/en/pacific/cephadm/osd/#the-advanced-case
there's still a note saying it's not implemented.
Question 2: If db_slots still *doesn't* work, is there a coherent
way to divide up a solid state DB drive for use by a bunch of OSDs
when the OSDs may not all be created in one go? At first I thought
it was related to limit, but re-reading the advanced specification
for a 4th time, I don't think that's the case. Of course this
question is moot if db_slots actually works.
Any advice or information would be appreciated.
-----
Edward Huyer
Golisano College of Computing and Information Sciences
Rochester Institute of Technology
Golisano 70-2373
152 Lomb Memorial Drive
Rochester, NY 14623
585-475-6651
erhvks@xxxxxxx<mailto:erhvks@xxxxxxx>
Obligatory Legalese:
The information transmitted, including attachments, is intended only
for the person(s) or entity to which it is addressed and may contain
confidential and/or privileged material. Any review, retransmission,
dissemination or other use of, or taking of any action in reliance
upon this information by persons or entities other than the intended
recipient is prohibited. If you received this in error, please
contact the sender and destroy any copies of this information.
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx