This is what leads me to believe it's other settings being referred to as well:
"There are dozens of documents floating around with long lists of Ceph
configurables that have been tuned for optimal performance on specific
hardware or for specific workloads. In most cases these ceph.conf
fragments tend to induce funny looks on developers’ faces because the
settings being adjusted seem counter-intuitive, unrelated to the
performance of the system, and/or outright dangerous. Our goal is to
make Ceph work as well as we can out of the box without requiring any
tuning at all, so we are always striving to choose sane defaults. And
generally, we discourage tuning by users. "
To me it's not just bluestore settings / sdd vs. hdd they're talking about ("dozens of documents floating around"... "our goal... without any tuning at all". Am I off base?
Regards
On Thu, Jul 12, 2018 at 9:12 PM, Konstantin Shalygin <k0ste@xxxxxxxx> wrote:
This mean that some "bluestore_*" settings tuned for nvme/hdd separately.I saw this in the Luminous release notes:
"Each OSD now adjusts its default configuration based on whether the
backing device is an HDD or SSD. Manual tuning generally not required"
Which tuning in particular? The ones in my configuration are
osd_op_threads, osd_disk_threads, osd_recovery_max_active,
osd_op_thread_suicide_timeout, and osd_crush_chooseleaf_type, among
others. Can I rip these out when I upgrade to
Luminous?
Also with Luminous we have:
osd_op_num_shards_(ssd|hdd)
osd_op_num_threads_per_shard_(ssd|hdd)
osd_recovery_sleep_(ssd|hdd)
k
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com