Hi Stefan,
We don’t have automatic conversion going on , and the « bluestore_fsck_quick_fix_on_mount » is not set .
So we did an offline compaction as suggested but this didn’t fix the problem os osd crush .
In the meantime we are rebuilding all OSDs on the cluster and it seems it improve the cluster overall performance and has removed the slow ops , so far …
But still I can’t figure out why same OSDs may still encounter random crush . Here below the last log from a crashed OSD :
{
"crash_id": "2022-08-29T23:25:21.xxxxx -4f61-8c1c-f3542f83568d",
"timestamp": "2022-08-29T23:25:21.501287Z",
"process_name": "ceph-osd",
"entity_name": "osd.39",
"ceph_version": "16.2.9",
"utsname_hostname": "xxxxx",
"utsname_sysname": "Linux",
"utsname_release": "4.15.0-162-generic",
"utsname_version": "#170-Ubuntu SMP Mon Oct 18 11:38:05 UTC 2021",
"utsname_machine": "x86_64",
"os_name": "Ubuntu",
"os_id": "ubuntu",
"os_version_id": "18.04",
"os_version": "18.04.6 LTS (Bionic Beaver)",
"backtrace": [
"/lib/x86_64-linux-gnu/libpthread.so.0(+0x12980) [0x7efef8051980]",
"/usr/bin/ceph-osd(+0xf89db0) [0x55a9a8bdddb0]",
"/usr/bin/ceph-osd(+0xf9472e) [0x55a9a8be872e]",
"/usr/bin/ceph-osd(+0xf950fd) [0x55a9a8be90fd]",
"(BlueStore::_collection_list(BlueStore::Collection*, ghobject_t const&, ghobject_t const&, int, bool, std::vector<ghobject_t, std::allocator<ghobject_t> >*, ghobject_t*)+0x13d2) [0x55a9a8c261a2]",
"(BlueStore::collection_list(boost::intrusive_ptr<ObjectStore::CollectionImpl>&, ghobject_t const&, ghobject_t const&, int, std::vector<ghobject_t, std::allocator<ghobject_t> >*, ghobject_t*)+0xad) [0x55a9a8c26f2d]",
"(PGBackend::objects_list_partial(hobject_t const&, int, int, std::vector<hobject_t, std::allocator<hobject_t> >*, hobject_t*)+0x68d) [0x55a9a88f0e4d]",
"(PgScrubber::select_range()+0x2c2) [0x55a9a8a72ba2]",
"(PgScrubber::select_range_n_notify()+0x24) [0x55a9a8a736d4]",
"(Scrub::NewChunk::NewChunk(boost::statechart::state<Scrub::NewChunk, Scrub::ActiveScrubbing, boost::mpl::list<mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na>, (boost::statechart::history_mode)0>::my_context)+0xf8) [0x55a9a8a8c888]",
"(boost::statechart::simple_state<Scrub::PendingTimer, Scrub::ActiveScrubbing, boost::mpl::list<mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na>, (boost::statechart::history_mode)0>::react_impl(boost::statechart::event_base const&, void const*)+0x16a) [0x55a9a8a9642a]",
"(boost::statechart::state_machine<Scrub::ScrubMachine, Scrub::NotActive, std::allocator<boost::statechart::none>, boost::statechart::null_exception_translator>::process_event(boost::statechart::event_base const&)+0x6b) [0x55a9a8a8863b]",
"(PgScrubber::send_scrub_resched(unsigned int)+0xef) [0x55a9a8a7fc4f]",
"(PG::forward_scrub_event(void (ScrubPgIF::*)(unsigned int), unsigned int, std::basic_string_view<char, std::char_traits<char> >)+0x78) [0x55a9a87cdf48]",
"(ceph::osd::scheduler::PGScrubResched::run(OSD*, OSDShard*, boost::intrusive_ptr<PG>&, ThreadPool::TPHandle&)+0x32) [0x55a9a897b2f2]",
"(OSD::ShardedOpWQ::_process(unsigned int, ceph::heartbeat_handle_d*)+0xd1e) [0x55a9a8736dbe]",
"(ShardedThreadPool::shardedthreadpool_worker(unsigned int)+0x4ac) [0x55a9a8dbc75c]",
"(ShardedThreadPool::WorkThreadSharded::entry()+0x10) [0x55a9a8dbfc20]",
"/lib/x86_64-linux-gnu/libpthread.so.0(+0x76db) [0x7efef80466db]",
"clone()"
]
}
Many Thanks
On 8/22/22 11: 50, Wissem MIMOUNA wrote: > Dear All, > > After updating our ceph cluster from Octopus to Pacific , we got a lot of a slow_ops on many osds ( which caused the cluster to become very slow ) . Do you have any automatic
ZjQcmQRYFpfptBannerStart
ZjQcmQRYFpfptBannerEnd
On 8/22/22 11:50, Wissem MIMOUNA wrote:
Dear All,
After updating our ceph cluster from Octopus to Pacific , we got a lot of a slow_ops on many osds ( which caused the cluster to become very slow ) .
Do you have any automatic conversion going on? What is the setting of
"bluestore_fsck_quick_fix_on_mount" on your cluster OSDs?
ceph daemon osd.$id config get bluestore_fsck_quick_fix_on_mount
We did our investiguation and search on the ceph-users list and we found that rebuilding all OSD scan improve ( or fix ) the issue ( we have a doubt that the cause is a defragmented bluestore filesystem ) ,
RocksDB itself might need to be compacted (offline operation). You might
want to stop all OSDs on a host and perform a compaction on them to see
if it helps with slow ops [1]:
df|grep "/var/lib/ceph/osd"|awk '{print $6}'|cut -d '-' -f 2|sort
-n|xargs -n 1 -P 10 -I OSD ceph-kvstore-tool bluestore-kv
/var/lib/ceph/osd/ceph-OSD compact
What is the status of your cluster (ceph -s)?
Gr. Stefan
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx