SOLVED
Short: I followed the procedure to replace an OSD.
Long: I reweighted the flapping OSD to 0 until it was done, then marked it out, and then unmounted it, then followed the replacement procedure[1], then restore weight. I attempted to recreate my actions from bash history from various terminals. Please excuse any mistakes:
admin:$ X=0 # ID of OSD
admin:$ ceph osd crush reweight osd.$X 0.0
admin:$ ceph -w # observe data migration
admin:$ ceph osd out $X
osd:$ systemctl stop ceph.target
admin:$ ceph osd destroy $X --yes-i-really-mean-it
osd:$ umount /dev/sdb1
osd:$ ceph-disk zap /dev/sdb
osd:$ ceph-disk prepare --bluestore /dev/sdb --osd-id 0 --osd-uuid `uuidgen`
osd:$ ceph-disk activate /dev/sdb1
admin:$ ceph osd crush reweight osd.0 9.09560
admin:$ ceph -w # observe data migration
# I also found it helpful to set 'nodown' for several minutes because two of the OSDs kept marking eachother down.
admin:$ ceph osd set nodown
# Once all was 'active+clean', I removed the 'nodown' flag.
admin:$ ceph osd unset nodown
And for the first time since upgrading from Kraken to all the Luminous releases through 12.1.2, my cluster is finally HEALTH_OK. Yay!
roger@desktop:~/ceph-cluster$ ceph -w
cluster:
id: eea7b78c-b138-40fc-9f3e-3d77afb770f0
health: HEALTH_OK
services:
mon: 3 daemons, quorum nuc1,nuc2,nuc3
mgr: nuc3(active), standbys: nuc1, nuc2
osd: 3 osds: 3 up, 3 in
rgw: 1 daemon active
data:
pools: 19 pools, 372 pgs
objects: 54278 objects, 71724 MB
usage: 121 GB used, 27820 GB / 27941 GB avail
pgs: 372 active+clean
On Wed, Aug 2, 2017 at 11:08 AM Roger Brown <rogerpbrown@xxxxxxxxx> wrote:
Hi,My OSD's were continuously crashing in cephx_verify_authorizer() while on Luminous v12.1.0 and v12.1.1, but the crashes stopped once I upgraded to v12.1.2.Now however, one of my OSDs is continuing to crash. Looking closer, the crash reason is different reason and started with v12.1.1.I've been troubleshooting with the aid of http://docs.ceph.com/docs/master/rados/troubleshooting/troubleshooting-osd/.I'm considering reweight to 0 and then redeploy that OSD from scratch, unless you can do a filesystem repair on bluestore/rocksdb somehow. Please advise.
Data follows...Log:roger@osd3:~$ sudo journalctl -u ceph-osd@0 --no-pager...Aug 02 10:38:47 osd3 systemd[1]: ceph-osd@0.service: Failed with result 'signal'.Aug 02 10:39:07 osd3 systemd[1]: ceph-osd@0.service: Service hold-off time over, scheduling restart.Aug 02 10:39:07 osd3 systemd[1]: Stopped Ceph object storage daemon osd.0.Aug 02 10:39:07 osd3 systemd[1]: Starting Ceph object storage daemon osd.0...Aug 02 10:39:07 osd3 systemd[1]: Started Ceph object storage daemon osd.0.Aug 02 10:39:07 osd3 ceph-osd[7413]: starting osd.0 at - osd_data /var/lib/ceph/osd/ceph-0 /var/lib/ceph/osd/ceph-0/journalAug 02 10:40:48 osd3 ceph-osd[7413]: 2017-08-02 10:40:48.583063 7f5262cc3e00 -1 osd.0 25924 log_to_monitors {default=true}Aug 02 10:43:32 osd3 ceph-osd[7413]: *** Caught signal (Aborted) **Aug 02 10:43:32 osd3 ceph-osd[7413]: in thread 7f524861b700 thread_name:tp_osd_tpAug 02 10:43:32 osd3 ceph-osd[7413]: ceph version 12.1.2 (b661348f156f148d764b998b65b90451f096cb27) luminous (rc)Aug 02 10:43:32 osd3 ceph-osd[7413]: 1: (()+0xa9a964) [0x5623f0a9c964]Aug 02 10:43:32 osd3 ceph-osd[7413]: 2: (()+0x11390) [0x7f52611a6390]Aug 02 10:43:32 osd3 ceph-osd[7413]: 3: (pread64()+0x33) [0x7f52611a5d43]Aug 02 10:43:32 osd3 ceph-osd[7413]: 4: (KernelDevice::direct_read_unaligned(unsigned long, unsigned long, char*)+0x81) [0x5623f0a7cfc1]Aug 02 10:43:32 osd3 ceph-osd[7413]: 5: (KernelDevice::read_random(unsigned long, unsigned long, char*, bool)+0x4f3) [0x5623f0a7da43]Aug 02 10:43:32 osd3 ceph-osd[7413]: 6: (BlueFS::_read_random(BlueFS::FileReader*, unsigned long, unsigned long, char*)+0x4fa) [0x5623f0a4d9ca]Aug 02 10:43:32 osd3 ceph-osd[7413]: 7: (BlueRocksRandomAccessFile::Read(unsigned long, unsigned long, rocksdb::Slice*, char*) const+0x20) [0x5623f0a77e10]Aug 02 10:43:32 osd3 ceph-osd[7413]: 8: (rocksdb::RandomAccessFileReader::Read(unsigned long, unsigned long, rocksdb::Slice*, char*) const+0xf8f) [0x5623f0e50acf]Aug 02 10:43:32 osd3 ceph-osd[7413]: 9: (rocksdb::ReadBlockContents(rocksdb::RandomAccessFileReader*, rocksdb::Footer const&, rocksdb::ReadOptions const&, rocksdb::BlockHandle const&, rocksdb::BlockContents*, rocksdb::ImmutableCFOptions const&, bool, rocksdb::Slice const&, rocksdb::PersistentCacheOptions const&)+0x5f3) [0x5623f0e21383]Aug 02 10:43:32 osd3 ceph-osd[7413]: 10: (()+0xe0f7c6) [0x5623f0e117c6]Aug 02 10:43:32 osd3 ceph-osd[7413]: 11: (rocksdb::BlockBasedTable::MaybeLoadDataBlockToCache(rocksdb::BlockBasedTable::Rep*, rocksdb::ReadOptions const&, rocksdb::BlockHandle const&, rocksdb::Slice, rocksdb::BlockBasedTable::CachableEntry<rocksdb::Block>*, bool)+0x2f8) [0x5623f0e13928]Aug 02 10:43:32 osd3 ceph-osd[7413]: 12: (rocksdb::BlockBasedTable::NewDataBlockIterator(rocksdb::BlockBasedTable::Rep*, rocksdb::ReadOptions const&, rocksdb::BlockHandle const&, rocksdb::BlockIter*, bool, rocksdb::Status)+0x2ac) [0x5623f0e13d2c]Aug 02 10:43:32 osd3 ceph-osd[7413]: 13: (rocksdb::BlockBasedTable::BlockEntryIteratorState::NewSecondaryIterator(rocksdb::Slice const&)+0x97) [0x5623f0e1c4a7]Aug 02 10:43:32 osd3 ceph-osd[7413]: 14: (()+0xe4576e) [0x5623f0e4776e]Aug 02 10:43:32 osd3 ceph-osd[7413]: 15: (()+0xe45836) [0x5623f0e47836]Aug 02 10:43:32 osd3 ceph-osd[7413]: 16: (()+0xe459b1) [0x5623f0e479b1]Aug 02 10:43:32 osd3 ceph-osd[7413]: 17: (rocksdb::MergingIterator::Next()+0x449) [0x5623f0e2ab09]Aug 02 10:43:32 osd3 ceph-osd[7413]: 18: (rocksdb::DBIter::FindNextUserEntryInternal(bool, bool)+0x182) [0x5623f0ec7ed2]Aug 02 10:43:32 osd3 ceph-osd[7413]: 19: (rocksdb::DBIter::Next()+0x1eb) [0x5623f0ec8c8b]Aug 02 10:43:32 osd3 ceph-osd[7413]: 20: (RocksDBStore::RocksDBWholeSpaceIteratorImpl::next()+0x9a) [0x5623f09dd58a]Aug 02 10:43:32 osd3 ceph-osd[7413]: 21: (BlueStore::_collection_list(BlueStore::Collection*, ghobject_t const&, ghobject_t const&, int, std::vector<ghobject_t, std::allocator<ghobject_t> >*, ghobject_t*)+0x1170) [0x5623f093d250]Aug 02 10:43:32 osd3 ceph-osd[7413]: 22: (BlueStore::collection_list(boost::intrusive_ptr<ObjectStore::CollectionImpl>&, ghobject_t const&, ghobject_t const&, int, std::vector<ghobject_t, std::allocator<ghobject_t> >*, ghobject_t*)+0x25a) [0x5623f093e6ea]Aug 02 10:43:32 osd3 ceph-osd[7413]: 23: (PGBackend::objects_list_range(hobject_t const&, hobject_t const&, snapid_t, std::vector<hobject_t, std::allocator<hobject_t> >*, std::vector<ghobject_t, std::allocator<ghobject_t> >*)+0x192) [0x5623f0700ef2]Aug 02 10:43:32 osd3 ceph-osd[7413]: 24: (PG::build_scrub_map_chunk(ScrubMap&, hobject_t, hobject_t, bool, unsigned int, ThreadPool::TPHandle&)+0x200) [0x5623f05a8b30]Aug 02 10:43:32 osd3 ceph-osd[7413]: 25: (PG::chunky_scrub(ThreadPool::TPHandle&)+0x3ea) [0x5623f05d61ca]Aug 02 10:43:32 osd3 ceph-osd[7413]: 26: (PG::scrub(unsigned int, ThreadPool::TPHandle&)+0x45c) [0x5623f05d7cec]Aug 02 10:43:32 osd3 ceph-osd[7413]: 27: (OSD::ShardedOpWQ::_process(unsigned int, ceph::heartbeat_handle_d*)+0x12d0) [0x5623f05179e0]Aug 02 10:43:32 osd3 ceph-osd[7413]: 28: (ShardedThreadPool::shardedthreadpool_worker(unsigned int)+0x884) [0x5623f0ae44e4]Aug 02 10:43:32 osd3 ceph-osd[7413]: 29: (ShardedThreadPool::WorkThreadSharded::entry()+0x10) [0x5623f0ae7520]Aug 02 10:43:32 osd3 ceph-osd[7413]: 30: (()+0x76ba) [0x7f526119c6ba]Aug 02 10:43:32 osd3 ceph-osd[7413]: 31: (clone()+0x6d) [0x7f52602133dd]Aug 02 10:43:32 osd3 ceph-osd[7413]: 2017-08-02 10:43:32.251841 7f524861b700 -1 *** Caught signal (Aborted) **Aug 02 10:43:32 osd3 ceph-osd[7413]: in thread 7f524861b700 thread_name:tp_osd_tpAug 02 10:43:32 osd3 ceph-osd[7413]: ceph version 12.1.2 (b661348f156f148d764b998b65b90451f096cb27) luminous (rc)Aug 02 10:43:32 osd3 ceph-osd[7413]: 1: (()+0xa9a964) [0x5623f0a9c964]Aug 02 10:43:32 osd3 ceph-osd[7413]: 2: (()+0x11390) [0x7f52611a6390]Aug 02 10:43:32 osd3 ceph-osd[7413]: 3: (pread64()+0x33) [0x7f52611a5d43]Aug 02 10:43:32 osd3 ceph-osd[7413]: 4: (KernelDevice::direct_read_unaligned(unsigned long, unsigned long, char*)+0x81) [0x5623f0a7cfc1]Aug 02 10:43:32 osd3 ceph-osd[7413]: 5: (KernelDevice::read_random(unsigned long, unsigned long, char*, bool)+0x4f3) [0x5623f0a7da43]Aug 02 10:43:32 osd3 ceph-osd[7413]: 6: (BlueFS::_read_random(BlueFS::FileReader*, unsigned long, unsigned long, char*)+0x4fa) [0x5623f0a4d9ca]Aug 02 10:43:32 osd3 ceph-osd[7413]: 7: (BlueRocksRandomAccessFile::Read(unsigned long, unsigned long, rocksdb::Slice*, char*) const+0x20) [0x5623f0a77e10]Aug 02 10:43:32 osd3 ceph-osd[7413]: 8: (rocksdb::RandomAccessFileReader::Read(unsigned long, unsigned long, rocksdb::Slice*, char*) const+0xf8f) [0x5623f0e50acf]Aug 02 10:43:32 osd3 ceph-osd[7413]: 9: (rocksdb::ReadBlockContents(rocksdb::RandomAccessFileReader*, rocksdb::Footer const&, rocksdb::ReadOptions const&, rocksdb::BlockHandle const&, rocksdb::BlockContents*, rocksdb::ImmutableCFOptions const&, bool, rocksdb::Slice const&, rocksdb::PersistentCacheOptions const&)+0x5f3) [0x5623f0e21383]Aug 02 10:43:32 osd3 ceph-osd[7413]: 10: (()+0xe0f7c6) [0x5623f0e117c6]Aug 02 10:43:32 osd3 ceph-osd[7413]: 11: (rocksdb::BlockBasedTable::MaybeLoadDataBlockToCache(rocksdb::BlockBasedTable::Rep*, rocksdb::ReadOptions const&, rocksdb::BlockHandle const&, rocksdb::Slice, rocksdb::BlockBasedTable::CachableEntry<rocksdb::Block>*, bool)+0x2f8) [0x5623f0e13928]Aug 02 10:43:32 osd3 ceph-osd[7413]: 12: (rocksdb::BlockBasedTable::NewDataBlockIterator(rocksdb::BlockBasedTable::Rep*, rocksdb::ReadOptions const&, rocksdb::BlockHandle const&, rocksdb::BlockIter*, bool, rocksdb::Status)+0x2ac) [0x5623f0e13d2c]Aug 02 10:43:32 osd3 ceph-osd[7413]: 13: (rocksdb::BlockBasedTable::BlockEntryIteratorState::NewSecondaryIterator(rocksdb::Slice const&)+0x97) [0x5623f0e1c4a7]Aug 02 10:43:32 osd3 ceph-osd[7413]: 14: (()+0xe4576e) [0x5623f0e4776e]Aug 02 10:43:32 osd3 ceph-osd[7413]: 15: (()+0xe45836) [0x5623f0e47836]Aug 02 10:43:32 osd3 ceph-osd[7413]: 16: (()+0xe459b1) [0x5623f0e479b1]Aug 02 10:43:32 osd3 ceph-osd[7413]: 17: (rocksdb::MergingIterator::Next()+0x449) [0x5623f0e2ab09]Aug 02 10:43:32 osd3 ceph-osd[7413]: 18: (rocksdb::DBIter::FindNextUserEntryInternal(bool, bool)+0x182) [0x5623f0ec7ed2]Aug 02 10:43:32 osd3 ceph-osd[7413]: 19: (rocksdb::DBIter::Next()+0x1eb) [0x5623f0ec8c8b]Aug 02 10:43:32 osd3 ceph-osd[7413]: 20: (RocksDBStore::RocksDBWholeSpaceIteratorImpl::next()+0x9a) [0x5623f09dd58a]Aug 02 10:43:32 osd3 ceph-osd[7413]: 21: (BlueStore::_collection_list(BlueStore::Collection*, ghobject_t const&, ghobject_t const&, int, std::vector<ghobject_t, std::allocator<ghobject_t> >*, ghobject_t*)+0x1170) [0x5623f093d250]Aug 02 10:43:32 osd3 ceph-osd[7413]: 22: (BlueStore::collection_list(boost::intrusive_ptr<ObjectStore::CollectionImpl>&, ghobject_t const&, ghobject_t const&, int, std::vector<ghobject_t, std::allocator<ghobject_t> >*, ghobject_t*)+0x25a) [0x5623f093e6ea]Aug 02 10:43:32 osd3 ceph-osd[7413]: 23: (PGBackend::objects_list_range(hobject_t const&, hobject_t const&, snapid_t, std::vector<hobject_t, std::allocator<hobject_t> >*, std::vector<ghobject_t, std::allocator<ghobject_t> >*)+0x192) [0x5623f0700ef2]Aug 02 10:43:32 osd3 ceph-osd[7413]: 24: (PG::build_scrub_map_chunk(ScrubMap&, hobject_t, hobject_t, bool, unsigned int, ThreadPool::TPHandle&)+0x200) [0x5623f05a8b30]Aug 02 10:43:32 osd3 ceph-osd[7413]: 25: (PG::chunky_scrub(ThreadPool::TPHandle&)+0x3ea) [0x5623f05d61ca]Aug 02 10:43:32 osd3 ceph-osd[7413]: 26: (PG::scrub(unsigned int, ThreadPool::TPHandle&)+0x45c) [0x5623f05d7cec]Aug 02 10:43:32 osd3 ceph-osd[7413]: 27: (OSD::ShardedOpWQ::_process(unsigned int, ceph::heartbeat_handle_d*)+0x12d0) [0x5623f05179e0]Aug 02 10:43:32 osd3 ceph-osd[7413]: 28: (ShardedThreadPool::shardedthreadpool_worker(unsigned int)+0x884) [0x5623f0ae44e4]Aug 02 10:43:32 osd3 ceph-osd[7413]: 29: (ShardedThreadPool::WorkThreadSharded::entry()+0x10) [0x5623f0ae7520]Aug 02 10:43:32 osd3 ceph-osd[7413]: 30: (()+0x76ba) [0x7f526119c6ba]Aug 02 10:43:32 osd3 ceph-osd[7413]: 31: (clone()+0x6d) [0x7f52602133dd]Aug 02 10:43:32 osd3 ceph-osd[7413]: NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed to interpret this.Aug 02 10:43:32 osd3 ceph-osd[7413]: 0> 2017-08-02 10:43:32.251841 7f524861b700 -1 *** Caught signal (Aborted) **Aug 02 10:43:32 osd3 ceph-osd[7413]: in thread 7f524861b700 thread_name:tp_osd_tpAug 02 10:43:32 osd3 ceph-osd[7413]: ceph version 12.1.2 (b661348f156f148d764b998b65b90451f096cb27) luminous (rc)Aug 02 10:43:32 osd3 ceph-osd[7413]: 1: (()+0xa9a964) [0x5623f0a9c964]Aug 02 10:43:32 osd3 ceph-osd[7413]: 2: (()+0x11390) [0x7f52611a6390]Aug 02 10:43:32 osd3 ceph-osd[7413]: 3: (pread64()+0x33) [0x7f52611a5d43]Aug 02 10:43:32 osd3 ceph-osd[7413]: 4: (KernelDevice::direct_read_unaligned(unsigned long, unsigned long, char*)+0x81) [0x5623f0a7cfc1]Aug 02 10:43:32 osd3 ceph-osd[7413]: 5: (KernelDevice::read_random(unsigned long, unsigned long, char*, bool)+0x4f3) [0x5623f0a7da43]Aug 02 10:43:32 osd3 ceph-osd[7413]: 6: (BlueFS::_read_random(BlueFS::FileReader*, unsigned long, unsigned long, char*)+0x4fa) [0x5623f0a4d9ca]Aug 02 10:43:32 osd3 ceph-osd[7413]: 7: (BlueRocksRandomAccessFile::Read(unsigned long, unsigned long, rocksdb::Slice*, char*) const+0x20) [0x5623f0a77e10]Aug 02 10:43:32 osd3 ceph-osd[7413]: 8: (rocksdb::RandomAccessFileReader::Read(unsigned long, unsigned long, rocksdb::Slice*, char*) const+0xf8f) [0x5623f0e50acf]Aug 02 10:43:32 osd3 ceph-osd[7413]: 9: (rocksdb::ReadBlockContents(rocksdb::RandomAccessFileReader*, rocksdb::Footer const&, rocksdb::ReadOptions const&, rocksdb::BlockHandle const&, rocksdb::BlockContents*, rocksdb::ImmutableCFOptions const&, bool, rocksdb::Slice const&, rocksdb::PersistentCacheOptions const&)+0x5f3) [0x5623f0e21383]Aug 02 10:43:32 osd3 ceph-osd[7413]: 10: (()+0xe0f7c6) [0x5623f0e117c6]Aug 02 10:43:32 osd3 ceph-osd[7413]: 11: (rocksdb::BlockBasedTable::MaybeLoadDataBlockToCache(rocksdb::BlockBasedTable::Rep*, rocksdb::ReadOptions const&, rocksdb::BlockHandle const&, rocksdb::Slice, rocksdb::BlockBasedTable::CachableEntry<rocksdb::Block>*, bool)+0x2f8) [0x5623f0e13928]Aug 02 10:43:32 osd3 ceph-osd[7413]: 12: (rocksdb::BlockBasedTable::NewDataBlockIterator(rocksdb::BlockBasedTable::Rep*, rocksdb::ReadOptions const&, rocksdb::BlockHandle const&, rocksdb::BlockIter*, bool, rocksdb::Status)+0x2ac) [0x5623f0e13d2c]Aug 02 10:43:32 osd3 ceph-osd[7413]: 13: (rocksdb::BlockBasedTable::BlockEntryIteratorState::NewSecondaryIterator(rocksdb::Slice const&)+0x97) [0x5623f0e1c4a7]Aug 02 10:43:32 osd3 ceph-osd[7413]: 14: (()+0xe4576e) [0x5623f0e4776e]Aug 02 10:43:32 osd3 ceph-osd[7413]: 15: (()+0xe45836) [0x5623f0e47836]Aug 02 10:43:32 osd3 ceph-osd[7413]: 16: (()+0xe459b1) [0x5623f0e479b1]Aug 02 10:43:32 osd3 ceph-osd[7413]: 17: (rocksdb::MergingIterator::Next()+0x449) [0x5623f0e2ab09]Aug 02 10:43:32 osd3 ceph-osd[7413]: 18: (rocksdb::DBIter::FindNextUserEntryInternal(bool, bool)+0x182) [0x5623f0ec7ed2]Aug 02 10:43:32 osd3 ceph-osd[7413]: 19: (rocksdb::DBIter::Next()+0x1eb) [0x5623f0ec8c8b]Aug 02 10:43:32 osd3 ceph-osd[7413]: 20: (RocksDBStore::RocksDBWholeSpaceIteratorImpl::next()+0x9a) [0x5623f09dd58a]Aug 02 10:43:32 osd3 ceph-osd[7413]: 21: (BlueStore::_collection_list(BlueStore::Collection*, ghobject_t const&, ghobject_t const&, int, std::vector<ghobject_t, std::allocator<ghobject_t> >*, ghobject_t*)+0x1170) [0x5623f093d250]Aug 02 10:43:32 osd3 ceph-osd[7413]: 22: (BlueStore::collection_list(boost::intrusive_ptr<ObjectStore::CollectionImpl>&, ghobject_t const&, ghobject_t const&, int, std::vector<ghobject_t, std::allocator<ghobject_t> >*, ghobject_t*)+0x25a) [0x5623f093e6ea]Aug 02 10:43:32 osd3 ceph-osd[7413]: 23: (PGBackend::objects_list_range(hobject_t const&, hobject_t const&, snapid_t, std::vector<hobject_t, std::allocator<hobject_t> >*, std::vector<ghobject_t, std::allocator<ghobject_t> >*)+0x192) [0x5623f0700ef2]Aug 02 10:43:32 osd3 ceph-osd[7413]: 24: (PG::build_scrub_map_chunk(ScrubMap&, hobject_t, hobject_t, bool, unsigned int, ThreadPool::TPHandle&)+0x200) [0x5623f05a8b30]Aug 02 10:43:32 osd3 ceph-osd[7413]: 25: (PG::chunky_scrub(ThreadPool::TPHandle&)+0x3ea) [0x5623f05d61ca]Aug 02 10:43:32 osd3 ceph-osd[7413]: 26: (PG::scrub(unsigned int, ThreadPool::TPHandle&)+0x45c) [0x5623f05d7cec]Aug 02 10:43:32 osd3 ceph-osd[7413]: 27: (OSD::ShardedOpWQ::_process(unsigned int, ceph::heartbeat_handle_d*)+0x12d0) [0x5623f05179e0]Aug 02 10:43:32 osd3 ceph-osd[7413]: 28: (ShardedThreadPool::shardedthreadpool_worker(unsigned int)+0x884) [0x5623f0ae44e4]Aug 02 10:43:32 osd3 ceph-osd[7413]: 29: (ShardedThreadPool::WorkThreadSharded::entry()+0x10) [0x5623f0ae7520]Aug 02 10:43:32 osd3 ceph-osd[7413]: 30: (()+0x76ba) [0x7f526119c6ba]Aug 02 10:43:32 osd3 ceph-osd[7413]: 31: (clone()+0x6d) [0x7f52602133dd]Aug 02 10:43:32 osd3 ceph-osd[7413]: NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed to interpret this.Aug 02 10:43:32 osd3 systemd[1]: ceph-osd@0.service: Main process exited, code=killed, status=6/ABRTAug 02 10:43:32 osd3 systemd[1]: ceph-osd@0.service: Unit entered failed state.Aug 02 10:43:32 osd3 systemd[1]: ceph-osd@0.service: Failed with result 'signal'.Aug 02 10:43:52 osd3 systemd[1]: ceph-osd@0.service: Service hold-off time over, scheduling restart.Aug 02 10:43:52 osd3 systemd[1]: Stopped Ceph object storage daemon osd.0.Aug 02 10:43:52 osd3 systemd[1]: Starting Ceph object storage daemon osd.0...Aug 02 10:43:52 osd3 systemd[1]: Started Ceph object storage daemon osd.0.Aug 02 10:43:52 osd3 ceph-osd[8322]: starting osd.0 at - osd_data /var/lib/ceph/osd/ceph-0 /var/lib/ceph/osd/ceph-0/journalroger@desktop:~$ ceph -scluster:id: eea7b78c-b138-40fc-9f3e-3d77afb770f0health: HEALTH_WARN1 osds down1 host (1 osds) downDegraded data redundancy: 43922/162834 objects degraded (26.973%), 300 pgs unclean, 305 pgs degraded114 pgs not deep-scrubbed for 86400155 pgs not scrubbed for 8640010 slow requests are blocked > 32 secservices:mon: 3 daemons, quorum nuc1,nuc2,nuc3mgr: nuc3(active), standbys: nuc2, nuc1osd: 3 osds: 2 up, 3 inrgw: 1 daemon activedata:pools: 19 pools, 372 pgsobjects: 54278 objects, 71724 MBusage: 122 GB used, 27819 GB / 27941 GB availpgs: 43922/162834 objects degraded (26.973%)303 active+undersized+degraded67 active+clean2 active+recovery_wait+degradedroger@desktop:~$ ceph osd treeID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF-1 27.28679 root default-5 9.09560 host osd13 hdd 9.09560 osd.3 up 1.00000 1.00000-6 9.09560 host osd24 hdd 9.09560 osd.4 up 1.00000 1.00000-2 9.09560 host osd30 hdd 9.09560 osd.0 down 1.00000 1.00000roger@desktop:~$ ceph mon versions{"ceph version 12.1.2 (b661348f156f148d764b998b65b90451f096cb27) luminous (rc)": 3}roger@desktop:~$ ceph osd versions{"ceph version 12.1.2 (b661348f156f148d764b998b65b90451f096cb27) luminous (rc)": 2}roger@osd3:~$ sudo ceph daemon osd.0 status{"cluster_fsid": "eea7b78c-b138-40fc-9f3e-3d77afb770f0","osd_fsid": "bdb31a03-e381-4bf8-82e3-18916c838308","whoami": 0,"state": "waiting_for_healthy","oldest_map": 25389,"newest_map": 25938,"num_pgs": 372}roger@desktop:~$ ceph dfGLOBAL:SIZE AVAIL RAW USED %RAW USED27941G 27819G 122G 0.44POOLS:NAME ID USED %USED MAX AVAIL OBJECTSdefault.rgw.rgw.gc 70 0 0 8807G 0default.rgw.buckets.non-ec 83 0 0 8807G 43default.rgw.control 85 0 0 8807G 8default.rgw.data.root 86 15601 0 8807G 49default.rgw.gc 87 0 0 8807G 32default.rgw.lc 88 0 0 8807G 32default.rgw.log 89 0 0 8807G 144default.rgw.users.uid 90 3346 0 8807G 14default.rgw.users.email 91 100 0 8807G 7default.rgw.users.keys 92 100 0 8807G 7default.rgw.buckets.index 93 0 0 8807G 39default.rgw.intent-log 95 0 0 8807G 0default.rgw.meta 96 0 0 8807G 0default.rgw.usage 97 0 0 8807G 0default.rgw.users.swift 98 39 0 8807G 4default.rgw.buckets.extra 99 0 0 8807G 0.rgw.root 100 1681 0 8807G 4default.rgw.reshard 101 0 0 8807G 17default.rgw.buckets.data 103 71724M 0.40 17614G 53878
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com