Hi,
Now however, one of my OSDs is continuing to crash. Looking closer, the crash reason is different reason and started with v12.1.1.
I've been troubleshooting with the aid of http://docs.ceph.com/docs/master/rados/troubleshooting/troubleshooting-osd/.
I'm considering reweight to 0 and then redeploy that OSD from scratch, unless you can do a filesystem repair on bluestore/rocksdb somehow. Please advise.
Data follows...
Log:
roger@osd3:~$ sudo journalctl -u ceph-osd@0 --no-pager
...
Aug 02 10:38:47 osd3 systemd[1]: ceph-osd@0.service: Failed with result 'signal'.
Aug 02 10:39:07 osd3 systemd[1]: ceph-osd@0.service: Service hold-off time over, scheduling restart.
Aug 02 10:39:07 osd3 systemd[1]: Stopped Ceph object storage daemon osd.0.
Aug 02 10:39:07 osd3 systemd[1]: Starting Ceph object storage daemon osd.0...
Aug 02 10:39:07 osd3 systemd[1]: Started Ceph object storage daemon osd.0.
Aug 02 10:39:07 osd3 ceph-osd[7413]: starting osd.0 at - osd_data /var/lib/ceph/osd/ceph-0 /var/lib/ceph/osd/ceph-0/journal
Aug 02 10:40:48 osd3 ceph-osd[7413]: 2017-08-02 10:40:48.583063 7f5262cc3e00 -1 osd.0 25924 log_to_monitors {default=true}
Aug 02 10:43:32 osd3 ceph-osd[7413]: *** Caught signal (Aborted) **
Aug 02 10:43:32 osd3 ceph-osd[7413]: in thread 7f524861b700 thread_name:tp_osd_tp
Aug 02 10:43:32 osd3 ceph-osd[7413]: ceph version 12.1.2 (b661348f156f148d764b998b65b90451f096cb27) luminous (rc)
Aug 02 10:43:32 osd3 ceph-osd[7413]: 1: (()+0xa9a964) [0x5623f0a9c964]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 2: (()+0x11390) [0x7f52611a6390]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 3: (pread64()+0x33) [0x7f52611a5d43]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 4: (KernelDevice::direct_read_unaligned(unsigned long, unsigned long, char*)+0x81) [0x5623f0a7cfc1]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 5: (KernelDevice::read_random(unsigned long, unsigned long, char*, bool)+0x4f3) [0x5623f0a7da43]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 6: (BlueFS::_read_random(BlueFS::FileReader*, unsigned long, unsigned long, char*)+0x4fa) [0x5623f0a4d9ca]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 7: (BlueRocksRandomAccessFile::Read(unsigned long, unsigned long, rocksdb::Slice*, char*) const+0x20) [0x5623f0a77e10]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 8: (rocksdb::RandomAccessFileReader::Read(unsigned long, unsigned long, rocksdb::Slice*, char*) const+0xf8f) [0x5623f0e50acf]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 9: (rocksdb::ReadBlockContents(rocksdb::RandomAccessFileReader*, rocksdb::Footer const&, rocksdb::ReadOptions const&, rocksdb::BlockHandle const&, rocksdb::BlockContents*, rocksdb::ImmutableCFOptions const&, bool, rocksdb::Slice const&, rocksdb::PersistentCacheOptions const&)+0x5f3) [0x5623f0e21383]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 10: (()+0xe0f7c6) [0x5623f0e117c6]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 11: (rocksdb::BlockBasedTable::MaybeLoadDataBlockToCache(rocksdb::BlockBasedTable::Rep*, rocksdb::ReadOptions const&, rocksdb::BlockHandle const&, rocksdb::Slice, rocksdb::BlockBasedTable::CachableEntry<rocksdb::Block>*, bool)+0x2f8) [0x5623f0e13928]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 12: (rocksdb::BlockBasedTable::NewDataBlockIterator(rocksdb::BlockBasedTable::Rep*, rocksdb::ReadOptions const&, rocksdb::BlockHandle const&, rocksdb::BlockIter*, bool, rocksdb::Status)+0x2ac) [0x5623f0e13d2c]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 13: (rocksdb::BlockBasedTable::BlockEntryIteratorState::NewSecondaryIterator(rocksdb::Slice const&)+0x97) [0x5623f0e1c4a7]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 14: (()+0xe4576e) [0x5623f0e4776e]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 15: (()+0xe45836) [0x5623f0e47836]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 16: (()+0xe459b1) [0x5623f0e479b1]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 17: (rocksdb::MergingIterator::Next()+0x449) [0x5623f0e2ab09]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 18: (rocksdb::DBIter::FindNextUserEntryInternal(bool, bool)+0x182) [0x5623f0ec7ed2]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 19: (rocksdb::DBIter::Next()+0x1eb) [0x5623f0ec8c8b]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 20: (RocksDBStore::RocksDBWholeSpaceIteratorImpl::next()+0x9a) [0x5623f09dd58a]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 21: (BlueStore::_collection_list(BlueStore::Collection*, ghobject_t const&, ghobject_t const&, int, std::vector<ghobject_t, std::allocator<ghobject_t> >*, ghobject_t*)+0x1170) [0x5623f093d250]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 22: (BlueStore::collection_list(boost::intrusive_ptr<ObjectStore::CollectionImpl>&, ghobject_t const&, ghobject_t const&, int, std::vector<ghobject_t, std::allocator<ghobject_t> >*, ghobject_t*)+0x25a) [0x5623f093e6ea]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 23: (PGBackend::objects_list_range(hobject_t const&, hobject_t const&, snapid_t, std::vector<hobject_t, std::allocator<hobject_t> >*, std::vector<ghobject_t, std::allocator<ghobject_t> >*)+0x192) [0x5623f0700ef2]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 24: (PG::build_scrub_map_chunk(ScrubMap&, hobject_t, hobject_t, bool, unsigned int, ThreadPool::TPHandle&)+0x200) [0x5623f05a8b30]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 25: (PG::chunky_scrub(ThreadPool::TPHandle&)+0x3ea) [0x5623f05d61ca]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 26: (PG::scrub(unsigned int, ThreadPool::TPHandle&)+0x45c) [0x5623f05d7cec]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 27: (OSD::ShardedOpWQ::_process(unsigned int, ceph::heartbeat_handle_d*)+0x12d0) [0x5623f05179e0]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 28: (ShardedThreadPool::shardedthreadpool_worker(unsigned int)+0x884) [0x5623f0ae44e4]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 29: (ShardedThreadPool::WorkThreadSharded::entry()+0x10) [0x5623f0ae7520]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 30: (()+0x76ba) [0x7f526119c6ba]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 31: (clone()+0x6d) [0x7f52602133dd]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 2017-08-02 10:43:32.251841 7f524861b700 -1 *** Caught signal (Aborted) **
Aug 02 10:43:32 osd3 ceph-osd[7413]: in thread 7f524861b700 thread_name:tp_osd_tp
Aug 02 10:43:32 osd3 ceph-osd[7413]: ceph version 12.1.2 (b661348f156f148d764b998b65b90451f096cb27) luminous (rc)
Aug 02 10:43:32 osd3 ceph-osd[7413]: 1: (()+0xa9a964) [0x5623f0a9c964]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 2: (()+0x11390) [0x7f52611a6390]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 3: (pread64()+0x33) [0x7f52611a5d43]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 4: (KernelDevice::direct_read_unaligned(unsigned long, unsigned long, char*)+0x81) [0x5623f0a7cfc1]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 5: (KernelDevice::read_random(unsigned long, unsigned long, char*, bool)+0x4f3) [0x5623f0a7da43]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 6: (BlueFS::_read_random(BlueFS::FileReader*, unsigned long, unsigned long, char*)+0x4fa) [0x5623f0a4d9ca]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 7: (BlueRocksRandomAccessFile::Read(unsigned long, unsigned long, rocksdb::Slice*, char*) const+0x20) [0x5623f0a77e10]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 8: (rocksdb::RandomAccessFileReader::Read(unsigned long, unsigned long, rocksdb::Slice*, char*) const+0xf8f) [0x5623f0e50acf]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 9: (rocksdb::ReadBlockContents(rocksdb::RandomAccessFileReader*, rocksdb::Footer const&, rocksdb::ReadOptions const&, rocksdb::BlockHandle const&, rocksdb::BlockContents*, rocksdb::ImmutableCFOptions const&, bool, rocksdb::Slice const&, rocksdb::PersistentCacheOptions const&)+0x5f3) [0x5623f0e21383]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 10: (()+0xe0f7c6) [0x5623f0e117c6]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 11: (rocksdb::BlockBasedTable::MaybeLoadDataBlockToCache(rocksdb::BlockBasedTable::Rep*, rocksdb::ReadOptions const&, rocksdb::BlockHandle const&, rocksdb::Slice, rocksdb::BlockBasedTable::CachableEntry<rocksdb::Block>*, bool)+0x2f8) [0x5623f0e13928]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 12: (rocksdb::BlockBasedTable::NewDataBlockIterator(rocksdb::BlockBasedTable::Rep*, rocksdb::ReadOptions const&, rocksdb::BlockHandle const&, rocksdb::BlockIter*, bool, rocksdb::Status)+0x2ac) [0x5623f0e13d2c]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 13: (rocksdb::BlockBasedTable::BlockEntryIteratorState::NewSecondaryIterator(rocksdb::Slice const&)+0x97) [0x5623f0e1c4a7]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 14: (()+0xe4576e) [0x5623f0e4776e]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 15: (()+0xe45836) [0x5623f0e47836]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 16: (()+0xe459b1) [0x5623f0e479b1]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 17: (rocksdb::MergingIterator::Next()+0x449) [0x5623f0e2ab09]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 18: (rocksdb::DBIter::FindNextUserEntryInternal(bool, bool)+0x182) [0x5623f0ec7ed2]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 19: (rocksdb::DBIter::Next()+0x1eb) [0x5623f0ec8c8b]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 20: (RocksDBStore::RocksDBWholeSpaceIteratorImpl::next()+0x9a) [0x5623f09dd58a]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 21: (BlueStore::_collection_list(BlueStore::Collection*, ghobject_t const&, ghobject_t const&, int, std::vector<ghobject_t, std::allocator<ghobject_t> >*, ghobject_t*)+0x1170) [0x5623f093d250]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 22: (BlueStore::collection_list(boost::intrusive_ptr<ObjectStore::CollectionImpl>&, ghobject_t const&, ghobject_t const&, int, std::vector<ghobject_t, std::allocator<ghobject_t> >*, ghobject_t*)+0x25a) [0x5623f093e6ea]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 23: (PGBackend::objects_list_range(hobject_t const&, hobject_t const&, snapid_t, std::vector<hobject_t, std::allocator<hobject_t> >*, std::vector<ghobject_t, std::allocator<ghobject_t> >*)+0x192) [0x5623f0700ef2]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 24: (PG::build_scrub_map_chunk(ScrubMap&, hobject_t, hobject_t, bool, unsigned int, ThreadPool::TPHandle&)+0x200) [0x5623f05a8b30]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 25: (PG::chunky_scrub(ThreadPool::TPHandle&)+0x3ea) [0x5623f05d61ca]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 26: (PG::scrub(unsigned int, ThreadPool::TPHandle&)+0x45c) [0x5623f05d7cec]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 27: (OSD::ShardedOpWQ::_process(unsigned int, ceph::heartbeat_handle_d*)+0x12d0) [0x5623f05179e0]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 28: (ShardedThreadPool::shardedthreadpool_worker(unsigned int)+0x884) [0x5623f0ae44e4]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 29: (ShardedThreadPool::WorkThreadSharded::entry()+0x10) [0x5623f0ae7520]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 30: (()+0x76ba) [0x7f526119c6ba]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 31: (clone()+0x6d) [0x7f52602133dd]
Aug 02 10:43:32 osd3 ceph-osd[7413]: NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed to interpret this.
Aug 02 10:43:32 osd3 ceph-osd[7413]: 0> 2017-08-02 10:43:32.251841 7f524861b700 -1 *** Caught signal (Aborted) **
Aug 02 10:43:32 osd3 ceph-osd[7413]: in thread 7f524861b700 thread_name:tp_osd_tp
Aug 02 10:43:32 osd3 ceph-osd[7413]: ceph version 12.1.2 (b661348f156f148d764b998b65b90451f096cb27) luminous (rc)
Aug 02 10:43:32 osd3 ceph-osd[7413]: 1: (()+0xa9a964) [0x5623f0a9c964]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 2: (()+0x11390) [0x7f52611a6390]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 3: (pread64()+0x33) [0x7f52611a5d43]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 4: (KernelDevice::direct_read_unaligned(unsigned long, unsigned long, char*)+0x81) [0x5623f0a7cfc1]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 5: (KernelDevice::read_random(unsigned long, unsigned long, char*, bool)+0x4f3) [0x5623f0a7da43]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 6: (BlueFS::_read_random(BlueFS::FileReader*, unsigned long, unsigned long, char*)+0x4fa) [0x5623f0a4d9ca]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 7: (BlueRocksRandomAccessFile::Read(unsigned long, unsigned long, rocksdb::Slice*, char*) const+0x20) [0x5623f0a77e10]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 8: (rocksdb::RandomAccessFileReader::Read(unsigned long, unsigned long, rocksdb::Slice*, char*) const+0xf8f) [0x5623f0e50acf]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 9: (rocksdb::ReadBlockContents(rocksdb::RandomAccessFileReader*, rocksdb::Footer const&, rocksdb::ReadOptions const&, rocksdb::BlockHandle const&, rocksdb::BlockContents*, rocksdb::ImmutableCFOptions const&, bool, rocksdb::Slice const&, rocksdb::PersistentCacheOptions const&)+0x5f3) [0x5623f0e21383]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 10: (()+0xe0f7c6) [0x5623f0e117c6]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 11: (rocksdb::BlockBasedTable::MaybeLoadDataBlockToCache(rocksdb::BlockBasedTable::Rep*, rocksdb::ReadOptions const&, rocksdb::BlockHandle const&, rocksdb::Slice, rocksdb::BlockBasedTable::CachableEntry<rocksdb::Block>*, bool)+0x2f8) [0x5623f0e13928]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 12: (rocksdb::BlockBasedTable::NewDataBlockIterator(rocksdb::BlockBasedTable::Rep*, rocksdb::ReadOptions const&, rocksdb::BlockHandle const&, rocksdb::BlockIter*, bool, rocksdb::Status)+0x2ac) [0x5623f0e13d2c]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 13: (rocksdb::BlockBasedTable::BlockEntryIteratorState::NewSecondaryIterator(rocksdb::Slice const&)+0x97) [0x5623f0e1c4a7]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 14: (()+0xe4576e) [0x5623f0e4776e]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 15: (()+0xe45836) [0x5623f0e47836]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 16: (()+0xe459b1) [0x5623f0e479b1]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 17: (rocksdb::MergingIterator::Next()+0x449) [0x5623f0e2ab09]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 18: (rocksdb::DBIter::FindNextUserEntryInternal(bool, bool)+0x182) [0x5623f0ec7ed2]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 19: (rocksdb::DBIter::Next()+0x1eb) [0x5623f0ec8c8b]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 20: (RocksDBStore::RocksDBWholeSpaceIteratorImpl::next()+0x9a) [0x5623f09dd58a]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 21: (BlueStore::_collection_list(BlueStore::Collection*, ghobject_t const&, ghobject_t const&, int, std::vector<ghobject_t, std::allocator<ghobject_t> >*, ghobject_t*)+0x1170) [0x5623f093d250]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 22: (BlueStore::collection_list(boost::intrusive_ptr<ObjectStore::CollectionImpl>&, ghobject_t const&, ghobject_t const&, int, std::vector<ghobject_t, std::allocator<ghobject_t> >*, ghobject_t*)+0x25a) [0x5623f093e6ea]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 23: (PGBackend::objects_list_range(hobject_t const&, hobject_t const&, snapid_t, std::vector<hobject_t, std::allocator<hobject_t> >*, std::vector<ghobject_t, std::allocator<ghobject_t> >*)+0x192) [0x5623f0700ef2]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 24: (PG::build_scrub_map_chunk(ScrubMap&, hobject_t, hobject_t, bool, unsigned int, ThreadPool::TPHandle&)+0x200) [0x5623f05a8b30]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 25: (PG::chunky_scrub(ThreadPool::TPHandle&)+0x3ea) [0x5623f05d61ca]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 26: (PG::scrub(unsigned int, ThreadPool::TPHandle&)+0x45c) [0x5623f05d7cec]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 27: (OSD::ShardedOpWQ::_process(unsigned int, ceph::heartbeat_handle_d*)+0x12d0) [0x5623f05179e0]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 28: (ShardedThreadPool::shardedthreadpool_worker(unsigned int)+0x884) [0x5623f0ae44e4]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 29: (ShardedThreadPool::WorkThreadSharded::entry()+0x10) [0x5623f0ae7520]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 30: (()+0x76ba) [0x7f526119c6ba]
Aug 02 10:43:32 osd3 ceph-osd[7413]: 31: (clone()+0x6d) [0x7f52602133dd]
Aug 02 10:43:32 osd3 ceph-osd[7413]: NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed to interpret this.
Aug 02 10:43:32 osd3 systemd[1]: ceph-osd@0.service: Main process exited, code=killed, status=6/ABRT
Aug 02 10:43:32 osd3 systemd[1]: ceph-osd@0.service: Unit entered failed state.
Aug 02 10:43:32 osd3 systemd[1]: ceph-osd@0.service: Failed with result 'signal'.
Aug 02 10:43:52 osd3 systemd[1]: ceph-osd@0.service: Service hold-off time over, scheduling restart.
Aug 02 10:43:52 osd3 systemd[1]: Stopped Ceph object storage daemon osd.0.
Aug 02 10:43:52 osd3 systemd[1]: Starting Ceph object storage daemon osd.0...
Aug 02 10:43:52 osd3 systemd[1]: Started Ceph object storage daemon osd.0.
Aug 02 10:43:52 osd3 ceph-osd[8322]: starting osd.0 at - osd_data /var/lib/ceph/osd/ceph-0 /var/lib/ceph/osd/ceph-0/journal
roger@desktop:~$ ceph -s
cluster:
id: eea7b78c-b138-40fc-9f3e-3d77afb770f0
health: HEALTH_WARN
1 osds down
1 host (1 osds) down
Degraded data redundancy: 43922/162834 objects degraded (26.973%), 300 pgs unclean, 305 pgs degraded
114 pgs not deep-scrubbed for 86400
155 pgs not scrubbed for 86400
10 slow requests are blocked > 32 sec
services:
mon: 3 daemons, quorum nuc1,nuc2,nuc3
mgr: nuc3(active), standbys: nuc2, nuc1
osd: 3 osds: 2 up, 3 in
rgw: 1 daemon active
data:
pools: 19 pools, 372 pgs
objects: 54278 objects, 71724 MB
usage: 122 GB used, 27819 GB / 27941 GB avail
pgs: 43922/162834 objects degraded (26.973%)
303 active+undersized+degraded
67 active+clean
2 active+recovery_wait+degraded
roger@desktop:~$ ceph osd tree
ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF
-1 27.28679 root default
-5 9.09560 host osd1
3 hdd 9.09560 osd.3 up 1.00000 1.00000
-6 9.09560 host osd2
4 hdd 9.09560 osd.4 up 1.00000 1.00000
-2 9.09560 host osd3
0 hdd 9.09560 osd.0 down 1.00000 1.00000
roger@desktop:~$ ceph mon versions
{
"ceph version 12.1.2 (b661348f156f148d764b998b65b90451f096cb27) luminous (rc)": 3
}
roger@desktop:~$ ceph osd versions
{
"ceph version 12.1.2 (b661348f156f148d764b998b65b90451f096cb27) luminous (rc)": 2
}
roger@osd3:~$ sudo ceph daemon osd.0 status
{
"cluster_fsid": "eea7b78c-b138-40fc-9f3e-3d77afb770f0",
"osd_fsid": "bdb31a03-e381-4bf8-82e3-18916c838308",
"whoami": 0,
"state": "waiting_for_healthy",
"oldest_map": 25389,
"newest_map": 25938,
"num_pgs": 372
}
roger@desktop:~$ ceph df
GLOBAL:
SIZE AVAIL RAW USED %RAW USED
27941G 27819G 122G 0.44
POOLS:
NAME ID USED %USED MAX AVAIL OBJECTS
default.rgw.rgw.gc 70 0 0 8807G 0
default.rgw.buckets.non-ec 83 0 0 8807G 43
default.rgw.control 85 0 0 8807G 8
default.rgw.data.root 86 15601 0 8807G 49
default.rgw.gc 87 0 0 8807G 32
default.rgw.lc 88 0 0 8807G 32
default.rgw.log 89 0 0 8807G 144
default.rgw.users.uid 90 3346 0 8807G 14
default.rgw.users.email 91 100 0 8807G 7
default.rgw.users.keys 92 100 0 8807G 7
default.rgw.buckets.index 93 0 0 8807G 39
default.rgw.intent-log 95 0 0 8807G 0
default.rgw.meta 96 0 0 8807G 0
default.rgw.usage 97 0 0 8807G 0
default.rgw.users.swift 98 39 0 8807G 4
default.rgw.buckets.extra 99 0 0 8807G 0
.rgw.root 100 1681 0 8807G 4
default.rgw.reshard 101 0 0 8807G 17
default.rgw.buckets.data 103 71724M 0.40 17614G 53878
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com