*databaseS in general (mysql, postgresql, etc.) Frédéric. ----- Le 30 Nov 24, à 15:14, Frédéric Nass frederic.nass@xxxxxxxxxxxxxxxx a écrit : > Hi Martin, > > Thanks for sharing your thoughts. I was wondering how database would handle low > disk space when receiving new requests. Maybe it's not much different than what > you've described below. > > Cheers, > Frédéric. > > ----- Le 29 Nov 24, à 17:46, Martin Konold <martin.konold@xxxxxxxxxx> a écrit : > >> Hi, > >> the traditional solution is to deny anything but deletions and either write the >> transaction log to another device or even Filesystem or add support deletions >> without a transaction log together with a force switch. > >> Regards >> --martin > >> Am 29.11.2024 12:03 schrieb Igor Fedotov <igor.fedotov@xxxxxxxx>: > >>> Hi Frederic, > >>> >My question was more about why bluefs would still fail to allocate 4k >>> chunks after being allowed to do so by >>> https://tracker.ceph.com/issues/53466 (John's case with v17.2.6 actually) > >>> My hypothesis is that it's facing real "no-space" case not the one I >>> explained above. But again - we haven't performed thorough analysis of >>> John's case, so we are just speculating... > >>> >Is BlueFS aware of the remaining space and maybe using some sort of >>> reserved blocks/chunks like other filesystems to handle full/near null >>> situations ? > >>> What remaining space are you talking here - if the one consisting of >>> small (<64K) extents only then BlueFS hopefully uses it since v17.2.6. >>> Reserving more spare space at Bluefs on mkfs? Previously at Bluestore >>> team we had discussions on something like that to permit easier recovery >>> from "no-space" cases. No final solution has been taken yet and in fact >>> this provides no complete fix for the issue anyway - that spare space >>> might end at some point as well... > >>> >If so, then it should never crash, right? >>> What we have here is not a real crash (although looks like that) - it's >>> an expected assertion. We just don't have good enough [automatic] >>> scenario to exit from this state. Hence "crying" about that aloud. > >>> The problem is that by its design RocksDB has to write out some data >>> (either due to the need for internal maintenance or to fulfill client >>> requests) on any update access. So at some point we have no space to >>> record such a transaction. How one can proceed in that case - refuse to >>> execute it and return an error? OK, but what's next? Any followup data >>> removal would need some transactions to be recorded in RocksDB as well. >>> And DB/FS would need more space for that. Use some reserved spare space? >>> But that's conceptually similar to stalling OSD writes at some free >>> space threshold we already have at OSD level - as you can see this >>> protection doesn't work from time to time. >>> So IMO it's better polish existing protection means then. > >>> Not to mention - I'm pretty sure that someone would abuse additional >>> spare space mechanics if any and finally face the same problem at some >>> point. Thus triggering another iteration of it.. ;) > >>> Thanks, >>> Igor > >>> On 29.11.2024 10:20, Frédéric Nass wrote: >>> > Hi Igor, > >>> > Thank you for taking the time to explains the fragmentation issue. I >>> > had figured out the most part of it by reading the tracker and the PR >>> > but it's always clearer when you explain it. > >>> > My question was more about why bluefs would still fail to allocate 4k >>> > chunks after being allowed to do so by >>> > https://tracker.ceph.com/issues/53466 (John's case with v17.2.6 actually) > >>> > Is BlueFS aware of the remaining space and maybe using some sort of >>> > reserved blocks/chunks like other filesystems to handle full/near null >>> > situations ? If so, then it should never crash, right? >>> > Like other filesystems don't crash, drives's firmwares dont crash, etc. > >>> > Thanks, >>> > Frédéric. > >>> > ----- Le 28 Nov 24, à 12:52, Igor Fedotov <igor.fedotov@xxxxxxxx> a >>> > écrit : > >>> > Hi Frederic, > >>> > here is an overview of the case when BlueFS ıs unable to allocate >>> > more space at main/shared device albeıt free space is available. >>> > Below I'm talking about stuff exısted before fıxıng >>> > https://tracker.ceph.com/issues/53466. > >>> > First of al - BlueFS's minimal allocation unit for shared device >>> > was bluefs_shared_alloc_size (=64K by default). Which means that >>> > it was unable to use e.g. 2x32K or 16x4K chunks when it needed >>> > additional 64K bytes. > >>> > Secondly - sometimes RocksDB performs recovery - and some other >>> > maintenance tasks that require space allocation - on startup. >>> > Which evidently triggers allocation of N*64K chunks from shared >>> > device. > >>> > Thirdly - a while ago we switched to 4K chunk allocations for user >>> > data (please not confuse with BlueFS allocation). Which >>> > potentially could result ın specific free space fragmentation >>> > pattern when there ıs limited (or even empty) set of long (>=64K) >>> > chunks free. Still technically having enough free space available. >>> > E.g. free extent list could look like (off~len, both in hex): > >>> > 0x0~1000, 0x2000~1000, 0x4000~2000, 0x10000~4000, 0x2000~1000, etc... > >>> > In that case original BlueFS allocator implementation was unable >>> > to locate more free space which in turn was effectively breaking >>> > both RockDB and OSD boot up. > >>> > One should realize that the above free space fragmentation depends >>> > on a bunch of factors, none of which is absolutely dominating: > >>> > 1. how user write/remove objects > >>> > 2. how allocator seeks for free space > >>> > 3. how much free space is available > >>> > So we don't have full control on 1. and 3. and have limited >>> > opportunities in tuning 2. > >>> > Small device sizes and high space utilization severely increase >>> > the probability for the issue to happen but theoretically even a >>> > large disk with mediocre utilization could reach "bad" state over >>> > time if used (by both clients and allocator) >>> > "improperly/inefficiently". Hence tuning thresholds can reduce the >>> > issue's probability to occur (at cost of additional spare space >>> > waste) but it isn't a silver bullet. > >>> > https://tracker.ceph.com/issues/53466 fixes (or rather works >>> > around) the issue by allowing BlueFS to use 4K extents. Plus we're >>> > working on making better resulting free space fragmentation on >>> > aged OSDs by improving allocation strategies, e.g. see : > >>> > - https://github.com/ceph/ceph/pull/52489 > >>> > - https://github.com/ceph/ceph/pull/57789 > >>> > - https://github.com/ceph/ceph/pull/60870 > >>> > Hope this is helpful. > >>> > Thanks, > >>> > Igor > > >>> > On 27.11.2024 16:31, Frédéric Nass wrote: > > > >>> > ----- Le 27 Nov 24, à 10:19, Igor Fedotov >>> > <igor.fedotov@xxxxxxxx> <mailto:igor.fedotov@xxxxxxxx> a écrit : > >>> > Hi Istvan, > >>> > first of all let me make a remark that we don't know why >>> > BlueStore is out of space at John's cluster. > >>> > It's just an unconfirmed hypothesis from Frederic that >>> > it's caused by high fragmentation and BlueFS'es inability >>> > to use chunks smaller than 64K. In fact fragmentation >>> > issue is fixed since 17.2.6 so I doubt that's the problem. > >>> > Hi Igor, > >>> > I wasn't actually pointing this as the root cause (since >>> > John's already using 17.2.6) but more to explain the context, >>> > but while we're at it... > >>> > Could you elaborate on circumstances that could prevent BlueFS >>> > from being able to allocate chunks in collocated OSDs >>> > scenario? Does this ability depend on near/full thresholds >>> > being reached or not? If so then icreasing these thresholds by >>> > 1-2% may help avoiding the crash, no? > >>> > Also, if BlueFS is aware of these thresholds, shouldn't an >>> > OSDs be able to start and live without crashing even when it's >>> > full and simply (maybe easier said than done...) refuse any >>> > I/Os? Sorry for the noob questions. :-) > >>> > This topic is particularly important when using NVMe drives as >>> > 'collocated' OSDs, expecially since they often host critical >>> > metadata pools (cephfs, rgw index). > >>> > Cheers, >>> > Frédéric. > >>> > Thanks, > >>> > Igor > >>> > On 27.11.2024 4:01, Szabo, Istvan (Agoda) wrote: > >>> > Hi, > >>> > This issue should not happen anymore from 17.2.8 am I >>> > correct? In this version all the fragmentation issue >>> > should have gone even with collocated wal+db+block. > >>> > ------------------------------------------------------------------------ >>> > *From:* Frédéric Nass <frederic.nass@xxxxxxxxxxxxxxxx> >>> > <mailto:frederic.nass@xxxxxxxxxxxxxxxx> >>> > *Sent:* Wednesday, November 27, 2024 6:12:46 AM >>> > *To:* John Jasen <jjasen@xxxxxxxxx> >>> > <mailto:jjasen@xxxxxxxxx> >>> > *Cc:* Igor Fedotov <igor.fedotov@xxxxxxxx> >>> > <mailto:igor.fedotov@xxxxxxxx>; ceph-users >>> > <ceph-users@xxxxxxx> <mailto:ceph-users@xxxxxxx> >>> > *Subject:* Re: down OSDs, Bluestore out >>> > of space, unable to restart > >>> > Email received from the internet. If in doubt, don't >>> > click any link nor open any attachment ! >>> > ________________________________ > >>> > Hi John, > >>> > That's about right. Two potential solutions exist: >>> > 1. Adding a new drive to the server and sharing it for >>> > RocksDB metadata, or >>> > 2. Repurposing one of the failed OSDs for the same >>> > purpose (if adding more drives isn't feasible). > >>> > Igor's post #6 [1] explains the challenges with >>> > co-located OSDs (DB+WAL+data on the same device) when >>> > they run out of space, where significant fragmentation >>> > occurs and BlueFS and BlueStore block sizes are >>> > misaligned. The solution (included in 17.2.6) was to >>> > allow BlueFS to allocate 4k extents when it couldn't >>> > find 64k contiguous extents. However, it seems that >>> > even with this fix, these OSDs still can't boot up. > >>> > Therefore, the recommendation is to extend the RocksDB >>> > volume to another device as a temporary workaround. > >>> > Before proceeding, I recommend checking the failed >>> > OSDs' bluefs_shared_alloc_size value. If it's 64k, you >>> > might want to try lowering this to 32k or even 4k, as >>> > some users reported [2] that reducing this value >>> > helped failed OSDs boot up and remain stable for a >>> > period of time. Might be worth checking and trying. > >>> > Regards, >>> > Frédéric. > >>> > [1] https://tracker.ceph.com/issues/53466#note-6 >>> > [2] >>> > https://github.com/rook/rook/issues/9885#issuecomment-1761076861" >>> > <https://github.com/rook/rook/issues/9885#issuecomment-1761076861"> > > >>> > ________________________________ >>> > De : John Jasen <jjasen@xxxxxxxxx> >>> > <mailto:jjasen@xxxxxxxxx> >>> > Envoyé : mardi 26 novembre 2024 18:50 >>> > À : Igor Fedotov >>> > Cc: ceph-users >>> > Objet : Re: down OSDs, Bluestore out of >>> > space, unable to restart > >>> > Let me see if I have the approach right'ish: > >>> > scrounge some more disk for the servers with full/down >>> > OSDs. >>> > partition the new disks into LVs for each downed OSD. >>> > Attach as a lvm new-db to the downed OSDs. >>> > Restart the OSDs. >>> > Profit. > >>> > Is that about right? > > >>> > On Tue, Nov 26, 2024 at 11:28 AM Igor Fedotov >>> > <igor.fedotov@xxxxxxxx> <mailto:igor.fedotov@xxxxxxxx> >>> > wrote: > >>> > > Well, so there is a single shared volume (disk) per >>> > OSD, right? > >>> > > If so one can add dedicated DB volume to such an OSD >>> > - one done OSD will >>> > > have two underlying devices: main(which is original >>> > shared disk) and new >>> > > dedicated DB ones. And hence this will effectively >>> > provide additional >>> > > space for BlueFS/RocksDB and permit OSD to start up. > >>> > > I'm not aware of all the details how to do that with >>> > cephadm (or whatever >>> > > RH uses) but on bare metal setup this could be >>> > achieved by issuing >>> > > 'ceph-volume lvm new-db' command which will attach >>> > new LV (provided by >>> > > user) to specific OSD. > > >>> > > Thanks, > >>> > > Igor > > >>> > > On 26.11.2024 19:16, John Jasen wrote: > >>> > > They're all bluefs_single_shared_device, if I >>> > understand your question. >>> > > There's no room left on the devices to expand. > >>> > > We started at quincy with this cluster, and didn't >>> > vary too much from the >>> > > Redhat Ceph storage 6 documentation for setting it up. > > >>> > > On Tue, Nov 26, 2024 at 4:48 AM Igor Fedotov >>> > <igor.fedotov@xxxxxxxx> <mailto:igor.fedotov@xxxxxxxx> >>> > > wrote: > >>> > >> Hi John, > >>> > >> you haven't described your OSD volume configuration >>> > but you might want >>> > >> to try adding standalone DB volume if OSD uses LVM >>> > and has single main >>> > >> device only. > >>> > >> 'ceph-volume lvm new-db' command is the preferred >>> > way of doing that, see > >>> > >> https://docs.ceph.com/en/quincy/ceph-volume/lvm/newdb/ > > >>> > >> Thanks, > >>> > >> Igor > >>> > >> On 25.11.2024 21:37, John Jasen wrote: >>> > >> > Ceph version 17.2.6 > >>> > >> > After a power loss event affecting my ceph >>> > cluster, I've been putting >>> > >> > humpty dumpty back together since. > >>> > >> > One problem I face is that with objects degraded, >>> > rebalancing doesn't >>> > >> run >>> > >> > -- and this resulted in several of my fast OSDs >>> > filling up. > >>> > >> > I have 8 OSDs currently down, 100% full >>> > (exceeding all the full ratio >>> > >> > settings on by default or I toggled to try and >>> > keep it together), and >>> > >> when >>> > >> > I try to restart them, they fail out. Is there >>> > any way to bring these >>> > >> back >>> > >> > from the dead? > >>> > >> > Here's some interesting output from journalctl >>> > -xeu on the failed OSD: > >>> > >> > ceph-osd[2383080]: >>> > bluestore::NCB::__restore_allocator::No Valid >>> > >> allocation >>> > >> > info on disk (empty file) >>> > >> > ceph-osd[2383080]: >>> > bluestore(/var/lib/ceph/osd/ceph-242) >>> > >> > _init_alloc::NCB::restore_allocator() failed! Run >>> > Full Recovery from >>> > >> ONodes >>> > >> > (might take a while) ... > >>> > >> > ceph-osd[2389725]: bluefs _allocate allocation >>> > failed, needed 0x3000 > > >>> > ceph-6ab85342-53d6-11ee-88a7-e43d1a153e91-osd-242[2389718]: >>> > -2> >>> > >> > 2024-11-25T18:31:42.070+0000 7f0adfdef540 -1 >>> > bluefs _flush_range_F >>> > >> > allocated: 0x0 offset: 0x0 length: 0x230f >>> > >> > ceph-osd[2389725]: bluefs _flush_range_F >>> > allocated: 0x0 offset: 0x0 >>> > >> length: >>> > >> > 0x230f > >>> > >> > Followed quickly by an abort: > > > >>> > /home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos8/DIST/centos8/MACHINE_SIZE/gigantic/release/17.2.6/rpm/el8/BUILD/ceph-17.2.6/src/os/bluestore/BlueFS.cc: >>> > >> > In funct> > > > >>> > /home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos8/DIST/centos8/MACHINE_SIZE/gigantic/release/17.2.6/rpm/el8/BUILD/ceph-17.2.6/src/os/bluestore/BlueFS.cc: >>> > >> > 3380: ce> > >>> > >> > ceph >>> > >> version >>> > >> > 17.2.6 (d7ff0d10654d2280e08f1ab989c7cdf3064446a5) >>> > quincy (stable) >>> > >> > 1: >>> > >> > (ceph::__ceph_abort(char const*, int, char const*, >>> > >> > std::__cxx11::basic_string<char, >>> > std::char_traits<char>, >>> > >> > std::allocator<char> > const&)+0xd7) [0x559bf4361d2f] >>> > >> > 2: >>> > >> > (BlueFS::_flush_range_F(BlueFS::FileWriter*, >>> > unsigned long, unsigned >>> > >> > long)+0x7a9) [0x559bf4b225f9] >>> > >> > 3: >>> > >> > (BlueFS::_flush_F(BlueFS::FileWriter*, bool, >>> > bool*)+0xa2) >>> > >> [0x559bf4b22812] >>> > >> > 4: >>> > >> > (BlueFS::fsync(BlueFS::FileWriter*)+0x8e) >>> > [0x559bf4b40c3e] >>> > >> > 5: >>> > >> > (BlueRocksWritableFile::Sync()+0x19) [0x559bf4b51ed9] >>> > >> > 6: > >>> > (rocksdb::LegacyWritableFileWrapper::Sync(rocksdb::IOOptions >>> > const&, >>> > >> > rocksdb::IODebugContext*)+0x22) [0x559bf507fbd2] >>> > >> > 7: > >>> > (rocksdb::WritableFileWriter::SyncInternal(bool)+0x5aa) >>> > [0x559bf51a880a] >>> > >> > 8: >>> > >> > (rocksdb::WritableFileWriter::Sync(bool)+0x100) >>> > [0x559bf51aa0a0] >>> > >> > 9: >>> > >> > (rocksdb::SyncManifest(rocksdb::Env*, >>> > rocksdb::ImmutableDBOptions >>> > >> const*, >>> > >> > rocksdb::WritableFileWriter*)+0x10b) [0x559bf51a3bfb] >>> > >> > 10: > > >>> > (rocksdb::VersionSet::ProcessManifestWrites(std::deque<rocksdb::VersionSet::ManifestWriter, > >>> > std::allocator<rocksdb::VersionSet::ManifestWriter> >&, >>> > >> > rocksdb::InstrumentedMutex*, >>> > rocksdb::FSDirectory*, bool, rocks> >>> > >> > 11: > > >>> > (rocksdb::VersionSet::LogAndApply(rocksdb::autovector<rocksdb::ColumnFamilyData*, >>> > >> > 8ul> const&, >>> > rocksdb::autovector<rocksdb::MutableCFOptions const*, 8ul> >>> > >> > const&, >>> > rocksdb::autovector<rocksdb::autovector<rocksdb::> >>> > >> > 12: > >>> > (rocksdb::VersionSet::LogAndApply(rocksdb::ColumnFamilyData*, >>> > >> > rocksdb::MutableCFOptions const&, >>> > rocksdb::VersionEdit*, >>> > >> > rocksdb::InstrumentedMutex*, >>> > rocksdb::FSDirectory*, bool, >>> > >> > rocksdb::ColumnFamilyOptions const> >>> > >> > 13: > >>> > (rocksdb::DBImpl::DeleteUnreferencedSstFiles()+0xa30) >>> > [0x559bf50bd250] >>> > >> > 14: > >>> > (rocksdb::DBImpl::Recover(std::vector<rocksdb::ColumnFamilyDescriptor, >>> > >> > std::allocator<rocksdb::ColumnFamilyDescriptor> > >>> > const&, bool, bool, >>> > >> bool, >>> > >> > unsigned long*)+0x13f1) [0x559bf50d3f21] >>> > >> > 15: >>> > >> > (rocksdb::DBImpl::Open(rocksdb::DBOptions const&, >>> > >> > std::__cxx11::basic_string<char, >>> > std::char_traits<char>, >>> > >> > std::allocator<char> > const&, >>> > >> std::vector<rocksdb::ColumnFamilyDescriptor, >>> > >> > std::allocator<rocksdb::Colu> >>> > >> > 16: >>> > >> > (rocksdb::DB::Open(rocksdb::DBOptions const&, >>> > >> > std::__cxx11::basic_string<char, >>> > std::char_traits<char>, >>> > >> > std::allocator<char> > const&, >>> > >> std::vector<rocksdb::ColumnFamilyDescriptor, >>> > >> > std::allocator<rocksdb::ColumnFa> >>> > >> > 17: >>> > >> > (RocksDBStore::do_open(std::ostream&, bool, bool, >>> > >> > std::__cxx11::basic_string<char, >>> > std::char_traits<char>, >>> > >> > std::allocator<char> > const&)+0x77a) >>> > [0x559bf503766a] >>> > >> > 18: >>> > >> > (BlueStore::_open_db(bool, bool, bool)+0xbb4) >>> > [0x559bf4a4bff4] >>> > >> > 19: >>> > >> > (BlueStore::_open_db_and_around(bool, >>> > bool)+0x500) [0x559bf4a766e0] >>> > >> > 20: >>> > >> > (BlueStore::_mount()+0x396) [0x559bf4a795d6] >>> > >> > 21: >>> > >> > (OSD::init()+0x556) [0x559bf44a0eb6] >>> > >> > 22: main() >>> > >> > 23: >>> > >> > __libc_start_main() >>> > >> > 24: >>> > >> _start() > >>> > >> > *** Caught signal (Aborted) ** >>> > >> > in thread >>> > >> > 7f0adfdef540 thread_name:ceph-osd > >>> > >> > ceph >>> > >> version >>> > >> > 17.2.6 (d7ff0d10654d2280e08f1ab989c7cdf3064446a5) >>> > quincy (stable) >>> > >> > 1: >>> > >> > /lib64/libpthread.so.0(+0x12cf0) [0x7f0addff1cf0] >>> > >> > 2: >>> > >> gsignal() >>> > >> > 3: abort() >>> > >> > 4: >>> > >> > (ceph::__ceph_abort(char const*, int, char const*, >>> > >> > std::__cxx11::basic_string<char, >>> > std::char_traits<char>, >>> > >> > std::allocator<char> > const&)+0x197) >>> > [0x559bf4361def] >>> > >> > 5: >>> > >> > (BlueFS::_flush_range_F(BlueFS::FileWriter*, >>> > unsigned long, unsigned >>> > >> > long)+0x7a9) [0x559bf4b225f9] >>> > >> > 6: >>> > >> > (BlueFS::_flush_F(BlueFS::FileWriter*, bool, >>> > bool*)+0xa2) >>> > >> [0x559bf4b22812] >>> > >> > 7: >>> > >> > (BlueFS::fsync(BlueFS::FileWriter*)+0x8e) >>> > [0x559bf4b40c3e] >>> > >> > 8: >>> > >> > (BlueRocksWritableFile::Sync()+0x19) [0x559bf4b51ed9] >>> > >> > 9: > >>> > (rocksdb::LegacyWritableFileWrapper::Sync(rocksdb::IOOptions >>> > const&, >>> > >> > rocksdb::IODebugContext*)+0x22) [0x559bf507fbd2] >>> > >> > 10: > >>> > (rocksdb::WritableFileWriter::SyncInternal(bool)+0x5aa) >>> > [0x559bf51a880a] >>> > >> > 11: >>> > >> > (rocksdb::WritableFileWriter::Sync(bool)+0x100) >>> > [0x559bf51aa0a0] >>> > >> > 12: >>> > >> > (rocksdb::SyncManifest(rocksdb::Env*, >>> > rocksdb::ImmutableDBOptions >>> > >> const*, >>> > >> > rocksdb::WritableFileWriter*)+0x10b) [0x559bf51a3bfb] >>> > >> > 13: > > >>> > (rocksdb::VersionSet::ProcessManifestWrites(std::deque<rocksdb::VersionSet::ManifestWriter, > >>> > std::allocator<rocksdb::VersionSet::ManifestWriter> >&, >>> > >> > rocksdb::InstrumentedMutex*, >>> > rocksdb::FSDirectory*, bool, rocks> >>> > >> > 14: > > >>> > (rocksdb::VersionSet::LogAndApply(rocksdb::autovector<rocksdb::ColumnFamilyData*, >>> > >> > 8ul> const&, >>> > rocksdb::autovector<rocksdb::MutableCFOptions const*, 8ul> >>> > >> > const&, >>> > rocksdb::autovector<rocksdb::autovector<rocksdb::> >>> > >> > 15: > >>> > (rocksdb::VersionSet::LogAndApply(rocksdb::ColumnFamilyData*, >>> > >> > rocksdb::MutableCFOptions const&, >>> > rocksdb::VersionEdit*, >>> > >> > rocksdb::InstrumentedMutex*, >>> > rocksdb::FSDirectory*, bool, >>> > >> > rocksdb::ColumnFamilyOptions const> >>> > >> > 16: > >>> > (rocksdb::DBImpl::DeleteUnreferencedSstFiles()+0xa30) >>> > [0x559bf50bd250] >>> > >> > 17: > >>> > (rocksdb::DBImpl::Recover(std::vector<rocksdb::ColumnFamilyDescriptor, >>> > >> > std::allocator<rocksdb::ColumnFamilyDescriptor> > >>> > const&, bool, bool, >>> > >> bool, >>> > >> > unsigned long*)+0x13f1) [0x559bf50d3f21] >>> > >> > 18: >>> > >> > (rocksdb::DBImpl::Open(rocksdb::DBOptions const&, >>> > >> > std::__cxx11::basic_string<char, >>> > std::char_traits<char>, >>> > >> > std::allocator<char> > const&, >>> > >> std::vector<rocksdb::ColumnFamilyDescriptor, >>> > >> > std::allocator<rocksdb::Colu> >>> > >> > 19: >>> > >> > (rocksdb::DB::Open(rocksdb::DBOptions const&, >>> > >> > std::__cxx11::basic_string<char, >>> > std::char_traits<char>, >>> > >> > std::allocator<char> > const&, >>> > >> std::vector<rocksdb::ColumnFamilyDescriptor, >>> > >> > std::allocator<rocksdb::ColumnFa> >>> > >> > 20: >>> > >> > (RocksDBStore::do_open(std::ostream&, bool, bool, >>> > >> > std::__cxx11::basic_string<char, >>> > std::char_traits<char>, >>> > >> > std::allocator<char> > const&)+0x77a) >>> > [0x559bf503766a] >>> > >> > 21: >>> > >> > (BlueStore::_open_db(bool, bool, bool)+0xbb4) >>> > [0x559bf4a4bff4] >>> > >> > 22: >>> > >> > (BlueStore::_open_db_and_around(bool, >>> > bool)+0x500) [0x559bf4a766e0] >>> > >> > 23: >>> > >> > (BlueStore::_mount()+0x396) [0x559bf4a795d6] >>> > >> > 24: >>> > >> > (OSD::init()+0x556) [0x559bf44a0eb6] >>> > >> > 25: main() >>> > >> > 26: >>> > >> > __libc_start_main() >>> > >> > 27: >>> > >> _start() >>> > >> > NOTE: a >>> > >> copy >>> > >> > of the executable, or `objdump -rdS <executable>` >>> > is needed to interpret >>> > >> > this. >>> > >> > _______________________________________________ >>> > >> > ceph-users mailing list -- ceph-users@xxxxxxx >>> > >> > To unsubscribe send an email to >>> > ceph-users-leave@xxxxxxx > > >>> > _______________________________________________ >>> > ceph-users mailing list -- ceph-users@xxxxxxx >>> > To unsubscribe send an email to ceph-users-leave@xxxxxxx >>> > _______________________________________________ >>> > ceph-users mailing list -- ceph-users@xxxxxxx >>> > To unsubscribe send an email to ceph-users-leave@xxxxxxx > > >>> > ------------------------------------------------------------------------ > > >>> > This message is confidential and is for the >>> > sole use of the intended recipient(s). It >>> > may also be privileged or otherwise >>> > protected by copyright or other legal rules. >>> > If you have received it by mistake please >>> > let us know by reply email and delete it >>> > from your system. It is prohibited to copy >>> > this message or disclose its content to >>> > anyone. Any confidentiality or privilege is >>> > not waived or lost by any mistaken delivery >>> > or unauthorized disclosure of the message. >>> > All messages sent to and from Agoda may be >>> > monitored to ensure compliance with company >>> > policies, to protect the company's interests >>> > and to remove potential malware. Electronic >>> > messages may be intercepted, amended, lost >>> > or deleted, or contain viruses. > > > >>> _______________________________________________ >>> ceph-users mailing list -- ceph-users@xxxxxxx >>> To unsubscribe send an email to ceph-users-leave@xxxxxxx > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx