Re: down OSDs, Bluestore out of space, unable to restart

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Martin, 

Thanks for sharing your thoughts. I was wondering how database would handle low disk space when receiving new requests. Maybe it's not much different than what you've described below. 

Cheers, 
Frédéric. 

----- Le 29 Nov 24, à 17:46, Martin Konold <martin.konold@xxxxxxxxxx> a écrit : 

> Hi,

> the traditional solution is to deny anything but deletions and either write the
> transaction log to another device or even Filesystem or add support deletions
> without a transaction log together with a force switch.

> Regards
> --martin

> Am 29.11.2024 12:03 schrieb Igor Fedotov <igor.fedotov@xxxxxxxx>:

>> Hi Frederic,

>> >My question was more about why bluefs would still fail to allocate 4k
>> chunks after being allowed to do so by
>> https://tracker.ceph.com/issues/53466 (John's case with v17.2.6 actually)

>> My hypothesis is that it's facing real "no-space" case not the one I
>> explained above. But again - we haven't performed thorough analysis of
>> John's case, so we are just speculating...

>> >Is BlueFS aware of the remaining space and maybe using some sort of
>> reserved blocks/chunks like other filesystems to handle full/near null
>> situations ?

>> What remaining space are you talking here - if the one consisting of
>> small (<64K) extents only then BlueFS hopefully uses it since v17.2.6.
>> Reserving more spare space at Bluefs on mkfs? Previously at Bluestore
>> team we had discussions on something like that to permit easier recovery
>> from "no-space" cases. No final solution has been taken yet and in fact
>> this provides no complete fix for the issue anyway - that spare space
>> might end at some point as well...

>> >If so, then it should never crash, right?
>> What we have here is not a real crash (although looks like that) - it's
>> an expected assertion. We just don't have good enough [automatic]
>> scenario to exit from this state. Hence "crying" about that aloud.

>> The problem is that by its design RocksDB has to write out some data
>> (either due to the need for internal maintenance or to fulfill client
>> requests) on any update access. So at some point we have no space to
>> record such a transaction. How one can proceed in that case - refuse to
>> execute it and return an error? OK, but what's next? Any followup data
>> removal would need some transactions to be recorded in RocksDB as well.
>> And DB/FS would need more space for that. Use some reserved spare space?
>> But that's conceptually similar to stalling OSD writes at some free
>> space threshold we already have at OSD level - as you can see this
>> protection doesn't work from time to time.
>> So IMO it's better polish existing protection means then.

>> Not to mention - I'm pretty sure that someone would abuse additional
>> spare space mechanics if any and finally face the same problem at some
>> point. Thus triggering another iteration of it.. ;)

>> Thanks,
>> Igor

>> On 29.11.2024 10:20, Frédéric Nass wrote:
>> > Hi Igor,

>> > Thank you for taking the time to explains the fragmentation issue. I
>> > had figured out the most part of it by reading the tracker and the PR
>> > but it's always clearer when you explain it.

>> > My question was more about why bluefs would still fail to allocate 4k
>> > chunks after being allowed to do so by
>> > https://tracker.ceph.com/issues/53466 (John's case with v17.2.6 actually)

>> > Is BlueFS aware of the remaining space and maybe using some sort of
>> > reserved blocks/chunks like other filesystems to handle full/near null
>> > situations ? If so, then it should never crash, right?
>> > Like other filesystems don't crash, drives's firmwares dont crash, etc.

>> > Thanks,
>> > Frédéric.

>> > ----- Le 28 Nov 24, à 12:52, Igor Fedotov <igor.fedotov@xxxxxxxx> a
>> > écrit :

>> > Hi Frederic,

>> > here is an overview of the case when BlueFS ıs unable to allocate
>> > more space at main/shared device albeıt free space is available.
>> > Below I'm talking about stuff exısted before fıxıng
>> > https://tracker.ceph.com/issues/53466.

>> > First of al - BlueFS's minimal allocation unit for shared device
>> > was bluefs_shared_alloc_size (=64K by default). Which means that
>> > it was unable to use e.g. 2x32K or 16x4K chunks when it needed
>> > additional 64K bytes.

>> > Secondly - sometimes RocksDB performs recovery - and some other
>> > maintenance tasks that require space allocation - on startup.
>> > Which evidently triggers allocation of N*64K chunks from shared
>> > device.

>> > Thirdly - a while ago we switched to 4K chunk allocations for user
>> > data (please not confuse with BlueFS allocation). Which
>> > potentially could result ın specific free space fragmentation
>> > pattern when there ıs limited (or even empty) set of long (>=64K)
>> > chunks free. Still technically having enough free space available.
>> > E.g. free extent list could look like (off~len, both in hex):

>> > 0x0~1000, 0x2000~1000, 0x4000~2000, 0x10000~4000, 0x2000~1000, etc...

>> > In that case original BlueFS allocator implementation was unable
>> > to locate more free space which in turn was effectively breaking
>> > both RockDB and OSD boot up.

>> > One should realize that the above free space fragmentation depends
>> > on a bunch of factors, none of which is absolutely dominating:

>> > 1. how user write/remove objects

>> > 2. how allocator seeks for free space

>> > 3. how much free space is available

>> > So we don't have full control on 1. and 3. and have limited
>> > opportunities in tuning 2.

>> > Small device sizes and high space utilization severely increase
>> > the probability for the issue to happen but theoretically even a
>> > large disk with mediocre utilization could reach "bad" state over
>> > time if used (by both clients and allocator)
>> > "improperly/inefficiently". Hence tuning thresholds can reduce the
>> > issue's probability to occur (at cost of additional spare space
>> > waste) but it isn't a silver bullet.

>> > https://tracker.ceph.com/issues/53466 fixes (or rather works
>> > around) the issue by allowing BlueFS to use 4K extents. Plus we're
>> > working on making better resulting free space fragmentation on
>> > aged OSDs by improving allocation strategies, e.g. see :

>> > - https://github.com/ceph/ceph/pull/52489

>> > - https://github.com/ceph/ceph/pull/57789

>> > - https://github.com/ceph/ceph/pull/60870

>> > Hope this is helpful.

>> > Thanks,

>> > Igor


>> > On 27.11.2024 16:31, Frédéric Nass wrote:



>> > ----- Le 27 Nov 24, à 10:19, Igor Fedotov
>> > <igor.fedotov@xxxxxxxx> <mailto:igor.fedotov@xxxxxxxx> a écrit :

>> > Hi Istvan,

>> > first of all let me make a remark that we don't know why
>> > BlueStore is out of space at John's cluster.

>> > It's just an unconfirmed hypothesis from Frederic that
>> > it's caused by high fragmentation and BlueFS'es inability
>> > to use chunks smaller than 64K. In fact fragmentation
>> > issue is fixed since 17.2.6 so I doubt that's the problem.

>> > Hi Igor,

>> > I wasn't actually pointing this as the root cause (since
>> > John's already using 17.2.6) but more to explain the context,
>> > but while we're at it...

>> > Could you elaborate on circumstances that could prevent BlueFS
>> > from being able to allocate chunks in collocated OSDs
>> > scenario? Does this ability depend on near/full thresholds
>> > being reached or not? If so then icreasing these thresholds by
>> > 1-2% may help avoiding the crash, no?

>> > Also, if BlueFS is aware of these thresholds, shouldn't an
>> > OSDs be able to start and live without crashing even when it's
>> > full and simply (maybe easier said than done...) refuse any
>> > I/Os? Sorry for the noob questions. :-)

>> > This topic is particularly important when using NVMe drives as
>> > 'collocated' OSDs, expecially since they often host critical
>> > metadata pools (cephfs, rgw index).

>> > Cheers,
>> > Frédéric.

>> > Thanks,

>> > Igor

>> > On 27.11.2024 4:01, Szabo, Istvan (Agoda) wrote:

>> > Hi,

>> > This issue should not happen anymore from 17.2.8 am I
>> > correct? In this version all the fragmentation issue
>> > should have gone even with collocated wal+db+block.

>> > ------------------------------------------------------------------------
>> > *From:* Frédéric Nass <frederic.nass@xxxxxxxxxxxxxxxx>
>> > <mailto:frederic.nass@xxxxxxxxxxxxxxxx>
>> > *Sent:* Wednesday, November 27, 2024 6:12:46 AM
>> > *To:* John Jasen <jjasen@xxxxxxxxx>
>> > <mailto:jjasen@xxxxxxxxx>
>> > *Cc:* Igor Fedotov <igor.fedotov@xxxxxxxx>
>> > <mailto:igor.fedotov@xxxxxxxx>; ceph-users
>> > <ceph-users@xxxxxxx> <mailto:ceph-users@xxxxxxx>
>> > *Subject:*  Re: down OSDs, Bluestore out
>> > of space, unable to restart

>> > Email received from the internet. If in doubt, don't
>> > click any link nor open any attachment !
>> > ________________________________

>> > Hi John,

>> > That's about right. Two potential solutions exist:
>> > 1. Adding a new drive to the server and sharing it for
>> > RocksDB metadata, or
>> > 2. Repurposing one of the failed OSDs for the same
>> > purpose (if adding more drives isn't feasible).

>> > Igor's post #6 [1] explains the challenges with
>> > co-located OSDs (DB+WAL+data on the same device) when
>> > they run out of space, where significant fragmentation
>> > occurs and BlueFS and BlueStore block sizes are
>> > misaligned. The solution (included in 17.2.6) was to
>> > allow BlueFS to allocate 4k extents when it couldn't
>> > find 64k contiguous extents. However, it seems that
>> > even with this fix, these OSDs still can't boot up.

>> > Therefore, the recommendation is to extend the RocksDB
>> > volume to another device as a temporary workaround.

>> > Before proceeding, I recommend checking the failed
>> > OSDs' bluefs_shared_alloc_size value. If it's 64k, you
>> > might want to try lowering this to 32k or even 4k, as
>> > some users reported [2] that reducing this value
>> > helped failed OSDs boot up and remain stable for a
>> > period of time. Might be worth checking and trying.

>> > Regards,
>> > Frédéric.

>> > [1] https://tracker.ceph.com/issues/53466#note-6
>> > [2]
>> > https://github.com/rook/rook/issues/9885#issuecomment-1761076861";
>> > <https://github.com/rook/rook/issues/9885#issuecomment-1761076861";>


>> > ________________________________
>> > De : John Jasen <jjasen@xxxxxxxxx>
>> > <mailto:jjasen@xxxxxxxxx>
>> > Envoyé : mardi 26 novembre 2024 18:50
>> > À : Igor Fedotov
>> > Cc: ceph-users
>> > Objet :  Re: down OSDs, Bluestore out of
>> > space, unable to restart

>> > Let me see if I have the approach right'ish:

>> > scrounge some more disk for the servers with full/down
>> > OSDs.
>> > partition the new disks into LVs for each downed OSD.
>> > Attach as a lvm new-db to the downed OSDs.
>> > Restart the OSDs.
>> > Profit.

>> > Is that about right?


>> > On Tue, Nov 26, 2024 at 11:28 AM Igor Fedotov
>> > <igor.fedotov@xxxxxxxx> <mailto:igor.fedotov@xxxxxxxx>
>> > wrote:

>> > > Well, so there is a single shared volume (disk) per
>> > OSD, right?

>> > > If so one can add dedicated DB volume to such an OSD
>> > - one done OSD will
>> > > have two underlying devices: main(which is original
>> > shared disk) and new
>> > > dedicated DB ones. And hence this will effectively
>> > provide additional
>> > > space for BlueFS/RocksDB and permit OSD to start up.

>> > > I'm not aware of all the details how to do that with
>> > cephadm (or whatever
>> > > RH uses) but on bare metal setup this could be
>> > achieved by issuing
>> > > 'ceph-volume lvm new-db' command which will attach
>> > new LV (provided by
>> > > user) to specific OSD.


>> > > Thanks,

>> > > Igor


>> > > On 26.11.2024 19:16, John Jasen wrote:

>> > > They're all bluefs_single_shared_device, if I
>> > understand your question.
>> > > There's no room left on the devices to expand.

>> > > We started at quincy with this cluster, and didn't
>> > vary too much from the
>> > > Redhat Ceph storage 6 documentation for setting it up.


>> > > On Tue, Nov 26, 2024 at 4:48 AM Igor Fedotov
>> > <igor.fedotov@xxxxxxxx> <mailto:igor.fedotov@xxxxxxxx>
>> > > wrote:

>> > >> Hi John,

>> > >> you haven't described your OSD volume configuration
>> > but you might want
>> > >> to try adding standalone DB volume if OSD uses LVM
>> > and has single main
>> > >> device only.

>> > >> 'ceph-volume lvm new-db' command is the preferred
>> > way of doing that, see

>> > >> https://docs.ceph.com/en/quincy/ceph-volume/lvm/newdb/


>> > >> Thanks,

>> > >> Igor

>> > >> On 25.11.2024 21:37, John Jasen wrote:
>> > >> > Ceph version 17.2.6

>> > >> > After a power loss event affecting my ceph
>> > cluster, I've been putting
>> > >> > humpty dumpty back together since.

>> > >> > One problem I face is that with objects degraded,
>> > rebalancing doesn't
>> > >> run
>> > >> > -- and this resulted in several of my fast OSDs
>> > filling up.

>> > >> > I have 8 OSDs currently down, 100% full
>> > (exceeding all the full ratio
>> > >> > settings on by default or I toggled to try and
>> > keep it together), and
>> > >> when
>> > >> > I try to restart them, they fail out. Is there
>> > any way to bring these
>> > >> back
>> > >> > from the dead?

>> > >> > Here's some interesting output from journalctl
>> > -xeu on the failed OSD:

>> > >> > ceph-osd[2383080]:
>> > bluestore::NCB::__restore_allocator::No Valid
>> > >> allocation
>> > >> > info on disk (empty file)
>> > >> > ceph-osd[2383080]:
>> > bluestore(/var/lib/ceph/osd/ceph-242)
>> > >> > _init_alloc::NCB::restore_allocator() failed! Run
>> > Full Recovery from
>> > >> ONodes
>> > >> > (might take a while) ...

>> > >> > ceph-osd[2389725]: bluefs _allocate allocation
>> > failed, needed 0x3000


>> > ceph-6ab85342-53d6-11ee-88a7-e43d1a153e91-osd-242[2389718]:
>> > -2>
>> > >> > 2024-11-25T18:31:42.070+0000 7f0adfdef540 -1
>> > bluefs _flush_range_F
>> > >> > allocated: 0x0 offset: 0x0 length: 0x230f
>> > >> > ceph-osd[2389725]: bluefs _flush_range_F
>> > allocated: 0x0 offset: 0x0
>> > >> length:
>> > >> > 0x230f

>> > >> > Followed quickly by an abort:



>> > /home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos8/DIST/centos8/MACHINE_SIZE/gigantic/release/17.2.6/rpm/el8/BUILD/ceph-17.2.6/src/os/bluestore/BlueFS.cc:
>> > >> > In funct>



>> > /home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos8/DIST/centos8/MACHINE_SIZE/gigantic/release/17.2.6/rpm/el8/BUILD/ceph-17.2.6/src/os/bluestore/BlueFS.cc:
>> > >> > 3380: ce>

>> > >> > ceph
>> > >> version
>> > >> > 17.2.6 (d7ff0d10654d2280e08f1ab989c7cdf3064446a5)
>> > quincy (stable)
>> > >> > 1:
>> > >> > (ceph::__ceph_abort(char const*, int, char const*,
>> > >> > std::__cxx11::basic_string<char,
>> > std::char_traits<char>,
>> > >> > std::allocator<char> > const&)+0xd7) [0x559bf4361d2f]
>> > >> > 2:
>> > >> > (BlueFS::_flush_range_F(BlueFS::FileWriter*,
>> > unsigned long, unsigned
>> > >> > long)+0x7a9) [0x559bf4b225f9]
>> > >> > 3:
>> > >> > (BlueFS::_flush_F(BlueFS::FileWriter*, bool,
>> > bool*)+0xa2)
>> > >> [0x559bf4b22812]
>> > >> > 4:
>> > >> > (BlueFS::fsync(BlueFS::FileWriter*)+0x8e)
>> > [0x559bf4b40c3e]
>> > >> > 5:
>> > >> > (BlueRocksWritableFile::Sync()+0x19) [0x559bf4b51ed9]
>> > >> > 6:

>> > (rocksdb::LegacyWritableFileWrapper::Sync(rocksdb::IOOptions
>> > const&,
>> > >> > rocksdb::IODebugContext*)+0x22) [0x559bf507fbd2]
>> > >> > 7:

>> > (rocksdb::WritableFileWriter::SyncInternal(bool)+0x5aa)
>> > [0x559bf51a880a]
>> > >> > 8:
>> > >> > (rocksdb::WritableFileWriter::Sync(bool)+0x100)
>> > [0x559bf51aa0a0]
>> > >> > 9:
>> > >> > (rocksdb::SyncManifest(rocksdb::Env*,
>> > rocksdb::ImmutableDBOptions
>> > >> const*,
>> > >> > rocksdb::WritableFileWriter*)+0x10b) [0x559bf51a3bfb]
>> > >> > 10:


>> > (rocksdb::VersionSet::ProcessManifestWrites(std::deque<rocksdb::VersionSet::ManifestWriter,

>> > std::allocator<rocksdb::VersionSet::ManifestWriter> >&,
>> > >> > rocksdb::InstrumentedMutex*,
>> > rocksdb::FSDirectory*, bool, rocks>
>> > >> > 11:


>> > (rocksdb::VersionSet::LogAndApply(rocksdb::autovector<rocksdb::ColumnFamilyData*,
>> > >> > 8ul> const&,
>> > rocksdb::autovector<rocksdb::MutableCFOptions const*, 8ul>
>> > >> > const&,
>> > rocksdb::autovector<rocksdb::autovector<rocksdb::>
>> > >> > 12:

>> > (rocksdb::VersionSet::LogAndApply(rocksdb::ColumnFamilyData*,
>> > >> > rocksdb::MutableCFOptions const&,
>> > rocksdb::VersionEdit*,
>> > >> > rocksdb::InstrumentedMutex*,
>> > rocksdb::FSDirectory*, bool,
>> > >> > rocksdb::ColumnFamilyOptions const>
>> > >> > 13:

>> > (rocksdb::DBImpl::DeleteUnreferencedSstFiles()+0xa30)
>> > [0x559bf50bd250]
>> > >> > 14:

>> > (rocksdb::DBImpl::Recover(std::vector<rocksdb::ColumnFamilyDescriptor,
>> > >> > std::allocator<rocksdb::ColumnFamilyDescriptor> >
>> > const&, bool, bool,
>> > >> bool,
>> > >> > unsigned long*)+0x13f1) [0x559bf50d3f21]
>> > >> > 15:
>> > >> > (rocksdb::DBImpl::Open(rocksdb::DBOptions const&,
>> > >> > std::__cxx11::basic_string<char,
>> > std::char_traits<char>,
>> > >> > std::allocator<char> > const&,
>> > >> std::vector<rocksdb::ColumnFamilyDescriptor,
>> > >> > std::allocator<rocksdb::Colu>
>> > >> > 16:
>> > >> > (rocksdb::DB::Open(rocksdb::DBOptions const&,
>> > >> > std::__cxx11::basic_string<char,
>> > std::char_traits<char>,
>> > >> > std::allocator<char> > const&,
>> > >> std::vector<rocksdb::ColumnFamilyDescriptor,
>> > >> > std::allocator<rocksdb::ColumnFa>
>> > >> > 17:
>> > >> > (RocksDBStore::do_open(std::ostream&, bool, bool,
>> > >> > std::__cxx11::basic_string<char,
>> > std::char_traits<char>,
>> > >> > std::allocator<char> > const&)+0x77a)
>> > [0x559bf503766a]
>> > >> > 18:
>> > >> > (BlueStore::_open_db(bool, bool, bool)+0xbb4)
>> > [0x559bf4a4bff4]
>> > >> > 19:
>> > >> > (BlueStore::_open_db_and_around(bool,
>> > bool)+0x500) [0x559bf4a766e0]
>> > >> > 20:
>> > >> > (BlueStore::_mount()+0x396) [0x559bf4a795d6]
>> > >> > 21:
>> > >> > (OSD::init()+0x556) [0x559bf44a0eb6]
>> > >> > 22: main()
>> > >> > 23:
>> > >> > __libc_start_main()
>> > >> > 24:
>> > >> _start()

>> > >> > *** Caught signal (Aborted) **
>> > >> > in thread
>> > >> > 7f0adfdef540 thread_name:ceph-osd

>> > >> > ceph
>> > >> version
>> > >> > 17.2.6 (d7ff0d10654d2280e08f1ab989c7cdf3064446a5)
>> > quincy (stable)
>> > >> > 1:
>> > >> > /lib64/libpthread.so.0(+0x12cf0) [0x7f0addff1cf0]
>> > >> > 2:
>> > >> gsignal()
>> > >> > 3: abort()
>> > >> > 4:
>> > >> > (ceph::__ceph_abort(char const*, int, char const*,
>> > >> > std::__cxx11::basic_string<char,
>> > std::char_traits<char>,
>> > >> > std::allocator<char> > const&)+0x197)
>> > [0x559bf4361def]
>> > >> > 5:
>> > >> > (BlueFS::_flush_range_F(BlueFS::FileWriter*,
>> > unsigned long, unsigned
>> > >> > long)+0x7a9) [0x559bf4b225f9]
>> > >> > 6:
>> > >> > (BlueFS::_flush_F(BlueFS::FileWriter*, bool,
>> > bool*)+0xa2)
>> > >> [0x559bf4b22812]
>> > >> > 7:
>> > >> > (BlueFS::fsync(BlueFS::FileWriter*)+0x8e)
>> > [0x559bf4b40c3e]
>> > >> > 8:
>> > >> > (BlueRocksWritableFile::Sync()+0x19) [0x559bf4b51ed9]
>> > >> > 9:

>> > (rocksdb::LegacyWritableFileWrapper::Sync(rocksdb::IOOptions
>> > const&,
>> > >> > rocksdb::IODebugContext*)+0x22) [0x559bf507fbd2]
>> > >> > 10:

>> > (rocksdb::WritableFileWriter::SyncInternal(bool)+0x5aa)
>> > [0x559bf51a880a]
>> > >> > 11:
>> > >> > (rocksdb::WritableFileWriter::Sync(bool)+0x100)
>> > [0x559bf51aa0a0]
>> > >> > 12:
>> > >> > (rocksdb::SyncManifest(rocksdb::Env*,
>> > rocksdb::ImmutableDBOptions
>> > >> const*,
>> > >> > rocksdb::WritableFileWriter*)+0x10b) [0x559bf51a3bfb]
>> > >> > 13:


>> > (rocksdb::VersionSet::ProcessManifestWrites(std::deque<rocksdb::VersionSet::ManifestWriter,

>> > std::allocator<rocksdb::VersionSet::ManifestWriter> >&,
>> > >> > rocksdb::InstrumentedMutex*,
>> > rocksdb::FSDirectory*, bool, rocks>
>> > >> > 14:


>> > (rocksdb::VersionSet::LogAndApply(rocksdb::autovector<rocksdb::ColumnFamilyData*,
>> > >> > 8ul> const&,
>> > rocksdb::autovector<rocksdb::MutableCFOptions const*, 8ul>
>> > >> > const&,
>> > rocksdb::autovector<rocksdb::autovector<rocksdb::>
>> > >> > 15:

>> > (rocksdb::VersionSet::LogAndApply(rocksdb::ColumnFamilyData*,
>> > >> > rocksdb::MutableCFOptions const&,
>> > rocksdb::VersionEdit*,
>> > >> > rocksdb::InstrumentedMutex*,
>> > rocksdb::FSDirectory*, bool,
>> > >> > rocksdb::ColumnFamilyOptions const>
>> > >> > 16:

>> > (rocksdb::DBImpl::DeleteUnreferencedSstFiles()+0xa30)
>> > [0x559bf50bd250]
>> > >> > 17:

>> > (rocksdb::DBImpl::Recover(std::vector<rocksdb::ColumnFamilyDescriptor,
>> > >> > std::allocator<rocksdb::ColumnFamilyDescriptor> >
>> > const&, bool, bool,
>> > >> bool,
>> > >> > unsigned long*)+0x13f1) [0x559bf50d3f21]
>> > >> > 18:
>> > >> > (rocksdb::DBImpl::Open(rocksdb::DBOptions const&,
>> > >> > std::__cxx11::basic_string<char,
>> > std::char_traits<char>,
>> > >> > std::allocator<char> > const&,
>> > >> std::vector<rocksdb::ColumnFamilyDescriptor,
>> > >> > std::allocator<rocksdb::Colu>
>> > >> > 19:
>> > >> > (rocksdb::DB::Open(rocksdb::DBOptions const&,
>> > >> > std::__cxx11::basic_string<char,
>> > std::char_traits<char>,
>> > >> > std::allocator<char> > const&,
>> > >> std::vector<rocksdb::ColumnFamilyDescriptor,
>> > >> > std::allocator<rocksdb::ColumnFa>
>> > >> > 20:
>> > >> > (RocksDBStore::do_open(std::ostream&, bool, bool,
>> > >> > std::__cxx11::basic_string<char,
>> > std::char_traits<char>,
>> > >> > std::allocator<char> > const&)+0x77a)
>> > [0x559bf503766a]
>> > >> > 21:
>> > >> > (BlueStore::_open_db(bool, bool, bool)+0xbb4)
>> > [0x559bf4a4bff4]
>> > >> > 22:
>> > >> > (BlueStore::_open_db_and_around(bool,
>> > bool)+0x500) [0x559bf4a766e0]
>> > >> > 23:
>> > >> > (BlueStore::_mount()+0x396) [0x559bf4a795d6]
>> > >> > 24:
>> > >> > (OSD::init()+0x556) [0x559bf44a0eb6]
>> > >> > 25: main()
>> > >> > 26:
>> > >> > __libc_start_main()
>> > >> > 27:
>> > >> _start()
>> > >> > NOTE: a
>> > >> copy
>> > >> > of the executable, or `objdump -rdS <executable>`
>> > is needed to interpret
>> > >> > this.
>> > >> > _______________________________________________
>> > >> > ceph-users mailing list -- ceph-users@xxxxxxx
>> > >> > To unsubscribe send an email to
>> > ceph-users-leave@xxxxxxx


>> > _______________________________________________
>> > ceph-users mailing list -- ceph-users@xxxxxxx
>> > To unsubscribe send an email to ceph-users-leave@xxxxxxx
>> > _______________________________________________
>> > ceph-users mailing list -- ceph-users@xxxxxxx
>> > To unsubscribe send an email to ceph-users-leave@xxxxxxx


>> > ------------------------------------------------------------------------


>> > This message is confidential and is for the
>> > sole use of the intended recipient(s). It
>> > may also be privileged or otherwise
>> > protected by copyright or other legal rules.
>> > If you have received it by mistake please
>> > let us know by reply email and delete it
>> > from your system. It is prohibited to copy
>> > this message or disclose its content to
>> > anyone. Any confidentiality or privilege is
>> > not waived or lost by any mistaken delivery
>> > or unauthorized disclosure of the message.
>> > All messages sent to and from Agoda may be
>> > monitored to ensure compliance with company
>> > policies, to protect the company's interests
>> > and to remove potential malware. Electronic
>> > messages may be intercepted, amended, lost
>> > or deleted, or contain viruses.



>> _______________________________________________
>> ceph-users mailing list -- ceph-users@xxxxxxx
>> To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux