Re: BlueStore not surviving power outage

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Thanks, everyone.

There is a RAID HBA in each of the machines in our clusters, to which
all SATA disks are attached. We configured the RAID HBA cache mode to
"write through", but, as I checked yesterday, the BBU of the RAID HBAs
are not charged. I'm not quite sure whether the BBU has something to
do with the data loss, as far as I know, all data should be persisted
to the underlying disk before acknowledging upper layer systems when
cache mode is "write through". Am I missing anything? Thanks:-)

On Tue, 27 Apr 2021 at 23:43, Martin Verges <martin.verges@xxxxxxxx> wrote:
>
> What drives do you use? Do they have PLP (power loss protection)? Is
> there any form of raid controller involved?
>
> --
> Martin Verges
> Managing director
>
> Mobile: +49 174 9335695
> E-Mail: martin.verges@xxxxxxxx
> Chat: https://t.me/MartinVerges
>
> croit GmbH, Freseniusstr. 31h, 81247 Munich
> CEO: Martin Verges - VAT-ID: DE310638492
> Com. register: Amtsgericht Munich HRB 231263
>
> Web: https://croit.io
> YouTube: https://goo.gl/PGE1Bx
>
> On Tue, 27 Apr 2021 at 10:54, Xuehan Xu <xxhdx1985126@xxxxxxxxx> wrote:
> >
> > Hi, everyone.
> >
> > Recently, one of our online cluster experienced a whole cluster power
> > outage, and after the power recovered, many osd started to log the
> > following error:
> >
> > 2021-04-27 15:38:05.503 2b372b957700 -1
> > bluestore(/var/lib/ceph/osd/ceph-3) _verify_csum bad crc32c/0x1000
> > checksum at blob offset 0x36000, got 0x41fe1397, expected 0x8d7f5975,
> > device location [0xa7e76000~1000], logical extent 0x1b6000~1000,
> > object #9:45a4e02a:::rbd_data.3b35df93038d.0000000000000095:head#
> > 2021-04-27 15:38:05.504 2b372b957700 -1
> > bluestore(/var/lib/ceph/osd/ceph-3) _verify_csum bad crc32c/0x1000
> > checksum at blob offset 0x36000, got 0x41fe1397, expected 0x8d7f5975,
> > device location [0xa7e76000~1000], logical extent 0x1b6000~1000,
> > object #9:45a4e02a:::rbd_data.3b35df93038d.0000000000000095:head#
> > 2021-04-27 15:38:05.505 2b372b957700 -1
> > bluestore(/var/lib/ceph/osd/ceph-3) _verify_csum bad crc32c/0x1000
> > checksum at blob offset 0x36000, got 0x41fe1397, expected 0x8d7f5975,
> > device location [0xa7e76000~1000], logical extent 0x1b6000~1000,
> > object #9:45a4e02a:::rbd_data.3b35df93038d.0000000000000095:head#
> > 2021-04-27 15:38:05.506 2b372b957700 -1
> > bluestore(/var/lib/ceph/osd/ceph-3) _verify_csum bad crc32c/0x1000
> > checksum at blob offset 0x36000, got 0x41fe1397, expected 0x8d7f5975,
> > device location [0xa7e76000~1000], logical extent 0x1b6000~1000,
> > object #9:45a4e02a:::rbd_data.3b35df93038d.0000000000000095:head#
> > 2021-04-27 15:38:28.379 2b372c158700 -1
> > bluestore(/var/lib/ceph/osd/ceph-3) _verify_csum bad crc32c/0x1000
> > checksum at blob offset 0x40000, got 0xce935e16, expected 0x9b502da7,
> > device location [0xa9a80000~1000], logical extent 0x80000~1000, object
> > #9:c2a6d9ae:::rbd_data.3b35df93038d.0000000000000696:head#
> >
> > We are using Nautilus 14.2.10 version, and we put rocksdb on top of
> > SSDs while bluestore data on SATA disks. It seems that the BlueStore
> > didn't survive the power outage, is it supposed to behave this way? Is
> > there any way to prevent it?
> >
> > Thanks:-)
> > _______________________________________________
> > Dev mailing list -- dev@xxxxxxx
> > To unsubscribe send an email to dev-leave@xxxxxxx
_______________________________________________
Dev mailing list -- dev@xxxxxxx
To unsubscribe send an email to dev-leave@xxxxxxx



[Index of Archives]     [CEPH Users]     [Ceph Devel]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]

  Powered by Linux