Re: Fwd: BlueFS spillover yet again

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Thank you for the insight.
> If you're using the default options for rocksdb, then the size of L3 will
be 25GB
Where this number comes from? Any documentation I can read?
I want to have a better understanding on how DB size is calculated.

ср, 5 февр. 2020 г. в 18:53, Moreno, Orlando <orlando.moreno@xxxxxxxxx>:

> Hi Vladimir,
>
> If you're using the default options for rocksdb, then the size of L3 will
> be 25GB. Since your block-db is only 20GB and L3 can only be filled if the
> entire level's size is available, bluefs will begin spillover. Like Igor
> said, having 30GB+ is recommended if you want to host up to 3 levels of
> rocksdb in the SSD.
>
> Thanks,
> Orlando
>
> -----Original Message-----
> From: Igor Fedotov <ifedotov@xxxxxxx>
> Sent: Wednesday, February 5, 2020 7:04 AM
> To: Vladimir Prokofev <v@xxxxxxxxxxx>; ceph-users@xxxxxxx
> Subject:  Re: Fwd: BlueFS spillover yet again
>
> Hi Vladimir,
>
> there were a plenty of discussions/recommendations around db volume size
> selection here.
>
> In short it's advised to have DB volume of 30 - 64GB for most of use cases.
>
> Thanks,
>
> Igor
>
> On 2/5/2020 4:21 PM, Vladimir Prokofev wrote:
> > Cluster upgraded from 12.2.12 to 14.2.5. All went smooth, except
> > BlueFS spillover warning.
> > We create OSDs with ceph-deploy, command goes like this:
> > ceph-deploy osd create --bluestore --data /dev/sdf --block-db
> > /dev/sdb5 --block-wal /dev/sdb6 ceph-osd3 where block-db and block-wal
> > are SSD partitions.
> > Default ceph-deploy settings created partitions ~1GB which is, of
> > course, too small. So we redeployed OSDs using manually partitioned
> > SSD for block-db/block-wal with sizes of 20G/5G respectively.
> > But now we still get BlueFS spillover warning for redeployed OSDs:
> >       osd.10 spilled over 2.4 GiB metadata from 'db' device (2.8 GiB
> > used of
> > 19 GiB) to slow device
> >       osd.19 spilled over 3.7 GiB metadata from 'db' device (2.7 GiB
> > used of
> > 19 GiB) to slow device
> >       osd.20 spilled over 4.2 GiB metadata from 'db' device (2.6 GiB
> > used of
> > 19 GiB) to slow device
> > osd size is 1.8 TiB.
> >
> > These OSDs are used primarily for RBD as a backup drives, so a lot of
> > snapshots held there. They also have RGW pool assigned to them, but it
> > has no data.
> > I know of sizing recommendations[1] for block-db/block-wal, but I
> > assumed since it's primarily RBD 1%(~20G) should be enough.
> > Also, compaction stats doesn't make sense to me[2]. It states that sum
> > of DB is only 5.08GB, that should be placed on block-db without a
> problem?
> > Am I understanding all this wrong? Should block-db size be greater in
> > my case?
> >
> > [1]
> > https://docs.ceph.com/docs/master/rados/configuration/bluestore-config
> > -ref/#sizing
> > [2] osd.10 logs as an example
> > https://pastebin.com/hC6w6jSn
> > _______________________________________________
> > ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an
> > email to ceph-users-leave@xxxxxxx
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an
> email to ceph-users-leave@xxxxxxx
>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux