Re: BlueStore and maximum number of objects per PG

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



> Op 22 februari 2017 om 11:51 schreef Wido den Hollander <wido@xxxxxxxx>:
> 
> 
> 
> > Op 22 februari 2017 om 3:53 schreef Mark Nelson <mnelson@xxxxxxxxxx>:
> > 
> > 
> > Hi Wido,
> > 
> > On 02/21/2017 02:04 PM, Wido den Hollander wrote:
> > > Hi,
> > >
> > > I'm about to start a test where I'll be putting a lot of objects into BlueStore and see how it holds.
> > >
> > > The reasoning behind is that I have a customer which has 165M objects in it's cluster which results in some PGs having 900k objects.
> > >
> > > For FileStore with XFS this is quite heavy. A simple scrub takes ages.
> > >
> > > The problem is that we can't simply increase the number of PGs since that will overload the OSDs as well.
> > >
> > > On the other hand we could add hardware, but that also takes time.
> > >
> > > So just for the sake of testing I'm looking at trying to replicate this situation using BlueStore from master.
> > >
> > > Is there anything I should take into account? I'll probably be just creating a lot (millions) of 100 byte objects in the cluster with just a few PGs.
> > 
> > Couple of general things:
> > 
> > I don't anticipate you'll run into the same kind of pg splitting 
> > slowdowns that you see with filestore, but you still may see some 
> > slowdown as the object count increases since rocksdb will have more 
> > key/value pairs to deal with.  I expect you'll see a lot of metadata 
> > movement between levels as it tries to keep things organized.  One thing 
> > to note is that it's possible you may see rocksdb bottlenecks as the OSD 
> > volume size increases.  This is one of the things the guys at Sandisk 
> > were trying to tackle with Zetascale.
> > 
> 
> Ah, ok!
> 
> > If you can put the rocksdb DB and WAL on SSDs that will likely help, but 
> > you'll want to be mindful of how full the SSDs are getting.  I'll be 
> > very curious to see how your tests go, it's been a while since we've 
> > thrown that many objects on a bluestore cluster (back around the 
> > newstore timeframe we filled bluestore with many 10s of millions of 
> > objects and from what I remember it did pretty well).
> > 
> 
> Thanks for the information! I'll try first with a few OSDs and size = 1 and just put a lot of small objects in the PG and see how it goes.
> 
> Will time the latency for writing and reading the objects afterwards to see how it goes.

First test, one OSD running inside VirtualBox with a 300GB disk and Luminous.

1 OSD, size = 1, pg_num = 8.

After 2.5M objects the disk was full... but the OSD was still working fine. Didn't experience any issues. Although the OSD was using 3.4GB of RAM at that moment while I stopped doing I/O.

2.5M objects of 128 bytes written to the disk.

Would like to scale this test out further, but I don't have hardware available to run it on.

Wido

> 
> Wido
> 
> > Mark
> > 
> > >
> > > Wido
> > > --
> > > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> > > the body of a message to majordomo@xxxxxxxxxxxxxxx
> > > More majordomo info at  http://vger.kernel.org/majordomo-info.html
> > >
> > --
> > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> > the body of a message to majordomo@xxxxxxxxxxxxxxx
> > More majordomo info at  http://vger.kernel.org/majordomo-info.html
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux