Re: question about rgw delete speed

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



>From what we have experienced, our delete speed scales with the CPU
available to the MDS. And the MDS only seems to scale to 2-4 CPUs per
daemon, so for our biggest filesystem, we have 5 active MDS daemons.
Migrations reduced performance a lot, but pinning fixed that. Even better
is just getting the fastest cores you can get.

On Thu., Nov. 12, 2020, 6:08 p.m. Brent Kennedy, <bkennedy@xxxxxxxxxx>
wrote:

> Ceph is definitely a good choice for storing millions of files.  It sounds
> like you plan to use this like s3, so my first question would be:  Are the
> deletes done for a specific reason?  ( e.g. the files are used for a
> process and discarded  )  If its an age thing, you can set the files to
> expire when putting them in, then ceph will automatically clear them.
>
> The more spinners you have the more performance you will end up with.
> Network 10Gb or higher?
>
> Octopus is production stable and contains many performance enhancements.
> Depending on the OS, you may not be able to upgrade from nautilus until
> they work out that process ( e.g. centos 7/8 ).
>
> Delete speed is not that great but you would have to test it with your
> cluster to see how it performs for your use case.  If you have enough space
> present, is there a process that breaks if the files are not deleted?
>
>
> Regards,
> -Brent
>
> Existing Clusters:
> Test: Ocotpus 15.2.5 ( all virtual on nvme )
> US Production(HDD): Nautilus 14.2.11 with 11 osd servers, 3 mons, 4
> gateways, 2 iscsi gateways
> UK Production(HDD): Nautilus 14.2.11 with 18 osd servers, 3 mons, 4
> gateways, 2 iscsi gateways
> US Production(SSD): Nautilus 14.2.11 with 6 osd servers, 3 mons, 4
> gateways, 2 iscsi gateways
> UK Production(SSD): Octopus 15.2.5 with 5 osd servers, 3 mons, 4 gateways
>
>
>
>
> -----Original Message-----
> From: Adrian Nicolae <adrian.nicolae@xxxxxxxxxx>
> Sent: Wednesday, November 11, 2020 3:42 PM
> To: ceph-users <ceph-users@xxxxxxx>
> Subject:  question about rgw delete speed
>
>
> Hey guys,
>
>
> I'm in charge of a local cloud-storage service. Our primary object storage
> is a vendor-based one and I want to replace it in the near future with Ceph
> with the following setup :
>
> - 6 OSD servers with 36 SATA 16TB drives each and 3 big NVME per server
> (1 big NVME for every 12 drives so I can reserve 300GB NVME storage for
> every SATA drive), 3 MON, 2 RGW with Epyc 7402p and 128GB RAM. So in the
> end we'll have ~ 3PB of raw data and 216 SATA drives.
>
> Currently we have ~ 100 millions of files on the primary storage with the
> following distribution :
>
> - ~10% = very small files ( less than 1MB - thumbnails, text&office files
> and so on)
>
> - ~60%= small files (between 1MB and 10MB)
>
> -  20% = medium files ( between 10MB and 1GB)
>
> - 10% = big files (over 1GB).
>
> My main concern is the speed of delete operations. We have around
> 500k-600k delete ops every 24 hours so quite a lot. Our current storage is
> not deleting all the files fast enough (it's always 1 week-10 days
> behind) , I guess is not only a software issue and probably the delete
> speed will get better if we add more drives (we now have 108).
>
> What do you think about Ceph delete speed ? I read on other threads that
> it's not very fast . I wonder if this hw setup can handle our current
> delete load better than our current storage. On RGW servers I want to use
> Swift , not S3.
>
> And another question :   can I start deploying in production directly the
> latest Ceph version (Octopus) or is it safer to start with Nautilus until
> Octopus will be more stable ?
>
> Any input would be greatly appreciated !
>
>
> Thanks,
>
> Adrian.
>
>
>
>
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an
> email to ceph-users-leave@xxxxxxx
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux