Re: question about rgw delete speed

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Brent,

Thanks for your input.

We will use Swift instead of S3. The deletes are mainly done by our customers using the sync app  (i.e they are syncing their folders with the storage accounts and every file change is translated to a delete in the cloud). We have a frontend cluster between the customers and the storage providing access via FTP/HTTP/Webdav and so on.

 The delete speed is important for us because we want to regain the 'deleted' storage capacity as free as possible so we can keep the costs down. I'm pretty obsessed with that because I went through some nightmares in the past having our storage full for some time :).

For the network we will probably use 1x10Gbps or 2x10Gbps for every OSD server.

Actually I have a Ceph cluster in production already but it's been used as a secondary storage. We are moving here the 'cold data' from the primary storage (i.e the big files) . We are using it as a secondary storage because it was deployed with 'zombie' OSD servers (old DDR2 servers recovered from other projects but with new SATA drives). It's working very  welll so far and it helped us to lower the costs with 30-40% per TB but we cannot use it as primary.



On 11/13/2020 1:07 AM, Brent Kennedy wrote:
Ceph is definitely a good choice for storing millions of files.  It sounds like you plan to use this like s3, so my first question would be:  Are the deletes done for a specific reason?  ( e.g. the files are used for a process and discarded  )  If its an age thing, you can set the files to expire when putting them in, then ceph will automatically clear them.

The more spinners you have the more performance you will end up with.  Network 10Gb or higher?

Octopus is production stable and contains many performance enhancements.  Depending on the OS, you may not be able to upgrade from nautilus until they work out that process ( e.g. centos 7/8 ).

Delete speed is not that great but you would have to test it with your cluster to see how it performs for your use case.  If you have enough space present, is there a process that breaks if the files are not deleted?


Regards,
-Brent

Existing Clusters:
Test: Ocotpus 15.2.5 ( all virtual on nvme )
US Production(HDD): Nautilus 14.2.11 with 11 osd servers, 3 mons, 4 gateways, 2 iscsi gateways
UK Production(HDD): Nautilus 14.2.11 with 18 osd servers, 3 mons, 4 gateways, 2 iscsi gateways
US Production(SSD): Nautilus 14.2.11 with 6 osd servers, 3 mons, 4 gateways, 2 iscsi gateways
UK Production(SSD): Octopus 15.2.5 with 5 osd servers, 3 mons, 4 gateways




-----Original Message-----
From: Adrian Nicolae <adrian.nicolae@xxxxxxxxxx>
Sent: Wednesday, November 11, 2020 3:42 PM
To: ceph-users <ceph-users@xxxxxxx>
Subject:  question about rgw delete speed


Hey guys,


I'm in charge of a local cloud-storage service. Our primary object storage is a vendor-based one and I want to replace it in the near future with Ceph with the following setup :

- 6 OSD servers with 36 SATA 16TB drives each and 3 big NVME per server
(1 big NVME for every 12 drives so I can reserve 300GB NVME storage for every SATA drive), 3 MON, 2 RGW with Epyc 7402p and 128GB RAM. So in the end we'll have ~ 3PB of raw data and 216 SATA drives.

Currently we have ~ 100 millions of files on the primary storage with the following distribution :

- ~10% = very small files ( less than 1MB - thumbnails, text&office files and so on)

- ~60%= small files (between 1MB and 10MB)

-  20% = medium files ( between 10MB and 1GB)

- 10% = big files (over 1GB).

My main concern is the speed of delete operations. We have around 500k-600k delete ops every 24 hours so quite a lot. Our current storage is not deleting all the files fast enough (it's always 1 week-10 days
behind) , I guess is not only a software issue and probably the delete speed will get better if we add more drives (we now have 108).

What do you think about Ceph delete speed ? I read on other threads that it's not very fast . I wonder if this hw setup can handle our current delete load better than our current storage. On RGW servers I want to use Swift , not S3.

And another question :   can I start deploying in production directly the latest Ceph version (Octopus) or is it safer to start with Nautilus until Octopus will be more stable ?

Any input would be greatly appreciated !


Thanks,

Adrian.




_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux