Hi Menguy-san, Thank you for your reply. Users who use large IO with tiny volumes are a nuisance to cloud providers. I confirmed my ceph cluster with 40 SSDs. Each OSD on 1TB SSD has about 50 placement groups in my cluster. Therefore, each PG has approximately 20GB of space. If we create a small 8GB volume, I had a feeling it wouldn't be distributed well, but it will be distributed well. Regards, -- Mitsumasa KONDO 2024年4月15日(月) 15:29 Etienne Menguy <etienne.menguy@xxxxxxxxxxx>: > Hi, > > Volume size doesn't affect performance, cloud providers apply a limit to > ensure they can deliver expected performances to all their customers. > > Étienne > ------------------------------ > *From:* Mitsumasa KONDO <kondo.mitsumasa@xxxxxxxxx> > *Sent:* Monday, 15 April 2024 06:06 > *To:* ceph-users@xxxxxxx <ceph-users@xxxxxxx> > *Subject:* Performance of volume size, not a block size > > [Some people who received this message don't often get email from > kondo.mitsumasa@xxxxxxxxx. Learn why this is important at > https://aka.ms/LearnAboutSenderIdentification ] > > Hi, > > In AWS EBS gp3, AWS says that small volume size cannot achieve best > performance. I think it's a feature or tendency of general > distributed storages including Ceph. Is that right in Ceph block storage? I > read many docs on ceph community. I never heard of Ceph storage. > > > https://can01.safelinks.protection.outlook.com/?url=https%3A%2F%2Fdocs.aws.amazon.com%2Febs%2Flatest%2Fuserguide%2Fgeneral-purpose.html&data=05%7C02%7Cetienne.menguy%40ubisoft.com%7C3076825a4d2a4897074208dc5d017852%7Ce01bd386fa514210a2a429e5ab6f7ab1%7C0%7C0%7C638487508098942744%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C0%7C%7C%7C&sdata=wOQKqG41uccTbyNHDIps62ojcTFBZYlyxxp3TzccsJI%3D&reserved=0 > <https://docs.aws.amazon.com/ebs/latest/userguide/general-purpose.html> > > Regard, > -- > Mitsumasa KONDO > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx