On 05/19/2015 11:31 AM, Srikanth Madugundi wrote:
Hi, I am seeing write performance hit with small files (60K) using radosgw. The radosgw is configured to run with 600 threads. Here is the write speed I get with file sizes of 60K # sudo ceph -s cluster e445e46e-4d84-4606-9923-16fff64446dc health HEALTH_OK monmap e1: 1 mons at {osd187=13.24.0.7:6789/0 <http://13.24.0.7:6789/0>}, election epoch 1, quorum 0 osd187 osdmap e205: 28 osds: 22 up, 22 in pgmap v17007: 1078 pgs, 9 pools, 154 GB data, 653 kobjects 292 GB used, 8709 GB / 9002 GB avail 1078 active+clean client io 1117 kB/s rd, *2878 kB/s wr*, 2513 op/s
It appears that you have 22 OSDs and between reads and writes, there are ~114 ops/s per OSD. How many ops/s per disk are you trying to achieve?
# If I run the same script with larger file sizes(1MB-3MB), I get a better write speed.
Generally larger files will do better for a variety of reasons, but the primary one is that the data will consistently be more sequentially laid out. Assuming your OSDs are on spinning disks, this is a big advantage.
# sudo ceph -s cluster e445e46e-4d84-4606-9923-16fff64446dc health HEALTH_OK monmap e1: 1 mons at {osd187=13.24.0.79:6789/0 <http://13.24.0.79:6789/0>}, election epoch 1, quorum 0 osd187 osdmap e205: 28 osds: 22 up, 22 in pgmap v16883: 1078 pgs, 9 pools, 125 GB data, 140 kobjects 192 GB used, 8809 GB / 9002 GB avail 1078 active+clean client io *105 MB/s wr*, 1839 op/s # My cluster has 2 OSD hosts running total of 20 osd daemons, 1 mon and 1 radosgw hosts. Is the bottleneck coming from the single radosgw process? If so, is it possible to run radosgw in multi process mode?
I think before anyone can answer your question, it might help to detail what your hardware setup is, how you are running the tests, and what kind of performance you'd like to achieve.
Regards Srikanth _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com