On Fri, Sep 20, 2013 at 3:11 PM, Mark Nelson <mark.nelson@xxxxxxxxxxx> wrote: > On 09/20/2013 07:27 AM, Andreas Joachim Peters wrote: >> >> Hi, > > > Hi Andreas! > > >> >> we made some benchmarks about object read/write latencies on the CERN ceph >> installation. >> >> The cluster has 44 nodes and ~1k disks, all on 10GE and the pool >> configuration has 3 copies. >> Client & Server is 0.67. >> >> The latencies we observe (using tiny objects ... 5 bytes) on the idle >> pool: >> >> write full object(sync) ~65-80ms >> append to object ~60-75ms >> set xattr object ~65-80ms >> lock object ~65-80ms >> stat object ~1ms >> >> We seem to saturate the pools writing ~ 20k objects/s (= internally >> 60k/s). > > > Out of curiosity, how much difference do you see with write latencies if you > do the same thing to a pool with 1 copy? # 3 copies: [root@p05151113777233 ~]# rados bench -p test 10 write -t 1 -b 1 Maintaining 1 concurrent writes of 1 bytes for up to 10 seconds or 0 objects ... Average Latency: 0.0655107 Stddev Latency: 0.0156095 Max latency: 0.113482 Min latency: 0.033944 # 1 copy: [root@p01001532149022 ~]# rados bench -p test 10 write -t 1 -b 5 Maintaining 1 concurrent writes of 5 bytes for up to 10 seconds or 0 objects ... Average Latency: 0.0470315 Stddev Latency: 0.0204646 Max latency: 0.097039 Min latency: 0.004141 Cheers, Dan > > >> >> Is there an easy explanation for 80 ms (quasi without payload) and a >> possible tuning to reduce that? >> I measured (append few bytes +fsync) on such a disk around 33ms which >> explains probably part of the latency. > > > I've been wanting to really dig into object write latency in RADOS but just > haven't had the time to devote to it yet. I've been doing some simple rados > bench tests to a 8-SSD test node and am topping out at about 8-9K write IOPS > and 26K read IOPS (no replication) though with little tuning. I suspect > there are many areas in the code where we could improve things. > > >> >> Then I tried with the async API to see if there is a difference in the >> measurement between wait_for_complete or wait_for_safe ... shouldn't >> wait_for_complete be much shorter, but I get always comparable results ... > > > Hrm, I'm going to let Sage or someone else comment on this. > >> >> Thanks, Andreas.-- >> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in >> the body of a message to majordomo@xxxxxxxxxxxxxxx >> More majordomo info at http://vger.kernel.org/majordomo-info.html > > > -- > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in > the body of a message to majordomo@xxxxxxxxxxxxxxx > More majordomo info at http://vger.kernel.org/majordomo-info.html -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html