Re: ceph write performance issue

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



From: ceph-users [mailto:ceph-users-bounces@xxxxxxxxxxxxxx] On Behalf Of min fang
Sent: 29 September 2016 10:34
To: ceph-users <ceph-users@xxxxxxxxxxxxxx>
Subject: ceph write performance issue

 

Hi, I created 40 osds ceph cluster with 8 PM863 960G SSD as journal. One ssd is used by 5 osd drives as journal.   The ssd 512 random write performance is about 450MB/s, but the whole cluster sequential write throughput is only 800MB/s. Any suggestion on improving sequential write performance? thanks.

 

Take a conservative figure of 50MB/s for each disk as writing in Ceph is not just straight sequential writes, there is a slight random nature to it.

(40x50MB/s)/3 = 666MB/s. Seems fine to me.

 


Testing result is here:
rados bench -p libvirt-pool 10 write --no-cleanup
Maintaining 16 concurrent writes of 4194304 bytes to objects of size 4194304 for up to 10 seconds or 0 objects
Object prefix: benchmark_data_redpower-sh-04_16462
  sec Cur ops   started  finished  avg MB/s  cur MB/s last lat(s)  avg lat(s)
    0       0         0         0         0         0           -           0
    1      15       189       174   695.968       696   0.0359122    0.082477
    2      16       395       379   757.938       820   0.0634079   0.0826266
    3      16       582       566   754.601       748   0.0401129   0.0830207
    4      16       796       780   779.934       856   0.0374938   0.0816794
    5      16       977       961   768.735       724   0.0489886   0.0827479
    6      16      1172      1156   770.601       780   0.0428639   0.0812062
    7      16      1387      1371   783.362       860   0.0461826   0.0811803
    8      16      1545      1529   764.433       632    0.238497   0.0831018
    9      16      1765      1749   777.265       880   0.0557358   0.0814399
   10      16      1971      1955   781.931       824   0.0321333   0.0814144
Total time run:         10.044813
Total writes made:      1972
Write size:             4194304
Object size:            4194304
Bandwidth (MB/sec):     785.281
Stddev Bandwidth:       80.8235
Max bandwidth (MB/sec): 880
Min bandwidth (MB/sec): 632
Average IOPS:           196
Stddev IOPS:            20
Max IOPS:               220
Min IOPS:               158
Average Latency(s):     0.081415
Stddev Latency(s):      0.0554568
Max latency(s):         0.345111
Min latency(s):         0.0230153

my ceph osd configuration:
sd_mkfs_type = xfs
osd_mount_options_xfs = rw,noatime,inode64,logbsize=256k
osd_mkfs_options_xfs = -f -i size=2048
filestore_max_inline_xattr_size = 254
filestore_max_inline_xattrs = 6
osd_op_threads = 20
filestore_queue_max_ops = 25000
journal_max_write_entries=10000
journal_queue_max_ops=50000
objecter_inflight_ops=10240
filestore_queue_max_bytes=1048576000
filestore_queue_committing_max_bytes =1048576000
journal_max_write_bytes=1073714824
journal_queue_max_bytes=10485760000
ms_dispatch_throttle_bytes=1048576000
objecter_infilght_op_bytes=1048576000
filestore_max_sync_interval=20
filestore_flusher=false
filestore_flush_min=0
filestore_sync_flush=true
journal_block_align = true
journal_dio = true
journal_aio = true
journal_force_aio = true
osd_op_num_shards=8
osd_op_num_threads_per_shard=2
filestore_wbthrottle_enable=false
filestore_fd_cache_size=1024
filestore_omap_header_cache_size=1024


_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux