Hi, I have ceph cluster currently unused and I have (to my mind) very low performances. I'm not an expert in benchs, here an example of quick bench: --------------------------------------------------------------- # fio --randrepeat=1 --ioengine=libaio --direct=1 --gtod_reduce=1 --name=readwrite --filename=rw.data --bs=4k --iodepth=64 --size=300MB --readwrite=randrw --rwmixread=50 readwrite: (g=0): rw=randrw, bs=4K-4K/4K-4K/4K-4K, ioengine=libaio, iodepth=64 fio-2.1.3 Starting 1 process readwrite: Laying out IO file(s) (1 file(s) / 300MB) Jobs: 1 (f=1): [m] [100.0% done] [2264KB/2128KB/0KB /s] [566/532/0 iops] [eta 00m:00s] readwrite: (groupid=0, jobs=1): err= 0: pid=3783: Fri Dec 18 02:01:13 2015 read : io=153640KB, bw=2302.9KB/s, iops=575, runt= 66719msec write: io=153560KB, bw=2301.7KB/s, iops=575, runt= 66719msec cpu : usr=0.77%, sys=3.07%, ctx=115432, majf=0, minf=604 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.1%, >=64=99.9% submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.1%, >=64=0.0% issued : total=r=38410/w=38390/d=0, short=r=0/w=0/d=0 Run status group 0 (all jobs): READ: io=153640KB, aggrb=2302KB/s, minb=2302KB/s, maxb=2302KB/s, mint=66719msec, maxt=66719msec WRITE: io=153560KB, aggrb=2301KB/s, minb=2301KB/s, maxb=2301KB/s, mint=66719msec, maxt=66719msec --------------------------------------------------------------- It seems to me very bad. Can I hope better results with my setup (explained below)? During the bench, I don't see particular symptoms (no CPU blocked at 100% etc). If you have advices to improve the perf and/or maybe to make smarter benchs, I'm really interested. Thanks in advance for your help. Here is my conf... I use Ubuntu 14.04 on each server with the 3.13 kernel (it's the same for the client ceph where I run my bench) and I use Ceph 9.2.0 (Infernalis). On the client, cephfs is mounted via cephfs-fuse with this in /etc/fstab: id=cephfs,keyring=/etc/ceph/ceph.client.cephfs.keyring,client_mountpoint=/ /mnt/cephfs fuse.ceph noatime,defaults,_netdev 0 0 I have 5 cluster node servers "Supermicro Motherboard X10SLM+-LN4 S1150" with one 1GbE port for the ceph public network and one 10GbE port for the ceph private network: - 1 x Intel Xeon E3-1265Lv3 - 1 SSD DC3710 Series 200GB (with partitions for the OS, the 3 OSD-journals and, just for ceph01, ceph02 and ceph03, the SSD contains too a partition for the workdir of a monitor - 3 HD 4TB Western Digital (WD) SATA 7200rpm - RAM 32GB - NO RAID controlleur - Each partition uses XFS with noatim option, except the OS partition in EXT4. Here is my ceph.conf : --------------------------------------------------------------- [global] fsid = xxxxxxxxxxxxxxxxxxxxxxxxxxxx cluster network = 192.168.22.0/24 public network = 10.0.2.0/24 auth cluster required = cephx auth service required = cephx auth client required = cephx filestore xattr use omap = true osd pool default size = 3 osd pool default min size = 1 osd pool default pg num = 64 osd pool default pgp num = 64 osd crush chooseleaf type = 1 osd journal size = 0 osd max backfills = 1 osd recovery max active = 1 osd client op priority = 63 osd recovery op priority = 1 osd op threads = 4 mds cache size = 1000000 osd scrub begin hour = 3 osd scrub end hour = 5 mon allow pool delete = false mon osd down out subtree limit = host mon osd min down reporters = 4 [mon.ceph01] host = ceph01 mon addr = 10.0.2.101 [mon.ceph02] host = ceph02 mon addr = 10.0.2.102 [mon.ceph03] host = ceph03 mon addr = 10.0.2.103 --------------------------------------------------------------- mds are in active/standby mode. -- François Lafont _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com