hello here the result : fio --ioengine=rbd --name=test --bs=4k --iodepth=1 --rw=randwrite --runtime=60 -pool=kube -rbdname=bench test: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=rbd, iodepth=1 fio-3.12 Starting 1 process Jobs: 1 (f=1): [w(1)][100.0%][w=248KiB/s][w=62 IOPS][eta 00m:00s] test: (groupid=0, jobs=1): err= 0: pid=1903256: Fri Aug 16 13:22:59 2019 write: IOPS=58, BW=232KiB/s (238kB/s)(13.6MiB/60011msec); 0 zone resets slat (usec): min=9, max=351, avg=52.82, stdev=23.02 clat (usec): min=1264, max=96970, avg=17156.70, stdev=6713.88 lat (usec): min=1276, max=97050, avg=17209.52, stdev=6715.06 clat percentiles (usec): | 1.00th=[ 2933], 5.00th=[ 3884], 10.00th=[11863], 20.00th=[13304], | 30.00th=[13960], 40.00th=[14484], 50.00th=[15008], 60.00th=[20579], | 70.00th=[22152], 80.00th=[23987], 90.00th=[25297], 95.00th=[25822], | 99.00th=[26346], 99.50th=[27395], 99.90th=[71828], 99.95th=[82314], | 99.99th=[96994] bw ( KiB/s): min= 104, max= 272, per=100.00%, avg=232.17, stdev=19.55, samples=120 iops : min= 26, max= 68, avg=57.97, stdev= 4.88, samples=120 lat (msec) : 2=0.06%, 4=5.51%, 10=3.41%, 20=50.22%, 50=40.69% lat (msec) : 100=0.11% cpu : usr=0.44%, sys=0.27%, ctx=3489, majf=0, minf=3582 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% issued rwts: total=0,3485,0,0 short=0,0,0,0 dropped=0,0,0,0 latency : target=0, window=0, percentile=100.00%, depth=1 Run status group 0 (all jobs): WRITE: bw=232KiB/s (238kB/s), 232KiB/s-232KiB/s (238kB/s-238kB/s), io=13.6MiB (14.3MB), run=60011-60011msec Disk stats (read/write): md2: ios=3/4611, merge=0/0, ticks=0/0, in_queue=0, util=0.00%, aggrios=3/8532, aggrmerge=0/2364, aggrticks=0/1346, aggrin_queue=52504, aggrutil=88.03% nvme1n1: ios=3/8529, merge=0/2295, ticks=1/1347, in_queue=52932, util=88.03% nvme0n1: ios=3/8535, merge=1/2434, ticks=0/1346, in_queue=52076, util=86.52% for you information my disk setup is : 2 500G nvme disk with : 10G raid 1 partition with the OS 160G raid0 partition for local docker data 387.00g partition on each disk for the osd. I got 5 physical devices with 12 core cpu and 32go RAM. Ceph is running on dedicated 1Gbps network. I got quite the same hardware setup with ceph 10.2.11 and I got much better performance. oau lsblk NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT rbd0 252:0 0 1G 0 disk /var/lib/kubelet/pods/db24eb3b-650c-42f3-bdf0-92ea0eaf37d8/volumes/kubernetes.io~csi/pvc-e70f3d74-c7bd-4652-983b-de3874d36117/mou rbd1 252:16 0 1G 0 disk /var/lib/kubelet/pods/a1b0b156-0bcc-4d0c-b2f8-26a74337baed/volumes/kubernetes.io~csi/pvc-14460f8a-da5a-44e3-a033-a3bf5054f967/mou rbd2 252:32 0 8G 0 disk /var/lib/kubelet/pods/0a5e3745-c6c2-49a9-971a-3ddac59af66c/volumes/kubernetes.io~csi/pvc-b8907f44-58d6-4599-b189-fafe65daed09/mou rbd3 252:48 0 50G 0 disk /var/lib/kubelet/pods/84df04af-a7dd-4035-a9a2-22d3d315fa60/volumes/kubernetes.io~csi/pvc-faf5cefa-ecce-450a-9f5c-42e7ca7d7fc2/mou rbd4 252:64 0 1G 0 disk /var/lib/kubelet/pods/872a3d7d-63d5-4567-86f6-bedab3fe0ad3/volumes/kubernetes.io~csi/pvc-83ea4013-b936-469b-b20f-7e703db2c871/mou nvme0n1 259:0 0 477G 0 disk ├─nvme0n1p1 259:1 0 511M 0 part /boot/efi ├─nvme0n1p2 259:2 0 9.8G 0 part │ └─md2 9:2 0 9.8G 0 raid1 / ├─nvme0n1p3 259:3 0 79.5G 0 part │ └─md3 9:3 0 158.9G 0 raid0 │ └─datavg-dockerlv 253:0 0 30G 0 lvm /var/lib/docker └─nvme0n1p4 259:8 0 387.2G 0 part └─ceph--52ce0eb9--9e69--4f29--8b87--9ab3fbb5df3e-osd--block--c26bdb06--2325--4fcd--9b8c--77a93ab46de0 253:1 0 387G 0 lvm nvme1n1 259:4 0 477G 0 disk ├─nvme1n1p1 259:5 0 511M 0 part ├─nvme1n1p2 259:6 0 9.8G 0 part │ └─md2 9:2 0 9.8G 0 raid1 / ├─nvme1n1p3 259:7 0 79.5G 0 part │ └─md3 9:3 0 158.9G 0 raid0 │ └─datavg-dockerlv 253:0 0 30G 0 lvm /var/lib/docker └─nvme1n1p4 259:9 0 387.2G 0 part └─ceph--4ef4eb77--4b23--4ffd--b7d2--0fd2cfa5e568-osd--block--28435fba--91c7--4424--99f2--ea86709a87ca 253:2 0 387G 0 lvm Le vendredi 16 août 2019 à 01:16 +0300, Vitaliy Filippov a écrit :
|
_______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx