Hi,
I am doing a PoC with 3 DELL R620 and 12 OSD , 3 SSD drives ( one on each server), bluestore
I configured the OSD using the following ( /dev/sda is my SSD drive)
ceph-disk prepare --zap-disk --cluster ceph --bluestore /dev/sde --block.wal /dev/sda --block.db /dev/sda
Unfortunately both fio and bench tests show much worse performance for the pools than for the individual disks
Example:
DISKS
fio
--filename=/dev/sda --direct=1 --sync=1 --rw=write --bs=4k --numjobs=14 --iodepth=1 --runtime=60 --time_based --group_reporting --name=journal-test
SSD drive
Jobs: 14 (f=14): [W(14)] [100.0% done] [0KB/465.2MB/0KB /s] [0/119K/0 iops] [eta 00m:00s]
HD drive
Jobs: 14 (f=14): [W(14)] [100.0% done] [0KB/179.2MB/0KB /s] [0/45.9K/0 iops] [eta 00m:00s]
POOL
fio write.fio
Jobs: 1 (f=0): [w(1)] [100.0% done] [0KB/51428KB/0KB /s] [0/12.9K/0 iops]
cat write.fio
[write-4M]
description="write test with 4k block"
ioengine=rbd
clientname=admin
pool=scbench
rbdname=image01
iodepth=32
runtime=120
rw=randwrite
bs=4k
rados bench -p scbench 12 write
Max bandwidth (MB/sec): 224
Min bandwidth (MB/sec): 0
Average IOPS: 26
Stddev IOPS: 24
Max IOPS: 56
Min IOPS: 0
Average Latency(s): 0.59819
Stddev Latency(s): 1.64017
Max latency(s): 10.8335
Min latency(s): 0.00475139
I must be missing something - any help/suggestions will be greatly appreciated
Here are some specific info
ceph -s
cluster:
id: 91118dde-f231-4e54-a5f0-a1037f3d5142
health: HEALTH_OK
services:
mon: 1 daemons, quorum mon01
mgr: mon01(active)
osd: 12 osds: 12 up, 12 in
data:
pools: 4 pools, 484 pgs
objects: 70082 objects, 273 GB
usage: 570 GB used, 6138 GB / 6708 GB avail
pgs: 484 active+clean
io:
client: 2558 B/s rd, 2 op/s rd, 0 op/s wr
ceph osd pool ls detail
pool 1 'test-replicated' replicated size 2 min_size 1 crush_rule 0 object_hash rjenkins pg_num 128 pgp_num 128 last_change 157 flags hashpspool stripe_width 0 application rbd
removed_snaps [1~3]
pool 2 'test-erasure' erasure size 3 min_size 3 crush_rule 1 object_hash rjenkins pg_num 128 pgp_num 128 last_change 334 flags hashpspool stripe_width 8192 application rbd
removed_snaps [1~5]
pool 3 'rbd' replicated size 2 min_size 1 crush_rule 0 object_hash rjenkins pg_num 128 pgp_num 128 last_change 200 flags hashpspool stripe_width 0 application rbd
removed_snaps [1~3]
pool 4 'scbench' replicated size 2 min_size 1 crush_rule 0 object_hash rjenkins pg_num 100 pgp_num 100 last_change 330 flags hashpspool stripe_width 0
removed_snaps [1~3]
[cephuser@ceph ceph-config]$ ceph osd df tree
ID CLASS WEIGHT REWEIGHT SIZE USE AVAIL %USE VAR PGS TYPE NAME
-1 6.55128 - 2237G 198G 2038G 0 0 - root default
-7 0 - 0 0 0 0 0 - host ods03
-3 2.18475 - 2237G 181G 2055G 8.12 0.96 - host osd01
3 hdd 0.54619 1.00000 559G 53890M 506G 9.41 1.11 90 osd.3
4 hdd 0.54619 1.00000 559G 30567M 529G 5.34 0.63 89 osd.4
5 hdd 0.54619 1.00000 559G 59385M 501G 10.37 1.22 93 osd.5
6 hdd 0.54619 1.00000 559G 42156M 518G 7.36 0.87 93 osd.6
-5 2.18178 - 2234G 189G 2044G 8.50 1.00 - host osd02
0 hdd 0.54520 1.00000 558G 32460M 526G 5.68 0.67 90 osd.0
1 hdd 0.54520 1.00000 558G 54578M 504G 9.55 1.12 89 osd.1
2 hdd 0.54520 1.00000 558G 47761M 511G 8.35 0.98 93 osd.2
7 hdd 0.54619 1.00000 559G 59584M 501G 10.40 1.22 92 osd.7
-9 2.18475 - 2237G 198G 2038G 8.88 1.04 - host osd03
8 hdd 0.54619 1.00000 559G 52462M 508G 9.16 1.08 99 osd.8
10 hdd 0.54619 1.00000 559G 35284M 524G 6.16 0.73 88 osd.10
11 hdd 0.54619 1.00000 559G 71739M 489G 12.53 1.47 87 osd.11
12 hdd 0.54619 1.00000 559G 43832M 516G 7.65 0.90 93 osd.12
TOTAL 6708G 570G 6138G 8.50
MIN/MAX VAR: 0.63/1.47 STDDEV: 2.06
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com