You're doing 4K direct IOs on a distributed storage system and then comparing it to what the local device does with 1GB blocks? :)
Try feeding Ceph with some larger IOs and check how it does.
-Greg
On Fri, May 18, 2018 at 1:22 PM Rhugga Harper <rhugga@xxxxxxxxx> wrote:
_______________________________________________We're evaluating persistent block providers for Kubernetes and looking at ceph at the moment.We aren't seeing performance anywhere near what we expect.I have a 50-node proof of concept cluster with 40 nodes available for storage and configured with rook/ceph. Each has 10GB nics and 8 x 1TB SSD's. (only 3 drives on each node have been allocated to ceph use)We are testing with replicated pools of size 1 and 3. I've been doing fio tests in parallel (pod setup to run fio) and it seems to average aggregate bandwidth around 150 MB/sec.I'm running the fio tests as follows:direct=1, fsync=8|16|32|64, readwrite=write, blocksize=4k, numjobs=4|8, size=10GRegardless of doing 1 stream or 40, the aggregate bandwidth as reported by "ceph -s" is ~150MB/sec:I'm creating my pool with pg_num/pgp_num=1024|2048|4096A baseline dd (100GB file using blocksize 1G) on these SSD's shows them capable of 1.6 GB/s.I can't seem to find any limitations or bottlenecks on the nodes or the network.Anyone have any idea where else I can look?I'm new to ceph and it just seems like this should be pushing more I/O. I've dug thru a lot of performance tuning sites and have m implemented most of the suggestions.# ceph -scluster:id: 949a8caf-9a9b-4f09-8711-1d5158a65bd8health: HEALTH_OKservices:mon: 7 daemons, quorum rook-ceph-mon1,rook-ceph-mon3,rook-ceph-mon0,rook-ceph-mon5,rook-ceph-mon4,rook-ceph-mon2,rook-ceph-mon6mgr: rook-ceph-mgr0(active)osd: 123 osds: 123 up, 123 indata:pools: 1 pools, 2048 pgsobjects: 134k objects, 508 GBusage: 1240 GB used, 110 TB / 112 TB availpgs: 2048 active+cleanio:client: 138 MB/s wr, 0 op/s rd, 71163 op/s wrThanks for any help,CC
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com