Hi, Your performance is not that bad, is it? What performance do you expect? I just ran the same test. 12 Node, SATA SSD Only: READ: bw=63.8MiB/s (66.9MB/s), 63.8MiB/s-63.8MiB/s (66.9MB/s-66.9MB/s), io=3070MiB (3219MB), run=48097-48097msec WRITE: bw=21.3MiB/s (22.4MB/s), 21.3MiB/s-21.3MiB/s (22.4MB/s-22.4MB/s), io=1026MiB (1076MB), run=48097-48097msec 6 Node, SAS Only: READ: bw=22.1MiB/s (23.2MB/s), 22.1MiB/s-22.1MiB/s (23.2MB/s-23.2MB/s), io=3070MiB (3219MB), run=138650-138650msec WRITE: bw=7578KiB/s (7759kB/s), 7578KiB/s-7578KiB/s (7759kB/s-7759kB/s), io=1026MiB (1076MB), run=138650-138650msec This is OpenStack Queens with Ceph FileStore (Luminous). Kind regards, Sinan Polat > Op 2 januari 2020 om 10:59 schreef Stefan Kooman <stefan@xxxxxx>: > > > Quoting Ignazio Cassano (ignaziocassano@xxxxxxxxx): > > Hello All, > > I installed ceph luminous with openstack, an using fio in a virtual machine > > I got slow random writes: > > > > fio --randrepeat=1 --ioengine=libaio --direct=1 --gtod_reduce=1 --name=test > > --filename=random_read_write.fio --bs=4k --iodepth=64 --size=4G > > --readwrite=randrw --rwmixread=75 > > Do you use virtio-scsi with a SCSI queue per virtual CPU core? How many > cores do you have? I suspect that the queue depth is hampering > throughput here ... but is throughput performance really interesting > anyway for your use case? Low latency generally matters most. > > Gr. Stefan > > > -- > | BIT BV https://www.bit.nl/ Kamer van Koophandel 09090351 > | GPG: 0xD14839C6 +31 318 648 688 / info@xxxxxx > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com