On Wed, Nov 6, 2019 at 5:57 PM Hermann Himmelbauer <hermann@xxxxxxx> wrote: > > Dear Vitaliy, dear Paul, > > Changing the block size for "dd" makes a huge difference. > > However, still some things are not fully clear to me: > > As recommended, I tried writing / reading directly to the rbd and this > is blazingly fast: > > fio -ioengine=rbd -name=test -direct=1 -rw=read -bs=4M -iodepth=16 > -pool=SATA -rbdname=vm-100-disk-0 > > write: IOPS=40, BW=160MiB/s (168MB/s)(4096MiB/25529msec) > read: IOPS=135, BW=542MiB/s (568MB/s)(4096MiB/7556msec) > > When I do the same within the virtual machine, I get the following results: > > fio --filename=/dev/vdb -name=test -direct=1 -rw=write -bs=4M -iodepth=16 --ioengine=aio Also, VirtIO block devices are usually slower than virtual SCSI disks on VirtIO SCSI controllers. Paul > > -------- cache = writeback ---------- > Blocksize: 4M: > read : io=4096.0MB, bw=97640KB/s, iops=23, runt= 42957msec > write: io=4096.0MB, bw=85250KB/s, iops=20, runt= 49200msec > > Blocksize: 4k: > read : io=4096.0MB, bw=3988.6KB/s, iops=997, runt=1051599msec > write: io=4096.0MB, bw=14529KB/s, iops=3632, runt=288686msec > ------------------------------------- > > The speeds are much slower whereas I don't really know why. > Nevertheless, this seams also really reasonable, although it's strange > that there's no difference between "cache = unsafe" and "cache = > writeback". Moreover, I find it strange that reading with 4k-blocks is > that slow, while writing is still o.k. > > My virtual machine is Debian 8, with a paravirtualized block device > (/dev/vdb), the process (and qemu parameters) look like the following: > > root 1854681 18.3 8.0 5813032 1980804 ? Sl 16:00 20:55 > /usr/bin/kvm -id 100 -name bya-backend -chardev > socket,id=qmp,path=/var/run/qemu-server/100.qmp,server,nowait -mon > chardev=qmp,mode=control -chardev > socket,id=qmp-event,path=/var/run/qmeventd.sock,reconnect=5 -mon > chardev=qmp-event,mode=control -pidfile /var/run/qemu-server/100.pid > -daemonize -smbios type=1,uuid=e53d6e2d-708e-4511-bb26-b0f1aefd81c6 -smp > 4,sockets=1,cores=4,maxcpus=4 -nodefaults -boot > menu=on,strict=on,reboot-timeout=1000,splash=/usr/share/qemu-server/bootsplash.jpg > -vnc unix:/var/run/qemu-server/100.vnc,password -cpu > kvm64,+lahf_lm,+sep,+kvm_pv_unhalt,+kvm_pv_eoi,enforce -m 4096 -device > pci-bridge,id=pci.1,chassis_nr=1,bus=pci.0,addr=0x1e -device > pci-bridge,id=pci.2,chassis_nr=2,bus=pci.0,addr=0x1f -device > piix3-usb-uhci,id=uhci,bus=pci.0,addr=0x1.0x2 -device > usb-tablet,id=tablet,bus=uhci.0,port=1 -device > VGA,id=vga,bus=pci.0,addr=0x2 -device > virtio-balloon-pci,id=balloon0,bus=pci.0,addr=0x3 -iscsi > initiator-name=iqn.1993-08.org.debian:01:7c6dc4c7e9f -drive > if=none,id=drive-ide2,media=cdrom,aio=threads -device > ide-cd,bus=ide.1,unit=0,drive=drive-ide2,id=ide2,bootindex=200 -drive > file=/mnt/pve/pontos-images/images/100/vm-100-disk-2.raw,if=none,id=drive-virtio0,cache=writeback,format=raw,aio=threads,detect-zeroes=on > -device > virtio-blk-pci,drive=drive-virtio0,id=virtio0,bus=pci.0,addr=0xa,bootindex=100 > -drive > file=rbd:SATA/vm-100-disk-0:conf=/etc/pve/ceph.conf:id=admin:keyring=/etc/pve/priv/ceph/SATA.keyring,if=none,id=drive-virtio1,cache=writeback,format=raw,aio=threads,detect-zeroes=on > -device virtio-blk-pci,drive=drive-virtio1,id=virtio1,bus=pci.0,addr=0xb > -netdev > type=tap,id=net0,ifname=tap100i0,script=/var/lib/qemu-server/pve-bridge,downscript=/var/lib/qemu-server/pve-bridgedown > -device > e1000,mac=46:22:36:C3:37:7E,netdev=net0,bus=pci.0,addr=0x12,id=net0,bootindex=300 > -machine type=pc > > In case you have some further speedup-hints for me, I'm glad. > > Nevertheless thank you a lot for help! > > Best Regards, > Hermann > > Am 05.11.19 um 12:49 schrieb Виталий Филиппов: > > Yes, cache=unsafe has no effect with RBD. Hm, that's strange, you should > > get ~40*6 MB/s linear write with 6 HDDs and Bluestore. > > > > Try to create a test image and test it with 'fio -ioengine=rbd > > -name=test -direct=1 -rw=write -bs=4M -iodepth=16 -pool=<pool> > > -rbdname=<rbd>' from outside a VM. > > > > If you still get 4 MB/s, something's wrong with your ceph. If you get > > adequate performance, something's wrong with your VM settings. > > > > 5 ноября 2019 г. 14:31:38 GMT+03:00, Hermann Himmelbauer > > <hermann@xxxxxxx> пишет: > > > > Hi, > > Thank you for your quick reply, Proxmox offers me "writeback" > > (cache=writeback) and "writeback unsafe" (cache=unsafe), however, for my > > "dd" test, this makes no difference at all. > > > > I still have write speeds of ~ 4,5 MB/s. > > > > Perhaps "dd" disables the write cache? > > > > Would it perhaps help to put the journal or something else on a SSD? > > > > Best Regards, > > Hermann > > > > Am 05.11.19 um 11:49 schrieb vitalif@xxxxxxxxxx: > > > > Use `cache=writeback` QEMU option for HDD clusters, that should > > solve > > your issue > > > > Hi, > > I recently upgraded my 3-node cluster to proxmox 6 / > > debian-10 and > > recreated my ceph cluster with a new release (14.2.4 > > bluestore) - > > basically hoping to gain some I/O speed. > > > > The installation went flawlessly, reading is faster than > > before (~ 80 > > MB/s), however, the write speed is still really slow (~ 3,5 > > MB/s). > > > > I wonder if I can do anything to speed things up? > > > > My Hardware is as the following: > > > > 3 Nodes with Supermicro X8DTT-HIBQF Mainboard each, > > 2 OSD per node (2TB SATA harddisks, WDC WD2000F9YZ-0), > > interconnected via Infiniband 40 > > > > The network should be reasonably fast, I measure ~ 16 GBit/s > > with iperf, > > so this seems fine. > > > > I use ceph for RBD only, so my measurement is simply doing a > > very simple > > "dd" read and write test within a virtual machine (Debian 8) > > like the > > following: > > > > read: > > dd if=/dev/vdb | pv | dd of=/dev/null > > -> 80 MB/s > > > > > > write: > > dd if=/dev/zero | pv | dd of=/dev/vdb > > -> 3.5 MB/s > > > > When I do the same on the virtual machine on a disk that is > > on a NFS > > storage, I get something about 30 MB/s for reading and writing. > > > > If I disable the write cache on all OSD disks via "hdparm -W 0 > > /dev/sdX", I gain a little bit of performance, write speed > > is then 4.3 > > MB/s. > > > > Thanks to your help from the list I plan to install a second > > ceph > > cluster which is SSD based (Samsung PM1725b) which should be > > much > > faster, however, I still wonder if there is any way to speed > > up my > > harddisk based cluster? > > > > Thank you in advance for any help, > > > > Best Regards, > > Hermann > > > > > > -- > > With best regards, > > Vitaliy Filippov > > -- > hermann@xxxxxxx > PGP/GPG: 299893C7 (on keyservers) > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx