Hi, same environment, after a test script, the io latency(get from sudo ceph --admin-daemon /run/ceph/guests/ceph-client.*.asok per dump) increase from about 4ms to 7.3ms qemu version: debian 2.1.2 kernel:3.10.45-openstack-amd64 system: debian 7.8 ceph: 0.94.5 VM CPU number: 4 (cpu MHz : 2599.998) VM memory size: 16GB 9 OSD storage servers, with 4 SSD OSD on each, total 36 OSDs. Test scripts in VM: # cat reproduce.sh #!/bin/bash times=20 for((i=1;i<=$times;i++)) do tmpdate=`date "+%F-%T"` echo "=======================$tmpdate($i/$times)=======================" tmp=$((i%2)) if [[ $tmp -eq 0 ]];then echo "############### fio /root/vdb.cfg ###############" fio /root/vdb.cfg else echo "############### fio /root/vdc.cfg ###############" fio /root/vdc.cfg fi done tmpdate=`date "+%F-%T"` echo "############### [$tmpdate] fio /root/vde.cfg ###############" fio /root/vde.cfg # cat vdb.cfg [global] rw=randwrite direct=1 numjobs=64 ioengine=sync bsrange=4k-4k runtime=180 group_reporting [disk01] filename=/dev/vdb # cat vdc.cfg [global] rw=randwrite direct=1 numjobs=64 ioengine=sync bsrange=4k-4k runtime=180 group_reporting [disk01] filename=/dev/vdc # cat vdd.cfg [global] rw=randwrite direct=1 numjobs=64 ioengine=sync bsrange=4k-4k runtime=180 group_reporting [disk01] filename=/dev/vdd # cat vde.cfg [global] rw=randwrite direct=1 numjobs=64 ioengine=sync bsrange=4k-4k runtime=180 group_reporting [disk01] filename=/dev/vde After run the scripts reproduce.sh, the disks in the VM's IOPS cutdown from 12k to 5k, the latency increase from 4ms to 7.3ms. run steps: 1. create a VM 2. create four volumes and attatch to the VM 3. sh reproduce.sh 4. in the runtime of reproduce.sh, run "fio vdd.cfg" or "fio vde.cfg" to checkt the performance After reproduce.sh finished, performance down. Anyone has the same problem or has some ideas about this? Thanks! -------------- hzwulibin 2015-11-02 _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com