It is CephFS mounted via ceph-fuse. I am getting the same results not depending on how many other clients are having this fs mounted and their activity. Cluster is working on Debian Wheezy, kernel 3.2.0-4-amd64. On Tue, Oct 21, 2014 at 1:44 PM, Gregory Farnum <greg@xxxxxxxxxxx> wrote: > Are these tests conducted using a local fs on RBD, or using CephFS? > If CephFS, do you have multiple clients mounting the FS, and what are > they doing? What client (kernel or ceph-fuse)? > -Greg > Software Engineer #42 @ http://inktank.com | http://ceph.com > > > On Tue, Oct 21, 2014 at 9:05 AM, Sergey Nazarov <natarajaya@xxxxxxxxx> wrote: >> Hi >> >> I just built a new cluster using this quickstart instructions: >> http://ceph.com/docs/master/start/ >> >> And here is what I am seeing: >> >> # time for i in {1..10}; do echo $i > $i.txt ; done >> real 0m0.081s >> user 0m0.000s >> sys 0m0.004s >> >> And if I try to repeat the same command (when files already created): >> >> # time for i in {1..10}; do echo $i > $i.txt ; done >> real 0m48.894s >> user 0m0.000s >> sys 0m0.004s >> >> I was very surprised and then just tried to rewrite a single file: >> >> # time echo 1 > 1.txt >> real 0m3.133s >> user 0m0.000s >> sys 0m0.000s >> >> BTW, I dont think it is the problem with OSD speed or network: >> >> # time sysbench --num-threads=1 --test=fileio --file-total-size=1G >> --file-test-mode=rndrw prepare >> 1073741824 bytes written in 23.52 seconds (43.54 MB/sec). >> >> Here is my ceph cluster status and verion: >> >> # ceph -w >> cluster d3dcacc3-89fb-4db0-9fa9-f1f6217280cb >> health HEALTH_OK >> monmap e4: 4 mons at >> {atl-fs10=10.44.101.70:6789/0,atl-fs11=10.44.101.91:6789/0,atl-fs12=10.44.101.92:6789/0,atl-fs9=10.44.101.69:6789/0}, >> election epoch 40, quorum 0,1,2,3 atl-fs9,atl-fs10,atl-fs11,atl-fs12 >> mdsmap e33: 1/1/1 up {0=atl-fs12=up:active}, 3 up:standby >> osdmap e92: 4 osds: 4 up, 4 in >> pgmap v8091: 192 pgs, 3 pools, 123 MB data, 1658 objects >> 881 GB used, 1683 GB / 2564 GB avail >> 192 active+clean >> client io 1820 B/s wr, 1 op/s >> >> # ceph -v >> ceph version 0.80.7 (6c0127fcb58008793d3c8b62d925bc91963672a3) >> >> All nodes connected with gigabit network. >> _______________________________________________ >> ceph-users mailing list >> ceph-users@xxxxxxxxxxxxxx >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com