Hello Cephers,
I am using CephFS, and running some benchmarks using fio.
After increasing the object_size to 33554432, when I try to run some read and write tests with different block sizes, when I get to block size of 64m and beyond, Ceph does not finish the operation (I tried letting it run for more than a day at least three times).
However, when I cancel the job and I expect to see no io operations, here is what I get:
[cephuser@node01 ~]$ ceph -s
cluster b7beebf6-ea9f-4560-a916-a58e106c6e8e
health HEALTH_OK
monmap e3: 3 mons at {node02=192.168.17.212:6789/0,node03=192.168.17.213:6789/0,node04=192.168.17.214:6789/0}
election epoch 8, quorum 0,1,2 node02,node03,node04
mdsmap e74: 1/1/1 up {0=node02=up:active}
osdmap e324: 14 osds: 14 up, 14 in
pgmap v155699: 768 pgs, 3 pools, 15285 MB data, 1772 objects
91283 MB used, 7700 GB / 7817 GB avail
768 active+clean
client io 2911 MB/s rd, 90 op/s
If I do ceph -w, it shows me that it is constantly doing reads, but I have no idea from where and when it would stop?
I had to remove my CephFS file system and the associated pools and start things from scratch.
1. Any idea what is happening?
2. When this happens, do you know a better way to get out of the situation without destroying the filesystem and the pools?
Thanks,
Hadi
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com