Re: long blocking with writes on rbds

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



As a follow-up to this issue, I'd like to point out some other things I've noticed.

First, per suggestions posted here, I've reduced the number of pgs per pool. This results in the following ceph status:

    cluster e96e10d3-ad2b-467f-9fe4-ab5269b70206
     health HEALTH_WARN too few pgs per osd (14 < min 20)
     monmap e1: 3 mons at {a=192.168.224.4:6789/0,b=192.168.232.4:6789/0,c=192.168.240.4:6789/0}, election epoch 8, quorum 0,1,2 a,b,c
     osdmap e238: 6 osds: 6 up, 6 in
      pgmap v1107: 86 pgs, 23 pools, 2511 MB data, 801 objects
            38288 MB used, 1467 GB / 1504 GB avail
                  86 active+clean

I'm not sure if I should be concerned about the HEALTH WARN.

However, this has not helped the performance issues. I've dug deeper to try to understand what is actually happening. It's curious because there isn't much data: our pools are about 5GB, so it really shouldn't take 30 minutes to an hour to run mkfs. Here some results taken from disk analysis tools while this delay is in progress:

From pt-diskstats:

  #ts device    rd_s rd_avkb rd_mb_s rd_mrg rd_cnc   rd_rt    wr_s wr_avkb wr_mb_s wr_mrg wr_cnc   wr_rt busy in_prg    io_s  qtime stime
  1.0 rbd0       0.0     0.0     0.0     0%    0.0     0.0     0.0     0.0     0.0     0%    0.0     0.0 100%      6     0.0    0.0   0.0

From iostat:

Device:         rrqm/s   wrqm/s     r/s     w/s    rkB/s    wkB/s avgrq-sz avgqu-sz   await r_await w_await  svctm  %util
rbd0              0.00     0.03    0.03    0.04     0.13    10.73   310.78     3.31 19730.41    0.40 37704.35 7073.59  49.47

These results correspond with my experience: the device is busy, as witnessed by the "busy" column in pt-diskstats and the "await" column in iostat. But both tools also attest to the fact that there isn't much reading or writing going on.  According to pt-diskstats, there isn't any. So my question is: what is ceph doing? It clearly isn't just blocking as a result of excess I/O load, something else is going on. Can anyone please explain?

Jeff


_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux