Re: Unexpected period of iowait, no obvious activity?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Fri, Jun 19, 2015 at 7:50 PM, Daniel Schneller
<daniel.schneller@xxxxxxxxxxxxxxxx> wrote:
> Hi!
>
> Recently over a few hours our 4 Ceph disk nodes showed unusually high
> and somewhat constant iowait times. Cluster runs 0.94.1 on Ubuntu
> 14.04.1.
>
> It started on one node, then - with maybe 15 minutes delay each - on the
> next and the next one. Overall duration of the phenomenon was about 90
> minutes on each machine, finishing in the same order they had started.
>
> We could not see any obvious cluster activity during that time,
> applications did not do anything out of the ordinary. Scrubbing and deep
> scrubbing were turned off long before this happened.
>
> We are using CephFS for shared administrator home directories on the
> system, RBD volumes for OpenStack and the Rados Gateway to manage
> application data via the Swift interface. Telemetry and logs from inside
> the VMs did not offer an explanation either.
>
> The fact that these readings were limited to OSD hosts, but none of the
> other (client) nodes in the system, suggests this must be some kind of
> Ceph behaviour. Any ideas? We would like to understand what the system
> was doing, but haven't found anything obvious in the logs.

Not necessarily — if they were truly offset it's unlikely Ceph did
that: everything Ceph does that involves changing read/write rates
involves doing it across multiple nodes. Writes from clients are
shared; reads only go to primaries but will not be so strongly
segregated across machines; there's nothing else but scrubbing and
recovery...

On the other hand, there are lots of administrative tasks that can run
and do something like this. The CERN guys had a lot of trouble with
some daemon which wanted to scan the OSD's entire store for tracking
changes, and was installed by their standard Ubuntu deployment. You
probably want to look for something like that, and try and figure out
which processes are doing all the IO during the time things get slow.
-Greg
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux