Re: rbd top

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Thu, Jun 11, 2015 at 12:33 PM, Robert LeBlanc <robert@xxxxxxxxxxxxx> wrote:
> -----BEGIN PGP SIGNED MESSAGE-----
> Hash: SHA256
>
> One feature we would like is an "rbd top" command that would be like
> top, but show usage of RBD volumes so that we can quickly identify
> high demand RBDs.
>
> Since I haven't done any programming for Ceph, I'm trying to think
> through the best way to approach this. I don't know if there are
> already perf counters that I can query that are at the client, RBD or
> the Rados layers. If these counters don't exist would it be best to
> implement them at the client layer and look for watchers on the RBD
> and query them? Is it better to handle it at the Rados layer and
> aggregate the I/O from all chunks? Of course this would need to scale
> out very large.
>
> It seems that if the client running rbd top requests the top 'X'
> number of objects from each OSD, then it would cut down on the data
> that the has to be moved around and processed. It wouldn't be an
> extremely accurate view, but might be enough.
>
> What are your thoughts?
>
> Also, what is the best way to get into the Ceph code? I've looked at
> several things and I find myself doing a lot of searching to find
> connecting pieces. My primary focus is not programming so picking up a
> new code base takes me a long time because I don't know many of the
> tricks that help people get to speed quickly.

The basic problem with a tool like this is that it requires gathering
real-time data from either all the OSDs, or all the clients. We do
something similar in order to display approximate IO going through the
system as a whole, but that is based on PGStat messages which come in
periodically and is both laggy and an approximation.

To do this, we'd need to get less-laggy data, and instead of scaling
with the number of OSDs/PGs it would scale with the number of RBD
volumes. You certainly couldn't send that through the monitor and I
shudder to think about the extra load it would invoke at all layers.

How up-to-date do you need the info to be, and how accurate? Does it
need to be queryable in the future or only online? You could perhaps
hook into one of the more precise HitSet implementations we
have...otherwise I think you'd need to add an online querying
framework, perhaps through the perfcounters (which...might scale to
something like this?) or a monitoring service (hopefully attached to
Calamari) that receives continuous updates.
-Greg
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux