On Mon, 2020-03-09 at 22:36 -0400, xiubli@xxxxxxxxxx wrote: > From: Xiubo Li <xiubli@xxxxxxxxxx> > > These will be very useful help diagnose problems. > > URL: https://tracker.ceph.com/issues/44533 > Signed-off-by: Xiubo Li <xiubli@xxxxxxxxxx> > --- > > The output will be like: > > # cat /sys/kernel/debug/ceph/19e31430-fc65-4aa1-99cf-2c8eaaafd451.client4347/metrics > item total sum_lat(us) avg_lat(us) min_lat(us) max_lat(us) > ------------------------------------------------------------------------------------- > read 27 297000 11000 2000 27000 > write 16 3860000 241250 175000 263000 > metadata 3 30000 10000 2000 16000 > > item total miss hit > ------------------------------------------------- > d_lease 2 0 1 > caps 2 0 3078 > > > > fs/ceph/debugfs.c | 27 ++++++++++++++++++++------ > fs/ceph/mds_client.c | 12 ++++++++++++ > fs/ceph/metric.h | 54 +++++++++++++++++++++++++++++++++++++++++++++++++++- > 3 files changed, 86 insertions(+), 7 deletions(-) > > > diff --git a/fs/ceph/metric.h b/fs/ceph/metric.h > index faba142..9f0d050 100644 > --- a/fs/ceph/metric.h > +++ b/fs/ceph/metric.h > @@ -2,6 +2,10 @@ > #ifndef _FS_CEPH_MDS_METRIC_H > #define _FS_CEPH_MDS_METRIC_H > > +#include <linux/atomic.h> > +#include <linux/percpu.h> > +#include <linux/spinlock.h> > + > /* This is the global metrics */ > struct ceph_client_metric { > atomic64_t total_dentries; > @@ -13,12 +17,21 @@ struct ceph_client_metric { > > struct percpu_counter total_reads; > struct percpu_counter read_latency_sum; > + spinlock_t read_latency_lock; > + atomic64_t read_latency_min; > + atomic64_t read_latency_max; > > struct percpu_counter total_writes; > struct percpu_counter write_latency_sum; > + spinlock_t write_latency_lock; > + atomic64_t write_latency_min; > + atomic64_t write_latency_max; > > struct percpu_counter total_metadatas; > struct percpu_counter metadata_latency_sum; > + spinlock_t metadata_latency_lock; > + atomic64_t metadata_latency_min; > + atomic64_t metadata_latency_max; > }; > > static inline void ceph_update_cap_hit(struct ceph_client_metric *m) > @@ -36,11 +49,24 @@ static inline void ceph_update_read_latency(struct ceph_client_metric *m, > unsigned long r_end, > int rc) > { > + unsigned long lat = r_end - r_start; > + > if (rc < 0 && rc != -ENOENT && rc != -ETIMEDOUT) > return; > > percpu_counter_inc(&m->total_reads); > - percpu_counter_add(&m->read_latency_sum, r_end - r_start); > + percpu_counter_add(&m->read_latency_sum, lat); > + > + if (lat >= atomic64_read(&m->read_latency_min) && > + lat <= atomic64_read(&m->read_latency_max)) > + return; > + > + spin_lock(&m->read_latency_lock); > + if (lat < atomic64_read(&m->read_latency_min)) > + atomic64_set(&m->read_latency_min, lat); > + if (lat > atomic64_read(&m->read_latency_max)) > + atomic64_set(&m->read_latency_max, lat); > + spin_unlock(&m->read_latency_lock); > } > Looks reasonable overall. I do sort of wonder if we really need spinlocks for these though. Might it be more efficient to use cmpxchg instead? i.e.: cur = atomic64_read(&m->read_latency_min); do { old = cur; if (likely(lat >= old)) break; } while ((cur = atomic_long_cmpxchg(&m->read_latency_min, old, lat)) != old); ...another idea might be to use a seqlock and non-atomic vars. Mostly this shouldn't matter much though as we'll almost always be hitting the non-locking fastpath. I'll plan to merge this as-is unless you want to rework it. -- Jeff Layton <jlayton@xxxxxxxxxx>