On Mon, Aug 25, 2008 at 01:27:19PM +0200, Peter Zijlstra wrote: > On Mon, 2008-08-25 at 16:50 +0530, Aneesh Kumar K.V wrote: > > fbc->count is of type s64. The change was introduced by > > 0216bfcffe424a5473daa4da47440881b36c1f4 which changed the type > > from long to s64. Moving to s64 also means on 32 bit architectures > > we can get wrong values on fbc->count. Since fbc->count is read > > more frequently and updated rarely use seqlocks. This should > > reduce the impact of locking in the read path for 32bit arch. > > > > percpu_counter_read is used within interrupt context also. So > > use the irq safe version of seqlock while reading > > > > Signed-off-by: Aneesh Kumar K.V <aneesh.kumar@xxxxxxxxxxxxxxxxxx> > > CC: Peter Zijlstra <a.p.zijlstra@xxxxxxxxx> > > CC: Andrew Morton <akpm@xxxxxxxxxxxxxxxxxxxx> > > --- > > include/linux/percpu_counter.h | 29 +++++++++++++++++++++++++---- > > lib/percpu_counter.c | 20 ++++++++++---------- > > 2 files changed, 35 insertions(+), 14 deletions(-) > > > > diff --git a/include/linux/percpu_counter.h b/include/linux/percpu_counter.h > > index 9007ccd..36f3d2d 100644 > > --- a/include/linux/percpu_counter.h > > +++ b/include/linux/percpu_counter.h > > @@ -6,7 +6,7 @@ > > * WARNING: these things are HUGE. 4 kbytes per counter on 32-way P4. > > */ > > > > -#include <linux/spinlock.h> > > +#include <linux/seqlock.h> > > #include <linux/smp.h> > > #include <linux/list.h> > > #include <linux/threads.h> > > @@ -16,7 +16,7 @@ > > #ifdef CONFIG_SMP > > > > struct percpu_counter { > > - spinlock_t lock; > > + seqlock_t lock; > > s64 count; > > #ifdef CONFIG_HOTPLUG_CPU > > struct list_head list; /* All percpu_counters are on a list */ > > @@ -53,10 +53,31 @@ static inline s64 percpu_counter_sum(struct percpu_counter *fbc) > > return __percpu_counter_sum(fbc); > > } > > > > -static inline s64 percpu_counter_read(struct percpu_counter *fbc) > > +#if BITS_PER_LONG == 64 > > +static inline s64 fbc_count(struct percpu_counter *fbc) > > { > > return fbc->count; > > } > > +#else > > +/* doesn't have atomic 64 bit operation */ > > +static inline s64 fbc_count(struct percpu_counter *fbc) > > +{ > > + s64 ret; > > + unsigned seq; > > + unsigned long flags; > > + do { > > + seq = read_seqbegin_irqsave(&fbc->lock, flags); > > + ret = fbc->count; > > + } while(read_seqretry_irqrestore(&fbc->lock, seq, flags)); > > Do we really need to disabled IRQs here? It seems to me the worst that > can happen is that the IRQ will change ->count and increase the sequence > number a bit - a case that is perfectly handled by the current retry > logic. > > And not doing the IRQ flags bit saves a lot of time on some archs. > Will update in the next version. BTW does it make sense to do the above unconditionally now ? ie to remove the #if ?. How much impact would it be to do read_seqbegin and read_seqretry on a 64bit machine too ? -aneesh -- To unsubscribe from this list: send the line "unsubscribe linux-ext4" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html