Hello, On Fri, Apr 01, 2016 at 11:09:37PM -0400, Waiman Long wrote: ... > +struct percpu_stats { > + unsigned long __percpu *stats; I'm not sure ulong is the best choice here. Atomic reads on 32bit are nice but people often need 64bit counters for stats. It probably is a better idea to use u64_stats_sync. > +/* > + * Reset the all statistics counts to 0 in the percpu_stats structure Proper function description please. > + */ > +static inline void percpu_stats_reset(struct percpu_stats *pcs) Why is this function inline? > +{ > + int cpu; > + > + for_each_possible_cpu(cpu) { > + unsigned long *pstats = per_cpu_ptr(pcs->stats, cpu); ^^ > + int stat; > + > + for (stat = 0; stat < pcs->nstats; stat++, pstats++) > + *pstats = 0; > + } > + > + /* > + * If a statistics count is in the middle of being updated, it > + * is possible that the above clearing may not work. So we need > + * to double check again to make sure that the counters are really > + * cleared. Still there is a still a very small chance that the > + * second clearing does not work. > + */ > + for_each_possible_cpu(cpu) { > + unsigned long *pstats = per_cpu_ptr(pcs->stats, cpu); > + int stat; > + > + for (stat = 0; stat < pcs->nstats; stat++, pstats++) > + if (*pstats) > + *pstats = 0; > + } I don't think this is acceptable. > +} > + > +static inline int percpu_stats_init(struct percpu_stats *pcs, int num) > +{ > + pcs->nstats = num; > + pcs->stats = __alloc_percpu(sizeof(unsigned long) * num, > + __alignof__(unsigned long)); > + if (!pcs->stats) > + return -ENOMEM; > + > + percpu_stats_reset(pcs); > + return 0; > +} > + > +static inline void percpu_stats_destroy(struct percpu_stats *pcs) > +{ > + free_percpu(pcs->stats); > + pcs->stats = NULL; > + pcs->nstats = 0; > +} Why inline the above functions? > +static inline void > +__percpu_stats_add(struct percpu_stats *pcs, int stat, int cnt) > +{ > + unsigned long *pstat; > + > + if ((unsigned int)stat >= pcs->nstats) > + return; This is a critical bug. Please don't fail silently. BUG_ON(), please. > + preempt_disable(); > + pstat = this_cpu_ptr(&pcs->stats[stat]); > + *pstat += cnt; > + preempt_enable(); > +} this_cpu_add() is atomic w.r.t. local operations. > +static inline unsigned long > +percpu_stats_sum(struct percpu_stats *pcs, int stat) > +{ > + int cpu; > + unsigned long sum = 0; > + > + if ((unsigned int)stat >= pcs->nstats) > + return sum; Ditto. > + for_each_possible_cpu(cpu) > + sum += per_cpu(pcs->stats[stat], cpu); > + return sum; > +} Thanks. -- tejun -- To unsubscribe from this list: send the line "unsubscribe linux-ext4" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html