On Thu, 6 Jun 2013 20:15:07 +0400 Glauber Costa <glommer@xxxxxxxxxxxxx> wrote: > On 06/06/2013 03:08 AM, Andrew Morton wrote: > >> + for_each_node_mask(nid, lru->active_nodes) { > >> > + struct list_lru_node *nlru = &lru->node[nid]; > >> > + > >> > + spin_lock(&nlru->lock); > >> > + BUG_ON(nlru->nr_items < 0); > > This is buggy. > > > > The bit in lru->active_nodes could be cleared by now. We can only make > > this assertion if we recheck lru->active_nodes[nid] inside the > > spinlocked region. > > > Sorry Andrew, how so ? > We will clear that flag if nr_items == 0. nr_items should *never* get to > be less than 0, it doesn't matter if the node is cleared or not. > > If the node is cleared, we would expected the following statement to > expand to > count += nlru->nr_items = 0; > spin_unlock(&nlru->lock); > > Which is actually cheaper than testing for the bit being still set. Well OK - I didn't actually look at the expression the BUG_ON() was testing. You got lucky ;) My point was that nlru->lock protects ->active_nodes and so the above code is racy due to a locking error. I now see that was incorrect - active_nodes has no locking. Well, it kinda has accidental locking - nrlu->lock happens to protect this nrlu's bit in active_nodes while permitting other nrlu's bits to concurrently change. The bottom line is that code which does if (node_isset(n, active_nodes)) use(n); can end up using a node which is no longer in the active_nodes, because there is no locking. This is a bit weird and worrisome and might lead to bugs in the future, at least. Perhaps we can improve the maintainability by documenting this at the active_nodes site, dunno. This code gets changed a lot in later patches and I didn't check to see if the problem remains in the final product. -- To unsubscribe from this list: send the line "unsubscribe linux-fsdevel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html