On Sat, Apr 27, 2013 at 03:19:07AM +0400, Glauber Costa wrote: > From: Dave Chinner <dchinner@xxxxxxxxxx> > > Now that we have an LRU list API, we can start to enhance the > implementation. This splits the single LRU list into per-node lists > and locks to enhance scalability. Items are placed on lists > according to the node the memory belongs to. To make scanning the > lists efficient, also track whether the per-node lists have entries > in them in a active nodemask. > > [ glommer: fixed warnings ] > Signed-off-by: Dave Chinner <dchinner@xxxxxxxxxx> > Signed-off-by: Glauber Costa <glommer@xxxxxxxxxx> > Reviewed-by: Greg Thelen <gthelen@xxxxxxxxxx> > --- > include/linux/list_lru.h | 14 ++-- > lib/list_lru.c | 162 +++++++++++++++++++++++++++++++++++------------ > 2 files changed, 130 insertions(+), 46 deletions(-) > > diff --git a/include/linux/list_lru.h b/include/linux/list_lru.h > index c0b796d..c422782 100644 > --- a/include/linux/list_lru.h > +++ b/include/linux/list_lru.h > @@ -8,6 +8,7 @@ > #define _LRU_LIST_H > > #include <linux/list.h> > +#include <linux/nodemask.h> > > enum lru_status { > LRU_REMOVED, /* item removed from list */ > @@ -17,20 +18,21 @@ enum lru_status { > internally, but has to return locked. */ > }; > > -struct list_lru { > +struct list_lru_node { > spinlock_t lock; > struct list_head list; > long nr_items; > +} ____cacheline_aligned_in_smp; > + > +struct list_lru { > + struct list_lru_node node[MAX_NUMNODES]; > + nodemask_t active_nodes; > }; struct list_lru is going to be large. 64K just for the list_lru_nodes on a distribution configuration that has NODES_SHIFT==10. On most machines it'll be mostly unused space. How big is super_block now with two of these things? xfs_buftarg? They are rarely allocated structures but it would be a little embarassing if we failed to mount a usb stick because kmalloc() of some large buffer failed on a laptop. You may need to convert "list_lru_node node" to be an array of MAX_NUMNODES pointers to list_lru_nodes. It'd need a lookup helper for list_lru_add and list_lru_del that lazily allocates the list_lru_nodes on first usage in case of node hot-add. You could allocate the online nodes at list_lru_init. It'd be awkward but avoid the need for a large kmalloc at runtime just because someone plugged in a USB stick. Otherwise I didn't spot a major problem. There are now per-node lists to walk but the overall size of the LRU for walkers should be similar and the additional overhead in list_lru_count is hardly going to be noticable. I liked the use of active_mask. -- Mel Gorman SUSE Labs -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxx. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>