On 03.07.2018 20:32, Kirill Tkhai wrote: > On 03.07.2018 20:00, Shakeel Butt wrote: >> On Tue, Jul 3, 2018 at 9:17 AM Kirill Tkhai <ktkhai@xxxxxxxxxxxxx> wrote: >>> >>> Hi, Shakeel, >>> >>> On 03.07.2018 18:46, Shakeel Butt wrote: >>>> On Tue, Jul 3, 2018 at 8:27 AM Matthew Wilcox <willy@xxxxxxxxxxxxx> wrote: >>>>> >>>>> On Tue, Jul 03, 2018 at 06:09:05PM +0300, Kirill Tkhai wrote: >>>>>> +++ b/mm/vmscan.c >>>>>> @@ -169,6 +169,49 @@ unsigned long vm_total_pages; >>>>>> static LIST_HEAD(shrinker_list); >>>>>> static DECLARE_RWSEM(shrinker_rwsem); >>>>>> >>>>>> +#ifdef CONFIG_MEMCG_KMEM >>>>>> +static DEFINE_IDR(shrinker_idr); >>>>>> +static int shrinker_nr_max; >>>>> >>>>> So ... we've now got a list_head (shrinker_list) which contains all of >>>>> the shrinkers, plus a shrinker_idr which contains the memcg-aware shrinkers? >>>>> >>>>> Why not replace the shrinker_list with the shrinker_idr? It's only used >>>>> twice in vmscan.c: >>>>> >>>>> void register_shrinker_prepared(struct shrinker *shrinker) >>>>> { >>>>> down_write(&shrinker_rwsem); >>>>> list_add_tail(&shrinker->list, &shrinker_list); >>>>> up_write(&shrinker_rwsem); >>>>> } >>>>> >>>>> list_for_each_entry(shrinker, &shrinker_list, list) { >>>>> ... >>>>> >>>>> The first is simply idr_alloc() and the second is >>>>> >>>>> idr_for_each_entry(&shrinker_idr, shrinker, id) { >>>>> >>>>> I understand there's a difference between allocating the shrinker's ID and >>>>> adding it to the list. You can do this by calling idr_alloc with NULL >>>>> as the pointer, and then using idr_replace() when you want to add the >>>>> shrinker to the list. idr_for_each_entry() skips over NULL entries. >>>>> >>>>> This will actually reduce the size of each shrinker and be more >>>>> cache-efficient when calling the shrinkers. I think we can also get >>>>> rid of the shrinker_rwsem eventually, but let's leave it for now. >>>> >>>> Can you explain how you envision shrinker_rwsem can be removed? I am >>>> very much interested in doing that. >>> >>> Have you tried to do some games with SRCU? It looks like we just need to >>> teach count_objects() and scan_objects() to work with semi-destructed >>> shrinkers. Though, this looks this will make impossible to introduce >>> shrinkers, which do synchronize_srcu() in scan_objects() for example. >>> Not sure, someone will actually use this, and this is possible to consider >>> as limitation. >>> >> >> Hi Kirill, I tried SRCU and the discussion is at >> https://lore.kernel.org/lkml/20171117173521.GA21692@xxxxxxxxxxxxx/T/#u >> >> Paul E. McKenney suggested to enable SRCU unconditionally. So, to use >> SRCU for shrinkers, we first have to push unconditional SRCU. > > First time, I read this, I though the talk goes about some new srcu_read_lock() > without an argument and it's need to rework SRCU in some huge way. Thanks > god, it was just a misreading :) >> Tetsuo had another lockless solution which was a bit involved but does >> not depend on SRCU. > > Ok, I see refcounters suggestion. Thanks for the link, Shakeel! Just returning to this theme. Since both of the suggested ways contain srcu synchronization, it may be better just to use percpu-rwsem, since there is the same functionality out-of-box. register/unregister_shrinker() will use two rw semaphores: register_shrinker() { down_write(&shrinker_rwsem); idr_alloc(); up_write(&shrinker_rwsem); } unregister_shrinker() { percpu_down_write(&percpu_shrinker_rwsem); down_write(&shrinker_rwsem); idr_remove(); up_write(&shrinker_rwsem); percpu_up_write(&percpu_shrinker_rwsem); } shrink_slab() { percpu_down_read(&percpu_shrinker_rwsem); rcu_read_lock(); shrinker = idr_find(); rcu_read_unlock(); do_shrink_slab(shrinker); percpu_up_read(&percpu_shrinker_rwsem); } 1)Here is a trick to make register_shrinker() not use percpu semaphore, i.e., not to wait RCU synchronization. This just makes register_shrinker() faster. So, we introduce 2 semaphores instead of 1: shrinker_rwsem to protect IDR and percpu_shrinker_rwsem. 2)rcu_read_lock() -- to synchronize idr_find() with idr_alloc(). Not sure, we really need this. It's possible, lockless idr_find() is OK in parallel with allocation of new ID. Parallel removing is not possible because of percpu rwsem. 3)Places, which are performance critical to unregister_shrinker() speed (e.g., like deactivate_locked_super(), as we want umount() to be fast), may just call it delayed from work: diff --git a/fs/super.c b/fs/super.c index 13647d4fd262..b4a98cb00166 100644 --- a/fs/super.c +++ b/fs/super.c @@ -324,19 +324,7 @@ void deactivate_locked_super(struct super_block *s) struct file_system_type *fs = s->s_type; if (atomic_dec_and_test(&s->s_active)) { cleancache_invalidate_fs(s); - unregister_shrinker(&s->s_shrink); - fs->kill_sb(s); - - /* - * Since list_lru_destroy() may sleep, we cannot call it from - * put_super(), where we hold the sb_lock. Therefore we destroy - * the lru lists right now. - */ - list_lru_destroy(&s->s_dentry_lru); - list_lru_destroy(&s->s_inode_lru); - - put_filesystem(fs); - put_super(s); + schedule_delayed_deactivate_super(s) } else { up_write(&s->s_umount); } Kirill