On Thu 19-07-18 10:33:29, Dave Chinner wrote: > On Tue, Jul 17, 2018 at 10:33:26AM +0200, Michal Hocko wrote: > > On Mon 16-07-18 16:40:32, Andrew Morton wrote: > > > On Mon, 16 Jul 2018 05:41:15 -0700 Matthew Wilcox <willy@xxxxxxxxxxxxx> wrote: > > > It's quite a small code change and would provide a mechanism for > > > implementing the hammer-cache-until-youve-freed-enough design above. > > > > > > > > > > > > Aside 2: if we *do* do something like the above __d_alloc() pseudo code > > > then perhaps it could be cast in terms of pages, not dentries. ie, > > > > > > __d_alloc() > > > { > > > ... > > > while (too many pages in dentry_cache) > > > call the dcache shrinker > > > ... > > > } > > Direct reclaim will result in all the people who care about long > tail latencies and/or highly concurent workloads starting to hate > you. Direct reclaim already hammers superblock shrinkers with > excessive concurrency, this would only make it worse. I can only confirm that! We have something similar in our SLES kernel. We have page cache soft limit implemented for many years and it is basically similar thing to above. We just shrink the page cache when we have too much of it. It turned out to be a complete PITA on large machines when hundreds of CPUs are fighting for locks. We have tried to address that but it is a complete whack a mole. More important lesson from this is that the original motivation for this functionality was to not allow too much page cache which would push a useful DB data out to swap. And as it turned out MM internals have changed a lot since the introduction and we do not really swap out in presence of the page cache anymore. Moreover we have a much more effective reclaim protection thanks to memcg low limit reclaim etc. While that is all good and nice there are still people tunning the pagecache limit based on some really old admin guides and the feature makes more harm than good and we see bug reports that system gets stalled... I really do not see why limiting (negative) dentries should be any different. > IOWs, anything like this needs to co-ordinate with other reclaim > operations in progress and, most likely, be done via background > reclaim processing rather than blocking new allocations > indefinitely. background processing can be done in bulk and as > efficiently as possible - concurrent direct reclaim in tiny batches > will just hammer dcache locks and destroy performance when there is > memory pressure. Absolutely agreed! > How many times do we have to learn this lesson the hard way? > > > > and, apart from the external name thing (grr), that should address > > > these fragmentation issues, no? I assume it's easy to ask slab how > > > many pages are presently in use for a particular cache. > > > > I remember Dave Chinner had an idea how to age dcache pages to push > > dentries with similar live time to the same page. Not sure what happened > > to that. > > Same thing that happened to all the "select the dentries on this > page for reclaim". i.e. it's referenced dentries that we can't > reclaim or move that are the issue, not the reclaimable dentries on > the page. > > Bsaically, without a hint at allocation time as to the expected life > time of the dentry, we can't be smart about how we select partial > pages to allocate from. And because we don't know at allocation time > if the dentry is going to remain a negative dentry or not, we can't > provide a hint about expected lifetime of teh object being > allocated. Can we allocate a new dentry at the time when we know the life time or the dentry pointer is so spread by that time that we cannot? -- Michal Hocko SUSE Labs