On Fri, Mar 23, 2018 at 06:20:28PM +0300, Andrey Ryabinin wrote: > We have separate LRU list for each memory cgroup. Memory reclaim iterates > over cgroups and calls shrink_inactive_list() every inactive LRU list. > Based on the state of a single LRU shrink_inactive_list() may flag > the whole node as dirty,congested or under writeback. This is obviously > wrong and hurtful. It's especially hurtful when we have possibly > small congested cgroup in system. Than *all* direct reclaims waste time > by sleeping in wait_iff_congested(). And the more memcgs in the system > we have the longer memory allocation stall is, because > wait_iff_congested() called on each lru-list scan. > > Sum reclaim stats across all visited LRUs on node and flag node as dirty, > congested or under writeback based on that sum. Also call > congestion_wait(), wait_iff_congested() once per pgdat scan, instead of > once per lru-list scan. > > This only fixes the problem for global reclaim case. Per-cgroup reclaim > may alter global pgdat flags too, which is wrong. But that is separate > issue and will be addressed in the next patch. > > This change will not have any effect on a systems with all workload > concentrated in a single cgroup. This makes a ton of sense, and I'm going to ack the patch, but here is one issue here: > @@ -2587,6 +2554,61 @@ static bool shrink_node(pg_data_t *pgdat, struct scan_control *sc) > if (sc->nr_reclaimed - nr_reclaimed) > reclaimable = true; > > + /* > + * If reclaim is isolating dirty pages under writeback, it > + * implies that the long-lived page allocation rate is exceeding > + * the page laundering rate. Either the global limits are not > + * being effective at throttling processes due to the page > + * distribution throughout zones or there is heavy usage of a > + * slow backing device. The only option is to throttle from > + * reclaim context which is not ideal as there is no guarantee > + * the dirtying process is throttled in the same way > + * balance_dirty_pages() manages. > + * > + * Once a node is flagged PGDAT_WRITEBACK, kswapd will count the > + * number of pages under pages flagged for immediate reclaim and > + * stall if any are encountered in the nr_immediate check below. > + */ > + if (sc->nr.writeback && sc->nr.writeback == sc->nr.file_taken) > + set_bit(PGDAT_WRITEBACK, &pgdat->flags); > + > + /* > + * Legacy memcg will stall in page writeback so avoid forcibly > + * stalling here. > + */ > + if (sane_reclaim(sc)) { > + /* > + * Tag a node as congested if all the dirty pages > + * scanned were backed by a congested BDI and > + * wait_iff_congested will stall. > + */ > + if (sc->nr.dirty && sc->nr.dirty == sc->nr.congested) > + set_bit(PGDAT_CONGESTED, &pgdat->flags); > + > + /* Allow kswapd to start writing pages during reclaim.*/ > + if (sc->nr.unqueued_dirty == sc->nr.file_taken) > + set_bit(PGDAT_DIRTY, &pgdat->flags); > + > + /* > + * If kswapd scans pages marked marked for immediate > + * reclaim and under writeback (nr_immediate), it > + * implies that pages are cycling through the LRU > + * faster than they are written so also forcibly stall. > + */ > + if (sc->nr.immediate) > + congestion_wait(BLK_RW_ASYNC, HZ/10); > + } This isn't quite equivalent to what we have right now. Yes, nr_dirty, nr_unqueued_dirty and nr_congested apply to file pages only. That part is about waking the flushers and avoiding writing files in 4k chunks from reclaim context. So those numbers do need to be compared against scanned *file* pages. But nr_writeback and nr_immediate is about throttling reclaim when we hit too many pages under writeout, and that applies to both file and anonymous/swap pages. We do want to throttle on swapout, too. So nr_writeback needs to check against all nr_taken, not just file.