---
mm/vmscan.c | 39 +++++++++++++++++++++------------------
1 file changed, 21 insertions(+), 18 deletions(-)
diff --git a/mm/vmscan.c b/mm/vmscan.c
index dca623d..c795fc3 100644
--- a/mm/vmscan.c
+++ b/mm/vmscan.c
@@ -238,7 +238,7 @@ static void unregister_memcg_shrinker(struct shrinker *shrinker)
up_write(&shrinker_rwsem);
}
-static bool cgroup_reclaim(struct scan_control *sc)
+static struct mem_cgroup *cgroup_reclaim(struct scan_control *sc)
{
return sc->target_mem_cgroup;
}
@@ -276,9 +276,9 @@ static void unregister_memcg_shrinker(struct shrinker *shrinker)
{
}
-static bool cgroup_reclaim(struct scan_control *sc)
+static struct mem_cgroup *cgroup_reclaim(struct scan_control *sc)
{
- return false;
+ return NULL;
}
static bool writeback_throttling_sane(struct scan_control *sc)
@@ -984,7 +984,7 @@ static enum page_references page_check_references(struct page *page,
int referenced_ptes, referenced_page;
unsigned long vm_flags;
- referenced_ptes = page_referenced(page, 1, sc->target_mem_cgroup,
+ referenced_ptes = page_referenced(page, 1, cgroup_reclaim(sc),
&vm_flags);
referenced_page = TestClearPageReferenced(page);
@@ -1422,7 +1422,7 @@ static unsigned long shrink_page_list(struct list_head *page_list,
count_vm_event(PGLAZYFREED);
count_memcg_page_event(page, PGLAZYFREED);
} else if (!mapping || !__remove_mapping(mapping, page, true,
- sc->target_mem_cgroup))
+ cgroup_reclaim(sc)))
goto keep_locked;
unlock_page(page);
@@ -1907,6 +1907,7 @@ static int current_may_throttle(void)
enum vm_event_item item;
struct pglist_data *pgdat = lruvec_pgdat(lruvec);
struct zone_reclaim_stat *reclaim_stat = &lruvec->reclaim_stat;
+ struct mem_cgroup *target_memcg = cgroup_reclaim(sc);
bool stalled = false;
while (unlikely(too_many_isolated(pgdat, file, sc))) {
@@ -1933,7 +1934,7 @@ static int current_may_throttle(void)
reclaim_stat->recent_scanned[file] += nr_taken;
item = current_is_kswapd() ? PGSCAN_KSWAPD : PGSCAN_DIRECT;
- if (!cgroup_reclaim(sc))
+ if (!target_memcg)
__count_vm_events(item, nr_scanned);
__count_memcg_events(lruvec_memcg(lruvec), item, nr_scanned);
spin_unlock_irq(&pgdat->lru_lock);
@@ -1947,7 +1948,7 @@ static int current_may_throttle(void)
spin_lock_irq(&pgdat->lru_lock);
item = current_is_kswapd() ? PGSTEAL_KSWAPD : PGSTEAL_DIRECT;
- if (!cgroup_reclaim(sc))
+ if (!target_memcg)
__count_vm_events(item, nr_reclaimed);
__count_memcg_events(lruvec_memcg(lruvec), item, nr_reclaimed);
reclaim_stat->recent_rotated[0] += stat.nr_activate[0];
@@ -2041,7 +2042,7 @@ static void shrink_active_list(unsigned long nr_to_scan,
}
}
- if (page_referenced(page, 0, sc->target_mem_cgroup,
+ if (page_referenced(page, 0, cgroup_reclaim(sc),
&vm_flags)) {
nr_rotated += hpage_nr_pages(page);
/*
@@ -2625,7 +2626,7 @@ static inline bool should_continue_reclaim(struct pglist_data *pgdat,
static void shrink_node_memcgs(pg_data_t *pgdat, struct scan_control *sc)
{
- struct mem_cgroup *target_memcg = sc->target_mem_cgroup;
+ struct mem_cgroup *target_memcg = cgroup_reclaim(sc);
struct mem_cgroup *memcg;
memcg = mem_cgroup_iter(target_memcg, NULL, NULL);
@@ -2686,10 +2687,11 @@ static void shrink_node(pg_data_t *pgdat, struct scan_control *sc)
struct reclaim_state *reclaim_state = current->reclaim_state;
unsigned long nr_reclaimed, nr_scanned;
struct lruvec *target_lruvec;
+ struct mem_cgroup *target_memcg = cgroup_reclaim(sc);
bool reclaimable = false;
unsigned long file;
- target_lruvec = mem_cgroup_lruvec(sc->target_mem_cgroup, pgdat);
+ target_lruvec = mem_cgroup_lruvec(target_memcg, pgdat);
again:
memset(&sc->nr, 0, sizeof(sc->nr));
@@ -2744,7 +2746,7 @@ static void shrink_node(pg_data_t *pgdat, struct scan_control *sc)
* thrashing file LRU becomes infinitely more attractive than
* anon pages. Try to detect this based on file LRU size.
*/
- if (!cgroup_reclaim(sc)) {
+ if (!target_memcg) {
unsigned long total_high_wmark = 0;
unsigned long free, anon;
int z;
@@ -2782,7 +2784,7 @@ static void shrink_node(pg_data_t *pgdat, struct scan_control *sc)
}
/* Record the subtree's reclaim efficiency */
- vmpressure(sc->gfp_mask, sc->target_mem_cgroup, true,
+ vmpressure(sc->gfp_mask, target_memcg, true,
sc->nr_scanned - nr_scanned,
sc->nr_reclaimed - nr_reclaimed);
@@ -2833,7 +2835,7 @@ static void shrink_node(pg_data_t *pgdat, struct scan_control *sc)
* stalling in wait_iff_congested().
*/
if ((current_is_kswapd() ||
- (cgroup_reclaim(sc) && writeback_throttling_sane(sc))) &&
+ (target_memcg && writeback_throttling_sane(sc))) &&
sc->nr.dirty && sc->nr.dirty == sc->nr.congested)
set_bit(LRUVEC_CONGESTED, &target_lruvec->flags);
@@ -3020,14 +3022,15 @@ static unsigned long do_try_to_free_pages(struct zonelist *zonelist,
pg_data_t *last_pgdat;
struct zoneref *z;
struct zone *zone;
+ struct mem_cgroup *target_memcg = cgroup_reclaim(sc);
retry:
delayacct_freepages_start();
- if (!cgroup_reclaim(sc))
+ if (!target_memcg)
__count_zid_vm_events(ALLOCSTALL, sc->reclaim_idx, 1);
do {
- vmpressure_prio(sc->gfp_mask, sc->target_mem_cgroup,
+ vmpressure_prio(sc->gfp_mask, target_memcg,
sc->priority);
sc->nr_scanned = 0;
shrink_zones(zonelist, sc);
@@ -3053,12 +3056,12 @@ static unsigned long do_try_to_free_pages(struct zonelist *zonelist,
continue;
last_pgdat = zone->zone_pgdat;
- snapshot_refaults(sc->target_mem_cgroup, zone->zone_pgdat);
+ snapshot_refaults(target_memcg, zone->zone_pgdat);
- if (cgroup_reclaim(sc)) {
+ if (target_memcg) {
struct lruvec *lruvec;
- lruvec = mem_cgroup_lruvec(sc->target_mem_cgroup,
+ lruvec = mem_cgroup_lruvec(target_memcg,
zone->zone_pgdat);
clear_bit(LRUVEC_CONGESTED, &lruvec->flags);
}
--
1.9.1