On Wed, May 11, 2022 at 03:06:56PM -0400, Johannes Weiner <hannes@xxxxxxxxxxx> wrote: > Correct. After which the uncompressed page is reclaimed and uncharged. > So the zswapout process will reduce the charge bottom line. A zswap object falling under memory.current was my first thinking, I was confused why it's exported as a separate counter memory.zswap.current (which IMO suggests disjoint counting) and it doubles a memory.stat:zswap entry. Is the separate memory.zswap.current good for anything? (Except maybe avoiding global rstat flush on memory.stat read but that'd be an undesired precendent.) (Ad the eventually reduced footprint, the transitional excursion above memcg's (or ancestor's) limit should be limited by number of parallel reclaims running (each one at most a page, right?), so it doesn't seem necessary to tackle (now).) > memory.zswap.* are there to configure zswap policy, within the > boundaries of available memory - it's by definition a subset. I see how the .max works when equal to 0 or "max". The intermediate values are more difficult to reason about. Also, I can see that on the global level, zswap is configured relatively (/sys/module/zswap/parameters/max_pool_percent). You wrote that the actual configured value is workload specific, would it be simpler to have also relative zswap limit per memcg? (Relative wrt memory.max, it'd be rather just a convenience with this simple ratio, however, it'd correspond to the top level limit. OTOH, the relatives would have counter-intuitive hierarchical behavior. I don't mean this should be changed, rather wondering why this variant was chosen.) > +bool obj_cgroup_may_zswap(struct obj_cgroup *objcg) > +{ > + struct mem_cgroup *memcg, *original_memcg; > + bool ret = true; > + > + original_memcg = get_mem_cgroup_from_objcg(objcg); > + for (memcg = original_memcg; memcg != root_mem_cgroup; > + memcg = parent_mem_cgroup(memcg)) { > + unsigned long max = READ_ONCE(memcg->zswap_max); > + unsigned long pages; > + > + if (max == PAGE_COUNTER_MAX) > + continue; > + if (max == 0) { > + ret = false; > + break; > + } > + > + cgroup_rstat_flush(memcg->css.cgroup); Here, I think it'd be better not to bypass mem_cgroup_flush_stats() (the mechanism is approximate and you traverse all ancestors anyway), i.e. mem_cgroup_flush_stats() before the loop instead of this. Thanks, Michal