Roman, Thanks for the review. On Mon, Jul 22, 2024 at 2:22 PM Roman Gushchin <roman.gushchin@xxxxxxxxx> wrote: > > On Mon, Jul 22, 2024 at 11:17:13AM -0400, David Finkel wrote: > > Other mechanisms for querying the peak memory usage of either a process > > or v1 memory cgroup allow for resetting the high watermark. Restore > > parity with those mechanisms, but with a less racy API. > > > > For example: > > - Any write to memory.max_usage_in_bytes in a cgroup v1 mount resets > > the high watermark. > > - writing "5" to the clear_refs pseudo-file in a processes's proc > > directory resets the peak RSS. > > > > This change is an evolution of a previous patch, which mostly copied the > > cgroup v1 behavior, however, there were concerns about races/ownership > > issues with a global reset, so instead this change makes the reset > > filedescriptor-local. > > > > Writing a specific string to the memory.peak and memory.swap.peak > > pseudo-files reset the high watermark to the current usage for > > subsequent reads through that same fd. > > > > Notably, following Johannes's suggestion, this implementation moves the > > O(fds that have written) behavior onto the fd write(2) path. Instead, on > > the page-allocation path, we simply add one additional watermark to > > conditionally bump per-hierarchy level in the page-counter. > > > > This behavior is particularly useful for work scheduling systems that > > need to track memory usage of worker processes/cgroups per-work-item. > > Since memory can't be squeezed like CPU can (the OOM-killer has > > opinions), these systems need to track the peak memory usage to compute > > system/container fullness when binpacking workitems. > > > > Most notably, Vimeo's use-case involves a system that's doing global > > binpacking across many Kubernetes pods/containers, and while we can use > > PSI for some local decisions about overload, we strive to avoid packing > > workloads too tightly in the first place. To facilitate this, we track > > the peak memory usage. However, since we run with long-lived workers (to > > amortize startup costs) we need a way to track the high watermark while > > a work-item is executing. Polling runs the risk of missing short spikes > > that last for timescales below the polling interval, and peak memory > > tracking at the cgroup level is otherwise perfect for this use-case. > > > > As this data is used to ensure that binpacked work ends up with > > sufficient headroom, this use-case mostly avoids the inaccuracies > > surrounding reclaimable memory. > > > > Suggested-by: Johannes Weiner <hannes@xxxxxxxxxxx> > > Signed-off-by: David Finkel <davidf@xxxxxxxxx> > > --- > > Documentation/admin-guide/cgroup-v2.rst | 26 +- > > include/linux/cgroup.h | 7 + > > include/linux/memcontrol.h | 5 + > > include/linux/page_counter.h | 6 + > > kernel/cgroup/cgroup-internal.h | 2 + > > kernel/cgroup/cgroup.c | 7 + > > mm/memcontrol.c | 165 ++++++++++++- > > mm/page_counter.c | 4 + > > tools/testing/selftests/cgroup/cgroup_util.c | 22 ++ > > tools/testing/selftests/cgroup/cgroup_util.h | 2 + > > .../selftests/cgroup/test_memcontrol.c | 227 +++++++++++++++++- > > 11 files changed, 448 insertions(+), 25 deletions(-) > > > > diff --git a/Documentation/admin-guide/cgroup-v2.rst b/Documentation/admin-guide/cgroup-v2.rst > > index 8fbb0519d556..10a2f919128f 100644 > > --- a/Documentation/admin-guide/cgroup-v2.rst > > +++ b/Documentation/admin-guide/cgroup-v2.rst > > @@ -1322,11 +1322,16 @@ PAGE_SIZE multiple when read back. > > reclaim induced by memory.reclaim. > > > > memory.peak > > - A read-only single value file which exists on non-root > > - cgroups. > > + A read-write single value file which exists on non-root cgroups. > > + > > + The max memory usage recorded for the cgroup and its descendants since > > + either the creation of the cgroup or the most recent reset for that fd. > > > > - The max memory usage recorded for the cgroup and its > > - descendants since the creation of the cgroup. > > + A write of the string "fd_local_reset" to this file resets it to the > > + current memory usage for subsequent reads through the same > > Hi David! > > Not a very strong preference, but with the current design, do we really expect > to have a non-local reset? If not, can we agree on a "reset" string instead > for a sake of simplicity? I put in "fd_local_reset" as a bit of a straw-man. I'm fine with changing it to "reset" since the consensus seemed to be that a global reset was problematic. > > > + file descriptor. > > + Attempts to write any other string will return EINVAL > > + (modulo leading and trailing whitespace). > > > > memory.oom.group > > A read-write single value file which exists on non-root > > @@ -1652,11 +1657,16 @@ PAGE_SIZE multiple when read back. > > Healthy workloads are not expected to reach this limit. > > > > memory.swap.peak > > - A read-only single value file which exists on non-root > > - cgroups. > > + A read-write single value file which exists on non-root cgroups. > > + > > + The max swap usage recorded for the cgroup and its descendants since > > + the creation of the cgroup or the most recent reset for that fd. > > > > - The max swap usage recorded for the cgroup and its > > - descendants since the creation of the cgroup. > > + A write of the string "fd_local_reset" to this file resets it to the > > + current memory usage for subsequent reads through the same > > + file descriptor. > > + Attempts to write any other string will return EINVAL > > + (modulo leading and trailing whitespace). > > > > memory.swap.max > > A read-write single value file which exists on non-root > > diff --git a/include/linux/cgroup.h b/include/linux/cgroup.h > > index 2150ca60394b..9bda441227ea 100644 > > --- a/include/linux/cgroup.h > > +++ b/include/linux/cgroup.h > > @@ -855,4 +855,11 @@ static inline void cgroup_bpf_put(struct cgroup *cgrp) {} > > > > struct cgroup *task_get_cgroup1(struct task_struct *tsk, int hierarchy_id); > > > > +struct memcg_peak_mem_ctx { > > + long local_watermark; > > + struct memcg_peak_mem_ctx *next, *prev; > > Please, take a look at include/linux/list.h and use it instead of > re-implementing list operations from scratch. I clearly didn't look hard enough for that. Thanks! I'll port this to use those helpers/macros. > > > +}; > > + > > +struct memcg_peak_mem_ctx *memcg_extract_peak_mem_ctx(struct kernfs_open_file *of); > > + > > #endif /* _LINUX_CGROUP_H */ > > diff --git a/include/linux/memcontrol.h b/include/linux/memcontrol.h > > index 030d34e9d117..6be7507c6fd3 100644 > > --- a/include/linux/memcontrol.h > > +++ b/include/linux/memcontrol.h > > @@ -198,6 +198,11 @@ struct mem_cgroup { > > struct page_counter kmem; /* v1 only */ > > struct page_counter tcpmem; /* v1 only */ > > > > + /* lists of memcg peak watching contexts on swap and memory */ > > + struct memcg_peak_mem_ctx *peak_memory_local_watermark_watchers; > > + struct memcg_peak_mem_ctx *peak_swap_local_watermark_watchers; > > + spinlock_t pagecounter_peak_watchers_lock; > > + > > /* Range enforcement for interrupt charges */ > > struct work_struct high_work; > > > > diff --git a/include/linux/page_counter.h b/include/linux/page_counter.h > > index 8cd858d912c4..047ceaece258 100644 > > --- a/include/linux/page_counter.h > > +++ b/include/linux/page_counter.h > > @@ -26,6 +26,7 @@ struct page_counter { > > atomic_long_t children_low_usage; > > > > unsigned long watermark; > > + unsigned long local_watermark; /* track min of fd-local resets */ > > unsigned long failcnt; > > > > /* Keep all the read most fields in a separete cacheline. */ > > @@ -81,4 +82,9 @@ static inline void page_counter_reset_watermark(struct page_counter *counter) > > counter->watermark = page_counter_read(counter); > > } > > > > +static inline void page_counter_reset_local_watermark(struct page_counter *counter) > > +{ > > + counter->local_watermark = page_counter_read(counter); > > +} > > + > > #endif /* _LINUX_PAGE_COUNTER_H */ > > diff --git a/kernel/cgroup/cgroup-internal.h b/kernel/cgroup/cgroup-internal.h > > index 520b90dd97ec..5a97ba08e976 100644 > > --- a/kernel/cgroup/cgroup-internal.h > > +++ b/kernel/cgroup/cgroup-internal.h > > @@ -81,6 +81,8 @@ struct cgroup_file_ctx { > > struct { > > struct cgroup_pidlist *pidlist; > > } procs1; > > + > > + struct memcg_peak_mem_ctx peak; > > }; > > > > /* > > diff --git a/kernel/cgroup/cgroup.c b/kernel/cgroup/cgroup.c > > index e32b6972c478..38b935ffa6cf 100644 > > --- a/kernel/cgroup/cgroup.c > > +++ b/kernel/cgroup/cgroup.c > > @@ -1964,6 +1964,13 @@ static int cgroup2_parse_param(struct fs_context *fc, struct fs_parameter *param > > return -EINVAL; > > } > > > > +struct memcg_peak_mem_ctx *memcg_extract_peak_mem_ctx(struct kernfs_open_file *of) > > +{ > > + struct cgroup_file_ctx *ctx = of->priv; > > + > > + return &ctx->peak; > > +} > > + > > static void apply_cgroup_root_flags(unsigned int root_flags) > > { > > if (current->nsproxy->cgroup_ns == &init_cgroup_ns) { > > diff --git a/mm/memcontrol.c b/mm/memcontrol.c > > index 8f2f1bb18c9c..eb6614236371 100644 > > --- a/mm/memcontrol.c > > +++ b/mm/memcontrol.c > > @@ -25,6 +25,7 @@ > > * Copyright (C) 2020 Alibaba, Inc, Alex Shi > > */ > > > > +#include <linux/cgroup-defs.h> > > #include <linux/page_counter.h> > > #include <linux/memcontrol.h> > > #include <linux/cgroup.h> > > @@ -5745,6 +5746,7 @@ static struct mem_cgroup *mem_cgroup_alloc(struct mem_cgroup *parent) > > vmpressure_init(&memcg->vmpressure); > > INIT_LIST_HEAD(&memcg->event_list); > > spin_lock_init(&memcg->event_list_lock); > > + spin_lock_init(&memcg->pagecounter_peak_watchers_lock); > > memcg->socket_pressure = jiffies; > > #ifdef CONFIG_MEMCG_KMEM > > memcg->kmemcg_id = -1; > > @@ -6907,12 +6909,130 @@ static u64 memory_current_read(struct cgroup_subsys_state *css, > > return (u64)page_counter_read(&memcg->memory) * PAGE_SIZE; > > } > > > > -static u64 memory_peak_read(struct cgroup_subsys_state *css, > > - struct cftype *cft) > > +static struct page_counter *memcg_memory_extract_page_counter(struct mem_cgroup *memcg) > > { > > + return &memcg->memory; > > +} > > + > > +static struct memcg_peak_mem_ctx **memcg_memory_extract_peak_watchers(struct mem_cgroup *memcg) > > +{ > > + return &memcg->peak_memory_local_watermark_watchers; > > +} > > + > > +inline int swap_memory_peak_show( > > + struct seq_file *sf, void *v, > > + struct page_counter *(*extract_pc)(struct mem_cgroup *memcg)) > > +{ > > + struct cgroup_subsys_state *css = seq_css(sf); > > struct mem_cgroup *memcg = mem_cgroup_from_css(css); > > + struct page_counter *pc = extract_pc(memcg); > > + > > + struct kernfs_open_file *of = sf->private; > > + struct memcg_peak_mem_ctx *ctx = memcg_extract_peak_mem_ctx(of); > > + s64 fd_peak = ctx->local_watermark; > > + > > + if (fd_peak == -1) { > > + seq_printf(sf, "%llu\n", (u64)pc->watermark * PAGE_SIZE); > > + return 0; > > + } > > + > > + s64 pc_peak = pc->local_watermark; > > + s64 wm = fd_peak > pc_peak ? fd_peak : pc_peak; > > + > > + seq_printf(sf, "%lld\n", wm * PAGE_SIZE); > > + return 0; > > +} > > + > > +static int memory_peak_show(struct seq_file *sf, void *v) > > +{ > > + return swap_memory_peak_show(sf, v, memcg_memory_extract_page_counter); > > I think it's really too complex. Why not pass a single boolean argument > which will define to use memory page_counter or swap page_counter? > It will eliminate a need to pass pointers to functions and also eliminate > a need for introducing these helper functions in general. Yeah, that's definitely a cleaner option with only two alternatives. I'll make that change. > > > +} > > + > > +static int swap_memory_peak_open(struct kernfs_open_file *of) > > +{ > > + struct memcg_peak_mem_ctx *ctx = memcg_extract_peak_mem_ctx(of); > > + > > + ctx->local_watermark = -1; > > + return 0; > > +} > > + > > +inline void swap_memory_peak_release( > > + struct kernfs_open_file *of, > > + struct memcg_peak_mem_ctx **(*extract_watchers)(struct mem_cgroup *memcg)) > > +{ > > + struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); > > + struct memcg_peak_mem_ctx *ctx = memcg_extract_peak_mem_ctx(of); > > + > > + if (ctx->local_watermark == -1) { > > + /* fast path (no writes on this fd)*/ > > + return; > > + } > > + spin_lock(&memcg->pagecounter_peak_watchers_lock); > > + if (ctx->next) { > > + ctx->next->prev = ctx->prev; > > + } > > + if (ctx->prev) { > > + ctx->prev->next = ctx->next; > > + } else { > > + struct memcg_peak_mem_ctx **watchers = extract_watchers(memcg); > > + > > + *watchers = ctx->next; > > + } > > + spin_unlock(&memcg->pagecounter_peak_watchers_lock); > > +} > > > > - return (u64)memcg->memory.watermark * PAGE_SIZE; > > +static void memory_peak_release(struct kernfs_open_file *of) > > +{ > > + swap_memory_peak_release(of, memcg_memory_extract_peak_watchers); > > +} > > + > > +inline ssize_t swap_memory_peak_write( > > + struct kernfs_open_file *of, > > + char *buf, size_t nbytes, loff_t off, > > + struct page_counter* (*extract_pc)(struct mem_cgroup *memcg), > > + struct memcg_peak_mem_ctx **(*extract_watchers)(struct mem_cgroup *memcg)) > > +{ > > + buf = strstrip(buf); > > + /* Only allow "fd_local_reset" to keep the API clear */ > > + if (strcmp(buf, "fd_local_reset")) > > + return -EINVAL; > > + struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); > > + struct memcg_peak_mem_ctx *ctx = memcg_extract_peak_mem_ctx(of); > > Please, don't mix variable definitions and code. Also, please, use Ok, I'll factor that out > scripts/checkpatch.pl for checking the code before submission. I guess > it will raise several issues in this patch. Thanks, I did run checkpatch.pl on this patch, but I might have made a few changes since my last pass. (it didn't complain about mixing cod and definitions) I'll make another pass before mailing the next version. > > > + > > + spin_lock(&memcg->pagecounter_peak_watchers_lock); > > + > > + struct page_counter *pc = extract_pc(memcg); > > + > > + page_counter_reset_local_watermark(pc); > > + const unsigned long cur = pc->local_watermark; > > + struct memcg_peak_mem_ctx **watchers = extract_watchers(memcg); > > + struct memcg_peak_mem_ctx *peer_ctx; > > + > > + for (peer_ctx = *watchers; peer_ctx; peer_ctx = peer_ctx->next) { > > + if (cur > peer_ctx->local_watermark) > > + peer_ctx->local_watermark = cur; > > + } > > + if (ctx->local_watermark == -1) { > > + /* only append to the list if we're not already there */ > > + if (peer_ctx) { > > + ctx->prev = peer_ctx; > > + peer_ctx->next = ctx; > > + } else { > > + *watchers = ctx; > > + } > > + } > > + ctx->local_watermark = cur; > > + spin_unlock(&memcg->pagecounter_peak_watchers_lock); > > + > > + return nbytes; > > +} > > + > > +static ssize_t memory_peak_write(struct kernfs_open_file *of, char *buf, > > + size_t nbytes, loff_t off) > > +{ > > + return swap_memory_peak_write(of, buf, nbytes, off, > > + memcg_memory_extract_page_counter, > > + memcg_memory_extract_peak_watchers); > > } > > > > static int memory_min_show(struct seq_file *m, void *v) > > @@ -7231,7 +7351,10 @@ static struct cftype memory_files[] = { > > { > > .name = "peak", > > .flags = CFTYPE_NOT_ON_ROOT, > > - .read_u64 = memory_peak_read, > > + .open = swap_memory_peak_open, > > + .release = memory_peak_release, > > + .seq_show = memory_peak_show, > > + .write = memory_peak_write, > > }, > > { > > .name = "min", > > @@ -8193,14 +8316,35 @@ static u64 swap_current_read(struct cgroup_subsys_state *css, > > return (u64)page_counter_read(&memcg->swap) * PAGE_SIZE; > > } > > > > -static u64 swap_peak_read(struct cgroup_subsys_state *css, > > - struct cftype *cft) > > + > > +static struct page_counter *memcg_swap_extract_page_counter(struct mem_cgroup *memcg) > > +{ > > + return &memcg->swap; > > +} > > + > > +static struct memcg_peak_mem_ctx **memcg_swap_extract_peak_watchers(struct mem_cgroup *memcg) > > { > > - struct mem_cgroup *memcg = mem_cgroup_from_css(css); > > + return &memcg->peak_swap_local_watermark_watchers; > > +} > > + > > +static int swap_peak_show(struct seq_file *sf, void *v) > > +{ > > + return swap_memory_peak_show(sf, v, memcg_swap_extract_page_counter); > > +} > > > > - return (u64)memcg->swap.watermark * PAGE_SIZE; > > +static ssize_t swap_peak_write(struct kernfs_open_file *of, char *buf, > > + size_t nbytes, loff_t off) > > +{ > > + return swap_memory_peak_write(of, buf, nbytes, off, > > + memcg_swap_extract_page_counter, > > + memcg_swap_extract_peak_watchers); > > +} > > +static void swap_peak_release(struct kernfs_open_file *of) > > +{ > > + swap_memory_peak_release(of, memcg_swap_extract_peak_watchers); > > } > > > > + > > static int swap_high_show(struct seq_file *m, void *v) > > { > > return seq_puts_memcg_tunable(m, > > @@ -8282,7 +8426,10 @@ static struct cftype swap_files[] = { > > { > > .name = "swap.peak", > > .flags = CFTYPE_NOT_ON_ROOT, > > - .read_u64 = swap_peak_read, > > + .open = swap_memory_peak_open, > > + .release = swap_peak_release, > > + .seq_show = swap_peak_show, > > + .write = swap_peak_write, > > }, > > { > > .name = "swap.events", > > diff --git a/mm/page_counter.c b/mm/page_counter.c > > index db20d6452b71..40d5f4990218 100644 > > --- a/mm/page_counter.c > > +++ b/mm/page_counter.c > > @@ -82,6 +82,8 @@ void page_counter_charge(struct page_counter *counter, unsigned long nr_pages) > > */ > > if (new > READ_ONCE(c->watermark)) > > WRITE_ONCE(c->watermark, new); > > + if (new > READ_ONCE(c->local_watermark)) > > + WRITE_ONCE(c->local_watermark, new); > > Hm, can't we have a single comparison on the hot path? > Also, we read and write c->local_watermark speculatively here, Idk if it's still > acceptable with an ability to reset watermarks "locally". Maybe it is, but > it definitely deserves at least a comment with an explanation. Unfortunately, since the two watermarks may be reset at different times I don't think we can consolidate. e.g. I think that if the usage peaked, dropped down a bit and then was going back up again when the "local_watermark" was reset, we'll continue only bumping local_watermark, but we don't want to touch "watermark" until we hit that watermark again. > > And btw thank you for including tests into the commit, it's really great to see. > I'd suggest you to extract them into a separate commit and post it as a series. Sure thing! I'll split them off into their own commit. (the tests are about half the line-delta) > > Thank you! Thanks for the prompt review! I'll try to get another revision out later today. -- David Finkel Senior Principal Software Engineer, Core Services