On Mon 11-09-23 10:21:24, Tejun Heo wrote: > Hello, > > On Mon, Sep 11, 2023 at 01:01:25PM -0700, Wei Xu wrote: > > Yes, it is the same test (10K contending readers). The kernel change > > is to remove stats_user_flush_mutex from mem_cgroup_user_flush_stats() > > so that the concurrent mem_cgroup_user_flush_stats() requests directly > > contend on cgroup_rstat_lock in cgroup_rstat_flush(). > > I don't think it'd be a good idea to twist rstat and other kernel internal > code to accommodate 10k parallel readers. I didn't mean to suggest optimizing for this specific scenario. I was mostly curious whether the pathological case of unbound high latency due to lock dropping is easy to trigger by huge number of readers. It seems it is not and the mutex might not be really needed as a prevention. > If we want to support that, let's > explicitly support that by implementing better batching in the read path. Well, we need to be able to handle those situations because stat files are generally readable and we do not want unrelated workloads to influence each other heavily through this path. [...] > When you have that many concurrent readers, most of them won't need to > actually flush. Agreed! -- Michal Hocko SUSE Labs