On Fri 24-06-22 10:04:30, Christian König wrote: > Hello everyone, > > To summarize the issue I'm trying to address here: Processes can allocate > resources through a file descriptor without being held responsible for it. > > I'm not explaining all the details again. See here for a more deeply > description of the problem: https://lwn.net/ml/linux-kernel/20220531100007.174649-1-christian.koenig@xxxxxxx/ > > With this iteration I'm trying to address a bunch of the comments Michal Hocko > (thanks a lot for that) gave as well as giving some new ideas. > > Changes made so far: > 1. Renamed the callback into file_rss(). This is at least a start to better > describe what this is all about. I've been going back and forth over the > naming here, if you have any better idea please speak up. > > 2. Cleanups, e.g. now providing a helper function in the fs layer to sum up > all the pages allocated by the files in a file descriptor table. > > 3. Using the actual number of allocated pages for the shmem implementation > instead of just the size. I also tried to ignore shmem files which are part > of tmpfs, cause that has a separate accounting/limitation approach. OK, this is better than the original approach there are still holes there though I am afraid. I am not sure your i_count hack is correct but that would be mostly an implementation detail. The scheme will over-account memory mapped files (including memfd). How much that matters will really differ. For the global OOM situations it is very likely that there will be barely any disk based page cache as it would be reclaimed by the time the oom killer is invoked. So this should be OK. Swap backed page cache (shmem and its users) is more tricky. It is swap bound and processes which map it will get "charged" in the form of swap entries while those which rely on read/write will just escape from the sight of the oom killer no matter how much memory they own via their shmem backed fd. This sounds rather serious to me and I hope I haven't missed anything subtle here that would keep those pages somehow visible. Anyway something to very carefully document. For the memcg OOM this gets even more tricky. Files can be shared among tasks accross memcgs. Something that is not really straightforward from the userspace POV because this is not strictly deterministic as first-one-first-charged logic is applied so a lot might depend on timing. This could also easily mean that a large part of the in memory state of the file is outside of the reclaim and therefore OOM scope of the memcg which is hitting the hard limit. This could result in tasks being killed just because they (co)operate on a large file outside of their memcg domain. To be honest I am not sure how big of a problem this would be in practice and the existing behavior has its own cons so to me it sounds like changing one set of deficiency with other. As we have discussed previously, there is unlikely a great solution but you a) need to document most prominent downsides so that people can at least see this is understood and documented behavior and b) think of the runaway situation wrt non mapped shmems memtioned above and see whether there is something we can do about that. -- Michal Hocko SUSE Labs