On Wed, Nov 29, 2023 at 10:21:49PM -0500, Kent Overstreet wrote: > On Thu, Nov 30, 2023 at 11:09:42AM +0800, Qi Zheng wrote: > > > > > > On 2023/11/30 07:11, Kent Overstreet wrote: > > > On Wed, Nov 29, 2023 at 10:14:54AM +0100, Michal Hocko wrote: > > > > On Tue 28-11-23 16:34:35, Roman Gushchin wrote: > > > > > On Tue, Nov 28, 2023 at 02:23:36PM +0800, Qi Zheng wrote: > > > > [...] > > > > > > Now I think adding this method might not be a good idea. If we allow > > > > > > shrinkers to report thier own private information, OOM logs may become > > > > > > cluttered. Most people only care about some general information when > > > > > > troubleshooting OOM problem, but not the private information of a > > > > > > shrinker. > > > > > > > > > > I agree with that. > > > > > > > > > > It seems that the feature is mostly useful for kernel developers and it's easily > > > > > achievable by attaching a bpf program to the oom handler. If it requires a bit > > > > > of work on the bpf side, we can do that instead, but probably not. And this > > > > > solution can potentially provide way more information in a more flexible way. > > > > > > > > > > So I'm not convinced it's a good idea to make the generic oom handling code > > > > > more complicated and fragile for everybody, as well as making oom reports differ > > > > > more between kernel versions and configurations. > > > > > > > > Completely agreed! From my many years of experience of oom reports > > > > analysing from production systems I would conclude the following categories > > > > - clear runaways (and/or memory leaks) > > > > - userspace consumers - either shmem or anonymous memory > > > > predominantly consumes the memory, swap is either depleted > > > > or not configured. > > > > OOM report is usually useful to pinpoint those as we > > > > have required counters available > > > > - kernel memory consumers - if we are lucky they are > > > > using slab allocator and unreclaimable slab is a huge > > > > part of the memory consumption. If this is a page > > > > allocator user the oom repport only helps to deduce > > > > the fact by looking at how much user + slab + page > > > > table etc. form. But identifying the root cause is > > > > close to impossible without something like page_owner > > > > or a crash dump. > > > > - misbehaving memory reclaim > > > > - minority of issues and the oom report is usually > > > > insufficient to drill down to the root cause. If the > > > > problem is reproducible then collecting vmstat data > > > > can give a much better clue. > > > > - high number of slab reclaimable objects or free swap > > > > are good indicators. Shrinkers data could be > > > > potentially helpful in the slab case but I really have > > > > hard time to remember any such situation. > > > > On non-production systems the situation is quite different. I can see > > > > how it could be very beneficial to add a very specific debugging data > > > > for subsystem/shrinker which is developed and could cause the OOM. For > > > > that purpose the proposed scheme is rather inflexible AFAICS. > > > > > > Considering that you're an MM guy, and that shrinkers are pretty much > > > universally used by _filesystem_ people - I'm not sure your experience > > > is the most relevant here? > > > > > > The general attitude I've been seeing in this thread has been one of > > > dismissiveness towards filesystem people. Roman too; back when he was > > > > Oh, please don't say that, it seems like you are the only one causing > > the fight. We deeply respect the opinions of file system developers, so > > I invited Dave to this thread from the beginning. And you didn't CC > > linux-fsdevel@xxxxxxxxxxxxxxx yourself. > > > > > working on his shrinker debug feature I reached out to him, explained > > > that I was working on my own, and asked about collaborating - got > > > crickets in response... > > > > > > Hmm.. > > > > > > Besides that, I haven't seen anything what-so-ever out of you guys to > > > make our lives easier, regarding OOM debugging, nor do you guys even > > > seem interested in the needs and perspectives of the filesytem people. > > > Roman, your feature didn't help one bit for OOM debuging - didn't even > > > come with documentation or hints as to what it's for. > > > > > > BPF? Please. > > > > (Disclaimer, no intention to start a fight, here are some objective > > views.) > > > > Why not? In addition to printk, there are many good debugging tools > > worth trying, such as BPF related tools, drgn, etc. > > > > For non-bcachefs developers, who knows what those statistics mean? > > > > You can use BPF or drgn to traverse in advance to get the address of the > > bcachefs shrinker structure, and then during OOM, find the bcachefs > > private structure through the shrinker->private_data member, and then > > dump the bcachefs private data. Is there any problem with this? > > No, BPF is not an excuse for improving our OOM/allocation failure > reports. BPF/tracing are secondary tools; whenever we're logging > information about a problem we should strive to log enough information > to debug the issue. Ok, a simple question then: why can't you dump /proc/slabinfo after the OOM? Unlike anon memory, slab memory (fs caches in particular) should not be heavily affected by killing some userspace task. Thanks.