On 11/17/22 10:01 PM, Johannes Weiner wrote: > On Thu, Nov 17, 2022 at 09:21:10PM +0530, Aneesh Kumar K V wrote: >> On 11/17/22 9:12 PM, Aneesh Kumar K V wrote: >>> On 11/17/22 8:42 PM, Johannes Weiner wrote: >>>> Hi Aneesh, >>>> >>>> On Thu, Nov 17, 2022 at 12:24:13PM +0530, Aneesh Kumar K.V wrote: >>>>> Currently, we don't pause in balance_dirty_pages with cgroup v1 when we >>>>> have task dirtying too many pages w.r.t to memory limit in the memcg. >>>>> This is because with cgroup v1 all the limits are checked against global >>>>> available resources. So on a system with a large amount of memory, a >>>>> cgroup with a smaller limit can easily hit OOM if the task within the >>>>> cgroup continuously dirty pages. >>>> >>>> Page reclaim has special writeback throttling for cgroup1, see the >>>> folio_wait_writeback() in shrink_folio_list(). It's not as smooth as >>>> proper dirty throttling, but it should prevent OOMs. >>>> >>>> Is this not working anymore? >>> >>> The test is a simple dd test on on a 256GB system. >>> >>> root@lp2:/sys/fs/cgroup/memory# mkdir test >>> root@lp2:/sys/fs/cgroup/memory# cd test/ >>> root@lp2:/sys/fs/cgroup/memory/test# echo 120M > memory.limit_in_bytes >>> root@lp2:/sys/fs/cgroup/memory/test# echo $$ > tasks >>> root@lp2:/sys/fs/cgroup/memory/test# dd if=/dev/zero of=/home/kvaneesh/test bs=1M >>> Killed >>> >>> >>> Will it hit the folio_wait_writeback, because it is sequential i/o and none of the folio >>> we are writing will be in writeback? >> >> Other way to look at this is, if the writeback is never started via balance_dirty_pages, >> will we be finding folios in shrink_folio_list that is in writeback? > > The flushers are started from reclaim if necessary. See this code from > shrink_inactive_list(): > > /* > * If dirty folios are scanned that are not queued for IO, it > * implies that flushers are not doing their job. This can > * happen when memory pressure pushes dirty folios to the end of > * the LRU before the dirty limits are breached and the dirty > * data has expired. It can also happen when the proportion of > * dirty folios grows not through writes but through memory > * pressure reclaiming all the clean cache. And in some cases, > * the flushers simply cannot keep up with the allocation > * rate. Nudge the flusher threads in case they are asleep. > */ > if (stat.nr_unqueued_dirty == nr_taken) > wakeup_flusher_threads(WB_REASON_VMSCAN); > > It sounds like there isn't enough time for writeback to commence > before the memcg already declares OOM. > > If you place a reclaim_throttle(VMSCAN_THROTTLE_WRITEBACK) after that > wakeup, does that fix the issue? yes. That helped. One thing I noticed is with that reclaim_throttle, we don't end up calling folio_wait_writeback() at all. But still the dd was able to continue till the file system got full. Without that reclaim_throttle(), we do end up calling folio_wait_writeback() but at some point hit OOM [ 78.274704] vmscan: memcg throttling [ 78.422914] dd invoked oom-killer: gfp_mask=0x101c4a(GFP_NOFS|__GFP_HIGHMEM|__GFP_HARDWALL|__GFP_MOVABLE|__GFP_WRITE), order=0, oom_score_adj=0 [ 78.422927] CPU: 33 PID: 1185 Comm: dd Not tainted 6.0.0-dirty #394 [ 78.422933] Call Trace: [ 78.422935] [c00000001d0ab1d0] [c000000000cbcba4] dump_stack_lvl+0x98/0xe0 (unreliable) [ 78.422947] [c00000001d0ab210] [c0000000004ef618] dump_header+0x68/0x470 [ 78.422955] [c00000001d0ab2a0] [c0000000004ed6e0] oom_kill_process+0x410/0x440 [ 78.422961] [c00000001d0ab2e0] [c0000000004eedf0] out_of_memory+0x230/0x950 [ 78.422968] [c00000001d0ab380] [c00000000063e748] mem_cgroup_out_of_memory+0x148/0x190 [ 78.422975] [c00000001d0ab410] [c00000000064b54c] try_charge_memcg+0x95c/0x9d0 [ 78.422982] [c00000001d0ab570] [c00000000064c83c] charge_memcg+0x6c/0x180 [ 78.422988] [c00000001d0ab5b0] [c00000000064f9b8] __mem_cgroup_charge+0x48/0xb0 [ 78.422993] [c00000001d0ab5f0] [c0000000004dfedc] __filemap_add_folio+0x2cc/0x870 [ 78.423000] [c00000001d0ab6b0] [c0000000004e04fc] filemap_add_folio+0x7c/0x130 [ 78.423006] [c00000001d0ab710] [c0000000004e1d4c] __filemap_get_folio+0x2dc/0xb00 [ 78.423012] [c00000001d0ab840] [c000000000771f64] iomap_write_begin+0x2a4/0xba0 [ 78.423018] [c00000001d0ab9a0] [c000000000772a28] iomap_file_buffered_write+0x1c8/0x460 [ 78.423024] [c00000001d0abb60] [c0000000009c1bf8] xfs_file_buffered_write+0x158/0x4f0