>On Wed 12-04-23 18:44:40, Jaewon Kim wrote: >> >On Wed 12-04-23 17:57:26, Jaewon Kim wrote: >> >> >Sorry for being late. I know there was some pre-existing discussion >> >> >around that but I didn't have time to participate. >> >> > >> >> >On Mon 10-04-23 16:32:28, Jaewon Kim wrote: >> >> >> @@ -350,6 +350,9 @@ static struct dma_buf *system_heap_allocate(struct dma_heap *heap, >> >> >> struct page *page, *tmp_page; >> >> >> int i, ret = -ENOMEM; >> >> >> >> >> >> + if (len / PAGE_SIZE > totalram_pages()) >> >> >> + return ERR_PTR(-ENOMEM); >> >> >> + >> >> > >> >> >This is an antipattern imho. Check 7661809d493b ("mm: don't allow >> >> >oversized kvmalloc() calls") how kvmalloc has dealt with a similar >> >> >> >> Hello Thank you for the information. >> >> >> >> I tried to search the macro of INT_MAX. >> >> >> >> include/vdso/limits.h >> >> #define INT_MAX ((int)(~0U >> 1)) >> >> >> >> AFAIK the dma-buf system heap user can request that huge size more than 2GB. >> > >> >Do you have any pointers? This all is unreclaimable memory, right? How >> >are those users constrained to not go overboard? >> >> Correct dma-buf system heap memory is unreclaimable. To avoid that huge request, >> this patch includes __GFP_RETRY_MAYFAIL. > >__GFP_RETRY_MAYFAIL doesn't avoud huge requests. It will drain the free >available memory to the edge of OOM (especially for low order requests) >so effectively anybody else requesting any memory (GFP_KERNEL like req.) >will hit the oom killer very likely). > >> #define LOW_ORDER_GFP (GFP_HIGHUSER | __GFP_ZERO | __GFP_RETRY_MAYFAIL) >> >> > >> >> So >> >> I think totalram_pages() is better than INT_MAX in this case. >> >> >> >> >issue. totalram_pages doesn't really tell you anything about incorrect >> >> >users. You might be on a low memory system where the request size is >> >> >sane normally, it just doesn't fit into memory on that particular >> >> >machine. >> >> >> >> Sorry maybe I'm not fully understand what you meant. User may requested >> >> a huge size like 3GB on 2GB ram device. But I think that should be rejected >> >> because it is bigger than the device ram size. >> > >> >Even totalram_pages/10 can be just unfeasible amount of data to be >> >allocated without a major disruption. totalram_pages is no measure of >> >the memory availability. >> >If you want to have a ballpark estimation then si_mem_available might be >> >something you are looking for. But I thought the sole purpose of this >> >patch is to catch obviously buggy callers (like sign overflow lenght >> >etc) rather than any memory consumption sanity check. >> >> Yes if we want to avoid some big size, si_mem_available could be one option. >> Actually I tried to do totalram_pages() / 2 like the old ion system heap in >> the previous patch version. Anyway totalram_pages in this patch is used to >> avoid the buggy size. > >So let me repeat that totalram_pages is a wrong thing to do(tm). > >This is not a subsystem I would feel like nacking a patch, but consider >this feedback as strong of a rejection as somebody external can give >you. A mm internal allocator would get an outright nack. > >What you are doing is just wrong and an antipattern to what other >allocators do. Either use something like INT_MAX to catch overflows or >do not try to catch buggy code but pretend a better memory consumer >citizen by using something like si_mem_available (ideally think of >other potential memory users so do not allow any request to use all >of it). The later might require much more involved interface and I do >rememeber some attempts to account and limit dmabuf memory better. > >> And as we discussed in v2 patch, __GFP_RETRY_MAYFAIL was added. And I think >> the gfp makes us feel better in memory perspective. > >wishful thinking that is. >-- >Michal Hocko >SUSE Labs Yes I think you're right. As a allocator, dma-buf system heap looks to be loose in memory allocation. Limiting dmabuf memory may be required. But I think there is no nice and reasonable way so far. And the dma-buf system heap is being widely used in Android mobile system. AFAIK the camera consumes huge memory through this dma-buf system heap. I actually even looked a huge size request over 2GB in one dma-buf request. Jaewon Kim