On Thu, May 09, 2024 at 10:23:16AM +0100, Daniel Stone wrote: > Hi, > > On Wed, 8 May 2024 at 16:49, Daniel Vetter <daniel@xxxxxxxx> wrote: > > On Wed, May 08, 2024 at 09:38:33AM +0100, Daniel Stone wrote: > > > Right now, if your platform requires CMA for display, then the app > > > needs access to the GPU render node and the display node too, in order > > > to allocate buffers which the compositor can scan out directly. If it > > > only has access to the render nodes and not the display node, it won't > > > be able to allocate correctly, so its content will need a composition > > > pass, i.e. performance penalty for sandboxing. But if it can allocate > > > correctly, then hey, it can exhaust CMA just like heaps can. > > > > > > Personally I think we'd be better off just allowing access and > > > figuring out cgroups later. It's not like the OOM story is great > > > generally, and hey, you can get there with just render nodes ... > > > > Imo the right fix is to ask the compositor to allocate the buffers in this > > case, and then maybe have some kind of revoke/purge behaviour on these > > buffers. Compositor has an actual idea of who's a candidate for direct > > scanout after all, not the app. Or well at least force migrate the memory > > from cma to shmem. > > > > If you only whack cgroups on this issue you're still stuck in the world > > where either all apps together can ddos the display or no one can > > realistically direct scanout. > > Mmm, back to DRI2. I can't say I'm wildly enthused about that, not > least because a client using GPU/codec/etc for those buffers would > have to communicate its requirements (alignment etc) forward to the > compositor in order for the compositor to allocate for it. Obviously > passing the constraints etc around isn't a solved problem yet, but it > is at least contained down in clients rather than making it back and > forth between client and compositor. I don't think you need the compositor to allocate the buffer from the requirements, you only need a protocol that a) allocates a buffer of a given size from a given heap and b) has some kinda of revoke provisions so that the compositor can claw back the memory again when it needs it. > I'm extremely not-wild about the compositor migrating memory from CMA > to shmem behind the client's back, and tbh I'm not sure how that would > even work if the client has it pinned through whatever API it's > imported into. Other option is revoke on cma buffers that are allocated by clients, for the case the compositor needs it. > Anyway, like Laurent says, if we're deciding that heaps can't be used > by generic apps (unlike DRM/V4L2/etc), then we need gralloc. gralloc doesn't really fix this, it's just abstraction around how/where you allocate? Anyway the current plan is that we all pretend this issue of CMA allocated buffers don't exist and we let clients allocate without limits. Given that we don't even have cgroups to sort out the mess for anything else I wouldn't worry too much ... -Sima -- Daniel Vetter Software Engineer, Intel Corporation http://blog.ffwll.ch