On Tue, Mar 8, 2022 at 5:17 AM Dmitry Osipenko <dmitry.osipenko@xxxxxxxxxxxxx> wrote: > > Add memory shrinker and new madvise IOCTL to the VirtIO-GPU driver. > Userspace (BO cache manager of Mesa driver) will mark BOs as "don't need" > using the new IOCTL to let shrinker purge the marked BOs on OOM, thus > shrinker will lower memory pressure and prevent OOM kills. > > Signed-off-by: Daniel Almeida <daniel.almeida@xxxxxxxxxxxxx> > Signed-off-by: Dmitry Osipenko <dmitry.osipenko@xxxxxxxxxxxxx> > --- > drivers/gpu/drm/virtio/Makefile | 3 +- > drivers/gpu/drm/virtio/virtgpu_drv.h | 26 +++- > drivers/gpu/drm/virtio/virtgpu_gem.c | 84 ++++++++++++ > drivers/gpu/drm/virtio/virtgpu_gem_shrinker.c | 124 ++++++++++++++++++ > drivers/gpu/drm/virtio/virtgpu_ioctl.c | 37 ++++++ > drivers/gpu/drm/virtio/virtgpu_kms.c | 10 ++ > drivers/gpu/drm/virtio/virtgpu_object.c | 7 + > drivers/gpu/drm/virtio/virtgpu_plane.c | 17 ++- > drivers/gpu/drm/virtio/virtgpu_vq.c | 15 +++ > include/uapi/drm/virtgpu_drm.h | 14 ++ > 10 files changed, 333 insertions(+), 4 deletions(-) > create mode 100644 drivers/gpu/drm/virtio/virtgpu_gem_shrinker.c > [snip] > diff --git a/drivers/gpu/drm/virtio/virtgpu_gem_shrinker.c b/drivers/gpu/drm/virtio/virtgpu_gem_shrinker.c > new file mode 100644 > index 000000000000..39eb9a3e7e4a > --- /dev/null > +++ b/drivers/gpu/drm/virtio/virtgpu_gem_shrinker.c > @@ -0,0 +1,124 @@ > +// SPDX-License-Identifier: GPL-2.0-only > +/* > + * Copyright (C) 2022 Collabora Ltd. > + */ > + > +#include <linux/dma-mapping.h> > +#include <linux/shmem_fs.h> > + > +#include "virtgpu_drv.h" > + > +static unsigned long > +virtio_gpu_gem_shrinker_count_objects(struct shrinker *shrinker, > + struct shrink_control *sc) > +{ > + struct drm_gem_shmem_object *shmem; > + struct virtio_gpu_device *vgdev; > + unsigned long count = 0; > + bool empty = true; > + > + vgdev = container_of(shrinker, struct virtio_gpu_device, > + vgshrinker.shrinker); > + > + if (!mutex_trylock(&vgdev->mm_lock)) > + return 0; One bit of advice from previously dealing with shrinker and heavy memory pressure situations (turns out 4GB chromebooks can be pretty much under *constant* memory pressure): You *really* want to make shrinker->count_objects lockless.. and minimize the lock contention on shrinker->scan_objects (ie. The problem is you can end up with shrinking going on on all CPU cores in parallel, you want to not funnel that thru one lock as much as possible. See in particular: 25ed38b3ed26 ("drm/msm: Drop mm_lock in scan loop") cc8a4d5a1bd8 ("drm/msm: Avoid mutex in shrinker_count()") BR, -R > + list_for_each_entry(shmem, &vgdev->vgshrinker.list, madv_list) { > + empty = false; > + > + if (!mutex_trylock(&shmem->pages_lock)) > + continue; > + > + if (drm_gem_shmem_is_purgeable(shmem)) > + count += shmem->base.size >> PAGE_SHIFT; > + > + mutex_unlock(&shmem->pages_lock); > + } > + > + mutex_unlock(&vgdev->mm_lock); > + > + return empty ? SHRINK_EMPTY : count; > +} > + > +static bool virtio_gpu_gem_shrinker_purge(struct virtio_gpu_device *vgdev, > + struct drm_gem_object *obj) > +{ > + struct virtio_gpu_object *bo = gem_to_virtio_gpu_obj(obj); > + struct drm_gem_shmem_object *shmem = &bo->base; > + int err; > + > + if (!dma_resv_test_signaled(obj->resv, true) || > + !drm_gem_shmem_is_purgeable(shmem) || > + refcount_read(&bo->pin_count)) > + return false; > + > + /* > + * Release host's memory before guest's memory is gone to ensure that > + * host won't touch released memory of the guest. > + */ > + err = virtio_gpu_gem_host_mem_release(bo); > + if (err) > + return false; > + > + list_del_init(&shmem->madv_list); > + drm_gem_shmem_purge_locked(shmem); > + > + return true; > +} > + > +static unsigned long > +virtio_gpu_gem_shrinker_scan_objects(struct shrinker *shrinker, > + struct shrink_control *sc) > +{ > + struct drm_gem_shmem_object *shmem, *tmp; > + struct virtio_gpu_device *vgdev; > + unsigned long freed = 0; > + > + vgdev = container_of(shrinker, struct virtio_gpu_device, > + vgshrinker.shrinker); > + > + if (!mutex_trylock(&vgdev->mm_lock)) > + return SHRINK_STOP; > + > + list_for_each_entry_safe(shmem, tmp, &vgdev->vgshrinker.list, madv_list) { > + if (freed >= sc->nr_to_scan) > + break; > + > + if (!dma_resv_trylock(shmem->base.resv)) > + continue; > + > + if (!mutex_trylock(&shmem->pages_lock)) > + goto resv_unlock; > + > + if (virtio_gpu_gem_shrinker_purge(vgdev, &shmem->base)) > + freed += shmem->base.size >> PAGE_SHIFT; > + > + mutex_unlock(&shmem->pages_lock); > +resv_unlock: > + dma_resv_unlock(shmem->base.resv); > + } > + > + mutex_unlock(&vgdev->mm_lock); > + > + return freed; > +} > + > +int virtio_gpu_gem_shrinker_init(struct virtio_gpu_device *vgdev) > +{ > + struct shrinker *shrinker = &vgdev->vgshrinker.shrinker; > + > + shrinker->count_objects = virtio_gpu_gem_shrinker_count_objects; > + shrinker->scan_objects = virtio_gpu_gem_shrinker_scan_objects; > + shrinker->seeks = DEFAULT_SEEKS; > + > + INIT_LIST_HEAD(&vgdev->vgshrinker.list); > + > + return register_shrinker(shrinker); > +} > + > +void virtio_gpu_gem_shrinker_fini(struct virtio_gpu_device *vgdev) > +{ > + struct shrinker *shrinker = &vgdev->vgshrinker.shrinker; > + > + unregister_shrinker(shrinker); > +} _______________________________________________ Virtualization mailing list Virtualization@xxxxxxxxxxxxxxxxxxxxxxxxxx https://lists.linuxfoundation.org/mailman/listinfo/virtualization