On Tue, 2 Apr 2019 10:41:15 +0800 Peter Xu <peterx@xxxxxxxxxx> wrote: > On Mon, Apr 01, 2019 at 02:16:52PM -0600, Alex Williamson wrote: > > [...] > > > @@ -1081,8 +1088,14 @@ static int vfio_dma_do_map(struct vfio_iommu *iommu, > > goto out_unlock; > > } > > > > + if (!atomic_add_unless(&iommu->dma_avail, -1, 0)) { > > + ret = -ENOSPC; > > + goto out_unlock; > > + } > > + > > dma = kzalloc(sizeof(*dma), GFP_KERNEL); > > if (!dma) { > > + atomic_inc(&iommu->dma_avail); > > This should be the only special path to revert the change. Not sure > whether this can be avoided by simply using atomic_read() or even > READ_ONCE() (I feel like we don't need atomic ops with dma_avail > because we've had the mutex but it of course it doesn't hurt...) to > replace atomic_add_unless() above to check against zero then we do > +1/-1 in vfio_[un]link_dma() only. But AFAICT this patch is correct. Thanks for the review, you're right, we're only twiddling this atomic while holding the iommu->lock mutex, so it appears unnecessary. Since we're within the mutex, I think we don't even need a READ_ONCE. We can simple test it before alloc and decrement after. Am I missing something that would specifically require READ_ONCE within our mutex critical section? Thanks, Alex