On Mon, Oct 21, 2019 at 11:17:11AM -0400, Alan Stern wrote: > On Fri, 18 Oct 2019, Johan Hovold wrote: > > > The custom ring-buffer implementation was merged without any locking > > whatsoever, but a spinlock was later added by commit 9d33efd9a791 > > ("USB: ldusb bugfix"). > > > > The lock did not cover the loads from the ring-buffer entry after > > determining the buffer was non-empty, nor the update of the tail index > > once the entry had been processed. The former could lead to stale data > > being returned, while the latter could lead to memory corruption on > > sufficiently weakly ordered architectures. > > Let's see if I understand this correctly. > > The completion routine stores a buffer-length value at the location > actual_buffer points to, and it stores the buffer contents themselves > in the immediately following bytes. All this happens while the > dev->rbsl spinlock is held. Right. > Later on the read routine loads a value from *actual_buffer while > holding the spinlock, but drops the spinlock before copying the > immediately following buffer contents to userspace. It doesn't currently hold the spinlock while reading *actual_buffer, only when checking if the ring-buffer is non-empty. The patch below extends the check to cover also the load from *actual_buffer. > Your question is whether the read routine needs to call smp_rmb() after > dropping the spinlock and before doing copy_to_user(), right? Right, or alternatively, if an smp_rmb() after dropping the spinlock and before loading *actual_buffer is needed. > The answer is: No, smp_rmb() isn't needed. All the data stored while > ld_usb_interrupt_in_callback() held the spinlock will be visible to > ld_usb_read() while it holds the spinlock and afterward (assuming the > critical section in ld_usb_read() runs after the critical section in > ld_usb_interrupt_in_callback() -- but you know this is true because of > the value you read from *actual_buffer). Did you mean the value "read from dev->ring_head" (which tells us the ring-buffer has been updated) here? We currently have something like this in ld_usb_read(): spin_lock_irq(&lock); while (head == tail) { spin_unlock(&lock); wait_event(event); spin_lock(&lock); } spin_unlock_irq(&lock); entry = &buffer[tail]; len = *entry; copy_to_user(buf, entry + 1, len); /* update tail */ And without an smp_rmb() after dropping the spinlock, what prevents the load from *entry from being done before the load from head? Nothing, right (the spin_unlock_irq() is only a compiler barrier for later loads)? But that's fine because all stores done by the completion handler under the spinlock would of course be visible at that point. So the current code is fine wrt to copy_to_user(), and only the tail bits below are actually needed. Thanks, Alan! Had myself confused there. Johan > > Fixes: 2824bd250f0b ("[PATCH] USB: add ldusb driver") > > Fixes: 9d33efd9a791 ("USB: ldusb bugfix") > > Cc: stable <stable@xxxxxxxxxxxxxxx> # 2.6.13 > > Signed-off-by: Johan Hovold <johan@xxxxxxxxxx> > > --- > > drivers/usb/misc/ldusb.c | 15 ++++++++++++--- > > 1 file changed, 12 insertions(+), 3 deletions(-) > > > > diff --git a/drivers/usb/misc/ldusb.c b/drivers/usb/misc/ldusb.c > > index 15b5f06fb0b3..6b5843b0071e 100644 > > --- a/drivers/usb/misc/ldusb.c > > +++ b/drivers/usb/misc/ldusb.c > > @@ -477,11 +477,11 @@ static ssize_t ld_usb_read(struct file *file, char __user *buffer, size_t count, > > > > spin_lock_irq(&dev->rbsl); > > } > > - spin_unlock_irq(&dev->rbsl); > > > > /* actual_buffer contains actual_length + interrupt_in_buffer */ > > actual_buffer = (size_t *)(dev->ring_buffer + dev->ring_tail * (sizeof(size_t)+dev->interrupt_in_endpoint_size)); > > if (*actual_buffer > dev->interrupt_in_endpoint_size) { > > + spin_unlock_irq(&dev->rbsl); > > retval = -EIO; > > goto unlock_exit; > > } > > @@ -489,17 +489,26 @@ static ssize_t ld_usb_read(struct file *file, char __user *buffer, size_t count, > > if (bytes_to_read < *actual_buffer) > > dev_warn(&dev->intf->dev, "Read buffer overflow, %zd bytes dropped\n", > > *actual_buffer-bytes_to_read); > > + spin_unlock_irq(&dev->rbsl); > > + > > + /* > > + * Pairs with spin_unlock_irqrestore() in > > + * ld_usb_interrupt_in_callback() and makes sure the ring-buffer entry > > + * has been updated before copy_to_user(). > > + */ > > + smp_rmb(); > > > > /* copy one interrupt_in_buffer from ring_buffer into userspace */ > > if (copy_to_user(buffer, actual_buffer+1, bytes_to_read)) { > > retval = -EFAULT; > > goto unlock_exit; > > } > > - dev->ring_tail = (dev->ring_tail+1) % ring_buffer_size; > > - > > retval = bytes_to_read; > > > > spin_lock_irq(&dev->rbsl); > > + > > + dev->ring_tail = (dev->ring_tail + 1) % ring_buffer_size; > > + > > if (dev->buffer_overflow) { > > dev->buffer_overflow = 0; > > spin_unlock_irq(&dev->rbsl); > > > >