Hi John,
On 11/22/22 6:37 PM, John Keeping wrote:
On Tue, Nov 22, 2022 at 05:56:56PM +0530, Udipto Goswami wrote:
On 11/22/22 5:17 PM, John Keeping wrote:
On Mon, Nov 21, 2022 at 09:52:43AM +0530, Udipto Goswami wrote:
Hi John
On 11/20/22 11:18 PM, John Keeping wrote:
On Sun, Nov 20, 2022 at 12:23:50PM +0530, Udipto Goswami wrote:
On 11/18/22 9:49 PM, John Keeping wrote:
On Wed, Nov 16, 2022 at 04:49:55PM +0530, Udipto Goswami wrote:
While performing fast composition switch, there is a possibility that the
process of ffs_ep0_write/ffs_ep0_read get into a race condition
due to ep0req being freed up from functionfs_unbind.
Consider the scenario that the ffs_ep0_write calls the ffs_ep0_queue_wait
by taking a lock &ffs->ev.waitq.lock. However, the functionfs_unbind isn't
bounded so it can go ahead and mark the ep0req to NULL, and since there
is no NULL check in ffs_ep0_queue_wait we will end up in use-after-free.
Fix this by making a serialized execution between the two functions using
a mutex_lock(ffs->mutex). Also, dequeue the ep0req to ensure that no
other function can use it after the free operation.
Fixes: ddf8abd25994 ("USB: f_fs: the FunctionFS driver")
Signed-off-by: Udipto Goswami <quic_ugoswami@xxxxxxxxxxx>
---
v2: Replaces spinlock with mutex & added dequeue operation in unbind.
drivers/usb/gadget/function/f_fs.c | 7 +++++++
1 file changed, 7 insertions(+)
diff --git a/drivers/usb/gadget/function/f_fs.c b/drivers/usb/gadget/function/f_fs.c
index 73dc10a77cde..1439449df39a 100644
--- a/drivers/usb/gadget/function/f_fs.c
+++ b/drivers/usb/gadget/function/f_fs.c
@@ -279,6 +279,9 @@ static int __ffs_ep0_queue_wait(struct ffs_data *ffs, char *data, size_t len)
struct usb_request *req = ffs->ep0req;
int ret;
+ if (!req)
+ return -EINVAL;
+
req->zero = len < le16_to_cpu(ffs->ev.setup.wLength);
spin_unlock_irq(&ffs->ev.waitq.lock);
@@ -1892,10 +1895,14 @@ static void functionfs_unbind(struct ffs_data *ffs)
ENTER();
if (!WARN_ON(!ffs->gadget)) {
+ mutex_lock(&ffs->mutex);
+ /* dequeue before freeing ep0req */
+ usb_ep_dequeue(ffs->gadget->ep0, ffs->ep0req);
usb_ep_free_request(ffs->gadget->ep0, ffs->ep0req);
ffs->ep0req = NULL;
ffs->gadget = NULL;
clear_bit(FFS_FL_BOUND, &ffs->flags);
+ mutex_unlock(&ffs->mutex);
There's now a deadlock here if some other thread is waiting in
__ffs_ep0_queue_wait() on ep0req_completion.
You need to dequeue before taking the lock.
That's a control request right, will it be async?
Anyway I see only 2 possible threads ep0_read/ep0_write who calls
ep0_queue_wait and waits for the completion of ep0req and both
ep0_read/write are prptected by the mutex lock so i guess execution won't
reach there right ?
Say functionfs_unbind ran first then ep0_read/write had to wait will the
functionfs_unbind is completed so ep_dequeue will ran, will get completed,
further free the request, mark in NULL. now ep0_read/write will have ep0req
as NULL so bail out.
Is reverse then functionfs_unbind will wait will the ep0_read/write is
completed.
What guarantee is there that the transfer completes?
If there is such a guarantee, then the request will not be queued, so
why is usb_ep_dequeue() necessary?
I Agree that we cannot say that for sure, but we see that
wait_for_completion in the ep0_queue_wait is also inside mutex which was
acquired in ep0_read/write right?
Correct.
I Though of maintaining the uniformity for the approaches.
What uniformity? If one process is blocked waiting for completion and
another process wants to cancel the operation, then the cancel
(usb_eq_dequeue()) must run concurrently with the wait, otherwise the
blocked process will never wake up.
I get that, we want to rely on the dequeue to get us unblocked.
But this is also true right that doing dequeue outside might cause this?
functionfs_unbind
ep0_dequeue
ffs_ep0_read
mutex_lock()
giveback ep0_queue
map request buffer
unmap buffer
This can affect the controller's list i.e the pending_list for ep0 or might
also result on controller accessing a stale memory address isn't it ?
Or does the mutex would let the ep0_read execute in atomic context? No
right. Will it not be able to execute parallely? If not then yah we can do
dequeue outside mutex for sure.
I would expect that if we're in unbind then any attempt to enqueue a new
request will fail, so if the mutex is taken in the case above ep_queue()
should fail with -ESHUTDOWN.
But I can't actually find an point to any code that ensures that is the
case!
This doesn't matter too much though - it's not going to result in any
access to stale memory because either:
ep0_dequeue
ffs_ep0_read
mutex_lock()
ep0_queue
... wait for response ...
read ep0req->status
mutex_unlock()
mutex_lock()
free_ep0_request
...
or:
ffs_ep0_read
mutex_lock()
ep0_queue
ep0_dequeue
wake up
read ep0req->status
mutex_unlock()
mutex_lock()
free_ep0_request
...
The first case isn't ideal as we don't want to be waiting for a request
while unbinding, but it's not unsafe.
Thanks for the clarification, i'll take the dequeue out of the mutex and
do some testing, will update it in v3.
Thanks again,
-Udipto