On Tue, Jan 07, 2025 at 09:10:09AM -0800, Nicolin Chen wrote: > +static ssize_t iommufd_veventq_fops_read(struct iommufd_eventq *eventq, > + char __user *buf, size_t count, > + loff_t *ppos) > +{ > + size_t done = 0; > + int rc = 0; > + > + if (*ppos) > + return -ESPIPE; > + > + mutex_lock(&eventq->mutex); > + while (!list_empty(&eventq->deliver) && count > done) { > + struct iommufd_vevent *cur = list_first_entry( > + &eventq->deliver, struct iommufd_vevent, node); > + > + if (cur->data_len > count - done) > + break; > + > + if (copy_to_user(buf + done, cur->event_data, cur->data_len)) { > + rc = -EFAULT; > + break; > + } Now that I look at this more closely, the fault path this is copied from is not great. This copy_to_user() can block while waiting on a page fault, possibily for a long time. While blocked the mutex is held and we can't add more entries to the list. That will cause the shared IRQ handler in the iommu driver to back up, which would cause a global DOS. This probably wants to be organized to look more like while (itm = eventq_get_next_item(eventq)) { if (..) { eventq_restore_failed_item(eventq); return -1; } } Where the next_item would just be a simple spinlock across the linked list manipulation. Jason