On Mon, Sep 23, 2024 at 3:34 AM Lai, Yi <yi1.lai@xxxxxxxxxxxxxxx> wrote: > > Hi Miklos Szeredi, > > Greetings! > > I used Syzkaller and found that there is WARNING in fuse_request_end in Linux-next tree - next-20240918. > > After bisection and the first bad commit is: > " > 5de8acb41c86 fuse: cleanup request queuing towards virtiofs > " > > All detailed into can be found at: > https://github.com/laifryiee/syzkaller_logs/tree/main/240922_114402_fuse_request_end > Syzkaller repro code: > https://github.com/laifryiee/syzkaller_logs/blob/main/240922_114402_fuse_request_end/repro.c > Syzkaller repro syscall steps: > https://github.com/laifryiee/syzkaller_logs/blob/main/240922_114402_fuse_request_end/repro.prog > Syzkaller report: > https://github.com/laifryiee/syzkaller_logs/blob/main/240922_114402_fuse_request_end/repro.report > Kconfig(make olddefconfig): > https://github.com/laifryiee/syzkaller_logs/blob/main/240922_114402_fuse_request_end/kconfig_origin > Bisect info: > https://github.com/laifryiee/syzkaller_logs/blob/main/240922_114402_fuse_request_end/bisect_info.log > bzImage: > https://github.com/laifryiee/syzkaller_logs/raw/main/240922_114402_fuse_request_end/bzImage_55bcd2e0d04c1171d382badef1def1fd04ef66c5 > Issue dmesg: > https://github.com/laifryiee/syzkaller_logs/blob/main/240922_114402_fuse_request_end/55bcd2e0d04c1171d382badef1def1fd04ef66c5_dmesg.log > > " > [ 31.577123] ------------[ cut here ]------------ > [ 31.578842] WARNING: CPU: 1 PID: 1186 at fs/fuse/dev.c:373 fuse_request_end+0x7d2/0x910 > [ 31.581269] Modules linked in: > [ 31.582553] CPU: 1 UID: 0 PID: 1186 Comm: repro Not tainted 6.11.0-next-20240918-55bcd2e0d04c #1 > [ 31.584332] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS rel-1.16.0-0-gd239552ce722-prebuilt.qemu.org 04/01/2014 > [ 31.586281] RIP: 0010:fuse_request_end+0x7d2/0x910 > [ 31.587001] Code: ff 48 8b 7d d0 e8 ae 0f 72 ff e9 c2 fb ff ff e8 a4 0f 72 ff e9 e7 fb ff ff e8 3a 3b 0a ff 0f 0b e9 17 fa ff ff e8 2e 3b 0a ff <0f> 0b e9 c1 f9 ff ff 4c 89 ff e8 af 0f 72 ff e9 82 f8 ff ff e8 a5 > [ 31.589442] RSP: 0018:ffff88802141f640 EFLAGS: 00010293 > [ 31.590198] RAX: 0000000000000000 RBX: 0000000000000201 RCX: ffffffff825d5bb2 > [ 31.591137] RDX: ffff888010b2a500 RSI: ffffffff825d61f2 RDI: 0000000000000001 > [ 31.592072] RBP: ffff88802141f680 R08: 0000000000000000 R09: ffffed100356f28e > [ 31.593010] R10: 0000000000000001 R11: 0000000000000001 R12: ffff88801ab79440 > [ 31.594062] R13: ffff88801ab79470 R14: ffff88801dcaa000 R15: ffff88800d71fa00 > [ 31.594820] FS: 00007f812eca2640(0000) GS:ffff88806c500000(0000) knlGS:0000000000000000 > [ 31.595670] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 > [ 31.596299] CR2: 000055b7baa16b20 CR3: 00000000109b0002 CR4: 0000000000770ef0 > [ 31.597054] DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000 > [ 31.597850] DR3: 0000000000000000 DR6: 00000000ffff07f0 DR7: 0000000000000400 > [ 31.598595] PKRU: 55555554 > [ 31.598902] Call Trace: > [ 31.599180] <TASK> > [ 31.599439] ? show_regs+0x6d/0x80 > [ 31.599805] ? __warn+0xf3/0x380 > [ 31.600137] ? report_bug+0x25e/0x4b0 > [ 31.600521] ? fuse_request_end+0x7d2/0x910 > [ 31.600885] ? report_bug+0x2cb/0x4b0 > [ 31.601204] ? fuse_request_end+0x7d2/0x910 > [ 31.601564] ? fuse_request_end+0x7d3/0x910 > [ 31.601956] ? handle_bug+0xf1/0x190 > [ 31.602275] ? exc_invalid_op+0x3c/0x80 > [ 31.602595] ? asm_exc_invalid_op+0x1f/0x30 > [ 31.602959] ? fuse_request_end+0x192/0x910 > [ 31.603301] ? fuse_request_end+0x7d2/0x910 > [ 31.603643] ? fuse_request_end+0x7d2/0x910 > [ 31.603988] ? do_raw_spin_unlock+0x15c/0x210 > [ 31.604366] fuse_dev_queue_req+0x23c/0x2b0 > [ 31.604714] fuse_send_one+0x1d1/0x360 > [ 31.605031] fuse_simple_request+0x348/0xd30 > [ 31.605385] ? lockdep_hardirqs_on+0x89/0x110 > [ 31.605755] fuse_send_open+0x234/0x2f0 > [ 31.606126] ? __pfx_fuse_send_open+0x10/0x10 > [ 31.606487] ? kasan_save_track+0x18/0x40 > [ 31.606834] ? lockdep_init_map_type+0x2df/0x810 > [ 31.607227] ? __kasan_check_write+0x18/0x20 > [ 31.607591] fuse_file_open+0x2bc/0x770 > [ 31.607921] fuse_do_open+0x5d/0xe0 > [ 31.608215] ? __sanitizer_cov_trace_const_cmp4+0x1a/0x20 > [ 31.608681] fuse_dir_open+0x138/0x220 > [ 31.609005] do_dentry_open+0x6be/0x1390 > [ 31.609358] ? __sanitizer_cov_trace_const_cmp4+0x1a/0x20 > [ 31.609861] ? __pfx_fuse_dir_open+0x10/0x10 > [ 31.610240] vfs_open+0x87/0x3f0 > [ 31.610523] ? may_open+0x205/0x430 > [ 31.610834] path_openat+0x23b7/0x32d0 > [ 31.611161] ? __pfx_path_openat+0x10/0x10 > [ 31.611502] ? lock_acquire.part.0+0x152/0x390 > [ 31.611874] ? __this_cpu_preempt_check+0x21/0x30 > [ 31.612266] ? lock_is_held_type+0xef/0x150 > [ 31.612611] ? __this_cpu_preempt_check+0x21/0x30 > [ 31.613002] do_filp_open+0x1cc/0x420 > [ 31.613316] ? __pfx_do_filp_open+0x10/0x10 > [ 31.613669] ? lock_release+0x441/0x870 > [ 31.614043] ? __pfx_lock_release+0x10/0x10 > [ 31.614404] ? do_raw_spin_unlock+0x15c/0x210 > [ 31.614784] do_sys_openat2+0x185/0x1f0 > [ 31.615105] ? __pfx_do_sys_openat2+0x10/0x10 > [ 31.615470] ? __this_cpu_preempt_check+0x21/0x30 > [ 31.615854] ? seqcount_lockdep_reader_access.constprop.0+0xb4/0xd0 > [ 31.616370] ? lockdep_hardirqs_on+0x89/0x110 > [ 31.616736] __x64_sys_openat+0x17a/0x240 > [ 31.617067] ? __pfx___x64_sys_openat+0x10/0x10 > [ 31.617447] ? __audit_syscall_entry+0x39c/0x500 > [ 31.617870] x64_sys_call+0x1a52/0x20d0 > [ 31.618194] do_syscall_64+0x6d/0x140 > [ 31.618504] entry_SYSCALL_64_after_hwframe+0x76/0x7e > [ 31.618917] RIP: 0033:0x7f812eb3e8c4 > [ 31.619225] Code: 24 20 eb 8f 66 90 44 89 54 24 0c e8 76 d3 f5 ff 44 8b 54 24 0c 44 89 e2 48 89 ee 41 89 c0 bf 9c ff ff ff b8 01 01 00 00 0f 05 <48> 3d 00 f0 ff ff 77 34 44 89 c7 89 44 24 0c e8 c8 d3 f5 ff 8b 44 > [ 31.620656] RSP: 002b:00007f812eca1b90 EFLAGS: 00000293 ORIG_RAX: 0000000000000101 > [ 31.621255] RAX: ffffffffffffffda RBX: 00007f812eca2640 RCX: 00007f812eb3e8c4 > [ 31.621864] RDX: 0000000000010000 RSI: 0000000020002080 RDI: 00000000ffffff9c > [ 31.622428] RBP: 0000000020002080 R08: 0000000000000000 R09: 0000000000000000 > [ 31.622987] R10: 0000000000000000 R11: 0000000000000293 R12: 0000000000010000 > [ 31.623549] R13: 0000000000000006 R14: 00007f812ea9f560 R15: 0000000000000000 > [ 31.624123] </TASK> > [ 31.624316] irq event stamp: 1655 > [ 31.624595] hardirqs last enabled at (1663): [<ffffffff8145cb85>] __up_console_sem+0x95/0xb0 > [ 31.625310] hardirqs last disabled at (1670): [<ffffffff8145cb6a>] __up_console_sem+0x7a/0xb0 > [ 31.626039] softirqs last enabled at (1466): [<ffffffff8128a889>] __irq_exit_rcu+0xa9/0x120 > [ 31.626726] softirqs last disabled at (1449): [<ffffffff8128a889>] __irq_exit_rcu+0xa9/0x120 > [ 31.627405] ---[ end trace 0000000000000000 ]--- > " > > I hope you find it useful. > > Regards, > Yi Lai > > --- > > If you don't need the following environment to reproduce the problem or if you > already have one reproduced environment, please ignore the following information. > > How to reproduce: > git clone https://gitlab.com/xupengfe/repro_vm_env.git > cd repro_vm_env > tar -xvf repro_vm_env.tar.gz > cd repro_vm_env; ./start3.sh // it needs qemu-system-x86_64 and I used v7.1.0 > // start3.sh will load bzImage_2241ab53cbb5cdb08a6b2d4688feb13971058f65 v6.2-rc5 kernel > // You could change the bzImage_xxx as you want > // Maybe you need to remove line "-drive if=pflash,format=raw,readonly=on,file=./OVMF_CODE.fd \" for different qemu version > You could use below command to log in, there is no password for root. > ssh -p 10023 root@localhost > > After login vm(virtual machine) successfully, you could transfer reproduced > binary to the vm by below way, and reproduce the problem in vm: > gcc -pthread -o repro repro.c > scp -P 10023 repro root@localhost:/root/ > > Get the bzImage for target kernel: > Please use target kconfig and copy it to kernel_src/.config > make olddefconfig > make -jx bzImage //x should equal or less than cpu num your pc has > > Fill the bzImage file into above start3.sh to load the target kernel in vm. > > Tips: > If you already have qemu-system-x86_64, please ignore below info. > If you want to install qemu v7.1.0 version: > git clone https://github.com/qemu/qemu.git > cd qemu > git checkout -f v7.1.0 > mkdir build > cd build > yum install -y ninja-build.x86_64 > yum -y install libslirp-devel.x86_64 > ../configure --target-list=x86_64-softmmu --enable-kvm --enable-vnc --enable-gtk --enable-sdl --enable-usb-redir --enable-slirp > make > make install > > On Wed, May 29, 2024 at 05:52:07PM +0200, Miklos Szeredi wrote: > > Virtiofs has its own queing mechanism, but still requests are first queued > > on fiq->pending to be immediately dequeued and queued onto the virtio > > queue. > > > > The queuing on fiq->pending is unnecessary and might even have some > > performance impact due to being a contention point. > > > > Forget requests are handled similarly. > > > > Move the queuing of requests and forgets into the fiq->ops->*. > > fuse_iqueue_ops are renamed to reflect the new semantics. > > > > Signed-off-by: Miklos Szeredi <mszeredi@xxxxxxxxxx> > > --- > > fs/fuse/dev.c | 159 ++++++++++++++++++++++++-------------------- > > fs/fuse/fuse_i.h | 19 ++---- > > fs/fuse/virtio_fs.c | 41 ++++-------- > > 3 files changed, 106 insertions(+), 113 deletions(-) > > > > diff --git a/fs/fuse/dev.c b/fs/fuse/dev.c > > index 9eb191b5c4de..a4f510f1b1a4 100644 > > --- a/fs/fuse/dev.c > > +++ b/fs/fuse/dev.c > > @@ -192,10 +192,22 @@ unsigned int fuse_len_args(unsigned int numargs, struct fuse_arg *args) > > } > > EXPORT_SYMBOL_GPL(fuse_len_args); > > > > -u64 fuse_get_unique(struct fuse_iqueue *fiq) > > +static u64 fuse_get_unique_locked(struct fuse_iqueue *fiq) > > { > > fiq->reqctr += FUSE_REQ_ID_STEP; > > return fiq->reqctr; > > + > > +} > > + > > +u64 fuse_get_unique(struct fuse_iqueue *fiq) > > +{ > > + u64 ret; > > + > > + spin_lock(&fiq->lock); > > + ret = fuse_get_unique_locked(fiq); > > + spin_unlock(&fiq->lock); > > + > > + return ret; > > } > > EXPORT_SYMBOL_GPL(fuse_get_unique); > > > > @@ -215,22 +227,67 @@ __releases(fiq->lock) > > spin_unlock(&fiq->lock); > > } > > > > +static void fuse_dev_queue_forget(struct fuse_iqueue *fiq, struct fuse_forget_link *forget) > > +{ > > + spin_lock(&fiq->lock); > > + if (fiq->connected) { > > + fiq->forget_list_tail->next = forget; > > + fiq->forget_list_tail = forget; > > + fuse_dev_wake_and_unlock(fiq); > > + } else { > > + kfree(forget); > > + spin_unlock(&fiq->lock); > > + } > > +} > > + > > +static void fuse_dev_queue_interrupt(struct fuse_iqueue *fiq, struct fuse_req *req) > > +{ > > + spin_lock(&fiq->lock); > > + if (list_empty(&req->intr_entry)) { > > + list_add_tail(&req->intr_entry, &fiq->interrupts); > > + /* > > + * Pairs with smp_mb() implied by test_and_set_bit() > > + * from fuse_request_end(). > > + */ > > + smp_mb(); > > + if (test_bit(FR_FINISHED, &req->flags)) { > > + list_del_init(&req->intr_entry); > > + spin_unlock(&fiq->lock); > > + } > > + fuse_dev_wake_and_unlock(fiq); > > + } else { > > + spin_unlock(&fiq->lock); > > + } > > +} > > + > > +static void fuse_dev_queue_req(struct fuse_iqueue *fiq, struct fuse_req *req) > > +{ > > + spin_lock(&fiq->lock); > > + if (fiq->connected) { > > + if (req->in.h.opcode != FUSE_NOTIFY_REPLY) > > + req->in.h.unique = fuse_get_unique_locked(fiq); > > + list_add_tail(&req->list, &fiq->pending); > > + fuse_dev_wake_and_unlock(fiq); > > + } else { > > + spin_unlock(&fiq->lock); > > + req->out.h.error = -ENOTCONN; > > + fuse_request_end(req); in the case where the connection has been aborted, this request will still have the FR_PENDING flag set on it when it calls fuse_request_end(). I think we can just call fuse_put_request() here instead. > > + } > > +} > > + > > const struct fuse_iqueue_ops fuse_dev_fiq_ops = { > > - .wake_forget_and_unlock = fuse_dev_wake_and_unlock, > > - .wake_interrupt_and_unlock = fuse_dev_wake_and_unlock, > > - .wake_pending_and_unlock = fuse_dev_wake_and_unlock, > > + .send_forget = fuse_dev_queue_forget, > > + .send_interrupt = fuse_dev_queue_interrupt, > > + .send_req = fuse_dev_queue_req, > > }; > > EXPORT_SYMBOL_GPL(fuse_dev_fiq_ops); > > > > -static void queue_request_and_unlock(struct fuse_iqueue *fiq, > > - struct fuse_req *req) > > -__releases(fiq->lock) > > +static void fuse_send_one(struct fuse_iqueue *fiq, struct fuse_req *req) > > { > > req->in.h.len = sizeof(struct fuse_in_header) + > > fuse_len_args(req->args->in_numargs, > > (struct fuse_arg *) req->args->in_args); > > - list_add_tail(&req->list, &fiq->pending); > > - fiq->ops->wake_pending_and_unlock(fiq); > > + fiq->ops->send_req(fiq, req); > > } > > > > void fuse_queue_forget(struct fuse_conn *fc, struct fuse_forget_link *forget, > > @@ -241,15 +298,7 @@ void fuse_queue_forget(struct fuse_conn *fc, struct fuse_forget_link *forget, > > forget->forget_one.nodeid = nodeid; > > forget->forget_one.nlookup = nlookup; > > > > - spin_lock(&fiq->lock); > > - if (fiq->connected) { > > - fiq->forget_list_tail->next = forget; > > - fiq->forget_list_tail = forget; > > - fiq->ops->wake_forget_and_unlock(fiq); > > - } else { > > - kfree(forget); > > - spin_unlock(&fiq->lock); > > - } > > + fiq->ops->send_forget(fiq, forget); > > } > > > > static void flush_bg_queue(struct fuse_conn *fc) > > @@ -263,9 +312,7 @@ static void flush_bg_queue(struct fuse_conn *fc) > > req = list_first_entry(&fc->bg_queue, struct fuse_req, list); > > list_del(&req->list); > > fc->active_background++; > > - spin_lock(&fiq->lock); > > - req->in.h.unique = fuse_get_unique(fiq); > > - queue_request_and_unlock(fiq, req); > > + fuse_send_one(fiq, req); > > } > > } > > > > @@ -335,29 +382,12 @@ static int queue_interrupt(struct fuse_req *req) > > { > > struct fuse_iqueue *fiq = &req->fm->fc->iq; > > > > - spin_lock(&fiq->lock); > > /* Check for we've sent request to interrupt this req */ > > - if (unlikely(!test_bit(FR_INTERRUPTED, &req->flags))) { > > - spin_unlock(&fiq->lock); > > + if (unlikely(!test_bit(FR_INTERRUPTED, &req->flags))) > > return -EINVAL; > > - } > > > > - if (list_empty(&req->intr_entry)) { > > - list_add_tail(&req->intr_entry, &fiq->interrupts); > > - /* > > - * Pairs with smp_mb() implied by test_and_set_bit() > > - * from fuse_request_end(). > > - */ > > - smp_mb(); > > - if (test_bit(FR_FINISHED, &req->flags)) { > > - list_del_init(&req->intr_entry); > > - spin_unlock(&fiq->lock); > > - return 0; > > - } > > - fiq->ops->wake_interrupt_and_unlock(fiq); > > - } else { > > - spin_unlock(&fiq->lock); > > - } > > + fiq->ops->send_interrupt(fiq, req); > > + > > return 0; > > } > > > > @@ -412,21 +442,15 @@ static void __fuse_request_send(struct fuse_req *req) > > struct fuse_iqueue *fiq = &req->fm->fc->iq; > > > > BUG_ON(test_bit(FR_BACKGROUND, &req->flags)); > > - spin_lock(&fiq->lock); > > - if (!fiq->connected) { > > - spin_unlock(&fiq->lock); > > - req->out.h.error = -ENOTCONN; > > - } else { > > - req->in.h.unique = fuse_get_unique(fiq); > > - /* acquire extra reference, since request is still needed > > - after fuse_request_end() */ > > - __fuse_get_request(req); > > - queue_request_and_unlock(fiq, req); > > > > - request_wait_answer(req); > > - /* Pairs with smp_wmb() in fuse_request_end() */ > > - smp_rmb(); > > - } > > + /* acquire extra reference, since request is still needed after > > + fuse_request_end() */ > > + __fuse_get_request(req); > > + fuse_send_one(fiq, req); > > + > > + request_wait_answer(req); > > + /* Pairs with smp_wmb() in fuse_request_end() */ > > + smp_rmb(); > > } > > > > static void fuse_adjust_compat(struct fuse_conn *fc, struct fuse_args *args) > > @@ -581,7 +605,6 @@ static int fuse_simple_notify_reply(struct fuse_mount *fm, > > { > > struct fuse_req *req; > > struct fuse_iqueue *fiq = &fm->fc->iq; > > - int err = 0; > > > > req = fuse_get_req(fm, false); > > if (IS_ERR(req)) > > @@ -592,16 +615,9 @@ static int fuse_simple_notify_reply(struct fuse_mount *fm, > > > > fuse_args_to_req(req, args); > > > > - spin_lock(&fiq->lock); > > - if (fiq->connected) { > > - queue_request_and_unlock(fiq, req); > > - } else { > > - err = -ENODEV; > > - spin_unlock(&fiq->lock); > > - fuse_put_request(req); > > - } > > + fuse_send_one(fiq, req); > > > > - return err; > > + return 0; > > } > > > > /* > > @@ -1076,9 +1092,9 @@ __releases(fiq->lock) > > return err ? err : reqsize; > > } > > > > -struct fuse_forget_link *fuse_dequeue_forget(struct fuse_iqueue *fiq, > > - unsigned int max, > > - unsigned int *countp) > > +static struct fuse_forget_link *fuse_dequeue_forget(struct fuse_iqueue *fiq, > > + unsigned int max, > > + unsigned int *countp) > > { > > struct fuse_forget_link *head = fiq->forget_list_head.next; > > struct fuse_forget_link **newhead = &head; > > @@ -1097,7 +1113,6 @@ struct fuse_forget_link *fuse_dequeue_forget(struct fuse_iqueue *fiq, > > > > return head; > > } > > -EXPORT_SYMBOL(fuse_dequeue_forget); > > > > static int fuse_read_single_forget(struct fuse_iqueue *fiq, > > struct fuse_copy_state *cs, > > @@ -1112,7 +1127,7 @@ __releases(fiq->lock) > > struct fuse_in_header ih = { > > .opcode = FUSE_FORGET, > > .nodeid = forget->forget_one.nodeid, > > - .unique = fuse_get_unique(fiq), > > + .unique = fuse_get_unique_locked(fiq), > > .len = sizeof(ih) + sizeof(arg), > > }; > > > > @@ -1143,7 +1158,7 @@ __releases(fiq->lock) > > struct fuse_batch_forget_in arg = { .count = 0 }; > > struct fuse_in_header ih = { > > .opcode = FUSE_BATCH_FORGET, > > - .unique = fuse_get_unique(fiq), > > + .unique = fuse_get_unique_locked(fiq), > > .len = sizeof(ih) + sizeof(arg), > > }; > > > > @@ -1822,7 +1837,7 @@ static void fuse_resend(struct fuse_conn *fc) > > spin_lock(&fiq->lock); > > /* iq and pq requests are both oldest to newest */ > > list_splice(&to_queue, &fiq->pending); > > - fiq->ops->wake_pending_and_unlock(fiq); > > + fuse_dev_wake_and_unlock(fiq); > > } > > > > static int fuse_notify_resend(struct fuse_conn *fc) > > diff --git a/fs/fuse/fuse_i.h b/fs/fuse/fuse_i.h > > index f23919610313..33b21255817e 100644 > > --- a/fs/fuse/fuse_i.h > > +++ b/fs/fuse/fuse_i.h > > @@ -449,22 +449,19 @@ struct fuse_iqueue; > > */ > > struct fuse_iqueue_ops { > > /** > > - * Signal that a forget has been queued > > + * Send one forget > > */ > > - void (*wake_forget_and_unlock)(struct fuse_iqueue *fiq) > > - __releases(fiq->lock); > > + void (*send_forget)(struct fuse_iqueue *fiq, struct fuse_forget_link *link); > > > > /** > > - * Signal that an INTERRUPT request has been queued > > + * Send interrupt for request > > */ > > - void (*wake_interrupt_and_unlock)(struct fuse_iqueue *fiq) > > - __releases(fiq->lock); > > + void (*send_interrupt)(struct fuse_iqueue *fiq, struct fuse_req *req); > > > > /** > > - * Signal that a request has been queued > > + * Send one request > > */ > > - void (*wake_pending_and_unlock)(struct fuse_iqueue *fiq) > > - __releases(fiq->lock); > > + void (*send_req)(struct fuse_iqueue *fiq, struct fuse_req *req); > > > > /** > > * Clean up when fuse_iqueue is destroyed > > @@ -1053,10 +1050,6 @@ void fuse_queue_forget(struct fuse_conn *fc, struct fuse_forget_link *forget, > > > > struct fuse_forget_link *fuse_alloc_forget(void); > > > > -struct fuse_forget_link *fuse_dequeue_forget(struct fuse_iqueue *fiq, > > - unsigned int max, > > - unsigned int *countp); > > - > > /* > > * Initialize READ or READDIR request > > */ > > diff --git a/fs/fuse/virtio_fs.c b/fs/fuse/virtio_fs.c > > index 1a52a51b6b07..690e508dbc4d 100644 > > --- a/fs/fuse/virtio_fs.c > > +++ b/fs/fuse/virtio_fs.c > > @@ -1089,22 +1089,13 @@ static struct virtio_driver virtio_fs_driver = { > > #endif > > }; > > > > -static void virtio_fs_wake_forget_and_unlock(struct fuse_iqueue *fiq) > > -__releases(fiq->lock) > > +static void virtio_fs_send_forget(struct fuse_iqueue *fiq, struct fuse_forget_link *link) > > { > > - struct fuse_forget_link *link; > > struct virtio_fs_forget *forget; > > struct virtio_fs_forget_req *req; > > - struct virtio_fs *fs; > > - struct virtio_fs_vq *fsvq; > > - u64 unique; > > - > > - link = fuse_dequeue_forget(fiq, 1, NULL); > > - unique = fuse_get_unique(fiq); > > - > > - fs = fiq->priv; > > - fsvq = &fs->vqs[VQ_HIPRIO]; > > - spin_unlock(&fiq->lock); > > + struct virtio_fs *fs = fiq->priv; > > + struct virtio_fs_vq *fsvq = &fs->vqs[VQ_HIPRIO]; > > + u64 unique = fuse_get_unique(fiq); > > > > /* Allocate a buffer for the request */ > > forget = kmalloc(sizeof(*forget), GFP_NOFS | __GFP_NOFAIL); > > @@ -1124,8 +1115,7 @@ __releases(fiq->lock) > > kfree(link); > > } > > > > -static void virtio_fs_wake_interrupt_and_unlock(struct fuse_iqueue *fiq) > > -__releases(fiq->lock) > > +static void virtio_fs_send_interrupt(struct fuse_iqueue *fiq, struct fuse_req *req) > > { > > /* > > * TODO interrupts. > > @@ -1134,7 +1124,6 @@ __releases(fiq->lock) > > * Exceptions are blocking lock operations; for example fcntl(F_SETLKW) > > * with shared lock between host and guest. > > */ > > - spin_unlock(&fiq->lock); > > } > > > > /* Count number of scatter-gather elements required */ > > @@ -1339,21 +1328,17 @@ static int virtio_fs_enqueue_req(struct virtio_fs_vq *fsvq, > > return ret; > > } > > > > -static void virtio_fs_wake_pending_and_unlock(struct fuse_iqueue *fiq) > > -__releases(fiq->lock) > > +static void virtio_fs_send_req(struct fuse_iqueue *fiq, struct fuse_req *req) > > { > > unsigned int queue_id; > > struct virtio_fs *fs; > > - struct fuse_req *req; > > struct virtio_fs_vq *fsvq; > > int ret; > > > > - WARN_ON(list_empty(&fiq->pending)); > > - req = list_last_entry(&fiq->pending, struct fuse_req, list); > > + if (req->in.h.opcode != FUSE_NOTIFY_REPLY) > > + req->in.h.unique = fuse_get_unique(fiq); > > + > > clear_bit(FR_PENDING, &req->flags); > > - list_del_init(&req->list); > > - WARN_ON(!list_empty(&fiq->pending)); > > - spin_unlock(&fiq->lock); > > > > fs = fiq->priv; > > queue_id = VQ_REQUEST + fs->mq_map[raw_smp_processor_id()]; > > @@ -1393,10 +1378,10 @@ __releases(fiq->lock) > > } > > > > static const struct fuse_iqueue_ops virtio_fs_fiq_ops = { > > - .wake_forget_and_unlock = virtio_fs_wake_forget_and_unlock, > > - .wake_interrupt_and_unlock = virtio_fs_wake_interrupt_and_unlock, > > - .wake_pending_and_unlock = virtio_fs_wake_pending_and_unlock, > > - .release = virtio_fs_fiq_release, > > + .send_forget = virtio_fs_send_forget, > > + .send_interrupt = virtio_fs_send_interrupt, > > + .send_req = virtio_fs_send_req, > > + .release = virtio_fs_fiq_release, > > }; > > > > static inline void virtio_fs_ctx_set_defaults(struct fuse_fs_context *ctx) > > -- > > 2.45.1 > > >