hello, > On Wed, Feb 15, 2023 at 08:41:22AM +0800, Xiaoguang Wang wrote: >> Currenly only one bpf_ublk_queue_sqe() ebpf is added, ublksrv target >> can use this helper to write ebpf prog to support ublk kernel & usersapce >> zero copy, please see ublksrv test codes for more info. >> >> */ >> + if ((req_op(req) == REQ_OP_WRITE) && ub->io_prep_prog) >> + return rq_bytes; > Can you explain a bit why READ isn't supported? Because WRITE zero > copy is supposed to be supported easily with splice based approach, > and I am more interested in READ zc actually. No special reason, READ op can also be supported. I'll add this support in patch set v2. For this RFC patch set, I just tried to show the idea, so I must admit that current codes are not mature enough :) > >> + >> if (req_op(req) != REQ_OP_WRITE && req_op(req) != REQ_OP_FLUSH) >> return rq_bytes; >> >> @@ -860,6 +921,89 @@ static void ublk_queue_cmd(struct ublk_queue *ubq, struct request *rq) >> } >> } >> >> >> + kbuf->bvec = bvec; >> + rq_for_each_bvec(tmp, rq, rq_iter) { >> + *bvec = tmp; >> + bvec++; >> + } >> + >> + kbuf->count = blk_rq_bytes(rq); >> + kbuf->nr_bvecs = nr_bvec; >> + data->kbuf = kbuf; >> + return 0; > bio/req bvec table is immutable, so here you can pass its reference > to kbuf directly. Yeah, thanks. > >> +} >> + >> +static int ublk_run_bpf_prog(struct ublk_queue *ubq, struct request *rq) >> +{ >> + int err; >> + struct ublk_device *ub = ubq->dev; >> + struct bpf_prog *prog = ub->io_prep_prog; >> + struct ublk_io_bpf_ctx *bpf_ctx; >> + >> + if (!prog) >> + return 0; >> + >> + bpf_ctx = kmalloc(sizeof(struct ublk_io_bpf_ctx), GFP_NOIO); >> + if (!bpf_ctx) >> + return -EIO; >> + >> + err = ublk_init_uring_kbuf(rq); >> + if (err < 0) { >> + kfree(bpf_ctx); >> + return -EIO; >> + } >> + bpf_ctx->ub = ub; >> + bpf_ctx->ctx.q_id = ubq->q_id; >> + bpf_ctx->ctx.tag = rq->tag; >> + bpf_ctx->ctx.op = req_op(rq); >> + bpf_ctx->ctx.nr_sectors = blk_rq_sectors(rq); >> + bpf_ctx->ctx.start_sector = blk_rq_pos(rq); > The above is for setting up target io parameter, which is supposed > to be from userspace, cause it is result of user space logic. If > these parameters are from kernel, the whole logic has to be done > in io_prep_prog. Yeah, it's designed that io_prep_prog implements user space io logic. > >> + bpf_prog_run_pin_on_cpu(prog, bpf_ctx); >> + >> + init_task_work(&bpf_ctx->work, ublk_bpf_io_submit_fn); >> + if (task_work_add(ubq->ubq_daemon, &bpf_ctx->work, TWA_SIGNAL_NO_IPI)) >> + kfree(bpf_ctx); > task_work_add() is only available in case of ublk builtin. Yeah, I'm thinking how to work around it. > >> + return 0; >> +} >> + >> static blk_status_t ublk_queue_rq(struct blk_mq_hw_ctx *hctx, >> const struct blk_mq_queue_data *bd) >> { >> @@ -872,6 +1016,9 @@ static blk_status_t ublk_queue_rq(struct blk_mq_hw_ctx *hctx, >> if (unlikely(res != BLK_STS_OK)) >> return BLK_STS_IOERR; >> >> + /* Currently just for test. */ >> + ublk_run_bpf_prog(ubq, rq); > Can you explain the above comment a bit? When is the io_prep_prog called > in the non-test version? Or can you post the non-test version in list > for review. Forgot to delete stale comments, sorry. I'm writing v2 patch set, > Here it is the key for understanding the whole idea, especially when > is io_prep_prog called finally? How to pass parameters to io_prep_prog? Let me explain more about the design: io_prep_prog has two types of parameters: 1) its call argument: struct ublk_bpf_ctx, see ublk.bpf.c. ublk_bpf_ctx will describe one kernel io requests about its op, qid, sectors info. io_prep_prog uses these info to map target io. 2) ebpf map structure, user space daemon can use map structure to pass much information from user space to io_prep_prog, which will help it to initialize target io if necessary. io_prep_prog is called when ublk_queue_rq() is called, this bpf prog will initialize one or more sqes according to user logic, and io_prep_prog will put these sqes in an ebpf map structure, then execute a task_work_add() to notify ubq_daemon to execute io_submit_prog. Note, we can not call io_uring_submit_sqe() in task context that calls ublk_queue_rq(), that context does not have io_uring instance owned by ubq_daemon. Later ubq_daemon will call io_submit_prog to submit sqes. > > Given it is ebpf prog, I don't think any userspace parameter can be > passed to io_prep_prog when submitting IO, that means all user logic has > to be done inside io_prep_prog? If yes, not sure if it is one good way, > cause ebpf prog is very limited programming environment, but the user > logic could be as complicated as using btree to map io, or communicating > with remote machine for figuring out the mapping. Loop is just the > simplest direct mapping. Currently, we can use ebpf map structure to pass user space parameter to io_prep_prog. Also I agree with you that complicated logic maybe hard to be implemented in ebpf prog, hope ebpf community will improve this situation gradually. For userspace target implementations I met so far, they just use userspace block device solutions to visit distributed filesystem, involves socket programming and have simple map io logic. We can prepare socket fd in ebpf map structure, and these map io logic should be easily implemented in ebpf prog, though I don't apply this ebpf method to our internal business yet. Thanks for review. Regards, Xiaoguang Wang > > > Thanks, > Ming