[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <Y2hZwWdY28bCn+iT@T590>
Date: Mon, 7 Nov 2022 09:05:05 +0800
From: Ming Lei <ming.lei@...hat.com>
To: Bernd Schubert <bernd.schubert@...tmail.fm>
Cc: Jens Axboe <axboe@...nel.dk>, io-uring@...r.kernel.org,
linux-block@...r.kernel.org, linux-kernel@...r.kernel.org,
linux-fsdevel@...r.kernel.org,
Miklos Szeredi <mszeredi@...hat.com>,
Stefan Hajnoczi <stefanha@...hat.com>,
ZiyangZhang <ZiyangZhang@...ux.alibaba.com>
Subject: Re: [RFC PATCH 4/4] ublk_drv: support splice based read/write zero
copy
On Sat, Nov 05, 2022 at 12:37:21AM +0100, Bernd Schubert wrote:
>
>
> On 11/4/22 01:44, Ming Lei wrote:
> > On Thu, Nov 03, 2022 at 11:28:29PM +0100, Bernd Schubert wrote:
> > >
> > >
> > > On 11/3/22 09:50, Ming Lei wrote:
> > > > Pass ublk block IO request pages to kernel backend IO handling code via
> > > > pipe, and request page copy can be avoided. So far, the existed
> > > > pipe/splice mechanism works for handling write request only.
> > > >
> > > > The initial idea of using splice for zero copy is from Miklos and Stefan.
> > > >
> > > > Read request's zero copy requires pipe's change to allow one read end to
> > > > produce buffers for another read end to consume. The added SPLICE_F_READ_TO_READ
> > > > flag is for supporting this feature.
> > > >
> > > > READ is handled by sending IORING_OP_SPLICE with SPLICE_F_DIRECT |
> > > > SPLICE_F_READ_TO_READ. WRITE is handled by sending IORING_OP_SPLICE with
> > > > SPLICE_F_DIRECT. Kernel internal pipe is used for simplifying userspace,
> > > > meantime potential info leak could be avoided.
> > >
> > >
> > > Sorry to ask, do you have an ublk branch that gives an example how to use
> > > this?
> >
> > Follows the ublk splice-zc branch:
> >
> > https://github.com/ming1/ubdsrv/commits/splice-zc
> >
> > which is mentioned in cover letter, but I guess it should be added to
> > here too, sorry for that, so far only ublk-loop supports it by:
> >
> > ublk add -t loop -f $BACKING -z
> >
> > without '-z', ublk-loop is created with zero copy disabled.
>
> Ah, thanks a lot! And sorry, I had missed this part in the cover letter.
>
> I will take a look on your new zero copy code on Monday.
>
>
> >
> > >
> > > I still have several things to fix in my branches, but I got basic fuse
> > > uring with copies working. Adding back splice would be next after posting
> > > rfc patches. My initial assumption was that I needed to duplicate everything
> > > splice does into the fuse .uring_cmd handler - obviously there is a better
> > > way with your patches.
> > >
> > > This week I have a few days off, by end of next week or the week after I
> > > might have patches in an rfc state (one thing I'm going to ask about is how
> > > do I know what is the next CQE in the kernel handler - ublk does this with
> > > tags through mq, but I don't understand yet where the tag is increased and
> > > what the relation between tag and right CQE order is).
> >
> > tag is one attribute of io request, which is originated from ublk
> > driver, and it is unique for each request among one queue. So ublksrv
> > won't change it at all, just use it, and ublk driver guarantees that
> > it is unique.
> >
> > In ublkserv implementation, the tag info is set in cqe->user_data, so
> > we can retrieve the io request via tag part of cqe->user_data.
>
> Yeah, this is the easy part I understood. At least I hope so :)
>
> >
> > Also I may not understand your question of 'the relation between tag and right
> > CQE order', io_uring provides IOSQE_IO_DRAIN/IOSQE_IO_LINK for ordering
> > SQE, and ublksrv only applies IOSQE_IO_LINK in ublk-qcow2, so care to
> > explain it in a bit details about the "the relation between tag and right
> > CQE order"?
>
>
> For fuse (kernel) a vfs request comes in and I need to choose a command in
> the ring queue. Right now this is just an atomic counter % queue_size
>
> fuse_request_alloc_ring()
> req_cnt = atomic_inc_return(&queue->req_cnt);
> tag = req_cnt & (fc->ring.queue_depth - 1); /* cnt % queue_depth */
>
> ring_req = &queue->ring_req[tag];
>
>
>
> I might be wrong, but I think that can be compared a bit to ublk_queue_rq().
> Looks like ublk_queue_rq gets called in blk-mq context and blk-mq seems to
> provide rq->tag, which then determines the command in the ring queue -
> completion of commands is done in tag-order provided by blk-mq? The part I
The two are not related, blk-mq tag number means nothing wrt. io
handling order:
- tag is allocated via sbitmap, which may return tag number in any
order, you may think the returned number is just random
- blk-mq may re-order requests and dispatch them with any order
- once requests are issued to io_uring, userspace may handles these IOs
with any order
- after backend io is queued via io_uring or libaio or whatever to kernel, it
could be completed at any order
> didn't figure out yet is where the tag value gets set.
> Also interesting is that there is no handler if the ring is already full -
> like the ublk_io command is currently busy in ublksrv (user space). Handled
> auto-magically with blk-mq?
For ublk, the queue has fixed depth, so the pre-allocated io_uring size is
enough, and blk-mq can throttle IOs from the beginning if the max queue depth is
reached, so ublk needn't to worry about io_uring size/depth.
But fuse may have to consider request throttle.
Thanks,
Ming
Powered by blists - more mailing lists