[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <87a6ieqssc.fsf@toke.dk>
Date: Mon, 08 Nov 2021 21:46:59 +0100
From: Toke Høiland-Jørgensen <toke@...hat.com>
To: Lorenzo Bianconi <lorenzo@...nel.org>
Cc: Jakub Kicinski <kuba@...nel.org>, bpf@...r.kernel.org,
netdev@...r.kernel.org, lorenzo.bianconi@...hat.com,
davem@...emloft.net, ast@...nel.org, daniel@...earbox.net,
shayagr@...zon.com, john.fastabend@...il.com, dsahern@...nel.org,
brouer@...hat.com, echaudro@...hat.com, jasowang@...hat.com,
alexander.duyck@...il.com, saeed@...nel.org,
maciej.fijalkowski@...el.com, magnus.karlsson@...el.com,
tirthendu.sarkar@...el.com
Subject: Re: [PATCH v17 bpf-next 12/23] bpf: add multi-buff support to the
bpf_xdp_adjust_tail() API
Lorenzo Bianconi <lorenzo@...nel.org> writes:
>> Lorenzo Bianconi <lorenzo@...nel.org> writes:
>>
>> >> On Thu, 4 Nov 2021 18:35:32 +0100 Lorenzo Bianconi wrote:
>> >> > This change adds support for tail growing and shrinking for XDP multi-buff.
>> >> >
>> >> > When called on a multi-buffer packet with a grow request, it will always
>> >> > work on the last fragment of the packet. So the maximum grow size is the
>> >> > last fragments tailroom, i.e. no new buffer will be allocated.
>> >> >
>> >> > When shrinking, it will work from the last fragment, all the way down to
>> >> > the base buffer depending on the shrinking size. It's important to mention
>> >> > that once you shrink down the fragment(s) are freed, so you can not grow
>> >> > again to the original size.
>> >>
>> >> > +static int bpf_xdp_mb_increase_tail(struct xdp_buff *xdp, int offset)
>> >> > +{
>> >> > + struct skb_shared_info *sinfo = xdp_get_shared_info_from_buff(xdp);
>> >> > + skb_frag_t *frag = &sinfo->frags[sinfo->nr_frags - 1];
>> >> > + int size, tailroom;
>> >> > +
>> >> > + tailroom = xdp->frame_sz - skb_frag_size(frag) - skb_frag_off(frag);
>> >>
>> >> I know I complained about this before but the assumption that we can
>> >> use all the space up to xdp->frame_sz makes me uneasy.
>> >>
>> >> Drivers may not expect the idea that core may decide to extend the
>> >> last frag.. I don't think the skb path would ever do this.
>> >>
>> >> How do you feel about any of these options:
>> >> - dropping this part for now (return an error for increase)
>> >> - making this an rxq flag or reading the "reserved frag size"
>> >> from rxq (so that drivers explicitly opt-in)
>> >> - adding a test that can be run on real NICs
>> >> ?
>> >
>> > I think this has been added to be symmetric with bpf_xdp_adjust_tail().
>> > I do think there is a real use-case for it so far so I am fine to just
>> > support the shrink part.
>> >
>> > @Eelco, Jesper, Toke: any comments on it?
>>
>> Well, tail adjust is useful for things like encapsulations that need to
>> add a trailer. Don't see why that wouldn't be something people would
>> want to do for jumboframes as well?
>>
>
> I agree this would be useful for protocols that add a trailer.
>
>> Not sure I get what the issue is with this either? But having a test
>> that can be run to validate this on hardware would be great in any case,
>> I suppose - we've been discussing more general "compliance tests" for
>> XDP before...
>
> what about option 2? We can add a frag_size field to rxq [0] that is set by
> the driver initializing the xdp_buff. frag_size set to 0 means we can use
> all the buffer.
>
> Regards,
> Lorenzo
>
> [0] pahole -C xdp_rxq_info vmlinux
> struct xdp_rxq_info {
> struct net_device * dev; /* 0 8 */
> u32 queue_index; /* 8 4 */
> u32 reg_state; /* 12 4 */
> struct xdp_mem_info mem; /* 16 8 */
> unsigned int napi_id; /* 24 4 */
>
> /* size: 64, cachelines: 1, members: 5 */
> /* padding: 36 */
> } __attribute__((__aligned__(64)));
Works for me :)
-Toke
Powered by blists - more mailing lists