[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20171103205610.GA31088@breakpoint.cc>
Date: Fri, 3 Nov 2017 21:56:10 +0100
From: Florian Westphal <fw@...len.de>
To: Pablo Neira Ayuso <pablo@...filter.org>
Cc: netfilter-devel@...r.kernel.org, netdev@...r.kernel.org
Subject: Re: [PATCH RFC,WIP 5/5] netfilter: nft_flow_offload: add ndo hooks
for hardware offload
Pablo Neira Ayuso <pablo@...filter.org> wrote:
> +static void flow_offload_work(struct work_struct *work)
> +{
> + struct flow_hw_offload *offload, *next;
> +
> + spin_lock_bh(&flow_hw_offload_lock);
> + list_for_each_entry_safe(offload, next, &flow_hw_offload_pending_list, list) {
> + do_flow_offload(offload->flow);
This should not offload flows that already have DYING bit set.
> + nf_conntrack_put(&offload->ct->ct_general);
> + list_del(&offload->list);
> + kfree(offload);
> + }
> + spin_unlock_bh(&flow_hw_offload_lock);
> +
> + queue_delayed_work(system_power_efficient_wq, &nft_flow_offload_dwork, HZ);
> +}
Missed this on first round, 1 second is quite large.
[..]
> static int nft_flow_route(const struct nft_pktinfo *pkt,
> const struct nf_conn *ct,
> union flow_gateway *orig_gw,
> @@ -211,6 +290,7 @@ static void nft_flow_offload_eval(const struct nft_expr *expr,
> union flow_gateway orig_gateway, reply_gateway;
> struct net_device *outdev = pkt->xt.state->out;
> struct net_device *indev = pkt->xt.state->in;
> + struct flow_hw_offload *offload;
> enum ip_conntrack_info ctinfo;
> struct flow_offload *flow;
> struct nf_conn *ct;
> @@ -250,6 +330,21 @@ static void nft_flow_offload_eval(const struct nft_expr *expr,
> if (ret < 0)
> goto err2;
>
> + if (!indev->netdev_ops->ndo_flow_add)
> + return;
> +
> + offload = kmalloc(sizeof(struct flow_hw_offload), GFP_ATOMIC);
> + if (!offload)
> + return;
> +
> + nf_conntrack_get(&ct->ct_general);
> + offload->ct = ct;
> + offload->flow = flow;
> +
> + spin_lock_bh(&flow_hw_offload_lock);
> + list_add_tail(&offload->list, &flow_hw_offload_pending_list);
> + spin_unlock_bh(&flow_hw_offload_lock);
> +
> return;
So this aims for lazy offloading (up to 1 second delay).
Is this intentional, e.g. to avoid offloading short-lived 'RR' flows?
I would have expected this to schedule the workqueue here, and not use
delayed wq at all (i.e., also no self-rescheduling from
flow_offload_work()).
Powered by blists - more mailing lists