[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CAEf4BzbwbqsWyc7cuk550Cg+F35rQFzAuLE+oJ6roBOcA=4jeA@mail.gmail.com>
Date: Mon, 3 May 2021 15:47:32 -0700
From: Andrii Nakryiko <andrii.nakryiko@...il.com>
To: Kumar Kartikeya Dwivedi <memxor@...il.com>
Cc: bpf <bpf@...r.kernel.org>,
Toke Høiland-Jørgensen <toke@...hat.com>,
Alexei Starovoitov <ast@...nel.org>,
Daniel Borkmann <daniel@...earbox.net>,
Andrii Nakryiko <andrii@...nel.org>,
Martin KaFai Lau <kafai@...com>,
Song Liu <songliubraving@...com>, Yonghong Song <yhs@...com>,
John Fastabend <john.fastabend@...il.com>,
KP Singh <kpsingh@...nel.org>,
"David S. Miller" <davem@...emloft.net>,
Jakub Kicinski <kuba@...nel.org>,
Jesper Dangaard Brouer <brouer@...hat.com>,
Shaun Crampton <shaun@...era.io>,
Networking <netdev@...r.kernel.org>
Subject: Re: [PATCH bpf-next v5 1/3] libbpf: add netlink helpers
On Fri, Apr 30, 2021 at 11:13 PM Kumar Kartikeya Dwivedi
<memxor@...il.com> wrote:
>
> On Sat, May 01, 2021 at 12:34:39AM IST, Andrii Nakryiko wrote:
> > On Wed, Apr 28, 2021 at 9:26 AM Kumar Kartikeya Dwivedi
> > <memxor@...il.com> wrote:
> > >
> > > This change introduces a few helpers to wrap open coded attribute
> > > preparation in netlink.c. It also adds a libbpf_nl_send_recv that is useful
> > > to wrap send + recv handling in a generic way. Subsequent patch will
> > > also use this function for sending and receiving a netlink response.
> > > The libbpf_nl_get_link helper has been removed instead, moving socket
> > > creation into the newly named libbpf_nl_send_recv.
> > >
> > > Every nested attribute's closure must happen using the helper
> > > nlattr_end_nested, which sets its length properly. NLA_F_NESTED is
> > > enforced using nlattr_begin_nested helper. Other simple attributes
> > > can be added directly.
> > >
> > > The maxsz parameter corresponds to the size of the request structure
> > > which is being filled in, so for instance with req being:
> > >
> > > struct {
> > > struct nlmsghdr nh;
> > > struct tcmsg t;
> > > char buf[4096];
> > > } req;
> > >
> > > Then, maxsz should be sizeof(req).
> > >
> > > This change also converts the open coded attribute preparation with the
> > > helpers. Note that the only failure the internal call to nlattr_add
> > > could result in the nested helper would be -EMSGSIZE, hence that is what
> > > we return to our caller.
> > >
> > > The libbpf_nl_send_recv call takes care of opening the socket, sending the
> > > netlink message, receiving the response, potentially invoking callbacks,
> > > and return errors if any, and then finally close the socket. This allows
> > > users to avoid identical socket setup code in different places. The only
> > > user of libbpf_nl_get_link has been converted to make use of it.
> > >
> > > __bpf_set_link_xdp_fd_replace has also been refactored to use it.
> > >
> > > Reviewed-by: Toke Høiland-Jørgensen <toke@...hat.com>
> > > Signed-off-by: Kumar Kartikeya Dwivedi <memxor@...il.com>
> > > ---
> > > tools/lib/bpf/netlink.c | 117 ++++++++++++++++++----------------------
> > > tools/lib/bpf/nlattr.h | 48 +++++++++++++++++
> > > 2 files changed, 100 insertions(+), 65 deletions(-)
> > >
> > > diff --git a/tools/lib/bpf/netlink.c b/tools/lib/bpf/netlink.c
> > > index d2cb28e9ef52..6daee6640725 100644
> > > --- a/tools/lib/bpf/netlink.c
> > > +++ b/tools/lib/bpf/netlink.c
> > > @@ -131,72 +131,53 @@ static int bpf_netlink_recv(int sock, __u32 nl_pid, int seq,
> > > return ret;
> > > }
> > >
> > > +static int libbpf_nl_send_recv(struct nlmsghdr *nh, __dump_nlmsg_t fn,
> > > + libbpf_dump_nlmsg_t _fn, void *cookie);
> > > +
> > > static int __bpf_set_link_xdp_fd_replace(int ifindex, int fd, int old_fd,
> > > __u32 flags)
> > > {
> > > - int sock, seq = 0, ret;
> > > - struct nlattr *nla, *nla_xdp;
> > > + struct nlattr *nla;
> > > + int ret;
> > > struct {
> > > struct nlmsghdr nh;
> > > struct ifinfomsg ifinfo;
> > > char attrbuf[64];
> > > } req;
> > > - __u32 nl_pid = 0;
> > > -
> > > - sock = libbpf_netlink_open(&nl_pid);
> > > - if (sock < 0)
> > > - return sock;
> > >
> > > memset(&req, 0, sizeof(req));
> > > req.nh.nlmsg_len = NLMSG_LENGTH(sizeof(struct ifinfomsg));
> > > req.nh.nlmsg_flags = NLM_F_REQUEST | NLM_F_ACK;
> > > req.nh.nlmsg_type = RTM_SETLINK;
> > > - req.nh.nlmsg_pid = 0;
> > > - req.nh.nlmsg_seq = ++seq;
> > > req.ifinfo.ifi_family = AF_UNSPEC;
> > > req.ifinfo.ifi_index = ifindex;
> > >
> > > /* started nested attribute for XDP */
> > > - nla = (struct nlattr *)(((char *)&req)
> > > - + NLMSG_ALIGN(req.nh.nlmsg_len));
> > > - nla->nla_type = NLA_F_NESTED | IFLA_XDP;
> > > - nla->nla_len = NLA_HDRLEN;
> > > + nla = nlattr_begin_nested(&req.nh, sizeof(req), IFLA_XDP);
> > > + if (!nla)
> > > + return -EMSGSIZE;
> > >
> > > /* add XDP fd */
> > > - nla_xdp = (struct nlattr *)((char *)nla + nla->nla_len);
> > > - nla_xdp->nla_type = IFLA_XDP_FD;
> > > - nla_xdp->nla_len = NLA_HDRLEN + sizeof(int);
> > > - memcpy((char *)nla_xdp + NLA_HDRLEN, &fd, sizeof(fd));
> > > - nla->nla_len += nla_xdp->nla_len;
> > > + ret = nlattr_add(&req.nh, sizeof(req), IFLA_XDP_FD, &fd, sizeof(fd));
> > > + if (ret < 0)
> > > + return ret;
> > >
> > > /* if user passed in any flags, add those too */
> > > if (flags) {
> > > - nla_xdp = (struct nlattr *)((char *)nla + nla->nla_len);
> > > - nla_xdp->nla_type = IFLA_XDP_FLAGS;
> > > - nla_xdp->nla_len = NLA_HDRLEN + sizeof(flags);
> > > - memcpy((char *)nla_xdp + NLA_HDRLEN, &flags, sizeof(flags));
> > > - nla->nla_len += nla_xdp->nla_len;
> > > + ret = nlattr_add(&req.nh, sizeof(req), IFLA_XDP_FLAGS, &flags, sizeof(flags));
> > > + if (ret < 0)
> > > + return ret;
> > > }
> > >
> > > if (flags & XDP_FLAGS_REPLACE) {
> > > - nla_xdp = (struct nlattr *)((char *)nla + nla->nla_len);
> > > - nla_xdp->nla_type = IFLA_XDP_EXPECTED_FD;
> > > - nla_xdp->nla_len = NLA_HDRLEN + sizeof(old_fd);
> > > - memcpy((char *)nla_xdp + NLA_HDRLEN, &old_fd, sizeof(old_fd));
> > > - nla->nla_len += nla_xdp->nla_len;
> > > + ret = nlattr_add(&req.nh, sizeof(req), IFLA_XDP_EXPECTED_FD, &flags, sizeof(flags));
> >
> > shouldn't old_fd be used here?
> >
>
> Ouch, yes, thanks for spotting this.
>
> > > + if (ret < 0)
> > > + return ret;
> > > }
> > >
> > > - req.nh.nlmsg_len += NLA_ALIGN(nla->nla_len);
> > > + nlattr_end_nested(&req.nh, nla);
> > >
> > > - if (send(sock, &req, req.nh.nlmsg_len, 0) < 0) {
> > > - ret = -errno;
> > > - goto cleanup;
> > > - }
> > > - ret = bpf_netlink_recv(sock, nl_pid, seq, NULL, NULL, NULL);
> > > -
> > > -cleanup:
> > > - close(sock);
> > > - return ret;
> > > + return libbpf_nl_send_recv(&req.nh, NULL, NULL, NULL);
> > > }
> > >
> > > int bpf_set_link_xdp_fd_opts(int ifindex, int fd, __u32 flags,
> >
> > [...]
> >
> > > -int libbpf_nl_get_link(int sock, unsigned int nl_pid,
> > > - libbpf_dump_nlmsg_t dump_link_nlmsg, void *cookie)
> > > +static int libbpf_nl_send_recv(struct nlmsghdr *nh, __dump_nlmsg_t fn,
> > > + libbpf_dump_nlmsg_t _fn, void *cookie)
> > > {
> > > - struct {
> > > - struct nlmsghdr nlh;
> > > - struct ifinfomsg ifm;
> > > - } req = {
> > > - .nlh.nlmsg_len = NLMSG_LENGTH(sizeof(struct ifinfomsg)),
> > > - .nlh.nlmsg_type = RTM_GETLINK,
> > > - .nlh.nlmsg_flags = NLM_F_DUMP | NLM_F_REQUEST,
> > > - .ifm.ifi_family = AF_PACKET,
> > > - };
> > > - int seq = time(NULL);
> > > + __u32 nl_pid = 0;
> > > + int sock, ret;
> > >
> > > - req.nlh.nlmsg_seq = seq;
> > > - if (send(sock, &req, req.nlh.nlmsg_len, 0) < 0)
> > > - return -errno;
> > > + if (!nh)
> > > + return -EINVAL;
> > > +
> > > + sock = libbpf_netlink_open(&nl_pid);
> > > + if (sock < 0)
> > > + return sock;
> > >
> > > - return bpf_netlink_recv(sock, nl_pid, seq, __dump_link_nlmsg,
> > > - dump_link_nlmsg, cookie);
> > > + nh->nlmsg_pid = 0;
> > > + nh->nlmsg_seq = time(NULL);
> > > + if (send(sock, nh, nh->nlmsg_len, 0) < 0) {
> > > + ret = -errno;
> > > + goto end;
> > > + }
> > > +
> > > + ret = bpf_netlink_recv(sock, nl_pid, nh->nlmsg_seq, fn, _fn, cookie);
> >
> > what's the difference between fn and _fn, can this be somehow
> > reflected in the name?
> >
>
> You can use fn as a common parsing function for the same RTM_GET* message, and
> then use _fn to parse a nested layer of attributes below it to fill in different
> kind of opts (through the cookie user data parameter).
>
> How about outer_cb, inner_cb?
so the outer thingy is "message" and internal one is "attribute" in
netlink lingo? If yes, then parse_msg and parse_attr would make more
sense, imo. If not, outer_cb/inner_cb is fine as well.
>
> > > +
> > > +end:
> > > + close(sock);
> > > + return ret;
> > > }
> > > diff --git a/tools/lib/bpf/nlattr.h b/tools/lib/bpf/nlattr.h
> > > index 6cc3ac91690f..1c94cdb6e89d 100644
> > > --- a/tools/lib/bpf/nlattr.h
> > > +++ b/tools/lib/bpf/nlattr.h
> > > @@ -10,7 +10,10 @@
> > > #define __LIBBPF_NLATTR_H
> > >
> > > #include <stdint.h>
> > > +#include <string.h>
> > > +#include <errno.h>
> > > #include <linux/netlink.h>
> > > +
> > > /* avoid multiple definition of netlink features */
> > > #define __LINUX_NETLINK_H
> > >
> > > @@ -103,4 +106,49 @@ int libbpf_nla_parse_nested(struct nlattr *tb[], int maxtype,
> > >
> > > int libbpf_nla_dump_errormsg(struct nlmsghdr *nlh);
> > >
> > > +static inline struct nlattr *nla_data(struct nlattr *nla)
> > > +{
> > > + return (struct nlattr *)((char *)nla + NLA_HDRLEN);
> > > +}
> > > +
> > > +static inline struct nlattr *nh_tail(struct nlmsghdr *nh)
> > > +{
> > > + return (struct nlattr *)((char *)nh + NLMSG_ALIGN(nh->nlmsg_len));
> > > +}
> > > +
> > > +static inline int nlattr_add(struct nlmsghdr *nh, size_t maxsz, int type,
> > > + const void *data, int len)
> > > +{
> > > + struct nlattr *nla;
> > > +
> > > + if (NLMSG_ALIGN(nh->nlmsg_len) + NLA_ALIGN(NLA_HDRLEN + len) > maxsz)
> > > + return -EMSGSIZE;
> > > + if ((!data && len) || (data && !len))
> >
> > we use !!data != !!len for this in at least few places
> >
>
> Ok.
>
> > > + return -EINVAL;
> > > +
> > > + nla = nh_tail(nh);
> > > + nla->nla_type = type;
> > > + nla->nla_len = NLA_HDRLEN + len;
> > > + if (data)
> > > + memcpy(nla_data(nla), data, len);
> > > + nh->nlmsg_len = NLMSG_ALIGN(nh->nlmsg_len) + NLA_ALIGN(nla->nla_len);
> > > + return 0;
> > > +}
> > > +
> >
> > [...]
>
> --
> Kartikeya
Powered by blists - more mailing lists