[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CAEf4BzaBw93YEU7CktQp8FkphsBYRGRHpsF-D-BjWHG1UXoakA@mail.gmail.com>
Date: Tue, 14 Jan 2020 22:39:45 -0800
From: Andrii Nakryiko <andrii.nakryiko@...il.com>
To: Martin Lau <kafai@...com>
Cc: bpf <bpf@...r.kernel.org>, Alexei Starovoitov <ast@...nel.org>,
Daniel Borkmann <daniel@...earbox.net>,
David Miller <davem@...emloft.net>,
Kernel Team <Kernel-team@...com>,
Networking <netdev@...r.kernel.org>
Subject: Re: [PATCH bpf-next 5/5] bpftool: Support dumping a map with btf_vmlinux_value_type_id
On Tue, Jan 14, 2020 at 10:04 PM Martin Lau <kafai@...com> wrote:
>
> On Tue, Jan 14, 2020 at 05:49:00PM -0800, Andrii Nakryiko wrote:
> > On Tue, Jan 14, 2020 at 2:46 PM Martin KaFai Lau <kafai@...com> wrote:
> > >
> > > This patch makes bpftool support dumping a map's value properly
> > > when the map's value type is a type of the running kernel's btf.
> > > (i.e. map_info.btf_vmlinux_value_type_id is set instead of
> > > map_info.btf_value_type_id). The first usecase is for the
> > > BPF_MAP_TYPE_STRUCT_OPS.
> > >
> > > Signed-off-by: Martin KaFai Lau <kafai@...com>
> > > ---
> > > tools/bpf/bpftool/map.c | 43 +++++++++++++++++++++++++++++++----------
> > > 1 file changed, 33 insertions(+), 10 deletions(-)
> > >
> > > diff --git a/tools/bpf/bpftool/map.c b/tools/bpf/bpftool/map.c
> > > index 4c5b15d736b6..d25f3b2355ad 100644
> > > --- a/tools/bpf/bpftool/map.c
> > > +++ b/tools/bpf/bpftool/map.c
> > > @@ -20,6 +20,7 @@
> > > #include "btf.h"
> > > #include "json_writer.h"
> > > #include "main.h"
> > > +#include "libbpf_internal.h"
> > >
> > > const char * const map_type_name[] = {
> > > [BPF_MAP_TYPE_UNSPEC] = "unspec",
> > > @@ -252,6 +253,7 @@ static int do_dump_btf(const struct btf_dumper *d,
> > > struct bpf_map_info *map_info, void *key,
> > > void *value)
> > > {
> > > + __u32 value_id;
> > > int ret;
> > >
> > > /* start of key-value pair */
> > > @@ -265,9 +267,12 @@ static int do_dump_btf(const struct btf_dumper *d,
> > > goto err_end_obj;
> > > }
> > >
> > > + value_id = map_info->btf_vmlinux_value_type_id ?
> > > + : map_info->btf_value_type_id;
> > > +
> > > if (!map_is_per_cpu(map_info->type)) {
> > > jsonw_name(d->jw, "value");
> > > - ret = btf_dumper_type(d, map_info->btf_value_type_id, value);
> > > + ret = btf_dumper_type(d, value_id, value);
> > > } else {
> > > unsigned int i, n, step;
> > >
> > > @@ -279,8 +284,7 @@ static int do_dump_btf(const struct btf_dumper *d,
> > > jsonw_start_object(d->jw);
> > > jsonw_int_field(d->jw, "cpu", i);
> > > jsonw_name(d->jw, "value");
> > > - ret = btf_dumper_type(d, map_info->btf_value_type_id,
> > > - value + i * step);
> > > + ret = btf_dumper_type(d, value_id, value + i * step);
> > > jsonw_end_object(d->jw);
> > > if (ret)
> > > break;
> > > @@ -932,6 +936,27 @@ static int maps_have_btf(int *fds, int nb_fds)
> > > return 1;
> > > }
> > >
> > > +static struct btf *get_map_kv_btf(const struct bpf_map_info *info)
> > > +{
> > > + struct btf *btf = NULL;
> > > +
> > > + if (info->btf_vmlinux_value_type_id) {
> > > + btf = bpf_find_kernel_btf();
> >
> > If there are multiple maps we are dumping, it might become quite
> > costly to re-read and re-parse kernel BTF all the time. Can we lazily
> > load it, when required,
> It is loaded lazily.
>
yeah, it was meant as "lazy load and cache" vs "pre-load always"
(which makes caching simpler).
> > and cache instead?
> Cache it in bpftool/map.c? Sure.
yeah, for the duration of dumping
>
> >
> > > + if (IS_ERR(btf))
> > > + p_err("failed to get kernel btf");
> > > + } else if (info->btf_value_type_id) {
> > > + int err;
> > > +
> > > + err = btf__get_from_id(info->btf_id, &btf);
> > > + if (err || !btf) {
> > > + p_err("failed to get btf");
> > > + btf = err ? ERR_PTR(err) : ERR_PTR(-ESRCH);
> > > + }
> > > + }
> > > +
> > > + return btf;
> > > +}
> > > +
> > > static int
> > > map_dump(int fd, struct bpf_map_info *info, json_writer_t *wtr,
> > > bool show_header)
> > > @@ -952,13 +977,11 @@ map_dump(int fd, struct bpf_map_info *info, json_writer_t *wtr,
> > > prev_key = NULL;
> > >
> > > if (wtr) {
> > > - if (info->btf_id) {
> > > - err = btf__get_from_id(info->btf_id, &btf);
> > > - if (err || !btf) {
> > > - err = err ? : -ESRCH;
> > > - p_err("failed to get btf");
> > > - goto exit_free;
> > > - }
> > > + btf = get_map_kv_btf(info);
> > > + if (IS_ERR(btf)) {
> > > + err = PTR_ERR(btf);
> > > + btf = NULL;
> > > + goto exit_free;
> > > }
> > >
> > > if (show_header) {
> > > --
> > > 2.17.1
> > >
Powered by blists - more mailing lists