[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <79a6e702-9fe7-3334-387e-82b5048b3275@fb.com>
Date: Thu, 9 Jul 2020 23:31:51 -0700
From: Yonghong Song <yhs@...com>
To: Andrii Nakryiko <andrii.nakryiko@...il.com>,
Alexei Starovoitov <alexei.starovoitov@...il.com>
CC: Linus Torvalds <torvalds@...ux-foundation.org>,
"David S. Miller" <davem@...emloft.net>,
Daniel Borkmann <daniel@...earbox.net>,
"Eric W. Biederman" <ebiederm@...ssion.com>,
Networking <netdev@...r.kernel.org>, bpf <bpf@...r.kernel.org>,
Kernel Team <kernel-team@...com>
Subject: Re: [PATCH bpf-next 3/3] bpf: Add kernel module with user mode driver
that populates bpffs.
On 7/8/20 8:15 PM, Andrii Nakryiko wrote:
> On Thu, Jul 2, 2020 at 1:04 PM Alexei Starovoitov
> <alexei.starovoitov@...il.com> wrote:
>>
>> From: Alexei Starovoitov <ast@...nel.org>
>>
>> Add kernel module with user mode driver that populates bpffs with
>> BPF iterators.
>>
>> $ mount bpffs /sys/fs/bpf/ -t bpf
>> $ ls -la /sys/fs/bpf/
>> total 4
>> drwxrwxrwt 2 root root 0 Jul 2 00:27 .
>> drwxr-xr-x 19 root root 4096 Jul 2 00:09 ..
>> -rw------- 1 root root 0 Jul 2 00:27 maps
>> -rw------- 1 root root 0 Jul 2 00:27 progs
>>
>> The user mode driver will load BPF Type Formats, create BPF maps, populate BPF
>> maps, load two BPF programs, attach them to BPF iterators, and finally send two
>> bpf_link IDs back to the kernel.
>> The kernel will pin two bpf_links into newly mounted bpffs instance under
>> names "progs" and "maps". These two files become human readable.
>>
>> $ cat /sys/fs/bpf/progs
>> id name pages attached
>> 11 dump_bpf_map 1 bpf_iter_bpf_map
>> 12 dump_bpf_prog 1 bpf_iter_bpf_prog
>> 27 test_pkt_access 1
>> 32 test_main 1 test_pkt_access test_pkt_access
>> 33 test_subprog1 1 test_pkt_access_subprog1 test_pkt_access
>> 34 test_subprog2 1 test_pkt_access_subprog2 test_pkt_access
>> 35 test_subprog3 1 test_pkt_access_subprog3 test_pkt_access
>> 36 new_get_skb_len 1 get_skb_len test_pkt_access
>> 37 new_get_skb_ifi 1 get_skb_ifindex test_pkt_access
>> 38 new_get_constan 1 get_constant test_pkt_access
>>
>> The BPF program dump_bpf_prog() in iterators.bpf.c is printing this data about
>> all BPF programs currently loaded in the system. This information is unstable
>> and will change from kernel to kernel.
>>
>> Signed-off-by: Alexei Starovoitov <ast@...nel.org>
>> ---
>
> [...]
>
>> +static int bpf_link_pin_kernel(struct dentry *parent,
>> + const char *name, struct bpf_link *link)
>> +{
>> + umode_t mode = S_IFREG | S_IRUSR | S_IWUSR;
>> + struct dentry *dentry;
>> + int ret;
>> +
>> + inode_lock(parent->d_inode);
>> + dentry = lookup_one_len(name, parent, strlen(name));
>> + if (IS_ERR(dentry)) {
>> + inode_unlock(parent->d_inode);
>> + return PTR_ERR(dentry);
>> + }
>> + ret = bpf_mkobj_ops(dentry, mode, link, &bpf_link_iops,
>> + &bpf_iter_fops);
>
> bpf_iter_fops only applies to bpf_iter links, while
> bpf_link_pin_kernel allows any link type. See bpf_mklink(), it checks
> bpf_link_is_iter() to decide between bpf_iter_fops and bpffs_obj_fops.
>
>
>> + dput(dentry);
>> + inode_unlock(parent->d_inode);
>> + return ret;
>> +}
>> +
>> static int bpf_obj_do_pin(const char __user *pathname, void *raw,
>> enum bpf_type type)
>> {
>> @@ -638,6 +659,57 @@ static int bpf_parse_param(struct fs_context *fc, struct fs_parameter *param)
>> return 0;
>> }
>>
>> +struct bpf_preload_ops bpf_preload_ops = { .info.driver_name = "bpf_preload" };
>> +EXPORT_SYMBOL_GPL(bpf_preload_ops);
>> +
>> +static int populate_bpffs(struct dentry *parent)
>
> So all the pinning has to happen from the kernel side because at the
> time that bpf_fill_super is called, user-space can't yet see the
> mounted BPFFS, do I understand the problem correctly? Would it be
> possible to add callback to fs_context_operations that would be called
> after FS is mounted and visible to user-space? At that point the
> kernel can spawn the user-mode blob and just instruct it to do both
> BPF object loading and pinning?
This is possible during bpf_fill_super() which is called when a `mount`
syscall is called. I experimented it a little bit when in my early
bpf_iter experiment with bpffs to re-populate every existing
iterators in a new bpffs mount.
In this case, we probably do not want to repopulate it in
every new bpffs mount. I think we just want to put them in a fixed
location. Since this is a fixed location, the system can go ahead
to do the mount, I think. But could just set up all necessary
data structures and do eventual mount after file system is up
in user space. Just my 2 cents.
>
> Or are there some other complications with such approach?
>
>> +{
>> + struct bpf_link *links[BPF_PRELOAD_LINKS] = {};
>> + u32 link_id[BPF_PRELOAD_LINKS] = {};
>> + int err = 0, i;
>> +
>> + mutex_lock(&bpf_preload_ops.lock);
>> + if (!bpf_preload_ops.do_preload) {
>> + mutex_unlock(&bpf_preload_ops.lock);
>> + request_module("bpf_preload");
>> + mutex_lock(&bpf_preload_ops.lock);
>> +
>> + if (!bpf_preload_ops.do_preload) {
>> + pr_err("bpf_preload module is missing.\n"
>> + "bpffs will not have iterators.\n");
>> + goto out;
>> + }
>> + }
>> +
>> + if (!bpf_preload_ops.info.tgid) {
>> + err = bpf_preload_ops.do_preload(link_id);
>> + if (err)
>> + goto out;
>> + for (i = 0; i < BPF_PRELOAD_LINKS; i++) {
>> + links[i] = bpf_link_by_id(link_id[i]);
>> + if (IS_ERR(links[i])) {
>> + err = PTR_ERR(links[i]);
>> + goto out;
>> + }
>> + }
>> + err = bpf_link_pin_kernel(parent, "maps", links[0]);
>> + if (err)
>> + goto out;
>> + err = bpf_link_pin_kernel(parent, "progs", links[1]);
>> + if (err)
>> + goto out;
>
> This hard coded "maps" -> link #0, "progs" -> link #1 mapping is what
> motivated the question above about letting user-space do all pinning.
> It would significantly simplify the kernel part, right?
>
>> + err = bpf_preload_ops.do_finish();
>> + if (err)
>> + goto out;
>> + }
>
> [...]
>
Powered by blists - more mailing lists