Re: [PATCH bpf-next 3/3] bpf: Add kernel module with user mode driver that populates bpffs.

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 





On 7/8/20 8:15 PM, Andrii Nakryiko wrote:
On Thu, Jul 2, 2020 at 1:04 PM Alexei Starovoitov
<alexei.starovoitov@xxxxxxxxx> wrote:

From: Alexei Starovoitov <ast@xxxxxxxxxx>

Add kernel module with user mode driver that populates bpffs with
BPF iterators.

$ mount bpffs /sys/fs/bpf/ -t bpf
$ ls -la /sys/fs/bpf/
total 4
drwxrwxrwt  2 root root    0 Jul  2 00:27 .
drwxr-xr-x 19 root root 4096 Jul  2 00:09 ..
-rw-------  1 root root    0 Jul  2 00:27 maps
-rw-------  1 root root    0 Jul  2 00:27 progs

The user mode driver will load BPF Type Formats, create BPF maps, populate BPF
maps, load two BPF programs, attach them to BPF iterators, and finally send two
bpf_link IDs back to the kernel.
The kernel will pin two bpf_links into newly mounted bpffs instance under
names "progs" and "maps". These two files become human readable.

$ cat /sys/fs/bpf/progs
   id name            pages attached
   11    dump_bpf_map     1 bpf_iter_bpf_map
   12   dump_bpf_prog     1 bpf_iter_bpf_prog
   27 test_pkt_access     1
   32       test_main     1 test_pkt_access test_pkt_access
   33   test_subprog1     1 test_pkt_access_subprog1 test_pkt_access
   34   test_subprog2     1 test_pkt_access_subprog2 test_pkt_access
   35   test_subprog3     1 test_pkt_access_subprog3 test_pkt_access
   36 new_get_skb_len     1 get_skb_len test_pkt_access
   37 new_get_skb_ifi     1 get_skb_ifindex test_pkt_access
   38 new_get_constan     1 get_constant test_pkt_access

The BPF program dump_bpf_prog() in iterators.bpf.c is printing this data about
all BPF programs currently loaded in the system. This information is unstable
and will change from kernel to kernel.

Signed-off-by: Alexei Starovoitov <ast@xxxxxxxxxx>
---

[...]

+static int bpf_link_pin_kernel(struct dentry *parent,
+                              const char *name, struct bpf_link *link)
+{
+       umode_t mode = S_IFREG | S_IRUSR | S_IWUSR;
+       struct dentry *dentry;
+       int ret;
+
+       inode_lock(parent->d_inode);
+       dentry = lookup_one_len(name, parent, strlen(name));
+       if (IS_ERR(dentry)) {
+               inode_unlock(parent->d_inode);
+               return PTR_ERR(dentry);
+       }
+       ret = bpf_mkobj_ops(dentry, mode, link, &bpf_link_iops,
+                           &bpf_iter_fops);

bpf_iter_fops only applies to bpf_iter links, while
bpf_link_pin_kernel allows any link type. See bpf_mklink(), it checks
bpf_link_is_iter() to decide between bpf_iter_fops and bpffs_obj_fops.


+       dput(dentry);
+       inode_unlock(parent->d_inode);
+       return ret;
+}
+
  static int bpf_obj_do_pin(const char __user *pathname, void *raw,
                           enum bpf_type type)
  {
@@ -638,6 +659,57 @@ static int bpf_parse_param(struct fs_context *fc, struct fs_parameter *param)
         return 0;
  }

+struct bpf_preload_ops bpf_preload_ops = { .info.driver_name = "bpf_preload" };
+EXPORT_SYMBOL_GPL(bpf_preload_ops);
+
+static int populate_bpffs(struct dentry *parent)

So all the pinning has to happen from the kernel side because at the
time that bpf_fill_super is called, user-space can't yet see the
mounted BPFFS, do I understand the problem correctly? Would it be
possible to add callback to fs_context_operations that would be called
after FS is mounted and visible to user-space? At that point the
kernel can spawn the user-mode blob and just instruct it to do both
BPF object loading and pinning?

This is possible during bpf_fill_super() which is called when a `mount`
syscall is called. I experimented it a little bit when in my early
bpf_iter experiment with bpffs to re-populate every existing
iterators in a new bpffs mount.

In this case, we probably do not want to repopulate it in
every new bpffs mount. I think we just want to put them in a fixed
location. Since this is a fixed location, the system can go ahead
to do the mount, I think. But could just set up all necessary
data structures and do eventual mount after file system is up
in user space. Just my 2 cents.


Or are there some other complications with such approach?

+{
+       struct bpf_link *links[BPF_PRELOAD_LINKS] = {};
+       u32 link_id[BPF_PRELOAD_LINKS] = {};
+       int err = 0, i;
+
+       mutex_lock(&bpf_preload_ops.lock);
+       if (!bpf_preload_ops.do_preload) {
+               mutex_unlock(&bpf_preload_ops.lock);
+               request_module("bpf_preload");
+               mutex_lock(&bpf_preload_ops.lock);
+
+               if (!bpf_preload_ops.do_preload) {
+                       pr_err("bpf_preload module is missing.\n"
+                              "bpffs will not have iterators.\n");
+                       goto out;
+               }
+       }
+
+       if (!bpf_preload_ops.info.tgid) {
+               err = bpf_preload_ops.do_preload(link_id);
+               if (err)
+                       goto out;
+               for (i = 0; i < BPF_PRELOAD_LINKS; i++) {
+                       links[i] = bpf_link_by_id(link_id[i]);
+                       if (IS_ERR(links[i])) {
+                               err = PTR_ERR(links[i]);
+                               goto out;
+                       }
+               }
+               err = bpf_link_pin_kernel(parent, "maps", links[0]);
+               if (err)
+                       goto out;
+               err = bpf_link_pin_kernel(parent, "progs", links[1]);
+               if (err)
+                       goto out;

This hard coded "maps" -> link #0, "progs" -> link #1 mapping is what
motivated the question above about letting user-space do all pinning.
It would significantly simplify the kernel part, right?

+               err = bpf_preload_ops.do_finish();
+               if (err)
+                       goto out;
+       }

[...]




[Index of Archives]     [Linux Samsung SoC]     [Linux Rockchip SoC]     [Linux Actions SoC]     [Linux for Synopsys ARC Processors]     [Linux NFS]     [Linux NILFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]


  Powered by Linux