v1->v2: - addressed feedback from Andrii and Eric. Thanks a lot for review! - added missing check at raw_tp attach time. - Andrii noticed that expected_attach_type cannot be reused. Had to introduce new field to bpf_attr. - cleaned up logging nicely by introducing bpf_log() helper. - rebased. Revolutionize bpf tracing and bpf C programming. C language allows any pointer to be typecasted to any other pointer or convert integer to a pointer. Though bpf verifier is operating at assembly level it has strict type checking for fixed number of types. Known types are defined in 'enum bpf_reg_type'. For example: PTR_TO_FLOW_KEYS is a pointer to 'struct bpf_flow_keys' PTR_TO_SOCKET is a pointer to 'struct bpf_sock', and so on. When it comes to bpf tracing there are no types to track. bpf+kprobe receives 'struct pt_regs' as input. bpf+raw_tracepoint receives raw kernel arguments as an array of u64 values. It was up to bpf program to interpret these integers. Typical tracing program looks like: int bpf_prog(struct pt_regs *ctx) { struct net_device *dev; struct sk_buff *skb; int ifindex; skb = (struct sk_buff *) ctx->di; bpf_probe_read(&dev, sizeof(dev), &skb->dev); bpf_probe_read(&ifindex, sizeof(ifindex), &dev->ifindex); } Addressing mistakes will not be caught by C compiler or by the verifier. The program above could have typecasted ctx->si to skb and page faulted on every bpf_probe_read(). bpf_probe_read() allows reading any address and suppresses page faults. Typical program has hundreds of bpf_probe_read() calls to walk kernel data structures. Not only tracing program would be slow, but there was always a risk that bpf_probe_read() would read mmio region of memory and cause unpredictable hw behavior. With introduction of Compile Once Run Everywhere technology in libbpf and in LLVM and BPF Type Format (BTF) the verifier is finally ready for the next step in program verification. Now it can use in-kernel BTF to type check bpf assembly code. Equivalent program will look like: struct trace_kfree_skb { struct sk_buff *skb; void *location; }; SEC("raw_tracepoint/kfree_skb") int trace_kfree_skb(struct trace_kfree_skb* ctx) { struct sk_buff *skb = ctx->skb; struct net_device *dev; int ifindex; __builtin_preserve_access_index(({ dev = skb->dev; ifindex = dev->ifindex; })); } These patches teach bpf verifier to recognize kfree_skb's first argument as 'struct sk_buff *' because this is what kernel C code is doing. The bpf program cannot 'cheat' and say that the first argument to kfree_skb raw_tracepoint is some other type. The verifier will catch such type mismatch between bpf program assumption of kernel code and the actual type in the kernel. Furthermore skb->dev access is type tracked as well. The verifier can see which field of skb is being read in bpf assembly. It will match offset to type. If bpf program has code: struct net_device *dev = (void *)skb->len; C compiler will not complain and generate bpf assembly code, but the verifier will recognize that integer 'len' field is being accessed at offsetof(struct sk_buff, len) and will reject further dereference of 'dev' variable because it contains integer value instead of a pointer. Such sophisticated type tracking allows calling networking bpf helpers from tracing programs. This patchset allows calling bpf_skb_event_output() that dumps skb data into perf ring buffer. It greatly improves observability. Now users can not only see packet lenth of the skb about to be freed in kfree_skb() kernel function, but can dump it to user space via perf ring buffer using bpf helper that was previously available only to TC and socket filters. See patch 10 for full example. The end result is safer and faster bpf tracing. Safer - because direct calls to bpf_probe_read() are disallowed and arbitrary addresses cannot be read. Faster - because normal loads are used to walk kernel data structures instead of bpf_probe_read() calls. Note that such loads can page fault and are supported by hidden bpf_probe_read() in interpreter and via exception table if program is JITed. See patches for details. Alexei Starovoitov (12): bpf: add typecast to raw_tracepoints to help BTF generation bpf: add typecast to bpf helpers to help BTF generation bpf: process in-kernel BTF bpf: add attach_btf_id attribute to program load libbpf: auto-detect btf_id of raw_tracepoint bpf: implement accurate raw_tp context access via BTF bpf: attach raw_tp program with BTF via type name bpf: add support for BTF pointers to interpreter bpf: add support for BTF pointers to x86 JIT bpf: check types of arguments passed into helpers bpf: disallow bpf_probe_read[_str] helpers selftests/bpf: add kfree_skb raw_tp test arch/x86/net/bpf_jit_comp.c | 97 +++++- include/linux/bpf.h | 39 ++- include/linux/bpf_verifier.h | 8 +- include/linux/btf.h | 1 + include/linux/extable.h | 10 + include/linux/filter.h | 6 +- include/trace/bpf_probe.h | 3 +- include/uapi/linux/bpf.h | 28 +- kernel/bpf/btf.c | 325 +++++++++++++++++- kernel/bpf/core.c | 39 ++- kernel/bpf/syscall.c | 85 +++-- kernel/bpf/verifier.c | 159 ++++++++- kernel/extable.c | 2 + kernel/trace/bpf_trace.c | 10 +- net/core/filter.c | 15 +- tools/include/uapi/linux/bpf.h | 28 +- tools/lib/bpf/bpf.c | 3 + tools/lib/bpf/libbpf.c | 17 + .../selftests/bpf/prog_tests/kfree_skb.c | 90 +++++ tools/testing/selftests/bpf/progs/kfree_skb.c | 74 ++++ 20 files changed, 975 insertions(+), 64 deletions(-) create mode 100644 tools/testing/selftests/bpf/prog_tests/kfree_skb.c create mode 100644 tools/testing/selftests/bpf/progs/kfree_skb.c -- 2.23.0