Re: [PATCH] um: vector: fix BPF loading in vector drivers

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 29/11/2019 23:12, Daniel Borkmann wrote:
On 11/29/19 12:54 PM, Anton Ivanov wrote:
On 29/11/2019 09:15, Daniel Borkmann wrote:
On 11/28/19 6:44 PM, anton.ivanov@xxxxxxxxxxxxxxxxxx wrote:
From: Anton Ivanov <anton.ivanov@xxxxxxxxxxxxxxxxxx>

This fixes a possible hang in bpf firmware loading in the
UML vector io drivers due to use of GFP_KERNEL while holding
a spinlock.

Based on a prposed fix by weiyongjun1@xxxxxxxxxx and suggestions for
improving it by dan.carpenter@xxxxxxxxxx

Signed-off-by: Anton Ivanov <anton.ivanov@xxxxxxxxxxxxxxxxxx>

Any reason why this BPF firmware loading mechanism in UML vector driver that was recently added [0] is plain old classic BPF? Quoting your commit log [0]:

It will allow whatever is allowed by sockfilter. Looking at the sockfilter implementation in the kernel it takes eBPF, however even the kernel docs still state BPF.

You are using SO_ATTACH_FILTER in uml_vector_attach_bpf() which is the old classic BPF (and not eBPF). The kernel internally moves that over to eBPF insns, but you'll be constrained forever with the abilities of cBPF. The later added SO_ATTACH_BPF is
the one for eBPF where you pass the prog fd from bpf().

I will switch to that in the next version.


   All vector drivers now allow a BPF program to be loaded and
   associated with the RX socket in the host kernel.

   1. The program can be loaded as an extra kernel command line
   option to any of the vector drivers.

   2. The program can also be loaded as "firmware", using the
   ethtool flash option. It is possible to turn this facility
   on or off using a command line option.

   A simplistic wrapper for generating the BPF firmware for the raw
   socket driver out of a tcpdump/libpcap filter expression can be
   found at: https://github.com/kot-begemot-uk/uml_vector_utilities/

... it tells what it does but /nothing/ about the original rationale / use case why it is needed. So what is the use case? And why is this only classic BPF? Is there any discussion to read up that lead you to this decision of only implementing
handling for classic BPF?

Moving processing out of the GUEST onto the HOST using a safe language. The firmware load is on the GUEST and your BPF is your virtual NIC "firmware" which runs on the HOST (in the host kernel in fact).

It is identical as an idea to what Netronome cards do in hardware.

I'm asking because classic BPF is /legacy/ stuff that is on feature freeze and only very limited in terms of functionality compared to native (e)BPF which is why you need this weird 'firmware' loader [1] which wraps around tcpdump to
parse the -ddd output into BPF insns ...

Because there is no other mechanism of retrieving it after it is compiled by libpcap in any of the common scripting languages.

The pcap Perl, Python, Go (or whatever else) wrappers do not give you access to the compiled code after the filter has been compiled.

Why is that ingenious design - you have to take it with their maintainers.

So if you want to start with pcap/tcpdump syntax and you do not want to rewrite that part of tcpdump as a dumper in C you have no other choice.

The starting point is chosen because the idea is at some point to replace the existing and very aged pcap network transport in UML. That takes pcap syntax on the kernel command line.

I admit it is a kludge, I will probably do the "do not want" bit and rewrite that in C.

Yeah, it would probably be about the same # of LOC in C.

In any case - the "loader" is only an example, you can compile BPF using LLVM or whatever else you like.

But did you try that with the code you have? Seems not, which is perhaps why there are some
wrong assumptions.

All of my tests were done using bpf generated by tcpdump out of a pcap expression. So the answer is no - I did not try LLVM because I did not need to for what I was aiming to achieve.

The pcap route matches 1:1 existing functionality in the uml pcap driver as well as existing functionality in the vector drivers for the cases where they need to avoid seeing their own xmits and cannot use features like QDISC_BYPASS.


You can't use LLVM's BPF backend here since you only allow to pass in cBPF, and LLVM emits an object file with native eBPF insns (you could use libbpf (in-tree under tools/lib/bpf/)
for loading that).

My initial aim was the same feature sets as pcap and achieve it using a virtual analogue of what cards like Netronome do - via the firmware route.

Switching to SO_ATTACH_BPF will come in the next revision.

A.


A.

_______________________________________________
linux-um mailing list
linux-um@xxxxxxxxxxxxxxxxxxx
http://lists.infradead.org/mailman/listinfo/linux-um


--
Anton R. Ivanov
Cambridgegreys Limited. Registered in England. Company Number 10273661
https://www.cambridgegreys.com/



[Index of Archives]     [Linux Samsung SoC]     [Linux Rockchip SoC]     [Linux Actions SoC]     [Linux for Synopsys ARC Processors]     [Linux NFS]     [Linux NILFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]


  Powered by Linux