Hi! Today, BPF x86-64 JIT is preserving all of the callee-saved registers for each BPF program being JITed, even when none of the R6-R9 registers are used by the BPF program. Furthermore the tail call counter is always pushed/popped to/from the stack even when there is no tail call usage in BPF program being JITed. Optimization can be introduced that would detect the usage of R6-R9 and based on that push/pop to/from the stack only what is needed. Same goes for tail call counter. Results look promising for such instruction reduction. Below are the numbers for xdp1 sample on FVL 40G NIC receiving traffic from pktgen: * With optimization: 22.3 Mpps * Without: 19.0 mpps So it's around 15% of performance improvement. Note that xdp1 is not using any of callee saved registers, nor the tail call, hence such speed-up. There is one detail that needs to be handled though. Currently, x86-64 JIT tail call implementation is skipping the prologue of target BPF program that has constant size. With the mentioned optimization implemented, each particular BPF program that might be inserted onto the prog array map and therefore be the target of tail call, could have various prologue size. Let's have some pseudo-code example: func1: pro code epi func2: pro code' epi func3: pro code'' epi Today, pro and epi are always the same (9/7) instructions. So a tail call from func1 to func2 is just a: jump func2 + sizeof pro in bytes (PROLOGUE_SIZE) With the optimization: func1: pro code epi func2: pro' code' epi' func3: pro'' code'' epi'' For making the tail calls up and running with the mentioned optimization in place, x86-64 JIT should emit the pop registers instructions that were pushed on prologue before the actual jump. Jump offset should skip the instructions that are handling rbp/rsp, not the whole prologue. A tail call within func1 would then need to be: epi -> pop what pro pushed, but no leave/ret instructions jump func2 + 16 // first push insn of pro'; if no push, then this would // a direct jump to code' Magic value of 16 comes from count of bytes that represent instructions that are skipped: 0f 1f 44 00 00 nopl 0x0(%rax,%rax,1) 55 push %rbp 48 89 e5 mov %rsp,%rbp 48 81 ec 08 00 00 00 sub $0x8,%rsp which would in many cases add *more* instructions for tailcalls. If none of callee-saved registers are used, then there would be no overhead with such optimization in place. I'm not sure how to measure properly the impact on the BPF programs that are utilizing tail calls. Any suggestions? Daniel, Alexei, what is your view on this? For implementation details, see commit message of included patch. Thank you, Maciej Maciej Fijalkowski (1): bpf, x64: optimize JIT prologue/epilogue generation arch/x86/net/bpf_jit_comp.c | 190 ++++++++++++++++++++++++++++-------- 1 file changed, 148 insertions(+), 42 deletions(-) -- 2.20.1