Re: perf BPF annotation output validation

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 




On 4/26/19 12:26 PM, Arnaldo Carvalho de Melo wrote:
> Em Fri, Apr 26, 2019 at 07:11:37PM +0000, Song Liu escreveu:
>>> On Apr 26, 2019, at 11:37 AM, Arnaldo Carvalho de Melo <arnaldo.melo@xxxxxxxxx> wrote:
>>>
>>> tldr; more to the end there are some annotation for all sorts of cache level
>>> misses trying to validate the 'perf annotation' output for disassembled
>>> BPF programs.
>>>
>>> [root@quaco ~]# uname -a
>>> Linux quaco 5.1.0-rc5+ #4 SMP Wed Apr 17 19:33:00 -03 2019 x86_64 x86_64 x86_64 GNU/Linux
>>>
>>> Running perf trace with:
>>>
>>> [root@quaco ~]# cat ~/.perfconfig
>>> [llvm]
>>> 	dump-obj = true
>>> 	clang-opt = -g
>>> [trace]
>>> 	add_events = /home/acme/git/perf/tools/perf/examples/bpf/augmented_raw_syscalls.c
>>> 	#add_events = /wb/augmented_raw_syscalls.o
>>> 	show_zeros = yes
>>> 	show_duration = no
>>> 	no_inherit = yes
>>> 	show_timestamp = no
>>> 	show_arg_names = no
>>> 	args_alignment = 40
>>> 	show_prefix = yes
>>>
>>> [annotate]
>>>
>>> 	hide_src_code = false
>>> [root@quaco ~]#
>>>
>>> clang/llvm 9:
>>>
>>> [root@quaco perf]# eu-readelf -winfo /home/acme/git/perf/tools/perf/examples/bpf/augmented_raw_syscalls.o | grep compile_unit
>>> [     b]  compile_unit         abbrev: 1
>>> [root@quaco perf]# eu-readelf -winfo /home/acme/git/perf/tools/perf/examples/bpf/augmented_raw_syscalls.o | grep compile_unit -B1 -A10
>>> Version: 4, Abbreviation section offset: 0, Address size: 8, Offset size: 4
>>> [     b]  compile_unit         abbrev: 1
>>>            producer             (strp) "clang version 9.0.0 (https://urldefense.proofpoint.com/v2/url?u=https-3A__git.llvm.org_git_clang.git_&d=DwIDaQ&c=5VD0RTtNlTh3ycd41b3MUw&r=dR8692q0_uaizy0jkrBJQM5k2hfm4CiFxYT8KaysFrg&m=8ymLhVNE6vfRn0KCyz7xlPkm7i1H4ttMrMYmeHEP1Cg&s=3_k_sPaZo7mwxQMoYhgd4hBUmCoS3TboKrw15yws6mA&e= 7906282d3afec5dfdc2b27943fd6c0309086c507) (https://urldefense.proofpoint.com/v2/url?u=https-3A__git.llvm.org_git_llvm.git_&d=DwIDaQ&c=5VD0RTtNlTh3ycd41b3MUw&r=dR8692q0_uaizy0jkrBJQM5k2hfm4CiFxYT8KaysFrg&m=8ymLhVNE6vfRn0KCyz7xlPkm7i1H4ttMrMYmeHEP1Cg&s=jutQ4xaAa5FePXthz5gHAgdx6kVlrPI64pnpqWWt59Q&e= a1b5de1ff8ae8bc79dc8e86e1f82565229bd0500)"
>>>            language             (data2) C99 (12)
>>>            name                 (strp) "/home/acme/git/perf/tools/perf/examples/bpf/augmented_raw_syscalls.c"
>>>            stmt_list            (sec_offset) 0
>>>            comp_dir             (strp) "/root"
>>>            low_pc               (addr) raw_syscalls:sys_enter+000000000000000000 <sys_enter>
>>>            ranges               (sec_offset) range list [     0]
>>> [    2a]    variable             abbrev: 2
>>>              name                 (strp) "__augmented_syscalls__"
>>>              type                 (ref4) [    3f]
>>> [root@quaco perf]#
>>>
>>> I then leave  this running doing system wide syscall tracing using that
>>> augmented_raw_syscalls.c eBPF program hooked into the syscall enter and
>>> exit tracepoints, so that I have something to profile and further test
>>> the BPF support in perf top/annotate/report (slides for LSFMM anyone?
>>> :-)) then I get:
>>>
>>> # perf trace -a sleep 100h
>>
>> wow, I never tried to run sleep with "100h". ;-)
> 
> This is me dreaming of the long night sleeps I don't have for about 3.3
> years by now ;-)
>   
>>> <SNIP>
>>> tmux: server/11597 ioctl(8</dev/ptmx>, TIOCLINUX, 0x7ffe1b49abf8) = 0
>>> tmux: server/11597 readv(8, 0x7ffe1b49ac00, 1)             = 101
>>> tmux: server/11597 ioctl(8</dev/ptmx>, TIOCSPGRP, 0x7ffe1b49a9f4) = 0
>>> tmux: server/11597 openat(AT_FDCWD, "/proc/30339/cmdline", O_RDONLY) = 5
>>> tmux: server/11597 fstat(5, 0x7ffe1b49a8e0)                = 0
>>> tmux: server/11597 read(5, 0x5571aaeb2c10, 1024)           = 10
>>> tmux: server/11597 close(5)                                = 0
>>> gnome-shell/19195  ... [continued]: ioctl())              = 0
>>> tmux: server/11597 poll(0x5571aad38e50, 6, 11997)       ...
>>> gnome-shell/19195 ioctl(11</dev/dri/card0>, DRM_MODE_RMFB, 0x7ffc2129c25c) = 0
>>> gnome-shell/19195 recvmsg(39<socket:[18671390]>, 0x7ffc2129c440, NONE) = -1 EAGAIN (Resource temporarily unavailable)
>>> <SNIP>
>>>
>>> With that running, I do:
>>>
>>> [root@quaco perf]# perf record -a sleep 2s
>>> [ perf record: Woken up 1 times to write data ]
>>> [ perf record: Captured and wrote 1.647 MB perf.data (3698 samples) ]
>>> [root@quaco perf]#
>>>
>>> I.e. collect 2 seconds worth of CPU cycles samples, system wide.
>>>
>>> [root@quaco perf]# perf record -a sleep 2s
>>> [ perf record: Woken up 1 times to write data ]
>>> [ perf record: Captured and wrote 2.328 MB perf.data (17764 samples) ]
>>> [root@quaco perf]# perf report --header-only perf.data | grep -i b[pt]f
>>> # event : name = cycles, , id = { 266608, 266609, 266610, 266611, 266612, 266613, 266614, 266615 }, size = 112, { sample_period, sample_freq } = 4000, sample_type = IP|TID|TIME|CPU|PERIOD, read_format = ID, disabled = 1, inherit = 1, mmap = 1, comm = 1, freq = 1, task = 1, precise_ip = 3, sample_id_all = 1, exclude_guest = 1, mmap2 = 1, comm_exec = 1, ksymbol = 1, bpf_event = 1
>>> # bpf_prog_info 13: bpf_prog_7be49e3934a125ba addr 0xffffffffc0437f01 size 229
>>> # bpf_prog_info 14: bpf_prog_2a142ef67aaad174 addr 0xffffffffc04393cd size 229
>>> # bpf_prog_info 17: bpf_prog_7be49e3934a125ba addr 0xffffffffc052ca3c size 229
>>> # bpf_prog_info 18: bpf_prog_2a142ef67aaad174 addr 0xffffffffc0531218 size 229
>>> # bpf_prog_info 21: bpf_prog_7be49e3934a125ba addr 0xffffffffc0cb8010 size 229
>>> # bpf_prog_info 22: bpf_prog_2a142ef67aaad174 addr 0xffffffffc0cba8be size 229
>>> # bpf_prog_info 65: bpf_prog_7be49e3934a125ba addr 0xffffffffc03f6bc2 size 229
>>> # bpf_prog_info 66: bpf_prog_2a142ef67aaad174 addr 0xffffffffc03f87f6 size 229
>>> # bpf_prog_info 125: bpf_prog_819967866022f1e1_sys_enter addr 0xffffffffc0c6e5a1 size 381
>>> # bpf_prog_info 126: bpf_prog_c1bd85c092d6e4aa_sys_exit addr 0xffffffffc0c953a8 size 191
>>> [root@quaco perf]#
>>>
>>> So everything seems in order, but then:
>>>
>>> [root@quaco perf]# perf report | grep bpf_prog
>>>      0.14%  perf             bpf_prog_c1bd85c092d6e4aa_sys_exit   [k] bpf_prog_c1bd85c092d6e4aa_sys_exit
>>>      0.08%  gnome-terminal-  bpf_prog_819967866022f1e1_sys_enter  [k] bpf_prog_819967866022f1e1_sys_enter
>>>      0.06%  perf             bpf_prog_819967866022f1e1_sys_enter  [k] bpf_prog_819967866022f1e1_sys_enter
>>>      0.04%  gnome-terminal-  bpf_prog_c1bd85c092d6e4aa_sys_exit   [k] bpf_prog_c1bd85c092d6e4aa_sys_exit
>>>      0.01%  alsa-sink-ALC25  bpf_prog_819967866022f1e1_sys_enter  [k] bpf_prog_819967866022f1e1_sys_enter
>>>      0.01%  Web Content      bpf_prog_c1bd85c092d6e4aa_sys_exit   [k] bpf_prog_c1bd85c092d6e4aa_sys_exit
>>>      0.01%  lvmetad          bpf_prog_819967866022f1e1_sys_enter  [k] bpf_prog_819967866022f1e1_sys_enter
>>>      0.00%  alsa-sink-ALC25  bpf_prog_c1bd85c092d6e4aa_sys_exit   [k] bpf_prog_c1bd85c092d6e4aa_sys_exit
>>>      0.00%  sssd_kcm         bpf_prog_c1bd85c092d6e4aa_sys_exit   [k] bpf_prog_c1bd85c092d6e4aa_sys_exit
>>>      0.00%  gnome-shell      bpf_prog_819967866022f1e1_sys_enter  [k] bpf_prog_819967866022f1e1_sys_enter
>>>      0.00%  firefox          bpf_prog_c1bd85c092d6e4aa_sys_exit   [k] bpf_prog_c1bd85c092d6e4aa_sys_exit
>>>      0.00%  weechat          bpf_prog_c1bd85c092d6e4aa_sys_exit   [k] bpf_prog_c1bd85c092d6e4aa_sys_exit
>>>      0.00%  WebExtensions    bpf_prog_c1bd85c092d6e4aa_sys_exit   [k] bpf_prog_c1bd85c092d6e4aa_sys_exit
>>>      0.00%  Timer            bpf_prog_819967866022f1e1_sys_enter  [k] bpf_prog_819967866022f1e1_sys_enter
>>>      0.00%  Web Content      bpf_prog_819967866022f1e1_sys_enter  [k] bpf_prog_819967866022f1e1_sys_enter
>>> [root@quaco perf]#
>>>
>>> Ok, lets check that bpf_prog_819967866022f1e1_sys_enter symbol, 38% of
>>> the samples at just at that %rbp, more in a previous run...
>>>
>>> [root@quaco perf]# perf annotate --stdio2 bpf_prog_819967866022f1e1_sys_enter
>>> Samples: 25  of event 'cycles', 4000 Hz, Event count (approx.): 7248318, [percent: local period]
>>> bpf_prog_819967866022f1e1_sys_enter() bpf_prog_819967866022f1e1_sys_enter
>>> 38.00         push   %rbp
>>>
>>>   2.72         mov    %rsp,%rbp
>>>                sub    $0x170,%rsp
>>> 13.19         sub    $0x28,%rbp
>>>   3.30         mov    %rbx,0x0(%rbp)
>>> 20.33         mov    %r13,0x8(%rbp)
>>>                mov    %r14,0x10(%rbp)
>>>                mov    %r15,0x18(%rbp)
>>>                xor    %eax,%eax
>>>                mov    %rax,0x20(%rbp)
>>>                mov    %rdi,%rbx
>>> 13.34       → callq  *ffffffffd359487f
>>>                mov    %eax,-0x148(%rbp)
>>>
>>> So I ended up trying with a event group to check what could be that:
>>>
>>> [root@quaco perf]# perf record -a -e '{cycles,instructions,cache-misses}' sleep 2s
>>> [ perf record: Woken up 18 times to write data ]
>>> [ perf record: Captured and wrote 6.619 MB perf.data (91964 samples) ]
>>> [root@quaco perf]# perf annotate --stdio2 bpf_prog_819967866022f1e1_sys_enter
>>> Samples: 314  of events 'anon group { cycles, instructions, cache-misses }', 4000 Hz, Event count (approx.): 118106063, [percent: local period]
>>> bpf_prog_819967866022f1e1_sys_enter() bpf_prog_819967866022f1e1_sys_enter
>>> 20.88  24.58  49.11         push   %rbp
>>>
>>> 47.37  31.62  26.71         mov    %rsp,%rbp
>>>   1.16   0.00   0.00         sub    $0x170,%rsp
>>>                              sub    $0x28,%rbp
>>>   0.58   2.48   0.00         mov    %rbx,0x0(%rbp)
>>>   7.10   5.92   0.00         mov    %r13,0x8(%rbp)
>>>   0.51   0.96   7.96         mov    %r14,0x10(%rbp)
>>>   0.00   1.10   0.00         mov    %r15,0x18(%rbp)
>>>   0.58   0.00   0.00         xor    %eax,%eax
>>>                              mov    %rax,0x20(%rbp)
>>>   2.39   0.85   0.00         mov    %rdi,%rbx
>>>   0.58   1.96   0.00       → callq  *ffffffffd359487f
>>>                              mov    %eax,-0x148(%rbp)
>>>   2.88   0.89   3.50         mov    %rbp,%rsi
>>>                              add    $0xfffffffffffffeb8,%rsi
>>>
>>> So seems ok? I.e. that big cost at right the start of the BPF program is
>>> just cache pressure, probably, right?
>>>
>>>
>>> [root@quaco perf]# perf record -a -e '{L1-icache-load-misses,L1-dcache-load-misses}' sleep 2s
>>> [ perf record: Woken up 14 times to write data ]
>>> [ perf record: Captured and wrote 5.557 MB perf.data (72026 samples) ]
>>> [root@quaco perf]# perf annotate --stdio2 bpf_prog_819967866022f1e1_sys_enter
>>> Samples: 376  of events 'anon group { L1-icache-load-misses, L1-dcache-load-misses }', 4000 Hz, Event count (approx.): 3715276, [percent: local p>
>>> bpf_prog_819967866022f1e1_sys_enter() bpf_prog_819967866022f1e1_sys_enter
>>>   1.54  63.01         push   %rbp
>>>
>>> 53.09  28.13         mov    %rsp,%rbp
>>>   0.58   0.00         sub    $0x170,%rsp
>>>   0.00   0.55         sub    $0x28,%rbp
>>>   6.96   0.00         mov    %rbx,0x0(%rbp)
>>> 10.25   0.00         mov    %r13,0x8(%rbp)
>>>   0.73   0.00         mov    %r14,0x10(%rbp)
>>>   3.04   0.00         mov    %r15,0x18(%rbp)
>>>
>>> Looks like. furthermore:
>>>
>>> [root@quaco perf]# perf record -a -e '{L1-icache-load-misses,L1-dcache-load-misses,dTLB-load-misses,iTLB-load-misses}' sleep 2s
>>> [ perf record: Woken up 21 times to write data ]
>>> [ perf record: Captured and wrote 7.708 MB perf.data (112323 samples) ]
>>> [root@quaco perf]# perf annotate --stdio2 bpf_prog_819967866022f1e1_sys_enter
>>> Samples: 391  of events 'anon group { L1-icache-load-misses, L1-dcache-load-misses, dTLB-load-misses, iTLB-load-misses }', 4000 Hz, Event count (>
>>> bpf_prog_819967866022f1e1_sys_enter() bpf_prog_819967866022f1e1_sys_enter
>>>   1.98  64.42 100.00  13.66         push   %rbp
>>>
>>> 50.79  23.91   0.00  82.23         mov    %rsp,%rbp
>>>                                     sub    $0x170,%rsp
>>>   0.24   0.00   0.00   0.00         sub    $0x28,%rbp
>>>   7.73   0.00   0.00   0.00         mov    %rbx,0x0(%rbp)
>>> 11.13   0.44   0.00   0.00         mov    %r13,0x8(%rbp)
>>>                                     mov    %r14,0x10(%rbp)
>>>   1.64   0.46   0.00   0.00         mov    %r15,0x18(%rbp)
>>>
>>> Do these numbers look sane to you guys?
>>
>>  From my experiments, mov is often the top time consumer. But
>>
>>     mov %rsp,%rbp
>>
>> is a little weird. It does not access cache, right?
> 
> Right, but its like the code for bpf_prog_819967866022f1e1_sys_enter is
> being sucked into L1 icache at that point, no?

do you need to enable pebs (precise event-based sampling)?

>   
>>> Song, I also noticed that source code is not being intermixed for the
>>> --stdio annotation, while it works, to some degree, for '--tui', i.e.
>>> when you do 'perf top', press '/bpf' to show just symbols with that
>>> substring and then press enter or 'A' to annotate, we can see the
>>> original C source code for the BPF program, but it is mangling the
>>> screen sometimes, I need to try and fix, please take a look if you have
>>> the time.
>>
>> I will take a look at this, most likely after LSFMM.
> 
> mmkay
>   
>>> Also things like the callq targets need some work to tell what function
>>> is that, which as I said isn't appearing on the --stdio2 output, but
>>> appears on the --tui, i.e. we need to resolve that symbol to check how
>>> to map back to a BPF helper or any othe callq target.
>   
>> I am aware of missing callq symbol. I was thinking they are less critical
>> with source code.
> 
> Some people dislikes intermixing C and assembly, Linus is one, IIRC, so
> if we can augment the assembly resolving those addresses to the helper,
> that would be best.
>   
>>> Also, what about those 'je 0', i.e. the target is being misinterpreted
>>> or is this some BPF construct I should've know about? :)
>>
>> I didn't notice the je 0 before. Let me try reproduce it.

Did you see these `je 0` when you use bpftool dump jited instructions?
With dumping source annotated jited/xlated instructions, you can get 
rough idea where this `je 0` is coming from. Want to give a try?

> 
> Ok!
>   
>> Thanks,
>> Song
>>
>>>
>>>   2.68   0.00   0.00   0.00         mov    %rdi,%rbx
>>>                                   → callq  *ffffffffd359487f
>>>                                     mov    %eax,-0x148(%rbp)
>>>   9.61   0.00   0.00   0.00         mov    %rbp,%rsi
>>>                                     add    $0xfffffffffffffeb8,%rsi
>>>                                     movabs $0xffff9d556c776c00,%rdi
>>>
>>>                                   → callq  *ffffffffd3595b2f
>>>                                     cmp    $0x0,%rax
>>>                                   → je     0
>>>   0.00   1.25   0.00   0.00         add    $0x38,%rax
>>>   0.80   0.21   0.00   0.00         xor    %r13d,%r13d
>>>                                     cmp    $0x0,%rax
>>>                                   → jne    0
>>>                                     mov    %rbp,%rdi
>>>                                     add    $0xfffffffffffffeb8,%rdi
>>>
>>> - Arnaldo
>>
> 




[Index of Archives]     [Linux Samsung SoC]     [Linux Rockchip SoC]     [Linux Actions SoC]     [Linux for Synopsys ARC Processors]     [Linux NFS]     [Linux NILFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]


  Powered by Linux