Hi, Here is a series to add nested-kprobes support to x86, arm64 and arm (which I can test). These make kprobes to accept 1-level nesting instead of incrementing missed count. Any kprobes hits in kprobes pre/post handler context can be nested at once. If the other kprobes hits in the nested pre/post handler context or in the single-stepping context, that will be still missed. The nest level is actually easily extended, but too many nest level can lead the overflow of the kernel stack (for each nest, the stack will be consumed by saving registers, handling kprobes and pre/post handlers.) Thus, at this moment it allows only one level nest. This feature allows BPF or ftrace user to put a kprobe on BPF jited code or ftrace internal code running in the kprobe context for debugging. We can test this feature on the kernel with CONFIG_KPROBE_EVENTS_ON_NOTRACE=y as below. # cd /sys/kernel/debug/tracing # echo p ring_buffer_lock_reserve > kprobe_events # echo p vfs_read >> kprobe_events # echo stacktrace > events/kprobes/p_ring_buffer_lock_reserve_0/trigger # echo 1 > events/kprobes/enable # cat trace ... cat-151 [000] ...1 48.669190: p_vfs_read_0: (vfs_read+0x0/0x160) cat-151 [000] ...2 48.669276: p_ring_buffer_lock_reserve_0: (ring_buffer_lock_reserve+0x0/0x400) cat-151 [000] ...2 48.669288: <stack trace> => kprobe_dispatcher => opt_pre_handler => optimized_callback => 0xffffffffa0002331 => ring_buffer_lock_reserve => kprobe_trace_func => kprobe_dispatcher => opt_pre_handler => optimized_callback => 0xffffffffa00023b0 => vfs_read => load_elf_phdrs => load_elf_binary => search_binary_handler.part.0 => __do_execve_file.isra.0 => __x64_sys_execve => do_syscall_64 => entry_SYSCALL_64_after_hwframe To check unoptimized code, disable optprobe and dump the log. # echo 0 > /proc/sys/debug/kprobes-optimization # echo > trace # cat trace cat-153 [000] d..1 140.581433: p_vfs_read_0: (vfs_read+0x0/0x160) cat-153 [000] d..2 140.581780: p_ring_buffer_lock_reserve_0: (ring_buffer_lock_reserve+0x0/0x400) cat-153 [000] d..2 140.581811: <stack trace> => kprobe_dispatcher => aggr_pre_handler => kprobe_int3_handler => do_int3 => int3 => ring_buffer_lock_reserve => kprobe_trace_func => kprobe_dispatcher => aggr_pre_handler => kprobe_int3_handler => do_int3 => int3 => vfs_read => load_elf_phdrs => load_elf_binary => search_binary_handler.part.0 => __do_execve_file.isra.0 => __x64_sys_execve => do_syscall_64 => entry_SYSCALL_64_after_hwframe So we can see the kprobe can be nested. Thank you, --- Masami Hiramatsu (3): x86/kprobes: Support nested kprobes arm64: kprobes: Support nested kprobes arm: kprobes: Support nested kprobes arch/arm/include/asm/kprobes.h | 5 ++ arch/arm/probes/kprobes/core.c | 79 ++++++++++++++---------------- arch/arm/probes/kprobes/core.h | 30 +++++++++++ arch/arm/probes/kprobes/opt-arm.c | 6 ++ arch/arm64/include/asm/kprobes.h | 5 ++ arch/arm64/kernel/probes/kprobes.c | 75 ++++++++++++++++------------ arch/x86/include/asm/kprobes.h | 5 ++ arch/x86/kernel/kprobes/common.h | 39 ++++++++++++++- arch/x86/kernel/kprobes/core.c | 96 +++++++++++++++--------------------- arch/x86/kernel/kprobes/ftrace.c | 6 ++ arch/x86/kernel/kprobes/opt.c | 13 +++-- 11 files changed, 214 insertions(+), 145 deletions(-) -- Masami Hiramatsu (Linaro) <mhiramat@xxxxxxxxxx>