On 08/28, Jiri Olsa wrote: > > On Tue, Aug 27, 2024 at 10:19:26PM +0200, Oleg Nesterov wrote: > > Hmm. Really? In this case these 2 different consumers will have the different > > trace_event_call's, so > > > > // consumer for task 1019 > > uretprobe_dispatcher > > uretprobe_perf_func > > __uprobe_perf_func > > perf_tp_event > > > > won't store the event because this_cpu_ptr(call->perf_events) should be > > hlist_empty() on this CPU, the perf_event for task 1019 wasn't scheduled in > > on this CPU... > > I'll double check on that, Yes, please. > but because there's no filter for uretprobe > I think it will be stored under 1018 event ... > I'm working on bpf selftests for above (uprobe and uprobe_multi paths) Meanwhile, I decided to try to test this case too ;) test.c: #include <unistd.h> int func(int i) { return i; } int main(void) { int i; for (i = 0;; ++i) { sleep(1); func(i); } return 0; } run_probe.c: #include "./include/uapi/linux/perf_event.h" #define _GNU_SOURCE #include <sys/syscall.h> #include <sys/ioctl.h> #include <assert.h> #include <unistd.h> #include <stdlib.h> #include <stdio.h> // cat /sys/bus/event_source/devices/uprobe/type #define UPROBE_TYPE 9 void run_probe(const char *file, unsigned offset, int pid) { struct perf_event_attr attr = { .type = UPROBE_TYPE, .config = 1, // ret-probe .uprobe_path = (unsigned long)file, .probe_offset = offset, .size = sizeof(struct perf_event_attr), }; int fd = syscall(__NR_perf_event_open, &attr, pid, 0, -1, 0); assert(fd >= 0); assert(ioctl(fd, PERF_EVENT_IOC_ENABLE, 0) == 0); for (;;) pause(); } int main(int argc, const char *argv[]) { int pid = atoi(argv[1]); run_probe("./test", 0x536, pid); return 0; } Now, with the kernel patch below applied, I do $ ./test & $ PID1=$! $ ./test & $ PID2=$! $ ./run_probe $PID1 & $ ./run_probe $PID2 & and the kernel prints: CHAIN trace_uprobe: HANDLER pid=46 consumers_target=46 stored=1 trace_uprobe: HANDLER pid=46 consumers_target=45 stored=0 CHAIN trace_uprobe: HANDLER pid=45 consumers_target=46 stored=0 trace_uprobe: HANDLER pid=45 consumers_target=45 stored=1 CHAIN trace_uprobe: HANDLER pid=46 consumers_target=46 stored=1 trace_uprobe: HANDLER pid=46 consumers_target=45 stored=0 CHAIN trace_uprobe: HANDLER pid=45 consumers_target=46 stored=0 trace_uprobe: HANDLER pid=45 consumers_target=45 stored=1 CHAIN trace_uprobe: HANDLER pid=46 consumers_target=46 stored=1 trace_uprobe: HANDLER pid=46 consumers_target=45 stored=0 CHAIN trace_uprobe: HANDLER pid=45 consumers_target=46 stored=0 trace_uprobe: HANDLER pid=45 consumers_target=45 stored=1 CHAIN trace_uprobe: HANDLER pid=46 consumers_target=46 stored=1 trace_uprobe: HANDLER pid=46 consumers_target=45 stored=0 CHAIN trace_uprobe: HANDLER pid=45 consumers_target=46 stored=0 trace_uprobe: HANDLER pid=45 consumers_target=45 stored=1 and so on. As you can see, perf_trace_buf_submit/etc is never called for the "unfiltered" consumer, so I still think that perf is fine wrt filtering. But I can be easily wrong, please check. Oleg. diff --git a/kernel/events/uprobes.c b/kernel/events/uprobes.c index acc73c1bc54c..14aa92a78d6d 100644 --- a/kernel/events/uprobes.c +++ b/kernel/events/uprobes.c @@ -2150,6 +2150,8 @@ handle_uretprobe_chain(struct return_instance *ri, struct pt_regs *regs) struct uprobe *uprobe = ri->uprobe; struct uprobe_consumer *uc; + pr_crit("CHAIN\n"); + rcu_read_lock_trace(); list_for_each_entry_rcu(uc, &uprobe->consumers, cons_node, rcu_read_lock_trace_held()) { if (uc->ret_handler) diff --git a/kernel/trace/trace_uprobe.c b/kernel/trace/trace_uprobe.c index f7443e996b1b..e4eaa0363742 100644 --- a/kernel/trace/trace_uprobe.c +++ b/kernel/trace/trace_uprobe.c @@ -1364,7 +1364,7 @@ static bool uprobe_perf_filter(struct uprobe_consumer *uc, struct mm_struct *mm) return ret; } -static void __uprobe_perf_func(struct trace_uprobe *tu, +static int __uprobe_perf_func(struct trace_uprobe *tu, unsigned long func, struct pt_regs *regs, struct uprobe_cpu_buffer **ucbp) { @@ -1375,6 +1375,7 @@ static void __uprobe_perf_func(struct trace_uprobe *tu, void *data; int size, esize; int rctx; + int ret = 0; #ifdef CONFIG_BPF_EVENTS if (bpf_prog_array_valid(call)) { @@ -1382,7 +1383,7 @@ static void __uprobe_perf_func(struct trace_uprobe *tu, ret = bpf_prog_run_array_uprobe(call->prog_array, regs, bpf_prog_run); if (!ret) - return; + return -1; } #endif /* CONFIG_BPF_EVENTS */ @@ -1392,12 +1393,13 @@ static void __uprobe_perf_func(struct trace_uprobe *tu, size = esize + ucb->dsize; size = ALIGN(size + sizeof(u32), sizeof(u64)) - sizeof(u32); if (WARN_ONCE(size > PERF_MAX_TRACE_SIZE, "profile buffer not large enough")) - return; + return -1; preempt_disable(); head = this_cpu_ptr(call->perf_events); if (hlist_empty(head)) goto out; + ret = 1; entry = perf_trace_buf_alloc(size, NULL, &rctx); if (!entry) @@ -1421,6 +1423,7 @@ static void __uprobe_perf_func(struct trace_uprobe *tu, head, NULL); out: preempt_enable(); + return ret; } /* uprobe profile handler */ @@ -1439,7 +1442,15 @@ static void uretprobe_perf_func(struct trace_uprobe *tu, unsigned long func, struct pt_regs *regs, struct uprobe_cpu_buffer **ucbp) { - __uprobe_perf_func(tu, func, regs, ucbp); + struct trace_uprobe_filter *filter = tu->tp.event->filter; + struct perf_event *event = list_first_entry(&filter->perf_events, + struct perf_event, hw.tp_list); + + int r = __uprobe_perf_func(tu, func, regs, ucbp); + + pr_crit("HANDLER pid=%d consumers_target=%d stored=%d\n", + current->pid, event->hw.target ? event->hw.target->pid : -1, r); + } int bpf_get_uprobe_info(const struct perf_event *event, u32 *fd_type,