On Wed, Nov 16, 2022 at 10:43:57PM +0800, Leizhen (ThunderTown) wrote: > On 2022/11/15 0:06, Paul E. McKenney wrote: > > On Mon, Nov 14, 2022 at 03:18:10PM +0800, Leizhen (ThunderTown) wrote: > >> On 2022/11/12 14:08, Paul E. McKenney wrote: > >>> On Sat, Nov 12, 2022 at 10:32:32AM +0800, Leizhen (ThunderTown) wrote: > >>>> On 2022/11/12 2:42, Paul E. McKenney wrote: > >>>>> On Fri, Nov 11, 2022 at 01:05:56PM +0000, Zhang, Qiang1 wrote: > >>>>>> On 2022/11/11 19:54, Zhang, Qiang1 wrote: > >>>>>>>> When a structure containing an RCU callback rhp is (incorrectly) > >>>>>>>> freed and reallocated after rhp is passed to call_rcu(), it is not > >>>>>>>> unusual for > >>>>>>>> rhp->func to be set to NULL. This defeats the debugging prints used > >>>>>>>> rhp->by > >>>>>>>> __call_rcu_common() in kernels built with > >>>>>>>> CONFIG_DEBUG_OBJECTS_RCU_HEAD=y, which expect to identify the > >>>>>>>> offending code using the identity of this function. > >>>>>>>> > >>>>>>>> And in kernels build without CONFIG_DEBUG_OBJECTS_RCU_HEAD=y, things > >>>>>>>> are even worse, as can be seen from this splat: > >>>>>>>> > >>>>>>>> Unable to handle kernel NULL pointer dereference at virtual address 0 > >>>>>>>> ... ... > >>>>>>>> PC is at 0x0 > >>>>>>>> LR is at rcu_do_batch+0x1c0/0x3b8 > >>>>>>>> ... ... > >>>>>>>> (rcu_do_batch) from (rcu_core+0x1d4/0x284) > >>>>>>>> (rcu_core) from (__do_softirq+0x24c/0x344) > >>>>>>>> (__do_softirq) from (__irq_exit_rcu+0x64/0x108) > >>>>>>>> (__irq_exit_rcu) from (irq_exit+0x8/0x10) > >>>>>>>> (irq_exit) from (__handle_domain_irq+0x74/0x9c) > >>>>>>>> (__handle_domain_irq) from (gic_handle_irq+0x8c/0x98) > >>>>>>>> (gic_handle_irq) from (__irq_svc+0x5c/0x94) > >>>>>>>> (__irq_svc) from (arch_cpu_idle+0x20/0x3c) > >>>>>>>> (arch_cpu_idle) from (default_idle_call+0x4c/0x78) > >>>>>>>> (default_idle_call) from (do_idle+0xf8/0x150) > >>>>>>>> (do_idle) from (cpu_startup_entry+0x18/0x20) > >>>>>>>> (cpu_startup_entry) from (0xc01530) > >>>>>>>> > >>>>>>>> This commit therefore adds calls to mem_dump_obj(rhp) to output some > >>>>>>>> information, for example: > >>>>>>>> > >>>>>>>> slab kmalloc-256 start ffff410c45019900 pointer offset 0 size 256 > >>>>>>>> > >>>>>>>> This provides the rough size of the memory block and the offset of > >>>>>>>> the rcu_head structure, which as least provides at least a few clues > >>>>>>>> to help locate the problem. If the problem is reproducible, > >>>>>>>> additional slab debugging can be enabled, for example, > >>>>>>>> CONFIG_DEBUG_SLAB=y, which can provide significantly more information. > >>>>>>>> > >>>>>>>> Signed-off-by: Zhen Lei <thunder.leizhen@xxxxxxxxxx> > >>>>>>>> Signed-off-by: Paul E. McKenney <paulmck@xxxxxxxxxx> > >>>>>>>> --- > >>>>>>>> kernel/rcu/rcu.h | 7 +++++++ > >>>>>>>> kernel/rcu/srcutiny.c | 1 + > >>>>>>>> kernel/rcu/srcutree.c | 1 + > >>>>>>>> kernel/rcu/tasks.h | 1 + > >>>>>>>> kernel/rcu/tiny.c | 1 + > >>>>>>>> kernel/rcu/tree.c | 1 + > >>>>>>>> 6 files changed, 12 insertions(+) > >>>>>>>> > >>>>>>>> v1 --> v2: > >>>>>>>> 1. Remove condition "(unsigned long)rhp->func & 0x3", it have problems on x86. > >>>>>>>> 2. Paul E. McKenney helped me update the commit message, thanks. > >>>>>>>> > >>>>>>> > >>>>>>> Hi, Zhen Lei > >>>>>>> > >>>>>>> Maybe the following scenarios should be considered: > >>>>>>> > >>>>>>> CPU 0 > >>>>>>> tasks context > >>>>>>> spin_lock(&vmap_area_lock) > >>>>>>> Interrupt > >>>>>>> RCU softirq > >>>>>>> rcu_do_batch > >>>>>>> mem_dump_obj > >>>>>>> vmalloc_dump_obj > >>>>>>> spin_lock(&vmap_area_lock) <-- deadlock > >>>>>> > >>>>>>> Right, thanks. I just saw the robot's report. So this patch should be dropped. > >>>>>>> I'll try to add an helper in mm, where I can check whether the lock has been held, and dump the content of memory object. > >>>>>> > >>>>>> This is a workaround, or maybe try a modification like the following, > >>>>>> of course, need to ask Paul's opinion. > >>>>> > >>>>> Another approach is to schedule a workqueue handler to do the > >>>>> mem_dump_obj(). This would allow mem_dump_obj() to run in a clean > >>>>> environment. > >>>> > >>>> It's about to panic, so no chance to schedule. > >>> > >>> It won't panic if you drop the callback on the floor. > >>> > >>> Though to your point, the ->next pointer is likely also trashed. So you > >>> could just drop the remainder of the callback list on the floor. That > >>> might provide a good (though not perfect) chance of getting decent output. > >> > >> OK, I think I understand what you mean. > >> if (!f) > >> schedule_work(&work); > >> else > >> f(rhp) > > > > Yes, except that the "schedule_work()" also needs to be accompanied > > by something that refuses to execute the rest of those callbacks. > > This needs to break out of the loop (or return) and to adjust counts, > > among other things. This might be as easy as setting count to the > > negative of the length of the "rcl" list, but does need some attention > > to the code following the callback-invocation loop. > > Yes, doing so would cause other problems. As you mentioned, the ->next > pointer is likely also trashed. Some nodes may need to be executed in > sequence. For such a weak debug function, it's not worth the risk, or > overly complicated thinking. >Do we have similar deadlock issues with the calls to mem_dump_obj() in >the call_rcu() code path? I think it exists, and also consider PREEMPT_RT kernel , because the vmap_area_lock spinlock convert to sleepable lock. I have sent a version of the patch. Thanks Zqiang > These are somewhat less concerning because >they are invoked under a Kconfig option that is (as far as I know) >rarely set, but still... > > Thanx, Paul > >>>>> This would allow vmalloc_dump_obj() to be called unconditionally. > >>>>> > >>>>> Other thoughts? > >>>> > >>>> locked = spin_is_locked(&vmap_area_lock); > >>>> if (!locked) > >>>> spin_lock(&vmap_area_lock) > >>>> > >>>> Careful analysis is required, which may cause other problems. > >>>> > >>>> Or in new function: > >>>> if (locked) > >>>> return; > >>>> spin_lock(&vmap_area_lock); > >>>> > >>>> If there is a chance to dump the data, dump the data. If there is no > >>>> chance to dump the data, do not dump the data. This is the fate of > >>>> debugging information. > >>> > >>> My concern is that there will be increasing numbers of special cases > >>> over time. > > The memory modules are mature and stable, so your concerns may not be true. > > >> > >> OK, I got it. > >> > >>> > >>> Thanx, Paul > >>> > >>>>>> diff --git a/mm/util.c b/mm/util.c > >>>>>> index 12984e76767e..86da0739fe5d 100644 > >>>>>> --- a/mm/util.c > >>>>>> +++ b/mm/util.c > >>>>>> @@ -1119,14 +1119,18 @@ void mem_dump_obj(void *object) > >>>>>> { > >>>>>> const char *type; > >>>>>> > >>>>>> + if (is_vmalloc_addr(object)) { > >>>>>> + if (in_task() && vmalloc_dump_obj(object)) > >>>>>> + return; > >>>>>> + type = "vmalloc memory"; > >>>>>> + goto end; > >>>>>> + } > >>>>>> + > >>>>>> if (kmem_valid_obj(object)) { > >>>>>> kmem_dump_obj(object); > >>>>>> return; > >>>>>> } > >>>>>> > >>>>>> - if (vmalloc_dump_obj(object)) > >>>>>> - return; > >>>>>> - > >>>>>> if (virt_addr_valid(object)) > >>>>>> type = "non-slab/vmalloc memory"; > >>>>>> else if (object == NULL) > >>>>>> @@ -1135,7 +1139,7 @@ void mem_dump_obj(void *object) > >>>>>> type = "zero-size pointer"; > >>>>>> else > >>>>>> type = "non-paged memory"; > >>>>>> - > >>>>>> +end: > >>>>>> pr_cont(" %s\n", type); > >>>>>> } > >>>>>> EXPORT_SYMBOL_GPL(mem_dump_obj); > >>>>>> > >>>>>> Thanks > >>>>>> Zqiang > >>>>>> > >>>>>> > >>>>>>> > >>>>>>> Thanks > >>>>>>> Zqiang > >>>>>>> > >>>>>>> > >>>>>>>> diff --git a/kernel/rcu/rcu.h b/kernel/rcu/rcu.h index > >>>>>>>> 65704cbc9df7b3d..32ab45fabf8eebf 100644 > >>>>>>>> --- a/kernel/rcu/rcu.h > >>>>>>>> +++ b/kernel/rcu/rcu.h > >>>>>>>> @@ -10,6 +10,7 @@ > >>>>>>>> #ifndef __LINUX_RCU_H > >>>>>>>> #define __LINUX_RCU_H > >>>>>>>> > >>>>>>>> +#include <linux/mm.h> > >>>>>>>> #include <trace/events/rcu.h> > >>>>>>>> > >>>>>>>> /* > >>>>>>>> @@ -211,6 +212,12 @@ static inline void debug_rcu_head_unqueue(struct > >>>>>>>> rcu_head *head) } > >>>>>>>> #endif /* #else !CONFIG_DEBUG_OBJECTS_RCU_HEAD */ > >>>>>>>> > >>>>>>>> +static inline void debug_rcu_head_callback(struct rcu_head *rhp) { > >>>>>>>> + if (unlikely(!rhp->func)) > >>>>>>>> + mem_dump_obj(rhp); > >>>>>>>> +} > >>>>>>>> + > >>>>>>>> extern int rcu_cpu_stall_suppress_at_boot; > >>>>>>>> > >>>>>>>> static inline bool rcu_stall_is_suppressed_at_boot(void) > >>>>>>>> diff --git a/kernel/rcu/srcutiny.c b/kernel/rcu/srcutiny.c index > >>>>>>>> 33adafdad261389..5e7f336baa06ae0 100644 > >>>>>>>> --- a/kernel/rcu/srcutiny.c > >>>>>>>> +++ b/kernel/rcu/srcutiny.c > >>>>>>>> @@ -138,6 +138,7 @@ void srcu_drive_gp(struct work_struct *wp) > >>>>>>>> while (lh) { > >>>>>>>> rhp = lh; > >>>>>>>> lh = lh->next; > >>>>>>>> + debug_rcu_head_callback(rhp); > >>>>>>>> local_bh_disable(); > >>>>>>>> rhp->func(rhp); > >>>>>>>> local_bh_enable(); > >>>>>>>> diff --git a/kernel/rcu/srcutree.c b/kernel/rcu/srcutree.c index > >>>>>>>> ca4b5dcec675bac..294972e66b31863 100644 > >>>>>>>> --- a/kernel/rcu/srcutree.c > >>>>>>>> +++ b/kernel/rcu/srcutree.c > >>>>>>>> @@ -1631,6 +1631,7 @@ static void srcu_invoke_callbacks(struct work_struct *work) > >>>>>>>> rhp = rcu_cblist_dequeue(&ready_cbs); > >>>>>>>> for (; rhp != NULL; rhp = rcu_cblist_dequeue(&ready_cbs)) { > >>>>>>>> debug_rcu_head_unqueue(rhp); > >>>>>>>> + debug_rcu_head_callback(rhp); > >>>>>>>> local_bh_disable(); > >>>>>>>> rhp->func(rhp); > >>>>>>>> local_bh_enable(); > >>>>>>>> diff --git a/kernel/rcu/tasks.h b/kernel/rcu/tasks.h index > >>>>>>>> b0b885e071fa8dc..b7f8c67c586cdc4 100644 > >>>>>>>> --- a/kernel/rcu/tasks.h > >>>>>>>> +++ b/kernel/rcu/tasks.h > >>>>>>>> @@ -478,6 +478,7 @@ static void rcu_tasks_invoke_cbs(struct rcu_tasks *rtp, struct rcu_tasks_percpu > >>>>>>>> raw_spin_unlock_irqrestore_rcu_node(rtpcp, flags); > >>>>>>>> len = rcl.len; > >>>>>>>> for (rhp = rcu_cblist_dequeue(&rcl); rhp; rhp = > >>>>>>>> rcu_cblist_dequeue(&rcl)) { > >>>>>>>> + debug_rcu_head_callback(rhp); > >>>>>>>> local_bh_disable(); > >>>>>>>> rhp->func(rhp); > >>>>>>>> local_bh_enable(); > >>>>>>>> diff --git a/kernel/rcu/tiny.c b/kernel/rcu/tiny.c index > >>>>>>>> bb8f7d270f01747..56e9a5d91d97ec5 100644 > >>>>>>>> --- a/kernel/rcu/tiny.c > >>>>>>>> +++ b/kernel/rcu/tiny.c > >>>>>>>> @@ -97,6 +97,7 @@ static inline bool rcu_reclaim_tiny(struct rcu_head > >>>>>>>> *head) > >>>>>>>> > >>>>>>>> trace_rcu_invoke_callback("", head); > >>>>>>>> f = head->func; > >>>>>>>> + debug_rcu_head_callback(head); > >>>>>>>> WRITE_ONCE(head->func, (rcu_callback_t)0L); > >>>>>>>> f(head); > >>>>>>>> rcu_lock_release(&rcu_callback_map); > >>>>>>>> diff --git a/kernel/rcu/tree.c b/kernel/rcu/tree.c index > >>>>>>>> 15aaff3203bf2d0..ed93ddb8203d42c 100644 > >>>>>>>> --- a/kernel/rcu/tree.c > >>>>>>>> +++ b/kernel/rcu/tree.c > >>>>>>>> @@ -2088,6 +2088,7 @@ static void rcu_do_batch(struct rcu_data *rdp) > >>>>>>>> trace_rcu_invoke_callback(rcu_state.name, rhp); > >>>>>>>> > >>>>>>>> f = rhp->func; > >>>>>>>> + debug_rcu_head_callback(rhp); > >>>>>>>> WRITE_ONCE(rhp->func, (rcu_callback_t)0L); > >>>>>>>> f(rhp); > >>>>>>>> > >>>>>>>> -- > >>>>>>>> 2.25.1 > >>>>>>> > >>>>>>> . > >>>>>>> > >>>>>> > >>>>>> -- > >>>>>> Regards, > >>>>>> Zhen Lei > >>>>> . > >>>>> > >>>> > >>>> -- > >>>> Regards, > >>>> Zhen Lei > >>> . > >>> > >> > >> -- > >> Regards, > >> Zhen Lei > > . > > > > -- > Regards, > Zhen Lei