Re: [PATCH v2 1/1] rcu: Reduce synchronize_rcu() waiting time

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Tue, Oct 03, 2023 at 08:04:03PM +0200, Uladzislau Rezki (Sony) wrote:
> A call to a synchronize_rcu() can be optimized from time point of
> view. Different workloads can be affected by this especially the
> ones which use this API in its time critical sections.
> 
> For example if CONFIG_RCU_NOCB_CPU is set, the wakeme_after_rcu()
> callback can be delayed and such delay depends on where in a nocb
> list it is located.
> 
> 1. On our Android devices i can easily trigger the scenario when
> it is a last in the list out of ~3600 callbacks:
> 
> <snip>
>   <...>-29      [001] d..1. 21950.145313: rcu_batch_start: rcu_preempt CBs=3613 bl=28
> ...
>   <...>-29      [001] ..... 21950.152578: rcu_invoke_callback: rcu_preempt rhp=00000000b2d6dee8 func=__free_vm_area_struct.cfi_jt
>   <...>-29      [001] ..... 21950.152579: rcu_invoke_callback: rcu_preempt rhp=00000000a446f607 func=__free_vm_area_struct.cfi_jt
>   <...>-29      [001] ..... 21950.152580: rcu_invoke_callback: rcu_preempt rhp=00000000a5cab03b func=__free_vm_area_struct.cfi_jt
>   <...>-29      [001] ..... 21950.152581: rcu_invoke_callback: rcu_preempt rhp=0000000013b7e5ee func=__free_vm_area_struct.cfi_jt
>   <...>-29      [001] ..... 21950.152582: rcu_invoke_callback: rcu_preempt rhp=000000000a8ca6f9 func=__free_vm_area_struct.cfi_jt
>   <...>-29      [001] ..... 21950.152583: rcu_invoke_callback: rcu_preempt rhp=000000008f162ca8 func=wakeme_after_rcu.cfi_jt
>   <...>-29      [001] d..1. 21950.152625: rcu_batch_end: rcu_preempt CBs-invoked=3612 idle=....
> <snip>
> 
> 2. On our Android devices we use cpuset/cgroup to classify tasks
> and assign them into different cgroups. For example "backgrond"
> group which binds tasks only to little CPUs or "foreground" that
> binds to all CPUs, i.e. tasks can be migrated between groups.
> 
> See below an example of how "surfaceflinger" task is migrated.
> Initially it is located in the "system-background" cgroup which
> allows to run only on little cores. In order to speedup it up
> it can be temporary moved into "foreground" cgroup which allows
> to use big CPUs:
> 
> cgroup_attach_task():
>  -> cgroup_migrate_execute()
>    -> cpuset_can_attach()
>      -> percpu_down_write()
>        -> rcu_sync_enter()
>          -> synchronize_rcu()
>    -> now move tasks to the new cgroup.
>  -> cgroup_migrate_finish()
> 
> <snip>
>          rcuop/1-29      [000] .....  7030.528570: rcu_invoke_callback: rcu_preempt rhp=00000000461605e0 func=wakeme_after_rcu.cfi_jt
>     PERFD-SERVER-1855    [000] d..1.  7030.530293: cgroup_attach_task: dst_root=3 dst_id=22 dst_level=1 dst_path=/foreground pid=1900 comm=surfaceflinger
>     PERFD-SERVER-1855    [000] d..1.  7030.530383: cgroup_attach_task: dst_root=3 dst_id=22 dst_level=1 dst_path=/foreground pid=1900 comm=surfaceflinger
>    TimerDispatch-2768    [002] d..5.  7030.537542: sched_migrate_task: comm=surfaceflinger pid=1900 prio=98 orig_cpu=0 dest_cpu=4
> <snip>
> 
> from this example it is clear that "a moving time" also depends
> on how fast synchronize_rcu() completes.
> 
> 3. This patch improves the synchronize_rcu() approximately by 30%-50%
> on synthetic tests. Apart of that i have tested app launch of camera
> app where i also see better perf. figures:
> 
> 542 vs 489 diff: 9%
> 540 vs 466 diff: 13%
> 518 vs 468 diff: 9%
> 531 vs 457 diff: 13%
> 548 vs 475 diff: 13%
> 509 vs 484 diff: 4%
> 
> Signed-off-by: Uladzislau Rezki (Sony) <urezki@xxxxxxxxx>
> ---
>  kernel/rcu/tree.c     | 151 +++++++++++++++++++++++++++++++++++++++++-
>  kernel/rcu/tree_exp.h |   2 +-
>  2 files changed, 151 insertions(+), 2 deletions(-)
> 
> diff --git a/kernel/rcu/tree.c b/kernel/rcu/tree.c
> index 78554e7181dd..a347c1f98f11 100644
> --- a/kernel/rcu/tree.c
> +++ b/kernel/rcu/tree.c
> @@ -1384,6 +1384,122 @@ static void rcu_poll_gp_seq_end_unlocked(unsigned long *snap)
>  		raw_spin_unlock_irqrestore_rcu_node(rnp, flags);
>  }
>  
> +/*
> + * There are three lists for handling synchronize_rcu() users.
> + * A first list corresponds to new coming users, second for users
> + * which wait for a grace period and third is for which a grace
> + * period is passed.
> + */
> +static struct sr_normal_state {
> +	struct llist_head curr;	/* request a GP users. */
> +	struct llist_head wait;	/* wait for GP users. */
> +	struct llist_head done;	/* ready for GP users. */
> +	struct llist_node *curr_tail;
> +	struct llist_node *wait_tail;
> +	atomic_t active;
> +} sr;
> +
> +/* Enable it by default. */
> +static int rcu_normal_wake_from_gp = 1;
> +module_param(rcu_normal_wake_from_gp, int, 0644);

Nice!

But could you please make this default to zero in order to avoid
surprising people for whom the old way works better?

							Thanx, Paul

> +static void rcu_sr_normal_complete(struct llist_node *node)
> +{
> +	struct rcu_synchronize *rs = container_of(
> +		(struct rcu_head *) node, struct rcu_synchronize, head);
> +	unsigned long oldstate = (unsigned long) rs->head.func;
> +
> +	if (!poll_state_synchronize_rcu(oldstate))
> +		WARN_ONCE(1, "A full grace period is not passed yet: %lu",
> +			rcu_seq_diff(get_state_synchronize_rcu(), oldstate));
> +
> +	/* Finally. */
> +	complete(&rs->completion);
> +}
> +
> +static void rcu_sr_normal_gp_cleanup_work(struct work_struct *work)
> +{
> +	struct llist_node *done, *rcu, *next;
> +
> +	done = llist_del_all(&sr.done);
> +	if (!done)
> +		return;
> +
> +	llist_for_each_safe(rcu, next, done)
> +		rcu_sr_normal_complete(rcu);
> +}
> +static DECLARE_WORK(sr_normal_gp_cleanup, rcu_sr_normal_gp_cleanup_work);
> +
> +/*
> + * Helper function for rcu_gp_cleanup().
> + */
> +static void rcu_sr_normal_gp_cleanup(void)
> +{
> +	struct llist_node *first, *tail;
> +
> +	tail = READ_ONCE(sr.wait_tail);
> +	first = llist_del_all(&sr.wait);
> +	if (!first)
> +		return;
> +
> +	/* Only one user? */
> +	if (!first->next) {
> +		rcu_sr_normal_complete(first);
> +		return;
> +	}
> +
> +	/* Can be not empty. */
> +	llist_add_batch(first, tail, &sr.done);
> +	queue_work(system_highpri_wq, &sr_normal_gp_cleanup);
> +}
> +
> +/*
> + * Helper function for rcu_gp_init().
> + */
> +static void rcu_sr_normal_gp_init(void)
> +{
> +	struct llist_node *llnode, *rcu;
> +	int ret;
> +
> +	if (llist_empty(&sr.curr))
> +		return;
> +
> +	/*
> +	 * A waiting list of GP should be empty on this step,
> +	 * since a GP-kthread, rcu_gp_init() -> gp_cleanup(),
> +	 * rolls it over. If not, it is a BUG, warn a user.
> +	 */
> +	WARN_ON_ONCE(!llist_empty(&sr.wait));
> +
> +	/*
> +	 * Obtain a tail of current active users. It is guaranteed
> +	 * that if we are only one active user and the list is not
> +	 * empty, the tail has already been updated.
> +	 */
> +	ret = atomic_inc_return(&sr.active);
> +	WRITE_ONCE(sr.wait_tail, (ret == 1) ? READ_ONCE(sr.curr_tail):NULL);
> +	llnode = llist_del_all(&sr.curr);
> +	atomic_dec(&sr.active);
> +
> +	if (ret != 1) {
> +		llist_for_each(rcu, llnode) {
> +			if (!rcu->next)
> +				WRITE_ONCE(sr.wait_tail, rcu);
> +		}
> +	}
> +
> +	llist_add_batch(llnode, READ_ONCE(sr.wait_tail), &sr.wait);
> +}
> +
> +static void rcu_sr_normal_add_req(struct rcu_synchronize *rs)
> +{
> +	atomic_inc(&sr.active);
> +	if (llist_add((struct llist_node *) &rs->head, &sr.curr))
> +		/* Set the tail. Only first and one user can do that. */
> +		WRITE_ONCE(sr.curr_tail, (struct llist_node *) &rs->head);
> +	atomic_dec(&sr.active);
> +}
> +
>  /*
>   * Initialize a new grace period.  Return false if no grace period required.
>   */
> @@ -1420,6 +1536,7 @@ static noinline_for_stack bool rcu_gp_init(void)
>  	ASSERT_EXCLUSIVE_WRITER(rcu_state.gp_seq);
>  	trace_rcu_grace_period(rcu_state.name, rcu_state.gp_seq, TPS("start"));
>  	rcu_poll_gp_seq_start(&rcu_state.gp_seq_polled_snap);
> +	rcu_sr_normal_gp_init();
>  	raw_spin_unlock_irq_rcu_node(rnp);
>  
>  	/*
> @@ -1787,6 +1904,9 @@ static noinline void rcu_gp_cleanup(void)
>  	}
>  	raw_spin_unlock_irq_rcu_node(rnp);
>  
> +	// Make synchronize_rcu() users aware of the end of old grace period.
> +	rcu_sr_normal_gp_cleanup();
> +
>  	// If strict, make all CPUs aware of the end of the old grace period.
>  	if (IS_ENABLED(CONFIG_RCU_STRICT_GRACE_PERIOD))
>  		on_each_cpu(rcu_strict_gp_boundary, NULL, 0);
> @@ -3500,6 +3620,35 @@ static int rcu_blocking_is_gp(void)
>  	return true;
>  }
>  
> +/*
> + * Helper function for the synchronize_rcu() API.
> + */
> +static void synchronize_rcu_normal(void)
> +{
> +	struct rcu_synchronize rs;
> +
> +	if (READ_ONCE(rcu_normal_wake_from_gp)) {
> +		init_rcu_head_on_stack(&rs.head);
> +		init_completion(&rs.completion);
> +
> +		/*
> +		 * This code might be preempted, therefore take a GP
> +		 * snapshot before adding a request.
> +		 */
> +		rs.head.func = (void *) get_state_synchronize_rcu();
> +		rcu_sr_normal_add_req(&rs);
> +
> +		/* Kick a GP and start waiting. */
> +		(void) start_poll_synchronize_rcu();
> +
> +		/* Now we can wait. */
> +		wait_for_completion(&rs.completion);
> +		destroy_rcu_head_on_stack(&rs.head);
> +	} else {
> +		wait_rcu_gp(call_rcu_hurry);
> +	}
> +}
> +
>  /**
>   * synchronize_rcu - wait until a grace period has elapsed.
>   *
> @@ -3551,7 +3700,7 @@ void synchronize_rcu(void)
>  		if (rcu_gp_is_expedited())
>  			synchronize_rcu_expedited();
>  		else
> -			wait_rcu_gp(call_rcu_hurry);
> +			synchronize_rcu_normal();
>  		return;
>  	}
>  
> diff --git a/kernel/rcu/tree_exp.h b/kernel/rcu/tree_exp.h
> index 6d7cea5d591f..279a37beb05a 100644
> --- a/kernel/rcu/tree_exp.h
> +++ b/kernel/rcu/tree_exp.h
> @@ -987,7 +987,7 @@ void synchronize_rcu_expedited(void)
>  
>  	/* If expedited grace periods are prohibited, fall back to normal. */
>  	if (rcu_gp_is_normal()) {
> -		wait_rcu_gp(call_rcu_hurry);
> +		synchronize_rcu_normal();
>  		return;
>  	}
>  
> -- 
> 2.30.2
> 



[Index of Archives]     [Linux Samsung SoC]     [Linux Rockchip SoC]     [Linux Actions SoC]     [Linux for Synopsys ARC Processors]     [Linux NFS]     [Linux NILFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]


  Powered by Linux