rcu: Avoid __call_rcu_core() root rcu_node ->lock acquisition
When __call_rcu_core() notices excessive numbers of callbacks pending on the current CPU, we know that at least one of them is not yet classified, namely the one that was just now queued. Therefore, it is not necessary to invoke rcu_start_gp() and thus not necessary to acquire the root rcu_node structure's ->lock. This commit therefore replaces the rcu_start_gp() with rcu_accelerate_cbs(), thus replacing an acquisition of the root rcu_node structure's ->lock with that of this CPU's leaf rcu_node structure. This decreases contention on the root rcu_node structure's ->lock. Reported-by: Nicholas Piggin <npiggin@gmail.com> Signed-off-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com> Tested-by: Nicholas Piggin <npiggin@gmail.com>
This commit is contained in:
parent
ec4eaccef4
commit
a6058d85a2
|
@ -2988,11 +2988,11 @@ static void __call_rcu_core(struct rcu_state *rsp, struct rcu_data *rdp,
|
||||||
|
|
||||||
/* Start a new grace period if one not already started. */
|
/* Start a new grace period if one not already started. */
|
||||||
if (!rcu_gp_in_progress(rsp)) {
|
if (!rcu_gp_in_progress(rsp)) {
|
||||||
struct rcu_node *rnp_root = rcu_get_root(rsp);
|
struct rcu_node *rnp = rdp->mynode;
|
||||||
|
|
||||||
raw_spin_lock_rcu_node(rnp_root);
|
raw_spin_lock_rcu_node(rnp);
|
||||||
needwake = rcu_start_gp(rsp);
|
needwake = rcu_accelerate_cbs(rsp, rnp, rdp);
|
||||||
raw_spin_unlock_rcu_node(rnp_root);
|
raw_spin_unlock_rcu_node(rnp);
|
||||||
if (needwake)
|
if (needwake)
|
||||||
rcu_gp_kthread_wake(rsp);
|
rcu_gp_kthread_wake(rsp);
|
||||||
} else {
|
} else {
|
||||||
|
|
Loading…
Reference in New Issue