rcu: Eliminate flavor scan in rcu_momentary_dyntick_idle()
The rcu_momentary_dyntick_idle() function scans the RCU flavors, checking that one of them still needs a quiescent state before doing an expensive atomic operation on the ->dynticks counter. However, this check reduces overhead only after a rare race condition, and increases complexity. This commit therefore removes the scan and the mechanism enabling the scan. Signed-off-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
This commit is contained in:
parent
9577df9a31
commit
0f9be8cabb
|
@ -1104,7 +1104,7 @@ Its fields are as follows:
|
||||||
1 int dynticks_nesting;
|
1 int dynticks_nesting;
|
||||||
2 int dynticks_nmi_nesting;
|
2 int dynticks_nmi_nesting;
|
||||||
3 atomic_t dynticks;
|
3 atomic_t dynticks;
|
||||||
4 int rcu_sched_qs_mask;
|
4 bool rcu_need_heavy_qs;
|
||||||
5 unsigned long rcu_qs_ctr;
|
5 unsigned long rcu_qs_ctr;
|
||||||
</pre>
|
</pre>
|
||||||
|
|
||||||
|
@ -1124,7 +1124,7 @@ CPU's transitions to and from dyntick-idle mode, so that this counter
|
||||||
has an even value when the CPU is in dyntick-idle mode and an odd
|
has an even value when the CPU is in dyntick-idle mode and an odd
|
||||||
value otherwise.
|
value otherwise.
|
||||||
|
|
||||||
</p><p>The <tt>->rcu_sched_qs_mask</tt> field is used
|
</p><p>The <tt>->rcu_need_heavy_qs</tt> field is used
|
||||||
to record the fact that the RCU core code would really like to
|
to record the fact that the RCU core code would really like to
|
||||||
see a quiescent state from the corresponding CPU, so much so that
|
see a quiescent state from the corresponding CPU, so much so that
|
||||||
it is willing to call for heavy-weight dyntick-counter operations.
|
it is willing to call for heavy-weight dyntick-counter operations.
|
||||||
|
|
|
@ -443,44 +443,14 @@ bool rcu_eqs_special_set(int cpu)
|
||||||
* memory barriers to let the RCU core know about it, regardless of what
|
* memory barriers to let the RCU core know about it, regardless of what
|
||||||
* this CPU might (or might not) do in the near future.
|
* this CPU might (or might not) do in the near future.
|
||||||
*
|
*
|
||||||
* We inform the RCU core by emulating a zero-duration dyntick-idle
|
* We inform the RCU core by emulating a zero-duration dyntick-idle period.
|
||||||
* period, which we in turn do by incrementing the ->dynticks counter
|
|
||||||
* by two.
|
|
||||||
*
|
*
|
||||||
* The caller must have disabled interrupts.
|
* The caller must have disabled interrupts.
|
||||||
*/
|
*/
|
||||||
static void rcu_momentary_dyntick_idle(void)
|
static void rcu_momentary_dyntick_idle(void)
|
||||||
{
|
{
|
||||||
struct rcu_data *rdp;
|
raw_cpu_write(rcu_dynticks.rcu_need_heavy_qs, false);
|
||||||
int resched_mask;
|
rcu_dynticks_momentary_idle();
|
||||||
struct rcu_state *rsp;
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Yes, we can lose flag-setting operations. This is OK, because
|
|
||||||
* the flag will be set again after some delay.
|
|
||||||
*/
|
|
||||||
resched_mask = raw_cpu_read(rcu_dynticks.rcu_sched_qs_mask);
|
|
||||||
raw_cpu_write(rcu_dynticks.rcu_sched_qs_mask, 0);
|
|
||||||
|
|
||||||
/* Find the flavor that needs a quiescent state. */
|
|
||||||
for_each_rcu_flavor(rsp) {
|
|
||||||
rdp = raw_cpu_ptr(rsp->rda);
|
|
||||||
if (!(resched_mask & rsp->flavor_mask))
|
|
||||||
continue;
|
|
||||||
smp_mb(); /* rcu_sched_qs_mask before cond_resched_completed. */
|
|
||||||
if (READ_ONCE(rdp->mynode->completed) !=
|
|
||||||
READ_ONCE(rdp->cond_resched_completed))
|
|
||||||
continue;
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Pretend to be momentarily idle for the quiescent state.
|
|
||||||
* This allows the grace-period kthread to record the
|
|
||||||
* quiescent state, with no need for this CPU to do anything
|
|
||||||
* further.
|
|
||||||
*/
|
|
||||||
rcu_dynticks_momentary_idle();
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@ -494,7 +464,7 @@ void rcu_note_context_switch(void)
|
||||||
trace_rcu_utilization(TPS("Start context switch"));
|
trace_rcu_utilization(TPS("Start context switch"));
|
||||||
rcu_sched_qs();
|
rcu_sched_qs();
|
||||||
rcu_preempt_note_context_switch();
|
rcu_preempt_note_context_switch();
|
||||||
if (unlikely(raw_cpu_read(rcu_dynticks.rcu_sched_qs_mask)))
|
if (unlikely(raw_cpu_read(rcu_dynticks.rcu_need_heavy_qs)))
|
||||||
rcu_momentary_dyntick_idle();
|
rcu_momentary_dyntick_idle();
|
||||||
trace_rcu_utilization(TPS("End context switch"));
|
trace_rcu_utilization(TPS("End context switch"));
|
||||||
barrier(); /* Avoid RCU read-side critical sections leaking up. */
|
barrier(); /* Avoid RCU read-side critical sections leaking up. */
|
||||||
|
@ -519,7 +489,7 @@ void rcu_all_qs(void)
|
||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
|
|
||||||
barrier(); /* Avoid RCU read-side critical sections leaking down. */
|
barrier(); /* Avoid RCU read-side critical sections leaking down. */
|
||||||
if (unlikely(raw_cpu_read(rcu_dynticks.rcu_sched_qs_mask))) {
|
if (unlikely(raw_cpu_read(rcu_dynticks.rcu_need_heavy_qs))) {
|
||||||
local_irq_save(flags);
|
local_irq_save(flags);
|
||||||
rcu_momentary_dyntick_idle();
|
rcu_momentary_dyntick_idle();
|
||||||
local_irq_restore(flags);
|
local_irq_restore(flags);
|
||||||
|
@ -1275,7 +1245,7 @@ static int rcu_implicit_dynticks_qs(struct rcu_data *rdp,
|
||||||
bool *isidle, unsigned long *maxj)
|
bool *isidle, unsigned long *maxj)
|
||||||
{
|
{
|
||||||
unsigned long jtsq;
|
unsigned long jtsq;
|
||||||
int *rcrmp;
|
bool *rnhqp;
|
||||||
unsigned long rjtsc;
|
unsigned long rjtsc;
|
||||||
struct rcu_node *rnp;
|
struct rcu_node *rnp;
|
||||||
|
|
||||||
|
@ -1332,7 +1302,7 @@ static int rcu_implicit_dynticks_qs(struct rcu_data *rdp,
|
||||||
* in-kernel CPU-bound tasks cannot advance grace periods.
|
* in-kernel CPU-bound tasks cannot advance grace periods.
|
||||||
* So if the grace period is old enough, make the CPU pay attention.
|
* So if the grace period is old enough, make the CPU pay attention.
|
||||||
* Note that the unsynchronized assignments to the per-CPU
|
* Note that the unsynchronized assignments to the per-CPU
|
||||||
* rcu_sched_qs_mask variable are safe. Yes, setting of
|
* rcu_need_heavy_qs variable are safe. Yes, setting of
|
||||||
* bits can be lost, but they will be set again on the next
|
* bits can be lost, but they will be set again on the next
|
||||||
* force-quiescent-state pass. So lost bit sets do not result
|
* force-quiescent-state pass. So lost bit sets do not result
|
||||||
* in incorrect behavior, merely in a grace period lasting
|
* in incorrect behavior, merely in a grace period lasting
|
||||||
|
@ -1346,16 +1316,11 @@ static int rcu_implicit_dynticks_qs(struct rcu_data *rdp,
|
||||||
* is set too high, we override with half of the RCU CPU stall
|
* is set too high, we override with half of the RCU CPU stall
|
||||||
* warning delay.
|
* warning delay.
|
||||||
*/
|
*/
|
||||||
rcrmp = &per_cpu(rcu_dynticks.rcu_sched_qs_mask, rdp->cpu);
|
rnhqp = &per_cpu(rcu_dynticks.rcu_need_heavy_qs, rdp->cpu);
|
||||||
if (time_after(jiffies, rdp->rsp->gp_start + jtsq) ||
|
if (!READ_ONCE(*rnhqp) &&
|
||||||
time_after(jiffies, rdp->rsp->jiffies_resched)) {
|
(time_after(jiffies, rdp->rsp->gp_start + jtsq) ||
|
||||||
if (!(READ_ONCE(*rcrmp) & rdp->rsp->flavor_mask)) {
|
time_after(jiffies, rdp->rsp->jiffies_resched))) {
|
||||||
WRITE_ONCE(rdp->cond_resched_completed,
|
WRITE_ONCE(*rnhqp, true);
|
||||||
READ_ONCE(rdp->mynode->completed));
|
|
||||||
smp_mb(); /* ->cond_resched_completed before *rcrmp. */
|
|
||||||
WRITE_ONCE(*rcrmp,
|
|
||||||
READ_ONCE(*rcrmp) + rdp->rsp->flavor_mask);
|
|
||||||
}
|
|
||||||
rdp->rsp->jiffies_resched += 5; /* Re-enable beating. */
|
rdp->rsp->jiffies_resched += 5; /* Re-enable beating. */
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -4169,7 +4134,6 @@ static void __init rcu_init_one(struct rcu_state *rsp)
|
||||||
static const char * const fqs[] = RCU_FQS_NAME_INIT;
|
static const char * const fqs[] = RCU_FQS_NAME_INIT;
|
||||||
static struct lock_class_key rcu_node_class[RCU_NUM_LVLS];
|
static struct lock_class_key rcu_node_class[RCU_NUM_LVLS];
|
||||||
static struct lock_class_key rcu_fqs_class[RCU_NUM_LVLS];
|
static struct lock_class_key rcu_fqs_class[RCU_NUM_LVLS];
|
||||||
static u8 fl_mask = 0x1;
|
|
||||||
|
|
||||||
int levelcnt[RCU_NUM_LVLS]; /* # nodes in each level. */
|
int levelcnt[RCU_NUM_LVLS]; /* # nodes in each level. */
|
||||||
int levelspread[RCU_NUM_LVLS]; /* kids/node in each level. */
|
int levelspread[RCU_NUM_LVLS]; /* kids/node in each level. */
|
||||||
|
@ -4191,8 +4155,6 @@ static void __init rcu_init_one(struct rcu_state *rsp)
|
||||||
for (i = 1; i < rcu_num_lvls; i++)
|
for (i = 1; i < rcu_num_lvls; i++)
|
||||||
rsp->level[i] = rsp->level[i - 1] + levelcnt[i - 1];
|
rsp->level[i] = rsp->level[i - 1] + levelcnt[i - 1];
|
||||||
rcu_init_levelspread(levelspread, levelcnt);
|
rcu_init_levelspread(levelspread, levelcnt);
|
||||||
rsp->flavor_mask = fl_mask;
|
|
||||||
fl_mask <<= 1;
|
|
||||||
|
|
||||||
/* Initialize the elements themselves, starting from the leaves. */
|
/* Initialize the elements themselves, starting from the leaves. */
|
||||||
|
|
||||||
|
|
|
@ -113,7 +113,7 @@ struct rcu_dynticks {
|
||||||
/* Process level is worth LLONG_MAX/2. */
|
/* Process level is worth LLONG_MAX/2. */
|
||||||
int dynticks_nmi_nesting; /* Track NMI nesting level. */
|
int dynticks_nmi_nesting; /* Track NMI nesting level. */
|
||||||
atomic_t dynticks; /* Even value for idle, else odd. */
|
atomic_t dynticks; /* Even value for idle, else odd. */
|
||||||
int rcu_sched_qs_mask; /* GP old, need heavy quiescent state. */
|
bool rcu_need_heavy_qs; /* GP old, need heavy quiescent state. */
|
||||||
unsigned long rcu_qs_ctr; /* Light universal quiescent state ctr. */
|
unsigned long rcu_qs_ctr; /* Light universal quiescent state ctr. */
|
||||||
#ifdef CONFIG_NO_HZ_FULL_SYSIDLE
|
#ifdef CONFIG_NO_HZ_FULL_SYSIDLE
|
||||||
long long dynticks_idle_nesting;
|
long long dynticks_idle_nesting;
|
||||||
|
@ -484,7 +484,6 @@ struct rcu_state {
|
||||||
struct rcu_node *level[RCU_NUM_LVLS + 1];
|
struct rcu_node *level[RCU_NUM_LVLS + 1];
|
||||||
/* Hierarchy levels (+1 to */
|
/* Hierarchy levels (+1 to */
|
||||||
/* shut bogus gcc warning) */
|
/* shut bogus gcc warning) */
|
||||||
u8 flavor_mask; /* bit in flavor mask. */
|
|
||||||
struct rcu_data __percpu *rda; /* pointer of percu rcu_data. */
|
struct rcu_data __percpu *rda; /* pointer of percu rcu_data. */
|
||||||
call_rcu_func_t call; /* call_rcu() flavor. */
|
call_rcu_func_t call; /* call_rcu() flavor. */
|
||||||
int ncpus; /* # CPUs seen so far. */
|
int ncpus; /* # CPUs seen so far. */
|
||||||
|
|
Loading…
Reference in New Issue