x86: Use generic idle loop
Signed-off-by: Thomas Gleixner <tglx@linutronix.de> Cc: Linus Torvalds <torvalds@linux-foundation.org> Cc: Rusty Russell <rusty@rustcorp.com.au> Cc: Paul McKenney <paulmck@linux.vnet.ibm.com> Cc: Peter Zijlstra <peterz@infradead.org> Reviewed-by: Srivatsa S. Bhat <srivatsa.bhat@linux.vnet.ibm.com> Cc: Magnus Damm <magnus.damm@gmail.com> Link: http://lkml.kernel.org/r/20130321215235.486594473@linutronix.de Signed-off-by: Thomas Gleixner <tglx@linutronix.de> Cc: x86@kernel.org
This commit is contained in:
parent
aba92c9e2c
commit
7d1a941731
|
@ -97,6 +97,7 @@ config X86
|
||||||
select GENERIC_IOMAP
|
select GENERIC_IOMAP
|
||||||
select DCACHE_WORD_ACCESS
|
select DCACHE_WORD_ACCESS
|
||||||
select GENERIC_SMP_IDLE_THREAD
|
select GENERIC_SMP_IDLE_THREAD
|
||||||
|
select GENERIC_IDLE_LOOP
|
||||||
select ARCH_WANT_IPC_PARSE_VERSION if X86_32
|
select ARCH_WANT_IPC_PARSE_VERSION if X86_32
|
||||||
select HAVE_ARCH_SECCOMP_FILTER
|
select HAVE_ARCH_SECCOMP_FILTER
|
||||||
select BUILDTIME_EXTABLE_SORT
|
select BUILDTIME_EXTABLE_SORT
|
||||||
|
|
|
@ -301,13 +301,7 @@ void exit_idle(void)
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
/*
|
void arch_cpu_idle_prepare(void)
|
||||||
* The idle thread. There's no useful work to be
|
|
||||||
* done, so just try to conserve power and have a
|
|
||||||
* low exit latency (ie sit in a loop waiting for
|
|
||||||
* somebody to say that they'd like to reschedule)
|
|
||||||
*/
|
|
||||||
void cpu_idle(void)
|
|
||||||
{
|
{
|
||||||
/*
|
/*
|
||||||
* If we're the non-boot CPU, nothing set the stack canary up
|
* If we're the non-boot CPU, nothing set the stack canary up
|
||||||
|
@ -317,71 +311,40 @@ void cpu_idle(void)
|
||||||
* canaries already on the stack wont ever trigger).
|
* canaries already on the stack wont ever trigger).
|
||||||
*/
|
*/
|
||||||
boot_init_stack_canary();
|
boot_init_stack_canary();
|
||||||
current_thread_info()->status |= TS_POLLING;
|
}
|
||||||
|
|
||||||
while (1) {
|
void arch_cpu_idle_enter(void)
|
||||||
tick_nohz_idle_enter();
|
{
|
||||||
|
local_touch_nmi();
|
||||||
|
enter_idle();
|
||||||
|
}
|
||||||
|
|
||||||
while (!need_resched()) {
|
void arch_cpu_idle_exit(void)
|
||||||
rmb();
|
{
|
||||||
|
__exit_idle();
|
||||||
|
}
|
||||||
|
|
||||||
if (cpu_is_offline(smp_processor_id()))
|
void arch_cpu_idle_dead(void)
|
||||||
play_dead();
|
{
|
||||||
|
play_dead();
|
||||||
/*
|
|
||||||
* Idle routines should keep interrupts disabled
|
|
||||||
* from here on, until they go to idle.
|
|
||||||
* Otherwise, idle callbacks can misfire.
|
|
||||||
*/
|
|
||||||
local_touch_nmi();
|
|
||||||
local_irq_disable();
|
|
||||||
|
|
||||||
enter_idle();
|
|
||||||
|
|
||||||
/* Don't trace irqs off for idle */
|
|
||||||
stop_critical_timings();
|
|
||||||
|
|
||||||
/* enter_idle() needs rcu for notifiers */
|
|
||||||
rcu_idle_enter();
|
|
||||||
|
|
||||||
if (cpuidle_idle_call())
|
|
||||||
x86_idle();
|
|
||||||
|
|
||||||
rcu_idle_exit();
|
|
||||||
start_critical_timings();
|
|
||||||
|
|
||||||
/* In many cases the interrupt that ended idle
|
|
||||||
has already called exit_idle. But some idle
|
|
||||||
loops can be woken up without interrupt. */
|
|
||||||
__exit_idle();
|
|
||||||
}
|
|
||||||
|
|
||||||
tick_nohz_idle_exit();
|
|
||||||
preempt_enable_no_resched();
|
|
||||||
schedule();
|
|
||||||
preempt_disable();
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* We use this if we don't have any better
|
* Called from the generic idle code.
|
||||||
* idle routine..
|
*/
|
||||||
|
void arch_cpu_idle(void)
|
||||||
|
{
|
||||||
|
if (cpuidle_idle_call())
|
||||||
|
x86_idle();
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* We use this if we don't have any better idle routine..
|
||||||
*/
|
*/
|
||||||
void default_idle(void)
|
void default_idle(void)
|
||||||
{
|
{
|
||||||
trace_cpu_idle_rcuidle(1, smp_processor_id());
|
trace_cpu_idle_rcuidle(1, smp_processor_id());
|
||||||
current_thread_info()->status &= ~TS_POLLING;
|
safe_halt();
|
||||||
/*
|
|
||||||
* TS_POLLING-cleared state must be visible before we
|
|
||||||
* test NEED_RESCHED:
|
|
||||||
*/
|
|
||||||
smp_mb();
|
|
||||||
|
|
||||||
if (!need_resched())
|
|
||||||
safe_halt(); /* enables interrupts racelessly */
|
|
||||||
else
|
|
||||||
local_irq_enable();
|
|
||||||
current_thread_info()->status |= TS_POLLING;
|
|
||||||
trace_cpu_idle_rcuidle(PWR_EVENT_EXIT, smp_processor_id());
|
trace_cpu_idle_rcuidle(PWR_EVENT_EXIT, smp_processor_id());
|
||||||
}
|
}
|
||||||
#ifdef CONFIG_APM_MODULE
|
#ifdef CONFIG_APM_MODULE
|
||||||
|
@ -411,20 +374,6 @@ void stop_this_cpu(void *dummy)
|
||||||
halt();
|
halt();
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
|
||||||
* On SMP it's slightly faster (but much more power-consuming!)
|
|
||||||
* to poll the ->work.need_resched flag instead of waiting for the
|
|
||||||
* cross-CPU IPI to arrive. Use this option with caution.
|
|
||||||
*/
|
|
||||||
static void poll_idle(void)
|
|
||||||
{
|
|
||||||
trace_cpu_idle_rcuidle(0, smp_processor_id());
|
|
||||||
local_irq_enable();
|
|
||||||
while (!need_resched())
|
|
||||||
cpu_relax();
|
|
||||||
trace_cpu_idle_rcuidle(PWR_EVENT_EXIT, smp_processor_id());
|
|
||||||
}
|
|
||||||
|
|
||||||
bool amd_e400_c1e_detected;
|
bool amd_e400_c1e_detected;
|
||||||
EXPORT_SYMBOL(amd_e400_c1e_detected);
|
EXPORT_SYMBOL(amd_e400_c1e_detected);
|
||||||
|
|
||||||
|
@ -489,10 +438,10 @@ static void amd_e400_idle(void)
|
||||||
void __cpuinit select_idle_routine(const struct cpuinfo_x86 *c)
|
void __cpuinit select_idle_routine(const struct cpuinfo_x86 *c)
|
||||||
{
|
{
|
||||||
#ifdef CONFIG_SMP
|
#ifdef CONFIG_SMP
|
||||||
if (x86_idle == poll_idle && smp_num_siblings > 1)
|
if (boot_option_idle_override == IDLE_POLL && smp_num_siblings > 1)
|
||||||
pr_warn_once("WARNING: polling idle and HT enabled, performance may degrade\n");
|
pr_warn_once("WARNING: polling idle and HT enabled, performance may degrade\n");
|
||||||
#endif
|
#endif
|
||||||
if (x86_idle)
|
if (x86_idle || boot_option_idle_override == IDLE_POLL)
|
||||||
return;
|
return;
|
||||||
|
|
||||||
if (cpu_has_amd_erratum(amd_erratum_400)) {
|
if (cpu_has_amd_erratum(amd_erratum_400)) {
|
||||||
|
@ -517,8 +466,8 @@ static int __init idle_setup(char *str)
|
||||||
|
|
||||||
if (!strcmp(str, "poll")) {
|
if (!strcmp(str, "poll")) {
|
||||||
pr_info("using polling idle threads\n");
|
pr_info("using polling idle threads\n");
|
||||||
x86_idle = poll_idle;
|
|
||||||
boot_option_idle_override = IDLE_POLL;
|
boot_option_idle_override = IDLE_POLL;
|
||||||
|
cpu_idle_poll_ctrl(true);
|
||||||
} else if (!strcmp(str, "halt")) {
|
} else if (!strcmp(str, "halt")) {
|
||||||
/*
|
/*
|
||||||
* When the boot option of idle=halt is added, halt is
|
* When the boot option of idle=halt is added, halt is
|
||||||
|
|
|
@ -284,7 +284,7 @@ notrace static void __cpuinit start_secondary(void *unused)
|
||||||
x86_cpuinit.setup_percpu_clockev();
|
x86_cpuinit.setup_percpu_clockev();
|
||||||
|
|
||||||
wmb();
|
wmb();
|
||||||
cpu_idle();
|
cpu_startup_entry(CPUHP_ONLINE);
|
||||||
}
|
}
|
||||||
|
|
||||||
void __init smp_store_boot_cpu_info(void)
|
void __init smp_store_boot_cpu_info(void)
|
||||||
|
|
|
@ -95,7 +95,7 @@ static void __cpuinit cpu_bringup(void)
|
||||||
static void __cpuinit cpu_bringup_and_idle(void)
|
static void __cpuinit cpu_bringup_and_idle(void)
|
||||||
{
|
{
|
||||||
cpu_bringup();
|
cpu_bringup();
|
||||||
cpu_idle();
|
cpu_startup_entry(CPUHP_ONLINE);
|
||||||
}
|
}
|
||||||
|
|
||||||
static int xen_smp_intr_init(unsigned int cpu)
|
static int xen_smp_intr_init(unsigned int cpu)
|
||||||
|
|
Loading…
Reference in New Issue