Index: sys/x86/x86/cpu_machdep.c =================================================================== --- sys/x86/x86/cpu_machdep.c +++ sys/x86/x86/cpu_machdep.c @@ -52,6 +52,7 @@ #include "opt_maxmem.h" #include "opt_mp_watchdog.h" #include "opt_platform.h" +#include "opt_sched.h" #ifdef __i386__ #include "opt_apic.h" #endif @@ -532,32 +533,24 @@ SYSCTL_INT(_machdep, OID_AUTO, idle_mwait, CTLFLAG_RWTUN, &idle_mwait, 0, "Use MONITOR/MWAIT for short idle"); -static void -cpu_idle_acpi(sbintime_t sbt) +static bool +cpu_idle_enter(int *statep, int newstate) { - int *state; + KASSERT(atomic_load_int(statep) == STATE_RUNNING, + ("%s: state %d", __func__, atomic_load_int(statep))); - state = &PCPU_PTR(monitorbuf)->idle_state; - atomic_store_int(state, STATE_SLEEPING); - - /* See comments in cpu_idle_hlt(). */ - disable_intr(); - if (sched_runnable()) - enable_intr(); - else if (cpu_idle_hook) - cpu_idle_hook(sbt); - else - acpi_cpu_c1(); - atomic_store_int(state, STATE_RUNNING); -} - -static void -cpu_idle_hlt(sbintime_t sbt) -{ - int *state; - - state = &PCPU_PTR(monitorbuf)->idle_state; - atomic_store_int(state, STATE_SLEEPING); + /* + * A fence is needed to prevent reordering of the load in + * sched_runnable() with this store to the idle state word. Without it, + * cpu_idle_wakeup() can observe the state as STATE_RUNNING after having + * added load to the queue, and elide an IPI. Then, sched_runnable() + * can observe tdq_load == 0, so the CPU ends up idling with pending + * work. + */ + atomic_store_int(statep, newstate); +#if defined(SCHED_ULE) && defined(SMP) + atomic_thread_fence_seq_cst(); +#endif /* * Since we may be in a critical section from cpu_idle(), if @@ -576,35 +569,62 @@ * interrupt. */ disable_intr(); - if (sched_runnable()) + if (sched_runnable()) { enable_intr(); - else - acpi_cpu_c1(); - atomic_store_int(state, STATE_RUNNING); + atomic_store_int(statep, STATE_RUNNING); + return (false); + } else { + return (true); + } } static void -cpu_idle_mwait(sbintime_t sbt) +cpu_idle_exit(int *statep) +{ + atomic_store_int(statep, STATE_RUNNING); +} + +static void +cpu_idle_acpi(sbintime_t sbt) { int *state; state = &PCPU_PTR(monitorbuf)->idle_state; - atomic_store_int(state, STATE_MWAIT); + if (cpu_idle_enter(state, STATE_SLEEPING)) { + if (cpu_idle_hook) + cpu_idle_hook(sbt); + else + acpi_cpu_c1(); + cpu_idle_exit(state); + } +} - /* See comments in cpu_idle_hlt(). */ - disable_intr(); - if (sched_runnable()) { +static void +cpu_idle_hlt(sbintime_t sbt) +{ + int *state; + + state = &PCPU_PTR(monitorbuf)->idle_state; + if (cpu_idle_enter(state, STATE_SLEEPING)) { + acpi_cpu_c1(); atomic_store_int(state, STATE_RUNNING); - enable_intr(); - return; } +} - cpu_monitor(state, 0, 0); - if (atomic_load_int(state) == STATE_MWAIT) - __asm __volatile("sti; mwait" : : "a" (MWAIT_C1), "c" (0)); - else - enable_intr(); - atomic_store_int(state, STATE_RUNNING); +static void +cpu_idle_mwait(sbintime_t sbt) +{ + int *state; + + state = &PCPU_PTR(monitorbuf)->idle_state; + if (cpu_idle_enter(state, STATE_MWAIT)) { + cpu_monitor(state, 0, 0); + if (atomic_load_int(state) == STATE_MWAIT) + __asm __volatile("sti; mwait" : : "a" (MWAIT_C1), "c" (0)); + else + enable_intr(); + cpu_idle_exit(state); + } } static void