sched/clock: Fix hotplug crash
Mike reported that he could trigger the WARN_ON_ONCE() in
set_sched_clock_stable() using hotplug.
This exposed a fundamental problem with the interface, we should never
mark the TSC stable if we ever find it to be unstable. Therefore
set_sched_clock_stable() is a broken interface.
The reason it existed is that not having it is a pain, it means all
relevant architecture code needs to call clear_sched_clock_stable()
where appropriate.
Of the three architectures that select HAVE_UNSTABLE_SCHED_CLOCK ia64
and parisc are trivial in that they never called
set_sched_clock_stable(), so add an unconditional call to
clear_sched_clock_stable() to them.
For x86 the story is a lot more involved, and what this patch tries to
do is ensure we preserve the status quo. So even is Cyrix or Transmeta
have usable TSC they never called set_sched_clock_stable() so they now
get an explicit mark unstable.
Reported-by: Mike Galbraith <efault@gmx.de>
Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Cc: Borislav Petkov <bp@alien8.de>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Fixes: 9881b024b7
("sched/clock: Delay switching sched_clock to stable")
Link: http://lkml.kernel.org/r/20170119133633.GB6536@twins.programming.kicks-ass.net
Signed-off-by: Ingo Molnar <mingo@kernel.org>
This commit is contained in:
Родитель
02cfdc95a0
Коммит
acb04058de
|
@ -619,6 +619,8 @@ setup_arch (char **cmdline_p)
|
||||||
check_sal_cache_flush();
|
check_sal_cache_flush();
|
||||||
#endif
|
#endif
|
||||||
paging_init();
|
paging_init();
|
||||||
|
|
||||||
|
clear_sched_clock_stable();
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
|
|
@ -36,6 +36,7 @@
|
||||||
#undef PCI_DEBUG
|
#undef PCI_DEBUG
|
||||||
#include <linux/proc_fs.h>
|
#include <linux/proc_fs.h>
|
||||||
#include <linux/export.h>
|
#include <linux/export.h>
|
||||||
|
#include <linux/sched.h>
|
||||||
|
|
||||||
#include <asm/processor.h>
|
#include <asm/processor.h>
|
||||||
#include <asm/sections.h>
|
#include <asm/sections.h>
|
||||||
|
@ -176,6 +177,7 @@ void __init setup_arch(char **cmdline_p)
|
||||||
conswitchp = &dummy_con; /* we use do_take_over_console() later ! */
|
conswitchp = &dummy_con; /* we use do_take_over_console() later ! */
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
|
clear_sched_clock_stable();
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
|
|
@ -548,8 +548,10 @@ static void early_init_amd(struct cpuinfo_x86 *c)
|
||||||
if (c->x86_power & (1 << 8)) {
|
if (c->x86_power & (1 << 8)) {
|
||||||
set_cpu_cap(c, X86_FEATURE_CONSTANT_TSC);
|
set_cpu_cap(c, X86_FEATURE_CONSTANT_TSC);
|
||||||
set_cpu_cap(c, X86_FEATURE_NONSTOP_TSC);
|
set_cpu_cap(c, X86_FEATURE_NONSTOP_TSC);
|
||||||
if (!check_tsc_unstable())
|
if (check_tsc_unstable())
|
||||||
set_sched_clock_stable();
|
clear_sched_clock_stable();
|
||||||
|
} else {
|
||||||
|
clear_sched_clock_stable();
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Bit 12 of 8000_0007 edx is accumulated power mechanism. */
|
/* Bit 12 of 8000_0007 edx is accumulated power mechanism. */
|
||||||
|
|
|
@ -1,5 +1,5 @@
|
||||||
#include <linux/bitops.h>
|
|
||||||
#include <linux/kernel.h>
|
#include <linux/sched.h>
|
||||||
|
|
||||||
#include <asm/cpufeature.h>
|
#include <asm/cpufeature.h>
|
||||||
#include <asm/e820.h>
|
#include <asm/e820.h>
|
||||||
|
@ -104,6 +104,8 @@ static void early_init_centaur(struct cpuinfo_x86 *c)
|
||||||
#ifdef CONFIG_X86_64
|
#ifdef CONFIG_X86_64
|
||||||
set_cpu_cap(c, X86_FEATURE_SYSENTER32);
|
set_cpu_cap(c, X86_FEATURE_SYSENTER32);
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
|
clear_sched_clock_stable();
|
||||||
}
|
}
|
||||||
|
|
||||||
static void init_centaur(struct cpuinfo_x86 *c)
|
static void init_centaur(struct cpuinfo_x86 *c)
|
||||||
|
|
|
@ -83,6 +83,7 @@ static void default_init(struct cpuinfo_x86 *c)
|
||||||
strcpy(c->x86_model_id, "386");
|
strcpy(c->x86_model_id, "386");
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
clear_sched_clock_stable();
|
||||||
}
|
}
|
||||||
|
|
||||||
static const struct cpu_dev default_cpu = {
|
static const struct cpu_dev default_cpu = {
|
||||||
|
@ -1055,6 +1056,8 @@ static void identify_cpu(struct cpuinfo_x86 *c)
|
||||||
*/
|
*/
|
||||||
if (this_cpu->c_init)
|
if (this_cpu->c_init)
|
||||||
this_cpu->c_init(c);
|
this_cpu->c_init(c);
|
||||||
|
else
|
||||||
|
clear_sched_clock_stable();
|
||||||
|
|
||||||
/* Disable the PN if appropriate */
|
/* Disable the PN if appropriate */
|
||||||
squash_the_stupid_serial_number(c);
|
squash_the_stupid_serial_number(c);
|
||||||
|
|
|
@ -9,6 +9,7 @@
|
||||||
#include <asm/pci-direct.h>
|
#include <asm/pci-direct.h>
|
||||||
#include <asm/tsc.h>
|
#include <asm/tsc.h>
|
||||||
#include <asm/cpufeature.h>
|
#include <asm/cpufeature.h>
|
||||||
|
#include <linux/sched.h>
|
||||||
|
|
||||||
#include "cpu.h"
|
#include "cpu.h"
|
||||||
|
|
||||||
|
@ -183,6 +184,7 @@ static void early_init_cyrix(struct cpuinfo_x86 *c)
|
||||||
set_cpu_cap(c, X86_FEATURE_CYRIX_ARR);
|
set_cpu_cap(c, X86_FEATURE_CYRIX_ARR);
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
|
clear_sched_clock_stable();
|
||||||
}
|
}
|
||||||
|
|
||||||
static void init_cyrix(struct cpuinfo_x86 *c)
|
static void init_cyrix(struct cpuinfo_x86 *c)
|
||||||
|
|
|
@ -124,8 +124,10 @@ static void early_init_intel(struct cpuinfo_x86 *c)
|
||||||
if (c->x86_power & (1 << 8)) {
|
if (c->x86_power & (1 << 8)) {
|
||||||
set_cpu_cap(c, X86_FEATURE_CONSTANT_TSC);
|
set_cpu_cap(c, X86_FEATURE_CONSTANT_TSC);
|
||||||
set_cpu_cap(c, X86_FEATURE_NONSTOP_TSC);
|
set_cpu_cap(c, X86_FEATURE_NONSTOP_TSC);
|
||||||
if (!check_tsc_unstable())
|
if (check_tsc_unstable())
|
||||||
set_sched_clock_stable();
|
clear_sched_clock_stable();
|
||||||
|
} else {
|
||||||
|
clear_sched_clock_stable();
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Penwell and Cloverview have the TSC which doesn't sleep on S3 */
|
/* Penwell and Cloverview have the TSC which doesn't sleep on S3 */
|
||||||
|
|
|
@ -1,4 +1,5 @@
|
||||||
#include <linux/kernel.h>
|
#include <linux/kernel.h>
|
||||||
|
#include <linux/sched.h>
|
||||||
#include <linux/mm.h>
|
#include <linux/mm.h>
|
||||||
#include <asm/cpufeature.h>
|
#include <asm/cpufeature.h>
|
||||||
#include <asm/msr.h>
|
#include <asm/msr.h>
|
||||||
|
@ -14,6 +15,8 @@ static void early_init_transmeta(struct cpuinfo_x86 *c)
|
||||||
if (xlvl >= 0x80860001)
|
if (xlvl >= 0x80860001)
|
||||||
c->x86_capability[CPUID_8086_0001_EDX] = cpuid_edx(0x80860001);
|
c->x86_capability[CPUID_8086_0001_EDX] = cpuid_edx(0x80860001);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
clear_sched_clock_stable();
|
||||||
}
|
}
|
||||||
|
|
||||||
static void init_transmeta(struct cpuinfo_x86 *c)
|
static void init_transmeta(struct cpuinfo_x86 *c)
|
||||||
|
|
|
@ -107,12 +107,12 @@ static inline void kvm_sched_clock_init(bool stable)
|
||||||
{
|
{
|
||||||
if (!stable) {
|
if (!stable) {
|
||||||
pv_time_ops.sched_clock = kvm_clock_read;
|
pv_time_ops.sched_clock = kvm_clock_read;
|
||||||
|
clear_sched_clock_stable();
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
kvm_sched_clock_offset = kvm_clock_read();
|
kvm_sched_clock_offset = kvm_clock_read();
|
||||||
pv_time_ops.sched_clock = kvm_sched_clock_read;
|
pv_time_ops.sched_clock = kvm_sched_clock_read;
|
||||||
set_sched_clock_stable();
|
|
||||||
|
|
||||||
printk(KERN_INFO "kvm-clock: using sched offset of %llu cycles\n",
|
printk(KERN_INFO "kvm-clock: using sched offset of %llu cycles\n",
|
||||||
kvm_sched_clock_offset);
|
kvm_sched_clock_offset);
|
||||||
|
|
|
@ -2547,7 +2547,6 @@ extern void sched_clock_init_late(void);
|
||||||
* is reliable after all:
|
* is reliable after all:
|
||||||
*/
|
*/
|
||||||
extern int sched_clock_stable(void);
|
extern int sched_clock_stable(void);
|
||||||
extern void set_sched_clock_stable(void);
|
|
||||||
extern void clear_sched_clock_stable(void);
|
extern void clear_sched_clock_stable(void);
|
||||||
|
|
||||||
extern void sched_clock_tick(void);
|
extern void sched_clock_tick(void);
|
||||||
|
|
|
@ -83,8 +83,15 @@ void sched_clock_init(void)
|
||||||
}
|
}
|
||||||
|
|
||||||
#ifdef CONFIG_HAVE_UNSTABLE_SCHED_CLOCK
|
#ifdef CONFIG_HAVE_UNSTABLE_SCHED_CLOCK
|
||||||
|
/*
|
||||||
|
* We must start with !__sched_clock_stable because the unstable -> stable
|
||||||
|
* transition is accurate, while the stable -> unstable transition is not.
|
||||||
|
*
|
||||||
|
* Similarly we start with __sched_clock_stable_early, thereby assuming we
|
||||||
|
* will become stable, such that there's only a single 1 -> 0 transition.
|
||||||
|
*/
|
||||||
static DEFINE_STATIC_KEY_FALSE(__sched_clock_stable);
|
static DEFINE_STATIC_KEY_FALSE(__sched_clock_stable);
|
||||||
static int __sched_clock_stable_early;
|
static int __sched_clock_stable_early = 1;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* We want: ktime_get_ns() + gtod_offset == sched_clock() + raw_offset
|
* We want: ktime_get_ns() + gtod_offset == sched_clock() + raw_offset
|
||||||
|
@ -132,24 +139,6 @@ static void __set_sched_clock_stable(void)
|
||||||
tick_dep_clear(TICK_DEP_BIT_CLOCK_UNSTABLE);
|
tick_dep_clear(TICK_DEP_BIT_CLOCK_UNSTABLE);
|
||||||
}
|
}
|
||||||
|
|
||||||
void set_sched_clock_stable(void)
|
|
||||||
{
|
|
||||||
__sched_clock_stable_early = 1;
|
|
||||||
|
|
||||||
smp_mb(); /* matches sched_clock_init_late() */
|
|
||||||
|
|
||||||
/*
|
|
||||||
* This really should only be called early (before
|
|
||||||
* sched_clock_init_late()) when guestimating our sched_clock() is
|
|
||||||
* solid.
|
|
||||||
*
|
|
||||||
* After that we test stability and we can negate our guess using
|
|
||||||
* clear_sched_clock_stable, possibly from a watchdog.
|
|
||||||
*/
|
|
||||||
if (WARN_ON_ONCE(sched_clock_running == 2))
|
|
||||||
__set_sched_clock_stable();
|
|
||||||
}
|
|
||||||
|
|
||||||
static void __clear_sched_clock_stable(struct work_struct *work)
|
static void __clear_sched_clock_stable(struct work_struct *work)
|
||||||
{
|
{
|
||||||
struct sched_clock_data *scd = this_scd();
|
struct sched_clock_data *scd = this_scd();
|
||||||
|
@ -199,8 +188,6 @@ void sched_clock_init_late(void)
|
||||||
|
|
||||||
if (__sched_clock_stable_early)
|
if (__sched_clock_stable_early)
|
||||||
__set_sched_clock_stable();
|
__set_sched_clock_stable();
|
||||||
else
|
|
||||||
__clear_sched_clock_stable(NULL);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
|
Загрузка…
Ссылка в новой задаче