x86/cpu: Clean up various files a bit
No code changes except printk levels (although some of the K6 mtrr code might be clearer if there were a few as would splitting out some of the intel cache code). Signed-off-by: Alan Cox <alan@linux.intel.com> LKML-Reference: <new-submission> Signed-off-by: Ingo Molnar <mingo@elte.hu>
This commit is contained in:
parent
e90476d3ba
commit
8bdbd962ec
|
@ -2,7 +2,7 @@
|
||||||
#include <linux/bitops.h>
|
#include <linux/bitops.h>
|
||||||
#include <linux/mm.h>
|
#include <linux/mm.h>
|
||||||
|
|
||||||
#include <asm/io.h>
|
#include <linux/io.h>
|
||||||
#include <asm/processor.h>
|
#include <asm/processor.h>
|
||||||
#include <asm/apic.h>
|
#include <asm/apic.h>
|
||||||
#include <asm/cpu.h>
|
#include <asm/cpu.h>
|
||||||
|
@ -87,9 +87,10 @@ static void __cpuinit init_amd_k6(struct cpuinfo_x86 *c)
|
||||||
d = d2-d;
|
d = d2-d;
|
||||||
|
|
||||||
if (d > 20*K6_BUG_LOOP)
|
if (d > 20*K6_BUG_LOOP)
|
||||||
printk("system stability may be impaired when more than 32 MB are used.\n");
|
printk(KERN_CONT
|
||||||
|
"system stability may be impaired when more than 32 MB are used.\n");
|
||||||
else
|
else
|
||||||
printk("probably OK (after B9730xxxx).\n");
|
printk(KERN_CONT "probably OK (after B9730xxxx).\n");
|
||||||
printk(KERN_INFO "Please see http://membres.lycos.fr/poulot/k6bug.html\n");
|
printk(KERN_INFO "Please see http://membres.lycos.fr/poulot/k6bug.html\n");
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -219,8 +220,9 @@ static void __cpuinit init_amd_k7(struct cpuinfo_x86 *c)
|
||||||
if ((c->x86_model == 8 && c->x86_mask >= 1) || (c->x86_model > 8)) {
|
if ((c->x86_model == 8 && c->x86_mask >= 1) || (c->x86_model > 8)) {
|
||||||
rdmsr(MSR_K7_CLK_CTL, l, h);
|
rdmsr(MSR_K7_CLK_CTL, l, h);
|
||||||
if ((l & 0xfff00000) != 0x20000000) {
|
if ((l & 0xfff00000) != 0x20000000) {
|
||||||
printk ("CPU: CLK_CTL MSR was %x. Reprogramming to %x\n", l,
|
printk(KERN_INFO
|
||||||
((l & 0x000fffff)|0x20000000));
|
"CPU: CLK_CTL MSR was %x. Reprogramming to %x\n",
|
||||||
|
l, ((l & 0x000fffff)|0x20000000));
|
||||||
wrmsr(MSR_K7_CLK_CTL, (l & 0x000fffff)|0x20000000, h);
|
wrmsr(MSR_K7_CLK_CTL, (l & 0x000fffff)|0x20000000, h);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -500,14 +502,17 @@ static void __cpuinit init_amd(struct cpuinfo_x86 *c)
|
||||||
}
|
}
|
||||||
|
|
||||||
#ifdef CONFIG_X86_32
|
#ifdef CONFIG_X86_32
|
||||||
static unsigned int __cpuinit amd_size_cache(struct cpuinfo_x86 *c, unsigned int size)
|
static unsigned int __cpuinit amd_size_cache(struct cpuinfo_x86 *c,
|
||||||
|
unsigned int size)
|
||||||
{
|
{
|
||||||
/* AMD errata T13 (order #21922) */
|
/* AMD errata T13 (order #21922) */
|
||||||
if ((c->x86 == 6)) {
|
if ((c->x86 == 6)) {
|
||||||
if (c->x86_model == 3 && c->x86_mask == 0) /* Duron Rev A0 */
|
/* Duron Rev A0 */
|
||||||
|
if (c->x86_model == 3 && c->x86_mask == 0)
|
||||||
size = 64;
|
size = 64;
|
||||||
|
/* Tbird rev A1/A2 */
|
||||||
if (c->x86_model == 4 &&
|
if (c->x86_model == 4 &&
|
||||||
(c->x86_mask == 0 || c->x86_mask == 1)) /* Tbird rev A1/A2 */
|
(c->x86_mask == 0 || c->x86_mask == 1))
|
||||||
size = 256;
|
size = 256;
|
||||||
}
|
}
|
||||||
return size;
|
return size;
|
||||||
|
|
|
@ -81,7 +81,7 @@ static void __init check_fpu(void)
|
||||||
|
|
||||||
boot_cpu_data.fdiv_bug = fdiv_bug;
|
boot_cpu_data.fdiv_bug = fdiv_bug;
|
||||||
if (boot_cpu_data.fdiv_bug)
|
if (boot_cpu_data.fdiv_bug)
|
||||||
printk("Hmm, FPU with FDIV bug.\n");
|
printk(KERN_WARNING "Hmm, FPU with FDIV bug.\n");
|
||||||
}
|
}
|
||||||
|
|
||||||
static void __init check_hlt(void)
|
static void __init check_hlt(void)
|
||||||
|
@ -98,7 +98,7 @@ static void __init check_hlt(void)
|
||||||
halt();
|
halt();
|
||||||
halt();
|
halt();
|
||||||
halt();
|
halt();
|
||||||
printk("OK.\n");
|
printk(KERN_CONT "OK.\n");
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@ -122,9 +122,9 @@ static void __init check_popad(void)
|
||||||
* CPU hard. Too bad.
|
* CPU hard. Too bad.
|
||||||
*/
|
*/
|
||||||
if (res != 12345678)
|
if (res != 12345678)
|
||||||
printk("Buggy.\n");
|
printk(KERN_CONT "Buggy.\n");
|
||||||
else
|
else
|
||||||
printk("OK.\n");
|
printk(KERN_CONT "OK.\n");
|
||||||
#endif
|
#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -156,7 +156,7 @@ void __init check_bugs(void)
|
||||||
{
|
{
|
||||||
identify_boot_cpu();
|
identify_boot_cpu();
|
||||||
#ifndef CONFIG_SMP
|
#ifndef CONFIG_SMP
|
||||||
printk("CPU: ");
|
printk(KERN_INFO "CPU: ");
|
||||||
print_cpu_info(&boot_cpu_data);
|
print_cpu_info(&boot_cpu_data);
|
||||||
#endif
|
#endif
|
||||||
check_config();
|
check_config();
|
||||||
|
|
|
@ -15,7 +15,7 @@ void __init check_bugs(void)
|
||||||
{
|
{
|
||||||
identify_boot_cpu();
|
identify_boot_cpu();
|
||||||
#if !defined(CONFIG_SMP)
|
#if !defined(CONFIG_SMP)
|
||||||
printk("CPU: ");
|
printk(KERN_INFO "CPU: ");
|
||||||
print_cpu_info(&boot_cpu_data);
|
print_cpu_info(&boot_cpu_data);
|
||||||
#endif
|
#endif
|
||||||
alternative_instructions();
|
alternative_instructions();
|
||||||
|
|
|
@ -18,8 +18,8 @@
|
||||||
#include <asm/hypervisor.h>
|
#include <asm/hypervisor.h>
|
||||||
#include <asm/processor.h>
|
#include <asm/processor.h>
|
||||||
#include <asm/sections.h>
|
#include <asm/sections.h>
|
||||||
#include <asm/topology.h>
|
#include <linux/topology.h>
|
||||||
#include <asm/cpumask.h>
|
#include <linux/cpumask.h>
|
||||||
#include <asm/pgtable.h>
|
#include <asm/pgtable.h>
|
||||||
#include <asm/atomic.h>
|
#include <asm/atomic.h>
|
||||||
#include <asm/proto.h>
|
#include <asm/proto.h>
|
||||||
|
@ -28,13 +28,13 @@
|
||||||
#include <asm/desc.h>
|
#include <asm/desc.h>
|
||||||
#include <asm/i387.h>
|
#include <asm/i387.h>
|
||||||
#include <asm/mtrr.h>
|
#include <asm/mtrr.h>
|
||||||
#include <asm/numa.h>
|
#include <linux/numa.h>
|
||||||
#include <asm/asm.h>
|
#include <asm/asm.h>
|
||||||
#include <asm/cpu.h>
|
#include <asm/cpu.h>
|
||||||
#include <asm/mce.h>
|
#include <asm/mce.h>
|
||||||
#include <asm/msr.h>
|
#include <asm/msr.h>
|
||||||
#include <asm/pat.h>
|
#include <asm/pat.h>
|
||||||
#include <asm/smp.h>
|
#include <linux/smp.h>
|
||||||
|
|
||||||
#ifdef CONFIG_X86_LOCAL_APIC
|
#ifdef CONFIG_X86_LOCAL_APIC
|
||||||
#include <asm/uv/uv.h>
|
#include <asm/uv/uv.h>
|
||||||
|
|
|
@ -3,10 +3,10 @@
|
||||||
#include <linux/delay.h>
|
#include <linux/delay.h>
|
||||||
#include <linux/pci.h>
|
#include <linux/pci.h>
|
||||||
#include <asm/dma.h>
|
#include <asm/dma.h>
|
||||||
#include <asm/io.h>
|
#include <linux/io.h>
|
||||||
#include <asm/processor-cyrix.h>
|
#include <asm/processor-cyrix.h>
|
||||||
#include <asm/processor-flags.h>
|
#include <asm/processor-flags.h>
|
||||||
#include <asm/timer.h>
|
#include <linux/timer.h>
|
||||||
#include <asm/pci-direct.h>
|
#include <asm/pci-direct.h>
|
||||||
#include <asm/tsc.h>
|
#include <asm/tsc.h>
|
||||||
|
|
||||||
|
@ -282,7 +282,8 @@ static void __cpuinit init_cyrix(struct cpuinfo_x86 *c)
|
||||||
* The 5510/5520 companion chips have a funky PIT.
|
* The 5510/5520 companion chips have a funky PIT.
|
||||||
*/
|
*/
|
||||||
if (vendor == PCI_VENDOR_ID_CYRIX &&
|
if (vendor == PCI_VENDOR_ID_CYRIX &&
|
||||||
(device == PCI_DEVICE_ID_CYRIX_5510 || device == PCI_DEVICE_ID_CYRIX_5520))
|
(device == PCI_DEVICE_ID_CYRIX_5510 ||
|
||||||
|
device == PCI_DEVICE_ID_CYRIX_5520))
|
||||||
mark_tsc_unstable("cyrix 5510/5520 detected");
|
mark_tsc_unstable("cyrix 5510/5520 detected");
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
@ -299,7 +300,8 @@ static void __cpuinit init_cyrix(struct cpuinfo_x86 *c)
|
||||||
* ? : 0x7x
|
* ? : 0x7x
|
||||||
* GX1 : 0x8x GX1 datasheet 56
|
* GX1 : 0x8x GX1 datasheet 56
|
||||||
*/
|
*/
|
||||||
if ((0x30 <= dir1 && dir1 <= 0x6f) || (0x80 <= dir1 && dir1 <= 0x8f))
|
if ((0x30 <= dir1 && dir1 <= 0x6f) ||
|
||||||
|
(0x80 <= dir1 && dir1 <= 0x8f))
|
||||||
geode_configure();
|
geode_configure();
|
||||||
return;
|
return;
|
||||||
} else { /* MediaGX */
|
} else { /* MediaGX */
|
||||||
|
@ -427,9 +429,12 @@ static void __cpuinit cyrix_identify(struct cpuinfo_x86 *c)
|
||||||
printk(KERN_INFO "Enabling CPUID on Cyrix processor.\n");
|
printk(KERN_INFO "Enabling CPUID on Cyrix processor.\n");
|
||||||
local_irq_save(flags);
|
local_irq_save(flags);
|
||||||
ccr3 = getCx86(CX86_CCR3);
|
ccr3 = getCx86(CX86_CCR3);
|
||||||
setCx86(CX86_CCR3, (ccr3 & 0x0f) | 0x10); /* enable MAPEN */
|
/* enable MAPEN */
|
||||||
setCx86_old(CX86_CCR4, getCx86_old(CX86_CCR4) | 0x80); /* enable cpuid */
|
setCx86(CX86_CCR3, (ccr3 & 0x0f) | 0x10);
|
||||||
setCx86(CX86_CCR3, ccr3); /* disable MAPEN */
|
/* enable cpuid */
|
||||||
|
setCx86_old(CX86_CCR4, getCx86_old(CX86_CCR4) | 0x80);
|
||||||
|
/* disable MAPEN */
|
||||||
|
setCx86(CX86_CCR3, ccr3);
|
||||||
local_irq_restore(flags);
|
local_irq_restore(flags);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -28,12 +28,11 @@
|
||||||
static inline void __cpuinit
|
static inline void __cpuinit
|
||||||
detect_hypervisor_vendor(struct cpuinfo_x86 *c)
|
detect_hypervisor_vendor(struct cpuinfo_x86 *c)
|
||||||
{
|
{
|
||||||
if (vmware_platform()) {
|
if (vmware_platform())
|
||||||
c->x86_hyper_vendor = X86_HYPER_VENDOR_VMWARE;
|
c->x86_hyper_vendor = X86_HYPER_VENDOR_VMWARE;
|
||||||
} else {
|
else
|
||||||
c->x86_hyper_vendor = X86_HYPER_VENDOR_NONE;
|
c->x86_hyper_vendor = X86_HYPER_VENDOR_NONE;
|
||||||
}
|
}
|
||||||
}
|
|
||||||
|
|
||||||
unsigned long get_hypervisor_tsc_freq(void)
|
unsigned long get_hypervisor_tsc_freq(void)
|
||||||
{
|
{
|
||||||
|
|
|
@ -7,17 +7,17 @@
|
||||||
#include <linux/sched.h>
|
#include <linux/sched.h>
|
||||||
#include <linux/thread_info.h>
|
#include <linux/thread_info.h>
|
||||||
#include <linux/module.h>
|
#include <linux/module.h>
|
||||||
|
#include <linux/uaccess.h>
|
||||||
|
|
||||||
#include <asm/processor.h>
|
#include <asm/processor.h>
|
||||||
#include <asm/pgtable.h>
|
#include <asm/pgtable.h>
|
||||||
#include <asm/msr.h>
|
#include <asm/msr.h>
|
||||||
#include <asm/uaccess.h>
|
|
||||||
#include <asm/ds.h>
|
#include <asm/ds.h>
|
||||||
#include <asm/bugs.h>
|
#include <asm/bugs.h>
|
||||||
#include <asm/cpu.h>
|
#include <asm/cpu.h>
|
||||||
|
|
||||||
#ifdef CONFIG_X86_64
|
#ifdef CONFIG_X86_64
|
||||||
#include <asm/topology.h>
|
#include <linux/topology.h>
|
||||||
#include <asm/numa_64.h>
|
#include <asm/numa_64.h>
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
|
@ -174,7 +174,8 @@ static void __cpuinit intel_workarounds(struct cpuinfo_x86 *c)
|
||||||
#ifdef CONFIG_X86_F00F_BUG
|
#ifdef CONFIG_X86_F00F_BUG
|
||||||
/*
|
/*
|
||||||
* All current models of Pentium and Pentium with MMX technology CPUs
|
* All current models of Pentium and Pentium with MMX technology CPUs
|
||||||
* have the F0 0F bug, which lets nonprivileged users lock up the system.
|
* have the F0 0F bug, which lets nonprivileged users lock up the
|
||||||
|
* system.
|
||||||
* Note that the workaround only should be initialized once...
|
* Note that the workaround only should be initialized once...
|
||||||
*/
|
*/
|
||||||
c->f00f_bug = 0;
|
c->f00f_bug = 0;
|
||||||
|
@ -283,7 +284,7 @@ static int __cpuinit intel_num_cpu_cores(struct cpuinfo_x86 *c)
|
||||||
/* Intel has a non-standard dependency on %ecx for this CPUID level. */
|
/* Intel has a non-standard dependency on %ecx for this CPUID level. */
|
||||||
cpuid_count(4, 0, &eax, &ebx, &ecx, &edx);
|
cpuid_count(4, 0, &eax, &ebx, &ecx, &edx);
|
||||||
if (eax & 0x1f)
|
if (eax & 0x1f)
|
||||||
return ((eax >> 26) + 1);
|
return (eax >> 26) + 1;
|
||||||
else
|
else
|
||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
|
|
|
@ -16,7 +16,7 @@
|
||||||
#include <linux/pci.h>
|
#include <linux/pci.h>
|
||||||
|
|
||||||
#include <asm/processor.h>
|
#include <asm/processor.h>
|
||||||
#include <asm/smp.h>
|
#include <linux/smp.h>
|
||||||
#include <asm/k8.h>
|
#include <asm/k8.h>
|
||||||
|
|
||||||
#define LVL_1_INST 1
|
#define LVL_1_INST 1
|
||||||
|
@ -25,14 +25,15 @@
|
||||||
#define LVL_3 4
|
#define LVL_3 4
|
||||||
#define LVL_TRACE 5
|
#define LVL_TRACE 5
|
||||||
|
|
||||||
struct _cache_table
|
struct _cache_table {
|
||||||
{
|
|
||||||
unsigned char descriptor;
|
unsigned char descriptor;
|
||||||
char cache_type;
|
char cache_type;
|
||||||
short size;
|
short size;
|
||||||
};
|
};
|
||||||
|
|
||||||
/* all the cache descriptor types we care about (no TLB or trace cache entries) */
|
/* All the cache descriptor types we care about (no TLB or
|
||||||
|
trace cache entries) */
|
||||||
|
|
||||||
static const struct _cache_table __cpuinitconst cache_table[] =
|
static const struct _cache_table __cpuinitconst cache_table[] =
|
||||||
{
|
{
|
||||||
{ 0x06, LVL_1_INST, 8 }, /* 4-way set assoc, 32 byte line size */
|
{ 0x06, LVL_1_INST, 8 }, /* 4-way set assoc, 32 byte line size */
|
||||||
|
@ -105,8 +106,7 @@ static const struct _cache_table __cpuinitconst cache_table[] =
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
||||||
enum _cache_type
|
enum _cache_type {
|
||||||
{
|
|
||||||
CACHE_TYPE_NULL = 0,
|
CACHE_TYPE_NULL = 0,
|
||||||
CACHE_TYPE_DATA = 1,
|
CACHE_TYPE_DATA = 1,
|
||||||
CACHE_TYPE_INST = 2,
|
CACHE_TYPE_INST = 2,
|
||||||
|
@ -350,7 +350,8 @@ static int __cpuinit find_num_cache_leaves(void)
|
||||||
|
|
||||||
unsigned int __cpuinit init_intel_cacheinfo(struct cpuinfo_x86 *c)
|
unsigned int __cpuinit init_intel_cacheinfo(struct cpuinfo_x86 *c)
|
||||||
{
|
{
|
||||||
unsigned int trace = 0, l1i = 0, l1d = 0, l2 = 0, l3 = 0; /* Cache sizes */
|
/* Cache sizes */
|
||||||
|
unsigned int trace = 0, l1i = 0, l1d = 0, l2 = 0, l3 = 0;
|
||||||
unsigned int new_l1d = 0, new_l1i = 0; /* Cache sizes from cpuid(4) */
|
unsigned int new_l1d = 0, new_l1i = 0; /* Cache sizes from cpuid(4) */
|
||||||
unsigned int new_l2 = 0, new_l3 = 0, i; /* Cache sizes from cpuid(4) */
|
unsigned int new_l2 = 0, new_l3 = 0, i; /* Cache sizes from cpuid(4) */
|
||||||
unsigned int l2_id = 0, l3_id = 0, num_threads_sharing, index_msb;
|
unsigned int l2_id = 0, l3_id = 0, num_threads_sharing, index_msb;
|
||||||
|
@ -395,7 +396,8 @@ unsigned int __cpuinit init_intel_cacheinfo(struct cpuinfo_x86 *c)
|
||||||
case 3:
|
case 3:
|
||||||
new_l3 = this_leaf.size/1024;
|
new_l3 = this_leaf.size/1024;
|
||||||
num_threads_sharing = 1 + this_leaf.eax.split.num_threads_sharing;
|
num_threads_sharing = 1 + this_leaf.eax.split.num_threads_sharing;
|
||||||
index_msb = get_count_order(num_threads_sharing);
|
index_msb = get_count_order(
|
||||||
|
num_threads_sharing);
|
||||||
l3_id = c->apicid >> index_msb;
|
l3_id = c->apicid >> index_msb;
|
||||||
break;
|
break;
|
||||||
default:
|
default:
|
||||||
|
@ -425,9 +427,9 @@ unsigned int __cpuinit init_intel_cacheinfo(struct cpuinfo_x86 *c)
|
||||||
cpuid(2, ®s[0], ®s[1], ®s[2], ®s[3]);
|
cpuid(2, ®s[0], ®s[1], ®s[2], ®s[3]);
|
||||||
|
|
||||||
/* If bit 31 is set, this is an unknown format */
|
/* If bit 31 is set, this is an unknown format */
|
||||||
for ( j = 0 ; j < 3 ; j++ ) {
|
for (j = 0 ; j < 3 ; j++)
|
||||||
if (regs[j] & (1 << 31)) regs[j] = 0;
|
if (regs[j] & (1 << 31))
|
||||||
}
|
regs[j] = 0;
|
||||||
|
|
||||||
/* Byte 0 is level count, not a descriptor */
|
/* Byte 0 is level count, not a descriptor */
|
||||||
for (j = 1 ; j < 16 ; j++) {
|
for (j = 1 ; j < 16 ; j++) {
|
||||||
|
@ -435,8 +437,7 @@ unsigned int __cpuinit init_intel_cacheinfo(struct cpuinfo_x86 *c)
|
||||||
unsigned char k = 0;
|
unsigned char k = 0;
|
||||||
|
|
||||||
/* look up this descriptor in the table */
|
/* look up this descriptor in the table */
|
||||||
while (cache_table[k].descriptor != 0)
|
while (cache_table[k].descriptor != 0) {
|
||||||
{
|
|
||||||
if (cache_table[k].descriptor == des) {
|
if (cache_table[k].descriptor == des) {
|
||||||
if (only_trace && cache_table[k].cache_type != LVL_TRACE)
|
if (only_trace && cache_table[k].cache_type != LVL_TRACE)
|
||||||
break;
|
break;
|
||||||
|
@ -493,9 +494,9 @@ unsigned int __cpuinit init_intel_cacheinfo(struct cpuinfo_x86 *c)
|
||||||
printk(KERN_INFO "CPU: L1 I cache: %dK", l1i);
|
printk(KERN_INFO "CPU: L1 I cache: %dK", l1i);
|
||||||
|
|
||||||
if (l1d)
|
if (l1d)
|
||||||
printk(", L1 D cache: %dK\n", l1d);
|
printk(KERN_CONT ", L1 D cache: %dK\n", l1d);
|
||||||
else
|
else
|
||||||
printk("\n");
|
printk(KERN_CONT "\n");
|
||||||
|
|
||||||
if (l2)
|
if (l2)
|
||||||
printk(KERN_INFO "CPU: L2 cache: %dK\n", l2);
|
printk(KERN_INFO "CPU: L2 cache: %dK\n", l2);
|
||||||
|
@ -558,8 +559,13 @@ static void __cpuinit cache_remove_shared_cpu_map(unsigned int cpu, int index)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
#else
|
#else
|
||||||
static void __cpuinit cache_shared_cpu_map_setup(unsigned int cpu, int index) {}
|
static void __cpuinit cache_shared_cpu_map_setup(unsigned int cpu, int index)
|
||||||
static void __cpuinit cache_remove_shared_cpu_map(unsigned int cpu, int index) {}
|
{
|
||||||
|
}
|
||||||
|
|
||||||
|
static void __cpuinit cache_remove_shared_cpu_map(unsigned int cpu, int index)
|
||||||
|
{
|
||||||
|
}
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
static void __cpuinit free_cache_attributes(unsigned int cpu)
|
static void __cpuinit free_cache_attributes(unsigned int cpu)
|
||||||
|
@ -925,9 +931,8 @@ static int __cpuinit cache_add_dev(struct sys_device * sys_dev)
|
||||||
per_cpu(cache_kobject, cpu),
|
per_cpu(cache_kobject, cpu),
|
||||||
"index%1lu", i);
|
"index%1lu", i);
|
||||||
if (unlikely(retval)) {
|
if (unlikely(retval)) {
|
||||||
for (j = 0; j < i; j++) {
|
for (j = 0; j < i; j++)
|
||||||
kobject_put(&(INDEX_KOBJECT_PTR(cpu, j)->kobj));
|
kobject_put(&(INDEX_KOBJECT_PTR(cpu, j)->kobj));
|
||||||
}
|
|
||||||
kobject_put(per_cpu(cache_kobject, cpu));
|
kobject_put(per_cpu(cache_kobject, cpu));
|
||||||
cpuid4_cache_sysfs_exit(cpu);
|
cpuid4_cache_sysfs_exit(cpu);
|
||||||
return retval;
|
return retval;
|
||||||
|
@ -977,8 +982,7 @@ static int __cpuinit cacheinfo_cpu_callback(struct notifier_block *nfb,
|
||||||
return NOTIFY_OK;
|
return NOTIFY_OK;
|
||||||
}
|
}
|
||||||
|
|
||||||
static struct notifier_block __cpuinitdata cacheinfo_cpu_notifier =
|
static struct notifier_block __cpuinitdata cacheinfo_cpu_notifier = {
|
||||||
{
|
|
||||||
.notifier_call = cacheinfo_cpu_callback,
|
.notifier_call = cacheinfo_cpu_callback,
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
|
@ -68,16 +68,16 @@ static inline unsigned int nmi_perfctr_msr_to_bit(unsigned int msr)
|
||||||
/* returns the bit offset of the performance counter register */
|
/* returns the bit offset of the performance counter register */
|
||||||
switch (boot_cpu_data.x86_vendor) {
|
switch (boot_cpu_data.x86_vendor) {
|
||||||
case X86_VENDOR_AMD:
|
case X86_VENDOR_AMD:
|
||||||
return (msr - MSR_K7_PERFCTR0);
|
return msr - MSR_K7_PERFCTR0;
|
||||||
case X86_VENDOR_INTEL:
|
case X86_VENDOR_INTEL:
|
||||||
if (cpu_has(&boot_cpu_data, X86_FEATURE_ARCH_PERFMON))
|
if (cpu_has(&boot_cpu_data, X86_FEATURE_ARCH_PERFMON))
|
||||||
return (msr - MSR_ARCH_PERFMON_PERFCTR0);
|
return msr - MSR_ARCH_PERFMON_PERFCTR0;
|
||||||
|
|
||||||
switch (boot_cpu_data.x86) {
|
switch (boot_cpu_data.x86) {
|
||||||
case 6:
|
case 6:
|
||||||
return (msr - MSR_P6_PERFCTR0);
|
return msr - MSR_P6_PERFCTR0;
|
||||||
case 15:
|
case 15:
|
||||||
return (msr - MSR_P4_BPU_PERFCTR0);
|
return msr - MSR_P4_BPU_PERFCTR0;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
return 0;
|
return 0;
|
||||||
|
@ -92,16 +92,16 @@ static inline unsigned int nmi_evntsel_msr_to_bit(unsigned int msr)
|
||||||
/* returns the bit offset of the event selection register */
|
/* returns the bit offset of the event selection register */
|
||||||
switch (boot_cpu_data.x86_vendor) {
|
switch (boot_cpu_data.x86_vendor) {
|
||||||
case X86_VENDOR_AMD:
|
case X86_VENDOR_AMD:
|
||||||
return (msr - MSR_K7_EVNTSEL0);
|
return msr - MSR_K7_EVNTSEL0;
|
||||||
case X86_VENDOR_INTEL:
|
case X86_VENDOR_INTEL:
|
||||||
if (cpu_has(&boot_cpu_data, X86_FEATURE_ARCH_PERFMON))
|
if (cpu_has(&boot_cpu_data, X86_FEATURE_ARCH_PERFMON))
|
||||||
return (msr - MSR_ARCH_PERFMON_EVENTSEL0);
|
return msr - MSR_ARCH_PERFMON_EVENTSEL0;
|
||||||
|
|
||||||
switch (boot_cpu_data.x86) {
|
switch (boot_cpu_data.x86) {
|
||||||
case 6:
|
case 6:
|
||||||
return (msr - MSR_P6_EVNTSEL0);
|
return msr - MSR_P6_EVNTSEL0;
|
||||||
case 15:
|
case 15:
|
||||||
return (msr - MSR_P4_BSU_ESCR0);
|
return msr - MSR_P4_BSU_ESCR0;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
return 0;
|
return 0;
|
||||||
|
@ -113,7 +113,7 @@ int avail_to_resrv_perfctr_nmi_bit(unsigned int counter)
|
||||||
{
|
{
|
||||||
BUG_ON(counter > NMI_MAX_COUNTER_BITS);
|
BUG_ON(counter > NMI_MAX_COUNTER_BITS);
|
||||||
|
|
||||||
return (!test_bit(counter, perfctr_nmi_owner));
|
return !test_bit(counter, perfctr_nmi_owner);
|
||||||
}
|
}
|
||||||
|
|
||||||
/* checks the an msr for availability */
|
/* checks the an msr for availability */
|
||||||
|
@ -124,7 +124,7 @@ int avail_to_resrv_perfctr_nmi(unsigned int msr)
|
||||||
counter = nmi_perfctr_msr_to_bit(msr);
|
counter = nmi_perfctr_msr_to_bit(msr);
|
||||||
BUG_ON(counter > NMI_MAX_COUNTER_BITS);
|
BUG_ON(counter > NMI_MAX_COUNTER_BITS);
|
||||||
|
|
||||||
return (!test_bit(counter, perfctr_nmi_owner));
|
return !test_bit(counter, perfctr_nmi_owner);
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(avail_to_resrv_perfctr_nmi_bit);
|
EXPORT_SYMBOL(avail_to_resrv_perfctr_nmi_bit);
|
||||||
|
|
||||||
|
@ -662,7 +662,8 @@ static int setup_intel_arch_watchdog(unsigned nmi_hz)
|
||||||
* NOTE: Corresponding bit = 0 in ebx indicates event present.
|
* NOTE: Corresponding bit = 0 in ebx indicates event present.
|
||||||
*/
|
*/
|
||||||
cpuid(10, &(eax.full), &ebx, &unused, &unused);
|
cpuid(10, &(eax.full), &ebx, &unused, &unused);
|
||||||
if ((eax.split.mask_length < (ARCH_PERFMON_UNHALTED_CORE_CYCLES_INDEX+1)) ||
|
if ((eax.split.mask_length <
|
||||||
|
(ARCH_PERFMON_UNHALTED_CORE_CYCLES_INDEX+1)) ||
|
||||||
(ebx & ARCH_PERFMON_UNHALTED_CORE_CYCLES_PRESENT))
|
(ebx & ARCH_PERFMON_UNHALTED_CORE_CYCLES_PRESENT))
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue