License cleanup: add SPDX GPL-2.0 license identifier to files with no license
Many source files in the tree are missing licensing information, which
makes it harder for compliance tools to determine the correct license.
By default all files without license information are under the default
license of the kernel, which is GPL version 2.
Update the files which contain no license information with the 'GPL-2.0'
SPDX license identifier. The SPDX identifier is a legally binding
shorthand, which can be used instead of the full boiler plate text.
This patch is based on work done by Thomas Gleixner and Kate Stewart and
Philippe Ombredanne.
How this work was done:
Patches were generated and checked against linux-4.14-rc6 for a subset of
the use cases:
- file had no licensing information it it.
- file was a */uapi/* one with no licensing information in it,
- file was a */uapi/* one with existing licensing information,
Further patches will be generated in subsequent months to fix up cases
where non-standard license headers were used, and references to license
had to be inferred by heuristics based on keywords.
The analysis to determine which SPDX License Identifier to be applied to
a file was done in a spreadsheet of side by side results from of the
output of two independent scanners (ScanCode & Windriver) producing SPDX
tag:value files created by Philippe Ombredanne. Philippe prepared the
base worksheet, and did an initial spot review of a few 1000 files.
The 4.13 kernel was the starting point of the analysis with 60,537 files
assessed. Kate Stewart did a file by file comparison of the scanner
results in the spreadsheet to determine which SPDX license identifier(s)
to be applied to the file. She confirmed any determination that was not
immediately clear with lawyers working with the Linux Foundation.
Criteria used to select files for SPDX license identifier tagging was:
- Files considered eligible had to be source code files.
- Make and config files were included as candidates if they contained >5
lines of source
- File already had some variant of a license header in it (even if <5
lines).
All documentation files were explicitly excluded.
The following heuristics were used to determine which SPDX license
identifiers to apply.
- when both scanners couldn't find any license traces, file was
considered to have no license information in it, and the top level
COPYING file license applied.
For non */uapi/* files that summary was:
SPDX license identifier # files
---------------------------------------------------|-------
GPL-2.0 11139
and resulted in the first patch in this series.
If that file was a */uapi/* path one, it was "GPL-2.0 WITH
Linux-syscall-note" otherwise it was "GPL-2.0". Results of that was:
SPDX license identifier # files
---------------------------------------------------|-------
GPL-2.0 WITH Linux-syscall-note 930
and resulted in the second patch in this series.
- if a file had some form of licensing information in it, and was one
of the */uapi/* ones, it was denoted with the Linux-syscall-note if
any GPL family license was found in the file or had no licensing in
it (per prior point). Results summary:
SPDX license identifier # files
---------------------------------------------------|------
GPL-2.0 WITH Linux-syscall-note 270
GPL-2.0+ WITH Linux-syscall-note 169
((GPL-2.0 WITH Linux-syscall-note) OR BSD-2-Clause) 21
((GPL-2.0 WITH Linux-syscall-note) OR BSD-3-Clause) 17
LGPL-2.1+ WITH Linux-syscall-note 15
GPL-1.0+ WITH Linux-syscall-note 14
((GPL-2.0+ WITH Linux-syscall-note) OR BSD-3-Clause) 5
LGPL-2.0+ WITH Linux-syscall-note 4
LGPL-2.1 WITH Linux-syscall-note 3
((GPL-2.0 WITH Linux-syscall-note) OR MIT) 3
((GPL-2.0 WITH Linux-syscall-note) AND MIT) 1
and that resulted in the third patch in this series.
- when the two scanners agreed on the detected license(s), that became
the concluded license(s).
- when there was disagreement between the two scanners (one detected a
license but the other didn't, or they both detected different
licenses) a manual inspection of the file occurred.
- In most cases a manual inspection of the information in the file
resulted in a clear resolution of the license that should apply (and
which scanner probably needed to revisit its heuristics).
- When it was not immediately clear, the license identifier was
confirmed with lawyers working with the Linux Foundation.
- If there was any question as to the appropriate license identifier,
the file was flagged for further research and to be revisited later
in time.
In total, over 70 hours of logged manual review was done on the
spreadsheet to determine the SPDX license identifiers to apply to the
source files by Kate, Philippe, Thomas and, in some cases, confirmation
by lawyers working with the Linux Foundation.
Kate also obtained a third independent scan of the 4.13 code base from
FOSSology, and compared selected files where the other two scanners
disagreed against that SPDX file, to see if there was new insights. The
Windriver scanner is based on an older version of FOSSology in part, so
they are related.
Thomas did random spot checks in about 500 files from the spreadsheets
for the uapi headers and agreed with SPDX license identifier in the
files he inspected. For the non-uapi files Thomas did random spot checks
in about 15000 files.
In initial set of patches against 4.14-rc6, 3 files were found to have
copy/paste license identifier errors, and have been fixed to reflect the
correct identifier.
Additionally Philippe spent 10 hours this week doing a detailed manual
inspection and review of the 12,461 patched files from the initial patch
version early this week with:
- a full scancode scan run, collecting the matched texts, detected
license ids and scores
- reviewing anything where there was a license detected (about 500+
files) to ensure that the applied SPDX license was correct
- reviewing anything where there was no detection but the patch license
was not GPL-2.0 WITH Linux-syscall-note to ensure that the applied
SPDX license was correct
This produced a worksheet with 20 files needing minor correction. This
worksheet was then exported into 3 different .csv files for the
different types of files to be modified.
These .csv files were then reviewed by Greg. Thomas wrote a script to
parse the csv files and add the proper SPDX tag to the file, in the
format that the file expected. This script was further refined by Greg
based on the output to detect more types of files automatically and to
distinguish between header and source .c files (which need different
comment types.) Finally Greg ran the script using the .csv files to
generate the patches.
Reviewed-by: Kate Stewart <kstewart@linuxfoundation.org>
Reviewed-by: Philippe Ombredanne <pombredanne@nexb.com>
Reviewed-by: Thomas Gleixner <tglx@linutronix.de>
Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
2017-11-01 22:07:57 +08:00
|
|
|
/* SPDX-License-Identifier: GPL-2.0 */
|
2005-09-24 03:08:58 +08:00
|
|
|
#ifndef __ASM_POWERPC_CPUTABLE_H
|
|
|
|
#define __ASM_POWERPC_CPUTABLE_H
|
|
|
|
|
2008-06-24 01:48:21 +08:00
|
|
|
|
2016-07-27 11:35:15 +08:00
|
|
|
#include <linux/types.h>
|
2012-10-09 16:47:26 +08:00
|
|
|
#include <uapi/asm/cputable.h>
|
2018-07-06 00:24:57 +08:00
|
|
|
#include <asm/asm-const.h>
|
2008-06-24 01:48:21 +08:00
|
|
|
|
2005-09-24 03:08:58 +08:00
|
|
|
#ifndef __ASSEMBLY__
|
|
|
|
|
|
|
|
/* This structure can grow, it's real size is used by head.S code
|
|
|
|
* via the mkdefs mechanism.
|
|
|
|
*/
|
|
|
|
struct cpu_spec;
|
|
|
|
|
|
|
|
typedef void (*cpu_setup_t)(unsigned long offset, struct cpu_spec* spec);
|
2006-08-11 13:07:08 +08:00
|
|
|
typedef void (*cpu_restore_t)(void);
|
2005-09-24 03:08:58 +08:00
|
|
|
|
2006-01-09 12:41:31 +08:00
|
|
|
enum powerpc_oprofile_type {
|
2006-01-13 20:35:49 +08:00
|
|
|
PPC_OPROFILE_INVALID = 0,
|
|
|
|
PPC_OPROFILE_RS64 = 1,
|
|
|
|
PPC_OPROFILE_POWER4 = 2,
|
|
|
|
PPC_OPROFILE_G4 = 3,
|
2008-02-05 08:27:55 +08:00
|
|
|
PPC_OPROFILE_FSL_EMB = 4,
|
[POWERPC] cell: Add oprofile support
Add PPU event-based and cycle-based profiling support to Oprofile for Cell.
Oprofile is expected to collect data on all CPUs simultaneously.
However, there is one set of performance counters per node. There are
two hardware threads or virtual CPUs on each node. Hence, OProfile must
multiplex in time the performance counter collection on the two virtual
CPUs.
The multiplexing of the performance counters is done by a virtual
counter routine. Initially, the counters are configured to collect data
on the even CPUs in the system, one CPU per node. In order to capture
the PC for the virtual CPU when the performance counter interrupt occurs
(the specified number of events between samples has occurred), the even
processors are configured to handle the performance counter interrupts
for their node. The virtual counter routine is called via a kernel
timer after the virtual sample time. The routine stops the counters,
saves the current counts, loads the last counts for the other virtual
CPU on the node, sets interrupts to be handled by the other virtual CPU
and restarts the counters, the virtual timer routine is scheduled to run
again. The virtual sample time is kept relatively small to make sure
sampling occurs on both CPUs on the node with a relatively small
granularity. Whenever the counters overflow, the performance counter
interrupt is called to collect the PC for the CPU where data is being
collected.
The oprofile driver relies on a firmware RTAS call to setup the debug bus
to route the desired signals to the performance counter hardware to be
counted. The RTAS call must set the routing registers appropriately in
each of the islands to pass the signals down the debug bus as well as
routing the signals from a particular island onto the bus. There is a
second firmware RTAS call to reset the debug bus to the non pass thru
state when the counters are not in use.
Signed-off-by: Carl Love <carll@us.ibm.com>
Signed-off-by: Maynard Johnson <mpjohn@us.ibm.com>
Signed-off-by: Arnd Bergmann <arnd.bergmann@de.ibm.com>
Signed-off-by: Paul Mackerras <paulus@samba.org>
2006-11-21 01:45:16 +08:00
|
|
|
PPC_OPROFILE_CELL = 5,
|
2007-04-18 14:38:21 +08:00
|
|
|
PPC_OPROFILE_PA6T = 6,
|
2006-01-09 12:41:31 +08:00
|
|
|
};
|
|
|
|
|
2007-01-29 11:23:54 +08:00
|
|
|
enum powerpc_pmc_type {
|
|
|
|
PPC_PMC_DEFAULT = 0,
|
|
|
|
PPC_PMC_IBM = 1,
|
|
|
|
PPC_PMC_PA6T = 2,
|
2008-08-18 12:23:51 +08:00
|
|
|
PPC_PMC_G4 = 3,
|
2007-01-29 11:23:54 +08:00
|
|
|
};
|
|
|
|
|
2007-12-21 12:39:21 +08:00
|
|
|
struct pt_regs;
|
|
|
|
|
|
|
|
extern int machine_check_generic(struct pt_regs *regs);
|
|
|
|
extern int machine_check_4xx(struct pt_regs *regs);
|
|
|
|
extern int machine_check_440A(struct pt_regs *regs);
|
2010-04-08 13:38:22 +08:00
|
|
|
extern int machine_check_e500mc(struct pt_regs *regs);
|
2007-12-21 12:39:21 +08:00
|
|
|
extern int machine_check_e500(struct pt_regs *regs);
|
|
|
|
extern int machine_check_e200(struct pt_regs *regs);
|
2010-03-05 11:43:18 +08:00
|
|
|
extern int machine_check_47x(struct pt_regs *regs);
|
2016-09-16 16:23:11 +08:00
|
|
|
int machine_check_8xx(struct pt_regs *regs);
|
2018-12-10 19:41:29 +08:00
|
|
|
int machine_check_83xx(struct pt_regs *regs);
|
2007-12-21 12:39:21 +08:00
|
|
|
|
2015-11-20 17:13:58 +08:00
|
|
|
extern void cpu_down_flush_e500v2(void);
|
|
|
|
extern void cpu_down_flush_e500mc(void);
|
|
|
|
extern void cpu_down_flush_e5500(void);
|
|
|
|
extern void cpu_down_flush_e6500(void);
|
|
|
|
|
[POWERPC] Fix performance monitor on machines with logical PVR
Some IBM machines supply a "logical" PVR (processor version register)
value in the device tree in the cpu nodes rather than the real PVR.
This is used for instance to indicate that the processors in a POWER6
partition have been configured by the hypervisor to run in POWER5+
mode rather than POWER6 mode. To cope with this, we call identify_cpu
a second time with the logical PVR value (the first call is with the
real PVR value in the very early setup code).
However, POWER5+ machines can also supply a logical PVR value, and use
the same value (the value that indicates a v2.04 architecture
compliant processor). This causes problems for code that uses the
performance monitor (such as oprofile), because the PMU registers are
different in POWER6 (even in POWER5+ mode) from the real POWER5+.
This change works around this problem by taking out the PMU
information from the cputable entries for the logical PVR values, and
changing identify_cpu so that the second call to it won't overwrite
the PMU information that was established by the first call (the one
with the real PVR), but does update the other fields. Specifically,
if the cputable entry for the logical PVR value has num_pmcs == 0,
none of the PMU-related fields get used.
So that we can create a mixed cputable entry, we now make cur_cpu_spec
point to a single static struct cpu_spec, and copy stuff from
cpu_specs[i] into it. This has the side-effect that we can now make
cpu_specs[] be initdata.
Ultimately it would be good to move the PMU-related fields out to a
separate structure, pointed to by the cputable entries, and change
identify_cpu so that it saves the PMU info pointer, copies the whole
structure, and restores the PMU info pointer, rather than identify_cpu
having to list all the fields that are *not* PMU-related.
Signed-off-by: Paul Mackerras <paulus@samba.org>
Acked-by: Benjamin Herrenschmidt <benh@kernel.crashing.org>
2007-10-04 12:18:01 +08:00
|
|
|
/* NOTE WELL: Update identify_cpu() if fields are added or removed! */
|
2005-09-24 03:08:58 +08:00
|
|
|
struct cpu_spec {
|
|
|
|
/* CPU is matched via (PVR & pvr_mask) == pvr_value */
|
|
|
|
unsigned int pvr_mask;
|
|
|
|
unsigned int pvr_value;
|
|
|
|
|
|
|
|
char *cpu_name;
|
|
|
|
unsigned long cpu_features; /* Kernel features */
|
|
|
|
unsigned int cpu_user_features; /* Userland features */
|
2013-04-18 01:33:11 +08:00
|
|
|
unsigned int cpu_user_features2; /* Userland features v2 */
|
2008-12-19 03:13:32 +08:00
|
|
|
unsigned int mmu_features; /* MMU features */
|
2005-09-24 03:08:58 +08:00
|
|
|
|
|
|
|
/* cache line sizes */
|
|
|
|
unsigned int icache_bsize;
|
|
|
|
unsigned int dcache_bsize;
|
|
|
|
|
2015-11-20 17:13:58 +08:00
|
|
|
/* flush caches inside the current cpu */
|
|
|
|
void (*cpu_down_flush)(void);
|
|
|
|
|
2005-09-24 03:08:58 +08:00
|
|
|
/* number of performance monitor counters */
|
|
|
|
unsigned int num_pmcs;
|
2007-01-29 11:23:54 +08:00
|
|
|
enum powerpc_pmc_type pmc_type;
|
2005-09-24 03:08:58 +08:00
|
|
|
|
|
|
|
/* this is called to initialize various CPU bits like L1 cache,
|
|
|
|
* BHT, SPD, etc... from head.S before branching to identify_machine
|
|
|
|
*/
|
|
|
|
cpu_setup_t cpu_setup;
|
2006-08-11 13:07:08 +08:00
|
|
|
/* Used to restore cpu setup on secondary processors and at resume */
|
|
|
|
cpu_restore_t cpu_restore;
|
2005-09-24 03:08:58 +08:00
|
|
|
|
|
|
|
/* Used by oprofile userspace to select the right counters */
|
|
|
|
char *oprofile_cpu_type;
|
|
|
|
|
|
|
|
/* Processor specific oprofile operations */
|
2006-01-09 12:41:31 +08:00
|
|
|
enum powerpc_oprofile_type oprofile_type;
|
2006-01-14 07:11:39 +08:00
|
|
|
|
2006-06-08 12:42:34 +08:00
|
|
|
/* Bit locations inside the mmcra change */
|
|
|
|
unsigned long oprofile_mmcra_sihv;
|
|
|
|
unsigned long oprofile_mmcra_sipr;
|
|
|
|
|
|
|
|
/* Bits to clear during an oprofile exception */
|
|
|
|
unsigned long oprofile_mmcra_clear;
|
|
|
|
|
2006-01-14 07:11:39 +08:00
|
|
|
/* Name of processor class, for the ELF AT_PLATFORM entry */
|
|
|
|
char *platform;
|
2007-12-21 12:39:21 +08:00
|
|
|
|
|
|
|
/* Processor specific machine check handling. Return negative
|
|
|
|
* if the error is fatal, 1 if it was fully recovered and 0 to
|
|
|
|
* pass up (not CPU originated) */
|
|
|
|
int (*machine_check)(struct pt_regs *regs);
|
2013-10-30 22:34:40 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Processor specific early machine check handler which is
|
|
|
|
* called in real mode to handle SLB and TLB errors.
|
|
|
|
*/
|
|
|
|
long (*machine_check_early)(struct pt_regs *regs);
|
2005-09-24 03:08:58 +08:00
|
|
|
};
|
|
|
|
|
|
|
|
extern struct cpu_spec *cur_cpu_spec;
|
|
|
|
|
2006-10-24 14:42:40 +08:00
|
|
|
extern unsigned int __start___ftr_fixup, __stop___ftr_fixup;
|
|
|
|
|
2017-05-09 11:16:52 +08:00
|
|
|
extern void set_cur_cpu_spec(struct cpu_spec *s);
|
2006-11-10 17:38:53 +08:00
|
|
|
extern struct cpu_spec *identify_cpu(unsigned long offset, unsigned int pvr);
|
2017-05-09 11:16:52 +08:00
|
|
|
extern void identify_cpu_name(unsigned int pvr);
|
2006-10-20 09:47:18 +08:00
|
|
|
extern void do_feature_fixups(unsigned long value, void *fixup_start,
|
|
|
|
void *fixup_end);
|
2005-10-06 10:06:20 +08:00
|
|
|
|
2008-07-16 07:58:51 +08:00
|
|
|
extern const char *powerpc_base_platform;
|
|
|
|
|
powerpc: Add option to use jump label for cpu_has_feature()
We do binary patching of asm code using CPU features, which is a
one-time operation, done during early boot. However checks of CPU
features in C code are currently done at run time, even though the set
of CPU features can never change after boot.
We can optimise this by using jump labels to implement cpu_has_feature(),
meaning checks in C code are binary patched into a single nop or branch.
For a C sequence along the lines of:
if (cpu_has_feature(FOO))
return 2;
The generated code before is roughly:
ld r9,-27640(r2)
ld r9,0(r9)
lwz r9,32(r9)
cmpwi cr7,r9,0
bge cr7, 1f
li r3,2
blr
1: ...
After (true):
nop
li r3,2
blr
After (false):
b 1f
li r3,2
blr
1: ...
mpe: Rename MAX_CPU_FEATURES as we already have a #define with that
name, and define it simply as a constant, rather than doing tricks with
sizeof and NULL pointers. Rename the array to cpu_feature_keys. Use the
kconfig we added to guard it. Add BUILD_BUG_ON() if the feature is not a
compile time constant. Rewrite the change log.
Signed-off-by: Kevin Hao <haokexin@gmail.com>
Signed-off-by: Aneesh Kumar K.V <aneesh.kumar@linux.vnet.ibm.com>
Signed-off-by: Michael Ellerman <mpe@ellerman.id.au>
2016-07-23 17:12:41 +08:00
|
|
|
#ifdef CONFIG_JUMP_LABEL_FEATURE_CHECKS
|
|
|
|
extern void cpu_feature_keys_init(void);
|
|
|
|
#else
|
|
|
|
static inline void cpu_feature_keys_init(void) { }
|
|
|
|
#endif
|
|
|
|
|
2005-09-24 03:08:58 +08:00
|
|
|
#endif /* __ASSEMBLY__ */
|
|
|
|
|
|
|
|
/* CPU kernel features */
|
|
|
|
|
2018-03-20 05:46:13 +08:00
|
|
|
/* Definitions for features that we have on both 32-bit and 64-bit chips */
|
2012-12-20 22:06:39 +08:00
|
|
|
#define CPU_FTR_COHERENT_ICACHE ASM_CONST(0x00000001)
|
2018-03-20 05:46:13 +08:00
|
|
|
#define CPU_FTR_ALTIVEC ASM_CONST(0x00000002)
|
|
|
|
#define CPU_FTR_DBELL ASM_CONST(0x00000004)
|
|
|
|
#define CPU_FTR_CAN_NAP ASM_CONST(0x00000008)
|
|
|
|
#define CPU_FTR_DEBUG_LVL_EXC ASM_CONST(0x00000010)
|
|
|
|
#define CPU_FTR_NODSISRALIGN ASM_CONST(0x00000020)
|
|
|
|
#define CPU_FTR_FPU_UNAVAILABLE ASM_CONST(0x00000040)
|
|
|
|
#define CPU_FTR_LWSYNC ASM_CONST(0x00000080)
|
|
|
|
#define CPU_FTR_NOEXECUTE ASM_CONST(0x00000100)
|
|
|
|
#define CPU_FTR_EMB_HV ASM_CONST(0x00000200)
|
|
|
|
|
|
|
|
/* Definitions for features that only exist on 32-bit chips */
|
|
|
|
#ifdef CONFIG_PPC32
|
|
|
|
#define CPU_FTR_L2CR ASM_CONST(0x00002000)
|
|
|
|
#define CPU_FTR_SPEC7450 ASM_CONST(0x00004000)
|
|
|
|
#define CPU_FTR_TAU ASM_CONST(0x00008000)
|
|
|
|
#define CPU_FTR_CAN_DOZE ASM_CONST(0x00010000)
|
|
|
|
#define CPU_FTR_L3CR ASM_CONST(0x00040000)
|
|
|
|
#define CPU_FTR_L3_DISABLE_NAP ASM_CONST(0x00080000)
|
|
|
|
#define CPU_FTR_NAP_DISABLE_L2_PR ASM_CONST(0x00100000)
|
|
|
|
#define CPU_FTR_DUAL_PLL_750FX ASM_CONST(0x00200000)
|
|
|
|
#define CPU_FTR_NO_DPM ASM_CONST(0x00400000)
|
|
|
|
#define CPU_FTR_476_DD2 ASM_CONST(0x00800000)
|
|
|
|
#define CPU_FTR_NEED_COHERENT ASM_CONST(0x01000000)
|
|
|
|
#define CPU_FTR_NO_BTIC ASM_CONST(0x02000000)
|
|
|
|
#define CPU_FTR_PPC_LE ASM_CONST(0x04000000)
|
|
|
|
#define CPU_FTR_SPE ASM_CONST(0x10000000)
|
|
|
|
#define CPU_FTR_NEED_PAIRED_STWCX ASM_CONST(0x20000000)
|
|
|
|
#define CPU_FTR_INDEXED_DCR ASM_CONST(0x40000000)
|
|
|
|
|
|
|
|
#else /* CONFIG_PPC32 */
|
|
|
|
/* Define these to 0 for the sake of tests in common code */
|
|
|
|
#define CPU_FTR_PPC_LE (0)
|
|
|
|
#endif
|
2005-09-24 03:08:58 +08:00
|
|
|
|
2006-06-28 11:50:39 +08:00
|
|
|
/*
|
2018-03-20 05:46:13 +08:00
|
|
|
* Definitions for the 64-bit processor unique features;
|
2006-06-28 11:50:39 +08:00
|
|
|
* on 32-bit, make the names available but defined to be 0.
|
|
|
|
*/
|
2005-09-24 03:08:58 +08:00
|
|
|
#ifdef __powerpc64__
|
2006-06-28 11:50:39 +08:00
|
|
|
#define LONG_ASM_CONST(x) ASM_CONST(x)
|
2005-09-24 03:08:58 +08:00
|
|
|
#else
|
2006-06-28 11:50:39 +08:00
|
|
|
#define LONG_ASM_CONST(x) 0
|
2005-09-24 03:08:58 +08:00
|
|
|
#endif
|
|
|
|
|
2018-03-20 05:46:13 +08:00
|
|
|
#define CPU_FTR_REAL_LE LONG_ASM_CONST(0x0000000000001000)
|
|
|
|
#define CPU_FTR_HVMODE LONG_ASM_CONST(0x0000000000002000)
|
|
|
|
#define CPU_FTR_ARCH_206 LONG_ASM_CONST(0x0000000000008000)
|
|
|
|
#define CPU_FTR_ARCH_207S LONG_ASM_CONST(0x0000000000010000)
|
|
|
|
#define CPU_FTR_ARCH_300 LONG_ASM_CONST(0x0000000000020000)
|
|
|
|
#define CPU_FTR_MMCRA LONG_ASM_CONST(0x0000000000040000)
|
|
|
|
#define CPU_FTR_CTRL LONG_ASM_CONST(0x0000000000080000)
|
|
|
|
#define CPU_FTR_SMT LONG_ASM_CONST(0x0000000000100000)
|
|
|
|
#define CPU_FTR_PAUSE_ZERO LONG_ASM_CONST(0x0000000000200000)
|
|
|
|
#define CPU_FTR_PURR LONG_ASM_CONST(0x0000000000400000)
|
|
|
|
#define CPU_FTR_CELL_TB_BUG LONG_ASM_CONST(0x0000000000800000)
|
|
|
|
#define CPU_FTR_SPURR LONG_ASM_CONST(0x0000000001000000)
|
|
|
|
#define CPU_FTR_DSCR LONG_ASM_CONST(0x0000000002000000)
|
|
|
|
#define CPU_FTR_VSX LONG_ASM_CONST(0x0000000004000000)
|
|
|
|
#define CPU_FTR_SAO LONG_ASM_CONST(0x0000000008000000)
|
|
|
|
#define CPU_FTR_CP_USE_DCBTZ LONG_ASM_CONST(0x0000000010000000)
|
|
|
|
#define CPU_FTR_UNALIGNED_LD_STD LONG_ASM_CONST(0x0000000020000000)
|
|
|
|
#define CPU_FTR_ASYM_SMT LONG_ASM_CONST(0x0000000040000000)
|
|
|
|
#define CPU_FTR_STCX_CHECKS_ADDRESS LONG_ASM_CONST(0x0000000080000000)
|
|
|
|
#define CPU_FTR_POPCNTB LONG_ASM_CONST(0x0000000100000000)
|
|
|
|
#define CPU_FTR_POPCNTD LONG_ASM_CONST(0x0000000200000000)
|
|
|
|
#define CPU_FTR_PKEY LONG_ASM_CONST(0x0000000400000000)
|
|
|
|
#define CPU_FTR_VMX_COPY LONG_ASM_CONST(0x0000000800000000)
|
|
|
|
#define CPU_FTR_TM LONG_ASM_CONST(0x0000001000000000)
|
|
|
|
#define CPU_FTR_CFAR LONG_ASM_CONST(0x0000002000000000)
|
|
|
|
#define CPU_FTR_HAS_PPR LONG_ASM_CONST(0x0000004000000000)
|
|
|
|
#define CPU_FTR_DAWR LONG_ASM_CONST(0x0000008000000000)
|
|
|
|
#define CPU_FTR_DABRX LONG_ASM_CONST(0x0000010000000000)
|
|
|
|
#define CPU_FTR_PMAO_BUG LONG_ASM_CONST(0x0000020000000000)
|
|
|
|
#define CPU_FTR_POWER9_DD2_1 LONG_ASM_CONST(0x0000080000000000)
|
2018-03-21 18:31:59 +08:00
|
|
|
#define CPU_FTR_P9_TM_HV_ASSIST LONG_ASM_CONST(0x0000100000000000)
|
|
|
|
#define CPU_FTR_P9_TM_XER_SO_BUG LONG_ASM_CONST(0x0000200000000000)
|
2019-09-24 11:52:52 +08:00
|
|
|
#define CPU_FTR_P9_TLBIE_STQ_BUG LONG_ASM_CONST(0x0000400000000000)
|
2018-05-11 14:12:57 +08:00
|
|
|
#define CPU_FTR_P9_TIDR LONG_ASM_CONST(0x0000800000000000)
|
2019-09-24 11:52:53 +08:00
|
|
|
#define CPU_FTR_P9_TLBIE_ERAT_BUG LONG_ASM_CONST(0x0001000000000000)
|
2006-06-28 11:50:39 +08:00
|
|
|
|
2005-09-24 03:08:58 +08:00
|
|
|
#ifndef __ASSEMBLY__
|
|
|
|
|
2011-04-07 03:48:50 +08:00
|
|
|
#define CPU_FTR_PPCAS_ARCH_V2 (CPU_FTR_NOEXECUTE | CPU_FTR_NODSISRALIGN)
|
|
|
|
|
2014-07-10 10:29:20 +08:00
|
|
|
#define MMU_FTR_PPCAS_ARCH_V2 (MMU_FTR_TLBIEL | MMU_FTR_16M_PAGE)
|
2005-09-24 03:08:58 +08:00
|
|
|
|
|
|
|
/* We only set the altivec features if the kernel was compiled with altivec
|
|
|
|
* support
|
|
|
|
*/
|
|
|
|
#ifdef CONFIG_ALTIVEC
|
|
|
|
#define CPU_FTR_ALTIVEC_COMP CPU_FTR_ALTIVEC
|
|
|
|
#define PPC_FEATURE_HAS_ALTIVEC_COMP PPC_FEATURE_HAS_ALTIVEC
|
|
|
|
#else
|
|
|
|
#define CPU_FTR_ALTIVEC_COMP 0
|
|
|
|
#define PPC_FEATURE_HAS_ALTIVEC_COMP 0
|
|
|
|
#endif
|
|
|
|
|
2008-06-25 12:07:18 +08:00
|
|
|
/* We only set the VSX features if the kernel was compiled with VSX
|
|
|
|
* support
|
|
|
|
*/
|
|
|
|
#ifdef CONFIG_VSX
|
|
|
|
#define CPU_FTR_VSX_COMP CPU_FTR_VSX
|
|
|
|
#define PPC_FEATURE_HAS_VSX_COMP PPC_FEATURE_HAS_VSX
|
|
|
|
#else
|
|
|
|
#define CPU_FTR_VSX_COMP 0
|
|
|
|
#define PPC_FEATURE_HAS_VSX_COMP 0
|
|
|
|
#endif
|
|
|
|
|
2007-09-13 14:44:20 +08:00
|
|
|
/* We only set the spe features if the kernel was compiled with spe
|
|
|
|
* support
|
|
|
|
*/
|
|
|
|
#ifdef CONFIG_SPE
|
|
|
|
#define CPU_FTR_SPE_COMP CPU_FTR_SPE
|
|
|
|
#define PPC_FEATURE_HAS_SPE_COMP PPC_FEATURE_HAS_SPE
|
|
|
|
#define PPC_FEATURE_HAS_EFP_SINGLE_COMP PPC_FEATURE_HAS_EFP_SINGLE
|
|
|
|
#define PPC_FEATURE_HAS_EFP_DOUBLE_COMP PPC_FEATURE_HAS_EFP_DOUBLE
|
|
|
|
#else
|
|
|
|
#define CPU_FTR_SPE_COMP 0
|
|
|
|
#define PPC_FEATURE_HAS_SPE_COMP 0
|
|
|
|
#define PPC_FEATURE_HAS_EFP_SINGLE_COMP 0
|
|
|
|
#define PPC_FEATURE_HAS_EFP_DOUBLE_COMP 0
|
|
|
|
#endif
|
|
|
|
|
2013-02-14 00:21:29 +08:00
|
|
|
/* We only set the TM feature if the kernel was compiled with TM supprt */
|
|
|
|
#ifdef CONFIG_PPC_TRANSACTIONAL_MEM
|
2015-06-12 09:06:32 +08:00
|
|
|
#define CPU_FTR_TM_COMP CPU_FTR_TM
|
|
|
|
#define PPC_FEATURE2_HTM_COMP PPC_FEATURE2_HTM
|
|
|
|
#define PPC_FEATURE2_HTM_NOSC_COMP PPC_FEATURE2_HTM_NOSC
|
2013-02-14 00:21:29 +08:00
|
|
|
#else
|
2015-06-12 09:06:32 +08:00
|
|
|
#define CPU_FTR_TM_COMP 0
|
|
|
|
#define PPC_FEATURE2_HTM_COMP 0
|
|
|
|
#define PPC_FEATURE2_HTM_NOSC_COMP 0
|
2013-02-14 00:21:29 +08:00
|
|
|
#endif
|
|
|
|
|
2007-09-15 04:32:14 +08:00
|
|
|
/* We need to mark all pages as being coherent if we're SMP or we have a
|
|
|
|
* 74[45]x and an MPC107 host bridge. Also 83xx and PowerQUICC II
|
|
|
|
* require it for PCI "streaming/prefetch" to work properly.
|
2009-03-17 23:17:50 +08:00
|
|
|
* This is also required by 52xx family.
|
2005-09-24 03:08:58 +08:00
|
|
|
*/
|
2006-02-22 23:46:02 +08:00
|
|
|
#if defined(CONFIG_SMP) || defined(CONFIG_MPC10X_BRIDGE) \
|
2009-03-17 23:17:50 +08:00
|
|
|
|| defined(CONFIG_PPC_83xx) || defined(CONFIG_8260) \
|
|
|
|
|| defined(CONFIG_PPC_MPC52xx)
|
2005-09-24 03:08:58 +08:00
|
|
|
#define CPU_FTR_COMMON CPU_FTR_NEED_COHERENT
|
|
|
|
#else
|
|
|
|
#define CPU_FTR_COMMON 0
|
|
|
|
#endif
|
|
|
|
|
|
|
|
/* The powersave features NAP & DOZE seems to confuse BDI when
|
|
|
|
debugging. So if a BDI is used, disable theses
|
|
|
|
*/
|
|
|
|
#ifndef CONFIG_BDI_SWITCH
|
|
|
|
#define CPU_FTR_MAYBE_CAN_DOZE CPU_FTR_CAN_DOZE
|
|
|
|
#define CPU_FTR_MAYBE_CAN_NAP CPU_FTR_CAN_NAP
|
|
|
|
#else
|
|
|
|
#define CPU_FTR_MAYBE_CAN_DOZE 0
|
|
|
|
#define CPU_FTR_MAYBE_CAN_NAP 0
|
|
|
|
#endif
|
|
|
|
|
2019-08-26 23:52:14 +08:00
|
|
|
#define CPU_FTRS_PPC601 (CPU_FTR_COMMON | \
|
2019-08-26 23:52:18 +08:00
|
|
|
CPU_FTR_COHERENT_ICACHE)
|
2018-03-20 05:46:11 +08:00
|
|
|
#define CPU_FTRS_603 (CPU_FTR_COMMON | CPU_FTR_MAYBE_CAN_DOZE | \
|
2018-11-29 01:21:10 +08:00
|
|
|
CPU_FTR_MAYBE_CAN_NAP | CPU_FTR_PPC_LE | CPU_FTR_NOEXECUTE)
|
2018-03-20 05:46:11 +08:00
|
|
|
#define CPU_FTRS_604 (CPU_FTR_COMMON | CPU_FTR_PPC_LE)
|
[POWERPC] Merge CPU features pertaining to icache coherency
Currently the powerpc kernel has a 64-bit only feature,
COHERENT_ICACHE used for those CPUS which maintain icache/dcache
coherency in hardware (POWER5, essentially). It also has a feature,
SPLIT_ID_CACHE, which is used on CPUs which have separate i and
d-caches, which is to say everything except 601 and Freescale E200.
In nearly all the places we check the SPLIT_ID_CACHE, what we actually
care about is whether the i and d-caches are coherent (which they will
be, trivially, if they're the same cache).
This tries to clarify the situation a little. The COHERENT_ICACHE
feature becomes availble on 32-bit and is set for all CPUs where i and
d-cache are effectively coherent, whether this is due to special logic
(POWER5) or because they're unified. We check this, instead of
SPLIT_ID_CACHE nearly everywhere.
The SPLIT_ID_CACHE feature itself is replaced by a UNIFIED_ID_CACHE
feature with reversed sense, set only on 601 and Freescale E200. In
the two places (one Freescale BookE specific) where we really care
whether it's a unified cache, not whether they're coherent, we check
this feature. The CPUs with unified cache are so few, we could
consider replacing this feature bit with explicit checks against the
PVR.
This will make unifying the 32-bit and 64-bit cache flush code a
little more straightforward.
Signed-off-by: David Gibson <dwg@au1.ibm.com>
Signed-off-by: Paul Mackerras <paulus@samba.org>
2007-06-13 12:52:57 +08:00
|
|
|
#define CPU_FTRS_740_NOTAU (CPU_FTR_COMMON | \
|
2018-03-20 05:46:11 +08:00
|
|
|
CPU_FTR_MAYBE_CAN_DOZE | CPU_FTR_L2CR | \
|
2008-12-19 03:13:32 +08:00
|
|
|
CPU_FTR_MAYBE_CAN_NAP | CPU_FTR_PPC_LE)
|
[POWERPC] Merge CPU features pertaining to icache coherency
Currently the powerpc kernel has a 64-bit only feature,
COHERENT_ICACHE used for those CPUS which maintain icache/dcache
coherency in hardware (POWER5, essentially). It also has a feature,
SPLIT_ID_CACHE, which is used on CPUs which have separate i and
d-caches, which is to say everything except 601 and Freescale E200.
In nearly all the places we check the SPLIT_ID_CACHE, what we actually
care about is whether the i and d-caches are coherent (which they will
be, trivially, if they're the same cache).
This tries to clarify the situation a little. The COHERENT_ICACHE
feature becomes availble on 32-bit and is set for all CPUs where i and
d-cache are effectively coherent, whether this is due to special logic
(POWER5) or because they're unified. We check this, instead of
SPLIT_ID_CACHE nearly everywhere.
The SPLIT_ID_CACHE feature itself is replaced by a UNIFIED_ID_CACHE
feature with reversed sense, set only on 601 and Freescale E200. In
the two places (one Freescale BookE specific) where we really care
whether it's a unified cache, not whether they're coherent, we check
this feature. The CPUs with unified cache are so few, we could
consider replacing this feature bit with explicit checks against the
PVR.
This will make unifying the 32-bit and 64-bit cache flush code a
little more straightforward.
Signed-off-by: David Gibson <dwg@au1.ibm.com>
Signed-off-by: Paul Mackerras <paulus@samba.org>
2007-06-13 12:52:57 +08:00
|
|
|
#define CPU_FTRS_740 (CPU_FTR_COMMON | \
|
2018-03-20 05:46:11 +08:00
|
|
|
CPU_FTR_MAYBE_CAN_DOZE | CPU_FTR_L2CR | \
|
2008-12-19 03:13:32 +08:00
|
|
|
CPU_FTR_TAU | CPU_FTR_MAYBE_CAN_NAP | \
|
2006-06-07 14:14:40 +08:00
|
|
|
CPU_FTR_PPC_LE)
|
[POWERPC] Merge CPU features pertaining to icache coherency
Currently the powerpc kernel has a 64-bit only feature,
COHERENT_ICACHE used for those CPUS which maintain icache/dcache
coherency in hardware (POWER5, essentially). It also has a feature,
SPLIT_ID_CACHE, which is used on CPUs which have separate i and
d-caches, which is to say everything except 601 and Freescale E200.
In nearly all the places we check the SPLIT_ID_CACHE, what we actually
care about is whether the i and d-caches are coherent (which they will
be, trivially, if they're the same cache).
This tries to clarify the situation a little. The COHERENT_ICACHE
feature becomes availble on 32-bit and is set for all CPUs where i and
d-cache are effectively coherent, whether this is due to special logic
(POWER5) or because they're unified. We check this, instead of
SPLIT_ID_CACHE nearly everywhere.
The SPLIT_ID_CACHE feature itself is replaced by a UNIFIED_ID_CACHE
feature with reversed sense, set only on 601 and Freescale E200. In
the two places (one Freescale BookE specific) where we really care
whether it's a unified cache, not whether they're coherent, we check
this feature. The CPUs with unified cache are so few, we could
consider replacing this feature bit with explicit checks against the
PVR.
This will make unifying the 32-bit and 64-bit cache flush code a
little more straightforward.
Signed-off-by: David Gibson <dwg@au1.ibm.com>
Signed-off-by: Paul Mackerras <paulus@samba.org>
2007-06-13 12:52:57 +08:00
|
|
|
#define CPU_FTRS_750 (CPU_FTR_COMMON | \
|
2018-03-20 05:46:11 +08:00
|
|
|
CPU_FTR_MAYBE_CAN_DOZE | CPU_FTR_L2CR | \
|
2008-12-19 03:13:32 +08:00
|
|
|
CPU_FTR_TAU | CPU_FTR_MAYBE_CAN_NAP | \
|
2006-06-07 14:14:40 +08:00
|
|
|
CPU_FTR_PPC_LE)
|
2008-12-19 03:13:32 +08:00
|
|
|
#define CPU_FTRS_750CL (CPU_FTRS_750)
|
2007-07-03 00:06:53 +08:00
|
|
|
#define CPU_FTRS_750FX1 (CPU_FTRS_750 | CPU_FTR_DUAL_PLL_750FX | CPU_FTR_NO_DPM)
|
|
|
|
#define CPU_FTRS_750FX2 (CPU_FTRS_750 | CPU_FTR_NO_DPM)
|
2008-12-19 03:13:32 +08:00
|
|
|
#define CPU_FTRS_750FX (CPU_FTRS_750 | CPU_FTR_DUAL_PLL_750FX)
|
2007-07-03 00:06:53 +08:00
|
|
|
#define CPU_FTRS_750GX (CPU_FTRS_750FX)
|
[POWERPC] Merge CPU features pertaining to icache coherency
Currently the powerpc kernel has a 64-bit only feature,
COHERENT_ICACHE used for those CPUS which maintain icache/dcache
coherency in hardware (POWER5, essentially). It also has a feature,
SPLIT_ID_CACHE, which is used on CPUs which have separate i and
d-caches, which is to say everything except 601 and Freescale E200.
In nearly all the places we check the SPLIT_ID_CACHE, what we actually
care about is whether the i and d-caches are coherent (which they will
be, trivially, if they're the same cache).
This tries to clarify the situation a little. The COHERENT_ICACHE
feature becomes availble on 32-bit and is set for all CPUs where i and
d-cache are effectively coherent, whether this is due to special logic
(POWER5) or because they're unified. We check this, instead of
SPLIT_ID_CACHE nearly everywhere.
The SPLIT_ID_CACHE feature itself is replaced by a UNIFIED_ID_CACHE
feature with reversed sense, set only on 601 and Freescale E200. In
the two places (one Freescale BookE specific) where we really care
whether it's a unified cache, not whether they're coherent, we check
this feature. The CPUs with unified cache are so few, we could
consider replacing this feature bit with explicit checks against the
PVR.
This will make unifying the 32-bit and 64-bit cache flush code a
little more straightforward.
Signed-off-by: David Gibson <dwg@au1.ibm.com>
Signed-off-by: Paul Mackerras <paulus@samba.org>
2007-06-13 12:52:57 +08:00
|
|
|
#define CPU_FTRS_7400_NOTAU (CPU_FTR_COMMON | \
|
2018-03-20 05:46:11 +08:00
|
|
|
CPU_FTR_MAYBE_CAN_DOZE | CPU_FTR_L2CR | \
|
2008-12-19 03:13:32 +08:00
|
|
|
CPU_FTR_ALTIVEC_COMP | \
|
2006-06-07 14:14:40 +08:00
|
|
|
CPU_FTR_MAYBE_CAN_NAP | CPU_FTR_PPC_LE)
|
[POWERPC] Merge CPU features pertaining to icache coherency
Currently the powerpc kernel has a 64-bit only feature,
COHERENT_ICACHE used for those CPUS which maintain icache/dcache
coherency in hardware (POWER5, essentially). It also has a feature,
SPLIT_ID_CACHE, which is used on CPUs which have separate i and
d-caches, which is to say everything except 601 and Freescale E200.
In nearly all the places we check the SPLIT_ID_CACHE, what we actually
care about is whether the i and d-caches are coherent (which they will
be, trivially, if they're the same cache).
This tries to clarify the situation a little. The COHERENT_ICACHE
feature becomes availble on 32-bit and is set for all CPUs where i and
d-cache are effectively coherent, whether this is due to special logic
(POWER5) or because they're unified. We check this, instead of
SPLIT_ID_CACHE nearly everywhere.
The SPLIT_ID_CACHE feature itself is replaced by a UNIFIED_ID_CACHE
feature with reversed sense, set only on 601 and Freescale E200. In
the two places (one Freescale BookE specific) where we really care
whether it's a unified cache, not whether they're coherent, we check
this feature. The CPUs with unified cache are so few, we could
consider replacing this feature bit with explicit checks against the
PVR.
This will make unifying the 32-bit and 64-bit cache flush code a
little more straightforward.
Signed-off-by: David Gibson <dwg@au1.ibm.com>
Signed-off-by: Paul Mackerras <paulus@samba.org>
2007-06-13 12:52:57 +08:00
|
|
|
#define CPU_FTRS_7400 (CPU_FTR_COMMON | \
|
2018-03-20 05:46:11 +08:00
|
|
|
CPU_FTR_MAYBE_CAN_DOZE | CPU_FTR_L2CR | \
|
2008-12-19 03:13:32 +08:00
|
|
|
CPU_FTR_TAU | CPU_FTR_ALTIVEC_COMP | \
|
2006-06-07 14:14:40 +08:00
|
|
|
CPU_FTR_MAYBE_CAN_NAP | CPU_FTR_PPC_LE)
|
[POWERPC] Merge CPU features pertaining to icache coherency
Currently the powerpc kernel has a 64-bit only feature,
COHERENT_ICACHE used for those CPUS which maintain icache/dcache
coherency in hardware (POWER5, essentially). It also has a feature,
SPLIT_ID_CACHE, which is used on CPUs which have separate i and
d-caches, which is to say everything except 601 and Freescale E200.
In nearly all the places we check the SPLIT_ID_CACHE, what we actually
care about is whether the i and d-caches are coherent (which they will
be, trivially, if they're the same cache).
This tries to clarify the situation a little. The COHERENT_ICACHE
feature becomes availble on 32-bit and is set for all CPUs where i and
d-cache are effectively coherent, whether this is due to special logic
(POWER5) or because they're unified. We check this, instead of
SPLIT_ID_CACHE nearly everywhere.
The SPLIT_ID_CACHE feature itself is replaced by a UNIFIED_ID_CACHE
feature with reversed sense, set only on 601 and Freescale E200. In
the two places (one Freescale BookE specific) where we really care
whether it's a unified cache, not whether they're coherent, we check
this feature. The CPUs with unified cache are so few, we could
consider replacing this feature bit with explicit checks against the
PVR.
This will make unifying the 32-bit and 64-bit cache flush code a
little more straightforward.
Signed-off-by: David Gibson <dwg@au1.ibm.com>
Signed-off-by: Paul Mackerras <paulus@samba.org>
2007-06-13 12:52:57 +08:00
|
|
|
#define CPU_FTRS_7450_20 (CPU_FTR_COMMON | \
|
2018-03-20 05:46:11 +08:00
|
|
|
CPU_FTR_L2CR | CPU_FTR_ALTIVEC_COMP | \
|
2008-12-19 03:13:32 +08:00
|
|
|
CPU_FTR_L3CR | CPU_FTR_SPEC7450 | \
|
2007-11-10 06:17:49 +08:00
|
|
|
CPU_FTR_NEED_COHERENT | CPU_FTR_PPC_LE | CPU_FTR_NEED_PAIRED_STWCX)
|
[POWERPC] Merge CPU features pertaining to icache coherency
Currently the powerpc kernel has a 64-bit only feature,
COHERENT_ICACHE used for those CPUS which maintain icache/dcache
coherency in hardware (POWER5, essentially). It also has a feature,
SPLIT_ID_CACHE, which is used on CPUs which have separate i and
d-caches, which is to say everything except 601 and Freescale E200.
In nearly all the places we check the SPLIT_ID_CACHE, what we actually
care about is whether the i and d-caches are coherent (which they will
be, trivially, if they're the same cache).
This tries to clarify the situation a little. The COHERENT_ICACHE
feature becomes availble on 32-bit and is set for all CPUs where i and
d-cache are effectively coherent, whether this is due to special logic
(POWER5) or because they're unified. We check this, instead of
SPLIT_ID_CACHE nearly everywhere.
The SPLIT_ID_CACHE feature itself is replaced by a UNIFIED_ID_CACHE
feature with reversed sense, set only on 601 and Freescale E200. In
the two places (one Freescale BookE specific) where we really care
whether it's a unified cache, not whether they're coherent, we check
this feature. The CPUs with unified cache are so few, we could
consider replacing this feature bit with explicit checks against the
PVR.
This will make unifying the 32-bit and 64-bit cache flush code a
little more straightforward.
Signed-off-by: David Gibson <dwg@au1.ibm.com>
Signed-off-by: Paul Mackerras <paulus@samba.org>
2007-06-13 12:52:57 +08:00
|
|
|
#define CPU_FTRS_7450_21 (CPU_FTR_COMMON | \
|
2006-03-23 14:36:59 +08:00
|
|
|
CPU_FTR_MAYBE_CAN_NAP | CPU_FTR_L2CR | CPU_FTR_ALTIVEC_COMP | \
|
2008-12-19 03:13:32 +08:00
|
|
|
CPU_FTR_L3CR | CPU_FTR_SPEC7450 | \
|
2006-03-23 14:36:59 +08:00
|
|
|
CPU_FTR_NAP_DISABLE_L2_PR | CPU_FTR_L3_DISABLE_NAP | \
|
2007-11-10 06:17:49 +08:00
|
|
|
CPU_FTR_NEED_COHERENT | CPU_FTR_PPC_LE | CPU_FTR_NEED_PAIRED_STWCX)
|
[POWERPC] Merge CPU features pertaining to icache coherency
Currently the powerpc kernel has a 64-bit only feature,
COHERENT_ICACHE used for those CPUS which maintain icache/dcache
coherency in hardware (POWER5, essentially). It also has a feature,
SPLIT_ID_CACHE, which is used on CPUs which have separate i and
d-caches, which is to say everything except 601 and Freescale E200.
In nearly all the places we check the SPLIT_ID_CACHE, what we actually
care about is whether the i and d-caches are coherent (which they will
be, trivially, if they're the same cache).
This tries to clarify the situation a little. The COHERENT_ICACHE
feature becomes availble on 32-bit and is set for all CPUs where i and
d-cache are effectively coherent, whether this is due to special logic
(POWER5) or because they're unified. We check this, instead of
SPLIT_ID_CACHE nearly everywhere.
The SPLIT_ID_CACHE feature itself is replaced by a UNIFIED_ID_CACHE
feature with reversed sense, set only on 601 and Freescale E200. In
the two places (one Freescale BookE specific) where we really care
whether it's a unified cache, not whether they're coherent, we check
this feature. The CPUs with unified cache are so few, we could
consider replacing this feature bit with explicit checks against the
PVR.
This will make unifying the 32-bit and 64-bit cache flush code a
little more straightforward.
Signed-off-by: David Gibson <dwg@au1.ibm.com>
Signed-off-by: Paul Mackerras <paulus@samba.org>
2007-06-13 12:52:57 +08:00
|
|
|
#define CPU_FTRS_7450_23 (CPU_FTR_COMMON | \
|
2018-03-20 05:46:11 +08:00
|
|
|
CPU_FTR_NEED_PAIRED_STWCX | \
|
2006-03-23 14:36:59 +08:00
|
|
|
CPU_FTR_MAYBE_CAN_NAP | CPU_FTR_L2CR | CPU_FTR_ALTIVEC_COMP | \
|
2008-12-19 03:13:32 +08:00
|
|
|
CPU_FTR_L3CR | CPU_FTR_SPEC7450 | \
|
2006-06-07 14:14:40 +08:00
|
|
|
CPU_FTR_NAP_DISABLE_L2_PR | CPU_FTR_NEED_COHERENT | CPU_FTR_PPC_LE)
|
[POWERPC] Merge CPU features pertaining to icache coherency
Currently the powerpc kernel has a 64-bit only feature,
COHERENT_ICACHE used for those CPUS which maintain icache/dcache
coherency in hardware (POWER5, essentially). It also has a feature,
SPLIT_ID_CACHE, which is used on CPUs which have separate i and
d-caches, which is to say everything except 601 and Freescale E200.
In nearly all the places we check the SPLIT_ID_CACHE, what we actually
care about is whether the i and d-caches are coherent (which they will
be, trivially, if they're the same cache).
This tries to clarify the situation a little. The COHERENT_ICACHE
feature becomes availble on 32-bit and is set for all CPUs where i and
d-cache are effectively coherent, whether this is due to special logic
(POWER5) or because they're unified. We check this, instead of
SPLIT_ID_CACHE nearly everywhere.
The SPLIT_ID_CACHE feature itself is replaced by a UNIFIED_ID_CACHE
feature with reversed sense, set only on 601 and Freescale E200. In
the two places (one Freescale BookE specific) where we really care
whether it's a unified cache, not whether they're coherent, we check
this feature. The CPUs with unified cache are so few, we could
consider replacing this feature bit with explicit checks against the
PVR.
This will make unifying the 32-bit and 64-bit cache flush code a
little more straightforward.
Signed-off-by: David Gibson <dwg@au1.ibm.com>
Signed-off-by: Paul Mackerras <paulus@samba.org>
2007-06-13 12:52:57 +08:00
|
|
|
#define CPU_FTRS_7455_1 (CPU_FTR_COMMON | \
|
2018-03-20 05:46:11 +08:00
|
|
|
CPU_FTR_NEED_PAIRED_STWCX | \
|
2006-03-23 14:36:59 +08:00
|
|
|
CPU_FTR_L2CR | CPU_FTR_ALTIVEC_COMP | CPU_FTR_L3CR | \
|
2008-12-19 03:13:32 +08:00
|
|
|
CPU_FTR_SPEC7450 | CPU_FTR_NEED_COHERENT | CPU_FTR_PPC_LE)
|
[POWERPC] Merge CPU features pertaining to icache coherency
Currently the powerpc kernel has a 64-bit only feature,
COHERENT_ICACHE used for those CPUS which maintain icache/dcache
coherency in hardware (POWER5, essentially). It also has a feature,
SPLIT_ID_CACHE, which is used on CPUs which have separate i and
d-caches, which is to say everything except 601 and Freescale E200.
In nearly all the places we check the SPLIT_ID_CACHE, what we actually
care about is whether the i and d-caches are coherent (which they will
be, trivially, if they're the same cache).
This tries to clarify the situation a little. The COHERENT_ICACHE
feature becomes availble on 32-bit and is set for all CPUs where i and
d-cache are effectively coherent, whether this is due to special logic
(POWER5) or because they're unified. We check this, instead of
SPLIT_ID_CACHE nearly everywhere.
The SPLIT_ID_CACHE feature itself is replaced by a UNIFIED_ID_CACHE
feature with reversed sense, set only on 601 and Freescale E200. In
the two places (one Freescale BookE specific) where we really care
whether it's a unified cache, not whether they're coherent, we check
this feature. The CPUs with unified cache are so few, we could
consider replacing this feature bit with explicit checks against the
PVR.
This will make unifying the 32-bit and 64-bit cache flush code a
little more straightforward.
Signed-off-by: David Gibson <dwg@au1.ibm.com>
Signed-off-by: Paul Mackerras <paulus@samba.org>
2007-06-13 12:52:57 +08:00
|
|
|
#define CPU_FTRS_7455_20 (CPU_FTR_COMMON | \
|
2018-03-20 05:46:11 +08:00
|
|
|
CPU_FTR_NEED_PAIRED_STWCX | \
|
2006-03-23 14:36:59 +08:00
|
|
|
CPU_FTR_MAYBE_CAN_NAP | CPU_FTR_L2CR | CPU_FTR_ALTIVEC_COMP | \
|
2008-12-19 03:13:32 +08:00
|
|
|
CPU_FTR_L3CR | CPU_FTR_SPEC7450 | \
|
2006-03-23 14:36:59 +08:00
|
|
|
CPU_FTR_NAP_DISABLE_L2_PR | CPU_FTR_L3_DISABLE_NAP | \
|
2008-12-19 03:13:32 +08:00
|
|
|
CPU_FTR_NEED_COHERENT | CPU_FTR_PPC_LE)
|
[POWERPC] Merge CPU features pertaining to icache coherency
Currently the powerpc kernel has a 64-bit only feature,
COHERENT_ICACHE used for those CPUS which maintain icache/dcache
coherency in hardware (POWER5, essentially). It also has a feature,
SPLIT_ID_CACHE, which is used on CPUs which have separate i and
d-caches, which is to say everything except 601 and Freescale E200.
In nearly all the places we check the SPLIT_ID_CACHE, what we actually
care about is whether the i and d-caches are coherent (which they will
be, trivially, if they're the same cache).
This tries to clarify the situation a little. The COHERENT_ICACHE
feature becomes availble on 32-bit and is set for all CPUs where i and
d-cache are effectively coherent, whether this is due to special logic
(POWER5) or because they're unified. We check this, instead of
SPLIT_ID_CACHE nearly everywhere.
The SPLIT_ID_CACHE feature itself is replaced by a UNIFIED_ID_CACHE
feature with reversed sense, set only on 601 and Freescale E200. In
the two places (one Freescale BookE specific) where we really care
whether it's a unified cache, not whether they're coherent, we check
this feature. The CPUs with unified cache are so few, we could
consider replacing this feature bit with explicit checks against the
PVR.
This will make unifying the 32-bit and 64-bit cache flush code a
little more straightforward.
Signed-off-by: David Gibson <dwg@au1.ibm.com>
Signed-off-by: Paul Mackerras <paulus@samba.org>
2007-06-13 12:52:57 +08:00
|
|
|
#define CPU_FTRS_7455 (CPU_FTR_COMMON | \
|
2006-03-23 14:36:59 +08:00
|
|
|
CPU_FTR_MAYBE_CAN_NAP | CPU_FTR_L2CR | CPU_FTR_ALTIVEC_COMP | \
|
2008-12-19 03:13:32 +08:00
|
|
|
CPU_FTR_L3CR | CPU_FTR_SPEC7450 | CPU_FTR_NAP_DISABLE_L2_PR | \
|
2007-11-10 06:17:49 +08:00
|
|
|
CPU_FTR_NEED_COHERENT | CPU_FTR_PPC_LE | CPU_FTR_NEED_PAIRED_STWCX)
|
[POWERPC] Merge CPU features pertaining to icache coherency
Currently the powerpc kernel has a 64-bit only feature,
COHERENT_ICACHE used for those CPUS which maintain icache/dcache
coherency in hardware (POWER5, essentially). It also has a feature,
SPLIT_ID_CACHE, which is used on CPUs which have separate i and
d-caches, which is to say everything except 601 and Freescale E200.
In nearly all the places we check the SPLIT_ID_CACHE, what we actually
care about is whether the i and d-caches are coherent (which they will
be, trivially, if they're the same cache).
This tries to clarify the situation a little. The COHERENT_ICACHE
feature becomes availble on 32-bit and is set for all CPUs where i and
d-cache are effectively coherent, whether this is due to special logic
(POWER5) or because they're unified. We check this, instead of
SPLIT_ID_CACHE nearly everywhere.
The SPLIT_ID_CACHE feature itself is replaced by a UNIFIED_ID_CACHE
feature with reversed sense, set only on 601 and Freescale E200. In
the two places (one Freescale BookE specific) where we really care
whether it's a unified cache, not whether they're coherent, we check
this feature. The CPUs with unified cache are so few, we could
consider replacing this feature bit with explicit checks against the
PVR.
This will make unifying the 32-bit and 64-bit cache flush code a
little more straightforward.
Signed-off-by: David Gibson <dwg@au1.ibm.com>
Signed-off-by: Paul Mackerras <paulus@samba.org>
2007-06-13 12:52:57 +08:00
|
|
|
#define CPU_FTRS_7447_10 (CPU_FTR_COMMON | \
|
2006-03-23 14:36:59 +08:00
|
|
|
CPU_FTR_MAYBE_CAN_NAP | CPU_FTR_L2CR | CPU_FTR_ALTIVEC_COMP | \
|
2008-12-19 03:13:32 +08:00
|
|
|
CPU_FTR_L3CR | CPU_FTR_SPEC7450 | CPU_FTR_NAP_DISABLE_L2_PR | \
|
2007-11-10 06:17:49 +08:00
|
|
|
CPU_FTR_NEED_COHERENT | CPU_FTR_NO_BTIC | CPU_FTR_PPC_LE | \
|
|
|
|
CPU_FTR_NEED_PAIRED_STWCX)
|
[POWERPC] Merge CPU features pertaining to icache coherency
Currently the powerpc kernel has a 64-bit only feature,
COHERENT_ICACHE used for those CPUS which maintain icache/dcache
coherency in hardware (POWER5, essentially). It also has a feature,
SPLIT_ID_CACHE, which is used on CPUs which have separate i and
d-caches, which is to say everything except 601 and Freescale E200.
In nearly all the places we check the SPLIT_ID_CACHE, what we actually
care about is whether the i and d-caches are coherent (which they will
be, trivially, if they're the same cache).
This tries to clarify the situation a little. The COHERENT_ICACHE
feature becomes availble on 32-bit and is set for all CPUs where i and
d-cache are effectively coherent, whether this is due to special logic
(POWER5) or because they're unified. We check this, instead of
SPLIT_ID_CACHE nearly everywhere.
The SPLIT_ID_CACHE feature itself is replaced by a UNIFIED_ID_CACHE
feature with reversed sense, set only on 601 and Freescale E200. In
the two places (one Freescale BookE specific) where we really care
whether it's a unified cache, not whether they're coherent, we check
this feature. The CPUs with unified cache are so few, we could
consider replacing this feature bit with explicit checks against the
PVR.
This will make unifying the 32-bit and 64-bit cache flush code a
little more straightforward.
Signed-off-by: David Gibson <dwg@au1.ibm.com>
Signed-off-by: Paul Mackerras <paulus@samba.org>
2007-06-13 12:52:57 +08:00
|
|
|
#define CPU_FTRS_7447 (CPU_FTR_COMMON | \
|
2006-03-23 14:36:59 +08:00
|
|
|
CPU_FTR_MAYBE_CAN_NAP | CPU_FTR_L2CR | CPU_FTR_ALTIVEC_COMP | \
|
2008-12-19 03:13:32 +08:00
|
|
|
CPU_FTR_L3CR | CPU_FTR_SPEC7450 | CPU_FTR_NAP_DISABLE_L2_PR | \
|
2007-11-10 06:17:49 +08:00
|
|
|
CPU_FTR_NEED_COHERENT | CPU_FTR_PPC_LE | CPU_FTR_NEED_PAIRED_STWCX)
|
[POWERPC] Merge CPU features pertaining to icache coherency
Currently the powerpc kernel has a 64-bit only feature,
COHERENT_ICACHE used for those CPUS which maintain icache/dcache
coherency in hardware (POWER5, essentially). It also has a feature,
SPLIT_ID_CACHE, which is used on CPUs which have separate i and
d-caches, which is to say everything except 601 and Freescale E200.
In nearly all the places we check the SPLIT_ID_CACHE, what we actually
care about is whether the i and d-caches are coherent (which they will
be, trivially, if they're the same cache).
This tries to clarify the situation a little. The COHERENT_ICACHE
feature becomes availble on 32-bit and is set for all CPUs where i and
d-cache are effectively coherent, whether this is due to special logic
(POWER5) or because they're unified. We check this, instead of
SPLIT_ID_CACHE nearly everywhere.
The SPLIT_ID_CACHE feature itself is replaced by a UNIFIED_ID_CACHE
feature with reversed sense, set only on 601 and Freescale E200. In
the two places (one Freescale BookE specific) where we really care
whether it's a unified cache, not whether they're coherent, we check
this feature. The CPUs with unified cache are so few, we could
consider replacing this feature bit with explicit checks against the
PVR.
This will make unifying the 32-bit and 64-bit cache flush code a
little more straightforward.
Signed-off-by: David Gibson <dwg@au1.ibm.com>
Signed-off-by: Paul Mackerras <paulus@samba.org>
2007-06-13 12:52:57 +08:00
|
|
|
#define CPU_FTRS_7447A (CPU_FTR_COMMON | \
|
2006-03-23 14:36:59 +08:00
|
|
|
CPU_FTR_MAYBE_CAN_NAP | CPU_FTR_L2CR | CPU_FTR_ALTIVEC_COMP | \
|
2008-12-19 03:13:32 +08:00
|
|
|
CPU_FTR_SPEC7450 | CPU_FTR_NAP_DISABLE_L2_PR | \
|
2007-11-10 06:17:49 +08:00
|
|
|
CPU_FTR_NEED_COHERENT | CPU_FTR_PPC_LE | CPU_FTR_NEED_PAIRED_STWCX)
|
[POWERPC] Merge CPU features pertaining to icache coherency
Currently the powerpc kernel has a 64-bit only feature,
COHERENT_ICACHE used for those CPUS which maintain icache/dcache
coherency in hardware (POWER5, essentially). It also has a feature,
SPLIT_ID_CACHE, which is used on CPUs which have separate i and
d-caches, which is to say everything except 601 and Freescale E200.
In nearly all the places we check the SPLIT_ID_CACHE, what we actually
care about is whether the i and d-caches are coherent (which they will
be, trivially, if they're the same cache).
This tries to clarify the situation a little. The COHERENT_ICACHE
feature becomes availble on 32-bit and is set for all CPUs where i and
d-cache are effectively coherent, whether this is due to special logic
(POWER5) or because they're unified. We check this, instead of
SPLIT_ID_CACHE nearly everywhere.
The SPLIT_ID_CACHE feature itself is replaced by a UNIFIED_ID_CACHE
feature with reversed sense, set only on 601 and Freescale E200. In
the two places (one Freescale BookE specific) where we really care
whether it's a unified cache, not whether they're coherent, we check
this feature. The CPUs with unified cache are so few, we could
consider replacing this feature bit with explicit checks against the
PVR.
This will make unifying the 32-bit and 64-bit cache flush code a
little more straightforward.
Signed-off-by: David Gibson <dwg@au1.ibm.com>
Signed-off-by: Paul Mackerras <paulus@samba.org>
2007-06-13 12:52:57 +08:00
|
|
|
#define CPU_FTRS_7448 (CPU_FTR_COMMON | \
|
2007-05-03 05:34:43 +08:00
|
|
|
CPU_FTR_MAYBE_CAN_NAP | CPU_FTR_L2CR | CPU_FTR_ALTIVEC_COMP | \
|
2008-12-19 03:13:32 +08:00
|
|
|
CPU_FTR_SPEC7450 | CPU_FTR_NAP_DISABLE_L2_PR | \
|
2007-11-10 06:17:49 +08:00
|
|
|
CPU_FTR_PPC_LE | CPU_FTR_NEED_PAIRED_STWCX)
|
2018-11-29 01:21:10 +08:00
|
|
|
#define CPU_FTRS_82XX (CPU_FTR_COMMON | CPU_FTR_MAYBE_CAN_DOZE | CPU_FTR_NOEXECUTE)
|
2007-09-15 04:32:14 +08:00
|
|
|
#define CPU_FTRS_G2_LE (CPU_FTR_COMMON | CPU_FTR_MAYBE_CAN_DOZE | \
|
2018-03-20 05:46:11 +08:00
|
|
|
CPU_FTR_MAYBE_CAN_NAP)
|
[POWERPC] Merge CPU features pertaining to icache coherency
Currently the powerpc kernel has a 64-bit only feature,
COHERENT_ICACHE used for those CPUS which maintain icache/dcache
coherency in hardware (POWER5, essentially). It also has a feature,
SPLIT_ID_CACHE, which is used on CPUs which have separate i and
d-caches, which is to say everything except 601 and Freescale E200.
In nearly all the places we check the SPLIT_ID_CACHE, what we actually
care about is whether the i and d-caches are coherent (which they will
be, trivially, if they're the same cache).
This tries to clarify the situation a little. The COHERENT_ICACHE
feature becomes availble on 32-bit and is set for all CPUs where i and
d-cache are effectively coherent, whether this is due to special logic
(POWER5) or because they're unified. We check this, instead of
SPLIT_ID_CACHE nearly everywhere.
The SPLIT_ID_CACHE feature itself is replaced by a UNIFIED_ID_CACHE
feature with reversed sense, set only on 601 and Freescale E200. In
the two places (one Freescale BookE specific) where we really care
whether it's a unified cache, not whether they're coherent, we check
this feature. The CPUs with unified cache are so few, we could
consider replacing this feature bit with explicit checks against the
PVR.
This will make unifying the 32-bit and 64-bit cache flush code a
little more straightforward.
Signed-off-by: David Gibson <dwg@au1.ibm.com>
Signed-off-by: Paul Mackerras <paulus@samba.org>
2007-06-13 12:52:57 +08:00
|
|
|
#define CPU_FTRS_E300 (CPU_FTR_MAYBE_CAN_DOZE | \
|
2018-03-20 05:46:11 +08:00
|
|
|
CPU_FTR_MAYBE_CAN_NAP | \
|
2018-11-29 01:21:10 +08:00
|
|
|
CPU_FTR_COMMON | CPU_FTR_NOEXECUTE)
|
[POWERPC] Merge CPU features pertaining to icache coherency
Currently the powerpc kernel has a 64-bit only feature,
COHERENT_ICACHE used for those CPUS which maintain icache/dcache
coherency in hardware (POWER5, essentially). It also has a feature,
SPLIT_ID_CACHE, which is used on CPUs which have separate i and
d-caches, which is to say everything except 601 and Freescale E200.
In nearly all the places we check the SPLIT_ID_CACHE, what we actually
care about is whether the i and d-caches are coherent (which they will
be, trivially, if they're the same cache).
This tries to clarify the situation a little. The COHERENT_ICACHE
feature becomes availble on 32-bit and is set for all CPUs where i and
d-cache are effectively coherent, whether this is due to special logic
(POWER5) or because they're unified. We check this, instead of
SPLIT_ID_CACHE nearly everywhere.
The SPLIT_ID_CACHE feature itself is replaced by a UNIFIED_ID_CACHE
feature with reversed sense, set only on 601 and Freescale E200. In
the two places (one Freescale BookE specific) where we really care
whether it's a unified cache, not whether they're coherent, we check
this feature. The CPUs with unified cache are so few, we could
consider replacing this feature bit with explicit checks against the
PVR.
This will make unifying the 32-bit and 64-bit cache flush code a
little more straightforward.
Signed-off-by: David Gibson <dwg@au1.ibm.com>
Signed-off-by: Paul Mackerras <paulus@samba.org>
2007-06-13 12:52:57 +08:00
|
|
|
#define CPU_FTRS_E300C2 (CPU_FTR_MAYBE_CAN_DOZE | \
|
2018-03-20 05:46:11 +08:00
|
|
|
CPU_FTR_MAYBE_CAN_NAP | \
|
2018-11-29 01:21:10 +08:00
|
|
|
CPU_FTR_COMMON | CPU_FTR_FPU_UNAVAILABLE | CPU_FTR_NOEXECUTE)
|
2018-03-20 05:46:11 +08:00
|
|
|
#define CPU_FTRS_CLASSIC32 (CPU_FTR_COMMON)
|
|
|
|
#define CPU_FTRS_8XX (CPU_FTR_NOEXECUTE)
|
|
|
|
#define CPU_FTRS_40X (CPU_FTR_NODSISRALIGN | CPU_FTR_NOEXECUTE)
|
|
|
|
#define CPU_FTRS_44X (CPU_FTR_NODSISRALIGN | CPU_FTR_NOEXECUTE)
|
|
|
|
#define CPU_FTRS_440x6 (CPU_FTR_NODSISRALIGN | CPU_FTR_NOEXECUTE | \
|
2008-12-19 03:13:22 +08:00
|
|
|
CPU_FTR_INDEXED_DCR)
|
2010-03-05 18:43:12 +08:00
|
|
|
#define CPU_FTRS_47X (CPU_FTRS_440x6)
|
2018-03-20 05:46:11 +08:00
|
|
|
#define CPU_FTRS_E200 (CPU_FTR_SPE_COMP | \
|
2007-09-13 14:44:20 +08:00
|
|
|
CPU_FTR_NODSISRALIGN | CPU_FTR_COHERENT_ICACHE | \
|
2019-08-26 23:52:18 +08:00
|
|
|
CPU_FTR_NOEXECUTE | \
|
2011-12-20 23:34:12 +08:00
|
|
|
CPU_FTR_DEBUG_LVL_EXC)
|
2018-03-20 05:46:11 +08:00
|
|
|
#define CPU_FTRS_E500 (CPU_FTR_MAYBE_CAN_DOZE | \
|
2008-12-12 14:33:25 +08:00
|
|
|
CPU_FTR_SPE_COMP | CPU_FTR_MAYBE_CAN_NAP | CPU_FTR_NODSISRALIGN | \
|
|
|
|
CPU_FTR_NOEXECUTE)
|
2018-03-20 05:46:11 +08:00
|
|
|
#define CPU_FTRS_E500_2 (CPU_FTR_MAYBE_CAN_DOZE | \
|
2008-12-19 03:13:32 +08:00
|
|
|
CPU_FTR_SPE_COMP | CPU_FTR_MAYBE_CAN_NAP | \
|
2008-12-12 14:33:25 +08:00
|
|
|
CPU_FTR_NODSISRALIGN | CPU_FTR_NOEXECUTE)
|
2018-03-20 05:46:11 +08:00
|
|
|
#define CPU_FTRS_E500MC (CPU_FTR_NODSISRALIGN | \
|
2018-03-20 05:46:12 +08:00
|
|
|
CPU_FTR_LWSYNC | CPU_FTR_NOEXECUTE | \
|
2011-12-20 23:34:47 +08:00
|
|
|
CPU_FTR_DBELL | CPU_FTR_DEBUG_LVL_EXC | CPU_FTR_EMB_HV)
|
2013-07-24 09:21:11 +08:00
|
|
|
/*
|
|
|
|
* e5500/e6500 erratum A-006958 is a timebase bug that can use the
|
|
|
|
* same workaround as CPU_FTR_CELL_TB_BUG.
|
|
|
|
*/
|
2018-03-20 05:46:11 +08:00
|
|
|
#define CPU_FTRS_E5500 (CPU_FTR_NODSISRALIGN | \
|
2018-03-20 05:46:12 +08:00
|
|
|
CPU_FTR_LWSYNC | CPU_FTR_NOEXECUTE | \
|
2011-04-06 13:18:48 +08:00
|
|
|
CPU_FTR_DBELL | CPU_FTR_POPCNTB | CPU_FTR_POPCNTD | \
|
2013-07-24 09:21:11 +08:00
|
|
|
CPU_FTR_DEBUG_LVL_EXC | CPU_FTR_EMB_HV | CPU_FTR_CELL_TB_BUG)
|
2018-03-20 05:46:11 +08:00
|
|
|
#define CPU_FTRS_E6500 (CPU_FTR_NODSISRALIGN | \
|
2018-03-20 05:46:12 +08:00
|
|
|
CPU_FTR_LWSYNC | CPU_FTR_NOEXECUTE | \
|
2011-11-07 01:51:07 +08:00
|
|
|
CPU_FTR_DBELL | CPU_FTR_POPCNTB | CPU_FTR_POPCNTD | \
|
2013-07-24 09:21:11 +08:00
|
|
|
CPU_FTR_DEBUG_LVL_EXC | CPU_FTR_EMB_HV | CPU_FTR_ALTIVEC_COMP | \
|
2011-12-08 15:20:27 +08:00
|
|
|
CPU_FTR_CELL_TB_BUG | CPU_FTR_SMT)
|
2006-03-23 14:36:59 +08:00
|
|
|
#define CPU_FTRS_GENERIC_32 (CPU_FTR_COMMON | CPU_FTR_NODSISRALIGN)
|
2006-11-23 07:46:46 +08:00
|
|
|
|
|
|
|
/* 64-bit CPUs */
|
2018-03-20 05:46:11 +08:00
|
|
|
#define CPU_FTRS_PPC970 (CPU_FTR_LWSYNC | \
|
2018-02-21 03:08:28 +08:00
|
|
|
CPU_FTR_PPCAS_ARCH_V2 | CPU_FTR_CTRL | \
|
2008-08-22 12:36:19 +08:00
|
|
|
CPU_FTR_ALTIVEC_COMP | CPU_FTR_CAN_NAP | CPU_FTR_MMCRA | \
|
powerpc, KVM: Split HVMODE_206 cpu feature bit into separate HV and architecture bits
This replaces the single CPU_FTR_HVMODE_206 bit with two bits, one to
indicate that we have a usable hypervisor mode, and another to indicate
that the processor conforms to PowerISA version 2.06. We also add
another bit to indicate that the processor conforms to ISA version 2.01
and set that for PPC970 and derivatives.
Some PPC970 chips (specifically those in Apple machines) have a
hypervisor mode in that MSR[HV] is always 1, but the hypervisor mode
is not useful in the sense that there is no way to run any code in
supervisor mode (HV=0 PR=0). On these processors, the LPES0 and LPES1
bits in HID4 are always 0, and we use that as a way of detecting that
hypervisor mode is not useful.
Where we have a feature section in assembly code around code that
only applies on POWER7 in hypervisor mode, we use a construct like
END_FTR_SECTION_IFSET(CPU_FTR_HVMODE | CPU_FTR_ARCH_206)
The definition of END_FTR_SECTION_IFSET is such that the code will
be enabled (not overwritten with nops) only if all bits in the
provided mask are set.
Note that the CPU feature check in __tlbie() only needs to check the
ARCH_206 bit, not the HVMODE bit, because __tlbie() can only get called
if we are running bare-metal, i.e. in hypervisor mode.
Signed-off-by: Paul Mackerras <paulus@samba.org>
Signed-off-by: Alexander Graf <agraf@suse.de>
2011-06-29 08:26:11 +08:00
|
|
|
CPU_FTR_CP_USE_DCBTZ | CPU_FTR_STCX_CHECKS_ADDRESS | \
|
2013-05-17 04:27:31 +08:00
|
|
|
CPU_FTR_HVMODE | CPU_FTR_DABRX)
|
2018-03-20 05:46:11 +08:00
|
|
|
#define CPU_FTRS_POWER5 (CPU_FTR_LWSYNC | \
|
2008-12-19 03:13:32 +08:00
|
|
|
CPU_FTR_PPCAS_ARCH_V2 | CPU_FTR_CTRL | \
|
2006-03-23 14:36:59 +08:00
|
|
|
CPU_FTR_MMCRA | CPU_FTR_SMT | \
|
2011-04-07 03:48:50 +08:00
|
|
|
CPU_FTR_COHERENT_ICACHE | CPU_FTR_PURR | \
|
2013-05-17 04:27:31 +08:00
|
|
|
CPU_FTR_STCX_CHECKS_ADDRESS | CPU_FTR_POPCNTB | CPU_FTR_DABRX)
|
2018-03-20 05:46:11 +08:00
|
|
|
#define CPU_FTRS_POWER6 (CPU_FTR_LWSYNC | \
|
2008-12-19 03:13:32 +08:00
|
|
|
CPU_FTR_PPCAS_ARCH_V2 | CPU_FTR_CTRL | \
|
2006-04-29 07:51:06 +08:00
|
|
|
CPU_FTR_MMCRA | CPU_FTR_SMT | \
|
2011-04-07 03:48:50 +08:00
|
|
|
CPU_FTR_COHERENT_ICACHE | \
|
2006-12-08 14:46:58 +08:00
|
|
|
CPU_FTR_PURR | CPU_FTR_SPURR | CPU_FTR_REAL_LE | \
|
2010-08-11 09:40:27 +08:00
|
|
|
CPU_FTR_DSCR | CPU_FTR_UNALIGNED_LD_STD | \
|
2013-05-17 04:27:31 +08:00
|
|
|
CPU_FTR_STCX_CHECKS_ADDRESS | CPU_FTR_POPCNTB | CPU_FTR_CFAR | \
|
|
|
|
CPU_FTR_DABRX)
|
2018-03-20 05:46:11 +08:00
|
|
|
#define CPU_FTRS_POWER7 (CPU_FTR_LWSYNC | \
|
powerpc, KVM: Split HVMODE_206 cpu feature bit into separate HV and architecture bits
This replaces the single CPU_FTR_HVMODE_206 bit with two bits, one to
indicate that we have a usable hypervisor mode, and another to indicate
that the processor conforms to PowerISA version 2.06. We also add
another bit to indicate that the processor conforms to ISA version 2.01
and set that for PPC970 and derivatives.
Some PPC970 chips (specifically those in Apple machines) have a
hypervisor mode in that MSR[HV] is always 1, but the hypervisor mode
is not useful in the sense that there is no way to run any code in
supervisor mode (HV=0 PR=0). On these processors, the LPES0 and LPES1
bits in HID4 are always 0, and we use that as a way of detecting that
hypervisor mode is not useful.
Where we have a feature section in assembly code around code that
only applies on POWER7 in hypervisor mode, we use a construct like
END_FTR_SECTION_IFSET(CPU_FTR_HVMODE | CPU_FTR_ARCH_206)
The definition of END_FTR_SECTION_IFSET is such that the code will
be enabled (not overwritten with nops) only if all bits in the
provided mask are set.
Note that the CPU feature check in __tlbie() only needs to check the
ARCH_206 bit, not the HVMODE bit, because __tlbie() can only get called
if we are running bare-metal, i.e. in hypervisor mode.
Signed-off-by: Paul Mackerras <paulus@samba.org>
Signed-off-by: Alexander Graf <agraf@suse.de>
2011-06-29 08:26:11 +08:00
|
|
|
CPU_FTR_PPCAS_ARCH_V2 | CPU_FTR_CTRL | CPU_FTR_ARCH_206 |\
|
2008-06-18 08:47:26 +08:00
|
|
|
CPU_FTR_MMCRA | CPU_FTR_SMT | \
|
2011-04-07 03:48:50 +08:00
|
|
|
CPU_FTR_COHERENT_ICACHE | \
|
2008-06-18 08:47:26 +08:00
|
|
|
CPU_FTR_PURR | CPU_FTR_SPURR | CPU_FTR_REAL_LE | \
|
2010-08-11 09:40:27 +08:00
|
|
|
CPU_FTR_DSCR | CPU_FTR_SAO | CPU_FTR_ASYM_SMT | \
|
2011-05-03 04:43:04 +08:00
|
|
|
CPU_FTR_STCX_CHECKS_ADDRESS | CPU_FTR_POPCNTB | CPU_FTR_POPCNTD | \
|
2017-10-19 12:08:19 +08:00
|
|
|
CPU_FTR_CFAR | CPU_FTR_HVMODE | \
|
2018-01-19 09:50:44 +08:00
|
|
|
CPU_FTR_VMX_COPY | CPU_FTR_HAS_PPR | CPU_FTR_DABRX | CPU_FTR_PKEY)
|
2018-03-20 05:46:11 +08:00
|
|
|
#define CPU_FTRS_POWER8 (CPU_FTR_LWSYNC | \
|
2012-10-31 03:34:15 +08:00
|
|
|
CPU_FTR_PPCAS_ARCH_V2 | CPU_FTR_CTRL | CPU_FTR_ARCH_206 |\
|
|
|
|
CPU_FTR_MMCRA | CPU_FTR_SMT | \
|
|
|
|
CPU_FTR_COHERENT_ICACHE | \
|
|
|
|
CPU_FTR_PURR | CPU_FTR_SPURR | CPU_FTR_REAL_LE | \
|
|
|
|
CPU_FTR_DSCR | CPU_FTR_SAO | \
|
|
|
|
CPU_FTR_STCX_CHECKS_ADDRESS | CPU_FTR_POPCNTB | CPU_FTR_POPCNTD | \
|
2017-10-19 12:08:19 +08:00
|
|
|
CPU_FTR_CFAR | CPU_FTR_HVMODE | CPU_FTR_VMX_COPY | \
|
2013-05-01 04:17:02 +08:00
|
|
|
CPU_FTR_DBELL | CPU_FTR_HAS_PPR | CPU_FTR_DAWR | \
|
2018-01-19 09:50:44 +08:00
|
|
|
CPU_FTR_ARCH_207S | CPU_FTR_TM_COMP | CPU_FTR_PKEY)
|
2014-03-14 13:00:28 +08:00
|
|
|
#define CPU_FTRS_POWER8E (CPU_FTRS_POWER8 | CPU_FTR_PMAO_BUG)
|
2018-03-20 05:46:11 +08:00
|
|
|
#define CPU_FTRS_POWER9 (CPU_FTR_LWSYNC | \
|
2016-02-19 08:16:24 +08:00
|
|
|
CPU_FTR_PPCAS_ARCH_V2 | CPU_FTR_CTRL | CPU_FTR_ARCH_206 |\
|
|
|
|
CPU_FTR_MMCRA | CPU_FTR_SMT | \
|
|
|
|
CPU_FTR_COHERENT_ICACHE | \
|
|
|
|
CPU_FTR_PURR | CPU_FTR_SPURR | CPU_FTR_REAL_LE | \
|
|
|
|
CPU_FTR_DSCR | CPU_FTR_SAO | \
|
|
|
|
CPU_FTR_STCX_CHECKS_ADDRESS | CPU_FTR_POPCNTB | CPU_FTR_POPCNTD | \
|
2017-04-19 10:27:37 +08:00
|
|
|
CPU_FTR_CFAR | CPU_FTR_HVMODE | CPU_FTR_VMX_COPY | \
|
2018-03-27 12:37:24 +08:00
|
|
|
CPU_FTR_DBELL | CPU_FTR_HAS_PPR | CPU_FTR_ARCH_207S | \
|
2018-03-28 19:59:50 +08:00
|
|
|
CPU_FTR_TM_COMP | CPU_FTR_ARCH_300 | CPU_FTR_PKEY | \
|
2019-09-24 11:52:53 +08:00
|
|
|
CPU_FTR_P9_TLBIE_STQ_BUG | CPU_FTR_P9_TLBIE_ERAT_BUG | CPU_FTR_P9_TIDR)
|
2017-11-15 11:25:42 +08:00
|
|
|
#define CPU_FTRS_POWER9_DD2_0 CPU_FTRS_POWER9
|
|
|
|
#define CPU_FTRS_POWER9_DD2_1 (CPU_FTRS_POWER9 | CPU_FTR_POWER9_DD2_1)
|
2018-04-05 13:57:55 +08:00
|
|
|
#define CPU_FTRS_POWER9_DD2_2 (CPU_FTRS_POWER9 | CPU_FTR_POWER9_DD2_1 | \
|
|
|
|
CPU_FTR_P9_TM_HV_ASSIST | \
|
2018-03-21 18:31:59 +08:00
|
|
|
CPU_FTR_P9_TM_XER_SO_BUG)
|
2018-03-20 05:46:11 +08:00
|
|
|
#define CPU_FTRS_CELL (CPU_FTR_LWSYNC | \
|
2008-12-19 03:13:32 +08:00
|
|
|
CPU_FTR_PPCAS_ARCH_V2 | CPU_FTR_CTRL | \
|
2006-03-23 14:36:59 +08:00
|
|
|
CPU_FTR_ALTIVEC_COMP | CPU_FTR_MMCRA | CPU_FTR_SMT | \
|
2011-04-07 03:48:50 +08:00
|
|
|
CPU_FTR_PAUSE_ZERO | CPU_FTR_CELL_TB_BUG | CPU_FTR_CP_USE_DCBTZ | \
|
2013-05-17 04:27:31 +08:00
|
|
|
CPU_FTR_UNALIGNED_LD_STD | CPU_FTR_DABRX)
|
2018-03-20 05:46:11 +08:00
|
|
|
#define CPU_FTRS_PA6T (CPU_FTR_LWSYNC | \
|
2011-04-07 03:48:50 +08:00
|
|
|
CPU_FTR_PPCAS_ARCH_V2 | CPU_FTR_ALTIVEC_COMP | \
|
2013-05-17 04:27:31 +08:00
|
|
|
CPU_FTR_PURR | CPU_FTR_REAL_LE | CPU_FTR_DABRX)
|
2018-03-20 05:46:11 +08:00
|
|
|
#define CPU_FTRS_COMPATIBLE (CPU_FTR_PPCAS_ARCH_V2)
|
2005-09-24 03:08:58 +08:00
|
|
|
|
2005-12-13 04:45:33 +08:00
|
|
|
#ifdef __powerpc64__
|
2011-04-06 13:11:06 +08:00
|
|
|
#ifdef CONFIG_PPC_BOOK3E
|
2014-08-06 16:26:28 +08:00
|
|
|
#define CPU_FTRS_POSSIBLE (CPU_FTRS_E6500 | CPU_FTRS_E5500)
|
2011-04-06 13:11:06 +08:00
|
|
|
#else
|
2018-02-21 03:08:31 +08:00
|
|
|
#ifdef CONFIG_CPU_LITTLE_ENDIAN
|
|
|
|
#define CPU_FTRS_POSSIBLE \
|
|
|
|
(CPU_FTRS_POWER7 | CPU_FTRS_POWER8E | CPU_FTRS_POWER8 | \
|
2018-07-11 14:02:58 +08:00
|
|
|
CPU_FTR_ALTIVEC_COMP | CPU_FTR_VSX_COMP | CPU_FTRS_POWER9 | \
|
2018-07-19 12:37:57 +08:00
|
|
|
CPU_FTRS_POWER9_DD2_1 | CPU_FTRS_POWER9_DD2_2)
|
2018-02-21 03:08:31 +08:00
|
|
|
#else
|
2006-03-23 14:36:59 +08:00
|
|
|
#define CPU_FTRS_POSSIBLE \
|
2018-02-21 03:08:29 +08:00
|
|
|
(CPU_FTRS_PPC970 | CPU_FTRS_POWER5 | \
|
2014-07-10 10:29:18 +08:00
|
|
|
CPU_FTRS_POWER6 | CPU_FTRS_POWER7 | CPU_FTRS_POWER8E | \
|
2018-07-11 14:02:58 +08:00
|
|
|
CPU_FTRS_POWER8 | CPU_FTRS_CELL | CPU_FTRS_PA6T | \
|
|
|
|
CPU_FTR_VSX_COMP | CPU_FTR_ALTIVEC_COMP | CPU_FTRS_POWER9 | \
|
2018-07-19 12:37:57 +08:00
|
|
|
CPU_FTRS_POWER9_DD2_1 | CPU_FTRS_POWER9_DD2_2)
|
2018-02-21 03:08:31 +08:00
|
|
|
#endif /* CONFIG_CPU_LITTLE_ENDIAN */
|
2011-04-06 13:11:06 +08:00
|
|
|
#endif
|
2005-12-13 04:45:33 +08:00
|
|
|
#else
|
2006-03-23 14:36:59 +08:00
|
|
|
enum {
|
|
|
|
CPU_FTRS_POSSIBLE =
|
2019-08-26 23:52:14 +08:00
|
|
|
#ifdef CONFIG_PPC_BOOK3S_601
|
|
|
|
CPU_FTRS_PPC601 |
|
|
|
|
#elif defined(CONFIG_PPC_BOOK3S_32)
|
2005-09-24 03:08:58 +08:00
|
|
|
CPU_FTRS_PPC601 | CPU_FTRS_603 | CPU_FTRS_604 | CPU_FTRS_740_NOTAU |
|
|
|
|
CPU_FTRS_740 | CPU_FTRS_750 | CPU_FTRS_750FX1 |
|
|
|
|
CPU_FTRS_750FX2 | CPU_FTRS_750FX | CPU_FTRS_750GX |
|
|
|
|
CPU_FTRS_7400_NOTAU | CPU_FTRS_7400 | CPU_FTRS_7450_20 |
|
|
|
|
CPU_FTRS_7450_21 | CPU_FTRS_7450_23 | CPU_FTRS_7455_1 |
|
|
|
|
CPU_FTRS_7455_20 | CPU_FTRS_7455 | CPU_FTRS_7447_10 |
|
|
|
|
CPU_FTRS_7447 | CPU_FTRS_7447A | CPU_FTRS_82XX |
|
2006-12-08 16:43:30 +08:00
|
|
|
CPU_FTRS_G2_LE | CPU_FTRS_E300 | CPU_FTRS_E300C2 |
|
|
|
|
CPU_FTRS_CLASSIC32 |
|
2005-09-24 03:08:58 +08:00
|
|
|
#else
|
|
|
|
CPU_FTRS_GENERIC_32 |
|
|
|
|
#endif
|
2017-08-08 19:58:54 +08:00
|
|
|
#ifdef CONFIG_PPC_8xx
|
2005-09-24 03:08:58 +08:00
|
|
|
CPU_FTRS_8XX |
|
|
|
|
#endif
|
|
|
|
#ifdef CONFIG_40x
|
|
|
|
CPU_FTRS_40X |
|
|
|
|
#endif
|
|
|
|
#ifdef CONFIG_44x
|
2008-12-19 03:13:22 +08:00
|
|
|
CPU_FTRS_44X | CPU_FTRS_440x6 |
|
2005-09-24 03:08:58 +08:00
|
|
|
#endif
|
2010-03-05 18:43:12 +08:00
|
|
|
#ifdef CONFIG_PPC_47x
|
2011-01-26 14:17:58 +08:00
|
|
|
CPU_FTRS_47X | CPU_FTR_476_DD2 |
|
2010-03-05 18:43:12 +08:00
|
|
|
#endif
|
2005-09-24 03:08:58 +08:00
|
|
|
#ifdef CONFIG_E200
|
|
|
|
CPU_FTRS_E200 |
|
|
|
|
#endif
|
|
|
|
#ifdef CONFIG_E500
|
2011-12-20 23:34:14 +08:00
|
|
|
CPU_FTRS_E500 | CPU_FTRS_E500_2 |
|
|
|
|
#endif
|
|
|
|
#ifdef CONFIG_PPC_E500MC
|
|
|
|
CPU_FTRS_E500MC | CPU_FTRS_E5500 | CPU_FTRS_E6500 |
|
2005-09-24 03:08:58 +08:00
|
|
|
#endif
|
|
|
|
0,
|
2006-03-23 14:36:59 +08:00
|
|
|
};
|
|
|
|
#endif /* __powerpc64__ */
|
2005-09-24 03:08:58 +08:00
|
|
|
|
2005-12-13 04:45:33 +08:00
|
|
|
#ifdef __powerpc64__
|
2011-04-06 13:11:06 +08:00
|
|
|
#ifdef CONFIG_PPC_BOOK3E
|
2014-08-06 16:26:28 +08:00
|
|
|
#define CPU_FTRS_ALWAYS (CPU_FTRS_E6500 & CPU_FTRS_E5500)
|
2011-04-06 13:11:06 +08:00
|
|
|
#else
|
powerpc/64s: Fix CPU_FTRS_ALWAYS vs DT CPU features
The cpu_has_feature() mechanism has an optimisation where at build
time we construct a mask of the CPU feature bits that will always be
true for the given .config, based on the platform/bitness/etc. that we
are building for.
That is incompatible with DT CPU features, where the set of CPU
features is dependent on feature flags that are given to us by
firmware.
The result is that some feature bits can not be *disabled* by DT CPU
features. Or more accurately, they can be disabled but they will still
appear in the ALWAYS mask, meaning cpu_has_feature() will always
return true for them.
In the past this hasn't really been a problem because on Book3S
64 (where we support DT CPU features), the set of ALWAYS bits has been
very small. That was because we always built for POWER4 and later,
meaning the set of common bits was small.
The only bit that could be cleared by DT CPU features that was also in
the ALWAYS mask was CPU_FTR_NODSISRALIGN, and that was only used in
the alignment handler to create a fake DSISR. That code was itself
deleted in 31bfdb036f12 ("powerpc: Use instruction emulation
infrastructure to handle alignment faults") (Sep 2017).
However the set of ALWAYS features changed with the recent commit
db5ae1c155af ("powerpc/64s: Refine feature sets for little endian
builds") which restricted the set of feature flags when building
little endian to Power7 or later. That caused the ALWAYS mask to
become much larger for little endian builds.
The result is that the following feature bits can currently not
be *disabled* by DT CPU features:
CPU_FTR_REAL_LE, CPU_FTR_MMCRA, CPU_FTR_CTRL, CPU_FTR_SMT,
CPU_FTR_PURR, CPU_FTR_SPURR, CPU_FTR_DSCR, CPU_FTR_PKEY,
CPU_FTR_VMX_COPY, CPU_FTR_CFAR, CPU_FTR_HAS_PPR.
To fix it we need to mask the set of ALWAYS features with the base set
of DT CPU features, ie. the features that are always enabled by DT CPU
features. That way there are no bits in the ALWAYS mask that are not
also always set by DT CPU features.
Fixes: db5ae1c155af ("powerpc/64s: Refine feature sets for little endian builds")
Reviewed-by: Nicholas Piggin <npiggin@gmail.com>
Signed-off-by: Michael Ellerman <mpe@ellerman.id.au>
2018-04-12 20:24:45 +08:00
|
|
|
|
|
|
|
#ifdef CONFIG_PPC_DT_CPU_FTRS
|
|
|
|
#define CPU_FTRS_DT_CPU_BASE \
|
|
|
|
(CPU_FTR_LWSYNC | \
|
|
|
|
CPU_FTR_FPU_UNAVAILABLE | \
|
|
|
|
CPU_FTR_NODSISRALIGN | \
|
|
|
|
CPU_FTR_NOEXECUTE | \
|
|
|
|
CPU_FTR_COHERENT_ICACHE | \
|
|
|
|
CPU_FTR_STCX_CHECKS_ADDRESS | \
|
|
|
|
CPU_FTR_POPCNTB | CPU_FTR_POPCNTD | \
|
|
|
|
CPU_FTR_DAWR | \
|
|
|
|
CPU_FTR_ARCH_206 | \
|
|
|
|
CPU_FTR_ARCH_207S)
|
|
|
|
#else
|
|
|
|
#define CPU_FTRS_DT_CPU_BASE (~0ul)
|
|
|
|
#endif
|
|
|
|
|
2018-02-21 03:08:31 +08:00
|
|
|
#ifdef CONFIG_CPU_LITTLE_ENDIAN
|
|
|
|
#define CPU_FTRS_ALWAYS \
|
|
|
|
(CPU_FTRS_POSSIBLE & ~CPU_FTR_HVMODE & CPU_FTRS_POWER7 & \
|
2018-07-19 12:37:57 +08:00
|
|
|
CPU_FTRS_POWER8E & CPU_FTRS_POWER8 & CPU_FTRS_POWER9 & \
|
|
|
|
CPU_FTRS_POWER9_DD2_1 & CPU_FTRS_DT_CPU_BASE)
|
2018-02-21 03:08:31 +08:00
|
|
|
#else
|
2006-03-23 14:36:59 +08:00
|
|
|
#define CPU_FTRS_ALWAYS \
|
2018-02-21 03:08:29 +08:00
|
|
|
(CPU_FTRS_PPC970 & CPU_FTRS_POWER5 & \
|
2014-07-10 10:29:18 +08:00
|
|
|
CPU_FTRS_POWER6 & CPU_FTRS_POWER7 & CPU_FTRS_CELL & \
|
2014-08-06 13:42:17 +08:00
|
|
|
CPU_FTRS_PA6T & CPU_FTRS_POWER8 & CPU_FTRS_POWER8E & \
|
2018-07-19 12:37:57 +08:00
|
|
|
~CPU_FTR_HVMODE & CPU_FTRS_POSSIBLE & CPU_FTRS_POWER9 & \
|
|
|
|
CPU_FTRS_POWER9_DD2_1 & CPU_FTRS_DT_CPU_BASE)
|
2018-02-21 03:08:31 +08:00
|
|
|
#endif /* CONFIG_CPU_LITTLE_ENDIAN */
|
2011-04-06 13:11:06 +08:00
|
|
|
#endif
|
2005-12-13 04:45:33 +08:00
|
|
|
#else
|
2006-03-23 14:36:59 +08:00
|
|
|
enum {
|
|
|
|
CPU_FTRS_ALWAYS =
|
2019-08-26 23:52:14 +08:00
|
|
|
#ifdef CONFIG_PPC_BOOK3S_601
|
|
|
|
CPU_FTRS_PPC601 &
|
|
|
|
#elif defined(CONFIG_PPC_BOOK3S_32)
|
|
|
|
CPU_FTRS_603 & CPU_FTRS_604 & CPU_FTRS_740_NOTAU &
|
2005-09-24 03:08:58 +08:00
|
|
|
CPU_FTRS_740 & CPU_FTRS_750 & CPU_FTRS_750FX1 &
|
|
|
|
CPU_FTRS_750FX2 & CPU_FTRS_750FX & CPU_FTRS_750GX &
|
|
|
|
CPU_FTRS_7400_NOTAU & CPU_FTRS_7400 & CPU_FTRS_7450_20 &
|
|
|
|
CPU_FTRS_7450_21 & CPU_FTRS_7450_23 & CPU_FTRS_7455_1 &
|
|
|
|
CPU_FTRS_7455_20 & CPU_FTRS_7455 & CPU_FTRS_7447_10 &
|
|
|
|
CPU_FTRS_7447 & CPU_FTRS_7447A & CPU_FTRS_82XX &
|
2006-12-08 16:43:30 +08:00
|
|
|
CPU_FTRS_G2_LE & CPU_FTRS_E300 & CPU_FTRS_E300C2 &
|
|
|
|
CPU_FTRS_CLASSIC32 &
|
2005-09-24 03:08:58 +08:00
|
|
|
#else
|
|
|
|
CPU_FTRS_GENERIC_32 &
|
|
|
|
#endif
|
2017-08-08 19:58:54 +08:00
|
|
|
#ifdef CONFIG_PPC_8xx
|
2005-09-24 03:08:58 +08:00
|
|
|
CPU_FTRS_8XX &
|
|
|
|
#endif
|
|
|
|
#ifdef CONFIG_40x
|
|
|
|
CPU_FTRS_40X &
|
|
|
|
#endif
|
|
|
|
#ifdef CONFIG_44x
|
2008-12-19 03:13:22 +08:00
|
|
|
CPU_FTRS_44X & CPU_FTRS_440x6 &
|
2005-09-24 03:08:58 +08:00
|
|
|
#endif
|
|
|
|
#ifdef CONFIG_E200
|
|
|
|
CPU_FTRS_E200 &
|
|
|
|
#endif
|
|
|
|
#ifdef CONFIG_E500
|
2011-12-20 23:34:14 +08:00
|
|
|
CPU_FTRS_E500 & CPU_FTRS_E500_2 &
|
|
|
|
#endif
|
|
|
|
#ifdef CONFIG_PPC_E500MC
|
|
|
|
CPU_FTRS_E500MC & CPU_FTRS_E5500 & CPU_FTRS_E6500 &
|
2005-09-24 03:08:58 +08:00
|
|
|
#endif
|
2011-12-20 23:34:47 +08:00
|
|
|
~CPU_FTR_EMB_HV & /* can be removed at runtime */
|
2005-09-24 03:08:58 +08:00
|
|
|
CPU_FTRS_POSSIBLE,
|
|
|
|
};
|
2006-03-23 14:36:59 +08:00
|
|
|
#endif /* __powerpc64__ */
|
2005-09-24 03:08:58 +08:00
|
|
|
|
2010-06-15 14:05:19 +08:00
|
|
|
#define HBP_NUM 1
|
|
|
|
|
2005-09-24 03:08:58 +08:00
|
|
|
#endif /* !__ASSEMBLY__ */
|
|
|
|
|
|
|
|
#endif /* __ASM_POWERPC_CPUTABLE_H */
|