x86/irq/32: Rename hard/softirq_stack to hard/softirq_stack_ptr
The percpu storage holds a pointer to the stack not the stack itself. Rename it before sharing struct irq_stack with 64-bit. No functional changes. Signed-off-by: Thomas Gleixner <tglx@linutronix.de> Signed-off-by: Borislav Petkov <bp@suse.de> Cc: Andrew Morton <akpm@linux-foundation.org> Cc: Andy Lutomirski <luto@kernel.org> Cc: Dave Hansen <dave.hansen@linux.intel.com> Cc: "H. Peter Anvin" <hpa@zytor.com> Cc: Ingo Molnar <mingo@redhat.com> Cc: Jiri Kosina <jkosina@suse.cz> Cc: Josh Poimboeuf <jpoimboe@redhat.com> Cc: Juergen Gross <jgross@suse.com> Cc: Nick Desaulniers <ndesaulniers@google.com> Cc: Nicolai Stange <nstange@suse.de> Cc: Sean Christopherson <sean.j.christopherson@intel.com> Cc: Vlastimil Babka <vbabka@suse.cz> Cc: x86-ml <x86@kernel.org> Link: https://lkml.kernel.org/r/20190414160145.824805922@linutronix.de
This commit is contained in:
parent
231c4846b1
commit
a754fe2b76
|
@ -425,8 +425,8 @@ struct irq_stack {
|
||||||
char stack[IRQ_STACK_SIZE];
|
char stack[IRQ_STACK_SIZE];
|
||||||
} __aligned(IRQ_STACK_SIZE);
|
} __aligned(IRQ_STACK_SIZE);
|
||||||
|
|
||||||
DECLARE_PER_CPU(struct irq_stack *, hardirq_stack);
|
DECLARE_PER_CPU(struct irq_stack *, hardirq_stack_ptr);
|
||||||
DECLARE_PER_CPU(struct irq_stack *, softirq_stack);
|
DECLARE_PER_CPU(struct irq_stack *, softirq_stack_ptr);
|
||||||
#endif /* X86_64 */
|
#endif /* X86_64 */
|
||||||
|
|
||||||
extern unsigned int fpu_kernel_xstate_size;
|
extern unsigned int fpu_kernel_xstate_size;
|
||||||
|
|
|
@ -34,7 +34,7 @@ const char *stack_type_name(enum stack_type type)
|
||||||
|
|
||||||
static bool in_hardirq_stack(unsigned long *stack, struct stack_info *info)
|
static bool in_hardirq_stack(unsigned long *stack, struct stack_info *info)
|
||||||
{
|
{
|
||||||
unsigned long *begin = (unsigned long *)this_cpu_read(hardirq_stack);
|
unsigned long *begin = (unsigned long *)this_cpu_read(hardirq_stack_ptr);
|
||||||
unsigned long *end = begin + (THREAD_SIZE / sizeof(long));
|
unsigned long *end = begin + (THREAD_SIZE / sizeof(long));
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@ -59,7 +59,7 @@ static bool in_hardirq_stack(unsigned long *stack, struct stack_info *info)
|
||||||
|
|
||||||
static bool in_softirq_stack(unsigned long *stack, struct stack_info *info)
|
static bool in_softirq_stack(unsigned long *stack, struct stack_info *info)
|
||||||
{
|
{
|
||||||
unsigned long *begin = (unsigned long *)this_cpu_read(softirq_stack);
|
unsigned long *begin = (unsigned long *)this_cpu_read(softirq_stack_ptr);
|
||||||
unsigned long *end = begin + (THREAD_SIZE / sizeof(long));
|
unsigned long *end = begin + (THREAD_SIZE / sizeof(long));
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
|
|
@ -51,8 +51,8 @@ static inline int check_stack_overflow(void) { return 0; }
|
||||||
static inline void print_stack_overflow(void) { }
|
static inline void print_stack_overflow(void) { }
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
DEFINE_PER_CPU(struct irq_stack *, hardirq_stack);
|
DEFINE_PER_CPU(struct irq_stack *, hardirq_stack_ptr);
|
||||||
DEFINE_PER_CPU(struct irq_stack *, softirq_stack);
|
DEFINE_PER_CPU(struct irq_stack *, softirq_stack_ptr);
|
||||||
|
|
||||||
static void call_on_stack(void *func, void *stack)
|
static void call_on_stack(void *func, void *stack)
|
||||||
{
|
{
|
||||||
|
@ -76,7 +76,7 @@ static inline int execute_on_irq_stack(int overflow, struct irq_desc *desc)
|
||||||
u32 *isp, *prev_esp, arg1;
|
u32 *isp, *prev_esp, arg1;
|
||||||
|
|
||||||
curstk = (struct irq_stack *) current_stack();
|
curstk = (struct irq_stack *) current_stack();
|
||||||
irqstk = __this_cpu_read(hardirq_stack);
|
irqstk = __this_cpu_read(hardirq_stack_ptr);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* this is where we switch to the IRQ stack. However, if we are
|
* this is where we switch to the IRQ stack. However, if we are
|
||||||
|
@ -113,21 +113,22 @@ void irq_ctx_init(int cpu)
|
||||||
{
|
{
|
||||||
struct irq_stack *irqstk;
|
struct irq_stack *irqstk;
|
||||||
|
|
||||||
if (per_cpu(hardirq_stack, cpu))
|
if (per_cpu(hardirq_stack_ptr, cpu))
|
||||||
return;
|
return;
|
||||||
|
|
||||||
irqstk = page_address(alloc_pages_node(cpu_to_node(cpu),
|
irqstk = page_address(alloc_pages_node(cpu_to_node(cpu),
|
||||||
THREADINFO_GFP,
|
THREADINFO_GFP,
|
||||||
THREAD_SIZE_ORDER));
|
THREAD_SIZE_ORDER));
|
||||||
per_cpu(hardirq_stack, cpu) = irqstk;
|
per_cpu(hardirq_stack_ptr, cpu) = irqstk;
|
||||||
|
|
||||||
irqstk = page_address(alloc_pages_node(cpu_to_node(cpu),
|
irqstk = page_address(alloc_pages_node(cpu_to_node(cpu),
|
||||||
THREADINFO_GFP,
|
THREADINFO_GFP,
|
||||||
THREAD_SIZE_ORDER));
|
THREAD_SIZE_ORDER));
|
||||||
per_cpu(softirq_stack, cpu) = irqstk;
|
per_cpu(softirq_stack_ptr, cpu) = irqstk;
|
||||||
|
|
||||||
printk(KERN_DEBUG "CPU %u irqstacks, hard=%p soft=%p\n",
|
pr_debug("CPU %u irqstacks, hard=%p soft=%p\n",
|
||||||
cpu, per_cpu(hardirq_stack, cpu), per_cpu(softirq_stack, cpu));
|
cpu, per_cpu(hardirq_stack_ptr, cpu),
|
||||||
|
per_cpu(softirq_stack_ptr, cpu));
|
||||||
}
|
}
|
||||||
|
|
||||||
void do_softirq_own_stack(void)
|
void do_softirq_own_stack(void)
|
||||||
|
@ -135,7 +136,7 @@ void do_softirq_own_stack(void)
|
||||||
struct irq_stack *irqstk;
|
struct irq_stack *irqstk;
|
||||||
u32 *isp, *prev_esp;
|
u32 *isp, *prev_esp;
|
||||||
|
|
||||||
irqstk = __this_cpu_read(softirq_stack);
|
irqstk = __this_cpu_read(softirq_stack_ptr);
|
||||||
|
|
||||||
/* build the stack frame on the softirq stack */
|
/* build the stack frame on the softirq stack */
|
||||||
isp = (u32 *) ((char *)irqstk + sizeof(*irqstk));
|
isp = (u32 *) ((char *)irqstk + sizeof(*irqstk));
|
||||||
|
|
Loading…
Reference in New Issue