Merge branch 'x86-urgent-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip
Pull x86 fixes from Thomas Gleixner: "A set of x86 specific fixes and updates: - The CR2 corruption fixes which store CR2 early in the entry code and hand the stored address to the fault handlers. - Revert a forgotten leftover of the dropped FSGSBASE series. - Plug a memory leak in the boot code. - Make the Hyper-V assist functionality robust by zeroing the shadow page. - Remove a useless check for dead processes with LDT - Update paravirt and VMware maintainers entries. - A few cleanup patches addressing various compiler warnings" * 'x86-urgent-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip: x86/entry/64: Prevent clobbering of saved CR2 value x86/hyper-v: Zero out the VP ASSIST PAGE on allocation x86, boot: Remove multiple copy of static function sanitize_boot_params() x86/boot/compressed/64: Remove unused variable x86/boot/efi: Remove unused variables x86/mm, tracing: Fix CR2 corruption x86/entry/64: Update comments and sanity tests for create_gap x86/entry/64: Simplify idtentry a little x86/entry/32: Simplify common_exception x86/paravirt: Make read_cr2() CALLEE_SAVE MAINTAINERS: Update PARAVIRT_OPS_INTERFACE and VMWARE_HYPERVISOR_INTERFACE x86/process: Delete useless check for dead process with LDT x86: math-emu: Hide clang warnings for 16-bit overflow x86/e820: Use proper booleans instead of 0/1 x86/apic: Silence -Wtype-limits compiler warnings x86/mm: Free sme_early_buffer after init x86/boot: Fix memory leak in default_get_smp_config() Revert "x86/ptrace: Prevent ptrace from clearing the FS/GS selector" and fix the test
This commit is contained in:
commit
c6dd78fcb8
|
@ -12133,7 +12133,8 @@ F: Documentation/driver-api/parport*.rst
|
||||||
|
|
||||||
PARAVIRT_OPS INTERFACE
|
PARAVIRT_OPS INTERFACE
|
||||||
M: Juergen Gross <jgross@suse.com>
|
M: Juergen Gross <jgross@suse.com>
|
||||||
M: Alok Kataria <akataria@vmware.com>
|
M: Thomas Hellstrom <thellstrom@vmware.com>
|
||||||
|
M: "VMware, Inc." <pv-drivers@vmware.com>
|
||||||
L: virtualization@lists.linux-foundation.org
|
L: virtualization@lists.linux-foundation.org
|
||||||
S: Supported
|
S: Supported
|
||||||
F: Documentation/virtual/paravirt_ops.txt
|
F: Documentation/virtual/paravirt_ops.txt
|
||||||
|
@ -17179,7 +17180,8 @@ S: Maintained
|
||||||
F: drivers/misc/vmw_balloon.c
|
F: drivers/misc/vmw_balloon.c
|
||||||
|
|
||||||
VMWARE HYPERVISOR INTERFACE
|
VMWARE HYPERVISOR INTERFACE
|
||||||
M: Alok Kataria <akataria@vmware.com>
|
M: Thomas Hellstrom <thellstrom@vmware.com>
|
||||||
|
M: "VMware, Inc." <pv-drivers@vmware.com>
|
||||||
L: virtualization@lists.linux-foundation.org
|
L: virtualization@lists.linux-foundation.org
|
||||||
S: Supported
|
S: Supported
|
||||||
F: arch/x86/kernel/cpu/vmware.c
|
F: arch/x86/kernel/cpu/vmware.c
|
||||||
|
|
|
@ -384,14 +384,11 @@ struct boot_params *make_boot_params(struct efi_config *c)
|
||||||
struct apm_bios_info *bi;
|
struct apm_bios_info *bi;
|
||||||
struct setup_header *hdr;
|
struct setup_header *hdr;
|
||||||
efi_loaded_image_t *image;
|
efi_loaded_image_t *image;
|
||||||
void *options, *handle;
|
void *handle;
|
||||||
efi_guid_t proto = LOADED_IMAGE_PROTOCOL_GUID;
|
efi_guid_t proto = LOADED_IMAGE_PROTOCOL_GUID;
|
||||||
int options_size = 0;
|
int options_size = 0;
|
||||||
efi_status_t status;
|
efi_status_t status;
|
||||||
char *cmdline_ptr;
|
char *cmdline_ptr;
|
||||||
u16 *s2;
|
|
||||||
u8 *s1;
|
|
||||||
int i;
|
|
||||||
unsigned long ramdisk_addr;
|
unsigned long ramdisk_addr;
|
||||||
unsigned long ramdisk_size;
|
unsigned long ramdisk_size;
|
||||||
|
|
||||||
|
@ -494,8 +491,6 @@ static void add_e820ext(struct boot_params *params,
|
||||||
struct setup_data *e820ext, u32 nr_entries)
|
struct setup_data *e820ext, u32 nr_entries)
|
||||||
{
|
{
|
||||||
struct setup_data *data;
|
struct setup_data *data;
|
||||||
efi_status_t status;
|
|
||||||
unsigned long size;
|
|
||||||
|
|
||||||
e820ext->type = SETUP_E820_EXT;
|
e820ext->type = SETUP_E820_EXT;
|
||||||
e820ext->len = nr_entries * sizeof(struct boot_e820_entry);
|
e820ext->len = nr_entries * sizeof(struct boot_e820_entry);
|
||||||
|
@ -677,8 +672,6 @@ static efi_status_t exit_boot_func(efi_system_table_t *sys_table_arg,
|
||||||
void *priv)
|
void *priv)
|
||||||
{
|
{
|
||||||
const char *signature;
|
const char *signature;
|
||||||
__u32 nr_desc;
|
|
||||||
efi_status_t status;
|
|
||||||
struct exit_boot_struct *p = priv;
|
struct exit_boot_struct *p = priv;
|
||||||
|
|
||||||
signature = efi_is_64bit() ? EFI64_LOADER_SIGNATURE
|
signature = efi_is_64bit() ? EFI64_LOADER_SIGNATURE
|
||||||
|
@ -747,7 +740,6 @@ struct boot_params *
|
||||||
efi_main(struct efi_config *c, struct boot_params *boot_params)
|
efi_main(struct efi_config *c, struct boot_params *boot_params)
|
||||||
{
|
{
|
||||||
struct desc_ptr *gdt = NULL;
|
struct desc_ptr *gdt = NULL;
|
||||||
efi_loaded_image_t *image;
|
|
||||||
struct setup_header *hdr = &boot_params->hdr;
|
struct setup_header *hdr = &boot_params->hdr;
|
||||||
efi_status_t status;
|
efi_status_t status;
|
||||||
struct desc_struct *desc;
|
struct desc_struct *desc;
|
||||||
|
|
|
@ -17,6 +17,7 @@
|
||||||
#include "pgtable.h"
|
#include "pgtable.h"
|
||||||
#include "../string.h"
|
#include "../string.h"
|
||||||
#include "../voffset.h"
|
#include "../voffset.h"
|
||||||
|
#include <asm/bootparam_utils.h>
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* WARNING!!
|
* WARNING!!
|
||||||
|
|
|
@ -23,7 +23,6 @@
|
||||||
#include <asm/page.h>
|
#include <asm/page.h>
|
||||||
#include <asm/boot.h>
|
#include <asm/boot.h>
|
||||||
#include <asm/bootparam.h>
|
#include <asm/bootparam.h>
|
||||||
#include <asm/bootparam_utils.h>
|
|
||||||
|
|
||||||
#define BOOT_CTYPE_H
|
#define BOOT_CTYPE_H
|
||||||
#include <linux/acpi.h>
|
#include <linux/acpi.h>
|
||||||
|
|
|
@ -40,7 +40,6 @@ int cmdline_find_option_bool(const char *option);
|
||||||
static unsigned long find_trampoline_placement(void)
|
static unsigned long find_trampoline_placement(void)
|
||||||
{
|
{
|
||||||
unsigned long bios_start = 0, ebda_start = 0;
|
unsigned long bios_start = 0, ebda_start = 0;
|
||||||
unsigned long trampoline_start;
|
|
||||||
struct boot_e820_entry *entry;
|
struct boot_e820_entry *entry;
|
||||||
char *signature;
|
char *signature;
|
||||||
int i;
|
int i;
|
||||||
|
|
|
@ -343,3 +343,9 @@ For 32-bit we have the following conventions - kernel is built with
|
||||||
.Lafter_call_\@:
|
.Lafter_call_\@:
|
||||||
#endif
|
#endif
|
||||||
.endm
|
.endm
|
||||||
|
|
||||||
|
#ifdef CONFIG_PARAVIRT_XXL
|
||||||
|
#define GET_CR2_INTO(reg) GET_CR2_INTO_AX ; _ASM_MOV %_ASM_AX, reg
|
||||||
|
#else
|
||||||
|
#define GET_CR2_INTO(reg) _ASM_MOV %cr2, reg
|
||||||
|
#endif
|
||||||
|
|
|
@ -294,9 +294,11 @@
|
||||||
.Lfinished_frame_\@:
|
.Lfinished_frame_\@:
|
||||||
.endm
|
.endm
|
||||||
|
|
||||||
.macro SAVE_ALL pt_regs_ax=%eax switch_stacks=0
|
.macro SAVE_ALL pt_regs_ax=%eax switch_stacks=0 skip_gs=0
|
||||||
cld
|
cld
|
||||||
|
.if \skip_gs == 0
|
||||||
PUSH_GS
|
PUSH_GS
|
||||||
|
.endif
|
||||||
FIXUP_FRAME
|
FIXUP_FRAME
|
||||||
pushl %fs
|
pushl %fs
|
||||||
pushl %es
|
pushl %es
|
||||||
|
@ -313,13 +315,13 @@
|
||||||
movl %edx, %es
|
movl %edx, %es
|
||||||
movl $(__KERNEL_PERCPU), %edx
|
movl $(__KERNEL_PERCPU), %edx
|
||||||
movl %edx, %fs
|
movl %edx, %fs
|
||||||
|
.if \skip_gs == 0
|
||||||
SET_KERNEL_GS %edx
|
SET_KERNEL_GS %edx
|
||||||
|
.endif
|
||||||
/* Switch to kernel stack if necessary */
|
/* Switch to kernel stack if necessary */
|
||||||
.if \switch_stacks > 0
|
.if \switch_stacks > 0
|
||||||
SWITCH_TO_KERNEL_STACK
|
SWITCH_TO_KERNEL_STACK
|
||||||
.endif
|
.endif
|
||||||
|
|
||||||
.endm
|
.endm
|
||||||
|
|
||||||
.macro SAVE_ALL_NMI cr3_reg:req
|
.macro SAVE_ALL_NMI cr3_reg:req
|
||||||
|
@ -1441,39 +1443,46 @@ BUILD_INTERRUPT3(hv_stimer0_callback_vector, HYPERV_STIMER0_VECTOR,
|
||||||
|
|
||||||
ENTRY(page_fault)
|
ENTRY(page_fault)
|
||||||
ASM_CLAC
|
ASM_CLAC
|
||||||
pushl $do_page_fault
|
pushl $0; /* %gs's slot on the stack */
|
||||||
ALIGN
|
|
||||||
jmp common_exception
|
SAVE_ALL switch_stacks=1 skip_gs=1
|
||||||
|
|
||||||
|
ENCODE_FRAME_POINTER
|
||||||
|
UNWIND_ESPFIX_STACK
|
||||||
|
|
||||||
|
/* fixup %gs */
|
||||||
|
GS_TO_REG %ecx
|
||||||
|
REG_TO_PTGS %ecx
|
||||||
|
SET_KERNEL_GS %ecx
|
||||||
|
|
||||||
|
GET_CR2_INTO(%ecx) # might clobber %eax
|
||||||
|
|
||||||
|
/* fixup orig %eax */
|
||||||
|
movl PT_ORIG_EAX(%esp), %edx # get the error code
|
||||||
|
movl $-1, PT_ORIG_EAX(%esp) # no syscall to restart
|
||||||
|
|
||||||
|
TRACE_IRQS_OFF
|
||||||
|
movl %esp, %eax # pt_regs pointer
|
||||||
|
call do_page_fault
|
||||||
|
jmp ret_from_exception
|
||||||
END(page_fault)
|
END(page_fault)
|
||||||
|
|
||||||
common_exception:
|
common_exception:
|
||||||
/* the function address is in %gs's slot on the stack */
|
/* the function address is in %gs's slot on the stack */
|
||||||
FIXUP_FRAME
|
SAVE_ALL switch_stacks=1 skip_gs=1
|
||||||
pushl %fs
|
|
||||||
pushl %es
|
|
||||||
pushl %ds
|
|
||||||
pushl %eax
|
|
||||||
movl $(__USER_DS), %eax
|
|
||||||
movl %eax, %ds
|
|
||||||
movl %eax, %es
|
|
||||||
movl $(__KERNEL_PERCPU), %eax
|
|
||||||
movl %eax, %fs
|
|
||||||
pushl %ebp
|
|
||||||
pushl %edi
|
|
||||||
pushl %esi
|
|
||||||
pushl %edx
|
|
||||||
pushl %ecx
|
|
||||||
pushl %ebx
|
|
||||||
SWITCH_TO_KERNEL_STACK
|
|
||||||
ENCODE_FRAME_POINTER
|
ENCODE_FRAME_POINTER
|
||||||
cld
|
|
||||||
UNWIND_ESPFIX_STACK
|
UNWIND_ESPFIX_STACK
|
||||||
|
|
||||||
|
/* fixup %gs */
|
||||||
GS_TO_REG %ecx
|
GS_TO_REG %ecx
|
||||||
movl PT_GS(%esp), %edi # get the function address
|
movl PT_GS(%esp), %edi # get the function address
|
||||||
movl PT_ORIG_EAX(%esp), %edx # get the error code
|
|
||||||
movl $-1, PT_ORIG_EAX(%esp) # no syscall to restart
|
|
||||||
REG_TO_PTGS %ecx
|
REG_TO_PTGS %ecx
|
||||||
SET_KERNEL_GS %ecx
|
SET_KERNEL_GS %ecx
|
||||||
|
|
||||||
|
/* fixup orig %eax */
|
||||||
|
movl PT_ORIG_EAX(%esp), %edx # get the error code
|
||||||
|
movl $-1, PT_ORIG_EAX(%esp) # no syscall to restart
|
||||||
|
|
||||||
TRACE_IRQS_OFF
|
TRACE_IRQS_OFF
|
||||||
movl %esp, %eax # pt_regs pointer
|
movl %esp, %eax # pt_regs pointer
|
||||||
CALL_NOSPEC %edi
|
CALL_NOSPEC %edi
|
||||||
|
|
|
@ -864,18 +864,84 @@ apicinterrupt IRQ_WORK_VECTOR irq_work_interrupt smp_irq_work_interrupt
|
||||||
*/
|
*/
|
||||||
#define CPU_TSS_IST(x) PER_CPU_VAR(cpu_tss_rw) + (TSS_ist + (x) * 8)
|
#define CPU_TSS_IST(x) PER_CPU_VAR(cpu_tss_rw) + (TSS_ist + (x) * 8)
|
||||||
|
|
||||||
|
.macro idtentry_part do_sym, has_error_code:req, read_cr2:req, paranoid:req, shift_ist=-1, ist_offset=0
|
||||||
|
|
||||||
|
.if \paranoid
|
||||||
|
call paranoid_entry
|
||||||
|
/* returned flag: ebx=0: need swapgs on exit, ebx=1: don't need it */
|
||||||
|
.else
|
||||||
|
call error_entry
|
||||||
|
.endif
|
||||||
|
UNWIND_HINT_REGS
|
||||||
|
|
||||||
|
.if \read_cr2
|
||||||
|
/*
|
||||||
|
* Store CR2 early so subsequent faults cannot clobber it. Use R12 as
|
||||||
|
* intermediate storage as RDX can be clobbered in enter_from_user_mode().
|
||||||
|
* GET_CR2_INTO can clobber RAX.
|
||||||
|
*/
|
||||||
|
GET_CR2_INTO(%r12);
|
||||||
|
.endif
|
||||||
|
|
||||||
|
.if \shift_ist != -1
|
||||||
|
TRACE_IRQS_OFF_DEBUG /* reload IDT in case of recursion */
|
||||||
|
.else
|
||||||
|
TRACE_IRQS_OFF
|
||||||
|
.endif
|
||||||
|
|
||||||
|
.if \paranoid == 0
|
||||||
|
testb $3, CS(%rsp)
|
||||||
|
jz .Lfrom_kernel_no_context_tracking_\@
|
||||||
|
CALL_enter_from_user_mode
|
||||||
|
.Lfrom_kernel_no_context_tracking_\@:
|
||||||
|
.endif
|
||||||
|
|
||||||
|
movq %rsp, %rdi /* pt_regs pointer */
|
||||||
|
|
||||||
|
.if \has_error_code
|
||||||
|
movq ORIG_RAX(%rsp), %rsi /* get error code */
|
||||||
|
movq $-1, ORIG_RAX(%rsp) /* no syscall to restart */
|
||||||
|
.else
|
||||||
|
xorl %esi, %esi /* no error code */
|
||||||
|
.endif
|
||||||
|
|
||||||
|
.if \shift_ist != -1
|
||||||
|
subq $\ist_offset, CPU_TSS_IST(\shift_ist)
|
||||||
|
.endif
|
||||||
|
|
||||||
|
.if \read_cr2
|
||||||
|
movq %r12, %rdx /* Move CR2 into 3rd argument */
|
||||||
|
.endif
|
||||||
|
|
||||||
|
call \do_sym
|
||||||
|
|
||||||
|
.if \shift_ist != -1
|
||||||
|
addq $\ist_offset, CPU_TSS_IST(\shift_ist)
|
||||||
|
.endif
|
||||||
|
|
||||||
|
.if \paranoid
|
||||||
|
/* this procedure expect "no swapgs" flag in ebx */
|
||||||
|
jmp paranoid_exit
|
||||||
|
.else
|
||||||
|
jmp error_exit
|
||||||
|
.endif
|
||||||
|
|
||||||
|
.endm
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* idtentry - Generate an IDT entry stub
|
* idtentry - Generate an IDT entry stub
|
||||||
* @sym: Name of the generated entry point
|
* @sym: Name of the generated entry point
|
||||||
* @do_sym: C function to be called
|
* @do_sym: C function to be called
|
||||||
* @has_error_code: True if this IDT vector has an error code on the stack
|
* @has_error_code: True if this IDT vector has an error code on the stack
|
||||||
* @paranoid: non-zero means that this vector may be invoked from
|
* @paranoid: non-zero means that this vector may be invoked from
|
||||||
* kernel mode with user GSBASE and/or user CR3.
|
* kernel mode with user GSBASE and/or user CR3.
|
||||||
* 2 is special -- see below.
|
* 2 is special -- see below.
|
||||||
* @shift_ist: Set to an IST index if entries from kernel mode should
|
* @shift_ist: Set to an IST index if entries from kernel mode should
|
||||||
* decrement the IST stack so that nested entries get a
|
* decrement the IST stack so that nested entries get a
|
||||||
* fresh stack. (This is for #DB, which has a nasty habit
|
* fresh stack. (This is for #DB, which has a nasty habit
|
||||||
* of recursing.)
|
* of recursing.)
|
||||||
|
* @create_gap: create a 6-word stack gap when coming from kernel mode.
|
||||||
|
* @read_cr2: load CR2 into the 3rd argument; done before calling any C code
|
||||||
*
|
*
|
||||||
* idtentry generates an IDT stub that sets up a usable kernel context,
|
* idtentry generates an IDT stub that sets up a usable kernel context,
|
||||||
* creates struct pt_regs, and calls @do_sym. The stub has the following
|
* creates struct pt_regs, and calls @do_sym. The stub has the following
|
||||||
|
@ -900,15 +966,19 @@ apicinterrupt IRQ_WORK_VECTOR irq_work_interrupt smp_irq_work_interrupt
|
||||||
* @paranoid == 2 is special: the stub will never switch stacks. This is for
|
* @paranoid == 2 is special: the stub will never switch stacks. This is for
|
||||||
* #DF: if the thread stack is somehow unusable, we'll still get a useful OOPS.
|
* #DF: if the thread stack is somehow unusable, we'll still get a useful OOPS.
|
||||||
*/
|
*/
|
||||||
.macro idtentry sym do_sym has_error_code:req paranoid=0 shift_ist=-1 ist_offset=0 create_gap=0
|
.macro idtentry sym do_sym has_error_code:req paranoid=0 shift_ist=-1 ist_offset=0 create_gap=0 read_cr2=0
|
||||||
ENTRY(\sym)
|
ENTRY(\sym)
|
||||||
UNWIND_HINT_IRET_REGS offset=\has_error_code*8
|
UNWIND_HINT_IRET_REGS offset=\has_error_code*8
|
||||||
|
|
||||||
/* Sanity check */
|
/* Sanity check */
|
||||||
.if \shift_ist != -1 && \paranoid == 0
|
.if \shift_ist != -1 && \paranoid != 1
|
||||||
.error "using shift_ist requires paranoid=1"
|
.error "using shift_ist requires paranoid=1"
|
||||||
.endif
|
.endif
|
||||||
|
|
||||||
|
.if \create_gap && \paranoid
|
||||||
|
.error "using create_gap requires paranoid=0"
|
||||||
|
.endif
|
||||||
|
|
||||||
ASM_CLAC
|
ASM_CLAC
|
||||||
|
|
||||||
.if \has_error_code == 0
|
.if \has_error_code == 0
|
||||||
|
@ -934,47 +1004,7 @@ ENTRY(\sym)
|
||||||
.Lfrom_usermode_no_gap_\@:
|
.Lfrom_usermode_no_gap_\@:
|
||||||
.endif
|
.endif
|
||||||
|
|
||||||
.if \paranoid
|
idtentry_part \do_sym, \has_error_code, \read_cr2, \paranoid, \shift_ist, \ist_offset
|
||||||
call paranoid_entry
|
|
||||||
.else
|
|
||||||
call error_entry
|
|
||||||
.endif
|
|
||||||
UNWIND_HINT_REGS
|
|
||||||
/* returned flag: ebx=0: need swapgs on exit, ebx=1: don't need it */
|
|
||||||
|
|
||||||
.if \paranoid
|
|
||||||
.if \shift_ist != -1
|
|
||||||
TRACE_IRQS_OFF_DEBUG /* reload IDT in case of recursion */
|
|
||||||
.else
|
|
||||||
TRACE_IRQS_OFF
|
|
||||||
.endif
|
|
||||||
.endif
|
|
||||||
|
|
||||||
movq %rsp, %rdi /* pt_regs pointer */
|
|
||||||
|
|
||||||
.if \has_error_code
|
|
||||||
movq ORIG_RAX(%rsp), %rsi /* get error code */
|
|
||||||
movq $-1, ORIG_RAX(%rsp) /* no syscall to restart */
|
|
||||||
.else
|
|
||||||
xorl %esi, %esi /* no error code */
|
|
||||||
.endif
|
|
||||||
|
|
||||||
.if \shift_ist != -1
|
|
||||||
subq $\ist_offset, CPU_TSS_IST(\shift_ist)
|
|
||||||
.endif
|
|
||||||
|
|
||||||
call \do_sym
|
|
||||||
|
|
||||||
.if \shift_ist != -1
|
|
||||||
addq $\ist_offset, CPU_TSS_IST(\shift_ist)
|
|
||||||
.endif
|
|
||||||
|
|
||||||
/* these procedures expect "no swapgs" flag in ebx */
|
|
||||||
.if \paranoid
|
|
||||||
jmp paranoid_exit
|
|
||||||
.else
|
|
||||||
jmp error_exit
|
|
||||||
.endif
|
|
||||||
|
|
||||||
.if \paranoid == 1
|
.if \paranoid == 1
|
||||||
/*
|
/*
|
||||||
|
@ -983,21 +1013,9 @@ ENTRY(\sym)
|
||||||
* run in real process context if user_mode(regs).
|
* run in real process context if user_mode(regs).
|
||||||
*/
|
*/
|
||||||
.Lfrom_usermode_switch_stack_\@:
|
.Lfrom_usermode_switch_stack_\@:
|
||||||
call error_entry
|
idtentry_part \do_sym, \has_error_code, \read_cr2, paranoid=0
|
||||||
|
|
||||||
movq %rsp, %rdi /* pt_regs pointer */
|
|
||||||
|
|
||||||
.if \has_error_code
|
|
||||||
movq ORIG_RAX(%rsp), %rsi /* get error code */
|
|
||||||
movq $-1, ORIG_RAX(%rsp) /* no syscall to restart */
|
|
||||||
.else
|
|
||||||
xorl %esi, %esi /* no error code */
|
|
||||||
.endif
|
.endif
|
||||||
|
|
||||||
call \do_sym
|
|
||||||
|
|
||||||
jmp error_exit
|
|
||||||
.endif
|
|
||||||
_ASM_NOKPROBE(\sym)
|
_ASM_NOKPROBE(\sym)
|
||||||
END(\sym)
|
END(\sym)
|
||||||
.endm
|
.endm
|
||||||
|
@ -1007,7 +1025,7 @@ idtentry overflow do_overflow has_error_code=0
|
||||||
idtentry bounds do_bounds has_error_code=0
|
idtentry bounds do_bounds has_error_code=0
|
||||||
idtentry invalid_op do_invalid_op has_error_code=0
|
idtentry invalid_op do_invalid_op has_error_code=0
|
||||||
idtentry device_not_available do_device_not_available has_error_code=0
|
idtentry device_not_available do_device_not_available has_error_code=0
|
||||||
idtentry double_fault do_double_fault has_error_code=1 paranoid=2
|
idtentry double_fault do_double_fault has_error_code=1 paranoid=2 read_cr2=1
|
||||||
idtentry coprocessor_segment_overrun do_coprocessor_segment_overrun has_error_code=0
|
idtentry coprocessor_segment_overrun do_coprocessor_segment_overrun has_error_code=0
|
||||||
idtentry invalid_TSS do_invalid_TSS has_error_code=1
|
idtentry invalid_TSS do_invalid_TSS has_error_code=1
|
||||||
idtentry segment_not_present do_segment_not_present has_error_code=1
|
idtentry segment_not_present do_segment_not_present has_error_code=1
|
||||||
|
@ -1179,10 +1197,10 @@ idtentry xendebug do_debug has_error_code=0
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
idtentry general_protection do_general_protection has_error_code=1
|
idtentry general_protection do_general_protection has_error_code=1
|
||||||
idtentry page_fault do_page_fault has_error_code=1
|
idtentry page_fault do_page_fault has_error_code=1 read_cr2=1
|
||||||
|
|
||||||
#ifdef CONFIG_KVM_GUEST
|
#ifdef CONFIG_KVM_GUEST
|
||||||
idtentry async_page_fault do_async_page_fault has_error_code=1
|
idtentry async_page_fault do_async_page_fault has_error_code=1 read_cr2=1
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#ifdef CONFIG_X86_MCE
|
#ifdef CONFIG_X86_MCE
|
||||||
|
@ -1281,18 +1299,9 @@ ENTRY(error_entry)
|
||||||
movq %rax, %rsp /* switch stack */
|
movq %rax, %rsp /* switch stack */
|
||||||
ENCODE_FRAME_POINTER
|
ENCODE_FRAME_POINTER
|
||||||
pushq %r12
|
pushq %r12
|
||||||
|
|
||||||
/*
|
|
||||||
* We need to tell lockdep that IRQs are off. We can't do this until
|
|
||||||
* we fix gsbase, and we should do it before enter_from_user_mode
|
|
||||||
* (which can take locks).
|
|
||||||
*/
|
|
||||||
TRACE_IRQS_OFF
|
|
||||||
CALL_enter_from_user_mode
|
|
||||||
ret
|
ret
|
||||||
|
|
||||||
.Lerror_entry_done:
|
.Lerror_entry_done:
|
||||||
TRACE_IRQS_OFF
|
|
||||||
ret
|
ret
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
|
|
@ -60,8 +60,17 @@ static int hv_cpu_init(unsigned int cpu)
|
||||||
if (!hv_vp_assist_page)
|
if (!hv_vp_assist_page)
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
if (!*hvp)
|
/*
|
||||||
*hvp = __vmalloc(PAGE_SIZE, GFP_KERNEL, PAGE_KERNEL);
|
* The VP ASSIST PAGE is an "overlay" page (see Hyper-V TLFS's Section
|
||||||
|
* 5.2.1 "GPA Overlay Pages"). Here it must be zeroed out to make sure
|
||||||
|
* we always write the EOI MSR in hv_apic_eoi_write() *after* the
|
||||||
|
* EOI optimization is disabled in hv_cpu_die(), otherwise a CPU may
|
||||||
|
* not be stopped in the case of CPU offlining and the VM will hang.
|
||||||
|
*/
|
||||||
|
if (!*hvp) {
|
||||||
|
*hvp = __vmalloc(PAGE_SIZE, GFP_KERNEL | __GFP_ZERO,
|
||||||
|
PAGE_KERNEL);
|
||||||
|
}
|
||||||
|
|
||||||
if (*hvp) {
|
if (*hvp) {
|
||||||
u64 val;
|
u64 val;
|
||||||
|
|
|
@ -49,7 +49,7 @@ static inline void generic_apic_probe(void)
|
||||||
|
|
||||||
#ifdef CONFIG_X86_LOCAL_APIC
|
#ifdef CONFIG_X86_LOCAL_APIC
|
||||||
|
|
||||||
extern unsigned int apic_verbosity;
|
extern int apic_verbosity;
|
||||||
extern int local_apic_timer_c2_ok;
|
extern int local_apic_timer_c2_ok;
|
||||||
|
|
||||||
extern int disable_apic;
|
extern int disable_apic;
|
||||||
|
|
|
@ -92,7 +92,7 @@ void kvm_async_pf_task_wait(u32 token, int interrupt_kernel);
|
||||||
void kvm_async_pf_task_wake(u32 token);
|
void kvm_async_pf_task_wake(u32 token);
|
||||||
u32 kvm_read_and_reset_pf_reason(void);
|
u32 kvm_read_and_reset_pf_reason(void);
|
||||||
extern void kvm_disable_steal_time(void);
|
extern void kvm_disable_steal_time(void);
|
||||||
void do_async_page_fault(struct pt_regs *regs, unsigned long error_code);
|
void do_async_page_fault(struct pt_regs *regs, unsigned long error_code, unsigned long address);
|
||||||
|
|
||||||
#ifdef CONFIG_PARAVIRT_SPINLOCKS
|
#ifdef CONFIG_PARAVIRT_SPINLOCKS
|
||||||
void __init kvm_spinlock_init(void);
|
void __init kvm_spinlock_init(void);
|
||||||
|
|
|
@ -116,7 +116,7 @@ static inline void write_cr0(unsigned long x)
|
||||||
|
|
||||||
static inline unsigned long read_cr2(void)
|
static inline unsigned long read_cr2(void)
|
||||||
{
|
{
|
||||||
return PVOP_CALL0(unsigned long, mmu.read_cr2);
|
return PVOP_CALLEE0(unsigned long, mmu.read_cr2);
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline void write_cr2(unsigned long x)
|
static inline void write_cr2(unsigned long x)
|
||||||
|
@ -910,13 +910,7 @@ extern void default_banner(void);
|
||||||
ANNOTATE_RETPOLINE_SAFE; \
|
ANNOTATE_RETPOLINE_SAFE; \
|
||||||
call PARA_INDIRECT(pv_ops+PV_CPU_swapgs); \
|
call PARA_INDIRECT(pv_ops+PV_CPU_swapgs); \
|
||||||
)
|
)
|
||||||
#endif
|
|
||||||
|
|
||||||
#define GET_CR2_INTO_RAX \
|
|
||||||
ANNOTATE_RETPOLINE_SAFE; \
|
|
||||||
call PARA_INDIRECT(pv_ops+PV_MMU_read_cr2);
|
|
||||||
|
|
||||||
#ifdef CONFIG_PARAVIRT_XXL
|
|
||||||
#define USERGS_SYSRET64 \
|
#define USERGS_SYSRET64 \
|
||||||
PARA_SITE(PARA_PATCH(PV_CPU_usergs_sysret64), \
|
PARA_SITE(PARA_PATCH(PV_CPU_usergs_sysret64), \
|
||||||
ANNOTATE_RETPOLINE_SAFE; \
|
ANNOTATE_RETPOLINE_SAFE; \
|
||||||
|
@ -930,9 +924,19 @@ extern void default_banner(void);
|
||||||
call PARA_INDIRECT(pv_ops+PV_IRQ_save_fl); \
|
call PARA_INDIRECT(pv_ops+PV_IRQ_save_fl); \
|
||||||
PV_RESTORE_REGS(clobbers | CLBR_CALLEE_SAVE);)
|
PV_RESTORE_REGS(clobbers | CLBR_CALLEE_SAVE);)
|
||||||
#endif
|
#endif
|
||||||
#endif
|
#endif /* CONFIG_PARAVIRT_XXL */
|
||||||
|
#endif /* CONFIG_X86_64 */
|
||||||
|
|
||||||
|
#ifdef CONFIG_PARAVIRT_XXL
|
||||||
|
|
||||||
|
#define GET_CR2_INTO_AX \
|
||||||
|
PARA_SITE(PARA_PATCH(PV_MMU_read_cr2), \
|
||||||
|
ANNOTATE_RETPOLINE_SAFE; \
|
||||||
|
call PARA_INDIRECT(pv_ops+PV_MMU_read_cr2); \
|
||||||
|
)
|
||||||
|
|
||||||
|
#endif /* CONFIG_PARAVIRT_XXL */
|
||||||
|
|
||||||
#endif /* CONFIG_X86_32 */
|
|
||||||
|
|
||||||
#endif /* __ASSEMBLY__ */
|
#endif /* __ASSEMBLY__ */
|
||||||
#else /* CONFIG_PARAVIRT */
|
#else /* CONFIG_PARAVIRT */
|
||||||
|
|
|
@ -220,7 +220,7 @@ struct pv_mmu_ops {
|
||||||
void (*exit_mmap)(struct mm_struct *mm);
|
void (*exit_mmap)(struct mm_struct *mm);
|
||||||
|
|
||||||
#ifdef CONFIG_PARAVIRT_XXL
|
#ifdef CONFIG_PARAVIRT_XXL
|
||||||
unsigned long (*read_cr2)(void);
|
struct paravirt_callee_save read_cr2;
|
||||||
void (*write_cr2)(unsigned long);
|
void (*write_cr2)(unsigned long);
|
||||||
|
|
||||||
unsigned long (*read_cr3)(void);
|
unsigned long (*read_cr3)(void);
|
||||||
|
|
|
@ -74,14 +74,14 @@ dotraplinkage void do_invalid_TSS(struct pt_regs *regs, long error_code);
|
||||||
dotraplinkage void do_segment_not_present(struct pt_regs *regs, long error_code);
|
dotraplinkage void do_segment_not_present(struct pt_regs *regs, long error_code);
|
||||||
dotraplinkage void do_stack_segment(struct pt_regs *regs, long error_code);
|
dotraplinkage void do_stack_segment(struct pt_regs *regs, long error_code);
|
||||||
#ifdef CONFIG_X86_64
|
#ifdef CONFIG_X86_64
|
||||||
dotraplinkage void do_double_fault(struct pt_regs *regs, long error_code);
|
dotraplinkage void do_double_fault(struct pt_regs *regs, long error_code, unsigned long address);
|
||||||
asmlinkage __visible notrace struct pt_regs *sync_regs(struct pt_regs *eregs);
|
asmlinkage __visible notrace struct pt_regs *sync_regs(struct pt_regs *eregs);
|
||||||
asmlinkage __visible notrace
|
asmlinkage __visible notrace
|
||||||
struct bad_iret_stack *fixup_bad_iret(struct bad_iret_stack *s);
|
struct bad_iret_stack *fixup_bad_iret(struct bad_iret_stack *s);
|
||||||
void __init trap_init(void);
|
void __init trap_init(void);
|
||||||
#endif
|
#endif
|
||||||
dotraplinkage void do_general_protection(struct pt_regs *regs, long error_code);
|
dotraplinkage void do_general_protection(struct pt_regs *regs, long error_code);
|
||||||
dotraplinkage void do_page_fault(struct pt_regs *regs, unsigned long error_code);
|
dotraplinkage void do_page_fault(struct pt_regs *regs, unsigned long error_code, unsigned long address);
|
||||||
dotraplinkage void do_spurious_interrupt_bug(struct pt_regs *regs, long error_code);
|
dotraplinkage void do_spurious_interrupt_bug(struct pt_regs *regs, long error_code);
|
||||||
dotraplinkage void do_coprocessor_error(struct pt_regs *regs, long error_code);
|
dotraplinkage void do_coprocessor_error(struct pt_regs *regs, long error_code);
|
||||||
dotraplinkage void do_alignment_check(struct pt_regs *regs, long error_code);
|
dotraplinkage void do_alignment_check(struct pt_regs *regs, long error_code);
|
||||||
|
|
|
@ -183,7 +183,7 @@ EXPORT_SYMBOL_GPL(local_apic_timer_c2_ok);
|
||||||
/*
|
/*
|
||||||
* Debug level, exported for io_apic.c
|
* Debug level, exported for io_apic.c
|
||||||
*/
|
*/
|
||||||
unsigned int apic_verbosity;
|
int apic_verbosity;
|
||||||
|
|
||||||
int pic_mode;
|
int pic_mode;
|
||||||
|
|
||||||
|
|
|
@ -76,6 +76,7 @@ static void __used common(void)
|
||||||
BLANK();
|
BLANK();
|
||||||
OFFSET(XEN_vcpu_info_mask, vcpu_info, evtchn_upcall_mask);
|
OFFSET(XEN_vcpu_info_mask, vcpu_info, evtchn_upcall_mask);
|
||||||
OFFSET(XEN_vcpu_info_pending, vcpu_info, evtchn_upcall_pending);
|
OFFSET(XEN_vcpu_info_pending, vcpu_info, evtchn_upcall_pending);
|
||||||
|
OFFSET(XEN_vcpu_info_arch_cr2, vcpu_info, arch.cr2);
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
BLANK();
|
BLANK();
|
||||||
|
|
|
@ -86,9 +86,9 @@ static bool _e820__mapped_any(struct e820_table *table,
|
||||||
continue;
|
continue;
|
||||||
if (entry->addr >= end || entry->addr + entry->size <= start)
|
if (entry->addr >= end || entry->addr + entry->size <= start)
|
||||||
continue;
|
continue;
|
||||||
return 1;
|
return true;
|
||||||
}
|
}
|
||||||
return 0;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
bool e820__mapped_raw_any(u64 start, u64 end, enum e820_type type)
|
bool e820__mapped_raw_any(u64 start, u64 end, enum e820_type type)
|
||||||
|
|
|
@ -29,9 +29,7 @@
|
||||||
#ifdef CONFIG_PARAVIRT_XXL
|
#ifdef CONFIG_PARAVIRT_XXL
|
||||||
#include <asm/asm-offsets.h>
|
#include <asm/asm-offsets.h>
|
||||||
#include <asm/paravirt.h>
|
#include <asm/paravirt.h>
|
||||||
#define GET_CR2_INTO(reg) GET_CR2_INTO_RAX ; movq %rax, reg
|
|
||||||
#else
|
#else
|
||||||
#define GET_CR2_INTO(reg) movq %cr2, reg
|
|
||||||
#define INTERRUPT_RETURN iretq
|
#define INTERRUPT_RETURN iretq
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
|
@ -323,7 +321,7 @@ early_idt_handler_common:
|
||||||
|
|
||||||
cmpq $14,%rsi /* Page fault? */
|
cmpq $14,%rsi /* Page fault? */
|
||||||
jnz 10f
|
jnz 10f
|
||||||
GET_CR2_INTO(%rdi) /* Can clobber any volatile register if pv */
|
GET_CR2_INTO(%rdi) /* can clobber %rax if pv */
|
||||||
call early_make_pgtable
|
call early_make_pgtable
|
||||||
andl %eax,%eax
|
andl %eax,%eax
|
||||||
jz 20f /* All good */
|
jz 20f /* All good */
|
||||||
|
|
|
@ -242,23 +242,23 @@ EXPORT_SYMBOL_GPL(kvm_read_and_reset_pf_reason);
|
||||||
NOKPROBE_SYMBOL(kvm_read_and_reset_pf_reason);
|
NOKPROBE_SYMBOL(kvm_read_and_reset_pf_reason);
|
||||||
|
|
||||||
dotraplinkage void
|
dotraplinkage void
|
||||||
do_async_page_fault(struct pt_regs *regs, unsigned long error_code)
|
do_async_page_fault(struct pt_regs *regs, unsigned long error_code, unsigned long address)
|
||||||
{
|
{
|
||||||
enum ctx_state prev_state;
|
enum ctx_state prev_state;
|
||||||
|
|
||||||
switch (kvm_read_and_reset_pf_reason()) {
|
switch (kvm_read_and_reset_pf_reason()) {
|
||||||
default:
|
default:
|
||||||
do_page_fault(regs, error_code);
|
do_page_fault(regs, error_code, address);
|
||||||
break;
|
break;
|
||||||
case KVM_PV_REASON_PAGE_NOT_PRESENT:
|
case KVM_PV_REASON_PAGE_NOT_PRESENT:
|
||||||
/* page is swapped out by the host. */
|
/* page is swapped out by the host. */
|
||||||
prev_state = exception_enter();
|
prev_state = exception_enter();
|
||||||
kvm_async_pf_task_wait((u32)read_cr2(), !user_mode(regs));
|
kvm_async_pf_task_wait((u32)address, !user_mode(regs));
|
||||||
exception_exit(prev_state);
|
exception_exit(prev_state);
|
||||||
break;
|
break;
|
||||||
case KVM_PV_REASON_PAGE_READY:
|
case KVM_PV_REASON_PAGE_READY:
|
||||||
rcu_irq_enter();
|
rcu_irq_enter();
|
||||||
kvm_async_pf_task_wake((u32)read_cr2());
|
kvm_async_pf_task_wake((u32)address);
|
||||||
rcu_irq_exit();
|
rcu_irq_exit();
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
|
|
|
@ -546,17 +546,15 @@ void __init default_get_smp_config(unsigned int early)
|
||||||
* local APIC has default address
|
* local APIC has default address
|
||||||
*/
|
*/
|
||||||
mp_lapic_addr = APIC_DEFAULT_PHYS_BASE;
|
mp_lapic_addr = APIC_DEFAULT_PHYS_BASE;
|
||||||
return;
|
goto out;
|
||||||
}
|
}
|
||||||
|
|
||||||
pr_info("Default MP configuration #%d\n", mpf->feature1);
|
pr_info("Default MP configuration #%d\n", mpf->feature1);
|
||||||
construct_default_ISA_mptable(mpf->feature1);
|
construct_default_ISA_mptable(mpf->feature1);
|
||||||
|
|
||||||
} else if (mpf->physptr) {
|
} else if (mpf->physptr) {
|
||||||
if (check_physptr(mpf, early)) {
|
if (check_physptr(mpf, early))
|
||||||
early_memunmap(mpf, sizeof(*mpf));
|
goto out;
|
||||||
return;
|
|
||||||
}
|
|
||||||
} else
|
} else
|
||||||
BUG();
|
BUG();
|
||||||
|
|
||||||
|
@ -565,7 +563,7 @@ void __init default_get_smp_config(unsigned int early)
|
||||||
/*
|
/*
|
||||||
* Only use the first configuration found.
|
* Only use the first configuration found.
|
||||||
*/
|
*/
|
||||||
|
out:
|
||||||
early_memunmap(mpf, sizeof(*mpf));
|
early_memunmap(mpf, sizeof(*mpf));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -370,7 +370,7 @@ struct paravirt_patch_template pv_ops = {
|
||||||
.mmu.exit_mmap = paravirt_nop,
|
.mmu.exit_mmap = paravirt_nop,
|
||||||
|
|
||||||
#ifdef CONFIG_PARAVIRT_XXL
|
#ifdef CONFIG_PARAVIRT_XXL
|
||||||
.mmu.read_cr2 = native_read_cr2,
|
.mmu.read_cr2 = __PV_IS_CALLEE_SAVE(native_read_cr2),
|
||||||
.mmu.write_cr2 = native_write_cr2,
|
.mmu.write_cr2 = native_write_cr2,
|
||||||
.mmu.read_cr3 = __native_read_cr3,
|
.mmu.read_cr3 = __native_read_cr3,
|
||||||
.mmu.write_cr3 = native_write_cr3,
|
.mmu.write_cr3 = native_write_cr3,
|
||||||
|
|
|
@ -143,17 +143,7 @@ void __show_regs(struct pt_regs *regs, enum show_regs_mode mode)
|
||||||
|
|
||||||
void release_thread(struct task_struct *dead_task)
|
void release_thread(struct task_struct *dead_task)
|
||||||
{
|
{
|
||||||
if (dead_task->mm) {
|
WARN_ON(dead_task->mm);
|
||||||
#ifdef CONFIG_MODIFY_LDT_SYSCALL
|
|
||||||
if (dead_task->mm->context.ldt) {
|
|
||||||
pr_warn("WARNING: dead process %s still has LDT? <%p/%d>\n",
|
|
||||||
dead_task->comm,
|
|
||||||
dead_task->mm->context.ldt->entries,
|
|
||||||
dead_task->mm->context.ldt->nr_entries);
|
|
||||||
BUG();
|
|
||||||
}
|
|
||||||
#endif
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
enum which_selector {
|
enum which_selector {
|
||||||
|
|
|
@ -369,12 +369,22 @@ static int putreg(struct task_struct *child,
|
||||||
case offsetof(struct user_regs_struct,fs_base):
|
case offsetof(struct user_regs_struct,fs_base):
|
||||||
if (value >= TASK_SIZE_MAX)
|
if (value >= TASK_SIZE_MAX)
|
||||||
return -EIO;
|
return -EIO;
|
||||||
x86_fsbase_write_task(child, value);
|
/*
|
||||||
|
* When changing the FS base, use do_arch_prctl_64()
|
||||||
|
* to set the index to zero and to set the base
|
||||||
|
* as requested.
|
||||||
|
*/
|
||||||
|
if (child->thread.fsbase != value)
|
||||||
|
return do_arch_prctl_64(child, ARCH_SET_FS, value);
|
||||||
return 0;
|
return 0;
|
||||||
case offsetof(struct user_regs_struct,gs_base):
|
case offsetof(struct user_regs_struct,gs_base):
|
||||||
|
/*
|
||||||
|
* Exactly the same here as the %fs handling above.
|
||||||
|
*/
|
||||||
if (value >= TASK_SIZE_MAX)
|
if (value >= TASK_SIZE_MAX)
|
||||||
return -EIO;
|
return -EIO;
|
||||||
x86_gsbase_write_task(child, value);
|
if (child->thread.gsbase != value)
|
||||||
|
return do_arch_prctl_64(child, ARCH_SET_GS, value);
|
||||||
return 0;
|
return 0;
|
||||||
#endif
|
#endif
|
||||||
}
|
}
|
||||||
|
|
|
@ -313,13 +313,10 @@ __visible void __noreturn handle_stack_overflow(const char *message,
|
||||||
|
|
||||||
#ifdef CONFIG_X86_64
|
#ifdef CONFIG_X86_64
|
||||||
/* Runs on IST stack */
|
/* Runs on IST stack */
|
||||||
dotraplinkage void do_double_fault(struct pt_regs *regs, long error_code)
|
dotraplinkage void do_double_fault(struct pt_regs *regs, long error_code, unsigned long cr2)
|
||||||
{
|
{
|
||||||
static const char str[] = "double fault";
|
static const char str[] = "double fault";
|
||||||
struct task_struct *tsk = current;
|
struct task_struct *tsk = current;
|
||||||
#ifdef CONFIG_VMAP_STACK
|
|
||||||
unsigned long cr2;
|
|
||||||
#endif
|
|
||||||
|
|
||||||
#ifdef CONFIG_X86_ESPFIX64
|
#ifdef CONFIG_X86_ESPFIX64
|
||||||
extern unsigned char native_irq_return_iret[];
|
extern unsigned char native_irq_return_iret[];
|
||||||
|
@ -415,7 +412,6 @@ dotraplinkage void do_double_fault(struct pt_regs *regs, long error_code)
|
||||||
* stack even if the actual trigger for the double fault was
|
* stack even if the actual trigger for the double fault was
|
||||||
* something else.
|
* something else.
|
||||||
*/
|
*/
|
||||||
cr2 = read_cr2();
|
|
||||||
if ((unsigned long)task_stack_page(tsk) - 1 - cr2 < PAGE_SIZE)
|
if ((unsigned long)task_stack_page(tsk) - 1 - cr2 < PAGE_SIZE)
|
||||||
handle_stack_overflow("kernel stack overflow (double-fault)", regs, cr2);
|
handle_stack_overflow("kernel stack overflow (double-fault)", regs, cr2);
|
||||||
#endif
|
#endif
|
||||||
|
|
|
@ -177,7 +177,7 @@ static inline void reg_copy(FPU_REG const *x, FPU_REG *y)
|
||||||
#define setexponentpos(x,y) { (*(short *)&((x)->exp)) = \
|
#define setexponentpos(x,y) { (*(short *)&((x)->exp)) = \
|
||||||
((y) + EXTENDED_Ebias) & 0x7fff; }
|
((y) + EXTENDED_Ebias) & 0x7fff; }
|
||||||
#define exponent16(x) (*(short *)&((x)->exp))
|
#define exponent16(x) (*(short *)&((x)->exp))
|
||||||
#define setexponent16(x,y) { (*(short *)&((x)->exp)) = (y); }
|
#define setexponent16(x,y) { (*(short *)&((x)->exp)) = (u16)(y); }
|
||||||
#define addexponent(x,y) { (*(short *)&((x)->exp)) += (y); }
|
#define addexponent(x,y) { (*(short *)&((x)->exp)) += (y); }
|
||||||
#define stdexp(x) { (*(short *)&((x)->exp)) += EXTENDED_Ebias; }
|
#define stdexp(x) { (*(short *)&((x)->exp)) += EXTENDED_Ebias; }
|
||||||
|
|
||||||
|
|
|
@ -18,7 +18,7 @@
|
||||||
#include "control_w.h"
|
#include "control_w.h"
|
||||||
|
|
||||||
#define MAKE_REG(s, e, l, h) { l, h, \
|
#define MAKE_REG(s, e, l, h) { l, h, \
|
||||||
((EXTENDED_Ebias+(e)) | ((SIGN_##s != 0)*0x8000)) }
|
(u16)((EXTENDED_Ebias+(e)) | ((SIGN_##s != 0)*0x8000)) }
|
||||||
|
|
||||||
FPU_REG const CONST_1 = MAKE_REG(POS, 0, 0x00000000, 0x80000000);
|
FPU_REG const CONST_1 = MAKE_REG(POS, 0, 0x00000000, 0x80000000);
|
||||||
#if 0
|
#if 0
|
||||||
|
|
|
@ -1490,9 +1490,8 @@ good_area:
|
||||||
NOKPROBE_SYMBOL(do_user_addr_fault);
|
NOKPROBE_SYMBOL(do_user_addr_fault);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* This routine handles page faults. It determines the address,
|
* Explicitly marked noinline such that the function tracer sees this as the
|
||||||
* and the problem, and then passes it off to one of the appropriate
|
* page_fault entry point.
|
||||||
* routines.
|
|
||||||
*/
|
*/
|
||||||
static noinline void
|
static noinline void
|
||||||
__do_page_fault(struct pt_regs *regs, unsigned long hw_error_code,
|
__do_page_fault(struct pt_regs *regs, unsigned long hw_error_code,
|
||||||
|
@ -1511,33 +1510,26 @@ __do_page_fault(struct pt_regs *regs, unsigned long hw_error_code,
|
||||||
}
|
}
|
||||||
NOKPROBE_SYMBOL(__do_page_fault);
|
NOKPROBE_SYMBOL(__do_page_fault);
|
||||||
|
|
||||||
static nokprobe_inline void
|
static __always_inline void
|
||||||
trace_page_fault_entries(unsigned long address, struct pt_regs *regs,
|
trace_page_fault_entries(struct pt_regs *regs, unsigned long error_code,
|
||||||
unsigned long error_code)
|
unsigned long address)
|
||||||
{
|
{
|
||||||
|
if (!trace_pagefault_enabled())
|
||||||
|
return;
|
||||||
|
|
||||||
if (user_mode(regs))
|
if (user_mode(regs))
|
||||||
trace_page_fault_user(address, regs, error_code);
|
trace_page_fault_user(address, regs, error_code);
|
||||||
else
|
else
|
||||||
trace_page_fault_kernel(address, regs, error_code);
|
trace_page_fault_kernel(address, regs, error_code);
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
dotraplinkage void
|
||||||
* We must have this function blacklisted from kprobes, tagged with notrace
|
do_page_fault(struct pt_regs *regs, unsigned long error_code, unsigned long address)
|
||||||
* and call read_cr2() before calling anything else. To avoid calling any
|
|
||||||
* kind of tracing machinery before we've observed the CR2 value.
|
|
||||||
*
|
|
||||||
* exception_{enter,exit}() contains all sorts of tracepoints.
|
|
||||||
*/
|
|
||||||
dotraplinkage void notrace
|
|
||||||
do_page_fault(struct pt_regs *regs, unsigned long error_code)
|
|
||||||
{
|
{
|
||||||
unsigned long address = read_cr2(); /* Get the faulting address */
|
|
||||||
enum ctx_state prev_state;
|
enum ctx_state prev_state;
|
||||||
|
|
||||||
prev_state = exception_enter();
|
prev_state = exception_enter();
|
||||||
if (trace_pagefault_enabled())
|
trace_page_fault_entries(regs, error_code, address);
|
||||||
trace_page_fault_entries(address, regs, error_code);
|
|
||||||
|
|
||||||
__do_page_fault(regs, error_code, address);
|
__do_page_fault(regs, error_code, address);
|
||||||
exception_exit(prev_state);
|
exception_exit(prev_state);
|
||||||
}
|
}
|
||||||
|
|
|
@ -41,7 +41,7 @@ EXPORT_SYMBOL_GPL(sev_enable_key);
|
||||||
bool sev_enabled __section(.data);
|
bool sev_enabled __section(.data);
|
||||||
|
|
||||||
/* Buffer used for early in-place encryption by BSP, no locking needed */
|
/* Buffer used for early in-place encryption by BSP, no locking needed */
|
||||||
static char sme_early_buffer[PAGE_SIZE] __aligned(PAGE_SIZE);
|
static char sme_early_buffer[PAGE_SIZE] __initdata __aligned(PAGE_SIZE);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* This routine does not change the underlying encryption setting of the
|
* This routine does not change the underlying encryption setting of the
|
||||||
|
|
|
@ -998,7 +998,8 @@ void __init xen_setup_vcpu_info_placement(void)
|
||||||
__PV_IS_CALLEE_SAVE(xen_irq_disable_direct);
|
__PV_IS_CALLEE_SAVE(xen_irq_disable_direct);
|
||||||
pv_ops.irq.irq_enable =
|
pv_ops.irq.irq_enable =
|
||||||
__PV_IS_CALLEE_SAVE(xen_irq_enable_direct);
|
__PV_IS_CALLEE_SAVE(xen_irq_enable_direct);
|
||||||
pv_ops.mmu.read_cr2 = xen_read_cr2_direct;
|
pv_ops.mmu.read_cr2 =
|
||||||
|
__PV_IS_CALLEE_SAVE(xen_read_cr2_direct);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -1307,16 +1307,6 @@ static void xen_write_cr2(unsigned long cr2)
|
||||||
this_cpu_read(xen_vcpu)->arch.cr2 = cr2;
|
this_cpu_read(xen_vcpu)->arch.cr2 = cr2;
|
||||||
}
|
}
|
||||||
|
|
||||||
static unsigned long xen_read_cr2(void)
|
|
||||||
{
|
|
||||||
return this_cpu_read(xen_vcpu)->arch.cr2;
|
|
||||||
}
|
|
||||||
|
|
||||||
unsigned long xen_read_cr2_direct(void)
|
|
||||||
{
|
|
||||||
return this_cpu_read(xen_vcpu_info.arch.cr2);
|
|
||||||
}
|
|
||||||
|
|
||||||
static noinline void xen_flush_tlb(void)
|
static noinline void xen_flush_tlb(void)
|
||||||
{
|
{
|
||||||
struct mmuext_op *op;
|
struct mmuext_op *op;
|
||||||
|
@ -2397,7 +2387,7 @@ static void xen_leave_lazy_mmu(void)
|
||||||
}
|
}
|
||||||
|
|
||||||
static const struct pv_mmu_ops xen_mmu_ops __initconst = {
|
static const struct pv_mmu_ops xen_mmu_ops __initconst = {
|
||||||
.read_cr2 = xen_read_cr2,
|
.read_cr2 = __PV_IS_CALLEE_SAVE(xen_read_cr2),
|
||||||
.write_cr2 = xen_write_cr2,
|
.write_cr2 = xen_write_cr2,
|
||||||
|
|
||||||
.read_cr3 = xen_read_cr3,
|
.read_cr3 = xen_read_cr3,
|
||||||
|
|
|
@ -10,6 +10,7 @@
|
||||||
#include <asm/percpu.h>
|
#include <asm/percpu.h>
|
||||||
#include <asm/processor-flags.h>
|
#include <asm/processor-flags.h>
|
||||||
#include <asm/frame.h>
|
#include <asm/frame.h>
|
||||||
|
#include <asm/asm.h>
|
||||||
|
|
||||||
#include <linux/linkage.h>
|
#include <linux/linkage.h>
|
||||||
|
|
||||||
|
@ -135,3 +136,18 @@ ENTRY(check_events)
|
||||||
FRAME_END
|
FRAME_END
|
||||||
ret
|
ret
|
||||||
ENDPROC(check_events)
|
ENDPROC(check_events)
|
||||||
|
|
||||||
|
ENTRY(xen_read_cr2)
|
||||||
|
FRAME_BEGIN
|
||||||
|
_ASM_MOV PER_CPU_VAR(xen_vcpu), %_ASM_AX
|
||||||
|
_ASM_MOV XEN_vcpu_info_arch_cr2(%_ASM_AX), %_ASM_AX
|
||||||
|
FRAME_END
|
||||||
|
ret
|
||||||
|
ENDPROC(xen_read_cr2);
|
||||||
|
|
||||||
|
ENTRY(xen_read_cr2_direct)
|
||||||
|
FRAME_BEGIN
|
||||||
|
_ASM_MOV PER_CPU_VAR(xen_vcpu_info) + XEN_vcpu_info_arch_cr2, %_ASM_AX
|
||||||
|
FRAME_END
|
||||||
|
ret
|
||||||
|
ENDPROC(xen_read_cr2_direct);
|
||||||
|
|
|
@ -134,6 +134,9 @@ __visible void xen_irq_disable_direct(void);
|
||||||
__visible unsigned long xen_save_fl_direct(void);
|
__visible unsigned long xen_save_fl_direct(void);
|
||||||
__visible void xen_restore_fl_direct(unsigned long);
|
__visible void xen_restore_fl_direct(unsigned long);
|
||||||
|
|
||||||
|
__visible unsigned long xen_read_cr2(void);
|
||||||
|
__visible unsigned long xen_read_cr2_direct(void);
|
||||||
|
|
||||||
/* These are not functions, and cannot be called normally */
|
/* These are not functions, and cannot be called normally */
|
||||||
__visible void xen_iret(void);
|
__visible void xen_iret(void);
|
||||||
__visible void xen_sysret32(void);
|
__visible void xen_sysret32(void);
|
||||||
|
|
|
@ -489,25 +489,11 @@ static void test_ptrace_write_gsbase(void)
|
||||||
* selector value is changed or not by the GSBASE write in
|
* selector value is changed or not by the GSBASE write in
|
||||||
* a ptracer.
|
* a ptracer.
|
||||||
*/
|
*/
|
||||||
if (gs != *shared_scratch) {
|
if (gs == 0 && base == 0xFF) {
|
||||||
nerrs++;
|
printf("[OK]\tGS was reset as expected\n");
|
||||||
printf("[FAIL]\tGS changed to %lx\n", gs);
|
|
||||||
|
|
||||||
/*
|
|
||||||
* On older kernels, poking a nonzero value into the
|
|
||||||
* base would zero the selector. On newer kernels,
|
|
||||||
* this behavior has changed -- poking the base
|
|
||||||
* changes only the base and, if FSGSBASE is not
|
|
||||||
* available, this may have no effect.
|
|
||||||
*/
|
|
||||||
if (gs == 0)
|
|
||||||
printf("\tNote: this is expected behavior on older kernels.\n");
|
|
||||||
} else if (have_fsgsbase && (base != 0xFF)) {
|
|
||||||
nerrs++;
|
|
||||||
printf("[FAIL]\tGSBASE changed to %lx\n", base);
|
|
||||||
} else {
|
} else {
|
||||||
printf("[OK]\tGS remained 0x%hx%s", *shared_scratch, have_fsgsbase ? " and GSBASE changed to 0xFF" : "");
|
nerrs++;
|
||||||
printf("\n");
|
printf("[FAIL]\tGS=0x%lx, GSBASE=0x%lx (should be 0, 0xFF)\n", gs, base);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue