Merge branch 'x86-platform-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip
Pull x86/platform changes from Ingo Molnar: "This cleans up some Xen-induced pagetable init code uglies, by generalizing new platform callbacks and state: x86_init.paging.*" * 'x86-platform-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip: x86: Document x86_init.paging.pagetable_init() x86: xen: Cleanup and remove x86_init.paging.pagetable_setup_done() x86: Move paging_init() call to x86_init.paging.pagetable_init() x86: Rename pagetable_setup_start() to pagetable_init() x86: Remove base argument from x86_init.paging.pagetable_setup_start
This commit is contained in:
commit
a5fa7b7d8f
|
@ -303,11 +303,9 @@ void set_pte_vaddr(unsigned long vaddr, pte_t pte);
|
||||||
|
|
||||||
extern void native_pagetable_reserve(u64 start, u64 end);
|
extern void native_pagetable_reserve(u64 start, u64 end);
|
||||||
#ifdef CONFIG_X86_32
|
#ifdef CONFIG_X86_32
|
||||||
extern void native_pagetable_setup_start(pgd_t *base);
|
extern void native_pagetable_init(void);
|
||||||
extern void native_pagetable_setup_done(pgd_t *base);
|
|
||||||
#else
|
#else
|
||||||
#define native_pagetable_setup_start x86_init_pgd_noop
|
#define native_pagetable_init paging_init
|
||||||
#define native_pagetable_setup_done x86_init_pgd_noop
|
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
struct seq_file;
|
struct seq_file;
|
||||||
|
|
|
@ -81,12 +81,13 @@ struct x86_init_mapping {
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* struct x86_init_paging - platform specific paging functions
|
* struct x86_init_paging - platform specific paging functions
|
||||||
* @pagetable_setup_start: platform specific pre paging_init() call
|
* @pagetable_init: platform specific paging initialization call to setup
|
||||||
* @pagetable_setup_done: platform specific post paging_init() call
|
* the kernel pagetables and prepare accessors functions.
|
||||||
|
* Callback must call paging_init(). Called once after the
|
||||||
|
* direct mapping for phys memory is available.
|
||||||
*/
|
*/
|
||||||
struct x86_init_paging {
|
struct x86_init_paging {
|
||||||
void (*pagetable_setup_start)(pgd_t *base);
|
void (*pagetable_init)(void);
|
||||||
void (*pagetable_setup_done)(pgd_t *base);
|
|
||||||
};
|
};
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
|
|
@ -961,9 +961,7 @@ void __init setup_arch(char **cmdline_p)
|
||||||
kvmclock_init();
|
kvmclock_init();
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
x86_init.paging.pagetable_setup_start(swapper_pg_dir);
|
x86_init.paging.pagetable_init();
|
||||||
paging_init();
|
|
||||||
x86_init.paging.pagetable_setup_done(swapper_pg_dir);
|
|
||||||
|
|
||||||
if (boot_cpu_data.cpuid_level >= 0) {
|
if (boot_cpu_data.cpuid_level >= 0) {
|
||||||
/* A CPU has %cr4 if and only if it has CPUID */
|
/* A CPU has %cr4 if and only if it has CPUID */
|
||||||
|
|
|
@ -26,7 +26,6 @@
|
||||||
|
|
||||||
void __cpuinit x86_init_noop(void) { }
|
void __cpuinit x86_init_noop(void) { }
|
||||||
void __init x86_init_uint_noop(unsigned int unused) { }
|
void __init x86_init_uint_noop(unsigned int unused) { }
|
||||||
void __init x86_init_pgd_noop(pgd_t *unused) { }
|
|
||||||
int __init iommu_init_noop(void) { return 0; }
|
int __init iommu_init_noop(void) { return 0; }
|
||||||
void iommu_shutdown_noop(void) { }
|
void iommu_shutdown_noop(void) { }
|
||||||
|
|
||||||
|
@ -68,8 +67,7 @@ struct x86_init_ops x86_init __initdata = {
|
||||||
},
|
},
|
||||||
|
|
||||||
.paging = {
|
.paging = {
|
||||||
.pagetable_setup_start = native_pagetable_setup_start,
|
.pagetable_init = native_pagetable_init,
|
||||||
.pagetable_setup_done = native_pagetable_setup_done,
|
|
||||||
},
|
},
|
||||||
|
|
||||||
.timers = {
|
.timers = {
|
||||||
|
|
|
@ -445,10 +445,10 @@ static inline void permanent_kmaps_init(pgd_t *pgd_base)
|
||||||
}
|
}
|
||||||
#endif /* CONFIG_HIGHMEM */
|
#endif /* CONFIG_HIGHMEM */
|
||||||
|
|
||||||
void __init native_pagetable_setup_start(pgd_t *base)
|
void __init native_pagetable_init(void)
|
||||||
{
|
{
|
||||||
unsigned long pfn, va;
|
unsigned long pfn, va;
|
||||||
pgd_t *pgd;
|
pgd_t *pgd, *base = swapper_pg_dir;
|
||||||
pud_t *pud;
|
pud_t *pud;
|
||||||
pmd_t *pmd;
|
pmd_t *pmd;
|
||||||
pte_t *pte;
|
pte_t *pte;
|
||||||
|
@ -475,10 +475,7 @@ void __init native_pagetable_setup_start(pgd_t *base)
|
||||||
pte_clear(NULL, va, pte);
|
pte_clear(NULL, va, pte);
|
||||||
}
|
}
|
||||||
paravirt_alloc_pmd(&init_mm, __pa(base) >> PAGE_SHIFT);
|
paravirt_alloc_pmd(&init_mm, __pa(base) >> PAGE_SHIFT);
|
||||||
}
|
paging_init();
|
||||||
|
|
||||||
void __init native_pagetable_setup_done(pgd_t *base)
|
|
||||||
{
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@ -493,7 +490,7 @@ void __init native_pagetable_setup_done(pgd_t *base)
|
||||||
* If we're booting paravirtualized under a hypervisor, then there are
|
* If we're booting paravirtualized under a hypervisor, then there are
|
||||||
* more options: we may already be running PAE, and the pagetable may
|
* more options: we may already be running PAE, and the pagetable may
|
||||||
* or may not be based in swapper_pg_dir. In any case,
|
* or may not be based in swapper_pg_dir. In any case,
|
||||||
* paravirt_pagetable_setup_start() will set up swapper_pg_dir
|
* paravirt_pagetable_init() will set up swapper_pg_dir
|
||||||
* appropriately for the rest of the initialization to work.
|
* appropriately for the rest of the initialization to work.
|
||||||
*
|
*
|
||||||
* In general, pagetable_init() assumes that the pagetable may already
|
* In general, pagetable_init() assumes that the pagetable may already
|
||||||
|
|
|
@ -1174,8 +1174,13 @@ static void xen_exit_mmap(struct mm_struct *mm)
|
||||||
spin_unlock(&mm->page_table_lock);
|
spin_unlock(&mm->page_table_lock);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void __init xen_pagetable_setup_start(pgd_t *base)
|
static void xen_post_allocator_init(void);
|
||||||
|
|
||||||
|
static void __init xen_pagetable_init(void)
|
||||||
{
|
{
|
||||||
|
paging_init();
|
||||||
|
xen_setup_shared_info();
|
||||||
|
xen_post_allocator_init();
|
||||||
}
|
}
|
||||||
|
|
||||||
static __init void xen_mapping_pagetable_reserve(u64 start, u64 end)
|
static __init void xen_mapping_pagetable_reserve(u64 start, u64 end)
|
||||||
|
@ -1192,14 +1197,6 @@ static __init void xen_mapping_pagetable_reserve(u64 start, u64 end)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
static void xen_post_allocator_init(void);
|
|
||||||
|
|
||||||
static void __init xen_pagetable_setup_done(pgd_t *base)
|
|
||||||
{
|
|
||||||
xen_setup_shared_info();
|
|
||||||
xen_post_allocator_init();
|
|
||||||
}
|
|
||||||
|
|
||||||
static void xen_write_cr2(unsigned long cr2)
|
static void xen_write_cr2(unsigned long cr2)
|
||||||
{
|
{
|
||||||
this_cpu_read(xen_vcpu)->arch.cr2 = cr2;
|
this_cpu_read(xen_vcpu)->arch.cr2 = cr2;
|
||||||
|
@ -2068,8 +2065,7 @@ static const struct pv_mmu_ops xen_mmu_ops __initconst = {
|
||||||
void __init xen_init_mmu_ops(void)
|
void __init xen_init_mmu_ops(void)
|
||||||
{
|
{
|
||||||
x86_init.mapping.pagetable_reserve = xen_mapping_pagetable_reserve;
|
x86_init.mapping.pagetable_reserve = xen_mapping_pagetable_reserve;
|
||||||
x86_init.paging.pagetable_setup_start = xen_pagetable_setup_start;
|
x86_init.paging.pagetable_init = xen_pagetable_init;
|
||||||
x86_init.paging.pagetable_setup_done = xen_pagetable_setup_done;
|
|
||||||
pv_mmu_ops = xen_mmu_ops;
|
pv_mmu_ops = xen_mmu_ops;
|
||||||
|
|
||||||
memset(dummy_mapping, 0xff, PAGE_SIZE);
|
memset(dummy_mapping, 0xff, PAGE_SIZE);
|
||||||
|
|
Loading…
Reference in New Issue