powerpc/mm: move update_mmu_cache() into book3s hash utils.
update_mmu_cache() is only for BOOK3S, and can be simplified for BOOK3S32. Move it out of mem.c into respective BOOK3S32 and BOOK3S64 files containing hash utils. BOOK3S64 version of hash_preload() is only used locally, declare it static. Remove the radix_enabled() stuff in BOOK3S32 version. Signed-off-by: Christophe Leroy <christophe.leroy@c-s.fr> Signed-off-by: Michael Ellerman <mpe@ellerman.id.au> Link: https://lore.kernel.org/r/107aaf43583a5f5d09e0d4e84c4c4390ecfcd512.1565933217.git.christophe.leroy@c-s.fr
This commit is contained in:
parent
4c1616ef03
commit
e5a1edb9fe
|
@ -309,6 +309,52 @@ void hash_preload(struct mm_struct *mm, unsigned long ea,
|
|||
add_hash_page(mm->context.id, ea, pmd_val(*pmd));
|
||||
}
|
||||
|
||||
/*
|
||||
* This is called at the end of handling a user page fault, when the
|
||||
* fault has been handled by updating a PTE in the linux page tables.
|
||||
* We use it to preload an HPTE into the hash table corresponding to
|
||||
* the updated linux PTE.
|
||||
*
|
||||
* This must always be called with the pte lock held.
|
||||
*/
|
||||
void update_mmu_cache(struct vm_area_struct *vma, unsigned long address,
|
||||
pte_t *ptep)
|
||||
{
|
||||
/*
|
||||
* We don't need to worry about _PAGE_PRESENT here because we are
|
||||
* called with either mm->page_table_lock held or ptl lock held
|
||||
*/
|
||||
unsigned long trap;
|
||||
bool is_exec;
|
||||
|
||||
/* We only want HPTEs for linux PTEs that have _PAGE_ACCESSED set */
|
||||
if (!pte_young(*ptep) || address >= TASK_SIZE)
|
||||
return;
|
||||
|
||||
/*
|
||||
* We try to figure out if we are coming from an instruction
|
||||
* access fault and pass that down to __hash_page so we avoid
|
||||
* double-faulting on execution of fresh text. We have to test
|
||||
* for regs NULL since init will get here first thing at boot.
|
||||
*
|
||||
* We also avoid filling the hash if not coming from a fault.
|
||||
*/
|
||||
|
||||
trap = current->thread.regs ? TRAP(current->thread.regs) : 0UL;
|
||||
switch (trap) {
|
||||
case 0x300:
|
||||
is_exec = false;
|
||||
break;
|
||||
case 0x400:
|
||||
is_exec = true;
|
||||
break;
|
||||
default:
|
||||
return;
|
||||
}
|
||||
|
||||
hash_preload(vma->vm_mm, address, is_exec, trap);
|
||||
}
|
||||
|
||||
/*
|
||||
* Initialize the hash table and patch the instructions in hashtable.S.
|
||||
*/
|
||||
|
|
|
@ -1519,8 +1519,8 @@ static bool should_hash_preload(struct mm_struct *mm, unsigned long ea)
|
|||
}
|
||||
#endif
|
||||
|
||||
void hash_preload(struct mm_struct *mm, unsigned long ea,
|
||||
bool is_exec, unsigned long trap)
|
||||
static void hash_preload(struct mm_struct *mm, unsigned long ea,
|
||||
bool is_exec, unsigned long trap)
|
||||
{
|
||||
int hugepage_shift;
|
||||
unsigned long vsid;
|
||||
|
@ -1600,6 +1600,57 @@ out_exit:
|
|||
local_irq_restore(flags);
|
||||
}
|
||||
|
||||
/*
|
||||
* This is called at the end of handling a user page fault, when the
|
||||
* fault has been handled by updating a PTE in the linux page tables.
|
||||
* We use it to preload an HPTE into the hash table corresponding to
|
||||
* the updated linux PTE.
|
||||
*
|
||||
* This must always be called with the pte lock held.
|
||||
*/
|
||||
void update_mmu_cache(struct vm_area_struct *vma, unsigned long address,
|
||||
pte_t *ptep)
|
||||
{
|
||||
/*
|
||||
* We don't need to worry about _PAGE_PRESENT here because we are
|
||||
* called with either mm->page_table_lock held or ptl lock held
|
||||
*/
|
||||
unsigned long trap;
|
||||
bool is_exec;
|
||||
|
||||
if (radix_enabled()) {
|
||||
prefetch((void *)address);
|
||||
return;
|
||||
}
|
||||
|
||||
/* We only want HPTEs for linux PTEs that have _PAGE_ACCESSED set */
|
||||
if (!pte_young(*ptep) || address >= TASK_SIZE)
|
||||
return;
|
||||
|
||||
/*
|
||||
* We try to figure out if we are coming from an instruction
|
||||
* access fault and pass that down to __hash_page so we avoid
|
||||
* double-faulting on execution of fresh text. We have to test
|
||||
* for regs NULL since init will get here first thing at boot.
|
||||
*
|
||||
* We also avoid filling the hash if not coming from a fault.
|
||||
*/
|
||||
|
||||
trap = current->thread.regs ? TRAP(current->thread.regs) : 0UL;
|
||||
switch (trap) {
|
||||
case 0x300:
|
||||
is_exec = false;
|
||||
break;
|
||||
case 0x400:
|
||||
is_exec = true;
|
||||
break;
|
||||
default:
|
||||
return;
|
||||
}
|
||||
|
||||
hash_preload(vma->vm_mm, address, is_exec, trap);
|
||||
}
|
||||
|
||||
#ifdef CONFIG_PPC_MEM_KEYS
|
||||
/*
|
||||
* Return the protection key associated with the given address and the
|
||||
|
|
|
@ -407,58 +407,6 @@ void flush_icache_user_range(struct vm_area_struct *vma, struct page *page,
|
|||
}
|
||||
EXPORT_SYMBOL(flush_icache_user_range);
|
||||
|
||||
/*
|
||||
* This is called at the end of handling a user page fault, when the
|
||||
* fault has been handled by updating a PTE in the linux page tables.
|
||||
* We use it to preload an HPTE into the hash table corresponding to
|
||||
* the updated linux PTE.
|
||||
*
|
||||
* This must always be called with the pte lock held.
|
||||
*/
|
||||
#ifdef CONFIG_PPC_BOOK3S
|
||||
void update_mmu_cache(struct vm_area_struct *vma, unsigned long address,
|
||||
pte_t *ptep)
|
||||
{
|
||||
/*
|
||||
* We don't need to worry about _PAGE_PRESENT here because we are
|
||||
* called with either mm->page_table_lock held or ptl lock held
|
||||
*/
|
||||
unsigned long trap;
|
||||
bool is_exec;
|
||||
|
||||
if (radix_enabled()) {
|
||||
prefetch((void *)address);
|
||||
return;
|
||||
}
|
||||
|
||||
/* We only want HPTEs for linux PTEs that have _PAGE_ACCESSED set */
|
||||
if (!pte_young(*ptep) || address >= TASK_SIZE)
|
||||
return;
|
||||
|
||||
/* We try to figure out if we are coming from an instruction
|
||||
* access fault and pass that down to __hash_page so we avoid
|
||||
* double-faulting on execution of fresh text. We have to test
|
||||
* for regs NULL since init will get here first thing at boot
|
||||
*
|
||||
* We also avoid filling the hash if not coming from a fault
|
||||
*/
|
||||
|
||||
trap = current->thread.regs ? TRAP(current->thread.regs) : 0UL;
|
||||
switch (trap) {
|
||||
case 0x300:
|
||||
is_exec = false;
|
||||
break;
|
||||
case 0x400:
|
||||
is_exec = true;
|
||||
break;
|
||||
default:
|
||||
return;
|
||||
}
|
||||
|
||||
hash_preload(vma->vm_mm, address, is_exec, trap);
|
||||
}
|
||||
#endif /* CONFIG_PPC_BOOK3S */
|
||||
|
||||
/*
|
||||
* System memory should not be in /proc/iomem but various tools expect it
|
||||
* (eg kdump).
|
||||
|
|
|
@ -82,10 +82,6 @@ static inline void print_system_hash_info(void) {}
|
|||
|
||||
#else /* CONFIG_PPC_MMU_NOHASH */
|
||||
|
||||
extern void hash_preload(struct mm_struct *mm, unsigned long ea,
|
||||
bool is_exec, unsigned long trap);
|
||||
|
||||
|
||||
extern void _tlbie(unsigned long address);
|
||||
extern void _tlbia(void);
|
||||
|
||||
|
@ -95,6 +91,9 @@ void print_system_hash_info(void);
|
|||
|
||||
#ifdef CONFIG_PPC32
|
||||
|
||||
void hash_preload(struct mm_struct *mm, unsigned long ea,
|
||||
bool is_exec, unsigned long trap);
|
||||
|
||||
extern void mapin_ram(void);
|
||||
extern void setbat(int index, unsigned long virt, phys_addr_t phys,
|
||||
unsigned int size, pgprot_t prot);
|
||||
|
|
Loading…
Reference in New Issue