powerpc/32s: Inline flush_tlb_range() and flush_tlb_kernel_range()
flush_tlb_range() and flush_tlb_kernel_range() are trivial calls to flush_range(). Make flush_range() global and inline flush_tlb_range() and flush_tlb_kernel_range(). Signed-off-by: Christophe Leroy <christophe.leroy@csgroup.eu> Signed-off-by: Michael Ellerman <mpe@ellerman.id.au> Link: https://lore.kernel.org/r/c7029a78e78709ad9272d7a44260e06b649169b2.1603348103.git.christophe.leroy@csgroup.eu
This commit is contained in:
parent
fd1b4b7f51
commit
1e83396f29
|
@ -8,9 +8,7 @@
|
||||||
*/
|
*/
|
||||||
void hash__flush_tlb_mm(struct mm_struct *mm);
|
void hash__flush_tlb_mm(struct mm_struct *mm);
|
||||||
void hash__flush_tlb_page(struct vm_area_struct *vma, unsigned long vmaddr);
|
void hash__flush_tlb_page(struct vm_area_struct *vma, unsigned long vmaddr);
|
||||||
extern void flush_tlb_range(struct vm_area_struct *vma, unsigned long start,
|
void flush_range(struct mm_struct *mm, unsigned long start, unsigned long end);
|
||||||
unsigned long end);
|
|
||||||
extern void flush_tlb_kernel_range(unsigned long start, unsigned long end);
|
|
||||||
|
|
||||||
#ifdef CONFIG_SMP
|
#ifdef CONFIG_SMP
|
||||||
void _tlbie(unsigned long address);
|
void _tlbie(unsigned long address);
|
||||||
|
@ -38,6 +36,17 @@ static inline void flush_tlb_page(struct vm_area_struct *vma, unsigned long vmad
|
||||||
_tlbie(vmaddr);
|
_tlbie(vmaddr);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static inline void
|
||||||
|
flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end)
|
||||||
|
{
|
||||||
|
flush_range(vma->vm_mm, start, end);
|
||||||
|
}
|
||||||
|
|
||||||
|
static inline void flush_tlb_kernel_range(unsigned long start, unsigned long end)
|
||||||
|
{
|
||||||
|
flush_range(&init_mm, start, end);
|
||||||
|
}
|
||||||
|
|
||||||
static inline void local_flush_tlb_page(struct vm_area_struct *vma,
|
static inline void local_flush_tlb_page(struct vm_area_struct *vma,
|
||||||
unsigned long vmaddr)
|
unsigned long vmaddr)
|
||||||
{
|
{
|
||||||
|
|
|
@ -71,8 +71,12 @@ void tlb_flush(struct mmu_gather *tlb)
|
||||||
* -- Cort
|
* -- Cort
|
||||||
*/
|
*/
|
||||||
|
|
||||||
static void flush_range(struct mm_struct *mm, unsigned long start,
|
/*
|
||||||
unsigned long end)
|
* For each address in the range, find the pte for the address
|
||||||
|
* and check _PAGE_HASHPTE bit; if it is set, find and destroy
|
||||||
|
* the corresponding HPTE.
|
||||||
|
*/
|
||||||
|
void flush_range(struct mm_struct *mm, unsigned long start, unsigned long end)
|
||||||
{
|
{
|
||||||
pmd_t *pmd;
|
pmd_t *pmd;
|
||||||
unsigned long pmd_end;
|
unsigned long pmd_end;
|
||||||
|
@ -105,15 +109,7 @@ static void flush_range(struct mm_struct *mm, unsigned long start,
|
||||||
++pmd;
|
++pmd;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
EXPORT_SYMBOL(flush_range);
|
||||||
/*
|
|
||||||
* Flush kernel TLB entries in the given range
|
|
||||||
*/
|
|
||||||
void flush_tlb_kernel_range(unsigned long start, unsigned long end)
|
|
||||||
{
|
|
||||||
flush_range(&init_mm, start, end);
|
|
||||||
}
|
|
||||||
EXPORT_SYMBOL(flush_tlb_kernel_range);
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Flush all the (user) entries for the address space described by mm.
|
* Flush all the (user) entries for the address space described by mm.
|
||||||
|
@ -145,18 +141,6 @@ void hash__flush_tlb_page(struct vm_area_struct *vma, unsigned long vmaddr)
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(hash__flush_tlb_page);
|
EXPORT_SYMBOL(hash__flush_tlb_page);
|
||||||
|
|
||||||
/*
|
|
||||||
* For each address in the range, find the pte for the address
|
|
||||||
* and check _PAGE_HASHPTE bit; if it is set, find and destroy
|
|
||||||
* the corresponding HPTE.
|
|
||||||
*/
|
|
||||||
void flush_tlb_range(struct vm_area_struct *vma, unsigned long start,
|
|
||||||
unsigned long end)
|
|
||||||
{
|
|
||||||
flush_range(vma->vm_mm, start, end);
|
|
||||||
}
|
|
||||||
EXPORT_SYMBOL(flush_tlb_range);
|
|
||||||
|
|
||||||
void __init early_init_mmu(void)
|
void __init early_init_mmu(void)
|
||||||
{
|
{
|
||||||
}
|
}
|
||||||
|
|
Loading…
Reference in New Issue