69 lines
1.6 KiB
C
69 lines
1.6 KiB
C
#ifndef _SPARC64_TLBFLUSH_H
|
|
#define _SPARC64_TLBFLUSH_H
|
|
|
|
#include <asm/mmu_context.h>
|
|
|
|
/* TSB flush operations. */
|
|
|
|
#define TLB_BATCH_NR 192
|
|
|
|
struct tlb_batch {
|
|
struct mm_struct *mm;
|
|
unsigned long tlb_nr;
|
|
unsigned long active;
|
|
unsigned long vaddrs[TLB_BATCH_NR];
|
|
};
|
|
|
|
void flush_tsb_kernel_range(unsigned long start, unsigned long end);
|
|
void flush_tsb_user(struct tlb_batch *tb);
|
|
void flush_tsb_user_page(struct mm_struct *mm, unsigned long vaddr);
|
|
|
|
/* TLB flush operations. */
|
|
|
|
static inline void flush_tlb_mm(struct mm_struct *mm)
|
|
{
|
|
}
|
|
|
|
static inline void flush_tlb_page(struct vm_area_struct *vma,
|
|
unsigned long vmaddr)
|
|
{
|
|
}
|
|
|
|
static inline void flush_tlb_range(struct vm_area_struct *vma,
|
|
unsigned long start, unsigned long end)
|
|
{
|
|
}
|
|
|
|
void flush_tlb_kernel_range(unsigned long start, unsigned long end);
|
|
|
|
#define __HAVE_ARCH_ENTER_LAZY_MMU_MODE
|
|
|
|
void flush_tlb_pending(void);
|
|
void arch_enter_lazy_mmu_mode(void);
|
|
void arch_leave_lazy_mmu_mode(void);
|
|
#define arch_flush_lazy_mmu_mode() do {} while (0)
|
|
|
|
/* Local cpu only. */
|
|
void __flush_tlb_all(void);
|
|
void __flush_tlb_page(unsigned long context, unsigned long vaddr);
|
|
void __flush_tlb_kernel_range(unsigned long start, unsigned long end);
|
|
|
|
#ifndef CONFIG_SMP
|
|
|
|
static inline void global_flush_tlb_page(struct mm_struct *mm, unsigned long vaddr)
|
|
{
|
|
__flush_tlb_page(CTX_HWBITS(mm->context), vaddr);
|
|
}
|
|
|
|
#else /* CONFIG_SMP */
|
|
|
|
void smp_flush_tlb_kernel_range(unsigned long start, unsigned long end);
|
|
void smp_flush_tlb_page(struct mm_struct *mm, unsigned long vaddr);
|
|
|
|
#define global_flush_tlb_page(mm, vaddr) \
|
|
smp_flush_tlb_page(mm, vaddr)
|
|
|
|
#endif /* ! CONFIG_SMP */
|
|
|
|
#endif /* _SPARC64_TLBFLUSH_H */
|