123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869 |
- #ifndef _SPARC64_TLBFLUSH_H
- #define _SPARC64_TLBFLUSH_H
- #include <asm/mmu_context.h>
- /* TSB flush operations. */
- #define TLB_BATCH_NR 192
- struct tlb_batch {
- bool huge;
- struct mm_struct *mm;
- unsigned long tlb_nr;
- unsigned long active;
- unsigned long vaddrs[TLB_BATCH_NR];
- };
- void flush_tsb_kernel_range(unsigned long start, unsigned long end);
- void flush_tsb_user(struct tlb_batch *tb);
- void flush_tsb_user_page(struct mm_struct *mm, unsigned long vaddr, bool huge);
- /* TLB flush operations. */
- static inline void flush_tlb_mm(struct mm_struct *mm)
- {
- }
- static inline void flush_tlb_page(struct vm_area_struct *vma,
- unsigned long vmaddr)
- {
- }
- static inline void flush_tlb_range(struct vm_area_struct *vma,
- unsigned long start, unsigned long end)
- {
- }
- void flush_tlb_kernel_range(unsigned long start, unsigned long end);
- #define __HAVE_ARCH_ENTER_LAZY_MMU_MODE
- void flush_tlb_pending(void);
- void arch_enter_lazy_mmu_mode(void);
- void arch_leave_lazy_mmu_mode(void);
- #define arch_flush_lazy_mmu_mode() do {} while (0)
- /* Local cpu only. */
- void __flush_tlb_all(void);
- void __flush_tlb_page(unsigned long context, unsigned long vaddr);
- void __flush_tlb_kernel_range(unsigned long start, unsigned long end);
- #ifndef CONFIG_SMP
- static inline void global_flush_tlb_page(struct mm_struct *mm, unsigned long vaddr)
- {
- __flush_tlb_page(CTX_HWBITS(mm->context), vaddr);
- }
- #else /* CONFIG_SMP */
- void smp_flush_tlb_kernel_range(unsigned long start, unsigned long end);
- void smp_flush_tlb_page(struct mm_struct *mm, unsigned long vaddr);
- #define global_flush_tlb_page(mm, vaddr) \
- smp_flush_tlb_page(mm, vaddr)
- #endif /* ! CONFIG_SMP */
- #endif /* _SPARC64_TLBFLUSH_H */
|