Lines Matching defs:vma

256  *	flush_tlb_range(vma,start,end)
264 * flush_tlb_page(vma, uaddr)
267 * - vma - vm_area_struct describing address range
413 __local_flush_tlb_page(struct vm_area_struct *vma, unsigned long uaddr)
418 uaddr = (uaddr & PAGE_MASK) | ASID(vma->vm_mm);
421 cpumask_test_cpu(smp_processor_id(), mm_cpumask(vma->vm_mm))) {
435 local_flush_tlb_page(struct vm_area_struct *vma, unsigned long uaddr)
439 uaddr = (uaddr & PAGE_MASK) | ASID(vma->vm_mm);
444 __local_flush_tlb_page(vma, uaddr);
452 __flush_tlb_page(struct vm_area_struct *vma, unsigned long uaddr)
456 uaddr = (uaddr & PAGE_MASK) | ASID(vma->vm_mm);
461 __local_flush_tlb_page(vma, uaddr);
597 #define local_flush_tlb_range(vma,start,end) __cpu_flush_user_tlb_range(start,end,vma)
611 extern void flush_tlb_page(struct vm_area_struct *vma, unsigned long uaddr);
613 extern void flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end);
625 void update_mmu_cache_range(struct vm_fault *vmf, struct vm_area_struct *vma,
629 struct vm_area_struct *vma, unsigned long addr, pte_t *ptep,
635 #define update_mmu_cache(vma, addr, ptep) \
636 update_mmu_cache_range(NULL, vma, addr, ptep, 1)
638 #define update_mmu_cache_pmd(vma, address, pmd) do { } while (0)
647 static inline void local_flush_tlb_page(struct vm_area_struct *vma, unsigned long uaddr) { }
649 static inline void local_flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) { }
655 extern void flush_tlb_page(struct vm_area_struct *vma, unsigned long uaddr);
657 extern void flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end);