/linux-master/arch/powerpc/include/asm/book3s/64/ |
H A D | hugetlb.h | 12 void radix__flush_hugetlb_page(struct vm_area_struct *vma, unsigned long vmaddr); 13 void radix__local_flush_hugetlb_page(struct vm_area_struct *vma, unsigned long vmaddr); 15 extern void radix__huge_ptep_modify_prot_commit(struct vm_area_struct *vma, 56 extern pte_t huge_ptep_modify_prot_start(struct vm_area_struct *vma, 60 extern void huge_ptep_modify_prot_commit(struct vm_area_struct *vma, 86 static inline void flush_hugetlb_page(struct vm_area_struct *vma, argument 90 return radix__flush_hugetlb_page(vma, vmaddr); 106 void flush_hugetlb_page(struct vm_area_struct *vma, unsigned long vmaddr);
|
H A D | tlbflush-radix.h | 63 extern void radix__flush_hugetlb_tlb_range(struct vm_area_struct *vma, 69 extern void radix__flush_pmd_tlb_range(struct vm_area_struct *vma, 71 extern void radix__flush_pud_tlb_range(struct vm_area_struct *vma, 73 extern void radix__flush_tlb_range(struct vm_area_struct *vma, unsigned long start, 79 extern void radix__local_flush_tlb_page(struct vm_area_struct *vma, unsigned long vmaddr); 86 extern void radix__flush_tlb_page(struct vm_area_struct *vma, unsigned long vmaddr); 92 #define radix__flush_tlb_page(vma,addr) radix__local_flush_tlb_page(vma,addr)
|
/linux-master/drivers/media/common/videobuf2/ |
H A D | videobuf2-memops.c | 84 * vb2_common_vm_open() - increase refcount of the vma 85 * @vma: virtual memory region for the mapping 87 * This function adds another user to the provided vma. It expects 88 * struct vb2_vmarea_handler pointer in vma->vm_private_data. 90 static void vb2_common_vm_open(struct vm_area_struct *vma) argument 92 struct vb2_vmarea_handler *h = vma->vm_private_data; 94 pr_debug("%s: %p, refcount: %d, vma: %08lx-%08lx\n", 95 __func__, h, refcount_read(h->refcount), vma->vm_start, 96 vma->vm_end); 102 * vb2_common_vm_close() - decrease refcount of the vma 108 vb2_common_vm_close(struct vm_area_struct *vma) argument [all...] |
/linux-master/include/linux/ |
H A D | mm.h | 185 * per a vma. In ELF, the number of sections is represented in unsigned short. 252 void __vm_area_free(struct vm_area_struct *vma); 295 #define VM_DONTCOPY 0x00020000 /* Do not copy this vma on fork */ 313 #define VM_HUGEPAGE 0x20000000 /* MADV_HUGEPAGE marked this vma */ 314 #define VM_NOHUGEPAGE 0x40000000 /* MADV_NOHUGEPAGE marked this vma */ 517 * vm_fault is filled by the pagefault handler and passed to the vma's 518 * ->fault function. The vma's ->fault is responsible for returning a bitmask 528 struct vm_area_struct *vma; /* Target VMA */ 530 pgoff_t pgoff; /* Logical page offset based on vma */ 592 int (*mprotect)(struct vm_area_struct *vma, unsigne 523 struct vm_area_struct *vma; /* Target VMA */ member in struct:vm_fault::__anon7 647 vma_numab_state_init(struct vm_area_struct *vma) argument 651 vma_numab_state_free(struct vm_area_struct *vma) argument 656 vma_numab_state_init(struct vm_area_struct *vma) argument 657 vma_numab_state_free(struct vm_area_struct *vma) argument 666 vma_start_read(struct vm_area_struct *vma) argument 699 vma_end_read(struct vm_area_struct *vma) argument 707 __is_vma_write_locked(struct vm_area_struct *vma, int *mm_lock_seq) argument 724 vma_start_write(struct vm_area_struct *vma) argument 742 vma_assert_write_locked(struct vm_area_struct *vma) argument 749 vma_assert_locked(struct vm_area_struct *vma) argument 755 vma_mark_detached(struct vm_area_struct *vma, bool detached) argument 784 vma_start_read(struct vm_area_struct *vma) argument 786 vma_end_read(struct vm_area_struct *vma) argument 787 vma_start_write(struct vm_area_struct *vma) argument 788 vma_assert_write_locked(struct vm_area_struct *vma) argument 790 vma_mark_detached(struct vm_area_struct *vma, bool detached) argument 799 vma_assert_locked(struct vm_area_struct *vma) argument 822 vma_init(struct vm_area_struct *vma, struct mm_struct *mm) argument 833 vm_flags_init(struct vm_area_struct *vma, vm_flags_t flags) argument 844 vm_flags_reset(struct vm_area_struct *vma, vm_flags_t flags) argument 851 vm_flags_reset_once(struct vm_area_struct *vma, vm_flags_t flags) argument 858 vm_flags_set(struct vm_area_struct *vma, vm_flags_t flags) argument 865 vm_flags_clear(struct vm_area_struct *vma, vm_flags_t flags) argument 876 __vm_flags_mod(struct vm_area_struct *vma, vm_flags_t set, vm_flags_t clear) argument 886 vm_flags_mod(struct vm_area_struct *vma, vm_flags_t set, vm_flags_t clear) argument 893 vma_set_anonymous(struct vm_area_struct *vma) argument 898 vma_is_anonymous(struct vm_area_struct *vma) argument 907 vma_is_initial_heap(const struct vm_area_struct *vma) argument 917 vma_is_initial_stack(const struct vm_area_struct *vma) argument 928 vma_is_temporary_stack(struct vm_area_struct *vma) argument 942 vma_is_foreign(struct vm_area_struct *vma) argument 953 vma_is_accessible(struct vm_area_struct *vma) argument 964 vma_is_shared_maywrite(struct vm_area_struct *vma) argument 1034 vma_iter_bulk_store(struct vma_iterator *vmi, struct vm_area_struct *vma) argument 1071 vma_is_shmem(struct vm_area_struct *vma) argument 1072 vma_is_anon_shmem(struct vm_area_struct *vma) argument 1367 maybe_mkwrite(pte_t pte, struct vm_area_struct *vma) argument 1759 vma_set_access_pid_bit(struct vm_area_struct *vma) argument 1818 vma_set_access_pid_bit(struct vm_area_struct *vma) argument 1982 folio_needs_cow_for_dma(struct vm_area_struct *vma, struct folio *folio) argument 2424 zap_vma_pages(struct vm_area_struct *vma) argument 2466 handle_mm_fault(struct vm_area_struct *vma, unsigned long address, unsigned int flags, struct pt_regs *regs) argument 2519 struct vm_area_struct *vma; local 2595 vma_wants_manual_pte_write_upgrade(struct vm_area_struct *vma) argument 3341 vma_modify_flags(struct vma_iterator *vmi, struct vm_area_struct *prev, struct vm_area_struct *vma, unsigned long start, unsigned long end, unsigned long new_flags) argument 3354 vma_modify_flags_name(struct vma_iterator *vmi, struct vm_area_struct *prev, struct vm_area_struct *vma, unsigned long start, unsigned long end, unsigned long new_flags, struct anon_vma_name *new_name) argument 3368 vma_modify_policy(struct vma_iterator *vmi, struct vm_area_struct *prev, struct vm_area_struct *vma, unsigned long start, unsigned long end, struct mempolicy *new_pol) argument 3380 vma_modify_flags_uffd(struct vma_iterator *vmi, struct vm_area_struct *prev, struct vm_area_struct *vma, unsigned long start, unsigned long end, unsigned long new_flags, struct vm_userfaultfd_ctx new_ctx) argument 3535 stack_guard_start_gap(struct vm_area_struct *vma) argument 3547 vm_start_gap(struct vm_area_struct *vma) argument 3558 vm_end_gap(struct vm_area_struct *vma) argument 3570 vma_pages(struct vm_area_struct *vma) argument 3579 struct vm_area_struct *vma = vma_lookup(mm, vm_start); local 3587 range_in_vma(struct vm_area_struct *vma, unsigned long start, unsigned long end) argument 3601 vma_set_page_prot(struct vm_area_struct *vma) argument 3637 vmf_insert_page(struct vm_area_struct *vma, unsigned long addr, struct page *page) argument 3651 io_remap_pfn_range(struct vm_area_struct *vma, unsigned long addr, unsigned long pfn, unsigned long size, pgprot_t prot) argument 3704 gup_can_follow_protnone(struct vm_area_struct *vma, unsigned int flags) argument 4144 vma_is_special_huge(const struct vm_area_struct *vma) argument 4194 seal_check_write(int seals, struct vm_area_struct *vma) argument [all...] |
/linux-master/arch/powerpc/tools/ |
H A D | head_check.sh | 52 vma=$(grep -e " [TA] _stext$" .tmp_symbols.txt | cut -d' ' -f1) 54 expected_start_head_addr="$vma" 66 top_vma=$(echo "$vma" | cut -d'0' -f1)
|
/linux-master/arch/mips/include/asm/ |
H A D | hugetlb.h | 46 static inline pte_t huge_ptep_clear_flush(struct vm_area_struct *vma, argument 56 pte = huge_ptep_get_and_clear(vma->vm_mm, addr, ptep); 57 flush_tlb_page(vma, addr); 69 static inline int huge_ptep_set_access_flags(struct vm_area_struct *vma, argument 77 set_pte_at(vma->vm_mm, addr, ptep, pte); 82 flush_tlb_range(vma, addr, addr + HPAGE_SIZE);
|
/linux-master/arch/powerpc/kernel/ |
H A D | proc_powerpc.c | 31 static int page_map_mmap( struct file *file, struct vm_area_struct *vma ) 33 if ((vma->vm_end - vma->vm_start) > PAGE_SIZE) 36 remap_pfn_range(vma, vma->vm_start, 38 PAGE_SIZE, vma->vm_page_prot);
|
/linux-master/arch/loongarch/include/asm/ |
H A D | hugetlb.h | 45 static inline pte_t huge_ptep_clear_flush(struct vm_area_struct *vma, argument 50 pte = huge_ptep_get_and_clear(vma->vm_mm, addr, ptep); 51 flush_tlb_page(vma, addr); 63 static inline int huge_ptep_set_access_flags(struct vm_area_struct *vma, argument 71 set_pte_at(vma->vm_mm, addr, ptep, pte); 76 flush_tlb_range(vma, addr, addr + HPAGE_SIZE);
|
/linux-master/arch/powerpc/include/asm/nohash/ |
H A D | tlbflush.h | 9 * - flush_tlb_page(vma, vmaddr) flushes one page 12 * - local_flush_tlb_page(vma, vmaddr) flushes one page on the local processor 13 * - flush_tlb_range(vma, start, end) flushes a range of pages 31 extern void flush_tlb_range(struct vm_area_struct *vma, unsigned long start, 43 static inline void local_flush_tlb_page(struct vm_area_struct *vma, unsigned long vmaddr) argument 66 extern void local_flush_tlb_page(struct vm_area_struct *vma, unsigned long vmaddr); 75 extern void flush_tlb_page(struct vm_area_struct *vma, unsigned long vmaddr); 80 #define flush_tlb_page(vma,addr) local_flush_tlb_page(vma,addr)
|
/linux-master/include/asm-generic/ |
H A D | cacheflush.h | 35 static inline void flush_cache_range(struct vm_area_struct *vma, argument 43 static inline void flush_cache_page(struct vm_area_struct *vma, argument 81 static inline void flush_icache_user_page(struct vm_area_struct *vma, argument 107 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ 111 flush_icache_user_page(vma, page, vaddr, len); \ 117 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \
|
/linux-master/include/xen/ |
H A D | xen-ops.h | 47 int xen_remap_pfn(struct vm_area_struct *vma, unsigned long addr, 51 static inline int xen_remap_pfn(struct vm_area_struct *vma, unsigned long addr, argument 64 int xen_xlate_remap_gfn_array(struct vm_area_struct *vma, 70 int xen_xlate_unmap_gfn_range(struct vm_area_struct *vma, 77 static inline int xen_xlate_remap_gfn_array(struct vm_area_struct *vma, argument 87 static inline int xen_xlate_unmap_gfn_range(struct vm_area_struct *vma, argument 94 int xen_remap_vma_range(struct vm_area_struct *vma, unsigned long addr, 99 * @vma: VMA to map the pages into 114 static inline int xen_remap_domain_gfn_array(struct vm_area_struct *vma, argument 122 return xen_xlate_remap_gfn_array(vma, add 150 xen_remap_domain_mfn_array(struct vm_area_struct *vma, unsigned long addr, xen_pfn_t *mfn, int nr, int *err_ptr, pgprot_t prot, unsigned int domid) argument 174 xen_remap_domain_gfn_range(struct vm_area_struct *vma, unsigned long addr, xen_pfn_t gfn, int nr, pgprot_t prot, unsigned int domid, struct page **pages) argument [all...] |
/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_pt.c | 266 * @vma: VMA being mapped 268 struct xe_vma *vma; member in struct:xe_pt_stage_bind_walk 408 if (xe_vma_is_null(xe_walk->vma)) 434 if (xe_vma_is_null(xe_walk->vma)) 478 u16 pat_index = xe_walk->vma->pat_index; 489 bool is_null = xe_vma_is_null(xe_walk->vma); 495 xe_walk->vma, pat_index, level); 504 xe_walk->vma->gpuva.flags |= XE_VMA_PTE_64K; 518 xe_walk->vma->gpuva.flags |= (XE_VMA_PTE_4K << level); 560 xe_walk->vma 597 xe_pt_stage_bind(struct xe_tile *tile, struct xe_vma *vma, struct xe_vm_pgtable_update *entries, u32 *num_entries) argument 773 xe_pt_zap_ptes(struct xe_tile *tile, struct xe_vma *vma) argument 813 xe_pt_abort_bind(struct xe_vma *vma, struct xe_vm_pgtable_update *entries, u32 num_entries) argument 829 xe_pt_commit_locks_assert(struct xe_vma *vma) argument 843 xe_pt_commit_bind(struct xe_vma *vma, struct xe_vm_pgtable_update *entries, u32 num_entries, bool rebind, struct llist_head *deferred) argument 880 xe_pt_prepare_bind(struct xe_tile *tile, struct xe_vma *vma, struct xe_vm_pgtable_update *entries, u32 *num_entries) argument 1078 struct xe_vma *vma; member in struct:invalidation_fence 1127 invalidation_fence_init(struct xe_gt *gt, struct invalidation_fence *ifence, struct dma_fence *fence, struct xe_vma *vma) argument 1164 xe_pt_calc_rfence_interval(struct xe_vma *vma, struct xe_pt_migrate_pt_update *update, struct xe_vm_pgtable_update *entries, u32 num_entries) argument 1208 __xe_pt_bind_vma(struct xe_tile *tile, struct xe_vma *vma, struct xe_exec_queue *q, struct xe_sync_entry *syncs, u32 num_syncs, bool rebind) argument 1467 xe_pt_stage_unbind(struct xe_tile *tile, struct xe_vma *vma, struct xe_vm_pgtable_update *entries) argument 1495 struct xe_vma *vma = pt_update->vma; local 1511 xe_pt_commit_unbind(struct xe_vma *vma, struct xe_vm_pgtable_update *entries, u32 num_entries, struct llist_head *deferred) argument 1571 __xe_pt_unbind_vma(struct xe_tile *tile, struct xe_vma *vma, struct xe_exec_queue *q, struct xe_sync_entry *syncs, u32 num_syncs) argument [all...] |
/linux-master/mm/ |
H A D | rmap.c | 150 static void anon_vma_chain_link(struct vm_area_struct *vma, argument 154 avc->vma = vma; 156 list_add(&avc->same_vma, &vma->anon_vma_chain); 162 * @vma: the memory region in question 164 * This makes sure the memory mapping described by 'vma' has 172 * reason for splitting a vma has been mprotect()), or we 175 * Anon-vma allocations are very subtle, because we may have 178 * allocated vma (it depends on RCU to make sure that the 186 int __anon_vma_prepare(struct vm_area_struct *vma) argument 334 anon_vma_fork(struct vm_area_struct *vma, struct vm_area_struct *pvma) argument 396 unlink_anon_vmas(struct vm_area_struct *vma) argument 774 page_address_in_vma(struct page *page, struct vm_area_struct *vma) argument 838 folio_referenced_one(struct folio *folio, struct vm_area_struct *vma, unsigned long address, void *arg) argument 928 invalid_folio_referenced_vma(struct vm_area_struct *vma, void *arg) argument 1005 struct vm_area_struct *vma = pvmw->vma; local 1064 page_mkclean_one(struct folio *folio, struct vm_area_struct *vma, unsigned long address, void *arg) argument 1075 invalid_mkclean_vma(struct vm_area_struct *vma, void *arg) argument 1120 pfn_mkclean_range(unsigned long pfn, unsigned long nr_pages, pgoff_t pgoff, struct vm_area_struct *vma) argument 1197 folio_move_anon_rmap(struct folio *folio, struct vm_area_struct *vma) argument 1220 __folio_set_anon(struct folio *folio, struct vm_area_struct *vma, unsigned long address, bool exclusive) argument 1252 __page_check_anon_rmap(struct folio *folio, struct page *page, struct vm_area_struct *vma, unsigned long address) argument 1272 __folio_add_anon_rmap(struct folio *folio, struct page *page, int nr_pages, struct vm_area_struct *vma, unsigned long address, rmap_t flags, enum rmap_level level) argument 1347 folio_add_anon_rmap_ptes(struct folio *folio, struct page *page, int nr_pages, struct vm_area_struct *vma, unsigned long address, rmap_t flags) argument 1368 folio_add_anon_rmap_pmd(struct folio *folio, struct page *page, struct vm_area_struct *vma, unsigned long address, rmap_t flags) argument 1392 folio_add_new_anon_rmap(struct folio *folio, struct vm_area_struct *vma, unsigned long address) argument 1434 __folio_add_file_rmap(struct folio *folio, struct page *page, int nr_pages, struct vm_area_struct *vma, enum rmap_level level) argument 1466 folio_add_file_rmap_ptes(struct folio *folio, struct page *page, int nr_pages, struct vm_area_struct *vma) argument 1482 folio_add_file_rmap_pmd(struct folio *folio, struct page *page, struct vm_area_struct *vma) argument 1492 __folio_remove_rmap(struct folio *folio, struct page *page, int nr_pages, struct vm_area_struct *vma, enum rmap_level level) argument 1592 folio_remove_rmap_ptes(struct folio *folio, struct page *page, int nr_pages, struct vm_area_struct *vma) argument 1608 folio_remove_rmap_pmd(struct folio *folio, struct page *page, struct vm_area_struct *vma) argument 1621 try_to_unmap_one(struct folio *folio, struct vm_area_struct *vma, unsigned long address, void *arg) argument 1928 invalid_migration_vma(struct vm_area_struct *vma, void *arg) argument 1970 try_to_migrate_one(struct folio *folio, struct vm_area_struct *vma, unsigned long address, void *arg) argument 2344 page_make_device_exclusive_one(struct folio *folio, struct vm_area_struct *vma, unsigned long address, void *priv) argument 2595 struct vm_area_struct *vma = avc->vma; local 2629 struct vm_area_struct *vma; local 2705 hugetlb_add_anon_rmap(struct folio *folio, struct vm_area_struct *vma, unsigned long address, rmap_t flags) argument 2719 hugetlb_add_new_anon_rmap(struct folio *folio, struct vm_area_struct *vma, unsigned long address) argument [all...] |
H A D | pagewalk.c | 154 if ((!walk->vma && (pmd_leaf(*pmd) || !pmd_present(*pmd))) || 159 if (walk->vma) 160 split_huge_pmd(walk->vma, pmd, addr); 208 if ((!walk->vma && (pud_leaf(*pud) || !pud_present(*pud))) || 213 if (walk->vma) 214 split_huge_pud(walk->vma, pud, addr); 312 struct vm_area_struct *vma = walk->vma; local 313 struct hstate *h = hstate_vma(vma); 321 hugetlb_vma_lock_read(vma); 355 struct vm_area_struct *vma = walk->vma; local 382 struct vm_area_struct *vma = walk->vma; local 412 process_vma_walk_lock(struct vm_area_struct *vma, enum page_walk_lock walk_lock) argument 476 struct vm_area_struct *vma; local 591 walk_page_range_vma(struct vm_area_struct *vma, unsigned long start, unsigned long end, const struct mm_walk_ops *ops, void *private) argument 612 walk_page_vma(struct vm_area_struct *vma, const struct mm_walk_ops *ops, void *private) argument 668 struct vm_area_struct *vma; local [all...] |
H A D | mlock.c | 333 struct vm_area_struct *vma, unsigned long start, 345 if (!(vma->vm_flags & VM_LOCKED)) 353 if (!folio_within_range(folio, vma, start, end)) 367 struct vm_area_struct *vma = walk->vma; local 375 ptl = pmd_trans_huge_lock(pmd, vma); 382 if (vma->vm_flags & VM_LOCKED) 389 start_pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); 399 folio = vm_normal_folio(vma, addr, ptent); 404 if (!allow_mlock_munlock(folio, vma, star 332 allow_mlock_munlock(struct folio *folio, struct vm_area_struct *vma, unsigned long start, unsigned long end, unsigned int step) argument 434 mlock_vma_pages_range(struct vm_area_struct *vma, unsigned long start, unsigned long end, vm_flags_t newflags) argument 477 mlock_fixup(struct vma_iterator *vmi, struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end, vm_flags_t newflags) argument 529 struct vm_area_struct *vma, *prev; local 585 struct vm_area_struct *vma; local 718 struct vm_area_struct *vma, *prev = NULL; local [all...] |
H A D | pgtable-generic.c | 68 int ptep_set_access_flags(struct vm_area_struct *vma, argument 74 set_pte_at(vma->vm_mm, address, ptep, entry); 75 flush_tlb_fix_spurious_fault(vma, address, ptep); 82 int ptep_clear_flush_young(struct vm_area_struct *vma, argument 86 young = ptep_test_and_clear_young(vma, address, ptep); 88 flush_tlb_page(vma, address); 94 pte_t ptep_clear_flush(struct vm_area_struct *vma, unsigned long address, argument 97 struct mm_struct *mm = (vma)->vm_mm; 101 flush_tlb_page(vma, address); 109 int pmdp_set_access_flags(struct vm_area_struct *vma, argument 124 pmdp_clear_flush_young(struct vm_area_struct *vma, unsigned long address, pmd_t *pmdp) argument 137 pmdp_huge_clear_flush(struct vm_area_struct *vma, unsigned long address, pmd_t *pmdp) argument 150 pudp_huge_clear_flush(struct vm_area_struct *vma, unsigned long address, pud_t *pudp) argument 198 pmdp_invalidate(struct vm_area_struct *vma, unsigned long address, pmd_t *pmdp) argument 209 pmdp_invalidate_ad(struct vm_area_struct *vma, unsigned long address, pmd_t *pmdp) argument 218 pmdp_collapse_flush(struct vm_area_struct *vma, unsigned long address, pmd_t *pmdp) argument [all...] |
/linux-master/arch/nios2/mm/ |
H A D | cacheflush.c | 77 struct vm_area_struct *vma; local 85 vma_interval_tree_foreach(vma, &mapping->i_mmap, pgoff, pgoff + nr - 1) { 88 if (vma->vm_mm != mm) 90 if (!(vma->vm_flags & VM_MAYSHARE)) 93 start = vma->vm_start + ((pgoff - vma->vm_pgoff) << PAGE_SHIFT); 94 flush_cache_range(vma, start, start + nr * PAGE_SIZE); 134 void flush_cache_range(struct vm_area_struct *vma, unsigned long start, argument 138 if (vma == NULL || (vma 142 flush_icache_pages(struct vm_area_struct *vma, struct page *page, unsigned int nr) argument 152 flush_cache_page(struct vm_area_struct *vma, unsigned long vmaddr, unsigned long pfn) argument 209 update_mmu_cache_range(struct vm_fault *vmf, struct vm_area_struct *vma, unsigned long address, pte_t *ptep, unsigned int nr) argument 261 copy_from_user_page(struct vm_area_struct *vma, struct page *page, unsigned long user_vaddr, void *dst, void *src, int len) argument 272 copy_to_user_page(struct vm_area_struct *vma, struct page *page, unsigned long user_vaddr, void *dst, void *src, int len) argument [all...] |
/linux-master/arch/arm/mm/ |
H A D | flush.c | 79 void flush_cache_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) argument 82 vivt_flush_cache_range(vma, start, end); 94 if (vma->vm_flags & VM_EXEC) 98 void flush_cache_pages(struct vm_area_struct *vma, unsigned long user_addr, unsigned long pfn, unsigned int nr) argument 101 vivt_flush_cache_pages(vma, user_addr, pfn, nr); 110 if (vma->vm_flags & VM_EXEC && icache_is_vivt_asid_tagged()) 159 void flush_ptrace_access(struct vm_area_struct *vma, struct page *page, argument 163 if (cpumask_test_cpu(smp_processor_id(), mm_cpumask(vma->vm_mm))) 165 if (vma->vm_flags & VM_EXEC) 185 void copy_to_user_page(struct vm_area_struct *vma, struc argument 241 struct vm_area_struct *vma; local 382 __flush_anon_page(struct vm_area_struct *vma, struct page *page, unsigned long vmaddr) argument [all...] |
/linux-master/drivers/char/ |
H A D | uv_mmtimer.c | 43 static int uv_mmtimer_mmap(struct file *file, struct vm_area_struct *vma); 142 * @vma: VMA to map the registers into 147 static int uv_mmtimer_mmap(struct file *file, struct vm_area_struct *vma) argument 151 if (vma->vm_end - vma->vm_start != PAGE_SIZE) 154 if (vma->vm_flags & VM_WRITE) 160 vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot); 166 if (remap_pfn_range(vma, vma [all...] |
/linux-master/kernel/dma/ |
H A D | ops_helpers.c | 34 int dma_common_mmap(struct device *dev, struct vm_area_struct *vma, argument 39 unsigned long user_count = vma_pages(vma); 41 unsigned long off = vma->vm_pgoff; 45 vma->vm_page_prot = dma_pgprot(dev, vma->vm_page_prot, attrs); 47 if (dma_mmap_from_dev_coherent(dev, vma, cpu_addr, size, &ret)) 53 return remap_pfn_range(vma, vma->vm_start, 54 page_to_pfn(page) + vma->vm_pgoff, 55 user_count << PAGE_SHIFT, vma [all...] |
/linux-master/drivers/xen/ |
H A D | privcmd-buf.c | 86 static void privcmd_buf_vma_open(struct vm_area_struct *vma) argument 88 struct privcmd_buf_vma_private *vma_priv = vma->vm_private_data; 98 static void privcmd_buf_vma_close(struct vm_area_struct *vma) argument 100 struct privcmd_buf_vma_private *vma_priv = vma->vm_private_data; 119 pr_debug("fault: vma=%p %lx-%lx, pgoff=%lx, uv=%p\n", 120 vmf->vma, vmf->vma->vm_start, vmf->vma->vm_end, 132 static int privcmd_buf_mmap(struct file *file, struct vm_area_struct *vma) argument 136 unsigned long count = vma_pages(vma); [all...] |
/linux-master/arch/loongarch/kernel/ |
H A D | vdso.c | 49 struct vm_area_struct *vma, struct vm_fault *vmf) 52 struct page *timens_page = find_timens_vvar_page(vma); 82 return vmf_insert_pfn(vma, vmf->address, pfn); 134 struct vm_area_struct *vma; local 139 for_each_vma(vmi, vma) { 140 if (vma_is_special_mapping(vma, &vdso_info.data_mapping)) 141 zap_vma_pages(vma); 166 struct vm_area_struct *vma; local 184 vma = _install_special_mapping(mm, data_addr, VVAR_SIZE, 187 if (IS_ERR(vma)) { 48 vvar_fault(const struct vm_special_mapping *sm, struct vm_area_struct *vma, struct vm_fault *vmf) argument [all...] |
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
H A D | uvmm.c | 113 struct nvkm_vma *vma; local 126 vma = nvkm_vmm_node_search(vmm, addr); 127 if (ret = -ENOENT, !vma || vma->addr != addr) { 129 addr, vma ? vma->addr : ~0ULL); 133 if (ret = -ENOENT, vma->busy) { 134 VMM_DEBUG(vmm, "denied %016llx: %d", addr, vma->busy); 138 if (ret = -EINVAL, !vma->memory) { 143 nvkm_vmm_unmap_locked(vmm, vma, fals 159 struct nvkm_vma *vma; local 238 struct nvkm_vma *vma; local 274 struct nvkm_vma *vma; local 394 struct nvkm_vma vma = { local [all...] |
/linux-master/arch/csky/kernel/ |
H A D | vdso.c | 89 const char *arch_vma_name(struct vm_area_struct *vma) argument 91 if (vma->vm_mm && (vma->vm_start == (long)vma->vm_mm->context.vdso)) 93 if (vma->vm_mm && (vma->vm_start == 94 (long)vma->vm_mm->context.vdso + PAGE_SIZE))
|
/linux-master/drivers/misc/ocxl/ |
H A D | context.c | 98 static vm_fault_t map_afu_irq(struct vm_area_struct *vma, unsigned long address, argument 108 return vmf_insert_pfn(vma, address, trigger_addr >> PAGE_SHIFT); 111 static vm_fault_t map_pp_mmio(struct vm_area_struct *vma, unsigned long address, argument 134 ret = vmf_insert_pfn(vma, address, pp_mmio_addr >> PAGE_SHIFT); 141 struct vm_area_struct *vma = vmf->vma; local 142 struct ocxl_context *ctx = vma->vm_file->private_data; 151 ret = map_pp_mmio(vma, vmf->address, offset, ctx); 153 ret = map_afu_irq(vma, vmf->address, offset, ctx); 162 struct vm_area_struct *vma) 161 check_mmap_afu_irq(struct ocxl_context *ctx, struct vm_area_struct *vma) argument 187 check_mmap_mmio(struct ocxl_context *ctx, struct vm_area_struct *vma) argument 196 ocxl_context_mmap(struct ocxl_context *ctx, struct vm_area_struct *vma) argument [all...] |