Searched refs:vma (Results 76 - 100 of 1013) sorted by relevance

1234567891011>>

/linux-master/arch/powerpc/include/asm/book3s/64/
H A Dhugetlb.h12 void radix__flush_hugetlb_page(struct vm_area_struct *vma, unsigned long vmaddr);
13 void radix__local_flush_hugetlb_page(struct vm_area_struct *vma, unsigned long vmaddr);
15 extern void radix__huge_ptep_modify_prot_commit(struct vm_area_struct *vma,
56 extern pte_t huge_ptep_modify_prot_start(struct vm_area_struct *vma,
60 extern void huge_ptep_modify_prot_commit(struct vm_area_struct *vma,
86 static inline void flush_hugetlb_page(struct vm_area_struct *vma, argument
90 return radix__flush_hugetlb_page(vma, vmaddr);
106 void flush_hugetlb_page(struct vm_area_struct *vma, unsigned long vmaddr);
H A Dtlbflush-radix.h63 extern void radix__flush_hugetlb_tlb_range(struct vm_area_struct *vma,
69 extern void radix__flush_pmd_tlb_range(struct vm_area_struct *vma,
71 extern void radix__flush_pud_tlb_range(struct vm_area_struct *vma,
73 extern void radix__flush_tlb_range(struct vm_area_struct *vma, unsigned long start,
79 extern void radix__local_flush_tlb_page(struct vm_area_struct *vma, unsigned long vmaddr);
86 extern void radix__flush_tlb_page(struct vm_area_struct *vma, unsigned long vmaddr);
92 #define radix__flush_tlb_page(vma,addr) radix__local_flush_tlb_page(vma,addr)
/linux-master/drivers/media/common/videobuf2/
H A Dvideobuf2-memops.c84 * vb2_common_vm_open() - increase refcount of the vma
85 * @vma: virtual memory region for the mapping
87 * This function adds another user to the provided vma. It expects
88 * struct vb2_vmarea_handler pointer in vma->vm_private_data.
90 static void vb2_common_vm_open(struct vm_area_struct *vma) argument
92 struct vb2_vmarea_handler *h = vma->vm_private_data;
94 pr_debug("%s: %p, refcount: %d, vma: %08lx-%08lx\n",
95 __func__, h, refcount_read(h->refcount), vma->vm_start,
96 vma->vm_end);
102 * vb2_common_vm_close() - decrease refcount of the vma
108 vb2_common_vm_close(struct vm_area_struct *vma) argument
[all...]
/linux-master/include/linux/
H A Dmm.h185 * per a vma. In ELF, the number of sections is represented in unsigned short.
252 void __vm_area_free(struct vm_area_struct *vma);
295 #define VM_DONTCOPY 0x00020000 /* Do not copy this vma on fork */
313 #define VM_HUGEPAGE 0x20000000 /* MADV_HUGEPAGE marked this vma */
314 #define VM_NOHUGEPAGE 0x40000000 /* MADV_NOHUGEPAGE marked this vma */
517 * vm_fault is filled by the pagefault handler and passed to the vma's
518 * ->fault function. The vma's ->fault is responsible for returning a bitmask
528 struct vm_area_struct *vma; /* Target VMA */
530 pgoff_t pgoff; /* Logical page offset based on vma */
592 int (*mprotect)(struct vm_area_struct *vma, unsigne
523 struct vm_area_struct *vma; /* Target VMA */ member in struct:vm_fault::__anon7
647 vma_numab_state_init(struct vm_area_struct *vma) argument
651 vma_numab_state_free(struct vm_area_struct *vma) argument
656 vma_numab_state_init(struct vm_area_struct *vma) argument
657 vma_numab_state_free(struct vm_area_struct *vma) argument
666 vma_start_read(struct vm_area_struct *vma) argument
699 vma_end_read(struct vm_area_struct *vma) argument
707 __is_vma_write_locked(struct vm_area_struct *vma, int *mm_lock_seq) argument
724 vma_start_write(struct vm_area_struct *vma) argument
742 vma_assert_write_locked(struct vm_area_struct *vma) argument
749 vma_assert_locked(struct vm_area_struct *vma) argument
755 vma_mark_detached(struct vm_area_struct *vma, bool detached) argument
784 vma_start_read(struct vm_area_struct *vma) argument
786 vma_end_read(struct vm_area_struct *vma) argument
787 vma_start_write(struct vm_area_struct *vma) argument
788 vma_assert_write_locked(struct vm_area_struct *vma) argument
790 vma_mark_detached(struct vm_area_struct *vma, bool detached) argument
799 vma_assert_locked(struct vm_area_struct *vma) argument
822 vma_init(struct vm_area_struct *vma, struct mm_struct *mm) argument
833 vm_flags_init(struct vm_area_struct *vma, vm_flags_t flags) argument
844 vm_flags_reset(struct vm_area_struct *vma, vm_flags_t flags) argument
851 vm_flags_reset_once(struct vm_area_struct *vma, vm_flags_t flags) argument
858 vm_flags_set(struct vm_area_struct *vma, vm_flags_t flags) argument
865 vm_flags_clear(struct vm_area_struct *vma, vm_flags_t flags) argument
876 __vm_flags_mod(struct vm_area_struct *vma, vm_flags_t set, vm_flags_t clear) argument
886 vm_flags_mod(struct vm_area_struct *vma, vm_flags_t set, vm_flags_t clear) argument
893 vma_set_anonymous(struct vm_area_struct *vma) argument
898 vma_is_anonymous(struct vm_area_struct *vma) argument
907 vma_is_initial_heap(const struct vm_area_struct *vma) argument
917 vma_is_initial_stack(const struct vm_area_struct *vma) argument
928 vma_is_temporary_stack(struct vm_area_struct *vma) argument
942 vma_is_foreign(struct vm_area_struct *vma) argument
953 vma_is_accessible(struct vm_area_struct *vma) argument
964 vma_is_shared_maywrite(struct vm_area_struct *vma) argument
1034 vma_iter_bulk_store(struct vma_iterator *vmi, struct vm_area_struct *vma) argument
1071 vma_is_shmem(struct vm_area_struct *vma) argument
1072 vma_is_anon_shmem(struct vm_area_struct *vma) argument
1367 maybe_mkwrite(pte_t pte, struct vm_area_struct *vma) argument
1759 vma_set_access_pid_bit(struct vm_area_struct *vma) argument
1818 vma_set_access_pid_bit(struct vm_area_struct *vma) argument
1982 folio_needs_cow_for_dma(struct vm_area_struct *vma, struct folio *folio) argument
2424 zap_vma_pages(struct vm_area_struct *vma) argument
2466 handle_mm_fault(struct vm_area_struct *vma, unsigned long address, unsigned int flags, struct pt_regs *regs) argument
2519 struct vm_area_struct *vma; local
2595 vma_wants_manual_pte_write_upgrade(struct vm_area_struct *vma) argument
3341 vma_modify_flags(struct vma_iterator *vmi, struct vm_area_struct *prev, struct vm_area_struct *vma, unsigned long start, unsigned long end, unsigned long new_flags) argument
3354 vma_modify_flags_name(struct vma_iterator *vmi, struct vm_area_struct *prev, struct vm_area_struct *vma, unsigned long start, unsigned long end, unsigned long new_flags, struct anon_vma_name *new_name) argument
3368 vma_modify_policy(struct vma_iterator *vmi, struct vm_area_struct *prev, struct vm_area_struct *vma, unsigned long start, unsigned long end, struct mempolicy *new_pol) argument
3380 vma_modify_flags_uffd(struct vma_iterator *vmi, struct vm_area_struct *prev, struct vm_area_struct *vma, unsigned long start, unsigned long end, unsigned long new_flags, struct vm_userfaultfd_ctx new_ctx) argument
3535 stack_guard_start_gap(struct vm_area_struct *vma) argument
3547 vm_start_gap(struct vm_area_struct *vma) argument
3558 vm_end_gap(struct vm_area_struct *vma) argument
3570 vma_pages(struct vm_area_struct *vma) argument
3579 struct vm_area_struct *vma = vma_lookup(mm, vm_start); local
3587 range_in_vma(struct vm_area_struct *vma, unsigned long start, unsigned long end) argument
3601 vma_set_page_prot(struct vm_area_struct *vma) argument
3637 vmf_insert_page(struct vm_area_struct *vma, unsigned long addr, struct page *page) argument
3651 io_remap_pfn_range(struct vm_area_struct *vma, unsigned long addr, unsigned long pfn, unsigned long size, pgprot_t prot) argument
3704 gup_can_follow_protnone(struct vm_area_struct *vma, unsigned int flags) argument
4144 vma_is_special_huge(const struct vm_area_struct *vma) argument
4194 seal_check_write(int seals, struct vm_area_struct *vma) argument
[all...]
/linux-master/arch/powerpc/tools/
H A Dhead_check.sh52 vma=$(grep -e " [TA] _stext$" .tmp_symbols.txt | cut -d' ' -f1)
54 expected_start_head_addr="$vma"
66 top_vma=$(echo "$vma" | cut -d'0' -f1)
/linux-master/arch/mips/include/asm/
H A Dhugetlb.h46 static inline pte_t huge_ptep_clear_flush(struct vm_area_struct *vma, argument
56 pte = huge_ptep_get_and_clear(vma->vm_mm, addr, ptep);
57 flush_tlb_page(vma, addr);
69 static inline int huge_ptep_set_access_flags(struct vm_area_struct *vma, argument
77 set_pte_at(vma->vm_mm, addr, ptep, pte);
82 flush_tlb_range(vma, addr, addr + HPAGE_SIZE);
/linux-master/arch/powerpc/kernel/
H A Dproc_powerpc.c31 static int page_map_mmap( struct file *file, struct vm_area_struct *vma )
33 if ((vma->vm_end - vma->vm_start) > PAGE_SIZE)
36 remap_pfn_range(vma, vma->vm_start,
38 PAGE_SIZE, vma->vm_page_prot);
/linux-master/arch/loongarch/include/asm/
H A Dhugetlb.h45 static inline pte_t huge_ptep_clear_flush(struct vm_area_struct *vma, argument
50 pte = huge_ptep_get_and_clear(vma->vm_mm, addr, ptep);
51 flush_tlb_page(vma, addr);
63 static inline int huge_ptep_set_access_flags(struct vm_area_struct *vma, argument
71 set_pte_at(vma->vm_mm, addr, ptep, pte);
76 flush_tlb_range(vma, addr, addr + HPAGE_SIZE);
/linux-master/arch/powerpc/include/asm/nohash/
H A Dtlbflush.h9 * - flush_tlb_page(vma, vmaddr) flushes one page
12 * - local_flush_tlb_page(vma, vmaddr) flushes one page on the local processor
13 * - flush_tlb_range(vma, start, end) flushes a range of pages
31 extern void flush_tlb_range(struct vm_area_struct *vma, unsigned long start,
43 static inline void local_flush_tlb_page(struct vm_area_struct *vma, unsigned long vmaddr) argument
66 extern void local_flush_tlb_page(struct vm_area_struct *vma, unsigned long vmaddr);
75 extern void flush_tlb_page(struct vm_area_struct *vma, unsigned long vmaddr);
80 #define flush_tlb_page(vma,addr) local_flush_tlb_page(vma,addr)
/linux-master/include/asm-generic/
H A Dcacheflush.h35 static inline void flush_cache_range(struct vm_area_struct *vma, argument
43 static inline void flush_cache_page(struct vm_area_struct *vma, argument
81 static inline void flush_icache_user_page(struct vm_area_struct *vma, argument
107 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \
111 flush_icache_user_page(vma, page, vaddr, len); \
117 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \
/linux-master/include/xen/
H A Dxen-ops.h47 int xen_remap_pfn(struct vm_area_struct *vma, unsigned long addr,
51 static inline int xen_remap_pfn(struct vm_area_struct *vma, unsigned long addr, argument
64 int xen_xlate_remap_gfn_array(struct vm_area_struct *vma,
70 int xen_xlate_unmap_gfn_range(struct vm_area_struct *vma,
77 static inline int xen_xlate_remap_gfn_array(struct vm_area_struct *vma, argument
87 static inline int xen_xlate_unmap_gfn_range(struct vm_area_struct *vma, argument
94 int xen_remap_vma_range(struct vm_area_struct *vma, unsigned long addr,
99 * @vma: VMA to map the pages into
114 static inline int xen_remap_domain_gfn_array(struct vm_area_struct *vma, argument
122 return xen_xlate_remap_gfn_array(vma, add
150 xen_remap_domain_mfn_array(struct vm_area_struct *vma, unsigned long addr, xen_pfn_t *mfn, int nr, int *err_ptr, pgprot_t prot, unsigned int domid) argument
174 xen_remap_domain_gfn_range(struct vm_area_struct *vma, unsigned long addr, xen_pfn_t gfn, int nr, pgprot_t prot, unsigned int domid, struct page **pages) argument
[all...]
/linux-master/drivers/gpu/drm/xe/
H A Dxe_pt.c266 * @vma: VMA being mapped
268 struct xe_vma *vma; member in struct:xe_pt_stage_bind_walk
408 if (xe_vma_is_null(xe_walk->vma))
434 if (xe_vma_is_null(xe_walk->vma))
478 u16 pat_index = xe_walk->vma->pat_index;
489 bool is_null = xe_vma_is_null(xe_walk->vma);
495 xe_walk->vma, pat_index, level);
504 xe_walk->vma->gpuva.flags |= XE_VMA_PTE_64K;
518 xe_walk->vma->gpuva.flags |= (XE_VMA_PTE_4K << level);
560 xe_walk->vma
597 xe_pt_stage_bind(struct xe_tile *tile, struct xe_vma *vma, struct xe_vm_pgtable_update *entries, u32 *num_entries) argument
773 xe_pt_zap_ptes(struct xe_tile *tile, struct xe_vma *vma) argument
813 xe_pt_abort_bind(struct xe_vma *vma, struct xe_vm_pgtable_update *entries, u32 num_entries) argument
829 xe_pt_commit_locks_assert(struct xe_vma *vma) argument
843 xe_pt_commit_bind(struct xe_vma *vma, struct xe_vm_pgtable_update *entries, u32 num_entries, bool rebind, struct llist_head *deferred) argument
880 xe_pt_prepare_bind(struct xe_tile *tile, struct xe_vma *vma, struct xe_vm_pgtable_update *entries, u32 *num_entries) argument
1078 struct xe_vma *vma; member in struct:invalidation_fence
1127 invalidation_fence_init(struct xe_gt *gt, struct invalidation_fence *ifence, struct dma_fence *fence, struct xe_vma *vma) argument
1164 xe_pt_calc_rfence_interval(struct xe_vma *vma, struct xe_pt_migrate_pt_update *update, struct xe_vm_pgtable_update *entries, u32 num_entries) argument
1208 __xe_pt_bind_vma(struct xe_tile *tile, struct xe_vma *vma, struct xe_exec_queue *q, struct xe_sync_entry *syncs, u32 num_syncs, bool rebind) argument
1467 xe_pt_stage_unbind(struct xe_tile *tile, struct xe_vma *vma, struct xe_vm_pgtable_update *entries) argument
1495 struct xe_vma *vma = pt_update->vma; local
1511 xe_pt_commit_unbind(struct xe_vma *vma, struct xe_vm_pgtable_update *entries, u32 num_entries, struct llist_head *deferred) argument
1571 __xe_pt_unbind_vma(struct xe_tile *tile, struct xe_vma *vma, struct xe_exec_queue *q, struct xe_sync_entry *syncs, u32 num_syncs) argument
[all...]
/linux-master/mm/
H A Drmap.c150 static void anon_vma_chain_link(struct vm_area_struct *vma, argument
154 avc->vma = vma;
156 list_add(&avc->same_vma, &vma->anon_vma_chain);
162 * @vma: the memory region in question
164 * This makes sure the memory mapping described by 'vma' has
172 * reason for splitting a vma has been mprotect()), or we
175 * Anon-vma allocations are very subtle, because we may have
178 * allocated vma (it depends on RCU to make sure that the
186 int __anon_vma_prepare(struct vm_area_struct *vma) argument
334 anon_vma_fork(struct vm_area_struct *vma, struct vm_area_struct *pvma) argument
396 unlink_anon_vmas(struct vm_area_struct *vma) argument
774 page_address_in_vma(struct page *page, struct vm_area_struct *vma) argument
838 folio_referenced_one(struct folio *folio, struct vm_area_struct *vma, unsigned long address, void *arg) argument
928 invalid_folio_referenced_vma(struct vm_area_struct *vma, void *arg) argument
1005 struct vm_area_struct *vma = pvmw->vma; local
1064 page_mkclean_one(struct folio *folio, struct vm_area_struct *vma, unsigned long address, void *arg) argument
1075 invalid_mkclean_vma(struct vm_area_struct *vma, void *arg) argument
1120 pfn_mkclean_range(unsigned long pfn, unsigned long nr_pages, pgoff_t pgoff, struct vm_area_struct *vma) argument
1197 folio_move_anon_rmap(struct folio *folio, struct vm_area_struct *vma) argument
1220 __folio_set_anon(struct folio *folio, struct vm_area_struct *vma, unsigned long address, bool exclusive) argument
1252 __page_check_anon_rmap(struct folio *folio, struct page *page, struct vm_area_struct *vma, unsigned long address) argument
1272 __folio_add_anon_rmap(struct folio *folio, struct page *page, int nr_pages, struct vm_area_struct *vma, unsigned long address, rmap_t flags, enum rmap_level level) argument
1347 folio_add_anon_rmap_ptes(struct folio *folio, struct page *page, int nr_pages, struct vm_area_struct *vma, unsigned long address, rmap_t flags) argument
1368 folio_add_anon_rmap_pmd(struct folio *folio, struct page *page, struct vm_area_struct *vma, unsigned long address, rmap_t flags) argument
1392 folio_add_new_anon_rmap(struct folio *folio, struct vm_area_struct *vma, unsigned long address) argument
1434 __folio_add_file_rmap(struct folio *folio, struct page *page, int nr_pages, struct vm_area_struct *vma, enum rmap_level level) argument
1466 folio_add_file_rmap_ptes(struct folio *folio, struct page *page, int nr_pages, struct vm_area_struct *vma) argument
1482 folio_add_file_rmap_pmd(struct folio *folio, struct page *page, struct vm_area_struct *vma) argument
1492 __folio_remove_rmap(struct folio *folio, struct page *page, int nr_pages, struct vm_area_struct *vma, enum rmap_level level) argument
1592 folio_remove_rmap_ptes(struct folio *folio, struct page *page, int nr_pages, struct vm_area_struct *vma) argument
1608 folio_remove_rmap_pmd(struct folio *folio, struct page *page, struct vm_area_struct *vma) argument
1621 try_to_unmap_one(struct folio *folio, struct vm_area_struct *vma, unsigned long address, void *arg) argument
1928 invalid_migration_vma(struct vm_area_struct *vma, void *arg) argument
1970 try_to_migrate_one(struct folio *folio, struct vm_area_struct *vma, unsigned long address, void *arg) argument
2344 page_make_device_exclusive_one(struct folio *folio, struct vm_area_struct *vma, unsigned long address, void *priv) argument
2595 struct vm_area_struct *vma = avc->vma; local
2629 struct vm_area_struct *vma; local
2705 hugetlb_add_anon_rmap(struct folio *folio, struct vm_area_struct *vma, unsigned long address, rmap_t flags) argument
2719 hugetlb_add_new_anon_rmap(struct folio *folio, struct vm_area_struct *vma, unsigned long address) argument
[all...]
H A Dpagewalk.c154 if ((!walk->vma && (pmd_leaf(*pmd) || !pmd_present(*pmd))) ||
159 if (walk->vma)
160 split_huge_pmd(walk->vma, pmd, addr);
208 if ((!walk->vma && (pud_leaf(*pud) || !pud_present(*pud))) ||
213 if (walk->vma)
214 split_huge_pud(walk->vma, pud, addr);
312 struct vm_area_struct *vma = walk->vma; local
313 struct hstate *h = hstate_vma(vma);
321 hugetlb_vma_lock_read(vma);
355 struct vm_area_struct *vma = walk->vma; local
382 struct vm_area_struct *vma = walk->vma; local
412 process_vma_walk_lock(struct vm_area_struct *vma, enum page_walk_lock walk_lock) argument
476 struct vm_area_struct *vma; local
591 walk_page_range_vma(struct vm_area_struct *vma, unsigned long start, unsigned long end, const struct mm_walk_ops *ops, void *private) argument
612 walk_page_vma(struct vm_area_struct *vma, const struct mm_walk_ops *ops, void *private) argument
668 struct vm_area_struct *vma; local
[all...]
H A Dmlock.c333 struct vm_area_struct *vma, unsigned long start,
345 if (!(vma->vm_flags & VM_LOCKED))
353 if (!folio_within_range(folio, vma, start, end))
367 struct vm_area_struct *vma = walk->vma; local
375 ptl = pmd_trans_huge_lock(pmd, vma);
382 if (vma->vm_flags & VM_LOCKED)
389 start_pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl);
399 folio = vm_normal_folio(vma, addr, ptent);
404 if (!allow_mlock_munlock(folio, vma, star
332 allow_mlock_munlock(struct folio *folio, struct vm_area_struct *vma, unsigned long start, unsigned long end, unsigned int step) argument
434 mlock_vma_pages_range(struct vm_area_struct *vma, unsigned long start, unsigned long end, vm_flags_t newflags) argument
477 mlock_fixup(struct vma_iterator *vmi, struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end, vm_flags_t newflags) argument
529 struct vm_area_struct *vma, *prev; local
585 struct vm_area_struct *vma; local
718 struct vm_area_struct *vma, *prev = NULL; local
[all...]
H A Dpgtable-generic.c68 int ptep_set_access_flags(struct vm_area_struct *vma, argument
74 set_pte_at(vma->vm_mm, address, ptep, entry);
75 flush_tlb_fix_spurious_fault(vma, address, ptep);
82 int ptep_clear_flush_young(struct vm_area_struct *vma, argument
86 young = ptep_test_and_clear_young(vma, address, ptep);
88 flush_tlb_page(vma, address);
94 pte_t ptep_clear_flush(struct vm_area_struct *vma, unsigned long address, argument
97 struct mm_struct *mm = (vma)->vm_mm;
101 flush_tlb_page(vma, address);
109 int pmdp_set_access_flags(struct vm_area_struct *vma, argument
124 pmdp_clear_flush_young(struct vm_area_struct *vma, unsigned long address, pmd_t *pmdp) argument
137 pmdp_huge_clear_flush(struct vm_area_struct *vma, unsigned long address, pmd_t *pmdp) argument
150 pudp_huge_clear_flush(struct vm_area_struct *vma, unsigned long address, pud_t *pudp) argument
198 pmdp_invalidate(struct vm_area_struct *vma, unsigned long address, pmd_t *pmdp) argument
209 pmdp_invalidate_ad(struct vm_area_struct *vma, unsigned long address, pmd_t *pmdp) argument
218 pmdp_collapse_flush(struct vm_area_struct *vma, unsigned long address, pmd_t *pmdp) argument
[all...]
/linux-master/arch/nios2/mm/
H A Dcacheflush.c77 struct vm_area_struct *vma; local
85 vma_interval_tree_foreach(vma, &mapping->i_mmap, pgoff, pgoff + nr - 1) {
88 if (vma->vm_mm != mm)
90 if (!(vma->vm_flags & VM_MAYSHARE))
93 start = vma->vm_start + ((pgoff - vma->vm_pgoff) << PAGE_SHIFT);
94 flush_cache_range(vma, start, start + nr * PAGE_SIZE);
134 void flush_cache_range(struct vm_area_struct *vma, unsigned long start, argument
138 if (vma == NULL || (vma
142 flush_icache_pages(struct vm_area_struct *vma, struct page *page, unsigned int nr) argument
152 flush_cache_page(struct vm_area_struct *vma, unsigned long vmaddr, unsigned long pfn) argument
209 update_mmu_cache_range(struct vm_fault *vmf, struct vm_area_struct *vma, unsigned long address, pte_t *ptep, unsigned int nr) argument
261 copy_from_user_page(struct vm_area_struct *vma, struct page *page, unsigned long user_vaddr, void *dst, void *src, int len) argument
272 copy_to_user_page(struct vm_area_struct *vma, struct page *page, unsigned long user_vaddr, void *dst, void *src, int len) argument
[all...]
/linux-master/arch/arm/mm/
H A Dflush.c79 void flush_cache_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) argument
82 vivt_flush_cache_range(vma, start, end);
94 if (vma->vm_flags & VM_EXEC)
98 void flush_cache_pages(struct vm_area_struct *vma, unsigned long user_addr, unsigned long pfn, unsigned int nr) argument
101 vivt_flush_cache_pages(vma, user_addr, pfn, nr);
110 if (vma->vm_flags & VM_EXEC && icache_is_vivt_asid_tagged())
159 void flush_ptrace_access(struct vm_area_struct *vma, struct page *page, argument
163 if (cpumask_test_cpu(smp_processor_id(), mm_cpumask(vma->vm_mm)))
165 if (vma->vm_flags & VM_EXEC)
185 void copy_to_user_page(struct vm_area_struct *vma, struc argument
241 struct vm_area_struct *vma; local
382 __flush_anon_page(struct vm_area_struct *vma, struct page *page, unsigned long vmaddr) argument
[all...]
/linux-master/drivers/char/
H A Duv_mmtimer.c43 static int uv_mmtimer_mmap(struct file *file, struct vm_area_struct *vma);
142 * @vma: VMA to map the registers into
147 static int uv_mmtimer_mmap(struct file *file, struct vm_area_struct *vma) argument
151 if (vma->vm_end - vma->vm_start != PAGE_SIZE)
154 if (vma->vm_flags & VM_WRITE)
160 vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot);
166 if (remap_pfn_range(vma, vma
[all...]
/linux-master/kernel/dma/
H A Dops_helpers.c34 int dma_common_mmap(struct device *dev, struct vm_area_struct *vma, argument
39 unsigned long user_count = vma_pages(vma);
41 unsigned long off = vma->vm_pgoff;
45 vma->vm_page_prot = dma_pgprot(dev, vma->vm_page_prot, attrs);
47 if (dma_mmap_from_dev_coherent(dev, vma, cpu_addr, size, &ret))
53 return remap_pfn_range(vma, vma->vm_start,
54 page_to_pfn(page) + vma->vm_pgoff,
55 user_count << PAGE_SHIFT, vma
[all...]
/linux-master/drivers/xen/
H A Dprivcmd-buf.c86 static void privcmd_buf_vma_open(struct vm_area_struct *vma) argument
88 struct privcmd_buf_vma_private *vma_priv = vma->vm_private_data;
98 static void privcmd_buf_vma_close(struct vm_area_struct *vma) argument
100 struct privcmd_buf_vma_private *vma_priv = vma->vm_private_data;
119 pr_debug("fault: vma=%p %lx-%lx, pgoff=%lx, uv=%p\n",
120 vmf->vma, vmf->vma->vm_start, vmf->vma->vm_end,
132 static int privcmd_buf_mmap(struct file *file, struct vm_area_struct *vma) argument
136 unsigned long count = vma_pages(vma);
[all...]
/linux-master/arch/loongarch/kernel/
H A Dvdso.c49 struct vm_area_struct *vma, struct vm_fault *vmf)
52 struct page *timens_page = find_timens_vvar_page(vma);
82 return vmf_insert_pfn(vma, vmf->address, pfn);
134 struct vm_area_struct *vma; local
139 for_each_vma(vmi, vma) {
140 if (vma_is_special_mapping(vma, &vdso_info.data_mapping))
141 zap_vma_pages(vma);
166 struct vm_area_struct *vma; local
184 vma = _install_special_mapping(mm, data_addr, VVAR_SIZE,
187 if (IS_ERR(vma)) {
48 vvar_fault(const struct vm_special_mapping *sm, struct vm_area_struct *vma, struct vm_fault *vmf) argument
[all...]
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/
H A Duvmm.c113 struct nvkm_vma *vma; local
126 vma = nvkm_vmm_node_search(vmm, addr);
127 if (ret = -ENOENT, !vma || vma->addr != addr) {
129 addr, vma ? vma->addr : ~0ULL);
133 if (ret = -ENOENT, vma->busy) {
134 VMM_DEBUG(vmm, "denied %016llx: %d", addr, vma->busy);
138 if (ret = -EINVAL, !vma->memory) {
143 nvkm_vmm_unmap_locked(vmm, vma, fals
159 struct nvkm_vma *vma; local
238 struct nvkm_vma *vma; local
274 struct nvkm_vma *vma; local
394 struct nvkm_vma vma = { local
[all...]
/linux-master/arch/csky/kernel/
H A Dvdso.c89 const char *arch_vma_name(struct vm_area_struct *vma) argument
91 if (vma->vm_mm && (vma->vm_start == (long)vma->vm_mm->context.vdso))
93 if (vma->vm_mm && (vma->vm_start ==
94 (long)vma->vm_mm->context.vdso + PAGE_SIZE))
/linux-master/drivers/misc/ocxl/
H A Dcontext.c98 static vm_fault_t map_afu_irq(struct vm_area_struct *vma, unsigned long address, argument
108 return vmf_insert_pfn(vma, address, trigger_addr >> PAGE_SHIFT);
111 static vm_fault_t map_pp_mmio(struct vm_area_struct *vma, unsigned long address, argument
134 ret = vmf_insert_pfn(vma, address, pp_mmio_addr >> PAGE_SHIFT);
141 struct vm_area_struct *vma = vmf->vma; local
142 struct ocxl_context *ctx = vma->vm_file->private_data;
151 ret = map_pp_mmio(vma, vmf->address, offset, ctx);
153 ret = map_afu_irq(vma, vmf->address, offset, ctx);
162 struct vm_area_struct *vma)
161 check_mmap_afu_irq(struct ocxl_context *ctx, struct vm_area_struct *vma) argument
187 check_mmap_mmio(struct ocxl_context *ctx, struct vm_area_struct *vma) argument
196 ocxl_context_mmap(struct ocxl_context *ctx, struct vm_area_struct *vma) argument
[all...]

Completed in 493 milliseconds

1234567891011>>