/linux-master/arch/m68k/mm/ |
H A D | fault.c | 74 struct vm_area_struct * vma; local 95 vma = find_vma(mm, address); 96 if (!vma) 98 if (vma->vm_start <= address) 100 if (!(vma->vm_flags & VM_GROWSDOWN)) 110 vma = expand_stack(mm, address); 111 if (!vma) 124 if (!(vma->vm_flags & VM_WRITE)) 131 if (unlikely(!vma_is_accessible(vma))) 141 fault = handle_mm_fault(vma, addres [all...] |
/linux-master/drivers/video/fbdev/core/ |
H A D | fb_io_fops.c | 141 int fb_io_mmap(struct fb_info *info, struct vm_area_struct *vma) argument 154 if (vma->vm_pgoff >= mmio_pgoff) { 158 vma->vm_pgoff -= mmio_pgoff; 163 vma->vm_page_prot = vm_get_page_prot(vma->vm_flags); 164 vma->vm_page_prot = pgprot_framebuffer(vma->vm_page_prot, vma->vm_start, 165 vma->vm_end, start); 167 return vm_iomap_memory(vma, star [all...] |
/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_trace.h | 387 TP_PROTO(struct xe_vma *vma), 388 TP_ARGS(vma), 391 __field(struct xe_vma *, vma) 399 __entry->vma = vma; 400 __entry->asid = xe_vma_vm(vma)->usm.asid; 401 __entry->start = xe_vma_start(vma); 402 __entry->end = xe_vma_end(vma) - 1; 403 __entry->ptr = xe_vma_userptr(vma); 406 TP_printk("vma [all...] |
H A D | xe_gt_pagefault.c | 69 static bool vma_is_valid(struct xe_tile *tile, struct xe_vma *vma) argument 71 return BIT(tile->id) & vma->tile_present && 72 !(BIT(tile->id) & vma->tile_invalidated); 75 static bool vma_matches(struct xe_vma *vma, u64 page_addr) argument 77 if (page_addr > xe_vma_end(vma) - 1 || 78 page_addr + SZ_4K - 1 < xe_vma_start(vma)) 86 struct xe_vma *vma = NULL; local 90 vma = vm->usm.last_fault_vma; 92 if (!vma) 93 vma 98 xe_pf_begin(struct drm_exec *exec, struct xe_vma *vma, bool atomic, unsigned int id) argument 135 struct xe_vma *vma = NULL; local 513 struct xe_vma *vma; local [all...] |
/linux-master/arch/sh/kernel/ |
H A D | sys_sh.c | 60 struct vm_area_struct *vma; local 73 vma = find_vma (current->mm, addr); 74 if (vma == NULL || addr < vma->vm_start || addr + len > vma->vm_end) {
|
/linux-master/drivers/gpu/drm/msm/ |
H A D | msm_gem.c | 289 struct vm_area_struct *vma = vmf->vma; local 290 struct drm_gem_object *obj = vma->vm_private_data; 321 pgoff = (vmf->address - vma->vm_start) >> PAGE_SHIFT; 328 ret = vmf_insert_pfn(vma, vmf->address, pfn); 369 struct msm_gem_vma *vma; local 373 vma = msm_gem_vma_new(aspace); 374 if (!vma) 377 list_add_tail(&vma->list, &msm_obj->vmas); 379 return vma; 386 struct msm_gem_vma *vma; local 398 del_vma(struct msm_gem_vma *vma) argument 417 struct msm_gem_vma *vma; local 435 struct msm_gem_vma *vma, *tmp; local 448 struct msm_gem_vma *vma; local 475 msm_gem_pin_vma_locked(struct drm_gem_object *obj, struct msm_gem_vma *vma) argument 538 struct msm_gem_vma *vma; local 587 struct msm_gem_vma *vma; local 605 struct msm_gem_vma *vma = lookup_vma(obj, aspace); local 633 struct msm_gem_vma *vma; local 655 struct msm_gem_vma *vma; local 927 struct msm_gem_vma *vma; local 1065 msm_gem_object_mmap(struct drm_gem_object *obj, struct vm_area_struct *vma) argument 1199 struct msm_gem_vma *vma; local [all...] |
/linux-master/mm/ |
H A D | migrate_device.c | 44 if (!vma_is_anonymous(walk->vma)) 63 struct vm_area_struct *vma = walk->vma; local 64 struct mm_struct *mm = vma->vm_mm; 85 split_huge_pmd(vma, pmdp, addr); 118 if (vma_is_anonymous(vma)) { 153 page = vm_normal_page(migrate->vma, addr, pte); 199 flush_cache_page(vma, addr, pte_pfn(pte)); 203 pte = ptep_clear_flush(vma, addr, ptep); 257 folio_remove_rmap_pte(folio, page, vma); 571 struct vm_area_struct *vma = migrate->vma; local [all...] |
/linux-master/arch/arc/mm/ |
H A D | tlb.c | 190 * (Android Binder ended up calling this for vma->mm != tsk->mm, 208 void local_flush_tlb_range(struct vm_area_struct *vma, unsigned long start, argument 222 local_flush_tlb_mm(vma->vm_mm); 235 if (asid_mm(vma->vm_mm, cpu) != MM_CTXT_NO_ASID) { 237 tlb_entry_erase(start | hw_pid(vma->vm_mm, cpu)); 278 void local_flush_tlb_page(struct vm_area_struct *vma, unsigned long page) argument 288 if (asid_mm(vma->vm_mm, cpu) != MM_CTXT_NO_ASID) { 289 tlb_entry_erase((page & PAGE_MASK) | hw_pid(vma->vm_mm, cpu)); 344 void flush_tlb_page(struct vm_area_struct *vma, unsigned long uaddr) argument 347 .ta_vma = vma, 354 flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) argument 367 flush_pmd_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) argument 394 create_tlb(struct vm_area_struct *vma, unsigned long vaddr, pte_t *ptep) argument 472 update_mmu_cache_range(struct vm_fault *vmf, struct vm_area_struct *vma, unsigned long vaddr_unaligned, pte_t *ptep, unsigned int nr) argument 530 update_mmu_cache_pmd(struct vm_area_struct *vma, unsigned long addr, pmd_t *pmd) argument 537 local_flush_pmd_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) argument [all...] |
/linux-master/arch/riscv/mm/ |
H A D | fault.c | 193 static inline bool access_error(unsigned long cause, struct vm_area_struct *vma) argument 197 if (!(vma->vm_flags & VM_EXEC)) { 203 if (!(vma->vm_flags & (VM_READ | VM_WRITE))) { 208 if (!(vma->vm_flags & VM_WRITE)) { 225 struct vm_area_struct *vma; local 289 vma = lock_vma_under_rcu(mm, addr); 290 if (!vma) 293 if (unlikely(access_error(cause, vma))) { 294 vma_end_read(vma); 298 fault = handle_mm_fault(vma, add [all...] |
H A D | pgtable.c | 8 int ptep_set_access_flags(struct vm_area_struct *vma, argument 13 __set_pte_at(vma->vm_mm, ptep, entry); 21 int ptep_test_and_clear_young(struct vm_area_struct *vma, argument 126 pmd_t pmdp_collapse_flush(struct vm_area_struct *vma, argument 129 pmd_t pmd = pmdp_huge_get_and_clear(vma->vm_mm, address, pmdp); 136 * valid leaf PTE at the level 1 page table. Since the sfence.vma 141 flush_tlb_mm(vma->vm_mm);
|
/linux-master/kernel/events/ |
H A D | uprobes.c | 105 * We keep the vma's vm_start rather than a pointer to the vma 107 * the vma go away, and we must handle that reasonably gracefully. 113 * valid_vma: Verify if the specified vma is an executable vma 118 * executable vma. 120 static bool valid_vma(struct vm_area_struct *vma, bool is_register) argument 127 return vma->vm_file && (vma->vm_flags & flags) == VM_MAYEXEC; 130 static unsigned long offset_to_vaddr(struct vm_area_struct *vma, loff_ argument 135 vaddr_to_offset(struct vm_area_struct *vma, unsigned long vaddr) argument 153 __replace_page(struct vm_area_struct *vma, unsigned long addr, struct page *old_page, struct page *new_page) argument 337 valid_ref_ctr_vma(struct uprobe *uprobe, struct vm_area_struct *vma) argument 464 struct vm_area_struct *vma; local 890 install_breakpoint(struct uprobe *uprobe, struct mm_struct *mm, struct vm_area_struct *vma, unsigned long vaddr) argument 962 struct vm_area_struct *vma; local 1045 struct vm_area_struct *vma; local 1235 struct vm_area_struct *vma; local 1288 build_probe_list(struct inode *inode, struct vm_area_struct *vma, unsigned long start, unsigned long end, struct list_head *head) argument 1323 delayed_ref_ctr_inc(struct vm_area_struct *vma) argument 1357 uprobe_mmap(struct vm_area_struct *vma) argument 1399 vma_has_uprobes(struct vm_area_struct *vma, unsigned long start, unsigned long end) argument 1420 uprobe_munmap(struct vm_area_struct *vma, unsigned long start, unsigned long end) argument 1439 struct vm_area_struct *vma; local 1988 struct vm_area_struct *vma; local 2043 struct vm_area_struct *vma; local [all...] |
/linux-master/include/linux/ |
H A D | pgtable.h | 283 extern int ptep_set_access_flags(struct vm_area_struct *vma, 290 extern int pmdp_set_access_flags(struct vm_area_struct *vma, 293 extern int pudp_set_access_flags(struct vm_area_struct *vma, 297 static inline int pmdp_set_access_flags(struct vm_area_struct *vma, argument 304 static inline int pudp_set_access_flags(struct vm_area_struct *vma, argument 350 static inline int ptep_test_and_clear_young(struct vm_area_struct *vma, argument 359 set_pte_at(vma->vm_mm, address, ptep, pte_mkold(pte)); 366 static inline int pmdp_test_and_clear_young(struct vm_area_struct *vma, argument 375 set_pmd_at(vma->vm_mm, address, pmdp, pmd_mkold(pmd)); 379 static inline int pmdp_test_and_clear_young(struct vm_area_struct *vma, argument 403 pmdp_clear_flush_young(struct vm_area_struct *vma, unsigned long address, pmd_t *pmdp) argument 437 arch_check_zapped_pte(struct vm_area_struct *vma, pte_t pte) argument 444 arch_check_zapped_pmd(struct vm_area_struct *vma, pmd_t pmd) argument 582 pmdp_huge_get_and_clear_full(struct vm_area_struct *vma, unsigned long address, pmd_t *pmdp, int full) argument 591 pudp_huge_get_and_clear_full(struct vm_area_struct *vma, unsigned long address, pud_t *pudp, int full) argument 689 update_mmu_tlb(struct vm_area_struct *vma, unsigned long address, pte_t *ptep) argument 727 pte_mkwrite(pte_t pte, struct vm_area_struct *vma) argument 734 pmd_mkwrite(pmd_t pmd, struct vm_area_struct *vma) argument 837 pmdp_collapse_flush(struct vm_area_struct *vma, unsigned long address, pmd_t *pmdp) argument 867 generic_pmdp_establish(struct vm_area_struct *vma, unsigned long address, pmd_t *pmdp, pmd_t pmd) argument 1022 arch_do_swap_page(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long addr, pte_t pte, pte_t oldpte) argument 1040 arch_unmap_one(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long addr, pte_t orig_pte) argument 1190 __ptep_modify_prot_start(struct vm_area_struct *vma, unsigned long addr, pte_t *ptep) argument 1202 __ptep_modify_prot_commit(struct vm_area_struct *vma, unsigned long addr, pte_t *ptep, pte_t pte) argument 1228 ptep_modify_prot_start(struct vm_area_struct *vma, unsigned long addr, pte_t *ptep) argument 1239 ptep_modify_prot_commit(struct vm_area_struct *vma, unsigned long addr, pte_t *ptep, pte_t old_pte, pte_t pte) argument 1406 track_pfn_remap(struct vm_area_struct *vma, pgprot_t *prot, unsigned long pfn, unsigned long addr, unsigned long size) argument 1417 track_pfn_insert(struct vm_area_struct *vma, pgprot_t *prot, pfn_t pfn) argument 1426 track_pfn_copy(struct vm_area_struct *vma) argument 1436 untrack_pfn(struct vm_area_struct *vma, unsigned long pfn, unsigned long size, bool mm_wr_locked) argument 1446 untrack_pfn_clear(struct vm_area_struct *vma) argument [all...] |
H A D | rmap.h | 24 * directly to a vma: instead it points to an anon_vma, on whose list 27 * After unlinking the last vma on the list, we must garbage collect 29 * pointing to this anon_vma once its vma list is empty. 36 * guarantee that the vma of page tables will exist for 83 struct vm_area_struct *vma; member in struct:anon_vma_chain 159 static inline int anon_vma_prepare(struct vm_area_struct *vma) argument 161 if (likely(vma->anon_vma)) 164 return __anon_vma_prepare(vma); 167 static inline void anon_vma_merge(struct vm_area_struct *vma, argument 170 VM_BUG_ON_VMA(vma 264 hugetlb_try_dup_anon_rmap(struct folio *folio, struct vm_area_struct *vma) argument 644 struct vm_area_struct *vma; member in struct:page_vma_mapped_walk [all...] |
H A D | ksm.h | 19 int ksm_madvise(struct vm_area_struct *vma, unsigned long start, 22 void ksm_add_vma(struct vm_area_struct *vma); 70 * no problem, it will be assigned to this vma's anon_vma; but thereafter, 76 * We'd like to make this conditional on vma->vm_flags & VM_MERGEABLE, 77 * but what if the vma was unmerged while the page was swapped out? 80 struct vm_area_struct *vma, unsigned long addr); 96 static inline void ksm_add_vma(struct vm_area_struct *vma) argument 126 static inline int ksm_madvise(struct vm_area_struct *vma, unsigned long start, argument 133 struct vm_area_struct *vma, unsigned long addr) 132 ksm_might_need_to_copy(struct folio *folio, struct vm_area_struct *vma, unsigned long addr) argument
|
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_execbuffer.c | 39 struct i915_vma *vma; member in struct:eb_vma 42 /** This vma's place in the execbuf reservation list */ 252 struct eb_vma *vma; member in struct:i915_execbuffer 262 /** identity of the batch obj/vma */ 275 /** list of vma not yet bound during reservation phase */ 278 /** list of vma that have execobj.relocation_count */ 382 const struct i915_vma *vma, 385 const u64 start = i915_vma_offset(vma); 386 const u64 size = i915_vma_size(vma); 407 !i915_vma_is_map_and_fenceable(vma)) 381 eb_vma_misplaced(const struct drm_i915_gem_exec_object2 *entry, const struct i915_vma *vma, unsigned int flags) argument 444 struct i915_vma *vma = ev->vma; local 502 eb_validate_vma(struct i915_execbuffer *eb, struct drm_i915_gem_exec_object2 *entry, struct i915_vma *vma) argument 563 eb_add_vma(struct i915_execbuffer *eb, unsigned int *current_batch, unsigned int i, struct i915_vma *vma) argument 662 struct i915_vma *vma = ev->vma; local 840 __eb_add_lut(struct i915_execbuffer *eb, u32 handle, struct i915_vma *vma) argument 896 struct i915_vma *vma; local 951 struct i915_vma *vma; local 1004 struct i915_vma *vma = ev->vma; local 1028 struct i915_vma *vma = ev->vma; local 1092 struct i915_vma *vma = ev->vma; local 1283 struct i915_vma *vma = ERR_PTR(-ENODEV); local 1351 reloc_vaddr(struct i915_vma *vma, struct i915_execbuffer *eb, unsigned long page) argument 1393 relocate_entry(struct i915_vma *vma, const struct drm_i915_gem_relocation_entry *reloc, struct i915_execbuffer *eb, const struct i915_vma *target) argument 1473 struct i915_vma *vma = target->vma; local 2012 struct i915_vma *vma = ev->vma; local 2104 struct i915_vma *vma = ev->vma; local 2253 struct i915_vma *vma; local 2267 eb_dispatch_secure(struct i915_execbuffer *eb, struct i915_vma *vma) argument [all...] |
/linux-master/arch/powerpc/mm/book3s64/ |
H A D | pgtable.c | 48 int pmdp_set_access_flags(struct vm_area_struct *vma, unsigned long address, argument 54 assert_spin_locked(pmd_lockptr(vma->vm_mm, pmdp)); 62 __ptep_set_access_flags(vma, pmdp_ptep(pmdp), 68 int pudp_set_access_flags(struct vm_area_struct *vma, unsigned long address, argument 74 assert_spin_locked(pud_lockptr(vma->vm_mm, pudp)); 82 __ptep_set_access_flags(vma, pudp_ptep(pudp), 89 int pmdp_test_and_clear_young(struct vm_area_struct *vma, argument 92 return __pmdp_test_and_clear_young(vma->vm_mm, address, pmdp); 95 int pudp_test_and_clear_young(struct vm_area_struct *vma, argument 98 return __pudp_test_and_clear_young(vma 168 pmdp_invalidate(struct vm_area_struct *vma, unsigned long address, pmd_t *pmdp) argument 178 pmdp_huge_get_and_clear_full(struct vm_area_struct *vma, unsigned long addr, pmd_t *pmdp, int full) argument 196 pudp_huge_get_and_clear_full(struct vm_area_struct *vma, unsigned long addr, pud_t *pudp, int full) argument 520 ptep_modify_prot_start(struct vm_area_struct *vma, unsigned long addr, pte_t *ptep) argument 536 ptep_modify_prot_commit(struct vm_area_struct *vma, unsigned long addr, pte_t *ptep, pte_t old_pte, pte_t pte) argument 558 pmd_move_must_withdraw(struct spinlock *new_pmd_ptl, struct spinlock *old_pmd_ptl, struct vm_area_struct *vma) argument [all...] |
/linux-master/fs/coda/ |
H A D | file.c | 123 coda_vm_open(struct vm_area_struct *vma) argument 126 container_of(vma->vm_ops, struct coda_vm_ops, vm_ops); 131 cvm_ops->host_vm_ops->open(vma); 135 coda_vm_close(struct vm_area_struct *vma) argument 138 container_of(vma->vm_ops, struct coda_vm_ops, vm_ops); 141 cvm_ops->host_vm_ops->close(vma); 144 vma->vm_ops = cvm_ops->host_vm_ops; 151 coda_file_mmap(struct file *coda_file, struct vm_area_struct *vma) argument 166 if (WARN_ON(coda_file != vma->vm_file)) 169 count = vma [all...] |
/linux-master/arch/powerpc/platforms/book3s/ |
H A D | vas-api.c | 397 struct vm_area_struct *vma = vmf->vma; local 398 struct file *fp = vma->vm_file; 423 if (txwin->task_ref.vma != vmf->vma) { 438 fault = vmf_insert_pfn(vma, vma->vm_start, 471 static int coproc_mmap(struct file *fp, struct vm_area_struct *vma) argument 482 if ((vma->vm_end - vma [all...] |
/linux-master/arch/powerpc/mm/ |
H A D | fault.c | 88 struct vm_area_struct *vma) 107 * 6. T1 : reaches here, sees vma_pkey(vma)=5, when we really 110 pkey = vma_pkey(vma); 231 struct vm_area_struct *vma) 238 if (!arch_vma_access_permitted(vma, is_write, is_exec, 0)) 244 static bool access_error(bool is_write, bool is_exec, struct vm_area_struct *vma) argument 257 return !(vma->vm_flags & VM_EXEC) && 259 !(vma->vm_flags & (VM_READ | VM_WRITE))); 263 if (unlikely(!(vma->vm_flags & VM_WRITE))) 274 if (unlikely(!vma_is_accessible(vma))) 87 bad_access_pkey(struct pt_regs *regs, unsigned long address, struct vm_area_struct *vma) argument 230 access_pkey_error(bool is_write, bool is_exec, bool is_pkey, struct vm_area_struct *vma) argument 409 struct vm_area_struct * vma; local [all...] |
H A D | copro_fault.c | 26 struct vm_area_struct *vma; local 36 vma = lock_mm_and_find_vma(mm, ea, NULL); 37 if (!vma) 43 if (!(vma->vm_flags & VM_WRITE)) 46 if (!(vma->vm_flags & (VM_READ | VM_EXEC))) 59 *flt = handle_mm_fault(vma, ea, is_write ? FAULT_FLAG_WRITE : 0, NULL);
|
/linux-master/arch/x86/kernel/cpu/sgx/ |
H A D | virt.c | 34 struct vm_area_struct *vma, unsigned long addr) 43 index = vma->vm_pgoff + PFN_DOWN(addr - vma->vm_start); 59 ret = vmf_insert_pfn(vma, addr, pfn); 76 struct vm_area_struct *vma = vmf->vma; local 77 struct sgx_vepc *vepc = vma->vm_private_data; 81 ret = __sgx_vepc_fault(vepc, vma, vmf->address); 88 mmap_read_unlock(vma->vm_mm); 99 static int sgx_vepc_mmap(struct file *file, struct vm_area_struct *vma) argument 33 __sgx_vepc_fault(struct sgx_vepc *vepc, struct vm_area_struct *vma, unsigned long addr) argument [all...] |
/linux-master/drivers/gpu/drm/exynos/ |
H A D | exynos_drm_gem.c | 23 static int exynos_drm_gem_mmap(struct drm_gem_object *obj, struct vm_area_struct *vma); 271 struct vm_area_struct *vma) 277 vm_flags_clear(vma, VM_PFNMAP); 278 vma->vm_pgoff = 0; 280 vm_size = vma->vm_end - vma->vm_start; 286 ret = dma_mmap_attrs(to_dma_dev(drm_dev), vma, exynos_gem->cookie, 363 static int exynos_drm_gem_mmap(struct drm_gem_object *obj, struct vm_area_struct *vma) argument 369 return dma_buf_mmap(obj->dma_buf, vma, 0); 371 vm_flags_set(vma, VM_I 270 exynos_drm_gem_mmap_buffer(struct exynos_drm_gem *exynos_gem, struct vm_area_struct *vma) argument [all...] |
/linux-master/arch/xtensa/include/asm/ |
H A D | tlbflush.h | 29 * - flush_tlb_page(vma, page) flushes a single page 30 * - flush_tlb_range(vma, vmaddr, end) flushes a range of pages 35 void local_flush_tlb_page(struct vm_area_struct *vma, 37 void local_flush_tlb_range(struct vm_area_struct *vma, 54 #define flush_tlb_page(vma, page) local_flush_tlb_page(vma, page) 55 #define flush_tlb_range(vma, vmaddr, end) local_flush_tlb_range(vma, vmaddr, \
|
H A D | cacheflush.h | 136 void local_flush_cache_range(struct vm_area_struct *vma, 138 void local_flush_cache_page(struct vm_area_struct *vma, 167 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ 174 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \
|
/linux-master/arch/parisc/kernel/ |
H A D | vdso.c | 30 struct vm_area_struct *vma) 32 current->mm->context.vdso_base = vma->vm_start; 49 * This is called from binfmt_elf, we create the special vma for the 59 struct vm_area_struct *vma; local 83 vma = _install_special_mapping(mm, vdso_text_start, vdso_text_len, 87 if (IS_ERR(vma)) { 89 rc = PTR_ERR(vma); 29 vdso_mremap(const struct vm_special_mapping *sm, struct vm_area_struct *vma) argument
|