/linux-master/arch/x86/um/ |
H A D | mem_64.c | 5 const char *arch_vma_name(struct vm_area_struct *vma) argument 7 if (vma->vm_mm && vma->vm_start == um_vdso_addr)
|
/linux-master/drivers/gpu/drm/i915/ |
H A D | i915_vma.h | 51 static inline bool i915_vma_is_active(const struct i915_vma *vma) argument 53 return !i915_active_is_idle(&vma->active); 60 int __must_check _i915_vma_move_to_active(struct i915_vma *vma, 65 i915_vma_move_to_active(struct i915_vma *vma, struct i915_request *rq, argument 68 return _i915_vma_move_to_active(vma, rq, &rq->fence, flags); 73 static inline bool i915_vma_is_ggtt(const struct i915_vma *vma) argument 75 return test_bit(I915_VMA_GGTT_BIT, __i915_vma_flags(vma)); 78 static inline bool i915_vma_is_dpt(const struct i915_vma *vma) argument 80 return i915_is_dpt(vma->vm); 83 static inline bool i915_vma_has_ggtt_write(const struct i915_vma *vma) argument 88 i915_vma_set_ggtt_write(struct i915_vma *vma) argument 94 i915_vma_unset_ggtt_write(struct i915_vma *vma) argument 102 i915_vma_is_map_and_fenceable(const struct i915_vma *vma) argument 107 i915_vma_set_userfault(struct i915_vma *vma) argument 113 i915_vma_unset_userfault(struct i915_vma *vma) argument 118 i915_vma_has_userfault(const struct i915_vma *vma) argument 123 i915_vma_is_closed(const struct i915_vma *vma) argument 129 __i915_vma_size(const struct i915_vma *vma) argument 144 i915_vma_size(const struct i915_vma *vma) argument 151 __i915_vma_offset(const struct i915_vma *vma) argument 167 i915_vma_offset(const struct i915_vma *vma) argument 173 i915_ggtt_offset(const struct i915_vma *vma) argument 183 i915_ggtt_pin_bias(struct i915_vma *vma) argument 188 i915_vma_get(struct i915_vma *vma) argument 194 i915_vma_tryget(struct i915_vma *vma) argument 202 i915_vma_put(struct i915_vma *vma) argument 208 i915_vma_compare(struct i915_vma *vma, struct i915_address_space *vm, const struct i915_gtt_view *view) argument 278 i915_vma_lock(struct i915_vma *vma) argument 283 i915_vma_unlock(struct i915_vma *vma) argument 293 i915_vma_pin(struct i915_vma *vma, u64 size, u64 alignment, u64 flags) argument 316 i915_vma_pin_count(const struct i915_vma *vma) argument 321 i915_vma_is_pinned(const struct i915_vma *vma) argument 326 __i915_vma_pin(struct i915_vma *vma) argument 332 __i915_vma_unpin(struct i915_vma *vma) argument 338 i915_vma_unpin(struct i915_vma *vma) argument 344 i915_vma_is_bound(const struct i915_vma *vma, unsigned int where) argument 400 __i915_vma_unpin_fence(struct i915_vma *vma) argument 415 i915_vma_unpin_fence(struct i915_vma *vma) argument 421 i915_vma_fence_id(const struct i915_vma *vma) argument 428 i915_vma_is_scanout(const struct i915_vma *vma) argument 433 i915_vma_mark_scanout(struct i915_vma *vma) argument 438 i915_vma_clear_scanout(struct i915_vma *vma) argument 466 i915_vma_sync(struct i915_vma *vma) argument 482 i915_vma_get_current_resource(struct i915_vma *vma) argument [all...] |
H A D | i915_vma.c | 48 static inline void assert_vma_held_evict(const struct i915_vma *vma) argument 55 if (kref_read(&vma->vm->ref)) 56 assert_object_held_shared(vma->obj); 66 static void i915_vma_free(struct i915_vma *vma) argument 68 return kmem_cache_free(slab_vmas, vma); 75 static void vma_print_allocator(struct i915_vma *vma, const char *reason) argument 79 if (!vma->node.stack) { 80 drm_dbg(vma->obj->base.dev, 81 "vma.node [%08llx + %08llx] %s: unknown owner\n", 82 vma 94 vma_print_allocator(struct i915_vma *vma, const char *reason) argument 107 struct i915_vma *vma = active_to_vma(ref); local 132 struct i915_vma *vma = active_to_vma(ref); local 151 struct i915_vma *vma; local 293 struct i915_vma *vma = rb_entry(rb, struct i915_vma, obj_node); local 327 struct i915_vma *vma; local 405 i915_vma_wait_for_bind(struct i915_vma *vma) argument 425 i915_vma_verify_bind_complete(struct i915_vma *vma) argument 447 i915_vma_resource_init_from_vma(struct i915_vma_resource *vma_res, struct i915_vma *vma) argument 472 i915_vma_bind(struct i915_vma *vma, unsigned int pat_index, u32 flags, struct i915_vma_work *work, struct i915_vma_resource *vma_res) argument 581 i915_vma_pin_iomap(struct i915_vma *vma) argument 649 i915_vma_flush_writes(struct i915_vma *vma) argument 655 i915_vma_unpin_iomap(struct i915_vma *vma) argument 669 struct i915_vma *vma; local 687 i915_vma_misplaced(const struct i915_vma *vma, u64 size, u64 alignment, u64 flags) argument 721 __i915_vma_set_map_and_fenceable(struct i915_vma *vma) argument 740 i915_gem_valid_gtt_space(struct i915_vma *vma, unsigned long color) argument 788 i915_vma_insert(struct i915_vma *vma, struct i915_gem_ww_ctx *ww, u64 size, u64 alignment, u64 flags) argument 927 i915_vma_detach(struct i915_vma *vma) argument 940 try_qad_pin(struct i915_vma *vma, unsigned int flags) argument 1300 __i915_vma_get_pages(struct i915_vma *vma) argument 1347 i915_vma_get_pages(struct i915_vma *vma) argument 1394 __vma_put_pages(struct i915_vma *vma, unsigned int count) argument 1410 i915_vma_put_pages(struct i915_vma *vma) argument 1418 vma_unbind_pages(struct i915_vma *vma) argument 1432 i915_vma_pin_ww(struct i915_vma *vma, struct i915_gem_ww_ctx *ww, u64 size, u64 alignment, u64 flags) argument 1621 __i915_ggtt_pin(struct i915_vma *vma, struct i915_gem_ww_ctx *ww, u32 align, unsigned int flags) argument 1656 i915_ggtt_pin(struct i915_vma *vma, struct i915_gem_ww_ctx *ww, u32 align, unsigned int flags) argument 1688 struct i915_vma *vma; local 1698 __vma_close(struct i915_vma *vma, struct intel_gt *gt) argument 1716 i915_vma_close(struct i915_vma *vma) argument 1733 __i915_vma_remove_closed(struct i915_vma *vma) argument 1738 i915_vma_reopen(struct i915_vma *vma) argument 1748 force_unbind(struct i915_vma *vma) argument 1758 release_references(struct i915_vma *vma, struct intel_gt *gt, bool vm_ddestroy) argument 1812 i915_vma_destroy_locked(struct i915_vma *vma) argument 1821 i915_vma_destroy(struct i915_vma *vma) argument 1840 struct i915_vma *vma, *next; local 1883 __i915_vma_iounmap(struct i915_vma *vma) argument 1897 i915_vma_revoke_mmap(struct i915_vma *vma) argument 1921 __i915_request_await_bind(struct i915_request *rq, struct i915_vma *vma) argument 1926 __i915_vma_move_to_active(struct i915_vma *vma, struct i915_request *rq) argument 1938 _i915_vma_move_to_active(struct i915_vma *vma, struct i915_request *rq, struct dma_fence *fence, unsigned int flags) argument 2013 __i915_vma_evict(struct i915_vma *vma, bool async) argument 2093 __i915_vma_unbind(struct i915_vma *vma) argument 2124 __i915_vma_unbind_async(struct i915_vma *vma) argument 2159 i915_vma_unbind(struct i915_vma *vma) argument 2197 i915_vma_unbind_async(struct i915_vma *vma, bool trylock_vm) argument 2259 i915_vma_unbind_unlocked(struct i915_vma *vma) argument 2270 i915_vma_make_unshrinkable(struct i915_vma *vma) argument 2276 i915_vma_make_shrinkable(struct i915_vma *vma) argument 2281 i915_vma_make_purgeable(struct i915_vma *vma) argument [all...] |
H A D | i915_gem_evict.c | 41 static bool dying_vma(struct i915_vma *vma) argument 43 return !kref_read(&vma->obj->base.refcount); 54 * Not everything in the GGTT is tracked via vma (otherwise we 67 static bool grab_vma(struct i915_vma *vma, struct i915_gem_ww_ctx *ww) argument 73 if (i915_gem_object_get_rcu(vma->obj)) { 74 if (!i915_gem_object_trylock(vma->obj, ww)) { 75 i915_gem_object_put(vma->obj); 80 atomic_and(~I915_VMA_PIN_MASK, &vma->flags); 86 static void ungrab_vma(struct i915_vma *vma) argument 88 if (dying_vma(vma)) 96 mark_free(struct drm_mm_scan *scan, struct i915_gem_ww_ctx *ww, struct i915_vma *vma, unsigned int flags, struct list_head *unwind) argument 112 defer_evict(struct i915_vma *vma) argument 157 struct i915_vma *vma, *next; local 337 struct i915_vma *vma, *next; local 478 struct i915_vma *vma, *vn; local [all...] |
/linux-master/drivers/gpu/drm/nouveau/ |
H A D | nouveau_vmm.c | 29 nouveau_vma_unmap(struct nouveau_vma *vma) argument 31 if (vma->mem) { 32 nvif_vmm_unmap(&vma->vmm->vmm, vma->addr); 33 vma->mem = NULL; 38 nouveau_vma_map(struct nouveau_vma *vma, struct nouveau_mem *mem) argument 40 struct nvif_vma tmp = { .addr = vma->addr }; 41 int ret = nouveau_mem_map(mem, &vma->vmm->vmm, &tmp); 44 vma->mem = mem; 51 struct nouveau_vma *vma; local 64 struct nouveau_vma *vma = *pvma; local 81 struct nouveau_vma *vma; local [all...] |
/linux-master/include/linux/ |
H A D | hugetlb_inline.h | 9 static inline bool is_vm_hugetlb_page(struct vm_area_struct *vma) argument 11 return !!(vma->vm_flags & VM_HUGETLB); 16 static inline bool is_vm_hugetlb_page(struct vm_area_struct *vma) argument
|
H A D | userfaultfd_k.h | 135 extern long uffd_wp_range(struct vm_area_struct *vma, 149 static inline bool is_mergeable_vm_userfaultfd_ctx(struct vm_area_struct *vma, argument 152 return vma->vm_userfaultfd_ctx.ctx == vm_ctx.ctx; 166 static inline bool uffd_disable_huge_pmd_share(struct vm_area_struct *vma) argument 168 return vma->vm_flags & (VM_UFFD_WP | VM_UFFD_MINOR); 178 static inline bool uffd_disable_fault_around(struct vm_area_struct *vma) argument 180 return vma->vm_flags & (VM_UFFD_WP | VM_UFFD_MINOR); 183 static inline bool userfaultfd_missing(struct vm_area_struct *vma) argument 185 return vma->vm_flags & VM_UFFD_MISSING; 188 static inline bool userfaultfd_wp(struct vm_area_struct *vma) argument 193 userfaultfd_minor(struct vm_area_struct *vma) argument 198 userfaultfd_pte_wp(struct vm_area_struct *vma, pte_t pte) argument 204 userfaultfd_huge_pmd_wp(struct vm_area_struct *vma, pmd_t pmd) argument 210 userfaultfd_armed(struct vm_area_struct *vma) argument 215 vma_can_userfault(struct vm_area_struct *vma, unsigned long vm_flags, bool wp_async) argument 276 uffd_wp_range(struct vm_area_struct *vma, unsigned long start, unsigned long len, bool enable_wp) argument 283 is_mergeable_vm_userfaultfd_ctx(struct vm_area_struct *vma, struct vm_userfaultfd_ctx vm_ctx) argument 289 userfaultfd_missing(struct vm_area_struct *vma) argument 294 userfaultfd_wp(struct vm_area_struct *vma) argument 299 userfaultfd_minor(struct vm_area_struct *vma) argument 304 userfaultfd_pte_wp(struct vm_area_struct *vma, pte_t pte) argument 310 userfaultfd_huge_pmd_wp(struct vm_area_struct *vma, pmd_t pmd) argument 317 userfaultfd_armed(struct vm_area_struct *vma) argument 322 dup_userfaultfd(struct vm_area_struct *vma, struct list_head *l) argument 332 mremap_userfaultfd_prep(struct vm_area_struct *vma, struct vm_userfaultfd_ctx *ctx) argument 344 userfaultfd_remove(struct vm_area_struct *vma, unsigned long start, unsigned long end) argument 351 userfaultfd_unmap_prep(struct vm_area_struct *vma, unsigned long start, unsigned long end, struct list_head *uf) argument 363 uffd_disable_fault_around(struct vm_area_struct *vma) argument 368 userfaultfd_wp_unpopulated(struct vm_area_struct *vma) argument 373 userfaultfd_wp_async(struct vm_area_struct *vma) argument 380 userfaultfd_wp_use_markers(struct vm_area_struct *vma) argument [all...] |
H A D | cacheflush.h | 21 static inline void flush_icache_pages(struct vm_area_struct *vma, argument 27 #define flush_icache_page(vma, page) flush_icache_pages(vma, page, 1)
|
/linux-master/arch/sparc/include/asm/ |
H A D | tlbflush_32.h | 11 #define flush_tlb_range(vma, start, end) \ 12 sparc32_cachetlb_ops->tlb_range(vma, start, end) 13 #define flush_tlb_page(vma, addr) \ 14 sparc32_cachetlb_ops->tlb_page(vma, addr)
|
/linux-master/fs/ocfs2/ |
H A D | mmap.h | 5 int ocfs2_mmap(struct file *file, struct vm_area_struct *vma);
|
/linux-master/arch/powerpc/mm/book3s64/ |
H A D | radix_hugetlbpage.c | 10 void radix__flush_hugetlb_page(struct vm_area_struct *vma, unsigned long vmaddr) argument 13 struct hstate *hstate = hstate_file(vma->vm_file); 16 radix__flush_tlb_page_psize(vma->vm_mm, vmaddr, psize); 19 void radix__local_flush_hugetlb_page(struct vm_area_struct *vma, unsigned long vmaddr) argument 22 struct hstate *hstate = hstate_file(vma->vm_file); 25 radix__local_flush_tlb_page_psize(vma->vm_mm, vmaddr, psize); 28 void radix__flush_hugetlb_tlb_range(struct vm_area_struct *vma, unsigned long start, argument 32 struct hstate *hstate = hstate_file(vma->vm_file); 39 radix__flush_tlb_pwc_range_psize(vma->vm_mm, start, end, psize); 41 radix__flush_tlb_range_psize(vma 45 radix__huge_ptep_modify_prot_commit(struct vm_area_struct *vma, unsigned long addr, pte_t *ptep, pte_t old_pte, pte_t pte) argument [all...] |
/linux-master/drivers/pci/ |
H A D | mmap.c | 25 struct vm_area_struct *vma, 32 if (vma->vm_pgoff + vma_pages(vma) > size) 36 vma->vm_page_prot = pgprot_writecombine(vma->vm_page_prot); 38 vma->vm_page_prot = pgprot_device(vma->vm_page_prot); 41 ret = pci_iobar_pfn(pdev, bar, vma); 45 vma->vm_pgoff += (pci_resource_start(pdev, bar) >> PAGE_SHIFT); 47 vma 24 pci_mmap_resource_range(struct pci_dev *pdev, int bar, struct vm_area_struct *vma, enum pci_mmap_state mmap_state, int write_combine) argument 59 pci_mmap_fits(struct pci_dev *pdev, int resno, struct vm_area_struct *vma, enum pci_mmap_api mmap_api) argument [all...] |
/linux-master/arch/arc/kernel/ |
H A D | arc_hostlink.c | 18 static int arc_hl_mmap(struct file *fp, struct vm_area_struct *vma) argument 20 vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot); 22 if (io_remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, 23 vma->vm_end - vma->vm_start, 24 vma->vm_page_prot)) {
|
/linux-master/drivers/gpu/drm/xe/display/ |
H A D | xe_dsb_buffer.c | 15 return xe_bo_ggtt_addr(dsb_buf->vma->bo); 20 iosys_map_wr(&dsb_buf->vma->bo->vmap, idx * 4, u32, val); 25 return iosys_map_rd(&dsb_buf->vma->bo->vmap, idx * 4, u32); 32 iosys_map_memset(&dsb_buf->vma->bo->vmap, idx * 4, val, size); 39 struct i915_vma *vma; local 41 vma = kzalloc(sizeof(*vma), GFP_KERNEL); 42 if (!vma) 51 kfree(vma); 55 vma [all...] |
/linux-master/drivers/gpu/drm/i915/gem/selftests/ |
H A D | igt_gem_utils.h | 25 igt_emit_store_dw(struct i915_vma *vma, 31 struct i915_vma *vma, u64 offset, 35 igt_vma_move_to_active_unlocked(struct i915_vma *vma, struct i915_request *rq, argument 40 i915_vma_lock(vma); 41 err = i915_vma_move_to_active(vma, rq, flags); 42 i915_vma_unlock(vma);
|
/linux-master/arch/sh/include/asm/ |
H A D | tlbflush.h | 10 * - flush_tlb_page(vma, vmaddr) flushes one page 11 * - flush_tlb_range(vma, start, end) flushes a range of pages 16 extern void local_flush_tlb_range(struct vm_area_struct *vma, 19 extern void local_flush_tlb_page(struct vm_area_struct *vma, 31 extern void flush_tlb_range(struct vm_area_struct *vma, unsigned long start, 33 extern void flush_tlb_page(struct vm_area_struct *vma, unsigned long page); 41 #define flush_tlb_page(vma, page) local_flush_tlb_page(vma, page) 44 #define flush_tlb_range(vma, start, end) \ 45 local_flush_tlb_range(vma, star [all...] |
/linux-master/arch/csky/include/asm/ |
H A D | tlbflush.h | 11 * - flush_tlb_page(vma, vmaddr) flushes one page 12 * - flush_tlb_range(vma, start, end) flushes a range of pages 17 extern void flush_tlb_page(struct vm_area_struct *vma, unsigned long page); 18 extern void flush_tlb_range(struct vm_area_struct *vma, unsigned long start,
|
/linux-master/drivers/gpu/drm/msm/ |
H A D | msm_gem_vma.c | 41 /* Actually unmap memory for the vma */ 42 void msm_gem_vma_purge(struct msm_gem_vma *vma) argument 44 struct msm_gem_address_space *aspace = vma->aspace; 45 unsigned size = vma->node.size; 48 if (!vma->mapped) 51 aspace->mmu->funcs->unmap(aspace->mmu, vma->iova, size); 53 vma->mapped = false; 56 /* Map and pin vma: */ 58 msm_gem_vma_map(struct msm_gem_vma *vma, int prot, argument 61 struct msm_gem_address_space *aspace = vma 94 msm_gem_vma_close(struct msm_gem_vma *vma) argument 112 struct msm_gem_vma *vma; local 124 msm_gem_vma_init(struct msm_gem_vma *vma, int size, u64 range_start, u64 range_end) argument [all...] |
/linux-master/drivers/gpu/drm/xe/compat-i915-headers/gem/ |
H A D | i915_gem_mman.h | 12 static inline int i915_gem_fb_mmap(struct xe_bo *bo, struct vm_area_struct *vma) argument 14 return drm_gem_prime_mmap(&bo->ttm.base, vma);
|
/linux-master/arch/nios2/include/asm/ |
H A D | tlbflush.h | 16 * - flush_tlb_range(vma, start, end) flushes a range of pages 17 * - flush_tlb_page(vma, address) flushes a page 21 * - reload_tlb_page(vma, address, pte) flushes the TLB for address like 26 extern void flush_tlb_range(struct vm_area_struct *vma, unsigned long start, 30 static inline void flush_tlb_page(struct vm_area_struct *vma, argument 33 flush_tlb_range(vma, address, address + PAGE_SIZE); 41 extern void reload_tlb_page(struct vm_area_struct *vma, unsigned long addr,
|
/linux-master/drivers/gpu/drm/i915/gt/ |
H A D | gen7_renderclear.h | 13 struct i915_vma * const vma);
|
/linux-master/drivers/gpu/drm/i915/display/ |
H A D | intel_fb_pin.c | 30 struct i915_vma *vma; local 74 vma = i915_vma_instance(obj, vm, view); 75 if (IS_ERR(vma)) { 76 ret = PTR_ERR(vma); 80 if (i915_vma_misplaced(vma, 0, alignment, 0)) { 81 ret = i915_vma_unbind(vma); 86 ret = i915_vma_pin_ww(vma, &ww, 0, alignment, PIN_GLOBAL); 91 vma = ERR_PTR(ret); 95 vma->display_alignment = max(vma 118 struct i915_vma *vma; local 231 intel_unpin_fb_vma(struct i915_vma *vma, unsigned long flags) argument 244 struct i915_vma *vma; local 296 struct i915_vma *vma; local [all...] |
/linux-master/arch/arc/include/asm/ |
H A D | tlbflush.h | 13 void local_flush_tlb_page(struct vm_area_struct *vma, unsigned long page); 15 void local_flush_tlb_range(struct vm_area_struct *vma, 18 void local_flush_pmd_tlb_range(struct vm_area_struct *vma, unsigned long start, 23 #define flush_tlb_range(vma, s, e) local_flush_tlb_range(vma, s, e) 24 #define flush_tlb_page(vma, page) local_flush_tlb_page(vma, page) 29 #define flush_pmd_tlb_range(vma, s, e) local_flush_pmd_tlb_range(vma, s, e) 32 extern void flush_tlb_range(struct vm_area_struct *vma, unsigne [all...] |
/linux-master/mm/ |
H A D | msync.c | 36 struct vm_area_struct *vma; local 63 vma = find_vma(mm, start); 70 if (!vma) 72 /* Here start < vma->vm_end. */ 73 if (start < vma->vm_start) { 76 start = vma->vm_start; 81 /* Here vma->vm_start <= start < vma->vm_end. */ 83 (vma->vm_flags & VM_LOCKED)) { 87 file = vma [all...] |
/linux-master/tools/testing/selftests/bpf/progs/ |
H A D | bpf_iter_task_vmas.c | 28 struct vm_area_struct *vma = ctx->vma; local 34 if (task == (void *)0 || vma == (void *)0) 37 file = vma->vm_file; 43 perm_str[0] = (vma->vm_flags & VM_READ) ? 'r' : '-'; 44 perm_str[1] = (vma->vm_flags & VM_WRITE) ? 'w' : '-'; 45 perm_str[2] = (vma->vm_flags & VM_EXEC) ? 'x' : '-'; 46 perm_str[3] = (vma->vm_flags & VM_MAYSHARE) ? 's' : 'p'; 47 BPF_SEQ_PRINTF(seq, "%08llx-%08llx %s ", vma->vm_start, vma [all...] |