Searched refs:vma (Results 1 - 25 of 1003) sorted by relevance

1234567891011>>

/linux-master/arch/x86/um/
H A Dmem_64.c5 const char *arch_vma_name(struct vm_area_struct *vma) argument
7 if (vma->vm_mm && vma->vm_start == um_vdso_addr)
/linux-master/drivers/gpu/drm/i915/
H A Di915_vma.h51 static inline bool i915_vma_is_active(const struct i915_vma *vma) argument
53 return !i915_active_is_idle(&vma->active);
60 int __must_check _i915_vma_move_to_active(struct i915_vma *vma,
65 i915_vma_move_to_active(struct i915_vma *vma, struct i915_request *rq, argument
68 return _i915_vma_move_to_active(vma, rq, &rq->fence, flags);
73 static inline bool i915_vma_is_ggtt(const struct i915_vma *vma) argument
75 return test_bit(I915_VMA_GGTT_BIT, __i915_vma_flags(vma));
78 static inline bool i915_vma_is_dpt(const struct i915_vma *vma) argument
80 return i915_is_dpt(vma->vm);
83 static inline bool i915_vma_has_ggtt_write(const struct i915_vma *vma) argument
88 i915_vma_set_ggtt_write(struct i915_vma *vma) argument
94 i915_vma_unset_ggtt_write(struct i915_vma *vma) argument
102 i915_vma_is_map_and_fenceable(const struct i915_vma *vma) argument
107 i915_vma_set_userfault(struct i915_vma *vma) argument
113 i915_vma_unset_userfault(struct i915_vma *vma) argument
118 i915_vma_has_userfault(const struct i915_vma *vma) argument
123 i915_vma_is_closed(const struct i915_vma *vma) argument
129 __i915_vma_size(const struct i915_vma *vma) argument
144 i915_vma_size(const struct i915_vma *vma) argument
151 __i915_vma_offset(const struct i915_vma *vma) argument
167 i915_vma_offset(const struct i915_vma *vma) argument
173 i915_ggtt_offset(const struct i915_vma *vma) argument
183 i915_ggtt_pin_bias(struct i915_vma *vma) argument
188 i915_vma_get(struct i915_vma *vma) argument
194 i915_vma_tryget(struct i915_vma *vma) argument
202 i915_vma_put(struct i915_vma *vma) argument
208 i915_vma_compare(struct i915_vma *vma, struct i915_address_space *vm, const struct i915_gtt_view *view) argument
278 i915_vma_lock(struct i915_vma *vma) argument
283 i915_vma_unlock(struct i915_vma *vma) argument
293 i915_vma_pin(struct i915_vma *vma, u64 size, u64 alignment, u64 flags) argument
316 i915_vma_pin_count(const struct i915_vma *vma) argument
321 i915_vma_is_pinned(const struct i915_vma *vma) argument
326 __i915_vma_pin(struct i915_vma *vma) argument
332 __i915_vma_unpin(struct i915_vma *vma) argument
338 i915_vma_unpin(struct i915_vma *vma) argument
344 i915_vma_is_bound(const struct i915_vma *vma, unsigned int where) argument
400 __i915_vma_unpin_fence(struct i915_vma *vma) argument
415 i915_vma_unpin_fence(struct i915_vma *vma) argument
421 i915_vma_fence_id(const struct i915_vma *vma) argument
428 i915_vma_is_scanout(const struct i915_vma *vma) argument
433 i915_vma_mark_scanout(struct i915_vma *vma) argument
438 i915_vma_clear_scanout(struct i915_vma *vma) argument
466 i915_vma_sync(struct i915_vma *vma) argument
482 i915_vma_get_current_resource(struct i915_vma *vma) argument
[all...]
H A Di915_vma.c48 static inline void assert_vma_held_evict(const struct i915_vma *vma) argument
55 if (kref_read(&vma->vm->ref))
56 assert_object_held_shared(vma->obj);
66 static void i915_vma_free(struct i915_vma *vma) argument
68 return kmem_cache_free(slab_vmas, vma);
75 static void vma_print_allocator(struct i915_vma *vma, const char *reason) argument
79 if (!vma->node.stack) {
80 drm_dbg(vma->obj->base.dev,
81 "vma.node [%08llx + %08llx] %s: unknown owner\n",
82 vma
94 vma_print_allocator(struct i915_vma *vma, const char *reason) argument
107 struct i915_vma *vma = active_to_vma(ref); local
132 struct i915_vma *vma = active_to_vma(ref); local
151 struct i915_vma *vma; local
293 struct i915_vma *vma = rb_entry(rb, struct i915_vma, obj_node); local
327 struct i915_vma *vma; local
405 i915_vma_wait_for_bind(struct i915_vma *vma) argument
425 i915_vma_verify_bind_complete(struct i915_vma *vma) argument
447 i915_vma_resource_init_from_vma(struct i915_vma_resource *vma_res, struct i915_vma *vma) argument
472 i915_vma_bind(struct i915_vma *vma, unsigned int pat_index, u32 flags, struct i915_vma_work *work, struct i915_vma_resource *vma_res) argument
581 i915_vma_pin_iomap(struct i915_vma *vma) argument
649 i915_vma_flush_writes(struct i915_vma *vma) argument
655 i915_vma_unpin_iomap(struct i915_vma *vma) argument
669 struct i915_vma *vma; local
687 i915_vma_misplaced(const struct i915_vma *vma, u64 size, u64 alignment, u64 flags) argument
721 __i915_vma_set_map_and_fenceable(struct i915_vma *vma) argument
740 i915_gem_valid_gtt_space(struct i915_vma *vma, unsigned long color) argument
788 i915_vma_insert(struct i915_vma *vma, struct i915_gem_ww_ctx *ww, u64 size, u64 alignment, u64 flags) argument
927 i915_vma_detach(struct i915_vma *vma) argument
940 try_qad_pin(struct i915_vma *vma, unsigned int flags) argument
1300 __i915_vma_get_pages(struct i915_vma *vma) argument
1347 i915_vma_get_pages(struct i915_vma *vma) argument
1394 __vma_put_pages(struct i915_vma *vma, unsigned int count) argument
1410 i915_vma_put_pages(struct i915_vma *vma) argument
1418 vma_unbind_pages(struct i915_vma *vma) argument
1432 i915_vma_pin_ww(struct i915_vma *vma, struct i915_gem_ww_ctx *ww, u64 size, u64 alignment, u64 flags) argument
1621 __i915_ggtt_pin(struct i915_vma *vma, struct i915_gem_ww_ctx *ww, u32 align, unsigned int flags) argument
1656 i915_ggtt_pin(struct i915_vma *vma, struct i915_gem_ww_ctx *ww, u32 align, unsigned int flags) argument
1688 struct i915_vma *vma; local
1698 __vma_close(struct i915_vma *vma, struct intel_gt *gt) argument
1716 i915_vma_close(struct i915_vma *vma) argument
1733 __i915_vma_remove_closed(struct i915_vma *vma) argument
1738 i915_vma_reopen(struct i915_vma *vma) argument
1748 force_unbind(struct i915_vma *vma) argument
1758 release_references(struct i915_vma *vma, struct intel_gt *gt, bool vm_ddestroy) argument
1812 i915_vma_destroy_locked(struct i915_vma *vma) argument
1821 i915_vma_destroy(struct i915_vma *vma) argument
1840 struct i915_vma *vma, *next; local
1883 __i915_vma_iounmap(struct i915_vma *vma) argument
1897 i915_vma_revoke_mmap(struct i915_vma *vma) argument
1921 __i915_request_await_bind(struct i915_request *rq, struct i915_vma *vma) argument
1926 __i915_vma_move_to_active(struct i915_vma *vma, struct i915_request *rq) argument
1938 _i915_vma_move_to_active(struct i915_vma *vma, struct i915_request *rq, struct dma_fence *fence, unsigned int flags) argument
2013 __i915_vma_evict(struct i915_vma *vma, bool async) argument
2093 __i915_vma_unbind(struct i915_vma *vma) argument
2124 __i915_vma_unbind_async(struct i915_vma *vma) argument
2159 i915_vma_unbind(struct i915_vma *vma) argument
2197 i915_vma_unbind_async(struct i915_vma *vma, bool trylock_vm) argument
2259 i915_vma_unbind_unlocked(struct i915_vma *vma) argument
2270 i915_vma_make_unshrinkable(struct i915_vma *vma) argument
2276 i915_vma_make_shrinkable(struct i915_vma *vma) argument
2281 i915_vma_make_purgeable(struct i915_vma *vma) argument
[all...]
H A Di915_gem_evict.c41 static bool dying_vma(struct i915_vma *vma) argument
43 return !kref_read(&vma->obj->base.refcount);
54 * Not everything in the GGTT is tracked via vma (otherwise we
67 static bool grab_vma(struct i915_vma *vma, struct i915_gem_ww_ctx *ww) argument
73 if (i915_gem_object_get_rcu(vma->obj)) {
74 if (!i915_gem_object_trylock(vma->obj, ww)) {
75 i915_gem_object_put(vma->obj);
80 atomic_and(~I915_VMA_PIN_MASK, &vma->flags);
86 static void ungrab_vma(struct i915_vma *vma) argument
88 if (dying_vma(vma))
96 mark_free(struct drm_mm_scan *scan, struct i915_gem_ww_ctx *ww, struct i915_vma *vma, unsigned int flags, struct list_head *unwind) argument
112 defer_evict(struct i915_vma *vma) argument
157 struct i915_vma *vma, *next; local
337 struct i915_vma *vma, *next; local
478 struct i915_vma *vma, *vn; local
[all...]
/linux-master/drivers/gpu/drm/nouveau/
H A Dnouveau_vmm.c29 nouveau_vma_unmap(struct nouveau_vma *vma) argument
31 if (vma->mem) {
32 nvif_vmm_unmap(&vma->vmm->vmm, vma->addr);
33 vma->mem = NULL;
38 nouveau_vma_map(struct nouveau_vma *vma, struct nouveau_mem *mem) argument
40 struct nvif_vma tmp = { .addr = vma->addr };
41 int ret = nouveau_mem_map(mem, &vma->vmm->vmm, &tmp);
44 vma->mem = mem;
51 struct nouveau_vma *vma; local
64 struct nouveau_vma *vma = *pvma; local
81 struct nouveau_vma *vma; local
[all...]
/linux-master/include/linux/
H A Dhugetlb_inline.h9 static inline bool is_vm_hugetlb_page(struct vm_area_struct *vma) argument
11 return !!(vma->vm_flags & VM_HUGETLB);
16 static inline bool is_vm_hugetlb_page(struct vm_area_struct *vma) argument
H A Duserfaultfd_k.h135 extern long uffd_wp_range(struct vm_area_struct *vma,
149 static inline bool is_mergeable_vm_userfaultfd_ctx(struct vm_area_struct *vma, argument
152 return vma->vm_userfaultfd_ctx.ctx == vm_ctx.ctx;
166 static inline bool uffd_disable_huge_pmd_share(struct vm_area_struct *vma) argument
168 return vma->vm_flags & (VM_UFFD_WP | VM_UFFD_MINOR);
178 static inline bool uffd_disable_fault_around(struct vm_area_struct *vma) argument
180 return vma->vm_flags & (VM_UFFD_WP | VM_UFFD_MINOR);
183 static inline bool userfaultfd_missing(struct vm_area_struct *vma) argument
185 return vma->vm_flags & VM_UFFD_MISSING;
188 static inline bool userfaultfd_wp(struct vm_area_struct *vma) argument
193 userfaultfd_minor(struct vm_area_struct *vma) argument
198 userfaultfd_pte_wp(struct vm_area_struct *vma, pte_t pte) argument
204 userfaultfd_huge_pmd_wp(struct vm_area_struct *vma, pmd_t pmd) argument
210 userfaultfd_armed(struct vm_area_struct *vma) argument
215 vma_can_userfault(struct vm_area_struct *vma, unsigned long vm_flags, bool wp_async) argument
276 uffd_wp_range(struct vm_area_struct *vma, unsigned long start, unsigned long len, bool enable_wp) argument
283 is_mergeable_vm_userfaultfd_ctx(struct vm_area_struct *vma, struct vm_userfaultfd_ctx vm_ctx) argument
289 userfaultfd_missing(struct vm_area_struct *vma) argument
294 userfaultfd_wp(struct vm_area_struct *vma) argument
299 userfaultfd_minor(struct vm_area_struct *vma) argument
304 userfaultfd_pte_wp(struct vm_area_struct *vma, pte_t pte) argument
310 userfaultfd_huge_pmd_wp(struct vm_area_struct *vma, pmd_t pmd) argument
317 userfaultfd_armed(struct vm_area_struct *vma) argument
322 dup_userfaultfd(struct vm_area_struct *vma, struct list_head *l) argument
332 mremap_userfaultfd_prep(struct vm_area_struct *vma, struct vm_userfaultfd_ctx *ctx) argument
344 userfaultfd_remove(struct vm_area_struct *vma, unsigned long start, unsigned long end) argument
351 userfaultfd_unmap_prep(struct vm_area_struct *vma, unsigned long start, unsigned long end, struct list_head *uf) argument
363 uffd_disable_fault_around(struct vm_area_struct *vma) argument
368 userfaultfd_wp_unpopulated(struct vm_area_struct *vma) argument
373 userfaultfd_wp_async(struct vm_area_struct *vma) argument
380 userfaultfd_wp_use_markers(struct vm_area_struct *vma) argument
[all...]
H A Dcacheflush.h21 static inline void flush_icache_pages(struct vm_area_struct *vma, argument
27 #define flush_icache_page(vma, page) flush_icache_pages(vma, page, 1)
/linux-master/arch/sparc/include/asm/
H A Dtlbflush_32.h11 #define flush_tlb_range(vma, start, end) \
12 sparc32_cachetlb_ops->tlb_range(vma, start, end)
13 #define flush_tlb_page(vma, addr) \
14 sparc32_cachetlb_ops->tlb_page(vma, addr)
/linux-master/fs/ocfs2/
H A Dmmap.h5 int ocfs2_mmap(struct file *file, struct vm_area_struct *vma);
/linux-master/arch/powerpc/mm/book3s64/
H A Dradix_hugetlbpage.c10 void radix__flush_hugetlb_page(struct vm_area_struct *vma, unsigned long vmaddr) argument
13 struct hstate *hstate = hstate_file(vma->vm_file);
16 radix__flush_tlb_page_psize(vma->vm_mm, vmaddr, psize);
19 void radix__local_flush_hugetlb_page(struct vm_area_struct *vma, unsigned long vmaddr) argument
22 struct hstate *hstate = hstate_file(vma->vm_file);
25 radix__local_flush_tlb_page_psize(vma->vm_mm, vmaddr, psize);
28 void radix__flush_hugetlb_tlb_range(struct vm_area_struct *vma, unsigned long start, argument
32 struct hstate *hstate = hstate_file(vma->vm_file);
39 radix__flush_tlb_pwc_range_psize(vma->vm_mm, start, end, psize);
41 radix__flush_tlb_range_psize(vma
45 radix__huge_ptep_modify_prot_commit(struct vm_area_struct *vma, unsigned long addr, pte_t *ptep, pte_t old_pte, pte_t pte) argument
[all...]
/linux-master/drivers/pci/
H A Dmmap.c25 struct vm_area_struct *vma,
32 if (vma->vm_pgoff + vma_pages(vma) > size)
36 vma->vm_page_prot = pgprot_writecombine(vma->vm_page_prot);
38 vma->vm_page_prot = pgprot_device(vma->vm_page_prot);
41 ret = pci_iobar_pfn(pdev, bar, vma);
45 vma->vm_pgoff += (pci_resource_start(pdev, bar) >> PAGE_SHIFT);
47 vma
24 pci_mmap_resource_range(struct pci_dev *pdev, int bar, struct vm_area_struct *vma, enum pci_mmap_state mmap_state, int write_combine) argument
59 pci_mmap_fits(struct pci_dev *pdev, int resno, struct vm_area_struct *vma, enum pci_mmap_api mmap_api) argument
[all...]
/linux-master/arch/arc/kernel/
H A Darc_hostlink.c18 static int arc_hl_mmap(struct file *fp, struct vm_area_struct *vma) argument
20 vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot);
22 if (io_remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff,
23 vma->vm_end - vma->vm_start,
24 vma->vm_page_prot)) {
/linux-master/drivers/gpu/drm/xe/display/
H A Dxe_dsb_buffer.c15 return xe_bo_ggtt_addr(dsb_buf->vma->bo);
20 iosys_map_wr(&dsb_buf->vma->bo->vmap, idx * 4, u32, val);
25 return iosys_map_rd(&dsb_buf->vma->bo->vmap, idx * 4, u32);
32 iosys_map_memset(&dsb_buf->vma->bo->vmap, idx * 4, val, size);
39 struct i915_vma *vma; local
41 vma = kzalloc(sizeof(*vma), GFP_KERNEL);
42 if (!vma)
51 kfree(vma);
55 vma
[all...]
/linux-master/drivers/gpu/drm/i915/gem/selftests/
H A Digt_gem_utils.h25 igt_emit_store_dw(struct i915_vma *vma,
31 struct i915_vma *vma, u64 offset,
35 igt_vma_move_to_active_unlocked(struct i915_vma *vma, struct i915_request *rq, argument
40 i915_vma_lock(vma);
41 err = i915_vma_move_to_active(vma, rq, flags);
42 i915_vma_unlock(vma);
/linux-master/arch/sh/include/asm/
H A Dtlbflush.h10 * - flush_tlb_page(vma, vmaddr) flushes one page
11 * - flush_tlb_range(vma, start, end) flushes a range of pages
16 extern void local_flush_tlb_range(struct vm_area_struct *vma,
19 extern void local_flush_tlb_page(struct vm_area_struct *vma,
31 extern void flush_tlb_range(struct vm_area_struct *vma, unsigned long start,
33 extern void flush_tlb_page(struct vm_area_struct *vma, unsigned long page);
41 #define flush_tlb_page(vma, page) local_flush_tlb_page(vma, page)
44 #define flush_tlb_range(vma, start, end) \
45 local_flush_tlb_range(vma, star
[all...]
/linux-master/arch/csky/include/asm/
H A Dtlbflush.h11 * - flush_tlb_page(vma, vmaddr) flushes one page
12 * - flush_tlb_range(vma, start, end) flushes a range of pages
17 extern void flush_tlb_page(struct vm_area_struct *vma, unsigned long page);
18 extern void flush_tlb_range(struct vm_area_struct *vma, unsigned long start,
/linux-master/drivers/gpu/drm/msm/
H A Dmsm_gem_vma.c41 /* Actually unmap memory for the vma */
42 void msm_gem_vma_purge(struct msm_gem_vma *vma) argument
44 struct msm_gem_address_space *aspace = vma->aspace;
45 unsigned size = vma->node.size;
48 if (!vma->mapped)
51 aspace->mmu->funcs->unmap(aspace->mmu, vma->iova, size);
53 vma->mapped = false;
56 /* Map and pin vma: */
58 msm_gem_vma_map(struct msm_gem_vma *vma, int prot, argument
61 struct msm_gem_address_space *aspace = vma
94 msm_gem_vma_close(struct msm_gem_vma *vma) argument
112 struct msm_gem_vma *vma; local
124 msm_gem_vma_init(struct msm_gem_vma *vma, int size, u64 range_start, u64 range_end) argument
[all...]
/linux-master/drivers/gpu/drm/xe/compat-i915-headers/gem/
H A Di915_gem_mman.h12 static inline int i915_gem_fb_mmap(struct xe_bo *bo, struct vm_area_struct *vma) argument
14 return drm_gem_prime_mmap(&bo->ttm.base, vma);
/linux-master/arch/nios2/include/asm/
H A Dtlbflush.h16 * - flush_tlb_range(vma, start, end) flushes a range of pages
17 * - flush_tlb_page(vma, address) flushes a page
21 * - reload_tlb_page(vma, address, pte) flushes the TLB for address like
26 extern void flush_tlb_range(struct vm_area_struct *vma, unsigned long start,
30 static inline void flush_tlb_page(struct vm_area_struct *vma, argument
33 flush_tlb_range(vma, address, address + PAGE_SIZE);
41 extern void reload_tlb_page(struct vm_area_struct *vma, unsigned long addr,
/linux-master/drivers/gpu/drm/i915/gt/
H A Dgen7_renderclear.h13 struct i915_vma * const vma);
/linux-master/drivers/gpu/drm/i915/display/
H A Dintel_fb_pin.c30 struct i915_vma *vma; local
74 vma = i915_vma_instance(obj, vm, view);
75 if (IS_ERR(vma)) {
76 ret = PTR_ERR(vma);
80 if (i915_vma_misplaced(vma, 0, alignment, 0)) {
81 ret = i915_vma_unbind(vma);
86 ret = i915_vma_pin_ww(vma, &ww, 0, alignment, PIN_GLOBAL);
91 vma = ERR_PTR(ret);
95 vma->display_alignment = max(vma
118 struct i915_vma *vma; local
231 intel_unpin_fb_vma(struct i915_vma *vma, unsigned long flags) argument
244 struct i915_vma *vma; local
296 struct i915_vma *vma; local
[all...]
/linux-master/arch/arc/include/asm/
H A Dtlbflush.h13 void local_flush_tlb_page(struct vm_area_struct *vma, unsigned long page);
15 void local_flush_tlb_range(struct vm_area_struct *vma,
18 void local_flush_pmd_tlb_range(struct vm_area_struct *vma, unsigned long start,
23 #define flush_tlb_range(vma, s, e) local_flush_tlb_range(vma, s, e)
24 #define flush_tlb_page(vma, page) local_flush_tlb_page(vma, page)
29 #define flush_pmd_tlb_range(vma, s, e) local_flush_pmd_tlb_range(vma, s, e)
32 extern void flush_tlb_range(struct vm_area_struct *vma, unsigne
[all...]
/linux-master/mm/
H A Dmsync.c36 struct vm_area_struct *vma; local
63 vma = find_vma(mm, start);
70 if (!vma)
72 /* Here start < vma->vm_end. */
73 if (start < vma->vm_start) {
76 start = vma->vm_start;
81 /* Here vma->vm_start <= start < vma->vm_end. */
83 (vma->vm_flags & VM_LOCKED)) {
87 file = vma
[all...]
/linux-master/tools/testing/selftests/bpf/progs/
H A Dbpf_iter_task_vmas.c28 struct vm_area_struct *vma = ctx->vma; local
34 if (task == (void *)0 || vma == (void *)0)
37 file = vma->vm_file;
43 perm_str[0] = (vma->vm_flags & VM_READ) ? 'r' : '-';
44 perm_str[1] = (vma->vm_flags & VM_WRITE) ? 'w' : '-';
45 perm_str[2] = (vma->vm_flags & VM_EXEC) ? 'x' : '-';
46 perm_str[3] = (vma->vm_flags & VM_MAYSHARE) ? 's' : 'p';
47 BPF_SEQ_PRINTF(seq, "%08llx-%08llx %s ", vma->vm_start, vma
[all...]

Completed in 813 milliseconds

1234567891011>>