Searched refs:vma (Results 176 - 200 of 1003) sorted by relevance

1234567891011>>

/linux-master/arch/m68k/mm/
H A Dfault.c74 struct vm_area_struct * vma; local
95 vma = find_vma(mm, address);
96 if (!vma)
98 if (vma->vm_start <= address)
100 if (!(vma->vm_flags & VM_GROWSDOWN))
110 vma = expand_stack(mm, address);
111 if (!vma)
124 if (!(vma->vm_flags & VM_WRITE))
131 if (unlikely(!vma_is_accessible(vma)))
141 fault = handle_mm_fault(vma, addres
[all...]
/linux-master/drivers/video/fbdev/core/
H A Dfb_io_fops.c141 int fb_io_mmap(struct fb_info *info, struct vm_area_struct *vma) argument
154 if (vma->vm_pgoff >= mmio_pgoff) {
158 vma->vm_pgoff -= mmio_pgoff;
163 vma->vm_page_prot = vm_get_page_prot(vma->vm_flags);
164 vma->vm_page_prot = pgprot_framebuffer(vma->vm_page_prot, vma->vm_start,
165 vma->vm_end, start);
167 return vm_iomap_memory(vma, star
[all...]
/linux-master/drivers/gpu/drm/xe/
H A Dxe_trace.h387 TP_PROTO(struct xe_vma *vma),
388 TP_ARGS(vma),
391 __field(struct xe_vma *, vma)
399 __entry->vma = vma;
400 __entry->asid = xe_vma_vm(vma)->usm.asid;
401 __entry->start = xe_vma_start(vma);
402 __entry->end = xe_vma_end(vma) - 1;
403 __entry->ptr = xe_vma_userptr(vma);
406 TP_printk("vma
[all...]
H A Dxe_gt_pagefault.c69 static bool vma_is_valid(struct xe_tile *tile, struct xe_vma *vma) argument
71 return BIT(tile->id) & vma->tile_present &&
72 !(BIT(tile->id) & vma->tile_invalidated);
75 static bool vma_matches(struct xe_vma *vma, u64 page_addr) argument
77 if (page_addr > xe_vma_end(vma) - 1 ||
78 page_addr + SZ_4K - 1 < xe_vma_start(vma))
86 struct xe_vma *vma = NULL; local
90 vma = vm->usm.last_fault_vma;
92 if (!vma)
93 vma
98 xe_pf_begin(struct drm_exec *exec, struct xe_vma *vma, bool atomic, unsigned int id) argument
135 struct xe_vma *vma = NULL; local
513 struct xe_vma *vma; local
[all...]
/linux-master/arch/sh/kernel/
H A Dsys_sh.c60 struct vm_area_struct *vma; local
73 vma = find_vma (current->mm, addr);
74 if (vma == NULL || addr < vma->vm_start || addr + len > vma->vm_end) {
/linux-master/drivers/gpu/drm/msm/
H A Dmsm_gem.c289 struct vm_area_struct *vma = vmf->vma; local
290 struct drm_gem_object *obj = vma->vm_private_data;
321 pgoff = (vmf->address - vma->vm_start) >> PAGE_SHIFT;
328 ret = vmf_insert_pfn(vma, vmf->address, pfn);
369 struct msm_gem_vma *vma; local
373 vma = msm_gem_vma_new(aspace);
374 if (!vma)
377 list_add_tail(&vma->list, &msm_obj->vmas);
379 return vma;
386 struct msm_gem_vma *vma; local
398 del_vma(struct msm_gem_vma *vma) argument
417 struct msm_gem_vma *vma; local
435 struct msm_gem_vma *vma, *tmp; local
448 struct msm_gem_vma *vma; local
475 msm_gem_pin_vma_locked(struct drm_gem_object *obj, struct msm_gem_vma *vma) argument
538 struct msm_gem_vma *vma; local
587 struct msm_gem_vma *vma; local
605 struct msm_gem_vma *vma = lookup_vma(obj, aspace); local
633 struct msm_gem_vma *vma; local
655 struct msm_gem_vma *vma; local
927 struct msm_gem_vma *vma; local
1065 msm_gem_object_mmap(struct drm_gem_object *obj, struct vm_area_struct *vma) argument
1199 struct msm_gem_vma *vma; local
[all...]
/linux-master/mm/
H A Dmigrate_device.c44 if (!vma_is_anonymous(walk->vma))
63 struct vm_area_struct *vma = walk->vma; local
64 struct mm_struct *mm = vma->vm_mm;
85 split_huge_pmd(vma, pmdp, addr);
118 if (vma_is_anonymous(vma)) {
153 page = vm_normal_page(migrate->vma, addr, pte);
199 flush_cache_page(vma, addr, pte_pfn(pte));
203 pte = ptep_clear_flush(vma, addr, ptep);
257 folio_remove_rmap_pte(folio, page, vma);
571 struct vm_area_struct *vma = migrate->vma; local
[all...]
/linux-master/arch/arc/mm/
H A Dtlb.c190 * (Android Binder ended up calling this for vma->mm != tsk->mm,
208 void local_flush_tlb_range(struct vm_area_struct *vma, unsigned long start, argument
222 local_flush_tlb_mm(vma->vm_mm);
235 if (asid_mm(vma->vm_mm, cpu) != MM_CTXT_NO_ASID) {
237 tlb_entry_erase(start | hw_pid(vma->vm_mm, cpu));
278 void local_flush_tlb_page(struct vm_area_struct *vma, unsigned long page) argument
288 if (asid_mm(vma->vm_mm, cpu) != MM_CTXT_NO_ASID) {
289 tlb_entry_erase((page & PAGE_MASK) | hw_pid(vma->vm_mm, cpu));
344 void flush_tlb_page(struct vm_area_struct *vma, unsigned long uaddr) argument
347 .ta_vma = vma,
354 flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) argument
367 flush_pmd_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) argument
394 create_tlb(struct vm_area_struct *vma, unsigned long vaddr, pte_t *ptep) argument
472 update_mmu_cache_range(struct vm_fault *vmf, struct vm_area_struct *vma, unsigned long vaddr_unaligned, pte_t *ptep, unsigned int nr) argument
530 update_mmu_cache_pmd(struct vm_area_struct *vma, unsigned long addr, pmd_t *pmd) argument
537 local_flush_pmd_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) argument
[all...]
/linux-master/arch/riscv/mm/
H A Dfault.c193 static inline bool access_error(unsigned long cause, struct vm_area_struct *vma) argument
197 if (!(vma->vm_flags & VM_EXEC)) {
203 if (!(vma->vm_flags & (VM_READ | VM_WRITE))) {
208 if (!(vma->vm_flags & VM_WRITE)) {
225 struct vm_area_struct *vma; local
289 vma = lock_vma_under_rcu(mm, addr);
290 if (!vma)
293 if (unlikely(access_error(cause, vma))) {
294 vma_end_read(vma);
298 fault = handle_mm_fault(vma, add
[all...]
H A Dpgtable.c8 int ptep_set_access_flags(struct vm_area_struct *vma, argument
13 __set_pte_at(vma->vm_mm, ptep, entry);
21 int ptep_test_and_clear_young(struct vm_area_struct *vma, argument
126 pmd_t pmdp_collapse_flush(struct vm_area_struct *vma, argument
129 pmd_t pmd = pmdp_huge_get_and_clear(vma->vm_mm, address, pmdp);
136 * valid leaf PTE at the level 1 page table. Since the sfence.vma
141 flush_tlb_mm(vma->vm_mm);
/linux-master/kernel/events/
H A Duprobes.c105 * We keep the vma's vm_start rather than a pointer to the vma
107 * the vma go away, and we must handle that reasonably gracefully.
113 * valid_vma: Verify if the specified vma is an executable vma
118 * executable vma.
120 static bool valid_vma(struct vm_area_struct *vma, bool is_register) argument
127 return vma->vm_file && (vma->vm_flags & flags) == VM_MAYEXEC;
130 static unsigned long offset_to_vaddr(struct vm_area_struct *vma, loff_ argument
135 vaddr_to_offset(struct vm_area_struct *vma, unsigned long vaddr) argument
153 __replace_page(struct vm_area_struct *vma, unsigned long addr, struct page *old_page, struct page *new_page) argument
337 valid_ref_ctr_vma(struct uprobe *uprobe, struct vm_area_struct *vma) argument
464 struct vm_area_struct *vma; local
890 install_breakpoint(struct uprobe *uprobe, struct mm_struct *mm, struct vm_area_struct *vma, unsigned long vaddr) argument
962 struct vm_area_struct *vma; local
1045 struct vm_area_struct *vma; local
1235 struct vm_area_struct *vma; local
1288 build_probe_list(struct inode *inode, struct vm_area_struct *vma, unsigned long start, unsigned long end, struct list_head *head) argument
1323 delayed_ref_ctr_inc(struct vm_area_struct *vma) argument
1357 uprobe_mmap(struct vm_area_struct *vma) argument
1399 vma_has_uprobes(struct vm_area_struct *vma, unsigned long start, unsigned long end) argument
1420 uprobe_munmap(struct vm_area_struct *vma, unsigned long start, unsigned long end) argument
1439 struct vm_area_struct *vma; local
1988 struct vm_area_struct *vma; local
2043 struct vm_area_struct *vma; local
[all...]
/linux-master/include/linux/
H A Dpgtable.h283 extern int ptep_set_access_flags(struct vm_area_struct *vma,
290 extern int pmdp_set_access_flags(struct vm_area_struct *vma,
293 extern int pudp_set_access_flags(struct vm_area_struct *vma,
297 static inline int pmdp_set_access_flags(struct vm_area_struct *vma, argument
304 static inline int pudp_set_access_flags(struct vm_area_struct *vma, argument
350 static inline int ptep_test_and_clear_young(struct vm_area_struct *vma, argument
359 set_pte_at(vma->vm_mm, address, ptep, pte_mkold(pte));
366 static inline int pmdp_test_and_clear_young(struct vm_area_struct *vma, argument
375 set_pmd_at(vma->vm_mm, address, pmdp, pmd_mkold(pmd));
379 static inline int pmdp_test_and_clear_young(struct vm_area_struct *vma, argument
403 pmdp_clear_flush_young(struct vm_area_struct *vma, unsigned long address, pmd_t *pmdp) argument
437 arch_check_zapped_pte(struct vm_area_struct *vma, pte_t pte) argument
444 arch_check_zapped_pmd(struct vm_area_struct *vma, pmd_t pmd) argument
582 pmdp_huge_get_and_clear_full(struct vm_area_struct *vma, unsigned long address, pmd_t *pmdp, int full) argument
591 pudp_huge_get_and_clear_full(struct vm_area_struct *vma, unsigned long address, pud_t *pudp, int full) argument
689 update_mmu_tlb(struct vm_area_struct *vma, unsigned long address, pte_t *ptep) argument
727 pte_mkwrite(pte_t pte, struct vm_area_struct *vma) argument
734 pmd_mkwrite(pmd_t pmd, struct vm_area_struct *vma) argument
837 pmdp_collapse_flush(struct vm_area_struct *vma, unsigned long address, pmd_t *pmdp) argument
867 generic_pmdp_establish(struct vm_area_struct *vma, unsigned long address, pmd_t *pmdp, pmd_t pmd) argument
1022 arch_do_swap_page(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long addr, pte_t pte, pte_t oldpte) argument
1040 arch_unmap_one(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long addr, pte_t orig_pte) argument
1190 __ptep_modify_prot_start(struct vm_area_struct *vma, unsigned long addr, pte_t *ptep) argument
1202 __ptep_modify_prot_commit(struct vm_area_struct *vma, unsigned long addr, pte_t *ptep, pte_t pte) argument
1228 ptep_modify_prot_start(struct vm_area_struct *vma, unsigned long addr, pte_t *ptep) argument
1239 ptep_modify_prot_commit(struct vm_area_struct *vma, unsigned long addr, pte_t *ptep, pte_t old_pte, pte_t pte) argument
1406 track_pfn_remap(struct vm_area_struct *vma, pgprot_t *prot, unsigned long pfn, unsigned long addr, unsigned long size) argument
1417 track_pfn_insert(struct vm_area_struct *vma, pgprot_t *prot, pfn_t pfn) argument
1426 track_pfn_copy(struct vm_area_struct *vma) argument
1436 untrack_pfn(struct vm_area_struct *vma, unsigned long pfn, unsigned long size, bool mm_wr_locked) argument
1446 untrack_pfn_clear(struct vm_area_struct *vma) argument
[all...]
H A Drmap.h24 * directly to a vma: instead it points to an anon_vma, on whose list
27 * After unlinking the last vma on the list, we must garbage collect
29 * pointing to this anon_vma once its vma list is empty.
36 * guarantee that the vma of page tables will exist for
83 struct vm_area_struct *vma; member in struct:anon_vma_chain
159 static inline int anon_vma_prepare(struct vm_area_struct *vma) argument
161 if (likely(vma->anon_vma))
164 return __anon_vma_prepare(vma);
167 static inline void anon_vma_merge(struct vm_area_struct *vma, argument
170 VM_BUG_ON_VMA(vma
264 hugetlb_try_dup_anon_rmap(struct folio *folio, struct vm_area_struct *vma) argument
644 struct vm_area_struct *vma; member in struct:page_vma_mapped_walk
[all...]
H A Dksm.h19 int ksm_madvise(struct vm_area_struct *vma, unsigned long start,
22 void ksm_add_vma(struct vm_area_struct *vma);
70 * no problem, it will be assigned to this vma's anon_vma; but thereafter,
76 * We'd like to make this conditional on vma->vm_flags & VM_MERGEABLE,
77 * but what if the vma was unmerged while the page was swapped out?
80 struct vm_area_struct *vma, unsigned long addr);
96 static inline void ksm_add_vma(struct vm_area_struct *vma) argument
126 static inline int ksm_madvise(struct vm_area_struct *vma, unsigned long start, argument
133 struct vm_area_struct *vma, unsigned long addr)
132 ksm_might_need_to_copy(struct folio *folio, struct vm_area_struct *vma, unsigned long addr) argument
/linux-master/drivers/gpu/drm/i915/gem/
H A Di915_gem_execbuffer.c39 struct i915_vma *vma; member in struct:eb_vma
42 /** This vma's place in the execbuf reservation list */
252 struct eb_vma *vma; member in struct:i915_execbuffer
262 /** identity of the batch obj/vma */
275 /** list of vma not yet bound during reservation phase */
278 /** list of vma that have execobj.relocation_count */
382 const struct i915_vma *vma,
385 const u64 start = i915_vma_offset(vma);
386 const u64 size = i915_vma_size(vma);
407 !i915_vma_is_map_and_fenceable(vma))
381 eb_vma_misplaced(const struct drm_i915_gem_exec_object2 *entry, const struct i915_vma *vma, unsigned int flags) argument
444 struct i915_vma *vma = ev->vma; local
502 eb_validate_vma(struct i915_execbuffer *eb, struct drm_i915_gem_exec_object2 *entry, struct i915_vma *vma) argument
563 eb_add_vma(struct i915_execbuffer *eb, unsigned int *current_batch, unsigned int i, struct i915_vma *vma) argument
662 struct i915_vma *vma = ev->vma; local
840 __eb_add_lut(struct i915_execbuffer *eb, u32 handle, struct i915_vma *vma) argument
896 struct i915_vma *vma; local
951 struct i915_vma *vma; local
1004 struct i915_vma *vma = ev->vma; local
1028 struct i915_vma *vma = ev->vma; local
1092 struct i915_vma *vma = ev->vma; local
1283 struct i915_vma *vma = ERR_PTR(-ENODEV); local
1351 reloc_vaddr(struct i915_vma *vma, struct i915_execbuffer *eb, unsigned long page) argument
1393 relocate_entry(struct i915_vma *vma, const struct drm_i915_gem_relocation_entry *reloc, struct i915_execbuffer *eb, const struct i915_vma *target) argument
1473 struct i915_vma *vma = target->vma; local
2012 struct i915_vma *vma = ev->vma; local
2104 struct i915_vma *vma = ev->vma; local
2253 struct i915_vma *vma; local
2267 eb_dispatch_secure(struct i915_execbuffer *eb, struct i915_vma *vma) argument
[all...]
/linux-master/arch/powerpc/mm/book3s64/
H A Dpgtable.c48 int pmdp_set_access_flags(struct vm_area_struct *vma, unsigned long address, argument
54 assert_spin_locked(pmd_lockptr(vma->vm_mm, pmdp));
62 __ptep_set_access_flags(vma, pmdp_ptep(pmdp),
68 int pudp_set_access_flags(struct vm_area_struct *vma, unsigned long address, argument
74 assert_spin_locked(pud_lockptr(vma->vm_mm, pudp));
82 __ptep_set_access_flags(vma, pudp_ptep(pudp),
89 int pmdp_test_and_clear_young(struct vm_area_struct *vma, argument
92 return __pmdp_test_and_clear_young(vma->vm_mm, address, pmdp);
95 int pudp_test_and_clear_young(struct vm_area_struct *vma, argument
98 return __pudp_test_and_clear_young(vma
168 pmdp_invalidate(struct vm_area_struct *vma, unsigned long address, pmd_t *pmdp) argument
178 pmdp_huge_get_and_clear_full(struct vm_area_struct *vma, unsigned long addr, pmd_t *pmdp, int full) argument
196 pudp_huge_get_and_clear_full(struct vm_area_struct *vma, unsigned long addr, pud_t *pudp, int full) argument
520 ptep_modify_prot_start(struct vm_area_struct *vma, unsigned long addr, pte_t *ptep) argument
536 ptep_modify_prot_commit(struct vm_area_struct *vma, unsigned long addr, pte_t *ptep, pte_t old_pte, pte_t pte) argument
558 pmd_move_must_withdraw(struct spinlock *new_pmd_ptl, struct spinlock *old_pmd_ptl, struct vm_area_struct *vma) argument
[all...]
/linux-master/fs/coda/
H A Dfile.c123 coda_vm_open(struct vm_area_struct *vma) argument
126 container_of(vma->vm_ops, struct coda_vm_ops, vm_ops);
131 cvm_ops->host_vm_ops->open(vma);
135 coda_vm_close(struct vm_area_struct *vma) argument
138 container_of(vma->vm_ops, struct coda_vm_ops, vm_ops);
141 cvm_ops->host_vm_ops->close(vma);
144 vma->vm_ops = cvm_ops->host_vm_ops;
151 coda_file_mmap(struct file *coda_file, struct vm_area_struct *vma) argument
166 if (WARN_ON(coda_file != vma->vm_file))
169 count = vma
[all...]
/linux-master/arch/powerpc/platforms/book3s/
H A Dvas-api.c397 struct vm_area_struct *vma = vmf->vma; local
398 struct file *fp = vma->vm_file;
423 if (txwin->task_ref.vma != vmf->vma) {
438 fault = vmf_insert_pfn(vma, vma->vm_start,
471 static int coproc_mmap(struct file *fp, struct vm_area_struct *vma) argument
482 if ((vma->vm_end - vma
[all...]
/linux-master/arch/powerpc/mm/
H A Dfault.c88 struct vm_area_struct *vma)
107 * 6. T1 : reaches here, sees vma_pkey(vma)=5, when we really
110 pkey = vma_pkey(vma);
231 struct vm_area_struct *vma)
238 if (!arch_vma_access_permitted(vma, is_write, is_exec, 0))
244 static bool access_error(bool is_write, bool is_exec, struct vm_area_struct *vma) argument
257 return !(vma->vm_flags & VM_EXEC) &&
259 !(vma->vm_flags & (VM_READ | VM_WRITE)));
263 if (unlikely(!(vma->vm_flags & VM_WRITE)))
274 if (unlikely(!vma_is_accessible(vma)))
87 bad_access_pkey(struct pt_regs *regs, unsigned long address, struct vm_area_struct *vma) argument
230 access_pkey_error(bool is_write, bool is_exec, bool is_pkey, struct vm_area_struct *vma) argument
409 struct vm_area_struct * vma; local
[all...]
H A Dcopro_fault.c26 struct vm_area_struct *vma; local
36 vma = lock_mm_and_find_vma(mm, ea, NULL);
37 if (!vma)
43 if (!(vma->vm_flags & VM_WRITE))
46 if (!(vma->vm_flags & (VM_READ | VM_EXEC)))
59 *flt = handle_mm_fault(vma, ea, is_write ? FAULT_FLAG_WRITE : 0, NULL);
/linux-master/arch/x86/kernel/cpu/sgx/
H A Dvirt.c34 struct vm_area_struct *vma, unsigned long addr)
43 index = vma->vm_pgoff + PFN_DOWN(addr - vma->vm_start);
59 ret = vmf_insert_pfn(vma, addr, pfn);
76 struct vm_area_struct *vma = vmf->vma; local
77 struct sgx_vepc *vepc = vma->vm_private_data;
81 ret = __sgx_vepc_fault(vepc, vma, vmf->address);
88 mmap_read_unlock(vma->vm_mm);
99 static int sgx_vepc_mmap(struct file *file, struct vm_area_struct *vma) argument
33 __sgx_vepc_fault(struct sgx_vepc *vepc, struct vm_area_struct *vma, unsigned long addr) argument
[all...]
/linux-master/drivers/gpu/drm/exynos/
H A Dexynos_drm_gem.c23 static int exynos_drm_gem_mmap(struct drm_gem_object *obj, struct vm_area_struct *vma);
271 struct vm_area_struct *vma)
277 vm_flags_clear(vma, VM_PFNMAP);
278 vma->vm_pgoff = 0;
280 vm_size = vma->vm_end - vma->vm_start;
286 ret = dma_mmap_attrs(to_dma_dev(drm_dev), vma, exynos_gem->cookie,
363 static int exynos_drm_gem_mmap(struct drm_gem_object *obj, struct vm_area_struct *vma) argument
369 return dma_buf_mmap(obj->dma_buf, vma, 0);
371 vm_flags_set(vma, VM_I
270 exynos_drm_gem_mmap_buffer(struct exynos_drm_gem *exynos_gem, struct vm_area_struct *vma) argument
[all...]
/linux-master/arch/xtensa/include/asm/
H A Dtlbflush.h29 * - flush_tlb_page(vma, page) flushes a single page
30 * - flush_tlb_range(vma, vmaddr, end) flushes a range of pages
35 void local_flush_tlb_page(struct vm_area_struct *vma,
37 void local_flush_tlb_range(struct vm_area_struct *vma,
54 #define flush_tlb_page(vma, page) local_flush_tlb_page(vma, page)
55 #define flush_tlb_range(vma, vmaddr, end) local_flush_tlb_range(vma, vmaddr, \
H A Dcacheflush.h136 void local_flush_cache_range(struct vm_area_struct *vma,
138 void local_flush_cache_page(struct vm_area_struct *vma,
167 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \
174 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \
/linux-master/arch/parisc/kernel/
H A Dvdso.c30 struct vm_area_struct *vma)
32 current->mm->context.vdso_base = vma->vm_start;
49 * This is called from binfmt_elf, we create the special vma for the
59 struct vm_area_struct *vma; local
83 vma = _install_special_mapping(mm, vdso_text_start, vdso_text_len,
87 if (IS_ERR(vma)) {
89 rc = PTR_ERR(vma);
29 vdso_mremap(const struct vm_special_mapping *sm, struct vm_area_struct *vma) argument

Completed in 226 milliseconds

1234567891011>>