/linux-master/drivers/gpu/drm/i915/gem/selftests/ |
H A D | huge_pages.c | 362 static int igt_check_page_sizes(struct i915_vma *vma) argument 364 struct drm_i915_private *i915 = vma->vm->i915; 366 struct drm_i915_gem_object *obj = vma->obj; 370 err = i915_vma_sync(vma); 374 if (!HAS_PAGE_SIZES(i915, vma->page_sizes.sg)) { 376 vma->page_sizes.sg & ~supported, supported); 380 if (!HAS_PAGE_SIZES(i915, vma->resource->page_sizes_gtt)) { 382 vma->resource->page_sizes_gtt & ~supported, supported); 386 if (vma->page_sizes.phys != obj->mm.page_sizes.phys) { 387 pr_err("vma 428 struct i915_vma *vma; local 505 struct i915_vma *vma; local 601 struct i915_vma *vma; local 745 struct i915_vma *vma; local 923 struct i915_vma *vma; local 1055 gpu_write(struct intel_context *ce, struct i915_vma *vma, u32 dw, u32 val) argument 1152 struct i915_vma *vma; local 1625 struct i915_vma *vma; local 1766 struct i915_vma *vma; local 1838 struct i915_vma *vma; local [all...] |
H A D | i915_gem_coherency.c | 85 struct i915_vma *vma; local 95 vma = i915_gem_object_ggtt_pin(ctx->obj, NULL, 0, 0, PIN_MAPPABLE); 96 if (IS_ERR(vma)) 97 return PTR_ERR(vma); 99 wakeref = intel_gt_pm_get(vma->vm->gt); 101 map = i915_vma_pin_iomap(vma); 102 i915_vma_unpin(vma); 109 i915_vma_unpin_iomap(vma); 112 intel_gt_pm_put(vma->vm->gt, wakeref); 119 struct i915_vma *vma; local 197 struct i915_vma *vma; local [all...] |
/linux-master/scripts/ |
H A D | makelst | 32 $3 -r --source --adjust-vma=${t6:-0} $1
|
/linux-master/include/asm-generic/ |
H A D | mm_hooks.h | 25 static inline bool arch_vma_access_permitted(struct vm_area_struct *vma, argument
|
/linux-master/arch/sh/kernel/vsyscall/ |
H A D | vsyscall.c | 87 const char *arch_vma_name(struct vm_area_struct *vma) argument 89 if (vma->vm_mm && vma->vm_start == (long)vma->vm_mm->context.vdso)
|
/linux-master/tools/testing/selftests/bpf/progs/ |
H A D | find_vma.c | 23 static long check_vma(struct task_struct *task, struct vm_area_struct *vma, argument 26 if (vma->vm_file) 28 vma->vm_file->f_path.dentry->d_iname); 31 if (vma->vm_flags & VM_EXEC)
|
H A D | find_vma_fail2.c | 12 static long write_task(struct task_struct *task, struct vm_area_struct *vma, argument
|
/linux-master/drivers/gpu/drm/i915/display/ |
H A D | intel_fb_pin.h | 23 void intel_unpin_fb_vma(struct i915_vma *vma, unsigned long flags);
|
H A D | intel_fbdev_fb.h | 19 struct drm_i915_gem_object *obj, struct i915_vma *vma);
|
/linux-master/arch/sparc/include/asm/ |
H A D | hugetlb.h | 26 static inline pte_t huge_ptep_clear_flush(struct vm_area_struct *vma, argument 41 static inline int huge_ptep_set_access_flags(struct vm_area_struct *vma, argument 47 __set_huge_pte_at(vma->vm_mm, addr, ptep, pte); 48 flush_tlb_page(vma, addr);
|
/linux-master/arch/hexagon/kernel/ |
H A D | vdso.c | 84 const char *arch_vma_name(struct vm_area_struct *vma) argument 86 if (vma->vm_mm && vma->vm_start == (long)vma->vm_mm->context.vdso)
|
/linux-master/drivers/gpu/drm/xe/display/ |
H A D | intel_fbdev_fb.h | 19 struct drm_i915_gem_object *obj, struct i915_vma *vma);
|
/linux-master/arch/arc/include/asm/ |
H A D | cacheflush.h | 50 #define flush_cache_page(vma, u_vaddr, pfn) /* PF handling/COW-break */ 59 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ 62 if (vma->vm_flags & VM_EXEC) \ 66 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \
|
/linux-master/include/linux/ |
H A D | buildid.h | 10 int build_id_parse(struct vm_area_struct *vma, unsigned char *build_id,
|
H A D | pgtable.h | 283 extern int ptep_set_access_flags(struct vm_area_struct *vma, 290 extern int pmdp_set_access_flags(struct vm_area_struct *vma, 293 extern int pudp_set_access_flags(struct vm_area_struct *vma, 297 static inline int pmdp_set_access_flags(struct vm_area_struct *vma, argument 304 static inline int pudp_set_access_flags(struct vm_area_struct *vma, argument 350 static inline int ptep_test_and_clear_young(struct vm_area_struct *vma, argument 359 set_pte_at(vma->vm_mm, address, ptep, pte_mkold(pte)); 366 static inline int pmdp_test_and_clear_young(struct vm_area_struct *vma, argument 375 set_pmd_at(vma->vm_mm, address, pmdp, pmd_mkold(pmd)); 379 static inline int pmdp_test_and_clear_young(struct vm_area_struct *vma, argument 403 pmdp_clear_flush_young(struct vm_area_struct *vma, unsigned long address, pmd_t *pmdp) argument 437 arch_check_zapped_pte(struct vm_area_struct *vma, pte_t pte) argument 444 arch_check_zapped_pmd(struct vm_area_struct *vma, pmd_t pmd) argument 481 clear_young_dirty_ptes(struct vm_area_struct *vma, unsigned long addr, pte_t *ptep, unsigned int nr, cydp_t flags) argument 626 pmdp_huge_get_and_clear_full(struct vm_area_struct *vma, unsigned long address, pmd_t *pmdp, int full) argument 635 pudp_huge_get_and_clear_full(struct vm_area_struct *vma, unsigned long address, pud_t *pudp, int full) argument 733 update_mmu_tlb(struct vm_area_struct *vma, unsigned long address, pte_t *ptep) argument 800 pte_mkwrite(pte_t pte, struct vm_area_struct *vma) argument 807 pmd_mkwrite(pmd_t pmd, struct vm_area_struct *vma) argument 910 pmdp_collapse_flush(struct vm_area_struct *vma, unsigned long address, pmd_t *pmdp) argument 940 generic_pmdp_establish(struct vm_area_struct *vma, unsigned long address, pmd_t *pmdp, pmd_t pmd) argument 1095 arch_do_swap_page(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long addr, pte_t pte, pte_t oldpte) argument 1113 arch_unmap_one(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long addr, pte_t orig_pte) argument 1263 __ptep_modify_prot_start(struct vm_area_struct *vma, unsigned long addr, pte_t *ptep) argument 1275 __ptep_modify_prot_commit(struct vm_area_struct *vma, unsigned long addr, pte_t *ptep, pte_t pte) argument 1301 ptep_modify_prot_start(struct vm_area_struct *vma, unsigned long addr, pte_t *ptep) argument 1312 ptep_modify_prot_commit(struct vm_area_struct *vma, unsigned long addr, pte_t *ptep, pte_t old_pte, pte_t pte) argument 1479 track_pfn_remap(struct vm_area_struct *vma, pgprot_t *prot, unsigned long pfn, unsigned long addr, unsigned long size) argument 1490 track_pfn_insert(struct vm_area_struct *vma, pgprot_t *prot, pfn_t pfn) argument 1499 track_pfn_copy(struct vm_area_struct *vma) argument 1509 untrack_pfn(struct vm_area_struct *vma, unsigned long pfn, unsigned long size, bool mm_wr_locked) argument 1519 untrack_pfn_clear(struct vm_area_struct *vma) argument [all...] |
/linux-master/arch/csky/abiv1/ |
H A D | mmap.c | 29 struct vm_area_struct *vma; local 63 vma = find_vma(mm, addr); 65 (!vma || addr + len <= vm_start_gap(vma)))
|
/linux-master/arch/arc/mm/ |
H A D | mmap.c | 29 struct vm_area_struct *vma; local 48 vma = find_vma(mm, addr); 50 (!vma || addr + len <= vm_start_gap(vma)))
|
/linux-master/arch/sh/mm/ |
H A D | tlbflush_32.c | 15 void local_flush_tlb_page(struct vm_area_struct *vma, unsigned long page) argument 19 if (vma->vm_mm && cpu_context(cpu, vma->vm_mm) != NO_CONTEXT) { 24 asid = cpu_asid(cpu, vma->vm_mm); 28 if (vma->vm_mm != current->mm) { 39 void local_flush_tlb_range(struct vm_area_struct *vma, unsigned long start, argument 42 struct mm_struct *mm = vma->vm_mm;
|
/linux-master/drivers/xen/xenbus/ |
H A D | xenbus_dev_backend.c | 92 static int xenbus_backend_mmap(struct file *file, struct vm_area_struct *vma) argument 94 size_t size = vma->vm_end - vma->vm_start; 99 if ((size > PAGE_SIZE) || (vma->vm_pgoff != 0)) 102 if (remap_pfn_range(vma, vma->vm_start, 104 size, vma->vm_page_prot))
|
/linux-master/arch/um/drivers/ |
H A D | mmapper_kern.c | 49 static int mmapper_mmap(struct file *file, struct vm_area_struct *vma) argument 54 if (vma->vm_pgoff != 0) 57 size = vma->vm_end - vma->vm_start; 65 if (remap_pfn_range(vma, vma->vm_start, p_buf >> PAGE_SHIFT, size, 66 vma->vm_page_prot))
|
/linux-master/arch/sh/include/asm/ |
H A D | pgtable.h | 100 extern void __update_cache(struct vm_area_struct *vma, 102 extern void __update_tlb(struct vm_area_struct *vma, 106 struct vm_area_struct *vma, unsigned long address, 110 __update_cache(vma, address, pte); 111 __update_tlb(vma, address, pte); 113 #define update_mmu_cache(vma, addr, ptep) \ 114 update_mmu_cache_range(NULL, vma, addr, ptep, 1) 105 update_mmu_cache_range(struct vm_fault *vmf, struct vm_area_struct *vma, unsigned long address, pte_t *ptep, unsigned int nr) argument
|
H A D | cacheflush.h | 14 * - flush_cache_range(vma, start, end) flushes a range of pages 18 * - flush_icache_pages(vma, pg, nr) flushes(invalidates) pages for icache 40 extern void flush_cache_page(struct vm_area_struct *vma, 42 extern void flush_cache_range(struct vm_area_struct *vma, 54 void flush_icache_pages(struct vm_area_struct *vma, struct page *page, 60 struct vm_area_struct *vma; member in struct:flusher_data 67 static inline void flush_anon_page(struct vm_area_struct *vma, argument 84 extern void copy_to_user_page(struct vm_area_struct *vma, 88 extern void copy_from_user_page(struct vm_area_struct *vma,
|
/linux-master/arch/hexagon/mm/ |
H A D | vm_fault.c | 39 struct vm_area_struct *vma; local 61 vma = lock_mm_and_find_vma(mm, address, regs); 62 if (unlikely(!vma)) 70 if (!(vma->vm_flags & VM_EXEC)) 74 if (!(vma->vm_flags & VM_READ)) 78 if (!(vma->vm_flags & VM_WRITE)) 84 fault = handle_mm_fault(vma, address, flags, regs);
|
/linux-master/fs/nilfs2/ |
H A D | file.c | 47 struct vm_area_struct *vma = vmf->vma; local 49 struct inode *inode = file_inode(vma->vm_file); 100 file_update_time(vma->vm_file); 101 ret = block_page_mkwrite(vma, vmf, nilfs_get_block); 128 static int nilfs_file_mmap(struct file *file, struct vm_area_struct *vma) argument 131 vma->vm_ops = &nilfs_file_vm_ops;
|
/linux-master/mm/ |
H A D | gup.c | 532 * results (0 v.s. -EMLINK) depending on whether vma is available. This is 534 * gup-slow to take the vma reference first. 536 static int gup_hugepte(struct vm_area_struct *vma, pte_t *ptep, unsigned long sz, argument 570 if (!pte_write(pte) && gup_must_unshare(vma, flags, &folio->page)) { 588 static int gup_hugepd(struct vm_area_struct *vma, hugepd_t hugepd, argument 601 ret = gup_hugepte(vma, ptep, sz, addr, end, flags, pages, nr); 609 static struct page *follow_hugepd(struct vm_area_struct *vma, hugepd_t hugepd, argument 621 if (WARN_ON_ONCE(!is_vm_hugetlb_page(vma))) 624 h = hstate_vma(vma); 626 ptl = huge_pte_lock(h, vma 642 gup_hugepd(struct vm_area_struct *vma, hugepd_t hugepd, unsigned long addr, unsigned int pdshift, unsigned long end, unsigned int flags, struct page **pages, int *nr) argument 650 follow_hugepd(struct vm_area_struct *vma, hugepd_t hugepd, unsigned long addr, unsigned int pdshift, unsigned int flags, struct follow_page_context *ctx) argument 660 no_page_table(struct vm_area_struct *vma, unsigned int flags, unsigned long address) argument 686 follow_huge_pud(struct vm_area_struct *vma, unsigned long addr, pud_t *pudp, int flags, struct follow_page_context *ctx) argument 742 can_follow_write_pmd(pmd_t pmd, struct page *page, struct vm_area_struct *vma, unsigned int flags) argument 779 follow_huge_pmd(struct vm_area_struct *vma, unsigned long addr, pmd_t *pmd, unsigned int flags, struct follow_page_context *ctx) argument 825 follow_huge_pud(struct vm_area_struct *vma, unsigned long addr, pud_t *pudp, int flags, struct follow_page_context *ctx) argument 832 follow_huge_pmd(struct vm_area_struct *vma, unsigned long addr, pmd_t *pmd, unsigned int flags, struct follow_page_context *ctx) argument 841 follow_pfn_pte(struct vm_area_struct *vma, unsigned long address, pte_t *pte, unsigned int flags) argument 863 can_follow_write_pte(pte_t pte, struct page *page, struct vm_area_struct *vma, unsigned int flags) argument 900 follow_page_pte(struct vm_area_struct *vma, unsigned long address, pmd_t *pmd, unsigned int flags, struct dev_pagemap **pgmap) argument 1012 follow_pmd_mask(struct vm_area_struct *vma, unsigned long address, pud_t *pudp, unsigned int flags, struct follow_page_context *ctx) argument 1067 follow_pud_mask(struct vm_area_struct *vma, unsigned long address, p4d_t *p4dp, unsigned int flags, struct follow_page_context *ctx) argument 1098 follow_p4d_mask(struct vm_area_struct *vma, unsigned long address, pgd_t *pgdp, unsigned int flags, struct follow_page_context *ctx) argument 1143 follow_page_mask(struct vm_area_struct *vma, unsigned long address, unsigned int flags, struct follow_page_context *ctx) argument 1169 follow_page(struct vm_area_struct *vma, unsigned long address, unsigned int foll_flags) argument 1191 get_gate_page(struct mm_struct *mm, unsigned long address, unsigned int gup_flags, struct vm_area_struct **vma, struct page **page) argument 1251 faultin_page(struct vm_area_struct *vma, unsigned long address, unsigned int *flags, bool unshare, int *locked) argument 1346 writable_file_mapping_allowed(struct vm_area_struct *vma, unsigned long gup_flags) argument 1364 check_vma_flags(struct vm_area_struct *vma, unsigned long gup_flags) argument 1524 struct vm_area_struct *vma = NULL; local 1677 vma_permits_fault(struct vm_area_struct *vma, unsigned int fault_flags) argument 1733 struct vm_area_struct *vma; local 1982 populate_vma_page_range(struct vm_area_struct *vma, unsigned long start, unsigned long end, int *locked) argument 2098 struct vm_area_struct *vma = NULL; local 2152 struct vm_area_struct *vma; local [all...] |