/linux-master/arch/riscv/include/asm/ |
H A D | tlbflush.h | 20 __asm__ __volatile__ ("sfence.vma" : : : "memory"); 50 void flush_tlb_page(struct vm_area_struct *vma, unsigned long addr); 51 void flush_tlb_range(struct vm_area_struct *vma, unsigned long start, 57 void flush_pmd_tlb_range(struct vm_area_struct *vma, unsigned long start,
|
H A D | cacheflush.h | 42 #define flush_icache_user_page(vma, pg, addr, len) \ 44 if (vma->vm_flags & VM_EXEC) \ 45 flush_icache_mm(vma->vm_mm, 0); \
|
/linux-master/drivers/gpu/drm/i915/selftests/ |
H A D | i915_gem_evict.c | 55 struct i915_vma *vma; local 62 vma = i915_gem_object_ggtt_pin(obj, NULL, 0, 0, 0); 63 if (IS_ERR(vma)) { 65 if (vma == ERR_PTR(-ENOSPC)) 68 return PTR_ERR(vma); 87 struct i915_vma *vma; local 89 list_for_each_entry(vma, &ggtt->vm.bound_list, vm_link) 90 if (i915_gem_object_has_tiling_quirk(vma->obj)) 91 i915_vma_unpin(vma); 158 struct i915_vma *vma; local 252 struct i915_vma *vma; local [all...] |
/linux-master/arch/powerpc/kernel/ |
H A D | vdso.c | 85 struct vm_area_struct *vma, struct vm_fault *vmf); 119 struct vm_area_struct *vma; local 122 for_each_vma(vmi, vma) { 123 if (vma_is_special_mapping(vma, &vvar_spec)) 124 zap_vma_pages(vma); 133 struct vm_area_struct *vma, struct vm_fault *vmf) 135 struct page *timens_page = find_timens_vvar_page(vma); 163 return vmf_insert_pfn(vma, vmf->address, pfn); 167 * This is called from binfmt_elf, we create the special vma for the 176 struct vm_area_struct *vma; local 132 vvar_fault(const struct vm_special_mapping *sm, struct vm_area_struct *vma, struct vm_fault *vmf) argument [all...] |
/linux-master/arch/sh/kernel/ |
H A D | smp.c | 380 struct vm_area_struct *vma; member in struct:flush_tlb_data 389 local_flush_tlb_range(fd->vma, fd->addr1, fd->addr2); 392 void flush_tlb_range(struct vm_area_struct *vma, argument 395 struct mm_struct *mm = vma->vm_mm; 401 fd.vma = vma; 411 local_flush_tlb_range(vma, start, end); 435 local_flush_tlb_page(fd->vma, fd->addr1); 438 void flush_tlb_page(struct vm_area_struct *vma, unsigned long page) argument 441 if ((atomic_read(&vma [all...] |
/linux-master/arch/sh/mm/ |
H A D | cache.c | 60 void copy_to_user_page(struct vm_area_struct *vma, struct page *page, argument 77 if (vma->vm_flags & VM_EXEC) 78 flush_cache_page(vma, vaddr, page_to_pfn(page)); 81 void copy_from_user_page(struct vm_area_struct *vma, struct page *page, argument 100 unsigned long vaddr, struct vm_area_struct *vma) 119 (vma->vm_flags & VM_EXEC)) 141 void __update_cache(struct vm_area_struct *vma, argument 200 void flush_cache_page(struct vm_area_struct *vma, unsigned long addr, argument 205 data.vma = vma; 99 copy_user_highpage(struct page *to, struct page *from, unsigned long vaddr, struct vm_area_struct *vma) argument 212 flush_cache_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) argument 243 flush_icache_pages(struct vm_area_struct *vma, struct page *page, unsigned int nr) argument [all...] |
/linux-master/mm/ |
H A D | ksm.c | 629 page = vm_normal_page(walk->vma, addr, ptent); 660 * We take great care only to touch a ksm page, in a VM_MERGEABLE vma, 666 * of the process that owns 'vma'. We also do not want to enforce 669 static int break_ksm(struct vm_area_struct *vma, unsigned long addr, bool lock_vma) argument 679 ksm_page = walk_page_range_vma(vma, addr, addr + 1, ops, NULL); 684 ret = handle_mm_fault(vma, addr, 716 static bool vma_ksm_compatible(struct vm_area_struct *vma) argument 718 if (vma->vm_flags & (VM_SHARED | VM_MAYSHARE | VM_PFNMAP | 723 if (vma_is_dax(vma)) 727 if (vma 741 struct vm_area_struct *vma; local 754 struct vm_area_struct *vma; local 773 struct vm_area_struct *vma; local 1070 unmerge_ksm_pages(struct vm_area_struct *vma, unsigned long start, unsigned long end, bool lock_vma) argument 1202 struct vm_area_struct *vma; local 1278 write_protect_page(struct vm_area_struct *vma, struct folio *folio, pte_t *orig_pte) argument 1371 replace_page(struct vm_area_struct *vma, struct page *page, struct page *kpage, pte_t orig_pte) argument 1476 try_to_merge_one_page(struct vm_area_struct *vma, struct page *page, struct page *kpage) argument 1545 struct vm_area_struct *vma; local 2384 struct vm_area_struct *vma; local 2574 struct vm_area_struct *vma; local 2803 __ksm_add_vma(struct vm_area_struct *vma) argument 2814 __ksm_del_vma(struct vm_area_struct *vma) argument 2835 ksm_add_vma(struct vm_area_struct *vma) argument 2845 struct vm_area_struct *vma; local 2854 struct vm_area_struct *vma; local 2933 ksm_madvise(struct vm_area_struct *vma, unsigned long start, unsigned long end, int advice, unsigned long *vm_flags) argument 3059 ksm_might_need_to_copy(struct folio *folio, struct vm_area_struct *vma, unsigned long addr) argument 3129 struct vm_area_struct *vma; local 3187 struct vm_area_struct *vma; local [all...] |
H A D | khugepaged.c | 83 * it would have happened if the vma was large enough during page 349 int hugepage_madvise(struct vm_area_struct *vma, argument 360 if (mm_has_pgste(vma->vm_mm)) 366 * If the vma become good for khugepaged to scan, 370 khugepaged_enter_vma(vma, *vm_flags); 377 * this vma even if we leave the mm registered in khugepaged if 451 void khugepaged_enter_vma(struct vm_area_struct *vma, argument 454 if (!test_bit(MMF_VM_HUGEPAGE, &vma->vm_mm->flags) && 456 if (thp_vma_allowable_order(vma, vm_flags, TVA_ENFORCE_SYSFS, 458 __khugepaged_enter(vma 541 __collapse_huge_page_isolate(struct vm_area_struct *vma, unsigned long address, pte_t *pte, struct collapse_control *cc, struct list_head *compound_pagelist) argument 687 __collapse_huge_page_copy_succeeded(pte_t *pte, struct vm_area_struct *vma, unsigned long address, spinlock_t *ptl, struct list_head *compound_pagelist) argument 740 __collapse_huge_page_copy_failed(pte_t *pte, pmd_t *pmd, pmd_t orig_pmd, struct vm_area_struct *vma, struct list_head *compound_pagelist) argument 779 __collapse_huge_page_copy(pte_t *pte, struct folio *folio, pmd_t *pmd, pmd_t orig_pmd, struct vm_area_struct *vma, unsigned long address, spinlock_t *ptl, struct list_head *compound_pagelist) argument 902 struct vm_area_struct *vma; local 973 __collapse_huge_page_swapin(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long haddr, pmd_t *pmd, int referenced) argument 1084 struct vm_area_struct *vma; local 1235 hpage_collapse_scan_pmd(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long address, bool *mmap_locked, struct collapse_control *cc) argument 1439 set_huge_pmd(struct vm_area_struct *vma, unsigned long addr, pmd_t *pmdp, struct page *hpage) argument 1477 struct vm_area_struct *vma = vma_lookup(mm, haddr); local 1678 struct vm_area_struct *vma; local 2059 struct vm_area_struct *vma; local 2324 struct vm_area_struct *vma; variable in typeref:struct:vm_area_struct 2689 madvise_collapse(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end) argument [all...] |
/linux-master/scripts/ |
H A D | extract-sys-certs.pl | 39 my $vma = Math::BigInt->new("0x" . $4); 45 vma => $vma, 109 my $s_vma = $sec->{vma}; 143 my $foff = $start - $s->{vma} + $s->{foff};
|
/linux-master/arch/powerpc/include/asm/ |
H A D | pkeys.h | 38 static inline int vma_pkey(struct vm_area_struct *vma) argument 42 return (vma->vm_flags & ARCH_VM_PKEY_FLAGS) >> VM_PKEY_SHIFT; 128 extern int __arch_override_mprotect_pkey(struct vm_area_struct *vma, 130 static inline int arch_override_mprotect_pkey(struct vm_area_struct *vma, argument 143 return __arch_override_mprotect_pkey(vma, prot, pkey);
|
/linux-master/arch/mips/kernel/ |
H A D | smp.c | 571 struct vm_area_struct *vma; member in struct:flush_tlb_data 580 local_flush_tlb_range(fd->vma, fd->addr1, fd->addr2); 583 void flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) argument 585 struct mm_struct *mm = vma->vm_mm; 607 .vma = vma, 613 local_flush_tlb_range(vma, start, end); 616 int exec = vma->vm_flags & VM_EXEC; 628 local_flush_tlb_range(vma, start, end); 654 local_flush_tlb_page(fd->vma, f 657 flush_tlb_page(struct vm_area_struct *vma, unsigned long page) argument [all...] |
/linux-master/arch/mips/include/asm/ |
H A D | cacheflush.h | 22 * - flush_cache_range(vma, start, end) flushes a range of pages 50 extern void (*flush_cache_range)(struct vm_area_struct *vma, 52 extern void (*flush_cache_page)(struct vm_area_struct *vma, unsigned long page, unsigned long pfn); 78 static inline void flush_anon_page(struct vm_area_struct *vma, argument 110 extern void copy_to_user_page(struct vm_area_struct *vma, 114 extern void copy_from_user_page(struct vm_area_struct *vma,
|
/linux-master/drivers/gpu/drm/i915/display/ |
H A D | intel_fbdev_fb.c | 70 struct drm_i915_gem_object *obj, struct i915_vma *vma) 90 (unsigned long)(ggtt->gmadr.start + i915_ggtt_offset(vma)); 91 info->fix.smem_len = vma->size; 95 ret = i915_gem_object_lock(vma->obj, &ww); 100 vaddr = i915_vma_pin_iomap(vma); 69 intel_fbdev_fb_fill_info(struct drm_i915_private *i915, struct fb_info *info, struct drm_i915_gem_object *obj, struct i915_vma *vma) argument
|
/linux-master/arch/s390/include/asm/ |
H A D | hugetlb.h | 57 static inline pte_t huge_ptep_clear_flush(struct vm_area_struct *vma, argument 60 return huge_ptep_get_and_clear(vma->vm_mm, address, ptep); 63 static inline int huge_ptep_set_access_flags(struct vm_area_struct *vma, argument 69 huge_ptep_get_and_clear(vma->vm_mm, addr, ptep); 70 __set_huge_pte_at(vma->vm_mm, addr, ptep, pte);
|
/linux-master/arch/powerpc/kvm/ |
H A D | book3s_hv_uvmem.c | 398 struct vm_area_struct *vma; local 408 vma = find_vma_intersection(kvm->mm, start, end); 409 if (!vma) { 413 vma_start_write(vma); 415 vm_flags = vma->vm_flags; 416 ret = ksm_madvise(vma, vma->vm_start, vma->vm_end, 422 vm_flags_reset(vma, vm_flags); 423 start = vma 513 __kvmppc_svm_page_out(struct vm_area_struct *vma, unsigned long start, unsigned long end, unsigned long page_shift, struct kvm *kvm, unsigned long gpa, struct page *fault_page) argument 586 kvmppc_svm_page_out(struct vm_area_struct *vma, unsigned long start, unsigned long end, unsigned long page_shift, struct kvm *kvm, unsigned long gpa, struct page *fault_page) argument 616 struct vm_area_struct *vma = NULL; local 740 kvmppc_svm_page_in(struct vm_area_struct *vma, unsigned long start, unsigned long end, unsigned long gpa, struct kvm *kvm, unsigned long page_shift, bool pagein) argument 798 struct vm_area_struct *vma; local 941 struct vm_area_struct *vma; local 1051 struct vm_area_struct *vma; local [all...] |
/linux-master/arch/um/kernel/ |
H A D | trap.c | 28 struct vm_area_struct *vma; local 47 vma = find_vma(mm, address); 48 if (!vma) 50 if (vma->vm_start <= address) 52 if (!(vma->vm_flags & VM_GROWSDOWN)) 56 vma = expand_stack(mm, address); 57 if (!vma) 63 if (!(vma->vm_flags & VM_WRITE)) 68 if (!(vma->vm_flags & (VM_READ | VM_EXEC))) 75 fault = handle_mm_fault(vma, addres [all...] |
/linux-master/drivers/char/ |
H A D | bsr.c | 114 static int bsr_mmap(struct file *filp, struct vm_area_struct *vma) argument 116 unsigned long size = vma->vm_end - vma->vm_start; 120 vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot); 124 ret = remap_4k_pfn(vma, vma->vm_start, dev->bsr_addr >> 12, 125 vma->vm_page_prot); 127 ret = io_remap_pfn_range(vma, vma [all...] |
/linux-master/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_doorbell.c | 107 struct vm_area_struct *vma) 116 if (vma->vm_end - vma->vm_start != kfd_doorbell_process_slice(dev->kfd)) 127 vm_flags_set(vma, VM_IO | VM_DONTCOPY | VM_DONTEXPAND | VM_NORESERVE | 130 vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot); 137 (unsigned long long) vma->vm_start, address, vma->vm_flags, 141 return io_remap_pfn_range(vma, 142 vma 106 kfd_doorbell_mmap(struct kfd_node *dev, struct kfd_process *process, struct vm_area_struct *vma) argument [all...] |
/linux-master/arch/powerpc/mm/book3s64/ |
H A D | pkeys.c | 379 static inline bool vma_is_pkey_exec_only(struct vm_area_struct *vma) argument 382 if ((vma->vm_flags & VM_ACCESS_FLAGS) != VM_EXEC) 385 return (vma_pkey(vma) == vma->vm_mm->context.execute_only_pkey); 391 int __arch_override_mprotect_pkey(struct vm_area_struct *vma, int prot, argument 398 if (vma_is_pkey_exec_only(vma) && (prot != PROT_EXEC)) 406 pkey = execute_only_pkey(vma->vm_mm); 412 return vma_pkey(vma); 447 bool arch_vma_access_permitted(struct vm_area_struct *vma, bool write, argument 453 * Do not enforce our key-permissions on a foreign vma [all...] |
/linux-master/arch/mips/mm/ |
H A D | tlb-r3k.c | 67 void local_flush_tlb_range(struct vm_area_struct *vma, unsigned long start, argument 71 struct mm_struct *mm = vma->vm_mm; 147 void local_flush_tlb_page(struct vm_area_struct *vma, unsigned long page) argument 152 if (cpu_context(cpu, vma->vm_mm) != 0) { 157 printk("[tlbpage<%lu,0x%08lx>]", cpu_context(cpu, vma->vm_mm), page); 159 newpid = cpu_context(cpu, vma->vm_mm) & asid_mask; 179 void __update_tlb(struct vm_area_struct *vma, unsigned long address, pte_t pte) argument 188 if (current->active_mm != vma->vm_mm) 194 if ((pid != (cpu_context(cpu, vma->vm_mm) & asid_mask)) || (cpu_context(cpu, vma [all...] |
/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_hmm.c | 118 * @uvma: the userptr vma which hold the scatter gather table 127 struct xe_vma *vma = &uvma->vma; local 128 bool write = !xe_vma_read_only(vma); 129 struct xe_vm *vm = xe_vma_vm(vma); 145 * @uvma: userptr vma which has information of the range to populate. 171 struct xe_vma *vma = &uvma->vma; local 172 u64 userptr_start = xe_vma_userptr(vma); 173 u64 userptr_end = userptr_start + xe_vma_size(vma); [all...] |
/linux-master/arch/nios2/mm/ |
H A D | fault.c | 46 struct vm_area_struct *vma = NULL; local 90 vma = lock_mm_and_find_vma(mm, address, regs); 91 if (!vma) 105 if (!(vma->vm_flags & VM_EXEC)) 109 if (!(vma->vm_flags & VM_READ)) 113 if (!(vma->vm_flags & VM_WRITE)) 124 fault = handle_mm_fault(vma, address, flags, regs);
|
/linux-master/arch/alpha/mm/ |
H A D | fault.c | 88 struct vm_area_struct * vma; local 122 vma = lock_mm_and_find_vma(mm, address, regs); 123 if (!vma) 130 if (!(vma->vm_flags & VM_EXEC)) 134 if (!(vma->vm_flags & (VM_READ | VM_WRITE))) 137 if (!(vma->vm_flags & VM_WRITE)) 145 fault = handle_mm_fault(vma, address, flags, regs);
|
/linux-master/drivers/uio/ |
H A D | uio.c | 672 static int uio_find_mem_index(struct vm_area_struct *vma) argument 674 struct uio_device *idev = vma->vm_private_data; 676 if (vma->vm_pgoff < MAX_UIO_MAPS) { 677 if (idev->info->mem[vma->vm_pgoff].size == 0) 679 return (int)vma->vm_pgoff; 686 struct uio_device *idev = vmf->vma->vm_private_data; 699 mi = uio_find_mem_index(vmf->vma); 729 static int uio_mmap_logical(struct vm_area_struct *vma) argument 731 vm_flags_set(vma, VM_DONTEXPAND | VM_DONTDUMP); 732 vma 742 uio_mmap_physical(struct vm_area_struct *vma) argument 777 uio_mmap_dma_coherent(struct vm_area_struct *vma) argument 820 uio_mmap(struct file *filep, struct vm_area_struct *vma) argument [all...] |
/linux-master/drivers/misc/cxl/ |
H A D | context.c | 128 struct vm_area_struct *vma = vmf->vma; local 129 struct cxl_context *ctx = vma->vm_file->private_data; 162 vma->vm_page_prot = pgprot_cached(vma->vm_page_prot); 168 ret = vmf_insert_pfn(vma, vmf->address, (area + offset) >> PAGE_SHIFT); 180 * Map a per-context mmio space into the given vma. 182 int cxl_context_iomap(struct cxl_context *ctx, struct vm_area_struct *vma) argument 184 u64 start = vma->vm_pgoff << PAGE_SHIFT; 185 u64 len = vma [all...] |