/linux-master/tools/mm/ |
H A D | thpmaps | 229 def thp_parse(vma, kpageflags, ranges, indexes, vfns, pfns, anons, heads): 285 anon_pmd_mapped = vma.stats['AnonHugePages']['value'] 286 file_pmd_mapped = vma.stats['ShmemPmdMapped']['value'] + \ 287 vma.stats['FilePmdMapped']['value'] 312 def cont_parse(vma, order, ranges, anons, heads): 346 anon_pmd_mapped = vma.stats['AnonHugePages']['value'] 347 file_pmd_mapped = vma.stats['ShmemPmdMapped']['value'] + \ 348 vma.stats['FilePmdMapped']['value'] 363 def vma_print(vma, pid): 368 pid, vma [all...] |
/linux-master/drivers/xen/ |
H A D | privcmd.c | 71 struct vm_area_struct *vma, 224 struct vm_area_struct *vma; member in struct:mmap_gfn_state 232 struct vm_area_struct *vma = st->vma; local 242 ((msg->va+(msg->npages<<PAGE_SHIFT)) > vma->vm_end)) 245 rc = xen_remap_domain_gfn_range(vma, 248 vma->vm_page_prot, 263 struct vm_area_struct *vma; local 293 vma = vma_lookup(mm, msg->va); 296 if (!vma || (ms 322 struct vm_area_struct *vma; member in struct:mmap_batch_state 346 struct vm_area_struct *vma = st->vma; local 426 alloc_empty_pages(struct vm_area_struct *vma, int numpgs) argument 457 struct vm_area_struct *vma; local 734 struct vm_area_struct *vma; local 1194 struct vm_area_struct *vma; local 1561 privcmd_close(struct vm_area_struct *vma) argument 1594 privcmd_mmap(struct file *file, struct vm_area_struct *vma) argument 1616 privcmd_vma_range_is_mapped( struct vm_area_struct *vma, unsigned long addr, unsigned long nr_pages) argument [all...] |
H A D | xlate_mmu.c | 69 struct vm_area_struct *vma; member in struct:remap_data 138 set_pte_at(info->vma->vm_mm, addr, ptep, pte); 143 int xen_xlate_remap_gfn_array(struct vm_area_struct *vma, argument 156 BUG_ON(!((vma->vm_flags & (VM_PFNMAP | VM_IO)) == (VM_PFNMAP | VM_IO))); 162 data.vma = vma; 168 err = apply_to_page_range(vma->vm_mm, addr, range, 183 int xen_xlate_unmap_gfn_range(struct vm_area_struct *vma, argument 285 int xen_remap_vma_range(struct vm_area_struct *vma, unsigned long addr, unsigned long len) argument 288 .mm = vma [all...] |
/linux-master/mm/ |
H A D | page_vma_mapped.c | 22 pvmw->pte = pte_offset_map_lock(pvmw->vma->vm_mm, pvmw->pmd, 35 pvmw->pte = pte_offset_map_nolock(pvmw->vma->vm_mm, pvmw->pmd, 150 * page_vma_mapped_walk - check if @pvmw->pfn is mapped in @pvmw->vma at 152 * @pvmw: pointer to struct page_vma_mapped_walk. page, vma, address and flags 155 * Returns true if the page is mapped in the vma. @pvmw->pmd and @pvmw->pte point 168 * the vma. @pvmw->ptl is unlocked and @pvmw->pte is unmapped. 175 struct vm_area_struct *vma = pvmw->vma; local 176 struct mm_struct *mm = vma->vm_mm; 188 if (unlikely(is_vm_hugetlb_page(vma))) { 326 page_mapped_in_vma(struct page *page, struct vm_area_struct *vma) argument [all...] |
H A D | debug.c | 166 void dump_vma(const struct vm_area_struct *vma) argument 168 pr_emerg("vma %px start %px end %px mm %px\n" 172 vma, (void *)vma->vm_start, (void *)vma->vm_end, vma->vm_mm, 173 (unsigned long)pgprot_val(vma->vm_page_prot), 174 vma->anon_vma, vma->vm_ops, vma [all...] |
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_domain.c | 61 struct i915_vma *vma; local 70 spin_lock(&obj->vma.lock); 71 for_each_ggtt_vma(vma, obj) 72 i915_vma_flush_writes(vma); 73 spin_unlock(&obj->vma.lock); 238 struct i915_vma *vma; local 244 spin_lock(&obj->vma.lock); 245 for_each_ggtt_vma(vma, obj) 246 if (i915_vma_is_bound(vma, I915_VMA_GLOBAL_BIND)) 247 i915_vma_set_ggtt_write(vma); 432 struct i915_vma *vma; local [all...] |
/linux-master/drivers/misc/sgi-gru/ |
H A D | grufile.c | 59 * and tables belonging to the vma. 61 static void gru_vma_close(struct vm_area_struct *vma) argument 67 if (!vma->vm_private_data) 70 vdata = vma->vm_private_data; 71 vma->vm_private_data = NULL; 72 gru_dbg(grudev, "vma %p, file %p, vdata %p\n", vma, vma->vm_file, 91 * Called when mmapping the device. Initializes the vma with a fault handler 95 static int gru_file_mmap(struct file *file, struct vm_area_struct *vma) argument 124 struct vm_area_struct *vma; local [all...] |
/linux-master/arch/x86/xen/ |
H A D | mmu.c | 42 int xen_unmap_domain_gfn_range(struct vm_area_struct *vma, argument 46 return xen_xlate_unmap_gfn_range(vma, nr, pages);
|
/linux-master/arch/sparc/include/asm/ |
H A D | tlbflush_64.h | 30 static inline void flush_tlb_page(struct vm_area_struct *vma, argument 35 static inline void flush_tlb_range(struct vm_area_struct *vma, argument
|
/linux-master/arch/arm/mm/ |
H A D | copypage-v4wb.c | 48 unsigned long vaddr, struct vm_area_struct *vma) 54 flush_cache_page(vma, vaddr, page_to_pfn(from)); 47 v4wb_copy_user_highpage(struct page *to, struct page *from, unsigned long vaddr, struct vm_area_struct *vma) argument
|
H A D | mmap.c | 34 struct vm_area_struct *vma; local 65 vma = find_vma(mm, addr); 67 (!vma || addr + len <= vm_start_gap(vma))) 85 struct vm_area_struct *vma; local 116 vma = find_vma(mm, addr); 118 (!vma || addr + len <= vm_start_gap(vma)))
|
H A D | copypage-xsc3.c | 65 unsigned long vaddr, struct vm_area_struct *vma) 71 flush_cache_page(vma, vaddr, page_to_pfn(from)); 64 xsc3_mc_copy_user_highpage(struct page *to, struct page *from, unsigned long vaddr, struct vm_area_struct *vma) argument
|
H A D | copypage-feroceon.c | 67 unsigned long vaddr, struct vm_area_struct *vma) 73 flush_cache_page(vma, vaddr, page_to_pfn(from)); 66 feroceon_copy_user_highpage(struct page *to, struct page *from, unsigned long vaddr, struct vm_area_struct *vma) argument
|
/linux-master/fs/ocfs2/ |
H A D | mmap.c | 33 struct vm_area_struct *vma = vmf->vma; local 41 trace_ocfs2_fault(OCFS2_I(vma->vm_file->f_mapping->host)->ip_blkno, 42 vma, vmf->page, vmf->pgoff); 116 struct inode *inode = file_inode(vmf->vma->vm_file); 144 ret = __ocfs2_page_mkwrite(vmf->vma->vm_file, di_bh, page); 162 int ocfs2_mmap(struct file *file, struct vm_area_struct *vma) argument 174 vma->vm_ops = &ocfs2_file_vm_ops;
|
/linux-master/arch/sh/mm/ |
H A D | mmap.c | 58 struct vm_area_struct *vma; local 85 vma = find_vma(mm, addr); 87 (!vma || addr + len <= vm_start_gap(vma))) 105 struct vm_area_struct *vma; local 135 vma = find_vma(mm, addr); 137 (!vma || addr + len <= vm_start_gap(vma)))
|
/linux-master/include/linux/ |
H A D | khugepaged.h | 15 extern void khugepaged_enter_vma(struct vm_area_struct *vma, 48 static inline void khugepaged_enter_vma(struct vm_area_struct *vma, argument
|
/linux-master/arch/x86/kernel/cpu/sgx/ |
H A D | driver.c | 84 static int sgx_mmap(struct file *file, struct vm_area_struct *vma) argument 89 ret = sgx_encl_may_map(encl, vma->vm_start, vma->vm_end, vma->vm_flags); 93 ret = sgx_encl_mm_add(encl, vma->vm_mm); 97 vma->vm_ops = &sgx_vm_ops; 98 vm_flags_set(vma, VM_PFNMAP | VM_DONTEXPAND | VM_DONTDUMP | VM_IO); 99 vma->vm_private_data = encl;
|
/linux-master/arch/arm64/include/asm/ |
H A D | page.h | 26 unsigned long vaddr, struct vm_area_struct *vma); 32 struct folio *vma_alloc_zeroed_movable_folio(struct vm_area_struct *vma,
|
/linux-master/arch/m68k/include/asm/ |
H A D | page_no.h | 16 #define vma_alloc_zeroed_movable_folio(vma, vaddr) \ 17 vma_alloc_folio(GFP_HIGHUSER_MOVABLE | __GFP_ZERO, 0, vma, vaddr, false)
|
/linux-master/arch/openrisc/mm/ |
H A D | cache.c | 42 void update_cache(struct vm_area_struct *vma, unsigned long address, argument 54 if ((vma->vm_flags & VM_EXEC) && dirty) {
|
/linux-master/arch/openrisc/include/asm/ |
H A D | cacheflush.h | 71 #define flush_icache_user_page(vma, page, addr, len) \ 73 if (vma->vm_flags & VM_EXEC) \
|
/linux-master/arch/parisc/include/asm/ |
H A D | hugetlb.h | 31 static inline pte_t huge_ptep_clear_flush(struct vm_area_struct *vma, argument 42 int huge_ptep_set_access_flags(struct vm_area_struct *vma,
|
/linux-master/arch/riscv/include/asm/ |
H A D | cacheflush.h | 36 #define flush_icache_user_page(vma, pg, addr, len) \ 37 flush_icache_mm(vma->vm_mm, 0)
|
/linux-master/arch/x86/kernel/ |
H A D | sys_x86_64.c | 120 struct vm_area_struct *vma; local 134 vma = find_vma(mm, addr); 136 (!vma || addr + len <= vm_start_gap(vma))) 158 struct vm_area_struct *vma; local 181 vma = find_vma(mm, addr); 182 if (!vma || addr + len <= vm_start_gap(vma))
|
/linux-master/drivers/misc/ |
H A D | open-dice.c | 89 static int open_dice_mmap(struct file *filp, struct vm_area_struct *vma) argument 93 if (vma->vm_flags & VM_MAYSHARE) { 95 if (vma->vm_flags & VM_WRITE) 98 vm_flags_clear(vma, VM_MAYWRITE); 102 vma->vm_page_prot = pgprot_writecombine(vma->vm_page_prot); 103 vm_flags_set(vma, VM_DONTCOPY | VM_DONTDUMP); 104 return vm_iomap_memory(vma, drvdata->rmem->base, drvdata->rmem->size);
|