Searched refs:vma (Results 201 - 225 of 1003) sorted by relevance

1234567891011>>

/linux-master/tools/mm/
H A Dthpmaps229 def thp_parse(vma, kpageflags, ranges, indexes, vfns, pfns, anons, heads):
285 anon_pmd_mapped = vma.stats['AnonHugePages']['value']
286 file_pmd_mapped = vma.stats['ShmemPmdMapped']['value'] + \
287 vma.stats['FilePmdMapped']['value']
312 def cont_parse(vma, order, ranges, anons, heads):
346 anon_pmd_mapped = vma.stats['AnonHugePages']['value']
347 file_pmd_mapped = vma.stats['ShmemPmdMapped']['value'] + \
348 vma.stats['FilePmdMapped']['value']
363 def vma_print(vma, pid):
368 pid, vma
[all...]
/linux-master/drivers/xen/
H A Dprivcmd.c71 struct vm_area_struct *vma,
224 struct vm_area_struct *vma; member in struct:mmap_gfn_state
232 struct vm_area_struct *vma = st->vma; local
242 ((msg->va+(msg->npages<<PAGE_SHIFT)) > vma->vm_end))
245 rc = xen_remap_domain_gfn_range(vma,
248 vma->vm_page_prot,
263 struct vm_area_struct *vma; local
293 vma = vma_lookup(mm, msg->va);
296 if (!vma || (ms
322 struct vm_area_struct *vma; member in struct:mmap_batch_state
346 struct vm_area_struct *vma = st->vma; local
426 alloc_empty_pages(struct vm_area_struct *vma, int numpgs) argument
457 struct vm_area_struct *vma; local
734 struct vm_area_struct *vma; local
1194 struct vm_area_struct *vma; local
1561 privcmd_close(struct vm_area_struct *vma) argument
1594 privcmd_mmap(struct file *file, struct vm_area_struct *vma) argument
1616 privcmd_vma_range_is_mapped( struct vm_area_struct *vma, unsigned long addr, unsigned long nr_pages) argument
[all...]
H A Dxlate_mmu.c69 struct vm_area_struct *vma; member in struct:remap_data
138 set_pte_at(info->vma->vm_mm, addr, ptep, pte);
143 int xen_xlate_remap_gfn_array(struct vm_area_struct *vma, argument
156 BUG_ON(!((vma->vm_flags & (VM_PFNMAP | VM_IO)) == (VM_PFNMAP | VM_IO)));
162 data.vma = vma;
168 err = apply_to_page_range(vma->vm_mm, addr, range,
183 int xen_xlate_unmap_gfn_range(struct vm_area_struct *vma, argument
285 int xen_remap_vma_range(struct vm_area_struct *vma, unsigned long addr, unsigned long len) argument
288 .mm = vma
[all...]
/linux-master/mm/
H A Dpage_vma_mapped.c22 pvmw->pte = pte_offset_map_lock(pvmw->vma->vm_mm, pvmw->pmd,
35 pvmw->pte = pte_offset_map_nolock(pvmw->vma->vm_mm, pvmw->pmd,
150 * page_vma_mapped_walk - check if @pvmw->pfn is mapped in @pvmw->vma at
152 * @pvmw: pointer to struct page_vma_mapped_walk. page, vma, address and flags
155 * Returns true if the page is mapped in the vma. @pvmw->pmd and @pvmw->pte point
168 * the vma. @pvmw->ptl is unlocked and @pvmw->pte is unmapped.
175 struct vm_area_struct *vma = pvmw->vma; local
176 struct mm_struct *mm = vma->vm_mm;
188 if (unlikely(is_vm_hugetlb_page(vma))) {
326 page_mapped_in_vma(struct page *page, struct vm_area_struct *vma) argument
[all...]
H A Ddebug.c166 void dump_vma(const struct vm_area_struct *vma) argument
168 pr_emerg("vma %px start %px end %px mm %px\n"
172 vma, (void *)vma->vm_start, (void *)vma->vm_end, vma->vm_mm,
173 (unsigned long)pgprot_val(vma->vm_page_prot),
174 vma->anon_vma, vma->vm_ops, vma
[all...]
/linux-master/drivers/gpu/drm/i915/gem/
H A Di915_gem_domain.c61 struct i915_vma *vma; local
70 spin_lock(&obj->vma.lock);
71 for_each_ggtt_vma(vma, obj)
72 i915_vma_flush_writes(vma);
73 spin_unlock(&obj->vma.lock);
238 struct i915_vma *vma; local
244 spin_lock(&obj->vma.lock);
245 for_each_ggtt_vma(vma, obj)
246 if (i915_vma_is_bound(vma, I915_VMA_GLOBAL_BIND))
247 i915_vma_set_ggtt_write(vma);
432 struct i915_vma *vma; local
[all...]
/linux-master/drivers/misc/sgi-gru/
H A Dgrufile.c59 * and tables belonging to the vma.
61 static void gru_vma_close(struct vm_area_struct *vma) argument
67 if (!vma->vm_private_data)
70 vdata = vma->vm_private_data;
71 vma->vm_private_data = NULL;
72 gru_dbg(grudev, "vma %p, file %p, vdata %p\n", vma, vma->vm_file,
91 * Called when mmapping the device. Initializes the vma with a fault handler
95 static int gru_file_mmap(struct file *file, struct vm_area_struct *vma) argument
124 struct vm_area_struct *vma; local
[all...]
/linux-master/arch/x86/xen/
H A Dmmu.c42 int xen_unmap_domain_gfn_range(struct vm_area_struct *vma, argument
46 return xen_xlate_unmap_gfn_range(vma, nr, pages);
/linux-master/arch/sparc/include/asm/
H A Dtlbflush_64.h30 static inline void flush_tlb_page(struct vm_area_struct *vma, argument
35 static inline void flush_tlb_range(struct vm_area_struct *vma, argument
/linux-master/arch/arm/mm/
H A Dcopypage-v4wb.c48 unsigned long vaddr, struct vm_area_struct *vma)
54 flush_cache_page(vma, vaddr, page_to_pfn(from));
47 v4wb_copy_user_highpage(struct page *to, struct page *from, unsigned long vaddr, struct vm_area_struct *vma) argument
H A Dmmap.c34 struct vm_area_struct *vma; local
65 vma = find_vma(mm, addr);
67 (!vma || addr + len <= vm_start_gap(vma)))
85 struct vm_area_struct *vma; local
116 vma = find_vma(mm, addr);
118 (!vma || addr + len <= vm_start_gap(vma)))
H A Dcopypage-xsc3.c65 unsigned long vaddr, struct vm_area_struct *vma)
71 flush_cache_page(vma, vaddr, page_to_pfn(from));
64 xsc3_mc_copy_user_highpage(struct page *to, struct page *from, unsigned long vaddr, struct vm_area_struct *vma) argument
H A Dcopypage-feroceon.c67 unsigned long vaddr, struct vm_area_struct *vma)
73 flush_cache_page(vma, vaddr, page_to_pfn(from));
66 feroceon_copy_user_highpage(struct page *to, struct page *from, unsigned long vaddr, struct vm_area_struct *vma) argument
/linux-master/fs/ocfs2/
H A Dmmap.c33 struct vm_area_struct *vma = vmf->vma; local
41 trace_ocfs2_fault(OCFS2_I(vma->vm_file->f_mapping->host)->ip_blkno,
42 vma, vmf->page, vmf->pgoff);
116 struct inode *inode = file_inode(vmf->vma->vm_file);
144 ret = __ocfs2_page_mkwrite(vmf->vma->vm_file, di_bh, page);
162 int ocfs2_mmap(struct file *file, struct vm_area_struct *vma) argument
174 vma->vm_ops = &ocfs2_file_vm_ops;
/linux-master/arch/sh/mm/
H A Dmmap.c58 struct vm_area_struct *vma; local
85 vma = find_vma(mm, addr);
87 (!vma || addr + len <= vm_start_gap(vma)))
105 struct vm_area_struct *vma; local
135 vma = find_vma(mm, addr);
137 (!vma || addr + len <= vm_start_gap(vma)))
/linux-master/include/linux/
H A Dkhugepaged.h15 extern void khugepaged_enter_vma(struct vm_area_struct *vma,
48 static inline void khugepaged_enter_vma(struct vm_area_struct *vma, argument
/linux-master/arch/x86/kernel/cpu/sgx/
H A Ddriver.c84 static int sgx_mmap(struct file *file, struct vm_area_struct *vma) argument
89 ret = sgx_encl_may_map(encl, vma->vm_start, vma->vm_end, vma->vm_flags);
93 ret = sgx_encl_mm_add(encl, vma->vm_mm);
97 vma->vm_ops = &sgx_vm_ops;
98 vm_flags_set(vma, VM_PFNMAP | VM_DONTEXPAND | VM_DONTDUMP | VM_IO);
99 vma->vm_private_data = encl;
/linux-master/arch/arm64/include/asm/
H A Dpage.h26 unsigned long vaddr, struct vm_area_struct *vma);
32 struct folio *vma_alloc_zeroed_movable_folio(struct vm_area_struct *vma,
/linux-master/arch/m68k/include/asm/
H A Dpage_no.h16 #define vma_alloc_zeroed_movable_folio(vma, vaddr) \
17 vma_alloc_folio(GFP_HIGHUSER_MOVABLE | __GFP_ZERO, 0, vma, vaddr, false)
/linux-master/arch/openrisc/mm/
H A Dcache.c42 void update_cache(struct vm_area_struct *vma, unsigned long address, argument
54 if ((vma->vm_flags & VM_EXEC) && dirty) {
/linux-master/arch/openrisc/include/asm/
H A Dcacheflush.h71 #define flush_icache_user_page(vma, page, addr, len) \
73 if (vma->vm_flags & VM_EXEC) \
/linux-master/arch/parisc/include/asm/
H A Dhugetlb.h31 static inline pte_t huge_ptep_clear_flush(struct vm_area_struct *vma, argument
42 int huge_ptep_set_access_flags(struct vm_area_struct *vma,
/linux-master/arch/riscv/include/asm/
H A Dcacheflush.h36 #define flush_icache_user_page(vma, pg, addr, len) \
37 flush_icache_mm(vma->vm_mm, 0)
/linux-master/arch/x86/kernel/
H A Dsys_x86_64.c120 struct vm_area_struct *vma; local
134 vma = find_vma(mm, addr);
136 (!vma || addr + len <= vm_start_gap(vma)))
158 struct vm_area_struct *vma; local
181 vma = find_vma(mm, addr);
182 if (!vma || addr + len <= vm_start_gap(vma))
/linux-master/drivers/misc/
H A Dopen-dice.c89 static int open_dice_mmap(struct file *filp, struct vm_area_struct *vma) argument
93 if (vma->vm_flags & VM_MAYSHARE) {
95 if (vma->vm_flags & VM_WRITE)
98 vm_flags_clear(vma, VM_MAYWRITE);
102 vma->vm_page_prot = pgprot_writecombine(vma->vm_page_prot);
103 vm_flags_set(vma, VM_DONTCOPY | VM_DONTDUMP);
104 return vm_iomap_memory(vma, drvdata->rmem->base, drvdata->rmem->size);

Completed in 235 milliseconds

1234567891011>>