Searched refs:vma (Results 351 - 375 of 1013) sorted by relevance

<<11121314151617181920>>

/linux-master/fs/romfs/
H A Dmmap-nommu.c64 static int romfs_mmap(struct file *file, struct vm_area_struct *vma) argument
66 return is_nommu_shared_mapping(vma->vm_flags) ? 0 : -ENOSYS;
/linux-master/drivers/gpu/drm/nouveau/nvif/
H A Dvmm.c65 nvif_vmm_put(struct nvif_vmm *vmm, struct nvif_vma *vma) argument
67 if (vma->size) {
70 .addr = vma->addr,
72 vma->size = 0;
78 u8 page, u8 align, u64 size, struct nvif_vma *vma)
83 args.version = vma->size = 0;
101 vma->addr = args.addr;
102 vma->size = args.size;
77 nvif_vmm_get(struct nvif_vmm *vmm, enum nvif_vmm_get type, bool sparse, u8 page, u8 align, u64 size, struct nvif_vma *vma) argument
/linux-master/arch/mips/alchemy/common/
H A Dsetup.c97 int io_remap_pfn_range(struct vm_area_struct *vma, unsigned long vaddr, argument
102 return remap_pfn_range(vma, vaddr, phys_addr >> PAGE_SHIFT, size, prot);
/linux-master/arch/x86/include/asm/
H A Dpage.h37 #define vma_alloc_zeroed_movable_folio(vma, vaddr) \
38 vma_alloc_folio(GFP_HIGHUSER_MOVABLE | __GFP_ZERO, 0, vma, vaddr, false)
/linux-master/arch/mips/mm/
H A Dfault.c42 struct vm_area_struct * vma = NULL; local
103 vma = lock_mm_and_find_vma(mm, address, regs);
104 if (!vma)
113 if (!(vma->vm_flags & VM_WRITE))
118 if (address == regs->cp0_epc && !(vma->vm_flags & VM_EXEC)) {
128 if (!(vma->vm_flags & VM_READ) &&
140 if (unlikely(!vma_is_accessible(vma)))
150 fault = handle_mm_fault(vma, address, flags, regs);
H A Dtlb-r4k.c48 static inline void flush_micro_tlb_vm(struct vm_area_struct *vma) argument
50 if (vma->vm_flags & VM_EXEC)
106 void local_flush_tlb_range(struct vm_area_struct *vma, unsigned long start, argument
109 struct mm_struct *mm = vma->vm_mm;
211 void local_flush_tlb_page(struct vm_area_struct *vma, unsigned long page) argument
215 if (cpu_context(cpu, vma->vm_mm) != 0) {
227 write_c0_memorymapid(cpu_asid(cpu, vma->vm_mm));
229 write_c0_entryhi(page | cpu_asid(cpu, vma->vm_mm));
250 flush_micro_tlb_vm(vma);
293 void __update_tlb(struct vm_area_struct * vma, unsigne argument
[all...]
/linux-master/include/linux/
H A Dmempolicy.h34 * its own state. All vma manipulation is somewhat protected by a down_read on
124 struct vm_area_struct *vma, struct mempolicy *mpol);
130 struct mempolicy *__get_vma_policy(struct vm_area_struct *vma,
132 struct mempolicy *get_vma_policy(struct vm_area_struct *vma,
134 bool vma_policy_mof(struct vm_area_struct *vma);
141 extern int huge_node(struct vm_area_struct *vma,
167 /* Check if a vma is migratable */
168 extern bool vma_migratable(struct vm_area_struct *vma);
224 static inline struct mempolicy *get_vma_policy(struct vm_area_struct *vma, argument
254 static inline int huge_node(struct vm_area_struct *vma, argument
[all...]
/linux-master/arch/xtensa/mm/
H A Dtlb.c89 void local_flush_tlb_range(struct vm_area_struct *vma, argument
93 struct mm_struct *mm = vma->vm_mm;
108 if (vma->vm_flags & VM_EXEC)
127 void local_flush_tlb_page(struct vm_area_struct *vma, unsigned long page) argument
130 struct mm_struct* mm = vma->vm_mm;
142 if (vma->vm_flags & VM_EXEC)
166 void update_mmu_tlb(struct vm_area_struct *vma, argument
169 local_flush_tlb_page(vma, address);
/linux-master/arch/riscv/include/asm/
H A Dhugetlb.h34 pte_t huge_ptep_clear_flush(struct vm_area_struct *vma,
42 int huge_ptep_set_access_flags(struct vm_area_struct *vma,
/linux-master/arch/x86/kernel/
H A Dsys_x86_64.c128 struct vm_area_struct *vma; local
142 vma = find_vma(mm, addr);
144 (!vma || addr + len <= vm_start_gap(vma)))
165 struct vm_area_struct *vma; local
188 vma = find_vma(mm, addr);
189 if (!vma || addr + len <= vm_start_gap(vma))
/linux-master/arch/s390/mm/
H A Dmmap.c88 struct vm_area_struct *vma; local
99 vma = find_vma(mm, addr);
101 (!vma || addr + len <= vm_start_gap(vma)))
122 struct vm_area_struct *vma; local
136 vma = find_vma(mm, addr);
138 (!vma || addr + len <= vm_start_gap(vma)))
H A Dfault.c284 struct vm_area_struct *vma; local
323 vma = lock_vma_under_rcu(mm, address);
324 if (!vma)
326 if (!(vma->vm_flags & access)) {
327 vma_end_read(vma);
331 fault = handle_mm_fault(vma, address, flags | FAULT_FLAG_VMA_LOCK, regs);
333 vma_end_read(vma);
365 vma = find_vma(mm, address);
366 if (!vma)
368 if (unlikely(vma
493 struct vm_area_struct *vma; local
[all...]
/linux-master/arch/sparc/mm/
H A Dfault_64.c271 struct vm_area_struct *vma; local
338 vma = find_vma(mm, address);
339 if (!vma)
352 (vma->vm_flags & VM_WRITE) != 0) {
371 if (vma->vm_start <= address)
373 if (!(vma->vm_flags & VM_GROWSDOWN))
389 vma = expand_stack(mm, address);
390 if (!vma)
402 if ((fault_code & FAULT_CODE_ITLB) && !(vma->vm_flags & VM_EXEC)) {
410 if (!(vma
[all...]
H A Dleon_mm.c194 void leon_flush_pcache_all(struct vm_area_struct *vma, unsigned long page) argument
196 if (vma->vm_flags & VM_EXEC)
282 static void leon_flush_cache_page(struct vm_area_struct *vma, unsigned long page) argument
284 leon_flush_pcache_all(vma, page);
287 static void leon_flush_cache_range(struct vm_area_struct *vma, argument
299 static void leon_flush_tlb_page(struct vm_area_struct *vma, argument
305 static void leon_flush_tlb_range(struct vm_area_struct *vma, argument
/linux-master/drivers/gpu/drm/rockchip/
H A Drockchip_drm_gem.c216 struct vm_area_struct *vma)
220 unsigned long user_count = vma_pages(vma);
225 return vm_map_pages(vma, rk_obj->pages, count);
229 struct vm_area_struct *vma)
234 return dma_mmap_attrs(drm->dev, vma, rk_obj->kvaddr, rk_obj->dma_addr,
239 struct vm_area_struct *vma)
248 vma->vm_pgoff = 0;
254 vm_flags_mod(vma, VM_IO | VM_DONTEXPAND | VM_DONTDUMP, VM_PFNMAP);
256 vma->vm_page_prot = pgprot_writecombine(vm_get_page_prot(vma
215 rockchip_drm_gem_object_mmap_iommu(struct drm_gem_object *obj, struct vm_area_struct *vma) argument
228 rockchip_drm_gem_object_mmap_dma(struct drm_gem_object *obj, struct vm_area_struct *vma) argument
238 rockchip_drm_gem_object_mmap(struct drm_gem_object *obj, struct vm_area_struct *vma) argument
[all...]
/linux-master/drivers/xen/
H A Dgntalloc.c448 static void gntalloc_vma_open(struct vm_area_struct *vma) argument
450 struct gntalloc_vma_private_data *priv = vma->vm_private_data;
460 static void gntalloc_vma_close(struct vm_area_struct *vma) argument
462 struct gntalloc_vma_private_data *priv = vma->vm_private_data;
491 static int gntalloc_mmap(struct file *filp, struct vm_area_struct *vma) argument
496 int count = vma_pages(vma);
499 if (!(vma->vm_flags & VM_SHARED)) {
511 priv, vm_priv, vma->vm_pgoff, count);
513 gref = find_grefs(priv, vma->vm_pgoff << PAGE_SHIFT, count);
526 vma
[all...]
H A Dgntdev.c499 static void gntdev_vma_open(struct vm_area_struct *vma) argument
501 struct gntdev_grant_map *map = vma->vm_private_data;
503 pr_debug("gntdev_vma_open %p\n", vma);
507 static void gntdev_vma_close(struct vm_area_struct *vma) argument
509 struct gntdev_grant_map *map = vma->vm_private_data;
510 struct file *file = vma->vm_file;
513 pr_debug("gntdev_vma_close %p\n", vma);
515 vma->vm_private_data = NULL;
519 static struct page *gntdev_vma_find_special_page(struct vm_area_struct *vma, argument
522 struct gntdev_grant_map *map = vma
693 struct vm_area_struct *vma; local
1033 gntdev_mmap(struct file *flip, struct vm_area_struct *vma) argument
[all...]
/linux-master/arch/arc/mm/
H A Dfault.c75 struct vm_area_struct *vma = NULL; local
117 vma = lock_mm_and_find_vma(mm, address, regs);
118 if (!vma)
130 if (!(vma->vm_flags & mask)) {
135 fault = handle_mm_fault(vma, address, flags, regs);
/linux-master/drivers/gpu/drm/i915/gt/uc/
H A Dintel_guc_log.c297 relay_reserve(log->relay.channel, log->vma->obj->base.size -
492 if (!log->vma || !log->buf_addr)
499 i915_gem_object_get(log->vma->obj);
509 i915_gem_object_put(log->vma->obj);
529 GEM_BUG_ON(!log->vma);
535 subbuf_size = log->vma->size - intel_guc_log_section_size_capture(log);
617 struct i915_vma *vma; local
622 GEM_BUG_ON(log->vma);
626 vma = intel_guc_allocate_vma(guc, guc_log_size);
627 if (IS_ERR(vma)) {
[all...]
/linux-master/kernel/dma/
H A Dcoherent.c235 struct vm_area_struct *vma, void *vaddr, size_t size, int *ret)
239 unsigned long off = vma->vm_pgoff;
241 unsigned long user_count = vma_pages(vma);
247 *ret = remap_pfn_range(vma, vma->vm_start, pfn,
249 vma->vm_page_prot);
259 * @vma: vm_area for the userspace memory
265 * coherent memory pool and if so, maps that memory to the provided vma.
271 int dma_mmap_from_dev_coherent(struct device *dev, struct vm_area_struct *vma, argument
276 return __dma_mmap_from_coherent(mem, vma, vadd
234 __dma_mmap_from_coherent(struct dma_coherent_mem *mem, struct vm_area_struct *vma, void *vaddr, size_t size, int *ret) argument
301 dma_mmap_from_global_coherent(struct vm_area_struct *vma, void *vaddr, size_t size, int *ret) argument
[all...]
/linux-master/ipc/
H A Dshm.c102 static void shm_open(struct vm_area_struct *vma);
103 static void shm_close(struct vm_area_struct *vma);
302 static void shm_open(struct vm_area_struct *vma) argument
304 struct file *file = vma->vm_file;
310 sfd->vm_ops->open(vma);
364 * remove the attach descriptor vma.
396 static void shm_close(struct vm_area_struct *vma) argument
398 struct file *file = vma->vm_file;
403 sfd->vm_ops->close(vma);
537 struct file *file = vmf->vma
543 shm_may_split(struct vm_area_struct *vma, unsigned long addr) argument
554 shm_pagesize(struct vm_area_struct *vma) argument
566 shm_set_policy(struct vm_area_struct *vma, struct mempolicy *mpol) argument
576 shm_get_policy(struct vm_area_struct *vma, unsigned long addr, pgoff_t *ilx) argument
588 shm_mmap(struct file *file, struct vm_area_struct *vma) argument
1726 struct vm_area_struct *vma; local
[all...]
/linux-master/fs/hugetlbfs/
H A Dinode.c96 static int hugetlbfs_file_mmap(struct file *file, struct vm_area_struct *vma) argument
106 * vma address alignment (but not the pgoff alignment) has
113 vm_flags_set(vma, VM_HUGETLB | VM_DONTEXPAND);
114 vma->vm_ops = &hugetlb_vm_ops;
116 ret = seal_check_write(info->seals, vma);
127 if (vma->vm_pgoff & PGOFF_LOFFT_MAX)
132 if (vma->vm_pgoff & (~huge_page_mask(h) >> PAGE_SHIFT))
135 vma_len = (loff_t)(vma->vm_end - vma->vm_start);
136 len = vma_len + ((loff_t)vma
225 struct vm_area_struct *vma; local
416 hugetlb_vma_maps_page(struct vm_area_struct *vma, unsigned long addr, struct page *page) argument
441 vma_offset_start(struct vm_area_struct *vma, pgoff_t start) argument
451 vma_offset_end(struct vm_area_struct *vma, pgoff_t end) argument
475 struct vm_area_struct *vma; local
555 struct vm_area_struct *vma; local
[all...]
/linux-master/fs/
H A Dexec.c203 struct vm_area_struct *vma = bprm->vma; local
212 if (write && pos < vma->vm_start) {
214 ret = expand_downwards(vma, pos);
235 acct_arg_size(bprm, vma_pages(vma));
252 flush_cache_page(bprm->vma, pos, page_to_pfn(page));
258 struct vm_area_struct *vma = NULL; local
261 bprm->vma = vma = vm_area_alloc(mm);
262 if (!vma)
695 shift_arg_pages(struct vm_area_struct *vma, unsigned long shift) argument
768 struct vm_area_struct *vma = bprm->vma; local
[all...]
/linux-master/arch/riscv/mm/
H A Dhugetlbpage.c31 struct vm_area_struct *vma,
57 if (want_pmd_share(vma, addr) && pud_none(pudp_get(pud)))
58 pte = huge_pmd_share(mm, vma, addr, pud);
175 struct vm_area_struct vma = TLB_FLUSH_VMA(mm, 0); local
179 flush_tlb_range(&vma, addr, addr + (PAGE_SIZE * pte_num));
206 struct vm_area_struct vma = TLB_FLUSH_VMA(mm, 0); local
212 flush_tlb_range(&vma, saddr, addr);
263 int huge_ptep_set_access_flags(struct vm_area_struct *vma, argument
269 struct mm_struct *mm = vma->vm_mm;
275 return ptep_set_access_flags(vma, add
30 huge_pte_alloc(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long addr, unsigned long sz) argument
334 huge_ptep_clear_flush(struct vm_area_struct *vma, unsigned long addr, pte_t *ptep) argument
[all...]
/linux-master/drivers/infiniband/hw/hfi1/
H A Dfile_ops.c38 static int hfi1_file_mmap(struct file *fp, struct vm_area_struct *vma);
309 ssize_t memlen, struct vm_area_struct *vma)
315 vma->vm_end - vma->vm_start, vma->vm_flags);
318 static int hfi1_file_mmap(struct file *fp, struct vm_area_struct *vma) argument
324 u64 token = vma->vm_pgoff << PAGE_SHIFT,
334 !(vma->vm_flags & VM_SHARED)) {
351 vma->vm_pgoff = 0;
352 flags = vma
307 mmap_cdbg(u16 ctxt, u8 subctxt, u8 type, u8 mapio, u8 vmf, u64 memaddr, void *memvirt, dma_addr_t memdma, ssize_t memlen, struct vm_area_struct *vma) argument
[all...]

Completed in 469 milliseconds

<<11121314151617181920>>