/linux-master/fs/romfs/ |
H A D | mmap-nommu.c | 64 static int romfs_mmap(struct file *file, struct vm_area_struct *vma) argument 66 return is_nommu_shared_mapping(vma->vm_flags) ? 0 : -ENOSYS;
|
/linux-master/drivers/gpu/drm/nouveau/nvif/ |
H A D | vmm.c | 65 nvif_vmm_put(struct nvif_vmm *vmm, struct nvif_vma *vma) argument 67 if (vma->size) { 70 .addr = vma->addr, 72 vma->size = 0; 78 u8 page, u8 align, u64 size, struct nvif_vma *vma) 83 args.version = vma->size = 0; 101 vma->addr = args.addr; 102 vma->size = args.size; 77 nvif_vmm_get(struct nvif_vmm *vmm, enum nvif_vmm_get type, bool sparse, u8 page, u8 align, u64 size, struct nvif_vma *vma) argument
|
/linux-master/arch/mips/alchemy/common/ |
H A D | setup.c | 97 int io_remap_pfn_range(struct vm_area_struct *vma, unsigned long vaddr, argument 102 return remap_pfn_range(vma, vaddr, phys_addr >> PAGE_SHIFT, size, prot);
|
/linux-master/arch/x86/include/asm/ |
H A D | page.h | 37 #define vma_alloc_zeroed_movable_folio(vma, vaddr) \ 38 vma_alloc_folio(GFP_HIGHUSER_MOVABLE | __GFP_ZERO, 0, vma, vaddr, false)
|
/linux-master/arch/mips/mm/ |
H A D | fault.c | 42 struct vm_area_struct * vma = NULL; local 103 vma = lock_mm_and_find_vma(mm, address, regs); 104 if (!vma) 113 if (!(vma->vm_flags & VM_WRITE)) 118 if (address == regs->cp0_epc && !(vma->vm_flags & VM_EXEC)) { 128 if (!(vma->vm_flags & VM_READ) && 140 if (unlikely(!vma_is_accessible(vma))) 150 fault = handle_mm_fault(vma, address, flags, regs);
|
H A D | tlb-r4k.c | 48 static inline void flush_micro_tlb_vm(struct vm_area_struct *vma) argument 50 if (vma->vm_flags & VM_EXEC) 106 void local_flush_tlb_range(struct vm_area_struct *vma, unsigned long start, argument 109 struct mm_struct *mm = vma->vm_mm; 211 void local_flush_tlb_page(struct vm_area_struct *vma, unsigned long page) argument 215 if (cpu_context(cpu, vma->vm_mm) != 0) { 227 write_c0_memorymapid(cpu_asid(cpu, vma->vm_mm)); 229 write_c0_entryhi(page | cpu_asid(cpu, vma->vm_mm)); 250 flush_micro_tlb_vm(vma); 293 void __update_tlb(struct vm_area_struct * vma, unsigne argument [all...] |
/linux-master/include/linux/ |
H A D | mempolicy.h | 34 * its own state. All vma manipulation is somewhat protected by a down_read on 124 struct vm_area_struct *vma, struct mempolicy *mpol); 130 struct mempolicy *__get_vma_policy(struct vm_area_struct *vma, 132 struct mempolicy *get_vma_policy(struct vm_area_struct *vma, 134 bool vma_policy_mof(struct vm_area_struct *vma); 141 extern int huge_node(struct vm_area_struct *vma, 167 /* Check if a vma is migratable */ 168 extern bool vma_migratable(struct vm_area_struct *vma); 224 static inline struct mempolicy *get_vma_policy(struct vm_area_struct *vma, argument 254 static inline int huge_node(struct vm_area_struct *vma, argument [all...] |
/linux-master/arch/xtensa/mm/ |
H A D | tlb.c | 89 void local_flush_tlb_range(struct vm_area_struct *vma, argument 93 struct mm_struct *mm = vma->vm_mm; 108 if (vma->vm_flags & VM_EXEC) 127 void local_flush_tlb_page(struct vm_area_struct *vma, unsigned long page) argument 130 struct mm_struct* mm = vma->vm_mm; 142 if (vma->vm_flags & VM_EXEC) 166 void update_mmu_tlb(struct vm_area_struct *vma, argument 169 local_flush_tlb_page(vma, address);
|
/linux-master/arch/riscv/include/asm/ |
H A D | hugetlb.h | 34 pte_t huge_ptep_clear_flush(struct vm_area_struct *vma, 42 int huge_ptep_set_access_flags(struct vm_area_struct *vma,
|
/linux-master/arch/x86/kernel/ |
H A D | sys_x86_64.c | 128 struct vm_area_struct *vma; local 142 vma = find_vma(mm, addr); 144 (!vma || addr + len <= vm_start_gap(vma))) 165 struct vm_area_struct *vma; local 188 vma = find_vma(mm, addr); 189 if (!vma || addr + len <= vm_start_gap(vma))
|
/linux-master/arch/s390/mm/ |
H A D | mmap.c | 88 struct vm_area_struct *vma; local 99 vma = find_vma(mm, addr); 101 (!vma || addr + len <= vm_start_gap(vma))) 122 struct vm_area_struct *vma; local 136 vma = find_vma(mm, addr); 138 (!vma || addr + len <= vm_start_gap(vma)))
|
H A D | fault.c | 284 struct vm_area_struct *vma; local 323 vma = lock_vma_under_rcu(mm, address); 324 if (!vma) 326 if (!(vma->vm_flags & access)) { 327 vma_end_read(vma); 331 fault = handle_mm_fault(vma, address, flags | FAULT_FLAG_VMA_LOCK, regs); 333 vma_end_read(vma); 365 vma = find_vma(mm, address); 366 if (!vma) 368 if (unlikely(vma 493 struct vm_area_struct *vma; local [all...] |
/linux-master/arch/sparc/mm/ |
H A D | fault_64.c | 271 struct vm_area_struct *vma; local 338 vma = find_vma(mm, address); 339 if (!vma) 352 (vma->vm_flags & VM_WRITE) != 0) { 371 if (vma->vm_start <= address) 373 if (!(vma->vm_flags & VM_GROWSDOWN)) 389 vma = expand_stack(mm, address); 390 if (!vma) 402 if ((fault_code & FAULT_CODE_ITLB) && !(vma->vm_flags & VM_EXEC)) { 410 if (!(vma [all...] |
H A D | leon_mm.c | 194 void leon_flush_pcache_all(struct vm_area_struct *vma, unsigned long page) argument 196 if (vma->vm_flags & VM_EXEC) 282 static void leon_flush_cache_page(struct vm_area_struct *vma, unsigned long page) argument 284 leon_flush_pcache_all(vma, page); 287 static void leon_flush_cache_range(struct vm_area_struct *vma, argument 299 static void leon_flush_tlb_page(struct vm_area_struct *vma, argument 305 static void leon_flush_tlb_range(struct vm_area_struct *vma, argument
|
/linux-master/drivers/gpu/drm/rockchip/ |
H A D | rockchip_drm_gem.c | 216 struct vm_area_struct *vma) 220 unsigned long user_count = vma_pages(vma); 225 return vm_map_pages(vma, rk_obj->pages, count); 229 struct vm_area_struct *vma) 234 return dma_mmap_attrs(drm->dev, vma, rk_obj->kvaddr, rk_obj->dma_addr, 239 struct vm_area_struct *vma) 248 vma->vm_pgoff = 0; 254 vm_flags_mod(vma, VM_IO | VM_DONTEXPAND | VM_DONTDUMP, VM_PFNMAP); 256 vma->vm_page_prot = pgprot_writecombine(vm_get_page_prot(vma 215 rockchip_drm_gem_object_mmap_iommu(struct drm_gem_object *obj, struct vm_area_struct *vma) argument 228 rockchip_drm_gem_object_mmap_dma(struct drm_gem_object *obj, struct vm_area_struct *vma) argument 238 rockchip_drm_gem_object_mmap(struct drm_gem_object *obj, struct vm_area_struct *vma) argument [all...] |
/linux-master/drivers/xen/ |
H A D | gntalloc.c | 448 static void gntalloc_vma_open(struct vm_area_struct *vma) argument 450 struct gntalloc_vma_private_data *priv = vma->vm_private_data; 460 static void gntalloc_vma_close(struct vm_area_struct *vma) argument 462 struct gntalloc_vma_private_data *priv = vma->vm_private_data; 491 static int gntalloc_mmap(struct file *filp, struct vm_area_struct *vma) argument 496 int count = vma_pages(vma); 499 if (!(vma->vm_flags & VM_SHARED)) { 511 priv, vm_priv, vma->vm_pgoff, count); 513 gref = find_grefs(priv, vma->vm_pgoff << PAGE_SHIFT, count); 526 vma [all...] |
H A D | gntdev.c | 499 static void gntdev_vma_open(struct vm_area_struct *vma) argument 501 struct gntdev_grant_map *map = vma->vm_private_data; 503 pr_debug("gntdev_vma_open %p\n", vma); 507 static void gntdev_vma_close(struct vm_area_struct *vma) argument 509 struct gntdev_grant_map *map = vma->vm_private_data; 510 struct file *file = vma->vm_file; 513 pr_debug("gntdev_vma_close %p\n", vma); 515 vma->vm_private_data = NULL; 519 static struct page *gntdev_vma_find_special_page(struct vm_area_struct *vma, argument 522 struct gntdev_grant_map *map = vma 693 struct vm_area_struct *vma; local 1033 gntdev_mmap(struct file *flip, struct vm_area_struct *vma) argument [all...] |
/linux-master/arch/arc/mm/ |
H A D | fault.c | 75 struct vm_area_struct *vma = NULL; local 117 vma = lock_mm_and_find_vma(mm, address, regs); 118 if (!vma) 130 if (!(vma->vm_flags & mask)) { 135 fault = handle_mm_fault(vma, address, flags, regs);
|
/linux-master/drivers/gpu/drm/i915/gt/uc/ |
H A D | intel_guc_log.c | 297 relay_reserve(log->relay.channel, log->vma->obj->base.size - 492 if (!log->vma || !log->buf_addr) 499 i915_gem_object_get(log->vma->obj); 509 i915_gem_object_put(log->vma->obj); 529 GEM_BUG_ON(!log->vma); 535 subbuf_size = log->vma->size - intel_guc_log_section_size_capture(log); 617 struct i915_vma *vma; local 622 GEM_BUG_ON(log->vma); 626 vma = intel_guc_allocate_vma(guc, guc_log_size); 627 if (IS_ERR(vma)) { [all...] |
/linux-master/kernel/dma/ |
H A D | coherent.c | 235 struct vm_area_struct *vma, void *vaddr, size_t size, int *ret) 239 unsigned long off = vma->vm_pgoff; 241 unsigned long user_count = vma_pages(vma); 247 *ret = remap_pfn_range(vma, vma->vm_start, pfn, 249 vma->vm_page_prot); 259 * @vma: vm_area for the userspace memory 265 * coherent memory pool and if so, maps that memory to the provided vma. 271 int dma_mmap_from_dev_coherent(struct device *dev, struct vm_area_struct *vma, argument 276 return __dma_mmap_from_coherent(mem, vma, vadd 234 __dma_mmap_from_coherent(struct dma_coherent_mem *mem, struct vm_area_struct *vma, void *vaddr, size_t size, int *ret) argument 301 dma_mmap_from_global_coherent(struct vm_area_struct *vma, void *vaddr, size_t size, int *ret) argument [all...] |
/linux-master/ipc/ |
H A D | shm.c | 102 static void shm_open(struct vm_area_struct *vma); 103 static void shm_close(struct vm_area_struct *vma); 302 static void shm_open(struct vm_area_struct *vma) argument 304 struct file *file = vma->vm_file; 310 sfd->vm_ops->open(vma); 364 * remove the attach descriptor vma. 396 static void shm_close(struct vm_area_struct *vma) argument 398 struct file *file = vma->vm_file; 403 sfd->vm_ops->close(vma); 537 struct file *file = vmf->vma 543 shm_may_split(struct vm_area_struct *vma, unsigned long addr) argument 554 shm_pagesize(struct vm_area_struct *vma) argument 566 shm_set_policy(struct vm_area_struct *vma, struct mempolicy *mpol) argument 576 shm_get_policy(struct vm_area_struct *vma, unsigned long addr, pgoff_t *ilx) argument 588 shm_mmap(struct file *file, struct vm_area_struct *vma) argument 1726 struct vm_area_struct *vma; local [all...] |
/linux-master/fs/hugetlbfs/ |
H A D | inode.c | 96 static int hugetlbfs_file_mmap(struct file *file, struct vm_area_struct *vma) argument 106 * vma address alignment (but not the pgoff alignment) has 113 vm_flags_set(vma, VM_HUGETLB | VM_DONTEXPAND); 114 vma->vm_ops = &hugetlb_vm_ops; 116 ret = seal_check_write(info->seals, vma); 127 if (vma->vm_pgoff & PGOFF_LOFFT_MAX) 132 if (vma->vm_pgoff & (~huge_page_mask(h) >> PAGE_SHIFT)) 135 vma_len = (loff_t)(vma->vm_end - vma->vm_start); 136 len = vma_len + ((loff_t)vma 225 struct vm_area_struct *vma; local 416 hugetlb_vma_maps_page(struct vm_area_struct *vma, unsigned long addr, struct page *page) argument 441 vma_offset_start(struct vm_area_struct *vma, pgoff_t start) argument 451 vma_offset_end(struct vm_area_struct *vma, pgoff_t end) argument 475 struct vm_area_struct *vma; local 555 struct vm_area_struct *vma; local [all...] |
/linux-master/fs/ |
H A D | exec.c | 203 struct vm_area_struct *vma = bprm->vma; local 212 if (write && pos < vma->vm_start) { 214 ret = expand_downwards(vma, pos); 235 acct_arg_size(bprm, vma_pages(vma)); 252 flush_cache_page(bprm->vma, pos, page_to_pfn(page)); 258 struct vm_area_struct *vma = NULL; local 261 bprm->vma = vma = vm_area_alloc(mm); 262 if (!vma) 695 shift_arg_pages(struct vm_area_struct *vma, unsigned long shift) argument 768 struct vm_area_struct *vma = bprm->vma; local [all...] |
/linux-master/arch/riscv/mm/ |
H A D | hugetlbpage.c | 31 struct vm_area_struct *vma, 57 if (want_pmd_share(vma, addr) && pud_none(pudp_get(pud))) 58 pte = huge_pmd_share(mm, vma, addr, pud); 175 struct vm_area_struct vma = TLB_FLUSH_VMA(mm, 0); local 179 flush_tlb_range(&vma, addr, addr + (PAGE_SIZE * pte_num)); 206 struct vm_area_struct vma = TLB_FLUSH_VMA(mm, 0); local 212 flush_tlb_range(&vma, saddr, addr); 263 int huge_ptep_set_access_flags(struct vm_area_struct *vma, argument 269 struct mm_struct *mm = vma->vm_mm; 275 return ptep_set_access_flags(vma, add 30 huge_pte_alloc(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long addr, unsigned long sz) argument 334 huge_ptep_clear_flush(struct vm_area_struct *vma, unsigned long addr, pte_t *ptep) argument [all...] |
/linux-master/drivers/infiniband/hw/hfi1/ |
H A D | file_ops.c | 38 static int hfi1_file_mmap(struct file *fp, struct vm_area_struct *vma); 309 ssize_t memlen, struct vm_area_struct *vma) 315 vma->vm_end - vma->vm_start, vma->vm_flags); 318 static int hfi1_file_mmap(struct file *fp, struct vm_area_struct *vma) argument 324 u64 token = vma->vm_pgoff << PAGE_SHIFT, 334 !(vma->vm_flags & VM_SHARED)) { 351 vma->vm_pgoff = 0; 352 flags = vma 307 mmap_cdbg(u16 ctxt, u8 subctxt, u8 type, u8 mapio, u8 vmf, u64 memaddr, void *memvirt, dma_addr_t memdma, ssize_t memlen, struct vm_area_struct *vma) argument [all...] |