/linux-master/kernel/sched/ |
H A D | fair.c | 3172 static bool vma_is_accessed(struct mm_struct *mm, struct vm_area_struct *vma) argument 3181 if ((READ_ONCE(current->mm->numa_scan_seq) - vma->numab_state->start_scan_seq) < 2) 3184 pids = vma->numab_state->pids_active[0] | vma->numab_state->pids_active[1]; 3192 if (mm->numa_scan_offset > vma->vm_start) { 3193 trace_sched_skip_vma_numa(mm, vma, NUMAB_SKIP_IGNORE_PID); 3212 struct vm_area_struct *vma; local 3281 vma = vma_next(&vmi); 3282 if (!vma) { 3286 vma [all...] |
/linux-master/drivers/dma/idxd/ |
H A D | cdev.c | 376 static int check_vma(struct idxd_wq *wq, struct vm_area_struct *vma, argument 381 if ((vma->vm_end - vma->vm_start) > PAGE_SIZE) { 385 vma->vm_end - vma->vm_start); 392 static int idxd_cdev_mmap(struct file *filp, struct vm_area_struct *vma) argument 403 rc = check_vma(wq, vma, __func__); 407 vm_flags_set(vma, VM_DONTCOPY); 410 vma->vm_page_prot = pgprot_noncached(vma [all...] |
/linux-master/arch/riscv/include/asm/ |
H A D | pgtable.h | 473 struct vm_area_struct *vma, unsigned long address, 486 #define update_mmu_cache(vma, addr, ptep) \ 487 update_mmu_cache_range(NULL, vma, addr, ptep, 1) 492 static inline void update_mmu_cache_pmd(struct vm_area_struct *vma, argument 497 update_mmu_cache(vma, address, ptep); 550 extern int ptep_set_access_flags(struct vm_area_struct *vma, unsigned long address, 553 extern int ptep_test_and_clear_young(struct vm_area_struct *vma, unsigned long address, 575 static inline int ptep_clear_flush_young(struct vm_area_struct *vma, argument 593 return ptep_test_and_clear_young(vma, address, ptep); 758 static inline int pmdp_set_access_flags(struct vm_area_struct *vma, argument 472 update_mmu_cache_range(struct vm_fault *vmf, struct vm_area_struct *vma, unsigned long address, pte_t *ptep, unsigned int nr) argument 766 pmdp_test_and_clear_young(struct vm_area_struct *vma, unsigned long address, pmd_t *pmdp) argument 791 pmdp_establish(struct vm_area_struct *vma, unsigned long address, pmd_t *pmdp, pmd_t pmd) argument [all...] |
/linux-master/mm/ |
H A D | hugetlb.c | 96 static void hugetlb_vma_lock_free(struct vm_area_struct *vma); 97 static void hugetlb_vma_lock_alloc(struct vm_area_struct *vma); 98 static void __hugetlb_vma_unlock_write_free(struct vm_area_struct *vma); 99 static void hugetlb_unshare_pmds(struct vm_area_struct *vma, 101 static struct resv_map *vma_resv_map(struct vm_area_struct *vma); 258 static inline struct hugepage_subpool *subpool_vma(struct vm_area_struct *vma) argument 260 return subpool_inode(file_inode(vma->vm_file)); 266 void hugetlb_vma_lock_read(struct vm_area_struct *vma) argument 268 if (__vma_shareable_lock(vma)) { 269 struct hugetlb_vma_lock *vma_lock = vma 279 hugetlb_vma_unlock_read(struct vm_area_struct *vma) argument 292 hugetlb_vma_lock_write(struct vm_area_struct *vma) argument 305 hugetlb_vma_unlock_write(struct vm_area_struct *vma) argument 318 hugetlb_vma_trylock_write(struct vm_area_struct *vma) argument 334 hugetlb_vma_assert_locked(struct vm_area_struct *vma) argument 357 struct vm_area_struct *vma = vma_lock->vma; local 370 __hugetlb_vma_unlock_write_free(struct vm_area_struct *vma) argument 384 hugetlb_vma_lock_free(struct vm_area_struct *vma) argument 400 hugetlb_vma_lock_alloc(struct vm_area_struct *vma) argument 990 vma_hugecache_offset(struct hstate *h, struct vm_area_struct *vma, unsigned long address) argument 1006 vma_kernel_pagesize(struct vm_area_struct *vma) argument 1020 vma_mmu_pagesize(struct vm_area_struct *vma) argument 1053 get_vma_private_data(struct vm_area_struct *vma) argument 1058 set_vma_private_data(struct vm_area_struct *vma, unsigned long value) argument 1148 vma_resv_map(struct vm_area_struct *vma) argument 1163 set_vma_resv_map(struct vm_area_struct *vma, struct resv_map *map) argument 1171 set_vma_resv_flags(struct vm_area_struct *vma, unsigned long flags) argument 1179 is_vma_resv_set(struct vm_area_struct *vma, unsigned long flag) argument 1186 __vma_private_lock(struct vm_area_struct *vma) argument 1193 hugetlb_dup_vma_private(struct vm_area_struct *vma) argument 1224 clear_vma_resv_huge_pages(struct vm_area_struct *vma) argument 1249 vma_has_reserves(struct vm_area_struct *vma, long chg) argument 1390 dequeue_hugetlb_folio_vma(struct hstate *h, struct vm_area_struct *vma, unsigned long address, int avoid_reserve, long chg) argument 2580 alloc_buddy_hugetlb_folio_with_mpol(struct hstate *h, struct vm_area_struct *vma, unsigned long addr) argument 2802 __vma_reservation_common(struct hstate *h, struct vm_area_struct *vma, unsigned long addr, enum vma_resv_mode mode) argument 2882 vma_needs_reservation(struct hstate *h, struct vm_area_struct *vma, unsigned long addr) argument 2888 vma_commit_reservation(struct hstate *h, struct vm_area_struct *vma, unsigned long addr) argument 2894 vma_end_reservation(struct hstate *h, struct vm_area_struct *vma, unsigned long addr) argument 2900 vma_add_reservation(struct hstate *h, struct vm_area_struct *vma, unsigned long addr) argument 2906 vma_del_reservation(struct hstate *h, struct vm_area_struct *vma, unsigned long addr) argument 2932 restore_reserve_on_error(struct hstate *h, struct vm_area_struct *vma, unsigned long address, struct folio *folio) argument 3132 alloc_hugetlb_folio(struct vm_area_struct *vma, unsigned long addr, int avoid_reserve) argument 5173 hugetlb_vm_op_open(struct vm_area_struct *vma) argument 5211 hugetlb_vm_op_close(struct vm_area_struct *vma) argument 5242 hugetlb_vm_op_split(struct vm_area_struct *vma, unsigned long addr) argument 5268 hugetlb_vm_op_pagesize(struct vm_area_struct *vma) argument 5300 make_huge_pte(struct vm_area_struct *vma, struct page *page, int writable) argument 5319 set_huge_ptep_writable(struct vm_area_struct *vma, unsigned long address, pte_t *ptep) argument 5356 hugetlb_install_folio(struct vm_area_struct *vma, pte_t *ptep, unsigned long addr, struct folio *new_folio, pte_t old, unsigned long sz) argument 5559 move_huge_pte(struct vm_area_struct *vma, unsigned long old_addr, unsigned long new_addr, pte_t *src_pte, pte_t *dst_pte, unsigned long sz) argument 5586 move_hugetlb_page_tables(struct vm_area_struct *vma, struct vm_area_struct *new_vma, unsigned long old_addr, unsigned long new_addr, unsigned long len) argument 5650 __unmap_hugepage_range(struct mmu_gather *tlb, struct vm_area_struct *vma, unsigned long start, unsigned long end, struct page *ref_page, zap_flags_t zap_flags) argument 5806 __hugetlb_zap_begin(struct vm_area_struct *vma, unsigned long *start, unsigned long *end) argument 5818 __hugetlb_zap_end(struct vm_area_struct *vma, struct zap_details *details) argument 5845 unmap_hugepage_range(struct vm_area_struct *vma, unsigned long start, unsigned long end, struct page *ref_page, zap_flags_t zap_flags) argument 5870 unmap_ref_private(struct mm_struct *mm, struct vm_area_struct *vma, struct page *page, unsigned long address) argument 5926 hugetlb_wp(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long address, pte_t *ptep, unsigned int flags, struct folio *pagecache_folio, spinlock_t *ptl, struct vm_fault *vmf) argument 6119 hugetlbfs_pagecache_present(struct hstate *h, struct vm_area_struct *vma, unsigned long address) argument 6196 hugetlb_no_page(struct mm_struct *mm, struct vm_area_struct *vma, struct address_space *mapping, pgoff_t idx, unsigned long address, pte_t *ptep, pte_t old_pte, unsigned int flags, struct vm_fault *vmf) argument 6432 hugetlb_fault(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long address, unsigned int flags) argument 6635 alloc_hugetlb_folio_vma(struct hstate *h, struct vm_area_struct *vma, unsigned long address) argument 6876 hugetlb_follow_page_mask(struct vm_area_struct *vma, unsigned long address, unsigned int flags, unsigned int *page_mask) argument 6947 hugetlb_change_protection(struct vm_area_struct *vma, unsigned long address, unsigned long end, pgprot_t newprot, unsigned long cp_flags) argument 7095 hugetlb_reserve_pages(struct inode *inode, long from, long to, struct vm_area_struct *vma, vm_flags_t vm_flags) argument 7297 page_table_shareable(struct vm_area_struct *svma, struct vm_area_struct *vma, unsigned long addr, pgoff_t idx) argument 7325 want_pmd_share(struct vm_area_struct *vma, unsigned long addr) argument 7351 adjust_range_if_pmd_sharing_possible(struct vm_area_struct *vma, unsigned long *start, unsigned long *end) argument 7382 huge_pmd_share(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long addr, pud_t *pud) argument 7439 huge_pmd_unshare(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long addr, pte_t *ptep) argument 7460 huge_pmd_share(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long addr, pud_t *pud) argument 7466 huge_pmd_unshare(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long addr, pte_t *ptep) argument 7472 adjust_range_if_pmd_sharing_possible(struct vm_area_struct *vma, unsigned long *start, unsigned long *end) argument 7477 want_pmd_share(struct vm_area_struct *vma, unsigned long addr) argument 7484 huge_pte_alloc(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long addr, unsigned long sz) argument 7689 hugetlb_unshare_pmds(struct vm_area_struct *vma, unsigned long start, unsigned long end) argument 7739 hugetlb_unshare_all_pmds(struct vm_area_struct *vma) argument [all...] |
H A D | shmem.c | 272 bool vma_is_anon_shmem(struct vm_area_struct *vma) argument 274 return vma->vm_ops == &shmem_anon_vm_ops; 277 bool vma_is_shmem(struct vm_area_struct *vma) argument 279 return vma_is_anon_shmem(vma) || vma->vm_ops == &shmem_vm_ops; 881 * given vma is swapped out. 886 unsigned long shmem_swap_usage(struct vm_area_struct *vma) argument 888 struct inode *inode = file_inode(vma->vm_file); 898 * the vma maps it whole. Then we can simply use the stats that we 904 if (!vma 1963 struct vm_area_struct *vma = vmf ? vmf->vma : NULL; local 2356 shmem_set_policy(struct vm_area_struct *vma, struct mempolicy *mpol) argument 2362 shmem_get_policy(struct vm_area_struct *vma, unsigned long addr, pgoff_t *ilx) argument 2427 shmem_mmap(struct file *file, struct vm_area_struct *vma) argument 4914 shmem_zero_setup(struct vm_area_struct *vma) argument [all...] |
H A D | memory-failure.c | 388 static unsigned long dev_pagemap_mapping_shift(struct vm_area_struct *vma, argument 399 VM_BUG_ON_VMA(address == -EFAULT, vma); 400 pgd = pgd_offset(vma->vm_mm, address); 444 struct vm_area_struct *vma, struct list_head *to_kill, 455 tk->addr = ksm_addr ? ksm_addr : page_address_in_vma(p, vma); 458 tk->addr = vma_pgoff_address(fsdax_pgoff, 1, vma); 459 tk->size_shift = dev_pagemap_mapping_shift(vma, tk->addr); 487 struct vm_area_struct *vma, 490 __add_to_kill(tsk, p, vma, to_kill, 0, FSDAX_INVALID_PGOFF); 507 struct vm_area_struct *vma, struc 443 __add_to_kill(struct task_struct *tsk, struct page *p, struct vm_area_struct *vma, struct list_head *to_kill, unsigned long ksm_addr, pgoff_t fsdax_pgoff) argument 486 add_to_kill_anon_file(struct task_struct *tsk, struct page *p, struct vm_area_struct *vma, struct list_head *to_kill) argument 506 add_to_kill_ksm(struct task_struct *tsk, struct page *p, struct vm_area_struct *vma, struct list_head *to_kill, unsigned long ksm_addr) argument 613 struct vm_area_struct *vma; local 650 struct vm_area_struct *vma; local 681 add_to_kill_fsdax(struct task_struct *tsk, struct page *p, struct vm_area_struct *vma, struct list_head *to_kill, pgoff_t pgoff) argument 695 struct vm_area_struct *vma; local [all...] |
H A D | madvise.c | 46 * Any behaviour which results in changes to the vma->vm_flags needs to 94 struct anon_vma_name *anon_vma_name(struct vm_area_struct *vma) argument 96 mmap_assert_locked(vma->vm_mm); 98 return vma->anon_name; 102 static int replace_anon_vma_name(struct vm_area_struct *vma, argument 105 struct anon_vma_name *orig_name = anon_vma_name(vma); 108 vma->anon_name = NULL; 116 vma->anon_name = anon_vma_name_reuse(anon_name); 122 static int replace_anon_vma_name(struct vm_area_struct *vma, argument 132 * Update the vm_flags on region of a vma, splittin 137 madvise_update_vma(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end, unsigned long new_flags, struct anon_vma_name *anon_name) argument 174 struct vm_area_struct *vma = walk->private; local 220 shmem_swapin_range(struct vm_area_struct *vma, unsigned long start, unsigned long end, struct address_space *mapping) argument 261 madvise_willneed(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end) argument 309 can_do_file_pageout(struct vm_area_struct *vma) argument 332 struct vm_area_struct *vma = walk->vma; local 538 madvise_cold_page_range(struct mmu_gather *tlb, struct vm_area_struct *vma, unsigned long addr, unsigned long end) argument 552 can_madv_lru_vma(struct vm_area_struct *vma) argument 557 madvise_cold(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start_addr, unsigned long end_addr) argument 576 madvise_pageout_page_range(struct mmu_gather *tlb, struct vm_area_struct *vma, unsigned long addr, unsigned long end) argument 590 madvise_pageout(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start_addr, unsigned long end_addr) argument 625 struct vm_area_struct *vma = walk->vma; local 759 madvise_free_single_vma(struct vm_area_struct *vma, unsigned long start_addr, unsigned long end_addr) argument 813 madvise_dontneed_single_vma(struct vm_area_struct *vma, unsigned long start, unsigned long end) argument 820 madvise_dontneed_free_valid_vma(struct vm_area_struct *vma, unsigned long start, unsigned long *end, int behavior) argument 850 madvise_dontneed_free(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end, int behavior) argument 904 madvise_populate(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end, int behavior) argument 952 madvise_remove(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end) argument 1002 madvise_vma_behavior(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end, unsigned long behavior) argument 1207 madvise_walk_vmas(struct mm_struct *mm, unsigned long start, unsigned long end, unsigned long arg, int (*visit)(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end, unsigned long arg)) argument 1213 struct vm_area_struct *vma; local 1266 madvise_vma_anon_name(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end, unsigned long anon_name) argument [all...] |
H A D | internal.h | 213 struct vm_area_struct *vma, 687 extern long populate_vma_page_range(struct vm_area_struct *vma, 710 folio_within_range(struct folio *folio, struct vm_area_struct *vma, argument 714 unsigned long vma_pglen = vma_pages(vma); 720 if (start < vma->vm_start) 721 start = vma->vm_start; 723 if (end > vma->vm_end) 724 end = vma->vm_end; 728 /* if folio start address is not in vma range */ 729 if (!in_range(pgoff, vma 738 folio_within_vma(struct folio *folio, struct vm_area_struct *vma) argument 753 mlock_vma_folio(struct folio *folio, struct vm_area_struct *vma) argument 769 munlock_vma_folio(struct folio *folio, struct vm_area_struct *vma) argument 797 vma_pgoff_address(pgoff_t pgoff, unsigned long nr_pages, struct vm_area_struct *vma) argument 823 vma_address(struct page *page, struct vm_area_struct *vma) argument 835 struct vm_area_struct *vma = pvmw->vma; local 1157 gup_must_unshare(struct vm_area_struct *vma, unsigned int flags, struct page *page) argument 1215 vma_set_range(struct vm_area_struct *vma, unsigned long start, unsigned long end, pgoff_t pgoff) argument 1224 vma_soft_dirty_enabled(struct vm_area_struct *vma) argument 1251 vma_iter_prealloc(struct vma_iterator *vmi, struct vm_area_struct *vma) argument 1268 vma_iter_store(struct vma_iterator *vmi, struct vm_area_struct *vma) argument 1295 vma_iter_store_gfp(struct vma_iterator *vmi, struct vm_area_struct *vma, gfp_t gfp) argument 1314 struct vm_area_struct *vma; member in struct:vma_prepare [all...] |
/linux-master/include/net/ |
H A D | sock.h | 1894 struct vm_area_struct *vma);
|
/linux-master/include/linux/ |
H A D | mm.h | 184 * per a vma. In ELF, the number of sections is represented in unsigned short. 251 void __vm_area_free(struct vm_area_struct *vma); 294 #define VM_DONTCOPY 0x00020000 /* Do not copy this vma on fork */ 312 #define VM_HUGEPAGE 0x20000000 /* MADV_HUGEPAGE marked this vma */ 313 #define VM_NOHUGEPAGE 0x40000000 /* MADV_NOHUGEPAGE marked this vma */ 511 * vm_fault is filled by the pagefault handler and passed to the vma's 512 * ->fault function. The vma's ->fault is responsible for returning a bitmask 522 struct vm_area_struct *vma; /* Target VMA */ member in struct:vm_fault::__anon82 524 pgoff_t pgoff; /* Logical page offset based on vma */ 586 int (*mprotect)(struct vm_area_struct *vma, unsigne 646 vma_numab_state_init(struct vm_area_struct *vma) argument 650 vma_numab_state_free(struct vm_area_struct *vma) argument 655 vma_numab_state_init(struct vm_area_struct *vma) argument 656 vma_numab_state_free(struct vm_area_struct *vma) argument 665 vma_start_read(struct vm_area_struct *vma) argument 698 vma_end_read(struct vm_area_struct *vma) argument 706 __is_vma_write_locked(struct vm_area_struct *vma, int *mm_lock_seq) argument 723 vma_start_write(struct vm_area_struct *vma) argument 741 vma_assert_write_locked(struct vm_area_struct *vma) argument 748 vma_assert_locked(struct vm_area_struct *vma) argument 754 vma_mark_detached(struct vm_area_struct *vma, bool detached) argument 783 vma_start_read(struct vm_area_struct *vma) argument 785 vma_end_read(struct vm_area_struct *vma) argument 786 vma_start_write(struct vm_area_struct *vma) argument 787 vma_assert_write_locked(struct vm_area_struct *vma) argument 789 vma_mark_detached(struct vm_area_struct *vma, bool detached) argument 798 vma_assert_locked(struct vm_area_struct *vma) argument 821 vma_init(struct vm_area_struct *vma, struct mm_struct *mm) argument 832 vm_flags_init(struct vm_area_struct *vma, vm_flags_t flags) argument 843 vm_flags_reset(struct vm_area_struct *vma, vm_flags_t flags) argument 850 vm_flags_reset_once(struct vm_area_struct *vma, vm_flags_t flags) argument 857 vm_flags_set(struct vm_area_struct *vma, vm_flags_t flags) argument 864 vm_flags_clear(struct vm_area_struct *vma, vm_flags_t flags) argument 875 __vm_flags_mod(struct vm_area_struct *vma, vm_flags_t set, vm_flags_t clear) argument 885 vm_flags_mod(struct vm_area_struct *vma, vm_flags_t set, vm_flags_t clear) argument 892 vma_set_anonymous(struct vm_area_struct *vma) argument 897 vma_is_anonymous(struct vm_area_struct *vma) argument 906 vma_is_initial_heap(const struct vm_area_struct *vma) argument 916 vma_is_initial_stack(const struct vm_area_struct *vma) argument 927 vma_is_temporary_stack(struct vm_area_struct *vma) argument 941 vma_is_foreign(struct vm_area_struct *vma) argument 952 vma_is_accessible(struct vm_area_struct *vma) argument 963 vma_is_shared_maywrite(struct vm_area_struct *vma) argument 1033 vma_iter_bulk_store(struct vma_iterator *vmi, struct vm_area_struct *vma) argument 1070 vma_is_shmem(struct vm_area_struct *vma) argument 1071 vma_is_anon_shmem(struct vm_area_struct *vma) argument 1362 maybe_mkwrite(pte_t pte, struct vm_area_struct *vma) argument 1759 vma_set_access_pid_bit(struct vm_area_struct *vma) argument 1818 vma_set_access_pid_bit(struct vm_area_struct *vma) argument 1982 folio_needs_cow_for_dma(struct vm_area_struct *vma, struct folio *folio) argument 2381 zap_vma_pages(struct vm_area_struct *vma) argument 2427 handle_mm_fault(struct vm_area_struct *vma, unsigned long address, unsigned int flags, struct pt_regs *regs) argument 2480 struct vm_area_struct *vma; local 2556 vma_wants_manual_pte_write_upgrade(struct vm_area_struct *vma) argument 3299 vma_modify_flags(struct vma_iterator *vmi, struct vm_area_struct *prev, struct vm_area_struct *vma, unsigned long start, unsigned long end, unsigned long new_flags) argument 3312 vma_modify_flags_name(struct vma_iterator *vmi, struct vm_area_struct *prev, struct vm_area_struct *vma, unsigned long start, unsigned long end, unsigned long new_flags, struct anon_vma_name *new_name) argument 3326 vma_modify_policy(struct vma_iterator *vmi, struct vm_area_struct *prev, struct vm_area_struct *vma, unsigned long start, unsigned long end, struct mempolicy *new_pol) argument 3338 vma_modify_flags_uffd(struct vma_iterator *vmi, struct vm_area_struct *prev, struct vm_area_struct *vma, unsigned long start, unsigned long end, unsigned long new_flags, struct vm_userfaultfd_ctx new_ctx) argument 3483 stack_guard_start_gap(struct vm_area_struct *vma) argument 3495 vm_start_gap(struct vm_area_struct *vma) argument 3506 vm_end_gap(struct vm_area_struct *vma) argument 3518 vma_pages(struct vm_area_struct *vma) argument 3527 struct vm_area_struct *vma = vma_lookup(mm, vm_start); local 3535 range_in_vma(struct vm_area_struct *vma, unsigned long start, unsigned long end) argument 3549 vma_set_page_prot(struct vm_area_struct *vma) argument 3585 vmf_insert_page(struct vm_area_struct *vma, unsigned long addr, struct page *page) argument 3599 io_remap_pfn_range(struct vm_area_struct *vma, unsigned long addr, unsigned long pfn, unsigned long size, pgprot_t prot) argument 3652 gup_can_follow_protnone(struct vm_area_struct *vma, unsigned int flags) argument 4095 vma_is_special_huge(const struct vm_area_struct *vma) argument 4145 seal_check_write(int seals, struct vm_area_struct *vma) argument [all...] |
/linux-master/fs/ntfs3/ |
H A D | file.c | 254 static int ntfs_file_mmap(struct file *file, struct vm_area_struct *vma) argument 259 u64 from = ((u64)vma->vm_pgoff << PAGE_SHIFT); 260 bool rw = vma->vm_flags & VM_WRITE; 283 from + vma->vm_end - vma->vm_start); 314 err = generic_file_mmap(file, vma);
|
/linux-master/fs/netfs/ |
H A D | buffered_write.c | 509 struct file *file = vmf->vma->vm_file;
|
/linux-master/drivers/video/fbdev/core/ |
H A D | fb_defio.c | 98 struct fb_info *info = vmf->vma->vm_private_data; 110 if (vmf->vma->vm_file) 111 page->mapping = vmf->vma->vm_file->f_mapping; 202 file_update_time(vmf->vma->vm_file); 210 struct fb_info *info = vmf->vma->vm_private_data; 224 int fb_deferred_io_mmap(struct fb_info *info, struct vm_area_struct *vma) argument 226 vma->vm_page_prot = pgprot_decrypted(vma->vm_page_prot); 228 vma->vm_ops = &fb_deferred_io_vm_ops; 229 vm_flags_set(vma, VM_DONTEXPAN [all...] |
/linux-master/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_svm.c | 1664 struct vm_area_struct *vma; local 1670 vma = vma_lookup(mm, addr); 1671 if (vma) { 1672 readonly = !(vma->vm_flags & VM_WRITE); 1674 next = min(vma->vm_end, end); 2657 struct vm_area_struct *vma; local 2662 vma = vma_lookup(p->mm, addr << PAGE_SHIFT); 2663 if (!vma) { 2668 *is_heap_stack = vma_is_initial_heap(vma) || vma_is_initial_stack(vma); 2873 svm_fault_allowed(struct vm_area_struct *vma, bool write_fault) argument 2900 struct vm_area_struct *vma; local 3275 struct vm_area_struct *vma; local [all...] |
H A D | kfd_process.c | 2062 struct vm_area_struct *vma) 2067 if ((vma->vm_end - vma->vm_start) != KFD_CWSR_TBA_TMA_SIZE) { 2084 vm_flags_set(vma, VM_IO | VM_DONTCOPY | VM_DONTEXPAND 2087 return remap_pfn_range(vma, vma->vm_start, 2089 KFD_CWSR_TBA_TMA_SIZE, vma->vm_page_prot); 2053 kfd_reserved_mem_mmap(struct kfd_node *dev, struct kfd_process *process, struct vm_area_struct *vma) argument
|
H A D | kfd_migrate.c | 243 svm_migrate_get_sys_page(struct vm_area_struct *vma, unsigned long addr) argument 247 page = alloc_page_vma(GFP_HIGHUSER, vma, addr); 382 struct vm_area_struct *vma, uint64_t start, 398 migrate.vma = vma; 421 dev_err(adev->dev, "%s: vma setup fail %d range [0x%lx 0x%lx]\n", 487 struct vm_area_struct *vma; local 531 vma = vma_lookup(mm, addr); 532 if (!vma) 535 next = min(vma 381 svm_migrate_vma_to_vram(struct kfd_node *node, struct svm_range *prange, struct vm_area_struct *vma, uint64_t start, uint64_t end, uint32_t trigger, uint64_t ttm_res_offset) argument 680 svm_migrate_vma_to_ram(struct kfd_node *node, struct svm_range *prange, struct vm_area_struct *vma, uint64_t start, uint64_t end, uint32_t trigger, struct page *fault_page) argument 792 struct vm_area_struct *vma; local [all...] |
/linux-master/arch/loongarch/mm/ |
H A D | fault.c | 141 struct vm_area_struct *vma = NULL; local 178 vma = lock_mm_and_find_vma(mm, address, regs); 179 if (unlikely(!vma)) 202 if (!(vma->vm_flags & VM_WRITE)) 205 if (!(vma->vm_flags & VM_EXEC) && address == exception_era(regs)) 207 if (!(vma->vm_flags & (VM_READ | VM_WRITE)) && address != exception_era(regs)) 216 fault = handle_mm_fault(vma, address, flags, regs);
|
/linux-master/arch/loongarch/include/asm/ |
H A D | tlb.h | 142 struct vm_area_struct vma; local 144 vma.vm_mm = tlb->mm; 145 vm_flags_init(&vma, 0); 151 flush_tlb_range(&vma, tlb->start, tlb->end);
|
/linux-master/arch/arc/mm/ |
H A D | tlb.c | 190 * (Android Binder ended up calling this for vma->mm != tsk->mm, 208 void local_flush_tlb_range(struct vm_area_struct *vma, unsigned long start, argument 222 local_flush_tlb_mm(vma->vm_mm); 235 if (asid_mm(vma->vm_mm, cpu) != MM_CTXT_NO_ASID) { 237 tlb_entry_erase(start | hw_pid(vma->vm_mm, cpu)); 278 void local_flush_tlb_page(struct vm_area_struct *vma, unsigned long page) argument 288 if (asid_mm(vma->vm_mm, cpu) != MM_CTXT_NO_ASID) { 289 tlb_entry_erase((page & PAGE_MASK) | hw_pid(vma->vm_mm, cpu)); 344 void flush_tlb_page(struct vm_area_struct *vma, unsigned long uaddr) argument 347 .ta_vma = vma, 354 flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) argument 367 flush_pmd_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) argument 394 create_tlb(struct vm_area_struct *vma, unsigned long vaddr, pte_t *ptep) argument 472 update_mmu_cache_range(struct vm_fault *vmf, struct vm_area_struct *vma, unsigned long vaddr_unaligned, pte_t *ptep, unsigned int nr) argument 530 update_mmu_cache_pmd(struct vm_area_struct *vma, unsigned long addr, pmd_t *pmd) argument 537 local_flush_pmd_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) argument [all...] |
/linux-master/arch/arc/include/asm/ |
H A D | pgtable-bits-arcv2.h | 104 void update_mmu_cache_range(struct vm_fault *vmf, struct vm_area_struct *vma, 107 #define update_mmu_cache(vma, addr, ptep) \ 108 update_mmu_cache_range(NULL, vma, addr, ptep, 1)
|
/linux-master/fs/btrfs/ |
H A D | inode.c | 8196 struct inode *inode = file_inode(vmf->vma->vm_file); 8232 ret2 = file_update_time(vmf->vma->vm_file);
|
/linux-master/virt/kvm/ |
H A D | kvm_main.c | 2684 struct vm_area_struct *vma; local 2694 vma = find_vma(current->mm, addr); 2695 if (!vma) 2698 size = vma_kernel_pagesize(vma); 2873 static bool vma_is_valid(struct vm_area_struct *vma, bool write_fault) argument 2875 if (unlikely(!(vma->vm_flags & VM_READ))) 2878 if (write_fault && (unlikely(!(vma->vm_flags & VM_WRITE)))) 2894 static int hva_to_pfn_remapped(struct vm_area_struct *vma, argument 2904 r = follow_pte(vma->vm_mm, addr, &ptep, &ptl); 2919 r = follow_pte(vma 2980 struct vm_area_struct *vma; local 4169 kvm_vcpu_mmap(struct file *file, struct vm_area_struct *vma) argument 4671 kvm_device_mmap(struct file *filp, struct vm_area_struct *vma) argument [all...] |
/linux-master/fs/sysfs/ |
H A D | file.c | 162 struct vm_area_struct *vma) 167 return battr->mmap(of->file, kobj, battr, vma); 161 sysfs_kf_bin_mmap(struct kernfs_open_file *of, struct vm_area_struct *vma) argument
|
/linux-master/fs/bcachefs/ |
H A D | fs.c | 1068 static int bch2_mmap(struct file *file, struct vm_area_struct *vma) argument 1072 vma->vm_ops = &bch_vm_ops;
|
/linux-master/drivers/android/ |
H A D | binder.c | 3245 s = (ret == -ESRCH) ? ": vma cleared, target dead or dying" 5563 static void binder_vma_open(struct vm_area_struct *vma) argument 5565 struct binder_proc *proc = vma->vm_private_data; 5568 "%d open vm area %lx-%lx (%ld K) vma %lx pagep %lx\n", 5569 proc->pid, vma->vm_start, vma->vm_end, 5570 (vma->vm_end - vma->vm_start) / SZ_1K, vma->vm_flags, 5571 (unsigned long)pgprot_val(vma 5574 binder_vma_close(struct vm_area_struct *vma) argument 5597 binder_mmap(struct file *filp, struct vm_area_struct *vma) argument [all...] |