/linux-master/arch/openrisc/include/asm/ |
H A D | mmu_context.h | 21 extern int init_new_context(struct task_struct *tsk, struct mm_struct *mm); 23 extern void destroy_context(struct mm_struct *mm); 33 extern volatile pgd_t *current_pgd[]; /* defined in arch/openrisc/mm/fault.c */
|
H A D | pgalloc.h | 20 #include <linux/mm.h> 28 #define pmd_populate_kernel(mm, pmd, pte) \ 31 static inline void pmd_populate(struct mm_struct *mm, pmd_t *pmd, argument 42 static inline pgd_t *pgd_alloc(struct mm_struct *mm) argument 58 * current_pgd (from mm->pgd) to load kernel pages so we need it 61 extern inline pgd_t *pgd_alloc(struct mm_struct *mm) 67 extern pte_t *pte_alloc_one_kernel(struct mm_struct *mm);
|
/linux-master/arch/nios2/include/asm/ |
H A D | mmu_context.h | 37 struct mm_struct *mm) 39 mm->context = 0; 47 * After we have set current->mm to a new value, this activates 48 * the context for the new mm so we see the new mappings. 36 init_new_context(struct task_struct *tsk, struct mm_struct *mm) argument
|
/linux-master/drivers/infiniband/hw/mlx5/ |
H A D | doorbell.c | 35 #include <linux/sched/mm.h> 45 struct mm_struct *mm; member in struct:mlx5_ib_user_db_page 57 if ((current->mm == page->mm) && 76 mmgrab(current->mm); 77 page->mm = current->mm; 99 mmdrop(db->u.user_page->mm);
|
/linux-master/arch/parisc/include/asm/ |
H A D | vdso.h | 12 #define VDSO64_SYMBOL(tsk, name) ((tsk)->mm->context.vdso_base + (vdso64_offset_##name)) 13 #define VDSO32_SYMBOL(tsk, name) ((tsk)->mm->context.vdso_base + (vdso32_offset_##name))
|
/linux-master/arch/um/kernel/skas/ |
H A D | mmu.c | 7 #include <linux/mm.h> 17 int init_new_context(struct task_struct *task, struct mm_struct *mm) argument 20 struct mm_context *to_mm = &mm->context; 29 if (current->mm != NULL && current->mm != &init_mm) 30 from_mm = ¤t->mm->context; 60 void destroy_context(struct mm_struct *mm) argument 62 struct mm_context *mmu = &mm->context;
|
/linux-master/drivers/infiniband/hw/qib/ |
H A D | qib_user_pages.c | 34 #include <linux/mm.h> 102 locked = atomic64_add_return(num_pages, ¤t->mm->pinned_vm); 109 mmap_read_lock(current->mm); 116 mmap_read_unlock(current->mm); 120 mmap_read_unlock(current->mm); 126 atomic64_sub(num_pages, ¤t->mm->pinned_vm); 134 /* during close after signal, mm can be NULL */ 135 if (current->mm) 136 atomic64_sub(num_pages, ¤t->mm->pinned_vm);
|
/linux-master/arch/m68k/include/asm/ |
H A D | mcf_pgalloc.h | 8 static inline void pte_free_kernel(struct mm_struct *mm, pte_t *pte) argument 15 static inline pte_t *pte_alloc_one_kernel(struct mm_struct *mm) argument 31 #define pmd_populate(mm, pmd, pte) (pmd_val(*pmd) = (unsigned long)(pte)) 44 static inline pgtable_t pte_alloc_one(struct mm_struct *mm) argument 60 static inline void pte_free(struct mm_struct *mm, pgtable_t pgtable) argument 72 #define pmd_free(mm, pmd) BUG() 74 static inline void pgd_free(struct mm_struct *mm, pgd_t *pgd) argument 79 static inline pgd_t *pgd_alloc(struct mm_struct *mm) argument
|
/linux-master/arch/loongarch/mm/ |
H A D | hugetlbpage.c | 7 #include <linux/mm.h> 16 pte_t *huge_pte_alloc(struct mm_struct *mm, struct vm_area_struct *vma, argument 24 pgd = pgd_offset(mm, addr); 25 p4d = p4d_alloc(mm, pgd, addr); 26 pud = pud_alloc(mm, p4d, addr); 28 pte = (pte_t *)pmd_alloc(mm, pud, addr); 33 pte_t *huge_pte_offset(struct mm_struct *mm, unsigned long addr, argument 41 pgd = pgd_offset(mm, addr);
|
/linux-master/include/linux/ |
H A D | mmu_context.h | 32 static inline unsigned long mm_untag_mask(struct mm_struct *mm) argument 39 static inline bool arch_pgtable_dma_compat(struct mm_struct *mm) argument
|
/linux-master/arch/riscv/include/asm/ |
H A D | tlb.h | 38 flush_tlb_mm(tlb->mm); 40 flush_tlb_mm_range(tlb->mm, tlb->start, tlb->end,
|
/linux-master/arch/powerpc/include/asm/book3s/32/ |
H A D | pgalloc.h | 8 static inline pgd_t *pgd_alloc(struct mm_struct *mm) argument 11 pgtable_gfp_flags(mm, GFP_KERNEL)); 14 static inline void pgd_free(struct mm_struct *mm, pgd_t *pgd) argument 23 /* #define pmd_alloc_one(mm,address) ({ BUG(); ((pmd_t *)2); }) */ 24 #define pmd_free(mm, x) do { } while (0) 26 /* #define pgd_populate(mm, pmd, pte) BUG() */ 28 static inline void pmd_populate_kernel(struct mm_struct *mm, pmd_t *pmdp, argument 34 static inline void pmd_populate(struct mm_struct *mm, pmd_t *pmdp, argument
|
/linux-master/arch/x86/um/vdso/ |
H A D | vma.c | 8 #include <linux/mm.h> 56 struct mm_struct *mm = current->mm; local 61 if (mmap_write_lock_killable(mm)) 64 err = install_special_mapping(mm, um_vdso_addr, PAGE_SIZE, 69 mmap_write_unlock(mm);
|
/linux-master/arch/sh/mm/ |
H A D | pgtable.c | 2 #include <linux/mm.h> 32 pgd_t *pgd_alloc(struct mm_struct *mm) argument 37 void pgd_free(struct mm_struct *mm, pgd_t *pgd) argument 43 void pud_populate(struct mm_struct *mm, pud_t *pud, pmd_t *pmd) argument 48 pmd_t *pmd_alloc_one(struct mm_struct *mm, unsigned long address) argument 53 void pmd_free(struct mm_struct *mm, pmd_t *pmd) argument
|
/linux-master/mm/ |
H A D | pgalloc-track.h | 6 static inline p4d_t *p4d_alloc_track(struct mm_struct *mm, pgd_t *pgd, argument 11 if (__p4d_alloc(mm, pgd, address)) 19 static inline pud_t *pud_alloc_track(struct mm_struct *mm, p4d_t *p4d, argument 24 if (__pud_alloc(mm, p4d, address)) 32 static inline pmd_t *pmd_alloc_track(struct mm_struct *mm, pud_t *pud, argument 37 if (__pmd_alloc(mm, pud, address))
|
/linux-master/arch/csky/abiv2/inc/abi/ |
H A D | cacheflush.h | 7 #include <linux/mm.h> 14 #define flush_cache_mm(mm) do { } while (0) 15 #define flush_cache_dup_mm(mm) do { } while (0) 39 void flush_icache_mm_range(struct mm_struct *mm, 41 void flush_icache_deferred(struct mm_struct *mm); 53 flush_icache_mm_range(current->mm, \
|
/linux-master/arch/sparc/include/asm/ |
H A D | cacheflush_32.h | 10 #define flush_cache_mm(mm) \ 11 sparc32_cachetlb_ops->cache_mm(mm) 12 #define flush_cache_dup_mm(mm) \ 13 sparc32_cachetlb_ops->cache_mm(mm) 33 #define flush_sig_insns(mm,insn_addr) \ 34 sparc32_cachetlb_ops->sig_insns(mm, insn_addr)
|
/linux-master/drivers/iommu/arm/arm-smmu-v3/ |
H A D | arm-smmu-v3-sva.c | 6 #include <linux/mm.h> 9 #include <linux/sched/mm.h> 27 struct mm_struct *mm; member in struct:arm_smmu_bond 61 arm_smmu_share_asid(struct mm_struct *mm, u16 asid) argument 73 if (cd->mm) { 74 if (WARN_ON(cd->mm != mm)) 76 /* All devices bound to this mm use the same cd struct. */ 108 static struct arm_smmu_ctx_desc *arm_smmu_alloc_shared_cd(struct mm_struct *mm) argument 116 /* Don't free the mm unti 214 arm_smmu_mm_arch_invalidate_secondary_tlbs(struct mmu_notifier *mn, struct mm_struct *mm, unsigned long start, unsigned long end) argument 252 arm_smmu_mm_release(struct mmu_notifier *mn, struct mm_struct *mm) argument 290 arm_smmu_mmu_notifier_get(struct arm_smmu_domain *smmu_domain, struct mm_struct *mm) argument 335 struct mm_struct *mm = smmu_mn->mn.mm; local 359 __arm_smmu_sva_bind(struct device *dev, ioasid_t pasid, struct mm_struct *mm) argument 543 struct mm_struct *mm = domain->mm; local 570 struct mm_struct *mm = domain->mm; local [all...] |
/linux-master/drivers/gpu/drm/ |
H A D | drm_mm.c | 118 static void show_leaks(struct drm_mm *mm) argument 127 list_for_each_entry(node, drm_mm_nodes(mm), node_list) { 146 static void show_leaks(struct drm_mm *mm) { } argument 157 __drm_mm_interval_first(const struct drm_mm *mm, u64 start, u64 last) argument 159 return drm_mm_interval_tree_iter_first((struct rb_root_cached *)&mm->interval_tree, 160 start, last) ?: (struct drm_mm_node *)&mm->head_node; 167 struct drm_mm *mm = hole_node->mm; local 190 link = &mm->interval_tree.rb_root.rb_node; 208 rb_insert_augmented_cached(&node->rb, &mm 268 struct drm_mm *mm = node->mm; local 305 best_hole(struct drm_mm *mm, u64 size) argument 330 find_hole_addr(struct drm_mm *mm, u64 addr, u64 size) argument 356 first_hole(struct drm_mm *mm, u64 start, u64 end, u64 size, enum drm_mm_insert_mode mode) argument 414 next_hole(struct drm_mm *mm, struct drm_mm_node *node, u64 size, enum drm_mm_insert_mode mode) argument 450 drm_mm_reserve_node(struct drm_mm *mm, struct drm_mm_node *node) argument 514 drm_mm_insert_node_in_range(struct drm_mm * const mm, struct drm_mm_node * const node, u64 size, u64 alignment, unsigned long color, u64 range_start, u64 range_end, enum drm_mm_insert_mode mode) argument 629 struct drm_mm *mm = node->mm; local 662 struct drm_mm *mm = old->mm; local 736 drm_mm_scan_init_with_range(struct drm_mm_scan *scan, struct drm_mm *mm, u64 size, u64 alignment, unsigned long color, u64 start, u64 end, enum drm_mm_insert_mode mode) argument 783 struct drm_mm *mm = scan->mm; local 914 struct drm_mm *mm = scan->mm; local 963 drm_mm_init(struct drm_mm *mm, u64 start, u64 size) argument 997 drm_mm_takedown(struct drm_mm *mm) argument 1023 drm_mm_print(const struct drm_mm *mm, struct drm_printer *p) argument [all...] |
/linux-master/arch/x86/mm/ |
H A D | mmap.c | 15 #include <linux/mm.h> 19 #include <linux/sched/mm.h> 129 void arch_pick_mmap_layout(struct mm_struct *mm, struct rlimit *rlim_stack) argument 132 mm->get_unmapped_area = arch_get_unmapped_area; 134 mm->get_unmapped_area = arch_get_unmapped_area_topdown; 136 arch_pick_mmap_base(&mm->mmap_base, &mm->mmap_legacy_base, 147 arch_pick_mmap_base(&mm->mmap_compat_base, &mm->mmap_compat_legacy_base, 155 struct mm_struct *mm local [all...] |
/linux-master/arch/arm/mm/ |
H A D | context.c | 3 * linux/arch/arm/mm/context.c 12 #include <linux/mm.h> 51 void a15_erratum_get_cpumask(int this_cpu, struct mm_struct *mm, argument 59 context_id = mm->context.id.counter; 172 * If we find one, then we can update our mm to use newasid 175 * of the old ASID are updated to reflect the mm. Failure to do 189 static u64 new_context(struct mm_struct *mm, unsigned int cpu) argument 192 u64 asid = atomic64_read(&mm->context.id); 233 cpumask_clear(mm_cpumask(mm)); 237 void check_and_switch_context(struct mm_struct *mm, struc argument [all...] |
/linux-master/arch/mips/mm/ |
H A D | context.c | 21 void get_new_mmu_context(struct mm_struct *mm) argument 42 set_cpu_context(cpu, mm, asid); 47 void check_mmu_context(struct mm_struct *mm) argument 59 if (!asid_versions_eq(cpu, cpu_context(cpu, mm), asid_cache(cpu))) 60 get_new_mmu_context(mm); 106 * If we find one, then we can update our mm to use newmmid 109 * of the old MMID are updated to reflect the mm. Failure to do 124 static u64 get_new_mmid(struct mm_struct *mm) argument 129 mmid = cpu_context(0, mm); 175 set_cpu_context(0, mm, mmi 179 check_switch_mmu_context(struct mm_struct *mm) argument [all...] |
/linux-master/kernel/ |
H A D | fork.c | 12 * management can be a bitch. See 'mm/memory.c': 'copy_page_range()' 18 #include <linux/sched/mm.h> 45 #include <linux/mm.h> 431 /* SLAB cache for mm_struct structures (tsk->mm) */ 463 struct vm_area_struct *vm_area_alloc(struct mm_struct *mm) argument 471 vma_init(vma, mm); 613 static void dup_mm_exe_file(struct mm_struct *mm, struct mm_struct *oldmm) argument 618 RCU_INIT_POINTER(mm->exe_file, exe_file); 628 static __latent_entropy int dup_mmap(struct mm_struct *mm, argument 635 VMA_ITERATOR(vmi, mm, 794 mm_alloc_pgd(struct mm_struct *mm) argument 802 mm_free_pgd(struct mm_struct *mm) argument 807 dup_mmap(struct mm_struct *mm, struct mm_struct *oldmm) argument 818 check_mm(struct mm_struct *mm) argument 847 struct mm_struct *mm = arg; local 854 struct mm_struct *mm = arg; local 863 cleanup_lazy_tlbs(struct mm_struct *mm) argument 911 __mmdrop(struct mm_struct *mm) argument 935 struct mm_struct *mm; local 941 mmdrop_async(struct mm_struct *mm) argument 1223 mm_init_aio(struct mm_struct *mm) argument 1231 mm_clear_owner(struct mm_struct *mm, struct task_struct *p) argument 1240 mm_init_owner(struct mm_struct *mm, struct task_struct *p) argument 1247 mm_init_uprobes_state(struct mm_struct *mm) argument 1254 mm_init(struct mm_struct *mm, struct task_struct *p, struct user_namespace *user_ns) argument 1328 struct mm_struct *mm; local 1338 __mmput(struct mm_struct *mm) argument 1363 mmput(struct mm_struct *mm) argument 1375 struct mm_struct *mm = container_of(work, struct mm_struct, local 1381 mmput_async(struct mm_struct *mm) argument 1404 set_mm_exe_file(struct mm_struct *mm, struct file *new_exe_file) argument 1441 replace_mm_exe_file(struct mm_struct *mm, struct file *new_exe_file) argument 1492 get_mm_exe_file(struct mm_struct *mm) argument 1513 struct mm_struct *mm; local 1537 struct mm_struct *mm; local 1554 struct mm_struct *mm; local 1618 mm_release(struct task_struct *tsk, struct mm_struct *mm) argument 1651 exit_mm_release(struct task_struct *tsk, struct mm_struct *mm) argument 1657 exec_mm_release(struct task_struct *tsk, struct mm_struct *mm) argument 1676 struct mm_struct *mm; local 1712 struct mm_struct *mm, *oldmm; local [all...] |
/linux-master/arch/powerpc/mm/nohash/ |
H A D | mmu_context.c | 10 * Derived from previous arch/powerpc/mm/mmu_context.c 21 * also clear mm->cpu_vm_mask bits when processes are migrated 25 #include <linux/mm.h> 38 #include <mm/mmu_decl.h> 99 struct mm_struct *mm; local 106 /* Pick up the victim mm */ 107 mm = context_mm[id]; 112 if (mm->context.active) { 119 /* Mark this mm has having no context anymore */ 120 mm 151 struct mm_struct *mm; local 184 struct mm_struct *mm; local 318 init_new_context(struct task_struct *t, struct mm_struct *mm) argument 329 destroy_context(struct mm_struct *mm) argument [all...] |
/linux-master/arch/arm64/kernel/ |
H A D | vdso.c | 17 #include <linux/mm.h> 78 current->mm->context.vdso = (void *)new_vma->vm_start; 131 struct mm_struct *mm = task->mm; local 133 VMA_ITERATOR(vmi, mm, 0); 135 mmap_read_lock(mm); 146 mmap_read_unlock(mm); 186 struct mm_struct *mm, 206 ret = _install_special_mapping(mm, vdso_base, VVAR_NR_PAGES * PAGE_SIZE, 216 mm 185 __setup_additional_pages(enum vdso_abi abi, struct mm_struct *mm, struct linux_binprm *bprm, int uses_interp) argument 338 aarch32_kuser_helpers_setup(struct mm_struct *mm) argument 357 aarch32_sigreturn_setup(struct mm_struct *mm) argument 387 struct mm_struct *mm = current->mm; local 438 struct mm_struct *mm = current->mm; local [all...] |