Searched refs:mm (Results 151 - 175 of 2900) sorted by relevance

1234567891011>>

/linux-master/arch/openrisc/include/asm/
H A Dmmu_context.h21 extern int init_new_context(struct task_struct *tsk, struct mm_struct *mm);
23 extern void destroy_context(struct mm_struct *mm);
33 extern volatile pgd_t *current_pgd[]; /* defined in arch/openrisc/mm/fault.c */
H A Dpgalloc.h20 #include <linux/mm.h>
28 #define pmd_populate_kernel(mm, pmd, pte) \
31 static inline void pmd_populate(struct mm_struct *mm, pmd_t *pmd, argument
42 static inline pgd_t *pgd_alloc(struct mm_struct *mm) argument
58 * current_pgd (from mm->pgd) to load kernel pages so we need it
61 extern inline pgd_t *pgd_alloc(struct mm_struct *mm)
67 extern pte_t *pte_alloc_one_kernel(struct mm_struct *mm);
/linux-master/arch/nios2/include/asm/
H A Dmmu_context.h37 struct mm_struct *mm)
39 mm->context = 0;
47 * After we have set current->mm to a new value, this activates
48 * the context for the new mm so we see the new mappings.
36 init_new_context(struct task_struct *tsk, struct mm_struct *mm) argument
/linux-master/drivers/infiniband/hw/mlx5/
H A Ddoorbell.c35 #include <linux/sched/mm.h>
45 struct mm_struct *mm; member in struct:mlx5_ib_user_db_page
57 if ((current->mm == page->mm) &&
76 mmgrab(current->mm);
77 page->mm = current->mm;
99 mmdrop(db->u.user_page->mm);
/linux-master/arch/parisc/include/asm/
H A Dvdso.h12 #define VDSO64_SYMBOL(tsk, name) ((tsk)->mm->context.vdso_base + (vdso64_offset_##name))
13 #define VDSO32_SYMBOL(tsk, name) ((tsk)->mm->context.vdso_base + (vdso32_offset_##name))
/linux-master/arch/um/kernel/skas/
H A Dmmu.c7 #include <linux/mm.h>
17 int init_new_context(struct task_struct *task, struct mm_struct *mm) argument
20 struct mm_context *to_mm = &mm->context;
29 if (current->mm != NULL && current->mm != &init_mm)
30 from_mm = &current->mm->context;
60 void destroy_context(struct mm_struct *mm) argument
62 struct mm_context *mmu = &mm->context;
/linux-master/drivers/infiniband/hw/qib/
H A Dqib_user_pages.c34 #include <linux/mm.h>
102 locked = atomic64_add_return(num_pages, &current->mm->pinned_vm);
109 mmap_read_lock(current->mm);
116 mmap_read_unlock(current->mm);
120 mmap_read_unlock(current->mm);
126 atomic64_sub(num_pages, &current->mm->pinned_vm);
134 /* during close after signal, mm can be NULL */
135 if (current->mm)
136 atomic64_sub(num_pages, &current->mm->pinned_vm);
/linux-master/arch/m68k/include/asm/
H A Dmcf_pgalloc.h8 static inline void pte_free_kernel(struct mm_struct *mm, pte_t *pte) argument
15 static inline pte_t *pte_alloc_one_kernel(struct mm_struct *mm) argument
31 #define pmd_populate(mm, pmd, pte) (pmd_val(*pmd) = (unsigned long)(pte))
44 static inline pgtable_t pte_alloc_one(struct mm_struct *mm) argument
60 static inline void pte_free(struct mm_struct *mm, pgtable_t pgtable) argument
72 #define pmd_free(mm, pmd) BUG()
74 static inline void pgd_free(struct mm_struct *mm, pgd_t *pgd) argument
79 static inline pgd_t *pgd_alloc(struct mm_struct *mm) argument
/linux-master/arch/loongarch/mm/
H A Dhugetlbpage.c7 #include <linux/mm.h>
16 pte_t *huge_pte_alloc(struct mm_struct *mm, struct vm_area_struct *vma, argument
24 pgd = pgd_offset(mm, addr);
25 p4d = p4d_alloc(mm, pgd, addr);
26 pud = pud_alloc(mm, p4d, addr);
28 pte = (pte_t *)pmd_alloc(mm, pud, addr);
33 pte_t *huge_pte_offset(struct mm_struct *mm, unsigned long addr, argument
41 pgd = pgd_offset(mm, addr);
/linux-master/include/linux/
H A Dmmu_context.h32 static inline unsigned long mm_untag_mask(struct mm_struct *mm) argument
39 static inline bool arch_pgtable_dma_compat(struct mm_struct *mm) argument
/linux-master/arch/riscv/include/asm/
H A Dtlb.h38 flush_tlb_mm(tlb->mm);
40 flush_tlb_mm_range(tlb->mm, tlb->start, tlb->end,
/linux-master/arch/powerpc/include/asm/book3s/32/
H A Dpgalloc.h8 static inline pgd_t *pgd_alloc(struct mm_struct *mm) argument
11 pgtable_gfp_flags(mm, GFP_KERNEL));
14 static inline void pgd_free(struct mm_struct *mm, pgd_t *pgd) argument
23 /* #define pmd_alloc_one(mm,address) ({ BUG(); ((pmd_t *)2); }) */
24 #define pmd_free(mm, x) do { } while (0)
26 /* #define pgd_populate(mm, pmd, pte) BUG() */
28 static inline void pmd_populate_kernel(struct mm_struct *mm, pmd_t *pmdp, argument
34 static inline void pmd_populate(struct mm_struct *mm, pmd_t *pmdp, argument
/linux-master/arch/x86/um/vdso/
H A Dvma.c8 #include <linux/mm.h>
56 struct mm_struct *mm = current->mm; local
61 if (mmap_write_lock_killable(mm))
64 err = install_special_mapping(mm, um_vdso_addr, PAGE_SIZE,
69 mmap_write_unlock(mm);
/linux-master/arch/sh/mm/
H A Dpgtable.c2 #include <linux/mm.h>
32 pgd_t *pgd_alloc(struct mm_struct *mm) argument
37 void pgd_free(struct mm_struct *mm, pgd_t *pgd) argument
43 void pud_populate(struct mm_struct *mm, pud_t *pud, pmd_t *pmd) argument
48 pmd_t *pmd_alloc_one(struct mm_struct *mm, unsigned long address) argument
53 void pmd_free(struct mm_struct *mm, pmd_t *pmd) argument
/linux-master/mm/
H A Dpgalloc-track.h6 static inline p4d_t *p4d_alloc_track(struct mm_struct *mm, pgd_t *pgd, argument
11 if (__p4d_alloc(mm, pgd, address))
19 static inline pud_t *pud_alloc_track(struct mm_struct *mm, p4d_t *p4d, argument
24 if (__pud_alloc(mm, p4d, address))
32 static inline pmd_t *pmd_alloc_track(struct mm_struct *mm, pud_t *pud, argument
37 if (__pmd_alloc(mm, pud, address))
/linux-master/arch/csky/abiv2/inc/abi/
H A Dcacheflush.h7 #include <linux/mm.h>
14 #define flush_cache_mm(mm) do { } while (0)
15 #define flush_cache_dup_mm(mm) do { } while (0)
39 void flush_icache_mm_range(struct mm_struct *mm,
41 void flush_icache_deferred(struct mm_struct *mm);
53 flush_icache_mm_range(current->mm, \
/linux-master/arch/sparc/include/asm/
H A Dcacheflush_32.h10 #define flush_cache_mm(mm) \
11 sparc32_cachetlb_ops->cache_mm(mm)
12 #define flush_cache_dup_mm(mm) \
13 sparc32_cachetlb_ops->cache_mm(mm)
33 #define flush_sig_insns(mm,insn_addr) \
34 sparc32_cachetlb_ops->sig_insns(mm, insn_addr)
/linux-master/drivers/iommu/arm/arm-smmu-v3/
H A Darm-smmu-v3-sva.c6 #include <linux/mm.h>
9 #include <linux/sched/mm.h>
27 struct mm_struct *mm; member in struct:arm_smmu_bond
61 arm_smmu_share_asid(struct mm_struct *mm, u16 asid) argument
73 if (cd->mm) {
74 if (WARN_ON(cd->mm != mm))
76 /* All devices bound to this mm use the same cd struct. */
108 static struct arm_smmu_ctx_desc *arm_smmu_alloc_shared_cd(struct mm_struct *mm) argument
116 /* Don't free the mm unti
214 arm_smmu_mm_arch_invalidate_secondary_tlbs(struct mmu_notifier *mn, struct mm_struct *mm, unsigned long start, unsigned long end) argument
252 arm_smmu_mm_release(struct mmu_notifier *mn, struct mm_struct *mm) argument
290 arm_smmu_mmu_notifier_get(struct arm_smmu_domain *smmu_domain, struct mm_struct *mm) argument
335 struct mm_struct *mm = smmu_mn->mn.mm; local
359 __arm_smmu_sva_bind(struct device *dev, ioasid_t pasid, struct mm_struct *mm) argument
543 struct mm_struct *mm = domain->mm; local
570 struct mm_struct *mm = domain->mm; local
[all...]
/linux-master/drivers/gpu/drm/
H A Ddrm_mm.c118 static void show_leaks(struct drm_mm *mm) argument
127 list_for_each_entry(node, drm_mm_nodes(mm), node_list) {
146 static void show_leaks(struct drm_mm *mm) { } argument
157 __drm_mm_interval_first(const struct drm_mm *mm, u64 start, u64 last) argument
159 return drm_mm_interval_tree_iter_first((struct rb_root_cached *)&mm->interval_tree,
160 start, last) ?: (struct drm_mm_node *)&mm->head_node;
167 struct drm_mm *mm = hole_node->mm; local
190 link = &mm->interval_tree.rb_root.rb_node;
208 rb_insert_augmented_cached(&node->rb, &mm
268 struct drm_mm *mm = node->mm; local
305 best_hole(struct drm_mm *mm, u64 size) argument
330 find_hole_addr(struct drm_mm *mm, u64 addr, u64 size) argument
356 first_hole(struct drm_mm *mm, u64 start, u64 end, u64 size, enum drm_mm_insert_mode mode) argument
414 next_hole(struct drm_mm *mm, struct drm_mm_node *node, u64 size, enum drm_mm_insert_mode mode) argument
450 drm_mm_reserve_node(struct drm_mm *mm, struct drm_mm_node *node) argument
514 drm_mm_insert_node_in_range(struct drm_mm * const mm, struct drm_mm_node * const node, u64 size, u64 alignment, unsigned long color, u64 range_start, u64 range_end, enum drm_mm_insert_mode mode) argument
629 struct drm_mm *mm = node->mm; local
662 struct drm_mm *mm = old->mm; local
736 drm_mm_scan_init_with_range(struct drm_mm_scan *scan, struct drm_mm *mm, u64 size, u64 alignment, unsigned long color, u64 start, u64 end, enum drm_mm_insert_mode mode) argument
783 struct drm_mm *mm = scan->mm; local
914 struct drm_mm *mm = scan->mm; local
963 drm_mm_init(struct drm_mm *mm, u64 start, u64 size) argument
997 drm_mm_takedown(struct drm_mm *mm) argument
1023 drm_mm_print(const struct drm_mm *mm, struct drm_printer *p) argument
[all...]
/linux-master/arch/x86/mm/
H A Dmmap.c15 #include <linux/mm.h>
19 #include <linux/sched/mm.h>
129 void arch_pick_mmap_layout(struct mm_struct *mm, struct rlimit *rlim_stack) argument
132 mm->get_unmapped_area = arch_get_unmapped_area;
134 mm->get_unmapped_area = arch_get_unmapped_area_topdown;
136 arch_pick_mmap_base(&mm->mmap_base, &mm->mmap_legacy_base,
147 arch_pick_mmap_base(&mm->mmap_compat_base, &mm->mmap_compat_legacy_base,
155 struct mm_struct *mm local
[all...]
/linux-master/arch/arm/mm/
H A Dcontext.c3 * linux/arch/arm/mm/context.c
12 #include <linux/mm.h>
51 void a15_erratum_get_cpumask(int this_cpu, struct mm_struct *mm, argument
59 context_id = mm->context.id.counter;
172 * If we find one, then we can update our mm to use newasid
175 * of the old ASID are updated to reflect the mm. Failure to do
189 static u64 new_context(struct mm_struct *mm, unsigned int cpu) argument
192 u64 asid = atomic64_read(&mm->context.id);
233 cpumask_clear(mm_cpumask(mm));
237 void check_and_switch_context(struct mm_struct *mm, struc argument
[all...]
/linux-master/arch/mips/mm/
H A Dcontext.c21 void get_new_mmu_context(struct mm_struct *mm) argument
42 set_cpu_context(cpu, mm, asid);
47 void check_mmu_context(struct mm_struct *mm) argument
59 if (!asid_versions_eq(cpu, cpu_context(cpu, mm), asid_cache(cpu)))
60 get_new_mmu_context(mm);
106 * If we find one, then we can update our mm to use newmmid
109 * of the old MMID are updated to reflect the mm. Failure to do
124 static u64 get_new_mmid(struct mm_struct *mm) argument
129 mmid = cpu_context(0, mm);
175 set_cpu_context(0, mm, mmi
179 check_switch_mmu_context(struct mm_struct *mm) argument
[all...]
/linux-master/kernel/
H A Dfork.c12 * management can be a bitch. See 'mm/memory.c': 'copy_page_range()'
18 #include <linux/sched/mm.h>
45 #include <linux/mm.h>
431 /* SLAB cache for mm_struct structures (tsk->mm) */
463 struct vm_area_struct *vm_area_alloc(struct mm_struct *mm) argument
471 vma_init(vma, mm);
613 static void dup_mm_exe_file(struct mm_struct *mm, struct mm_struct *oldmm) argument
618 RCU_INIT_POINTER(mm->exe_file, exe_file);
628 static __latent_entropy int dup_mmap(struct mm_struct *mm, argument
635 VMA_ITERATOR(vmi, mm,
794 mm_alloc_pgd(struct mm_struct *mm) argument
802 mm_free_pgd(struct mm_struct *mm) argument
807 dup_mmap(struct mm_struct *mm, struct mm_struct *oldmm) argument
818 check_mm(struct mm_struct *mm) argument
847 struct mm_struct *mm = arg; local
854 struct mm_struct *mm = arg; local
863 cleanup_lazy_tlbs(struct mm_struct *mm) argument
911 __mmdrop(struct mm_struct *mm) argument
935 struct mm_struct *mm; local
941 mmdrop_async(struct mm_struct *mm) argument
1223 mm_init_aio(struct mm_struct *mm) argument
1231 mm_clear_owner(struct mm_struct *mm, struct task_struct *p) argument
1240 mm_init_owner(struct mm_struct *mm, struct task_struct *p) argument
1247 mm_init_uprobes_state(struct mm_struct *mm) argument
1254 mm_init(struct mm_struct *mm, struct task_struct *p, struct user_namespace *user_ns) argument
1328 struct mm_struct *mm; local
1338 __mmput(struct mm_struct *mm) argument
1363 mmput(struct mm_struct *mm) argument
1375 struct mm_struct *mm = container_of(work, struct mm_struct, local
1381 mmput_async(struct mm_struct *mm) argument
1404 set_mm_exe_file(struct mm_struct *mm, struct file *new_exe_file) argument
1441 replace_mm_exe_file(struct mm_struct *mm, struct file *new_exe_file) argument
1492 get_mm_exe_file(struct mm_struct *mm) argument
1513 struct mm_struct *mm; local
1537 struct mm_struct *mm; local
1554 struct mm_struct *mm; local
1618 mm_release(struct task_struct *tsk, struct mm_struct *mm) argument
1651 exit_mm_release(struct task_struct *tsk, struct mm_struct *mm) argument
1657 exec_mm_release(struct task_struct *tsk, struct mm_struct *mm) argument
1676 struct mm_struct *mm; local
1712 struct mm_struct *mm, *oldmm; local
[all...]
/linux-master/arch/powerpc/mm/nohash/
H A Dmmu_context.c10 * Derived from previous arch/powerpc/mm/mmu_context.c
21 * also clear mm->cpu_vm_mask bits when processes are migrated
25 #include <linux/mm.h>
38 #include <mm/mmu_decl.h>
99 struct mm_struct *mm; local
106 /* Pick up the victim mm */
107 mm = context_mm[id];
112 if (mm->context.active) {
119 /* Mark this mm has having no context anymore */
120 mm
151 struct mm_struct *mm; local
184 struct mm_struct *mm; local
318 init_new_context(struct task_struct *t, struct mm_struct *mm) argument
329 destroy_context(struct mm_struct *mm) argument
[all...]
/linux-master/arch/arm64/kernel/
H A Dvdso.c17 #include <linux/mm.h>
78 current->mm->context.vdso = (void *)new_vma->vm_start;
131 struct mm_struct *mm = task->mm; local
133 VMA_ITERATOR(vmi, mm, 0);
135 mmap_read_lock(mm);
146 mmap_read_unlock(mm);
186 struct mm_struct *mm,
206 ret = _install_special_mapping(mm, vdso_base, VVAR_NR_PAGES * PAGE_SIZE,
216 mm
185 __setup_additional_pages(enum vdso_abi abi, struct mm_struct *mm, struct linux_binprm *bprm, int uses_interp) argument
338 aarch32_kuser_helpers_setup(struct mm_struct *mm) argument
357 aarch32_sigreturn_setup(struct mm_struct *mm) argument
387 struct mm_struct *mm = current->mm; local
438 struct mm_struct *mm = current->mm; local
[all...]

Completed in 229 milliseconds

1234567891011>>