Searched refs:mm (Results 401 - 425 of 2900) sorted by relevance

<<11121314151617181920>>

/linux-master/arch/s390/mm/
H A Dgmap.c103 * @mm: pointer to the parent mm_struct
108 struct gmap *gmap_create(struct mm_struct *mm, unsigned long limit) argument
116 gmap->mm = mm;
117 spin_lock(&mm->context.lock);
118 list_add_rcu(&gmap->list, &mm->context.gmap_list);
119 if (list_is_singular(&mm->context.gmap_list))
123 WRITE_ONCE(mm->context.gmap_asce, gmap_asce);
124 spin_unlock(&mm->context.lock);
266 /* Remove gmap from the pre-mm lis
523 gmap_unlink(struct mm_struct *mm, unsigned long *table, unsigned long vmaddr) argument
554 struct mm_struct *mm; local
895 struct mm_struct *mm = gmap->mm; local
2267 ptep_notify(struct mm_struct *mm, unsigned long vmaddr, pte_t *pte, unsigned long bits) argument
2334 gmap_pmdp_clear(struct mm_struct *mm, unsigned long vmaddr, int purge) argument
2366 gmap_pmdp_invalidate(struct mm_struct *mm, unsigned long vmaddr) argument
2377 gmap_pmdp_csp(struct mm_struct *mm, unsigned long vmaddr) argument
2388 gmap_pmdp_idte_local(struct mm_struct *mm, unsigned long vmaddr) argument
2423 gmap_pmdp_idte_global(struct mm_struct *mm, unsigned long vmaddr) argument
2535 thp_split_mm(struct mm_struct *mm) argument
2547 thp_split_mm(struct mm_struct *mm) argument
2592 struct mm_struct *mm = current->mm; local
2680 struct mm_struct *mm = current->mm; local
2716 s390_reset_cmma(struct mm_struct *mm) argument
2785 __s390_uv_destroy_range(struct mm_struct *mm, unsigned long start, unsigned long end, bool interruptible) argument
[all...]
/linux-master/fs/
H A Dexec.c30 #include <linux/mm.h>
36 #include <linux/sched/mm.h>
181 * The nascent bprm->mm is not visible until exec_mmap() but it can
182 * use a lot of memory, account these pages in current->mm temporary
188 struct mm_struct *mm = current->mm; local
191 if (!mm || !diff)
195 add_mm_counter(mm, MM_ANONPAGES, diff);
203 struct mm_struct *mm = bprm->mm; local
258 struct mm_struct *mm = bprm->mm; local
372 struct mm_struct *mm = NULL; local
686 struct mm_struct *mm = vma->vm_mm; local
756 struct mm_struct *mm = current->mm; local
996 exec_mmap(struct mm_struct *mm) argument
2102 struct mm_struct *mm = current->mm; local
2116 set_dumpable(struct mm_struct *mm, int value) argument
[all...]
/linux-master/arch/um/kernel/
H A Dtrap.c6 #include <linux/mm.h>
27 struct mm_struct *mm = current->mm; local
46 mmap_read_lock(mm);
47 vma = find_vma(mm, address);
56 vma = expand_stack(mm, address);
101 pmd = pmd_off(mm, address);
118 mmap_read_unlock(mm);
127 mmap_read_unlock(mm);
217 else if (current->mm
[all...]
/linux-master/arch/csky/mm/
H A Dfault.c100 static inline void bad_area_nosemaphore(struct pt_regs *regs, struct mm_struct *mm, int code, unsigned long addr) argument
191 struct mm_struct *mm; local
198 mm = tsk->mm;
227 if (unlikely(faulthandler_disabled() || !mm)) {
240 vma = lock_mm_and_find_vma(mm, addr, regs);
242 bad_area_nosemaphore(regs, mm, code, addr);
253 mmap_read_unlock(mm);
254 bad_area_nosemaphore(regs, mm, code, addr);
268 * would already be released in __lock_page_or_retry in mm/filema
[all...]
/linux-master/arch/powerpc/include/asm/nohash/
H A Dpgtable.h6 static inline pte_basic_t pte_update(struct mm_struct *mm, unsigned long addr, pte_t *p,
50 static inline pte_basic_t pte_update(struct mm_struct *mm, unsigned long addr, pte_t *p, argument
66 assert_pte_locked(mm, addr);
84 static inline void ptep_set_wrprotect(struct mm_struct *mm, unsigned long addr, argument
87 pte_update(mm, addr, ptep, _PAGE_WRITE, 0, 0);
92 static inline pte_t ptep_get_and_clear(struct mm_struct *mm, unsigned long addr, argument
95 return __pte(pte_update(mm, addr, ptep, ~0UL, 0, 0));
99 static inline void pte_clear(struct mm_struct *mm, unsigned long addr, pte_t *ptep) argument
101 pte_update(mm, addr, ptep, ~0UL, 0, 0);
279 static inline void __set_pte_at(struct mm_struct *mm, unsigne argument
[all...]
/linux-master/arch/loongarch/mm/
H A Dtlb.c8 #include <linux/mm.h>
38 * All entries common to a mm share an asid. To effectively flush
41 void local_flush_tlb_mm(struct mm_struct *mm) argument
49 if (asid_valid(mm, cpu))
50 drop_mmu_context(mm, cpu);
52 cpumask_clear_cpu(cpu, mm_cpumask(mm));
60 struct mm_struct *mm = vma->vm_mm; local
63 if (asid_valid(mm, cpu)) {
73 int asid = cpu_asid(cpu, mm);
80 drop_mmu_context(mm, cp
[all...]
/linux-master/arch/arm64/include/asm/
H A Dtlbflush.h199 * flush_tlb_mm(mm)
201 * The 'mm' argument identifies the ASID to invalidate.
205 * CPUs for the user address space corresponding to 'vma->mm'.
218 * address space corresponding to 'vma->mm'. Note that this
236 * CPUs for the user address space corresponding to 'vma->mm'.
266 static inline void flush_tlb_mm(struct mm_struct *mm) argument
271 asid = __TLBI_VADDR(0, ASID(mm));
275 mmu_notifier_arch_invalidate_secondary_tlbs(mm, 0, -1UL);
278 static inline void __flush_tlb_page_nosync(struct mm_struct *mm, argument
284 addr = __TLBI_VADDR(uaddr, ASID(mm));
304 arch_tlbbatch_should_defer(struct mm_struct *mm) argument
318 arch_tlbbatch_add_pending(struct arch_tlbflush_unmap_batch *batch, struct mm_struct *mm, unsigned long uaddr) argument
330 arch_flush_tlb_batched_pending(struct mm_struct *mm) argument
[all...]
H A Dpgtable.h101 #define __pte_clear(mm, addr, ptep) \
153 #define pte_accessible(mm, pte) \
154 (mm_tlb_flush_pending(mm) ? pte_present(pte) : pte_valid(pte))
312 static inline void __check_safe_pte_update(struct mm_struct *mm, pte_t *ptep, argument
324 if (mm != current->active_mm && atomic_read(&mm->mm_users) <= 1)
375 static inline void __set_ptes(struct mm_struct *mm, argument
379 page_table_check_ptes_set(mm, ptep, pte, nr);
383 __check_safe_pte_update(mm, ptep, pte);
551 static inline void __set_pte_at(struct mm_struct *mm, argument
560 set_pmd_at(struct mm_struct *mm, unsigned long addr, pmd_t *pmdp, pmd_t pmd) argument
568 set_pud_at(struct mm_struct *mm, unsigned long addr, pud_t *pudp, pud_t pud) argument
786 mm_pud_folded(const struct mm_struct *mm) argument
910 mm_p4d_folded(const struct mm_struct *mm) argument
1143 __ptep_get_and_clear(struct mm_struct *mm, unsigned long address, pte_t *ptep) argument
1153 __clear_full_ptes(struct mm_struct *mm, unsigned long addr, pte_t *ptep, unsigned int nr, int full) argument
1165 __get_and_clear_full_ptes(struct mm_struct *mm, unsigned long addr, pte_t *ptep, unsigned int nr, int full) argument
1186 pmdp_huge_get_and_clear(struct mm_struct *mm, unsigned long address, pmd_t *pmdp) argument
1197 ___ptep_set_wrprotect(struct mm_struct *mm, unsigned long address, pte_t *ptep, pte_t pte) argument
1215 __ptep_set_wrprotect(struct mm_struct *mm, unsigned long address, pte_t *ptep) argument
1221 __wrprotect_ptes(struct mm_struct *mm, unsigned long address, pte_t *ptep, unsigned int nr) argument
1232 pmdp_set_wrprotect(struct mm_struct *mm, unsigned long address, pmd_t *pmdp) argument
1396 contpte_try_fold(struct mm_struct *mm, unsigned long addr, pte_t *ptep, pte_t pte) argument
1419 contpte_try_unfold(struct mm_struct *mm, unsigned long addr, pte_t *ptep, pte_t pte) argument
1491 set_ptes(struct mm_struct *mm, unsigned long addr, pte_t *ptep, pte_t pte, unsigned int nr) argument
1505 pte_clear(struct mm_struct *mm, unsigned long addr, pte_t *ptep) argument
1513 clear_full_ptes(struct mm_struct *mm, unsigned long addr, pte_t *ptep, unsigned int nr, int full) argument
1525 get_and_clear_full_ptes(struct mm_struct *mm, unsigned long addr, pte_t *ptep, unsigned int nr, int full) argument
1542 ptep_get_and_clear(struct mm_struct *mm, unsigned long addr, pte_t *ptep) argument
1574 wrprotect_ptes(struct mm_struct *mm, unsigned long addr, pte_t *ptep, unsigned int nr) argument
1598 ptep_set_wrprotect(struct mm_struct *mm, unsigned long addr, pte_t *ptep) argument
[all...]
/linux-master/arch/powerpc/mm/book3s64/
H A Dslb.c32 static long slb_allocate_user(struct mm_struct *mm, unsigned long ea);
334 struct mm_struct *mm = current->mm; local
351 * after we switch to the mm and have already preloaded the SLBEs.
366 slb_allocate_user(mm, exec);
370 if (!is_kernel_addr(mm->mmap_base)) {
371 if (preload_add(ti, mm->mmap_base))
372 slb_allocate_user(mm, mm->mmap_base);
384 struct mm_struct *mm local
445 switch_slb(struct task_struct *tsk, struct mm_struct *mm) argument
786 slb_allocate_user(struct mm_struct *mm, unsigned long ea) argument
858 struct mm_struct *mm = current->mm; local
[all...]
/linux-master/fs/proc/
H A Darray.c65 #include <linux/sched/mm.h>
73 #include <linux/mm.h>
420 static inline void task_thp_status(struct seq_file *m, struct mm_struct *mm) argument
425 thp_enabled = !test_bit(MMF_DISABLE_THP, &mm->flags);
429 static inline void task_untag_mask(struct seq_file *m, struct mm_struct *mm) argument
431 seq_printf(m, "untag_mask:\t%#lx\n", mm_untag_mask(mm));
442 struct mm_struct *mm = get_task_mm(task); local
450 if (mm) {
451 task_mem(m, mm);
453 task_thp_status(m, mm);
478 struct mm_struct *mm; local
686 struct mm_struct *mm = get_task_mm(task); local
[all...]
/linux-master/arch/x86/math-emu/
H A Dfpu_system.h18 #include <linux/mm.h>
30 mutex_lock(&current->mm->context.lock);
31 if (current->mm->context.ldt && seg < current->mm->context.ldt->nr_entries)
32 ret = current->mm->context.ldt->entries[seg];
33 mutex_unlock(&current->mm->context.lock);
/linux-master/arch/x86/mm/
H A Dhugetlbpage.c10 #include <linux/mm.h>
11 #include <linux/sched/mm.h>
117 struct mm_struct *mm = current->mm; local
138 vma = find_vma(mm, addr);
144 if (mm->get_unmapped_area == arch_get_unmapped_area)
/linux-master/arch/powerpc/include/asm/nohash/32/
H A Dpte-8xx.h123 static inline pte_basic_t pte_update(struct mm_struct *mm, unsigned long addr, pte_t *p,
126 static inline void ptep_set_wrprotect(struct mm_struct *mm, unsigned long addr, pte_t *ptep) argument
128 pte_update(mm, addr, ptep, 0, _PAGE_RO, 0);
173 static pmd_t *pmd_off(struct mm_struct *mm, unsigned long addr);
188 static inline pte_basic_t pte_update(struct mm_struct *mm, unsigned long addr, pte_t *p, argument
195 pmd_t *pmd = pmd_off(mm, addr);
/linux-master/arch/x86/kernel/
H A Dstep.c7 #include <linux/mm.h>
38 mutex_lock(&child->mm->context.lock);
39 if (unlikely(!child->mm->context.ldt ||
40 seg >= child->mm->context.ldt->nr_entries))
43 desc = &child->mm->context.ldt->entries[seg];
51 mutex_unlock(&child->mm->context.lock);
H A Dshstk.c12 #include <linux/mm.h>
104 struct mm_struct *mm = current->mm; local
110 mmap_write_lock(mm);
113 mmap_write_unlock(mm);
314 mmap_read_lock_killable(current->mm);
321 vma = find_vma(current->mm, *ssp);
327 mmap_read_unlock(current->mm);
343 mmap_read_unlock(current->mm);
417 * the same mm struc
[all...]
/linux-master/tools/testing/selftests/ptrace/
H A Dvmaccess.c27 char mm[64]; local
38 sprintf(mm, "/proc/%d/mem", pid);
39 f = open(mm, O_RDONLY);
/linux-master/arch/arm/kernel/
H A Dsuspend.c22 struct mm_struct *mm = current->active_mm; local
47 cpu_switch_mm(mm->pgd, mm);
/linux-master/tools/testing/selftests/kvm/x86_64/
H A Dnx_huge_pages_test.sh14 HUGE_PAGES=$(cat /sys/kernel/mm/hugepages/hugepages-2048kB/nr_hugepages)
32 sudo_echo "$(( $HUGE_PAGES + 3 ))" /sys/kernel/mm/hugepages/hugepages-2048kB/nr_hugepages
56 sudo_echo "$HUGE_PAGES" /sys/kernel/mm/hugepages/hugepages-2048kB/nr_hugepages
/linux-master/arch/sh/include/asm/
H A Dcacheflush.h5 #include <linux/mm.h>
11 * - flush_cache_mm(mm) flushes the specified mm context's cache lines
12 * - flush_cache_dup mm(mm) handles cache flushing when forking
13 * - flush_cache_page(mm, vmaddr, pfn) flushes a single page
38 extern void flush_cache_mm(struct mm_struct *mm);
39 extern void flush_cache_dup_mm(struct mm_struct *mm);
/linux-master/arch/mips/ath25/
H A Dearly_printk.c9 #include <linux/mm.h>
/linux-master/arch/sparc/include/asm/
H A Duprobes.h40 extern int arch_uprobe_analyze_insn(struct arch_uprobe *aup, struct mm_struct *mm, unsigned long addr);
/linux-master/arch/arm/mm/
H A Dproc-syms.c3 * linux/arch/arm/mm/proc-syms.c
8 #include <linux/mm.h>
/linux-master/include/linux/ceph/
H A Dbuffer.h6 #include <linux/mm.h>
/linux-master/security/
H A Dmin_addr.c3 #include <linux/mm.h>
/linux-master/arch/m68k/include/asm/
H A Dcacheflush_no.h8 #include <linux/mm.h>

Completed in 420 milliseconds

<<11121314151617181920>>