/linux-master/arch/s390/mm/ |
H A D | gmap.c | 103 * @mm: pointer to the parent mm_struct 108 struct gmap *gmap_create(struct mm_struct *mm, unsigned long limit) argument 116 gmap->mm = mm; 117 spin_lock(&mm->context.lock); 118 list_add_rcu(&gmap->list, &mm->context.gmap_list); 119 if (list_is_singular(&mm->context.gmap_list)) 123 WRITE_ONCE(mm->context.gmap_asce, gmap_asce); 124 spin_unlock(&mm->context.lock); 266 /* Remove gmap from the pre-mm lis 523 gmap_unlink(struct mm_struct *mm, unsigned long *table, unsigned long vmaddr) argument 554 struct mm_struct *mm; local 895 struct mm_struct *mm = gmap->mm; local 2267 ptep_notify(struct mm_struct *mm, unsigned long vmaddr, pte_t *pte, unsigned long bits) argument 2334 gmap_pmdp_clear(struct mm_struct *mm, unsigned long vmaddr, int purge) argument 2366 gmap_pmdp_invalidate(struct mm_struct *mm, unsigned long vmaddr) argument 2377 gmap_pmdp_csp(struct mm_struct *mm, unsigned long vmaddr) argument 2388 gmap_pmdp_idte_local(struct mm_struct *mm, unsigned long vmaddr) argument 2423 gmap_pmdp_idte_global(struct mm_struct *mm, unsigned long vmaddr) argument 2535 thp_split_mm(struct mm_struct *mm) argument 2547 thp_split_mm(struct mm_struct *mm) argument 2592 struct mm_struct *mm = current->mm; local 2680 struct mm_struct *mm = current->mm; local 2716 s390_reset_cmma(struct mm_struct *mm) argument 2785 __s390_uv_destroy_range(struct mm_struct *mm, unsigned long start, unsigned long end, bool interruptible) argument [all...] |
/linux-master/fs/ |
H A D | exec.c | 30 #include <linux/mm.h> 36 #include <linux/sched/mm.h> 181 * The nascent bprm->mm is not visible until exec_mmap() but it can 182 * use a lot of memory, account these pages in current->mm temporary 188 struct mm_struct *mm = current->mm; local 191 if (!mm || !diff) 195 add_mm_counter(mm, MM_ANONPAGES, diff); 203 struct mm_struct *mm = bprm->mm; local 258 struct mm_struct *mm = bprm->mm; local 372 struct mm_struct *mm = NULL; local 686 struct mm_struct *mm = vma->vm_mm; local 756 struct mm_struct *mm = current->mm; local 996 exec_mmap(struct mm_struct *mm) argument 2102 struct mm_struct *mm = current->mm; local 2116 set_dumpable(struct mm_struct *mm, int value) argument [all...] |
/linux-master/arch/um/kernel/ |
H A D | trap.c | 6 #include <linux/mm.h> 27 struct mm_struct *mm = current->mm; local 46 mmap_read_lock(mm); 47 vma = find_vma(mm, address); 56 vma = expand_stack(mm, address); 101 pmd = pmd_off(mm, address); 118 mmap_read_unlock(mm); 127 mmap_read_unlock(mm); 217 else if (current->mm [all...] |
/linux-master/arch/csky/mm/ |
H A D | fault.c | 100 static inline void bad_area_nosemaphore(struct pt_regs *regs, struct mm_struct *mm, int code, unsigned long addr) argument 191 struct mm_struct *mm; local 198 mm = tsk->mm; 227 if (unlikely(faulthandler_disabled() || !mm)) { 240 vma = lock_mm_and_find_vma(mm, addr, regs); 242 bad_area_nosemaphore(regs, mm, code, addr); 253 mmap_read_unlock(mm); 254 bad_area_nosemaphore(regs, mm, code, addr); 268 * would already be released in __lock_page_or_retry in mm/filema [all...] |
/linux-master/arch/powerpc/include/asm/nohash/ |
H A D | pgtable.h | 6 static inline pte_basic_t pte_update(struct mm_struct *mm, unsigned long addr, pte_t *p, 50 static inline pte_basic_t pte_update(struct mm_struct *mm, unsigned long addr, pte_t *p, argument 66 assert_pte_locked(mm, addr); 84 static inline void ptep_set_wrprotect(struct mm_struct *mm, unsigned long addr, argument 87 pte_update(mm, addr, ptep, _PAGE_WRITE, 0, 0); 92 static inline pte_t ptep_get_and_clear(struct mm_struct *mm, unsigned long addr, argument 95 return __pte(pte_update(mm, addr, ptep, ~0UL, 0, 0)); 99 static inline void pte_clear(struct mm_struct *mm, unsigned long addr, pte_t *ptep) argument 101 pte_update(mm, addr, ptep, ~0UL, 0, 0); 279 static inline void __set_pte_at(struct mm_struct *mm, unsigne argument [all...] |
/linux-master/arch/loongarch/mm/ |
H A D | tlb.c | 8 #include <linux/mm.h> 38 * All entries common to a mm share an asid. To effectively flush 41 void local_flush_tlb_mm(struct mm_struct *mm) argument 49 if (asid_valid(mm, cpu)) 50 drop_mmu_context(mm, cpu); 52 cpumask_clear_cpu(cpu, mm_cpumask(mm)); 60 struct mm_struct *mm = vma->vm_mm; local 63 if (asid_valid(mm, cpu)) { 73 int asid = cpu_asid(cpu, mm); 80 drop_mmu_context(mm, cp [all...] |
/linux-master/arch/arm64/include/asm/ |
H A D | tlbflush.h | 199 * flush_tlb_mm(mm) 201 * The 'mm' argument identifies the ASID to invalidate. 205 * CPUs for the user address space corresponding to 'vma->mm'. 218 * address space corresponding to 'vma->mm'. Note that this 236 * CPUs for the user address space corresponding to 'vma->mm'. 266 static inline void flush_tlb_mm(struct mm_struct *mm) argument 271 asid = __TLBI_VADDR(0, ASID(mm)); 275 mmu_notifier_arch_invalidate_secondary_tlbs(mm, 0, -1UL); 278 static inline void __flush_tlb_page_nosync(struct mm_struct *mm, argument 284 addr = __TLBI_VADDR(uaddr, ASID(mm)); 304 arch_tlbbatch_should_defer(struct mm_struct *mm) argument 318 arch_tlbbatch_add_pending(struct arch_tlbflush_unmap_batch *batch, struct mm_struct *mm, unsigned long uaddr) argument 330 arch_flush_tlb_batched_pending(struct mm_struct *mm) argument [all...] |
H A D | pgtable.h | 101 #define __pte_clear(mm, addr, ptep) \ 153 #define pte_accessible(mm, pte) \ 154 (mm_tlb_flush_pending(mm) ? pte_present(pte) : pte_valid(pte)) 312 static inline void __check_safe_pte_update(struct mm_struct *mm, pte_t *ptep, argument 324 if (mm != current->active_mm && atomic_read(&mm->mm_users) <= 1) 375 static inline void __set_ptes(struct mm_struct *mm, argument 379 page_table_check_ptes_set(mm, ptep, pte, nr); 383 __check_safe_pte_update(mm, ptep, pte); 551 static inline void __set_pte_at(struct mm_struct *mm, argument 560 set_pmd_at(struct mm_struct *mm, unsigned long addr, pmd_t *pmdp, pmd_t pmd) argument 568 set_pud_at(struct mm_struct *mm, unsigned long addr, pud_t *pudp, pud_t pud) argument 786 mm_pud_folded(const struct mm_struct *mm) argument 910 mm_p4d_folded(const struct mm_struct *mm) argument 1143 __ptep_get_and_clear(struct mm_struct *mm, unsigned long address, pte_t *ptep) argument 1153 __clear_full_ptes(struct mm_struct *mm, unsigned long addr, pte_t *ptep, unsigned int nr, int full) argument 1165 __get_and_clear_full_ptes(struct mm_struct *mm, unsigned long addr, pte_t *ptep, unsigned int nr, int full) argument 1186 pmdp_huge_get_and_clear(struct mm_struct *mm, unsigned long address, pmd_t *pmdp) argument 1197 ___ptep_set_wrprotect(struct mm_struct *mm, unsigned long address, pte_t *ptep, pte_t pte) argument 1215 __ptep_set_wrprotect(struct mm_struct *mm, unsigned long address, pte_t *ptep) argument 1221 __wrprotect_ptes(struct mm_struct *mm, unsigned long address, pte_t *ptep, unsigned int nr) argument 1232 pmdp_set_wrprotect(struct mm_struct *mm, unsigned long address, pmd_t *pmdp) argument 1396 contpte_try_fold(struct mm_struct *mm, unsigned long addr, pte_t *ptep, pte_t pte) argument 1419 contpte_try_unfold(struct mm_struct *mm, unsigned long addr, pte_t *ptep, pte_t pte) argument 1491 set_ptes(struct mm_struct *mm, unsigned long addr, pte_t *ptep, pte_t pte, unsigned int nr) argument 1505 pte_clear(struct mm_struct *mm, unsigned long addr, pte_t *ptep) argument 1513 clear_full_ptes(struct mm_struct *mm, unsigned long addr, pte_t *ptep, unsigned int nr, int full) argument 1525 get_and_clear_full_ptes(struct mm_struct *mm, unsigned long addr, pte_t *ptep, unsigned int nr, int full) argument 1542 ptep_get_and_clear(struct mm_struct *mm, unsigned long addr, pte_t *ptep) argument 1574 wrprotect_ptes(struct mm_struct *mm, unsigned long addr, pte_t *ptep, unsigned int nr) argument 1598 ptep_set_wrprotect(struct mm_struct *mm, unsigned long addr, pte_t *ptep) argument [all...] |
/linux-master/arch/powerpc/mm/book3s64/ |
H A D | slb.c | 32 static long slb_allocate_user(struct mm_struct *mm, unsigned long ea); 334 struct mm_struct *mm = current->mm; local 351 * after we switch to the mm and have already preloaded the SLBEs. 366 slb_allocate_user(mm, exec); 370 if (!is_kernel_addr(mm->mmap_base)) { 371 if (preload_add(ti, mm->mmap_base)) 372 slb_allocate_user(mm, mm->mmap_base); 384 struct mm_struct *mm local 445 switch_slb(struct task_struct *tsk, struct mm_struct *mm) argument 786 slb_allocate_user(struct mm_struct *mm, unsigned long ea) argument 858 struct mm_struct *mm = current->mm; local [all...] |
/linux-master/fs/proc/ |
H A D | array.c | 65 #include <linux/sched/mm.h> 73 #include <linux/mm.h> 420 static inline void task_thp_status(struct seq_file *m, struct mm_struct *mm) argument 425 thp_enabled = !test_bit(MMF_DISABLE_THP, &mm->flags); 429 static inline void task_untag_mask(struct seq_file *m, struct mm_struct *mm) argument 431 seq_printf(m, "untag_mask:\t%#lx\n", mm_untag_mask(mm)); 442 struct mm_struct *mm = get_task_mm(task); local 450 if (mm) { 451 task_mem(m, mm); 453 task_thp_status(m, mm); 478 struct mm_struct *mm; local 686 struct mm_struct *mm = get_task_mm(task); local [all...] |
/linux-master/arch/x86/math-emu/ |
H A D | fpu_system.h | 18 #include <linux/mm.h> 30 mutex_lock(¤t->mm->context.lock); 31 if (current->mm->context.ldt && seg < current->mm->context.ldt->nr_entries) 32 ret = current->mm->context.ldt->entries[seg]; 33 mutex_unlock(¤t->mm->context.lock);
|
/linux-master/arch/x86/mm/ |
H A D | hugetlbpage.c | 10 #include <linux/mm.h> 11 #include <linux/sched/mm.h> 117 struct mm_struct *mm = current->mm; local 138 vma = find_vma(mm, addr); 144 if (mm->get_unmapped_area == arch_get_unmapped_area)
|
/linux-master/arch/powerpc/include/asm/nohash/32/ |
H A D | pte-8xx.h | 123 static inline pte_basic_t pte_update(struct mm_struct *mm, unsigned long addr, pte_t *p, 126 static inline void ptep_set_wrprotect(struct mm_struct *mm, unsigned long addr, pte_t *ptep) argument 128 pte_update(mm, addr, ptep, 0, _PAGE_RO, 0); 173 static pmd_t *pmd_off(struct mm_struct *mm, unsigned long addr); 188 static inline pte_basic_t pte_update(struct mm_struct *mm, unsigned long addr, pte_t *p, argument 195 pmd_t *pmd = pmd_off(mm, addr);
|
/linux-master/arch/x86/kernel/ |
H A D | step.c | 7 #include <linux/mm.h> 38 mutex_lock(&child->mm->context.lock); 39 if (unlikely(!child->mm->context.ldt || 40 seg >= child->mm->context.ldt->nr_entries)) 43 desc = &child->mm->context.ldt->entries[seg]; 51 mutex_unlock(&child->mm->context.lock);
|
H A D | shstk.c | 12 #include <linux/mm.h> 104 struct mm_struct *mm = current->mm; local 110 mmap_write_lock(mm); 113 mmap_write_unlock(mm); 314 mmap_read_lock_killable(current->mm); 321 vma = find_vma(current->mm, *ssp); 327 mmap_read_unlock(current->mm); 343 mmap_read_unlock(current->mm); 417 * the same mm struc [all...] |
/linux-master/tools/testing/selftests/ptrace/ |
H A D | vmaccess.c | 27 char mm[64]; local 38 sprintf(mm, "/proc/%d/mem", pid); 39 f = open(mm, O_RDONLY);
|
/linux-master/arch/arm/kernel/ |
H A D | suspend.c | 22 struct mm_struct *mm = current->active_mm; local 47 cpu_switch_mm(mm->pgd, mm);
|
/linux-master/tools/testing/selftests/kvm/x86_64/ |
H A D | nx_huge_pages_test.sh | 14 HUGE_PAGES=$(cat /sys/kernel/mm/hugepages/hugepages-2048kB/nr_hugepages) 32 sudo_echo "$(( $HUGE_PAGES + 3 ))" /sys/kernel/mm/hugepages/hugepages-2048kB/nr_hugepages 56 sudo_echo "$HUGE_PAGES" /sys/kernel/mm/hugepages/hugepages-2048kB/nr_hugepages
|
/linux-master/arch/sh/include/asm/ |
H A D | cacheflush.h | 5 #include <linux/mm.h> 11 * - flush_cache_mm(mm) flushes the specified mm context's cache lines 12 * - flush_cache_dup mm(mm) handles cache flushing when forking 13 * - flush_cache_page(mm, vmaddr, pfn) flushes a single page 38 extern void flush_cache_mm(struct mm_struct *mm); 39 extern void flush_cache_dup_mm(struct mm_struct *mm);
|
/linux-master/arch/mips/ath25/ |
H A D | early_printk.c | 9 #include <linux/mm.h>
|
/linux-master/arch/sparc/include/asm/ |
H A D | uprobes.h | 40 extern int arch_uprobe_analyze_insn(struct arch_uprobe *aup, struct mm_struct *mm, unsigned long addr);
|
/linux-master/arch/arm/mm/ |
H A D | proc-syms.c | 3 * linux/arch/arm/mm/proc-syms.c 8 #include <linux/mm.h>
|
/linux-master/include/linux/ceph/ |
H A D | buffer.h | 6 #include <linux/mm.h>
|
/linux-master/security/ |
H A D | min_addr.c | 3 #include <linux/mm.h>
|
/linux-master/arch/m68k/include/asm/ |
H A D | cacheflush_no.h | 8 #include <linux/mm.h>
|