/linux-master/arch/csky/include/asm/ |
H A D | mmu_context.h | 17 #define cpu_asid(mm) (atomic64_read(&mm->context.asid) & ASID_MASK) 19 #define init_new_context(tsk,mm) ({ atomic64_set(&(mm)->context.asid, 0); 0; }) 21 void check_and_switch_context(struct mm_struct *mm, unsigned int cpu);
|
/linux-master/arch/microblaze/include/asm/ |
H A D | pgalloc.h | 29 #define pgd_alloc(mm) get_pgd() 31 extern pte_t *pte_alloc_one_kernel(struct mm_struct *mm); 33 #define __pte_free_tlb(tlb, pte, addr) pte_free((tlb)->mm, (pte)) 35 #define pmd_populate(mm, pmd, pte) \ 38 #define pmd_populate_kernel(mm, pmd, pte) \
|
/linux-master/arch/alpha/include/asm/ |
H A D | cacheflush.h | 5 #include <linux/mm.h> 30 that icache entries are tagged with the ASN and load a new mm context. */ 42 struct mm_struct *mm = vma->vm_mm; local 43 if (current->active_mm == mm) 44 __load_new_mm_context(mm); 46 mm->context[smp_processor_id()] = 0;
|
/linux-master/mm/ |
H A D | msync.c | 3 * linux/mm/msync.c 12 #include <linux/mm.h> 35 struct mm_struct *mm = current->mm; local 62 mmap_read_lock(mm); 63 vma = find_vma(mm, start); 95 mmap_read_unlock(mm); 100 mmap_read_lock(mm); 101 vma = find_vma(mm, start); 107 vma = find_vma(mm, vm [all...] |
/linux-master/arch/loongarch/include/asm/ |
H A D | pgalloc.h | 8 #include <linux/mm.h> 15 static inline void pmd_populate_kernel(struct mm_struct *mm, argument 21 static inline void pmd_populate(struct mm_struct *mm, pmd_t *pmd, pgtable_t pte) argument 28 static inline void pud_populate(struct mm_struct *mm, pud_t *pud, pmd_t *pmd) argument 36 static inline void p4d_populate(struct mm_struct *mm, p4d_t *p4d, pud_t *pud) argument 45 extern pgd_t *pgd_alloc(struct mm_struct *mm); 55 static inline pmd_t *pmd_alloc_one(struct mm_struct *mm, unsigned long address) argument 74 #define __pmd_free_tlb(tlb, x, addr) pmd_free((tlb)->mm, x) 80 static inline pud_t *pud_alloc_one(struct mm_struct *mm, unsigned long address) argument 94 #define __pud_free_tlb(tlb, x, addr) pud_free((tlb)->mm, [all...] |
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_amdkfd_fence.c | 29 #include <linux/sched/mm.h> 64 struct mm_struct *mm, 74 mmgrab(mm); 75 fence->mm = mm; 130 if (!kgd2kfd_schedule_evict_and_restore_process(fence->mm, f)) 157 mmdrop(fence->mm); 162 * amdkfd_fence_check_mm - Check whether to prevent eviction of @f by @mm 165 * @mm: [IN] mm tha 63 amdgpu_amdkfd_fence_create(u64 context, struct mm_struct *mm, struct svm_range_bo *svm_bo) argument 171 amdkfd_fence_check_mm(struct dma_fence *f, struct mm_struct *mm) argument [all...] |
/linux-master/arch/csky/kernel/ |
H A D | vdso.c | 7 #include <linux/mm.h> 48 struct mm_struct *mm = current->mm; local 54 mmap_write_lock(mm); 62 * Put vDSO base into mm struct. We need to do this before calling 66 mm->context.vdso = (void *)vdso_base; 69 install_special_mapping(mm, vdso_base, vdso_pages << PAGE_SHIFT, 74 mm->context.vdso = NULL; 79 ret = install_special_mapping(mm, vdso_base, PAGE_SIZE, 83 mm [all...] |
/linux-master/include/trace/events/ |
H A D | huge_memory.h | 58 TP_PROTO(struct mm_struct *mm, struct page *page, bool writable, 61 TP_ARGS(mm, page, writable, referenced, none_or_zero, status, unmapped), 64 __field(struct mm_struct *, mm) 74 __entry->mm = mm; 83 TP_printk("mm=%p, scan_pfn=0x%lx, writable=%d, referenced=%d, none_or_zero=%d, status=%s, unmapped=%d", 84 __entry->mm, 95 TP_PROTO(struct mm_struct *mm, int isolated, int status), 97 TP_ARGS(mm, isolated, status), 100 __field(struct mm_struct *, mm) [all...] |
/linux-master/arch/powerpc/mm/book3s64/ |
H A D | subpage_prot.c | 22 void subpage_prot_free(struct mm_struct *mm) argument 24 struct subpage_prot_table *spt = mm_ctx_subpage_prot(&mm->context); 53 static void hpte_flush_range(struct mm_struct *mm, unsigned long addr, argument 63 pgd = pgd_offset(mm, addr); 73 pte = pte_offset_map_lock(mm, pmd, addr, &ptl); 78 pte_update(mm, addr, pte, 0, 0, 0); 92 struct mm_struct *mm = current->mm; local 99 mmap_write_lock(mm); 101 spt = mm_ctx_subpage_prot(&mm 151 subpage_mark_vma_nohuge(struct mm_struct *mm, unsigned long addr, unsigned long len) argument 167 subpage_mark_vma_nohuge(struct mm_struct *mm, unsigned long addr, unsigned long len) argument 187 struct mm_struct *mm = current->mm; local [all...] |
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_shrinker.c | 8 #include <linux/sched/mm.h> 37 return swap_available() || obj->mm.madv == I915_MADV_DONTNEED; 91 * (like cpu mmaps) or the mm core has reused them before we could grab them. 113 { &i915->mm.purge_list, ~0u }, 115 &i915->mm.shrink_list, 194 spin_lock_irqsave(&i915->mm.obj_lock, flags); 198 mm.link))) { 199 list_move_tail(&obj->mm.link, &still_in_list); 202 !is_vmalloc_addr(obj->mm.mapping)) 215 spin_unlock_irqrestore(&i915->mm [all...] |
/linux-master/arch/powerpc/include/asm/book3s/64/ |
H A D | pgalloc.h | 25 static inline pgd_t *radix__pgd_alloc(struct mm_struct *mm) argument 28 return (pgd_t *)__get_free_page(pgtable_gfp_flags(mm, PGALLOC_GFP)); 31 page = alloc_pages(pgtable_gfp_flags(mm, PGALLOC_GFP | __GFP_RETRY_MAYFAIL), 39 static inline void radix__pgd_free(struct mm_struct *mm, pgd_t *pgd) argument 48 static inline pgd_t *pgd_alloc(struct mm_struct *mm) argument 53 return radix__pgd_alloc(mm); 56 pgtable_gfp_flags(mm, GFP_KERNEL)); 81 static inline void pgd_free(struct mm_struct *mm, pgd_t *pgd) argument 84 return radix__pgd_free(mm, pgd); 88 static inline void p4d_populate(struct mm_struct *mm, p4d_ argument 93 pud_alloc_one(struct mm_struct *mm, unsigned long addr) argument 126 pud_free(struct mm_struct *mm, pud_t *pud) argument 131 pud_populate(struct mm_struct *mm, pud_t *pud, pmd_t *pmd) argument 142 pmd_alloc_one(struct mm_struct *mm, unsigned long addr) argument 147 pmd_free(struct mm_struct *mm, pmd_t *pmd) argument 158 pmd_populate_kernel(struct mm_struct *mm, pmd_t *pmd, pte_t *pte) argument 164 pmd_populate(struct mm_struct *mm, pmd_t *pmd, pgtable_t pte_page) argument [all...] |
/linux-master/arch/arm64/include/asm/ |
H A D | pgalloc.h | 29 static inline void pud_populate(struct mm_struct *mm, pud_t *pudp, pmd_t *pmdp) argument 33 pudval |= (mm == &init_mm) ? PUD_TABLE_UXN : PUD_TABLE_PXN; 51 static inline void p4d_populate(struct mm_struct *mm, p4d_t *p4dp, pud_t *pudp) argument 55 p4dval |= (mm == &init_mm) ? P4D_TABLE_UXN : P4D_TABLE_PXN; 59 static inline void pud_free(struct mm_struct *mm, pud_t *pud) argument 63 __pud_free(mm, pud); 80 static inline void pgd_populate(struct mm_struct *mm, pgd_t *pgdp, p4d_t *p4dp) argument 84 pgdval |= (mm == &init_mm) ? PGD_TABLE_UXN : PGD_TABLE_PXN; 88 static inline p4d_t *p4d_alloc_one(struct mm_struct *mm, unsigned long addr) argument 92 if (mm 97 p4d_free(struct mm_struct *mm, p4d_t *p4d) argument 127 pmd_populate_kernel(struct mm_struct *mm, pmd_t *pmdp, pte_t *ptep) argument 134 pmd_populate(struct mm_struct *mm, pmd_t *pmdp, pgtable_t ptep) argument [all...] |
/linux-master/arch/s390/mm/ |
H A D | pgtable.c | 11 #include <linux/mm.h> 47 static inline void ptep_ipte_local(struct mm_struct *mm, unsigned long addr, argument 54 asce = READ_ONCE(mm->context.gmap_asce); 58 asce = asce ? : mm->context.asce; 67 static inline void ptep_ipte_global(struct mm_struct *mm, unsigned long addr, argument 74 asce = READ_ONCE(mm->context.gmap_asce); 78 asce = asce ? : mm->context.asce; 87 static inline pte_t ptep_flush_direct(struct mm_struct *mm, argument 96 atomic_inc(&mm->context.flush_count); 98 cpumask_equal(mm_cpumask(mm), cpumask_o 106 ptep_flush_lazy(struct mm_struct *mm, unsigned long addr, pte_t *ptep, int nodat) argument 164 pgste_update_all(pte_t pte, pgste_t pgste, struct mm_struct *mm) argument 185 pgste_set_key(pte_t *ptep, pgste_t pgste, pte_t entry, struct mm_struct *mm) argument 230 pgste_pte_notify(struct mm_struct *mm, unsigned long addr, pte_t *ptep, pgste_t pgste) argument 246 ptep_xchg_start(struct mm_struct *mm, unsigned long addr, pte_t *ptep) argument 258 ptep_xchg_commit(struct mm_struct *mm, unsigned long addr, pte_t *ptep, pgste_t pgste, pte_t old, pte_t new) argument 279 ptep_xchg_direct(struct mm_struct *mm, unsigned long addr, pte_t *ptep, pte_t new) argument 300 ptep_reset_dat_prot(struct mm_struct *mm, unsigned long addr, pte_t *ptep, pte_t new) argument 321 ptep_xchg_lazy(struct mm_struct *mm, unsigned long addr, pte_t *ptep, pte_t new) argument 344 struct mm_struct *mm = vma->vm_mm; local 361 struct mm_struct *mm = vma->vm_mm; local 376 pmdp_idte_local(struct mm_struct *mm, unsigned long addr, pmd_t *pmdp) argument 388 pmdp_idte_global(struct mm_struct *mm, unsigned long addr, pmd_t *pmdp) argument 407 pmdp_flush_direct(struct mm_struct *mm, unsigned long addr, pmd_t *pmdp) argument 425 pmdp_flush_lazy(struct mm_struct *mm, unsigned long addr, pmd_t *pmdp) argument 448 pmd_lookup(struct mm_struct *mm, unsigned long addr, pmd_t **pmdp) argument 481 pmdp_xchg_direct(struct mm_struct *mm, unsigned long addr, pmd_t *pmdp, pmd_t new) argument 494 pmdp_xchg_lazy(struct mm_struct *mm, unsigned long addr, pmd_t *pmdp, pmd_t new) argument 507 pudp_idte_local(struct mm_struct *mm, unsigned long addr, pud_t *pudp) argument 517 pudp_idte_global(struct mm_struct *mm, unsigned long addr, pud_t *pudp) argument 533 pudp_flush_direct(struct mm_struct *mm, unsigned long addr, pud_t *pudp) argument 551 pudp_xchg_direct(struct mm_struct *mm, unsigned long addr, pud_t *pudp, pud_t new) argument 565 pgtable_trans_huge_deposit(struct mm_struct *mm, pmd_t *pmdp, pgtable_t pgtable) argument 580 pgtable_trans_huge_withdraw(struct mm_struct *mm, pmd_t *pmdp) argument 606 ptep_set_pte_at(struct mm_struct *mm, unsigned long addr, pte_t *ptep, pte_t entry) argument 621 ptep_set_notify(struct mm_struct *mm, unsigned long addr, pte_t *ptep) argument 643 ptep_force_prot(struct mm_struct *mm, unsigned long addr, pte_t *ptep, int prot, unsigned long bit) argument 678 ptep_shadow_pte(struct mm_struct *mm, unsigned long saddr, pte_t *sptep, pte_t *tptep, pte_t pte) argument 705 ptep_unshadow_pte(struct mm_struct *mm, unsigned long saddr, pte_t *ptep) argument 719 ptep_zap_swap_entry(struct mm_struct *mm, swp_entry_t entry) argument 731 ptep_zap_unused(struct mm_struct *mm, unsigned long addr, pte_t *ptep, int reset) argument 755 ptep_zap_key(struct mm_struct *mm, unsigned long addr, pte_t *ptep) argument 775 ptep_test_and_clear_uc(struct mm_struct *mm, unsigned long addr, pte_t *ptep) argument 802 set_guest_storage_key(struct mm_struct *mm, unsigned long addr, unsigned char key, bool nq) argument 882 cond_set_guest_storage_key(struct mm_struct *mm, unsigned long addr, unsigned char key, unsigned char *oldkey, bool nq, bool mr, bool mc) argument 913 reset_guest_reference_bit(struct mm_struct *mm, unsigned long addr) argument 975 get_guest_storage_key(struct mm_struct *mm, unsigned long addr, unsigned char *key) argument 1042 pgste_perform_essa(struct mm_struct *mm, unsigned long hva, int orc, unsigned long *oldpte, unsigned long *oldpgste) argument 1148 set_pgste_bits(struct mm_struct *mm, unsigned long hva, unsigned long bits, unsigned long value) argument 1181 get_pgste(struct mm_struct *mm, unsigned long hva, unsigned long *pgstep) argument [all...] |
/linux-master/arch/sh/include/asm/ |
H A D | mmu_context.h | 40 #define cpu_context(cpu, mm) ((mm)->context.id[cpu]) 42 #define cpu_asid(cpu, mm) \ 43 (cpu_context((cpu), (mm)) & MMU_CONTEXT_ASID_MASK) 55 static inline void get_mmu_context(struct mm_struct *mm, unsigned int cpu) argument 60 if (((cpu_context(cpu, mm) ^ asid) & MMU_CONTEXT_VERSION_MASK) == 0) 80 cpu_context(cpu, mm) = asid_cache(cpu) = asid; 89 struct mm_struct *mm) 94 cpu_context(i, mm) = NO_CONTEXT; 100 * After we have set current->mm t 88 init_new_context(struct task_struct *tsk, struct mm_struct *mm) argument 103 activate_context(struct mm_struct *mm, unsigned int cpu) argument [all...] |
/linux-master/arch/arm/include/asm/ |
H A D | pgalloc.h | 26 static inline void pud_populate(struct mm_struct *mm, pud_t *pud, pmd_t *pmd) argument 37 #define pmd_alloc_one(mm,addr) ({ BUG(); ((pmd_t *)2); }) 38 #define pmd_free(mm, pmd) do { } while (0) 41 #define pud_populate(mm,pmd,pte) do { } while (0) 43 #define pud_populate(mm,pmd,pte) BUG() 47 extern pgd_t *pgd_alloc(struct mm_struct *mm); 48 extern void pgd_free(struct mm_struct *mm, pgd_t *pgd); 78 pte_alloc_one_kernel(struct mm_struct *mm) argument 80 pte_t *pte = __pte_alloc_one_kernel(mm); 95 pte_alloc_one(struct mm_struct *mm) argument 125 pmd_populate_kernel(struct mm_struct *mm, pmd_t *pmdp, pte_t *ptep) argument 134 pmd_populate(struct mm_struct *mm, pmd_t *pmdp, pgtable_t ptep) argument [all...] |
/linux-master/arch/powerpc/mm/ |
H A D | copro_fault.c | 3 * CoProcessor (SPU/AFU) mm fault handler 11 #include <linux/mm.h> 23 int copro_handle_mm_fault(struct mm_struct *mm, unsigned long ea, argument 30 if (mm == NULL) 33 if (mm->pgd == NULL) 36 vma = lock_mm_and_find_vma(mm, ea, NULL); 77 mmap_read_unlock(mm); 83 int copro_calculate_slb(struct mm_struct *mm, u64 ea, struct copro_slb *slb) argument 91 if (mm == NULL) 93 psize = get_slice_psize(mm, e 139 copro_flush_all_slbs(struct mm_struct *mm) argument [all...] |
/linux-master/arch/riscv/include/asm/ |
H A D | pgalloc.h | 10 #include <linux/mm.h> 18 static inline void pmd_populate_kernel(struct mm_struct *mm, argument 26 static inline void pmd_populate(struct mm_struct *mm, argument 35 static inline void pud_populate(struct mm_struct *mm, pud_t *pud, pmd_t *pmd) argument 42 static inline void p4d_populate(struct mm_struct *mm, p4d_t *p4d, pud_t *pud) argument 51 static inline void p4d_populate_safe(struct mm_struct *mm, p4d_t *p4d, argument 62 static inline void pgd_populate(struct mm_struct *mm, pgd_t *pgd, p4d_t *p4d) argument 71 static inline void pgd_populate_safe(struct mm_struct *mm, pgd_t *pgd, argument 83 static inline pud_t *pud_alloc_one(struct mm_struct *mm, unsigned long addr) argument 86 return __pud_alloc_one(mm, add 92 pud_free(struct mm_struct *mm, pud_t *pud) argument 113 p4d_alloc_one(struct mm_struct *mm, unsigned long addr) argument 126 __p4d_free(struct mm_struct *mm, p4d_t *p4d) argument 133 p4d_free(struct mm_struct *mm, p4d_t *p4d) argument 158 pgd_alloc(struct mm_struct *mm) argument [all...] |
H A D | tlbflush.h | 37 void flush_tlb_mm(struct mm_struct *mm); 38 void flush_tlb_mm_range(struct mm_struct *mm, unsigned long start, 51 bool arch_tlbbatch_should_defer(struct mm_struct *mm); 53 struct mm_struct *mm, 55 void arch_flush_tlb_batched_pending(struct mm_struct *mm); 76 #define flush_tlb_mm(mm) flush_tlb_all() 77 #define flush_tlb_mm_range(mm, start, end, page_size) flush_tlb_all()
|
/linux-master/arch/sh/kernel/vsyscall/ |
H A D | vsyscall.c | 10 #include <linux/mm.h> 60 struct mm_struct *mm = current->mm; local 64 if (mmap_write_lock_killable(mm)) 73 ret = install_special_mapping(mm, addr, PAGE_SIZE, 80 current->mm->context.vdso = (void *)addr; 83 mmap_write_unlock(mm);
|
/linux-master/include/drm/ |
H A D | drm_mm.h | 165 struct drm_mm *mm; member in struct:drm_mm_node 229 struct drm_mm *mm; member in struct:drm_mm_scan 265 * @mm: drm_mm to check 274 * True if the @mm is initialized. 276 static inline bool drm_mm_initialized(const struct drm_mm *mm) argument 278 return READ_ONCE(mm->hole_stack.next); 343 * @mm: the struct drm_mm range manager 353 #define drm_mm_nodes(mm) (&(mm)->head_node.node_list) 358 * @mm 434 drm_mm_insert_node_generic(struct drm_mm *mm, struct drm_mm_node *node, u64 size, u64 alignment, unsigned long color, enum drm_mm_insert_mode mode) argument 458 drm_mm_insert_node(struct drm_mm *mm, struct drm_mm_node *node, u64 size) argument 478 drm_mm_clean(const struct drm_mm *mm) argument 533 drm_mm_scan_init(struct drm_mm_scan *scan, struct drm_mm *mm, u64 size, u64 alignment, unsigned long color, enum drm_mm_insert_mode mode) argument [all...] |
/linux-master/arch/sparc/include/asm/ |
H A D | mman.h | 16 struct mm_struct *mm = arg; local 22 if (current->mm == mm) { 36 if (!current->mm->context.adi) { 39 current->mm->context.adi = true; 40 on_each_cpu_mask(mm_cpumask(current->mm), 41 ipi_set_tstate_mcde, current->mm, 0);
|
/linux-master/arch/hexagon/include/asm/ |
H A D | pgalloc.h | 21 static inline pgd_t *pgd_alloc(struct mm_struct *mm) argument 36 mm->context.generation = kmap_generation; 39 mm->context.ptbase = __pa(pgd); 44 static inline void pmd_populate(struct mm_struct *mm, pmd_t *pmd, argument 64 static inline void pmd_populate_kernel(struct mm_struct *mm, pmd_t *pmd, argument 73 mm->context.generation = kmap_generation; 83 pmdindex = (pgd_t *)pmd - mm->pgd;
|
/linux-master/arch/hexagon/kernel/ |
H A D | vdso.c | 9 #include <linux/mm.h> 54 struct mm_struct *mm = current->mm; local 56 if (mmap_write_lock_killable(mm)) 69 ret = install_special_mapping(mm, vdso_base, PAGE_SIZE, 77 mm->context.vdso = (void *)vdso_base; 80 mmap_write_unlock(mm);
|
/linux-master/arch/csky/abiv2/ |
H A D | cacheflush.c | 6 #include <linux/mm.h> 39 void flush_icache_deferred(struct mm_struct *mm) argument 42 cpumask_t *mask = &mm->context.icache_stale_mask; 55 void flush_icache_mm_range(struct mm_struct *mm, argument 64 if (mm == current->mm) { 72 mask = &mm->context.icache_stale_mask; 84 cpumask_andnot(&others, mm_cpumask(mm), cpumask_of(cpu)); 86 if (mm != current->active_mm || !cpumask_empty(&others)) {
|
/linux-master/arch/powerpc/include/asm/nohash/32/ |
H A D | pgalloc.h | 12 /* #define pmd_alloc_one(mm,address) ({ BUG(); ((pmd_t *)2); }) */ 13 #define pmd_free(mm, x) do { } while (0) 15 /* #define pgd_populate(mm, pmd, pte) BUG() */ 17 static inline void pmd_populate_kernel(struct mm_struct *mm, pmd_t *pmdp, argument 26 static inline void pmd_populate(struct mm_struct *mm, pmd_t *pmdp, argument
|