/linux-master/arch/x86/kernel/ |
H A D | ldt.c | 19 #include <linux/mm.h> 42 void load_mm_ldt(struct mm_struct *mm) argument 47 ldt = READ_ONCE(mm->context.ldt); 50 * Any change to mm->context.ldt is followed by an IPI to all 51 * CPUs with the mm active. The LDT will not be freed until 93 * Load the LDT if either the old or new mm had an LDT. 95 * An mm will never go from having an LDT to not having an LDT. Two 138 struct mm_struct *mm = __mm; local 140 if (this_cpu_read(cpu_tlbstate.loaded_mm) != mm) 143 load_mm_ldt(mm); 189 do_sanity_check(struct mm_struct *mm, bool had_kernel_mapping, bool had_user_mapping) argument 234 map_ldt_struct_to_user(struct mm_struct *mm) argument 247 sanity_check_ldt_mapping(struct mm_struct *mm) argument 264 map_ldt_struct_to_user(struct mm_struct *mm) argument 272 sanity_check_ldt_mapping(struct mm_struct *mm) argument 288 map_ldt_struct(struct mm_struct *mm, struct ldt_struct *ldt, int slot) argument 349 unmap_ldt_struct(struct mm_struct *mm, struct ldt_struct *ldt) argument 383 map_ldt_struct(struct mm_struct *mm, struct ldt_struct *ldt, int slot) argument 388 unmap_ldt_struct(struct mm_struct *mm, struct ldt_struct *ldt) argument 393 free_ldt_pgtables(struct mm_struct *mm) argument 421 install_ldt(struct mm_struct *mm, struct ldt_struct *ldt) argument 429 on_each_cpu_mask(mm_cpumask(mm), flush_ldt, mm, true); local 451 ldt_dup_context(struct mm_struct *old_mm, struct mm_struct *mm) argument 491 destroy_context_ldt(struct mm_struct *mm) argument 497 ldt_arch_exit_mmap(struct mm_struct *mm) argument 504 struct mm_struct *mm = current->mm; local 580 struct mm_struct *mm = current->mm; local [all...] |
/linux-master/include/linux/sched/ |
H A D | mm.h | 19 * @mm: The &struct mm_struct to pin. 21 * Make sure that @mm will not get freed even after the owning task 26 * This is a preferred way to pin @mm for a longer/unbounded amount 31 * See also <Documentation/mm/active_mm.rst> for an in-depth explanation 34 static inline void mmgrab(struct mm_struct *mm) argument 36 atomic_inc(&mm->mm_count); 44 extern void __mmdrop(struct mm_struct *mm); 46 static inline void mmdrop(struct mm_struct *mm) argument 53 if (unlikely(atomic_dec_and_test(&mm->mm_count))) 54 __mmdrop(mm); 64 struct mm_struct *mm = container_of(rhp, struct mm_struct, delayed_drop); local 73 mmdrop_sched(struct mm_struct *mm) argument 80 mmdrop_sched(struct mm_struct *mm) argument 87 mmgrab_lazy_tlb(struct mm_struct *mm) argument 93 mmdrop_lazy_tlb(struct mm_struct *mm) argument 106 mmdrop_lazy_tlb_sched(struct mm_struct *mm) argument 130 mmget(struct mm_struct *mm) argument 135 mmget_not_zero(struct mm_struct *mm) argument 165 mm_update_next_owner(struct mm_struct *mm) argument 198 arch_pick_mmap_layout(struct mm_struct *mm, struct rlimit *rlim_stack) argument 526 membarrier_mm_sync_core_before_usermode(struct mm_struct *mm) argument 548 membarrier_exec_mmap(struct mm_struct *mm) argument 551 membarrier_mm_sync_core_before_usermode(struct mm_struct *mm) argument [all...] |
/linux-master/include/linux/ |
H A D | sync_core.h | 30 static inline void prepare_sync_core_cmd(struct mm_struct *mm) argument
|
H A D | aio.h | 15 extern void exit_aio(struct mm_struct *mm); 18 static inline void exit_aio(struct mm_struct *mm) { } argument
|
/linux-master/arch/arm64/include/asm/ |
H A D | mmu_context.h | 55 void cpu_do_switch_mm(phys_addr_t pgd_phys, struct mm_struct *mm); 57 static inline void cpu_switch_mm(pgd_t *pgd, struct mm_struct *mm) argument 60 cpu_do_switch_mm(virt_to_phys(pgd),mm); local 88 * Remove the idmap from TTBR0_EL1 and install the pgd of the active mm. 95 * If current is a not a user task, the mm covers the TTBR1_EL1 page tables, 101 struct mm_struct *mm = current->active_mm; local 107 if (mm != &init_mm && !system_uses_ttbr0_pan()) 108 cpu_switch_mm(mm->pgd, mm); 165 * freeing the ASID from the context of the dying mm ( 174 init_new_context(struct task_struct *tsk, struct mm_struct *mm) argument 182 update_saved_ttbr0(struct task_struct *tsk, struct mm_struct *mm) argument 198 update_saved_ttbr0(struct task_struct *tsk, struct mm_struct *mm) argument 206 enter_lazy_tlb(struct mm_struct *mm, struct task_struct *tsk) argument 265 mm_untag_mask(struct mm_struct *mm) argument [all...] |
/linux-master/arch/powerpc/mm/ |
H A D | pgtable-frag.c | 10 #include <linux/mm.h> 33 static pte_t *get_pte_from_cache(struct mm_struct *mm) argument 40 spin_lock(&mm->page_table_lock); 41 ret = pte_frag_get(&mm->context); 49 pte_frag_set(&mm->context, pte_frag); 51 spin_unlock(&mm->page_table_lock); 55 static pte_t *__alloc_for_ptecache(struct mm_struct *mm, int kernel) argument 83 spin_lock(&mm->page_table_lock); 89 if (likely(!pte_frag_get(&mm->context))) { 91 pte_frag_set(&mm 98 pte_fragment_alloc(struct mm_struct *mm, int kernel) argument 137 pte_free_defer(struct mm_struct *mm, pgtable_t pgtable) argument [all...] |
/linux-master/arch/sparc/include/asm/ |
H A D | pgalloc_32.h | 26 #define pgd_free(mm, pgd) free_pgd_fast(pgd) 27 #define pgd_alloc(mm) get_pgd_fast() 38 static inline pmd_t *pmd_alloc_one(struct mm_struct *mm, argument 50 #define pmd_free(mm, pmd) free_pmd_fast(pmd) 51 #define __pmd_free_tlb(tlb, pmd, addr) pmd_free((tlb)->mm, pmd) 53 #define pmd_populate(mm, pmd, pte) pmd_set(pmd, pte) 58 pgtable_t pte_alloc_one(struct mm_struct *mm); 60 static inline pte_t *pte_alloc_one_kernel(struct mm_struct *mm) argument 72 #define pte_free_kernel(mm, pte) free_pte_fast(pte) 74 void pte_free(struct mm_struct * mm, pgtable_ [all...] |
H A D | tlbflush_32.h | 9 #define flush_tlb_mm(mm) \ 10 sparc32_cachetlb_ops->tlb_mm(mm)
|
/linux-master/arch/powerpc/mm/book3s32/ |
H A D | mmu_context.c | 9 * Derived from arch/ppc/mm/init.c: 16 * Derived from "arch/i386/mm/init.c" 20 #include <linux/mm.h> 69 int init_new_context(struct task_struct *t, struct mm_struct *mm) argument 71 mm->context.id = __init_new_context(); 72 mm->context.sr0 = CTX_TO_VSID(mm->context.id, 0); 75 mm->context.sr0 |= SR_NX; 77 mm->context.sr0 |= SR_KS; 94 void destroy_context(struct mm_struct *mm) argument [all...] |
/linux-master/arch/mips/include/asm/ |
H A D | pgalloc.h | 13 #include <linux/mm.h> 21 static inline void pmd_populate_kernel(struct mm_struct *mm, pmd_t *pmd, argument 27 static inline void pmd_populate(struct mm_struct *mm, pmd_t *pmd, argument 40 static inline void pud_populate(struct mm_struct *mm, pud_t *pud, pmd_t *pmd) argument 50 extern pgd_t *pgd_alloc(struct mm_struct *mm); 52 static inline void pgd_free(struct mm_struct *mm, pgd_t *pgd) argument 65 static inline pmd_t *pmd_alloc_one(struct mm_struct *mm, unsigned long address) argument 84 #define __pmd_free_tlb(tlb, x, addr) pmd_free((tlb)->mm, x) 90 static inline pud_t *pud_alloc_one(struct mm_struct *mm, unsigned long address) argument 105 static inline void p4d_populate(struct mm_struct *mm, p4d_ argument [all...] |
/linux-master/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_mqd_manager.c | 91 void free_mqd_hiq_sdma(struct mqd_manager *mm, void *mqd, argument 98 void mqd_symmetrically_map_cu_mask(struct mqd_manager *mm, argument 102 struct amdgpu_cu_info *cu_info = &mm->dev->adev->gfx.cu_info; 103 struct amdgpu_gfx_config *gfx_info = &mm->dev->adev->gfx.config; 105 bool wgp_mode_req = KFD_GC_VERSION(mm->dev) >= IP_VERSION(10, 0, 0); 109 int inc = cu_inc * NUM_XCC(mm->dev->xcc_mask); 110 int xcc_inst = inst + ffs(mm->dev->xcc_mask) - 1; 112 cu_active_per_node = cu_info->number / mm->dev->kfd->num_nodes; 131 cu_bitmap_sh_mul = (KFD_GC_VERSION(mm->dev) >= IP_VERSION(11, 0, 0) && 132 KFD_GC_VERSION(mm 206 kfd_hiq_load_mqd_kiq(struct mqd_manager *mm, void *mqd, uint32_t pipe_id, uint32_t queue_id, struct queue_properties *p, struct mm_struct *mms) argument 214 kfd_destroy_mqd_cp(struct mqd_manager *mm, void *mqd, enum kfd_preempt_type type, unsigned int timeout, uint32_t pipe_id, uint32_t queue_id) argument 222 kfd_free_mqd_cp(struct mqd_manager *mm, void *mqd, struct kfd_mem_obj *mqd_mem_obj) argument 233 kfd_is_occupied_cp(struct mqd_manager *mm, void *mqd, uint64_t queue_address, uint32_t pipe_id, uint32_t queue_id) argument 241 kfd_load_mqd_sdma(struct mqd_manager *mm, void *mqd, uint32_t pipe_id, uint32_t queue_id, struct queue_properties *p, struct mm_struct *mms) argument 254 kfd_destroy_mqd_sdma(struct mqd_manager *mm, void *mqd, enum kfd_preempt_type type, unsigned int timeout, uint32_t pipe_id, uint32_t queue_id) argument 262 kfd_is_occupied_sdma(struct mqd_manager *mm, void *mqd, uint64_t queue_address, uint32_t pipe_id, uint32_t queue_id) argument 288 kfd_mqd_stride(struct mqd_manager *mm, struct queue_properties *q) argument [all...] |
H A D | kfd_mqd_manager.h | 74 void (*init_mqd)(struct mqd_manager *mm, void **mqd, 78 int (*load_mqd)(struct mqd_manager *mm, void *mqd, 83 void (*update_mqd)(struct mqd_manager *mm, void *mqd, 87 int (*destroy_mqd)(struct mqd_manager *mm, void *mqd, 92 void (*free_mqd)(struct mqd_manager *mm, void *mqd, 95 bool (*is_occupied)(struct mqd_manager *mm, void *mqd, 99 int (*get_wave_state)(struct mqd_manager *mm, void *mqd, 105 void (*get_checkpoint_info)(struct mqd_manager *mm, void *mqd, uint32_t *ctl_stack_size); 107 void (*checkpoint_mqd)(struct mqd_manager *mm, 112 void (*restore_mqd)(struct mqd_manager *mm, voi [all...] |
/linux-master/arch/parisc/mm/ |
H A D | hugetlbpage.c | 11 #include <linux/mm.h> 12 #include <linux/sched/mm.h> 47 pte_t *huge_pte_alloc(struct mm_struct *mm, struct vm_area_struct *vma, argument 63 pgd = pgd_offset(mm, addr); 65 pud = pud_alloc(mm, p4d, addr); 67 pmd = pmd_alloc(mm, pud, addr); 69 pte = pte_alloc_huge(mm, pmd, addr); 74 pte_t *huge_pte_offset(struct mm_struct *mm, argument 85 pgd = pgd_offset(mm, addr); 104 static inline void purge_tlb_entries_huge(struct mm_struct *mm, unsigne argument 122 __set_huge_pte_at(struct mm_struct *mm, unsigned long addr, pte_t *ptep, pte_t entry) argument 142 set_huge_pte_at(struct mm_struct *mm, unsigned long addr, pte_t *ptep, pte_t entry, unsigned long sz) argument 149 huge_ptep_get_and_clear(struct mm_struct *mm, unsigned long addr, pte_t *ptep) argument 161 huge_ptep_set_wrprotect(struct mm_struct *mm, unsigned long addr, pte_t *ptep) argument 175 struct mm_struct *mm = vma->vm_mm; local [all...] |
/linux-master/drivers/gpu/drm/ |
H A D | drm_buddy.c | 14 static struct drm_buddy_block *drm_block_alloc(struct drm_buddy *mm, argument 35 static void drm_block_free(struct drm_buddy *mm, argument 41 static void list_insert_sorted(struct drm_buddy *mm, argument 47 head = &mm->free_list[drm_buddy_block_order(block)]; 68 static void mark_free(struct drm_buddy *mm, argument 74 list_insert_sorted(mm, block); 88 * @mm: DRM buddy manager to initialize 97 int drm_buddy_init(struct drm_buddy *mm, u64 size, u64 chunk_size) argument 113 mm->size = size; 114 mm 187 drm_buddy_fini(struct drm_buddy *mm) argument 203 split_block(struct drm_buddy *mm, struct drm_buddy_block *block) argument 263 __drm_buddy_free(struct drm_buddy *mm, struct drm_buddy_block *block) argument 293 drm_buddy_free_block(struct drm_buddy *mm, struct drm_buddy_block *block) argument 308 drm_buddy_free_list(struct drm_buddy *mm, struct list_head *objects) argument 331 alloc_range_bias(struct drm_buddy *mm, u64 start, u64 end, unsigned int order) argument 418 get_maxblock(struct drm_buddy *mm, unsigned int order) argument 444 alloc_from_freelist(struct drm_buddy *mm, unsigned int order, unsigned long flags) argument 490 __alloc_range(struct drm_buddy *mm, struct list_head *dfs, u64 start, u64 size, struct list_head *blocks, u64 *total_allocated_on_err) argument 583 __drm_buddy_alloc_range(struct drm_buddy *mm, u64 start, u64 size, u64 *total_allocated_on_err, struct list_head *blocks) argument 599 __alloc_contig_try_harder(struct drm_buddy *mm, u64 size, u64 min_block_size, struct list_head *blocks) argument 671 drm_buddy_block_trim(struct drm_buddy *mm, u64 new_size, struct list_head *blocks) argument 742 drm_buddy_alloc_blocks(struct drm_buddy *mm, u64 start, u64 end, u64 size, u64 min_block_size, struct list_head *blocks, unsigned long flags) argument 881 drm_buddy_block_print(struct drm_buddy *mm, struct drm_buddy_block *block, struct drm_printer *p) argument 898 drm_buddy_print(struct drm_buddy *mm, struct drm_printer *p) argument [all...] |
/linux-master/arch/powerpc/mm/book3s64/ |
H A D | slice.c | 15 #include <linux/mm.h> 21 #include <linux/sched/mm.h> 86 static int slice_area_is_free(struct mm_struct *mm, unsigned long addr, argument 91 if ((mm_ctx_slb_addr_limit(&mm->context) - len) < addr) 93 vma = find_vma(mm, addr); 97 static int slice_low_has_vma(struct mm_struct *mm, unsigned long slice) argument 99 return !slice_area_is_free(mm, slice << SLICE_LOW_SHIFT, 103 static int slice_high_has_vma(struct mm_struct *mm, unsigned long slice) argument 114 return !slice_area_is_free(mm, start, end - start); 117 static void slice_mask_for_free(struct mm_struct *mm, struc argument 138 slice_check_range_fits(struct mm_struct *mm, const struct slice_mask *available, unsigned long start, unsigned long len) argument 173 struct mm_struct *mm = parm; local 187 slice_convert(struct mm_struct *mm, const struct slice_mask *mask, int psize) argument 278 slice_find_area_bottomup(struct mm_struct *mm, unsigned long addr, unsigned long len, const struct slice_mask *available, int psize, unsigned long high_limit) argument 322 slice_find_area_topdown(struct mm_struct *mm, unsigned long addr, unsigned long len, const struct slice_mask *available, int psize, unsigned long high_limit) argument 380 slice_find_area(struct mm_struct *mm, unsigned long len, const struct slice_mask *mask, int psize, int topdown, unsigned long high_limit) argument 436 struct mm_struct *mm = current->mm; local 664 get_slice_psize(struct mm_struct *mm, unsigned long addr) argument 683 slice_init_new_context_exec(struct mm_struct *mm) argument 719 struct mm_struct *mm = current->mm; local 729 slice_set_range_psize(struct mm_struct *mm, unsigned long start, unsigned long len, unsigned int psize) argument 760 slice_is_hugepage_only_range(struct mm_struct *mm, unsigned long addr, unsigned long len) argument [all...] |
/linux-master/drivers/gpu/drm/tests/ |
H A D | drm_buddy_test.c | 29 struct drm_buddy mm; local 39 KUNIT_ASSERT_FALSE_MSG(test, drm_buddy_init(&mm, mm_size, ps), 63 drm_buddy_alloc_blocks(&mm, bias_start, 72 drm_buddy_alloc_blocks(&mm, bias_start, 81 drm_buddy_alloc_blocks(&mm, bias_start + ps, 90 drm_buddy_alloc_blocks(&mm, bias_start + ps, 100 drm_buddy_alloc_blocks(&mm, bias_start, 106 drm_buddy_free_list(&mm, &tmp); 110 drm_buddy_alloc_blocks(&mm, bias_start, 116 drm_buddy_free_list(&mm, 232 struct drm_buddy mm; local 320 struct drm_buddy mm; local 400 struct drm_buddy mm; local 495 struct drm_buddy mm; local 541 struct drm_buddy mm; local [all...] |
H A D | drm_mm_test.c | 38 static bool assert_no_holes(struct kunit *test, const struct drm_mm *mm) argument 45 drm_mm_for_each_hole(hole, mm, hole_start, hole_end) 53 drm_mm_for_each_node(hole, mm) { 63 static bool assert_one_hole(struct kunit *test, const struct drm_mm *mm, u64 start, u64 end) argument 74 drm_mm_for_each_hole(hole, mm, hole_start, hole_end) { 78 "empty mm has incorrect hole, found (%llx, %llx), expect (%llx, %llx)\n", 103 static bool assert_node(struct kunit *test, struct drm_mm_node *node, struct drm_mm *mm, argument 108 if (!drm_mm_node_allocated(node) || node->mm != mm) { 138 struct drm_mm mm; local 192 struct drm_mm mm; local 217 expect_insert(struct kunit *test, struct drm_mm *mm, struct drm_mm_node *node, u64 size, u64 alignment, unsigned long color, const struct insert_mode *mode) argument 243 struct drm_mm mm; local 290 struct drm_mm mm; local [all...] |
/linux-master/drivers/gpu/drm/nouveau/nvkm/core/ |
H A D | mm.c | 24 #include <core/mm.h> 26 #define node(root, dir) ((root)->nl_entry.dir == &mm->nodes) ? NULL : \ 30 nvkm_mm_dump(struct nvkm_mm *mm, const char *header) argument 36 list_for_each_entry(node, &mm->nodes, nl_entry) { 41 list_for_each_entry(node, &mm->free, fl_entry) { 48 nvkm_mm_free(struct nvkm_mm *mm, struct nvkm_mm_node **pthis) argument 72 list_for_each_entry(prev, &mm->free, fl_entry) { 86 region_head(struct nvkm_mm *mm, struct nvkm_mm_node *a, u32 size) argument 111 nvkm_mm_head(struct nvkm_mm *mm, u8 heap, u8 type, u32 size_max, u32 size_min, argument 121 list_for_each_entry(this, &mm 161 region_tail(struct nvkm_mm *mm, struct nvkm_mm_node *a, u32 size) argument 186 nvkm_mm_tail(struct nvkm_mm *mm, u8 heap, u8 type, u32 size_max, u32 size_min, u32 align, struct nvkm_mm_node **pnode) argument 240 nvkm_mm_init(struct nvkm_mm *mm, u8 heap, u32 offset, u32 length, u32 block) argument 283 nvkm_mm_fini(struct nvkm_mm *mm) argument [all...] |
/linux-master/arch/alpha/include/asm/ |
H A D | tlb.h | 7 #define __pte_free_tlb(tlb, pte, address) pte_free((tlb)->mm, pte) 8 #define __pmd_free_tlb(tlb, pmd, address) pmd_free((tlb)->mm, pmd)
|
/linux-master/arch/um/include/asm/ |
H A D | tlbflush.h | 9 #include <linux/mm.h> 14 * - flush_tlb() flushes the current mm struct TLBs 16 * - flush_tlb_mm(mm) flushes the specified mm context TLB's 23 extern void flush_tlb_mm(struct mm_struct *mm);
|
/linux-master/arch/parisc/include/asm/ |
H A D | tlb.h | 8 #define __pmd_free_tlb(tlb, pmd, addr) pmd_free((tlb)->mm, pmd) 10 #define __pte_free_tlb(tlb, pte, addr) pte_free((tlb)->mm, pte)
|
/linux-master/arch/mips/mm/ |
H A D | hugetlbpage.c | 15 #include <linux/mm.h> 24 pte_t *huge_pte_alloc(struct mm_struct *mm, struct vm_area_struct *vma, argument 32 pgd = pgd_offset(mm, addr); 33 p4d = p4d_alloc(mm, pgd, addr); 34 pud = pud_alloc(mm, p4d, addr); 36 pte = (pte_t *)pmd_alloc(mm, pud, addr); 41 pte_t *huge_pte_offset(struct mm_struct *mm, unsigned long addr, argument 49 pgd = pgd_offset(mm, addr);
|
/linux-master/drivers/infiniband/hw/hfi1/ |
H A D | user_pages.c | 6 #include <linux/mm.h> 29 bool hfi1_can_pin_pages(struct hfi1_devdata *dd, struct mm_struct *mm, argument 47 if (atomic64_read(&mm->pinned_vm) + npages > ulimit_pages) 80 int hfi1_acquire_user_pages(struct mm_struct *mm, unsigned long vaddr, size_t npages, argument 90 atomic64_add(ret, &mm->pinned_vm); 95 void hfi1_release_user_pages(struct mm_struct *mm, struct page **p, argument 100 if (mm) { /* during close after signal, mm can be NULL */ 101 atomic64_sub(npages, &mm->pinned_vm);
|
/linux-master/arch/x86/include/asm/ |
H A D | pgalloc.h | 6 #include <linux/mm.h> /* for struct page */ 13 static inline int __paravirt_pgd_alloc(struct mm_struct *mm) { return 0; } argument 18 #define paravirt_pgd_alloc(mm) __paravirt_pgd_alloc(mm) 19 static inline void paravirt_pgd_free(struct mm_struct *mm, pgd_t *pgd) {} argument 20 static inline void paravirt_alloc_pte(struct mm_struct *mm, unsigned long pfn) {} argument 21 static inline void paravirt_alloc_pmd(struct mm_struct *mm, unsigned long pfn) {} argument 24 static inline void paravirt_alloc_pud(struct mm_struct *mm, unsigned long pfn) {} argument 25 static inline void paravirt_alloc_p4d(struct mm_struct *mm, unsigned long pfn) {} argument 52 extern void pgd_free(struct mm_struct *mm, pgd_ 64 pmd_populate_kernel(struct mm_struct *mm, pmd_t *pmd, pte_t *pte) argument 71 pmd_populate_kernel_safe(struct mm_struct *mm, pmd_t *pmd, pte_t *pte) argument 78 pmd_populate(struct mm_struct *mm, pmd_t *pmd, struct page *pte) argument 99 pud_populate(struct mm_struct *mm, pud_t *pud, pmd_t *pmd) argument 105 pud_populate_safe(struct mm_struct *mm, pud_t *pud, pmd_t *pmd) argument 113 p4d_populate(struct mm_struct *mm, p4d_t *p4d, pud_t *pud) argument 119 p4d_populate_safe(struct mm_struct *mm, p4d_t *p4d, pud_t *pud) argument 134 pgd_populate(struct mm_struct *mm, pgd_t *pgd, p4d_t *p4d) argument 142 pgd_populate_safe(struct mm_struct *mm, pgd_t *pgd, p4d_t *p4d) argument 150 p4d_alloc_one(struct mm_struct *mm, unsigned long addr) argument 159 p4d_free(struct mm_struct *mm, p4d_t *p4d) argument [all...] |
/linux-master/arch/xtensa/include/asm/ |
H A D | mmu_context.h | 70 static inline void get_new_mmu_context(struct mm_struct *mm, unsigned int cpu) argument 82 mm->context.asid[cpu] = asid; 83 mm->context.cpu = cpu; 86 static inline void get_mmu_context(struct mm_struct *mm, unsigned int cpu) argument 92 if (mm) { 93 unsigned long asid = mm->context.asid[cpu]; 97 get_new_mmu_context(mm, cpu); 101 static inline void activate_context(struct mm_struct *mm, unsigned int cpu) argument 103 get_mmu_context(mm, cpu); 104 set_rasid_register(ASID_INSERT(mm 115 init_new_context(struct task_struct *tsk, struct mm_struct *mm) argument 145 destroy_context(struct mm_struct *mm) argument [all...] |