Searched refs:mm (Results 76 - 100 of 2900) sorted by relevance

1234567891011>>

/linux-master/arch/x86/kernel/
H A Dldt.c19 #include <linux/mm.h>
42 void load_mm_ldt(struct mm_struct *mm) argument
47 ldt = READ_ONCE(mm->context.ldt);
50 * Any change to mm->context.ldt is followed by an IPI to all
51 * CPUs with the mm active. The LDT will not be freed until
93 * Load the LDT if either the old or new mm had an LDT.
95 * An mm will never go from having an LDT to not having an LDT. Two
138 struct mm_struct *mm = __mm; local
140 if (this_cpu_read(cpu_tlbstate.loaded_mm) != mm)
143 load_mm_ldt(mm);
189 do_sanity_check(struct mm_struct *mm, bool had_kernel_mapping, bool had_user_mapping) argument
234 map_ldt_struct_to_user(struct mm_struct *mm) argument
247 sanity_check_ldt_mapping(struct mm_struct *mm) argument
264 map_ldt_struct_to_user(struct mm_struct *mm) argument
272 sanity_check_ldt_mapping(struct mm_struct *mm) argument
288 map_ldt_struct(struct mm_struct *mm, struct ldt_struct *ldt, int slot) argument
349 unmap_ldt_struct(struct mm_struct *mm, struct ldt_struct *ldt) argument
383 map_ldt_struct(struct mm_struct *mm, struct ldt_struct *ldt, int slot) argument
388 unmap_ldt_struct(struct mm_struct *mm, struct ldt_struct *ldt) argument
393 free_ldt_pgtables(struct mm_struct *mm) argument
421 install_ldt(struct mm_struct *mm, struct ldt_struct *ldt) argument
429 on_each_cpu_mask(mm_cpumask(mm), flush_ldt, mm, true); local
451 ldt_dup_context(struct mm_struct *old_mm, struct mm_struct *mm) argument
491 destroy_context_ldt(struct mm_struct *mm) argument
497 ldt_arch_exit_mmap(struct mm_struct *mm) argument
504 struct mm_struct *mm = current->mm; local
580 struct mm_struct *mm = current->mm; local
[all...]
/linux-master/include/linux/sched/
H A Dmm.h19 * @mm: The &struct mm_struct to pin.
21 * Make sure that @mm will not get freed even after the owning task
26 * This is a preferred way to pin @mm for a longer/unbounded amount
31 * See also <Documentation/mm/active_mm.rst> for an in-depth explanation
34 static inline void mmgrab(struct mm_struct *mm) argument
36 atomic_inc(&mm->mm_count);
44 extern void __mmdrop(struct mm_struct *mm);
46 static inline void mmdrop(struct mm_struct *mm) argument
53 if (unlikely(atomic_dec_and_test(&mm->mm_count)))
54 __mmdrop(mm);
64 struct mm_struct *mm = container_of(rhp, struct mm_struct, delayed_drop); local
73 mmdrop_sched(struct mm_struct *mm) argument
80 mmdrop_sched(struct mm_struct *mm) argument
87 mmgrab_lazy_tlb(struct mm_struct *mm) argument
93 mmdrop_lazy_tlb(struct mm_struct *mm) argument
106 mmdrop_lazy_tlb_sched(struct mm_struct *mm) argument
130 mmget(struct mm_struct *mm) argument
135 mmget_not_zero(struct mm_struct *mm) argument
165 mm_update_next_owner(struct mm_struct *mm) argument
198 arch_pick_mmap_layout(struct mm_struct *mm, struct rlimit *rlim_stack) argument
526 membarrier_mm_sync_core_before_usermode(struct mm_struct *mm) argument
548 membarrier_exec_mmap(struct mm_struct *mm) argument
551 membarrier_mm_sync_core_before_usermode(struct mm_struct *mm) argument
[all...]
/linux-master/include/linux/
H A Dsync_core.h30 static inline void prepare_sync_core_cmd(struct mm_struct *mm) argument
H A Daio.h15 extern void exit_aio(struct mm_struct *mm);
18 static inline void exit_aio(struct mm_struct *mm) { } argument
/linux-master/arch/arm64/include/asm/
H A Dmmu_context.h55 void cpu_do_switch_mm(phys_addr_t pgd_phys, struct mm_struct *mm);
57 static inline void cpu_switch_mm(pgd_t *pgd, struct mm_struct *mm) argument
60 cpu_do_switch_mm(virt_to_phys(pgd),mm); local
88 * Remove the idmap from TTBR0_EL1 and install the pgd of the active mm.
95 * If current is a not a user task, the mm covers the TTBR1_EL1 page tables,
101 struct mm_struct *mm = current->active_mm; local
107 if (mm != &init_mm && !system_uses_ttbr0_pan())
108 cpu_switch_mm(mm->pgd, mm);
165 * freeing the ASID from the context of the dying mm (
174 init_new_context(struct task_struct *tsk, struct mm_struct *mm) argument
182 update_saved_ttbr0(struct task_struct *tsk, struct mm_struct *mm) argument
198 update_saved_ttbr0(struct task_struct *tsk, struct mm_struct *mm) argument
206 enter_lazy_tlb(struct mm_struct *mm, struct task_struct *tsk) argument
265 mm_untag_mask(struct mm_struct *mm) argument
[all...]
/linux-master/arch/powerpc/mm/
H A Dpgtable-frag.c10 #include <linux/mm.h>
33 static pte_t *get_pte_from_cache(struct mm_struct *mm) argument
40 spin_lock(&mm->page_table_lock);
41 ret = pte_frag_get(&mm->context);
49 pte_frag_set(&mm->context, pte_frag);
51 spin_unlock(&mm->page_table_lock);
55 static pte_t *__alloc_for_ptecache(struct mm_struct *mm, int kernel) argument
83 spin_lock(&mm->page_table_lock);
89 if (likely(!pte_frag_get(&mm->context))) {
91 pte_frag_set(&mm
98 pte_fragment_alloc(struct mm_struct *mm, int kernel) argument
137 pte_free_defer(struct mm_struct *mm, pgtable_t pgtable) argument
[all...]
/linux-master/arch/sparc/include/asm/
H A Dpgalloc_32.h26 #define pgd_free(mm, pgd) free_pgd_fast(pgd)
27 #define pgd_alloc(mm) get_pgd_fast()
38 static inline pmd_t *pmd_alloc_one(struct mm_struct *mm, argument
50 #define pmd_free(mm, pmd) free_pmd_fast(pmd)
51 #define __pmd_free_tlb(tlb, pmd, addr) pmd_free((tlb)->mm, pmd)
53 #define pmd_populate(mm, pmd, pte) pmd_set(pmd, pte)
58 pgtable_t pte_alloc_one(struct mm_struct *mm);
60 static inline pte_t *pte_alloc_one_kernel(struct mm_struct *mm) argument
72 #define pte_free_kernel(mm, pte) free_pte_fast(pte)
74 void pte_free(struct mm_struct * mm, pgtable_
[all...]
H A Dtlbflush_32.h9 #define flush_tlb_mm(mm) \
10 sparc32_cachetlb_ops->tlb_mm(mm)
/linux-master/arch/powerpc/mm/book3s32/
H A Dmmu_context.c9 * Derived from arch/ppc/mm/init.c:
16 * Derived from "arch/i386/mm/init.c"
20 #include <linux/mm.h>
69 int init_new_context(struct task_struct *t, struct mm_struct *mm) argument
71 mm->context.id = __init_new_context();
72 mm->context.sr0 = CTX_TO_VSID(mm->context.id, 0);
75 mm->context.sr0 |= SR_NX;
77 mm->context.sr0 |= SR_KS;
94 void destroy_context(struct mm_struct *mm) argument
[all...]
/linux-master/arch/mips/include/asm/
H A Dpgalloc.h13 #include <linux/mm.h>
21 static inline void pmd_populate_kernel(struct mm_struct *mm, pmd_t *pmd, argument
27 static inline void pmd_populate(struct mm_struct *mm, pmd_t *pmd, argument
40 static inline void pud_populate(struct mm_struct *mm, pud_t *pud, pmd_t *pmd) argument
50 extern pgd_t *pgd_alloc(struct mm_struct *mm);
52 static inline void pgd_free(struct mm_struct *mm, pgd_t *pgd) argument
65 static inline pmd_t *pmd_alloc_one(struct mm_struct *mm, unsigned long address) argument
84 #define __pmd_free_tlb(tlb, x, addr) pmd_free((tlb)->mm, x)
90 static inline pud_t *pud_alloc_one(struct mm_struct *mm, unsigned long address) argument
105 static inline void p4d_populate(struct mm_struct *mm, p4d_ argument
[all...]
/linux-master/drivers/gpu/drm/amd/amdkfd/
H A Dkfd_mqd_manager.c91 void free_mqd_hiq_sdma(struct mqd_manager *mm, void *mqd, argument
98 void mqd_symmetrically_map_cu_mask(struct mqd_manager *mm, argument
102 struct amdgpu_cu_info *cu_info = &mm->dev->adev->gfx.cu_info;
103 struct amdgpu_gfx_config *gfx_info = &mm->dev->adev->gfx.config;
105 bool wgp_mode_req = KFD_GC_VERSION(mm->dev) >= IP_VERSION(10, 0, 0);
109 int inc = cu_inc * NUM_XCC(mm->dev->xcc_mask);
110 int xcc_inst = inst + ffs(mm->dev->xcc_mask) - 1;
112 cu_active_per_node = cu_info->number / mm->dev->kfd->num_nodes;
131 cu_bitmap_sh_mul = (KFD_GC_VERSION(mm->dev) >= IP_VERSION(11, 0, 0) &&
132 KFD_GC_VERSION(mm
206 kfd_hiq_load_mqd_kiq(struct mqd_manager *mm, void *mqd, uint32_t pipe_id, uint32_t queue_id, struct queue_properties *p, struct mm_struct *mms) argument
214 kfd_destroy_mqd_cp(struct mqd_manager *mm, void *mqd, enum kfd_preempt_type type, unsigned int timeout, uint32_t pipe_id, uint32_t queue_id) argument
222 kfd_free_mqd_cp(struct mqd_manager *mm, void *mqd, struct kfd_mem_obj *mqd_mem_obj) argument
233 kfd_is_occupied_cp(struct mqd_manager *mm, void *mqd, uint64_t queue_address, uint32_t pipe_id, uint32_t queue_id) argument
241 kfd_load_mqd_sdma(struct mqd_manager *mm, void *mqd, uint32_t pipe_id, uint32_t queue_id, struct queue_properties *p, struct mm_struct *mms) argument
254 kfd_destroy_mqd_sdma(struct mqd_manager *mm, void *mqd, enum kfd_preempt_type type, unsigned int timeout, uint32_t pipe_id, uint32_t queue_id) argument
262 kfd_is_occupied_sdma(struct mqd_manager *mm, void *mqd, uint64_t queue_address, uint32_t pipe_id, uint32_t queue_id) argument
288 kfd_mqd_stride(struct mqd_manager *mm, struct queue_properties *q) argument
[all...]
H A Dkfd_mqd_manager.h74 void (*init_mqd)(struct mqd_manager *mm, void **mqd,
78 int (*load_mqd)(struct mqd_manager *mm, void *mqd,
83 void (*update_mqd)(struct mqd_manager *mm, void *mqd,
87 int (*destroy_mqd)(struct mqd_manager *mm, void *mqd,
92 void (*free_mqd)(struct mqd_manager *mm, void *mqd,
95 bool (*is_occupied)(struct mqd_manager *mm, void *mqd,
99 int (*get_wave_state)(struct mqd_manager *mm, void *mqd,
105 void (*get_checkpoint_info)(struct mqd_manager *mm, void *mqd, uint32_t *ctl_stack_size);
107 void (*checkpoint_mqd)(struct mqd_manager *mm,
112 void (*restore_mqd)(struct mqd_manager *mm, voi
[all...]
/linux-master/arch/parisc/mm/
H A Dhugetlbpage.c11 #include <linux/mm.h>
12 #include <linux/sched/mm.h>
47 pte_t *huge_pte_alloc(struct mm_struct *mm, struct vm_area_struct *vma, argument
63 pgd = pgd_offset(mm, addr);
65 pud = pud_alloc(mm, p4d, addr);
67 pmd = pmd_alloc(mm, pud, addr);
69 pte = pte_alloc_huge(mm, pmd, addr);
74 pte_t *huge_pte_offset(struct mm_struct *mm, argument
85 pgd = pgd_offset(mm, addr);
104 static inline void purge_tlb_entries_huge(struct mm_struct *mm, unsigne argument
122 __set_huge_pte_at(struct mm_struct *mm, unsigned long addr, pte_t *ptep, pte_t entry) argument
142 set_huge_pte_at(struct mm_struct *mm, unsigned long addr, pte_t *ptep, pte_t entry, unsigned long sz) argument
149 huge_ptep_get_and_clear(struct mm_struct *mm, unsigned long addr, pte_t *ptep) argument
161 huge_ptep_set_wrprotect(struct mm_struct *mm, unsigned long addr, pte_t *ptep) argument
175 struct mm_struct *mm = vma->vm_mm; local
[all...]
/linux-master/drivers/gpu/drm/
H A Ddrm_buddy.c14 static struct drm_buddy_block *drm_block_alloc(struct drm_buddy *mm, argument
35 static void drm_block_free(struct drm_buddy *mm, argument
41 static void list_insert_sorted(struct drm_buddy *mm, argument
47 head = &mm->free_list[drm_buddy_block_order(block)];
68 static void mark_free(struct drm_buddy *mm, argument
74 list_insert_sorted(mm, block);
88 * @mm: DRM buddy manager to initialize
97 int drm_buddy_init(struct drm_buddy *mm, u64 size, u64 chunk_size) argument
113 mm->size = size;
114 mm
187 drm_buddy_fini(struct drm_buddy *mm) argument
203 split_block(struct drm_buddy *mm, struct drm_buddy_block *block) argument
263 __drm_buddy_free(struct drm_buddy *mm, struct drm_buddy_block *block) argument
293 drm_buddy_free_block(struct drm_buddy *mm, struct drm_buddy_block *block) argument
308 drm_buddy_free_list(struct drm_buddy *mm, struct list_head *objects) argument
331 alloc_range_bias(struct drm_buddy *mm, u64 start, u64 end, unsigned int order) argument
418 get_maxblock(struct drm_buddy *mm, unsigned int order) argument
444 alloc_from_freelist(struct drm_buddy *mm, unsigned int order, unsigned long flags) argument
490 __alloc_range(struct drm_buddy *mm, struct list_head *dfs, u64 start, u64 size, struct list_head *blocks, u64 *total_allocated_on_err) argument
583 __drm_buddy_alloc_range(struct drm_buddy *mm, u64 start, u64 size, u64 *total_allocated_on_err, struct list_head *blocks) argument
599 __alloc_contig_try_harder(struct drm_buddy *mm, u64 size, u64 min_block_size, struct list_head *blocks) argument
671 drm_buddy_block_trim(struct drm_buddy *mm, u64 new_size, struct list_head *blocks) argument
742 drm_buddy_alloc_blocks(struct drm_buddy *mm, u64 start, u64 end, u64 size, u64 min_block_size, struct list_head *blocks, unsigned long flags) argument
881 drm_buddy_block_print(struct drm_buddy *mm, struct drm_buddy_block *block, struct drm_printer *p) argument
898 drm_buddy_print(struct drm_buddy *mm, struct drm_printer *p) argument
[all...]
/linux-master/arch/powerpc/mm/book3s64/
H A Dslice.c15 #include <linux/mm.h>
21 #include <linux/sched/mm.h>
86 static int slice_area_is_free(struct mm_struct *mm, unsigned long addr, argument
91 if ((mm_ctx_slb_addr_limit(&mm->context) - len) < addr)
93 vma = find_vma(mm, addr);
97 static int slice_low_has_vma(struct mm_struct *mm, unsigned long slice) argument
99 return !slice_area_is_free(mm, slice << SLICE_LOW_SHIFT,
103 static int slice_high_has_vma(struct mm_struct *mm, unsigned long slice) argument
114 return !slice_area_is_free(mm, start, end - start);
117 static void slice_mask_for_free(struct mm_struct *mm, struc argument
138 slice_check_range_fits(struct mm_struct *mm, const struct slice_mask *available, unsigned long start, unsigned long len) argument
173 struct mm_struct *mm = parm; local
187 slice_convert(struct mm_struct *mm, const struct slice_mask *mask, int psize) argument
278 slice_find_area_bottomup(struct mm_struct *mm, unsigned long addr, unsigned long len, const struct slice_mask *available, int psize, unsigned long high_limit) argument
322 slice_find_area_topdown(struct mm_struct *mm, unsigned long addr, unsigned long len, const struct slice_mask *available, int psize, unsigned long high_limit) argument
380 slice_find_area(struct mm_struct *mm, unsigned long len, const struct slice_mask *mask, int psize, int topdown, unsigned long high_limit) argument
436 struct mm_struct *mm = current->mm; local
664 get_slice_psize(struct mm_struct *mm, unsigned long addr) argument
683 slice_init_new_context_exec(struct mm_struct *mm) argument
719 struct mm_struct *mm = current->mm; local
729 slice_set_range_psize(struct mm_struct *mm, unsigned long start, unsigned long len, unsigned int psize) argument
760 slice_is_hugepage_only_range(struct mm_struct *mm, unsigned long addr, unsigned long len) argument
[all...]
/linux-master/drivers/gpu/drm/tests/
H A Ddrm_buddy_test.c29 struct drm_buddy mm; local
39 KUNIT_ASSERT_FALSE_MSG(test, drm_buddy_init(&mm, mm_size, ps),
63 drm_buddy_alloc_blocks(&mm, bias_start,
72 drm_buddy_alloc_blocks(&mm, bias_start,
81 drm_buddy_alloc_blocks(&mm, bias_start + ps,
90 drm_buddy_alloc_blocks(&mm, bias_start + ps,
100 drm_buddy_alloc_blocks(&mm, bias_start,
106 drm_buddy_free_list(&mm, &tmp);
110 drm_buddy_alloc_blocks(&mm, bias_start,
116 drm_buddy_free_list(&mm,
232 struct drm_buddy mm; local
320 struct drm_buddy mm; local
400 struct drm_buddy mm; local
495 struct drm_buddy mm; local
541 struct drm_buddy mm; local
[all...]
H A Ddrm_mm_test.c38 static bool assert_no_holes(struct kunit *test, const struct drm_mm *mm) argument
45 drm_mm_for_each_hole(hole, mm, hole_start, hole_end)
53 drm_mm_for_each_node(hole, mm) {
63 static bool assert_one_hole(struct kunit *test, const struct drm_mm *mm, u64 start, u64 end) argument
74 drm_mm_for_each_hole(hole, mm, hole_start, hole_end) {
78 "empty mm has incorrect hole, found (%llx, %llx), expect (%llx, %llx)\n",
103 static bool assert_node(struct kunit *test, struct drm_mm_node *node, struct drm_mm *mm, argument
108 if (!drm_mm_node_allocated(node) || node->mm != mm) {
138 struct drm_mm mm; local
192 struct drm_mm mm; local
217 expect_insert(struct kunit *test, struct drm_mm *mm, struct drm_mm_node *node, u64 size, u64 alignment, unsigned long color, const struct insert_mode *mode) argument
243 struct drm_mm mm; local
290 struct drm_mm mm; local
[all...]
/linux-master/drivers/gpu/drm/nouveau/nvkm/core/
H A Dmm.c24 #include <core/mm.h>
26 #define node(root, dir) ((root)->nl_entry.dir == &mm->nodes) ? NULL : \
30 nvkm_mm_dump(struct nvkm_mm *mm, const char *header) argument
36 list_for_each_entry(node, &mm->nodes, nl_entry) {
41 list_for_each_entry(node, &mm->free, fl_entry) {
48 nvkm_mm_free(struct nvkm_mm *mm, struct nvkm_mm_node **pthis) argument
72 list_for_each_entry(prev, &mm->free, fl_entry) {
86 region_head(struct nvkm_mm *mm, struct nvkm_mm_node *a, u32 size) argument
111 nvkm_mm_head(struct nvkm_mm *mm, u8 heap, u8 type, u32 size_max, u32 size_min, argument
121 list_for_each_entry(this, &mm
161 region_tail(struct nvkm_mm *mm, struct nvkm_mm_node *a, u32 size) argument
186 nvkm_mm_tail(struct nvkm_mm *mm, u8 heap, u8 type, u32 size_max, u32 size_min, u32 align, struct nvkm_mm_node **pnode) argument
240 nvkm_mm_init(struct nvkm_mm *mm, u8 heap, u32 offset, u32 length, u32 block) argument
283 nvkm_mm_fini(struct nvkm_mm *mm) argument
[all...]
/linux-master/arch/alpha/include/asm/
H A Dtlb.h7 #define __pte_free_tlb(tlb, pte, address) pte_free((tlb)->mm, pte)
8 #define __pmd_free_tlb(tlb, pmd, address) pmd_free((tlb)->mm, pmd)
/linux-master/arch/um/include/asm/
H A Dtlbflush.h9 #include <linux/mm.h>
14 * - flush_tlb() flushes the current mm struct TLBs
16 * - flush_tlb_mm(mm) flushes the specified mm context TLB's
23 extern void flush_tlb_mm(struct mm_struct *mm);
/linux-master/arch/parisc/include/asm/
H A Dtlb.h8 #define __pmd_free_tlb(tlb, pmd, addr) pmd_free((tlb)->mm, pmd)
10 #define __pte_free_tlb(tlb, pte, addr) pte_free((tlb)->mm, pte)
/linux-master/arch/mips/mm/
H A Dhugetlbpage.c15 #include <linux/mm.h>
24 pte_t *huge_pte_alloc(struct mm_struct *mm, struct vm_area_struct *vma, argument
32 pgd = pgd_offset(mm, addr);
33 p4d = p4d_alloc(mm, pgd, addr);
34 pud = pud_alloc(mm, p4d, addr);
36 pte = (pte_t *)pmd_alloc(mm, pud, addr);
41 pte_t *huge_pte_offset(struct mm_struct *mm, unsigned long addr, argument
49 pgd = pgd_offset(mm, addr);
/linux-master/drivers/infiniband/hw/hfi1/
H A Duser_pages.c6 #include <linux/mm.h>
29 bool hfi1_can_pin_pages(struct hfi1_devdata *dd, struct mm_struct *mm, argument
47 if (atomic64_read(&mm->pinned_vm) + npages > ulimit_pages)
80 int hfi1_acquire_user_pages(struct mm_struct *mm, unsigned long vaddr, size_t npages, argument
90 atomic64_add(ret, &mm->pinned_vm);
95 void hfi1_release_user_pages(struct mm_struct *mm, struct page **p, argument
100 if (mm) { /* during close after signal, mm can be NULL */
101 atomic64_sub(npages, &mm->pinned_vm);
/linux-master/arch/x86/include/asm/
H A Dpgalloc.h6 #include <linux/mm.h> /* for struct page */
13 static inline int __paravirt_pgd_alloc(struct mm_struct *mm) { return 0; } argument
18 #define paravirt_pgd_alloc(mm) __paravirt_pgd_alloc(mm)
19 static inline void paravirt_pgd_free(struct mm_struct *mm, pgd_t *pgd) {} argument
20 static inline void paravirt_alloc_pte(struct mm_struct *mm, unsigned long pfn) {} argument
21 static inline void paravirt_alloc_pmd(struct mm_struct *mm, unsigned long pfn) {} argument
24 static inline void paravirt_alloc_pud(struct mm_struct *mm, unsigned long pfn) {} argument
25 static inline void paravirt_alloc_p4d(struct mm_struct *mm, unsigned long pfn) {} argument
52 extern void pgd_free(struct mm_struct *mm, pgd_
64 pmd_populate_kernel(struct mm_struct *mm, pmd_t *pmd, pte_t *pte) argument
71 pmd_populate_kernel_safe(struct mm_struct *mm, pmd_t *pmd, pte_t *pte) argument
78 pmd_populate(struct mm_struct *mm, pmd_t *pmd, struct page *pte) argument
99 pud_populate(struct mm_struct *mm, pud_t *pud, pmd_t *pmd) argument
105 pud_populate_safe(struct mm_struct *mm, pud_t *pud, pmd_t *pmd) argument
113 p4d_populate(struct mm_struct *mm, p4d_t *p4d, pud_t *pud) argument
119 p4d_populate_safe(struct mm_struct *mm, p4d_t *p4d, pud_t *pud) argument
134 pgd_populate(struct mm_struct *mm, pgd_t *pgd, p4d_t *p4d) argument
142 pgd_populate_safe(struct mm_struct *mm, pgd_t *pgd, p4d_t *p4d) argument
150 p4d_alloc_one(struct mm_struct *mm, unsigned long addr) argument
159 p4d_free(struct mm_struct *mm, p4d_t *p4d) argument
[all...]
/linux-master/arch/xtensa/include/asm/
H A Dmmu_context.h70 static inline void get_new_mmu_context(struct mm_struct *mm, unsigned int cpu) argument
82 mm->context.asid[cpu] = asid;
83 mm->context.cpu = cpu;
86 static inline void get_mmu_context(struct mm_struct *mm, unsigned int cpu) argument
92 if (mm) {
93 unsigned long asid = mm->context.asid[cpu];
97 get_new_mmu_context(mm, cpu);
101 static inline void activate_context(struct mm_struct *mm, unsigned int cpu) argument
103 get_mmu_context(mm, cpu);
104 set_rasid_register(ASID_INSERT(mm
115 init_new_context(struct task_struct *tsk, struct mm_struct *mm) argument
145 destroy_context(struct mm_struct *mm) argument
[all...]

Completed in 218 milliseconds

1234567891011>>