Lines Matching refs:mm
24 void __check_vmalloc_seq(struct mm_struct *mm);
27 static inline void check_vmalloc_seq(struct mm_struct *mm)
30 unlikely(atomic_read(&mm->context.vmalloc_seq) !=
32 __check_vmalloc_seq(mm);
38 void check_and_switch_context(struct mm_struct *mm, struct task_struct *tsk);
42 init_new_context(struct task_struct *tsk, struct mm_struct *mm)
44 atomic64_set(&mm->context.id, 0);
49 void a15_erratum_get_cpumask(int this_cpu, struct mm_struct *mm,
52 static inline void a15_erratum_get_cpumask(int this_cpu, struct mm_struct *mm,
62 static inline void check_and_switch_context(struct mm_struct *mm,
65 check_vmalloc_seq(mm);
71 * running with the old mm. Since we only support UP systems
72 * on non-ASID CPUs, the old mm will remain valid until the
75 mm->context.switch_pending = 1;
77 cpu_switch_mm(mm->pgd, mm);
85 struct mm_struct *mm = current->mm;
87 if (mm && mm->context.switch_pending) {
92 * switch to this mm was already done.
95 if (mm->context.switch_pending) {
96 mm->context.switch_pending = 0;
97 cpu_switch_mm(mm->pgd, mm);
111 * This is the actual mm switch as far as the scheduler
113 * calling the CPU specific function when the mm hasn't
142 static inline void enter_lazy_tlb(struct mm_struct *mm, struct task_struct *tsk)
144 if (mm != &init_mm)
145 check_vmalloc_seq(mm);