Lines Matching refs:mm
18 #include <linux/sched/mm.h>
19 #include <linux/mm.h>
145 /* All kernel threads share the same mm context. */
626 struct mm_struct *mm = x;
627 if (mm == current->active_mm && !asn_locked())
628 flush_tlb_current(mm);
630 flush_tlb_other(mm);
634 flush_tlb_mm(struct mm_struct *mm)
638 if (mm == current->active_mm) {
639 flush_tlb_current(mm);
640 if (atomic_read(&mm->mm_users) <= 1) {
645 if (mm->context[cpu])
646 mm->context[cpu] = 0;
653 smp_call_function(ipi_flush_tlb_mm, mm, 1);
661 struct mm_struct *mm;
669 struct mm_struct * mm = data->mm;
671 if (mm == current->active_mm && !asn_locked())
672 flush_tlb_current_page(mm, data->vma, data->addr);
674 flush_tlb_other(mm);
681 struct mm_struct *mm = vma->vm_mm;
685 if (mm == current->active_mm) {
686 flush_tlb_current_page(mm, vma, addr);
687 if (atomic_read(&mm->mm_users) <= 1) {
692 if (mm->context[cpu])
693 mm->context[cpu] = 0;
701 data.mm = mm;
721 struct mm_struct *mm = (struct mm_struct *) x;
722 if (mm == current->active_mm && !asn_locked())
723 __load_new_mm_context(mm);
725 flush_tlb_other(mm);
732 struct mm_struct *mm = vma->vm_mm;
739 if (mm == current->active_mm) {
740 __load_new_mm_context(mm);
741 if (atomic_read(&mm->mm_users) <= 1) {
746 if (mm->context[cpu])
747 mm->context[cpu] = 0;
754 smp_call_function(ipi_flush_icache_page, mm, 1);