Lines Matching defs:mm

119 void pte_free_defer(struct mm_struct *mm, pgtable_t pgtable);
144 * a shortcut to get a pgd_t in a given mm
147 #define pgd_offset(mm, address) pgd_offset_pgd((mm)->pgd, (address))
163 static inline pmd_t *pmd_off(struct mm_struct *mm, unsigned long va)
165 return pmd_offset(pud_offset(p4d_offset(pgd_offset(mm, va), va), va), va);
248 * @mm: Address space to map the pages into.
264 static inline void set_ptes(struct mm_struct *mm, unsigned long addr,
267 page_table_check_ptes_set(mm, ptep, pte, nr);
280 #define set_pte_at(mm, addr, ptep, pte) set_ptes(mm, addr, ptep, pte, 1)
451 static inline pte_t ptep_get_and_clear(struct mm_struct *mm,
456 pte_clear(mm, address, ptep);
457 page_table_check_pte_clear(mm, pte);
466 * @mm: Address space the pages are mapped into.
506 static inline void ptep_clear(struct mm_struct *mm, unsigned long addr,
509 ptep_get_and_clear(mm, addr, ptep);
597 static inline pmd_t pmdp_huge_get_and_clear(struct mm_struct *mm,
604 page_table_check_pmd_clear(mm, pmd);
610 static inline pud_t pudp_huge_get_and_clear(struct mm_struct *mm,
617 page_table_check_pud_clear(mm, pud);
645 static inline pte_t ptep_get_and_clear_full(struct mm_struct *mm,
649 return ptep_get_and_clear(mm, address, ptep);
657 * @mm: Address space the pages are mapped into.
661 * @full: Whether we are clearing a full mm.
673 static inline pte_t get_and_clear_full_ptes(struct mm_struct *mm,
678 pte = ptep_get_and_clear_full(mm, addr, ptep, full);
682 tmp_pte = ptep_get_and_clear_full(mm, addr, ptep, full);
696 * @mm: Address space the pages are mapped into.
700 * @full: Whether we are clearing a full mm.
711 static inline void clear_full_ptes(struct mm_struct *mm, unsigned long addr,
715 ptep_get_and_clear_full(mm, addr, ptep, full);
746 static inline void pte_clear_not_present_full(struct mm_struct *mm,
751 pte_clear(mm, address, ptep);
759 * @mm: Address space the ptes represent.
763 * @full: Whether we are clearing a full mm.
771 static inline void clear_not_present_full_ptes(struct mm_struct *mm,
775 pte_clear_not_present_full(mm, addr, ptep, full);
815 static inline void ptep_set_wrprotect(struct mm_struct *mm, unsigned long address, pte_t *ptep)
818 set_pte_at(mm, address, ptep, pte_wrprotect(old_pte));
826 * @mm: Address space the pages are mapped into.
840 static inline void wrprotect_ptes(struct mm_struct *mm, unsigned long addr,
844 ptep_set_wrprotect(mm, addr, ptep);
871 static inline void pmdp_set_wrprotect(struct mm_struct *mm,
875 set_pmd_at(mm, address, pmdp, pmd_wrprotect(old_pmd));
878 static inline void pmdp_set_wrprotect(struct mm_struct *mm,
888 static inline void pudp_set_wrprotect(struct mm_struct *mm,
893 set_pud_at(mm, address, pudp, pud_wrprotect(old_pud));
896 static inline void pudp_set_wrprotect(struct mm_struct *mm,
922 extern void pgtable_trans_huge_deposit(struct mm_struct *mm, pmd_t *pmdp,
927 extern pgtable_t pgtable_trans_huge_withdraw(struct mm_struct *mm, pmd_t *pmdp);
1095 static inline void arch_do_swap_page(struct mm_struct *mm,
1113 static inline int arch_unmap_one(struct mm_struct *mm,
1151 #define pgd_offset_gate(mm, addr) pgd_offset(mm, addr)
1159 # define pte_accessible(mm, pte) ((void)(pte), 1)
1201 * Do the tests inline, but report and clear the bad entry in mm/memory.c.
1820 * On some architectures it depends on the mm if the p4d/pud or pmd
1824 #define mm_p4d_folded(mm) __is_defined(__PAGETABLE_P4D_FOLDED)
1828 #define mm_pud_folded(mm) __is_defined(__PAGETABLE_PUD_FOLDED)
1832 #define mm_pmd_folded(mm) __is_defined(__PAGETABLE_PMD_FOLDED)