Searched refs:pte_entry (Results 1 - 9 of 9) sorted by relevance
/linux-master/arch/openrisc/kernel/ |
H A D | dma.c | 46 .pte_entry = page_set_nocache, 65 .pte_entry = page_clear_nocache,
|
/linux-master/include/linux/ |
H A D | pagewalk.h | 28 * @pte_entry: if set, called for each PTE (lowest-level) entry, 67 int (*pte_entry)(pte_t *pte, unsigned long addr, member in struct:mm_walk_ops
|
/linux-master/mm/ |
H A D | pagewalk.c | 30 err = ops->pte_entry(pte, addr, addr + PAGE_SIZE, walk); 85 if (!ops->pte_entry) 96 err = ops->pte_entry(pte, addr, addr + page_size, walk); 156 !(ops->pte_entry)) 210 !(ops->pmd_entry || ops->pte_entry)) 255 else if (ops->pud_entry || ops->pmd_entry || ops->pte_entry) 292 else if (ops->p4d_entry || ops->pud_entry || ops->pmd_entry || ops->pte_entry) 441 * pte_entry(), and/or hugetlb_entry(). If you don't set up for some of these 540 * will also not lock the PTEs for the pte_entry() callback. This is useful for
|
H A D | mapping_dirty_helpers.c | 232 .pte_entry = clean_record_pte, 241 .pte_entry = wp_pte,
|
H A D | ptdump.c | 148 .pte_entry = ptdump_pte_entry,
|
H A D | hugetlb_vmemmap.c | 156 .pte_entry = vmemmap_pte_entry,
|
H A D | mprotect.c | 570 .pte_entry = prot_none_pte_entry,
|
/linux-master/arch/riscv/mm/ |
H A D | pageattr.c | 90 .pte_entry = pageattr_pte_entry,
|
/linux-master/arch/s390/mm/ |
H A D | gmap.c | 2673 .pte_entry = __s390_enable_skey_pte, 2712 .pte_entry = __s390_reset_cmma, 2749 .pte_entry = s390_gather_pages,
|
Completed in 404 milliseconds