Searched refs:pmd_entry (Results 1 - 21 of 21) sorted by relevance
/linux-master/arch/x86/kernel/ |
H A D | head64.c | 152 pmdval_t *pmd, pmd_entry; local 227 pmd_entry = __PAGE_KERNEL_LARGE_EXEC & ~_PAGE_GLOBAL; 229 pmd_entry &= RIP_REL_REF(__supported_pte_mask); 230 pmd_entry += sme_get_me_mask(); 231 pmd_entry += physaddr; 236 pmd[idx % PTRS_PER_PMD] = pmd_entry + i * PMD_SIZE;
|
/linux-master/include/linux/ |
H A D | pagewalk.h | 24 * @pmd_entry: if set, called for each non-empty PMD entry 65 int (*pmd_entry)(pmd_t *pmd, unsigned long addr, member in struct:mm_walk_ops 83 * Action for pud_entry / pmd_entry callbacks.
|
/linux-master/mm/ |
H A D | mapping_dirty_helpers.c | 233 .pmd_entry = wp_clean_pmd_entry, 242 .pmd_entry = wp_clean_pmd_entry,
|
H A D | pagewalk.c | 139 * This implies that each ->pmd_entry() handler 142 if (ops->pmd_entry) 143 err = ops->pmd_entry(pmd, addr, next, walk); 210 !(ops->pmd_entry || ops->pte_entry)) 255 else if (ops->pud_entry || ops->pmd_entry || ops->pte_entry) 292 else if (ops->p4d_entry || ops->pud_entry || ops->pmd_entry || ops->pte_entry) 440 * some caller-specific works for each entry, by setting up pmd_entry(),
|
H A D | ptdump.c | 147 .pmd_entry = ptdump_pmd_entry,
|
H A D | mincore.c | 176 .pmd_entry = mincore_pte_range,
|
H A D | madvise.c | 216 .pmd_entry = swapin_walk_pmd_entry, 534 .pmd_entry = madvise_cold_or_pageout_pte_range, 755 .pmd_entry = madvise_free_pte_range,
|
H A D | hugetlb_vmemmap.c | 155 .pmd_entry = vmemmap_pmd_entry,
|
H A D | hmm.c | 561 .pmd_entry = hmm_vma_walk_pmd,
|
H A D | migrate_device.c | 283 .pmd_entry = migrate_vma_collect_pmd,
|
H A D | mlock.c | 438 .pmd_entry = mlock_pte_range,
|
H A D | ksm.c | 647 .pmd_entry = break_ksm_pmd_entry, 652 .pmd_entry = break_ksm_pmd_entry,
|
H A D | mempolicy.c | 732 .pmd_entry = queue_folios_pte_range, 739 .pmd_entry = queue_folios_pte_range,
|
H A D | memory-failure.c | 846 .pmd_entry = hwpoison_pte_range,
|
H A D | memcontrol.c | 6268 .pmd_entry = mem_cgroup_count_precharge_pte_range, 6548 .pmd_entry = mem_cgroup_move_charge_pte_range,
|
/linux-master/arch/powerpc/mm/book3s64/ |
H A D | subpage_prot.c | 147 .pmd_entry = subpage_walk_pmd_entry,
|
/linux-master/mm/damon/ |
H A D | vaddr.c | 391 .pmd_entry = damon_mkold_pmd_entry, 535 .pmd_entry = damon_young_pmd_entry,
|
/linux-master/fs/ |
H A D | dax.c | 1212 pmd_t pmd_entry; local 1240 pmd_entry = mk_pmd(zero_page, vmf->vma->vm_page_prot); 1241 pmd_entry = pmd_mkhuge(pmd_entry); 1242 set_pmd_at(vmf->vma->vm_mm, pmd_addr, vmf->pmd, pmd_entry);
|
/linux-master/fs/proc/ |
H A D | task_mmu.c | 753 .pmd_entry = smaps_pte_range, 759 .pmd_entry = smaps_pte_range, 1240 .pmd_entry = clear_refs_pte_range, 1615 .pmd_entry = pagemap_pmd_range, 2331 .pmd_entry = pagemap_scan_pmd_entry, 2690 .pmd_entry = gather_pte_stats,
|
/linux-master/arch/riscv/mm/ |
H A D | pageattr.c | 89 .pmd_entry = pageattr_pmd_entry,
|
/linux-master/arch/s390/mm/ |
H A D | gmap.c | 2531 .pmd_entry = thp_split_walk_pmd_entry, 2583 .pmd_entry = __zap_zero_pages, 2674 .pmd_entry = __s390_enable_skey_pmd,
|
Completed in 355 milliseconds