/linux-master/arch/csky/mm/ |
H A D | fault.c | 119 pmd_t *pmd, *pmd_k; local 154 pmd = pmd_offset(pud, addr); 160 set_pmd(pmd, *pmd_k);
|
/linux-master/mm/ |
H A D | mincore.c | 100 static int mincore_pte_range(pmd_t *pmd, unsigned long addr, unsigned long end, argument 109 ptl = pmd_trans_huge_lock(pmd, vma); 116 ptep = pte_offset_map_lock(walk->mm, pmd, addr, &ptl);
|
H A D | mremap.c | 59 pmd_t *pmd; local 65 pmd = pmd_offset(pud, addr); 66 if (pmd_none(*pmd)) 69 return pmd; 90 pmd_t *pmd; local 96 pmd = pmd_alloc(mm, pud, addr); 97 if (!pmd) 100 VM_BUG_ON(pmd_trans_huge(*pmd)); 102 return pmd; 241 pmd_t pmd; local [all...] |
H A D | mlock.c | 363 static int mlock_pte_range(pmd_t *pmd, unsigned long addr, argument 375 ptl = pmd_trans_huge_lock(pmd, vma); 377 if (!pmd_present(*pmd)) 379 if (is_huge_zero_pmd(*pmd)) 381 folio = pmd_folio(*pmd); 389 start_pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl);
|
H A D | vmalloc.c | 94 static int vmap_pte_range(pmd_t *pmd, unsigned long addr, unsigned long end, argument 104 pte = pte_alloc_kernel_track(pmd, addr, mask); 134 static int vmap_try_huge_pmd(pmd_t *pmd, unsigned long addr, unsigned long end, argument 153 if (pmd_present(*pmd) && !pmd_free_pte_page(pmd, addr)) 156 return pmd_set_huge(pmd, phys_addr, prot); 163 pmd_t *pmd; local 166 pmd = pmd_alloc_track(&init_mm, pud, addr, mask); 167 if (!pmd) 172 if (vmap_try_huge_pmd(pmd, add 349 vunmap_pte_range(pmd_t *pmd, unsigned long addr, unsigned long end, pgtbl_mod_mask *mask) argument 365 pmd_t *pmd; local 486 vmap_pages_pte_range(pmd_t *pmd, unsigned long addr, unsigned long end, pgprot_t prot, struct page **pages, int *nr, pgtbl_mod_mask *mask) argument 521 pmd_t *pmd; local 746 pmd_t *pmd; local [all...] |
H A D | mapping_dirty_helpers.c | 118 * wp_clean_pmd_entry - The pagewalk pmd callback. 121 * WARN() if encountering a dirty huge pmd. 126 static int wp_clean_pmd_entry(pmd_t *pmd, unsigned long addr, unsigned long end, argument 129 pmd_t pmdval = pmdp_get_lockless(pmd); 131 /* Do not split a huge pmd, present or migrated */
|
/linux-master/arch/x86/boot/compressed/ |
H A D | ident_map_64.c | 198 pmd_t pmd; local 224 pmd = __pmd((unsigned long)pte | info->kernpg_flag); 225 set_pmd(pmdp, pmd);
|
/linux-master/include/linux/ |
H A D | userfaultfd_k.h | 156 * Never enable huge pmd sharing on some uffd registered vmas: 163 * with huge pmd sharing this would *also* setup the second UFFD-registered 205 pmd_t pmd) 207 return userfaultfd_wp(vma) && pmd_uffd_wp(pmd); 311 pmd_t pmd) 204 userfaultfd_huge_pmd_wp(struct vm_area_struct *vma, pmd_t pmd) argument 310 userfaultfd_huge_pmd_wp(struct vm_area_struct *vma, pmd_t pmd) argument
|
H A D | pfn_t.h | 113 pmd_t pmd_mkdevmap(pmd_t pmd);
|
/linux-master/arch/mips/mm/ |
H A D | fault.c | 289 pmd_t *pmd, *pmd_k; local 309 pmd = pmd_offset(pud, address); 313 set_pmd(pmd, *pmd_k);
|
/linux-master/arch/sh/mm/ |
H A D | cache-sh4.c | 221 pmd_t *pmd; local 234 pmd = pmd_off(vma->vm_mm, address); 235 pte = pte_offset_kernel(pmd, address);
|
/linux-master/mm/kasan/ |
H A D | shadow.c | 191 pmd_t *pmd; local 204 pmd = pmd_offset(pud, addr); 205 if (pmd_none(*pmd)) 207 if (pmd_leaf(*pmd)) 209 pte = pte_offset_kernel(pmd, addr);
|
/linux-master/arch/arm64/mm/ |
H A D | hugetlbpage.c | 302 pmd_t *pmdp, pmd; local 325 pmd = READ_ONCE(*pmdp); 327 pmd_none(pmd)) 329 if (pmd_leaf(pmd) || !pmd_present(pmd))
|
/linux-master/arch/x86/mm/ |
H A D | ioremap.c | 862 pmd_t *pmd = pmd_offset(pud, addr); local 864 return pmd; 879 pmd_t *pmd; local 889 pmd = early_ioremap_pmd(fix_to_virt(FIX_BTMAP_BEGIN)); 891 pmd_populate_kernel(&init_mm, pmd, bm_pte); 901 if (pmd != early_ioremap_pmd(fix_to_virt(FIX_BTMAP_END))) { 903 printk(KERN_WARNING "pmd %p != %p\n", 904 pmd, early_ioremap_pmd(fix_to_virt(FIX_BTMAP_END)));
|
H A D | mem_encrypt_amd.c | 160 pmdval_t pmd_flags, pmd; local 166 pmd = map ? (paddr & PMD_MASK) + pmd_flags : 0; 167 __early_make_pgtable((unsigned long)vaddr, pmd);
|
H A D | kmmio.c | 133 static void clear_pmd_presence(pmd_t *pmd, bool clear, pmdval_t *old) argument 136 pmdval_t v = pmd_val(*pmd); 139 new_pmd = pmd_mkinvalid(*pmd); 144 set_pmd(pmd, new_pmd);
|
/linux-master/fs/ |
H A D | dax.c | 709 * pte or pmd without holding the respective lock, so we are 1230 ptl = pmd_lock(vmf->vma->vm_mm, vmf->pmd); 1231 if (!pmd_none(*(vmf->pmd))) { 1237 pgtable_trans_huge_deposit(vma->vm_mm, vmf->pmd, pgtable); 1242 set_pmd_at(vmf->vma->vm_mm, pmd_addr, vmf->pmd, pmd_entry); 1640 * @pmd: distinguish whether it is a pmd fault 1644 struct xa_state *xas, void **entry, bool pmd) 1648 size_t size = pmd ? PMD_SIZE : PAGE_SIZE; 1651 unsigned long entry_flags = pmd 1642 dax_fault_iter(struct vm_fault *vmf, const struct iomap_iter *iter, pfn_t *pfnp, struct xa_state *xas, void **entry, bool pmd) argument [all...] |
/linux-master/arch/loongarch/mm/ |
H A D | kasan_init.c | 30 #define __pmd_none(early, pmd) (early ? (pmd_val(pmd) == 0) : \ 31 (__pa(pmd_val(pmd)) == (unsigned long)__pa(kasan_early_shadow_pte)))
|
/linux-master/drivers/net/ethernet/intel/ice/devlink/ |
H A D | devlink_port.c | 95 "%-6u", options[i].pmd); 207 if (count == options[j].pmd) { 283 attrs->lanes = max_t(int, attrs->lanes, options[i].pmd);
|
/linux-master/arch/x86/kernel/ |
H A D | tboot.c | 119 pmd_t *pmd; local 129 pmd = pmd_alloc(&tboot_mm, pud, vaddr); 130 if (!pmd) 132 pte = pte_alloc_map(&tboot_mm, pmd, vaddr);
|
/linux-master/arch/arm/mach-sa1100/ |
H A D | assabet.c | 606 pmd_t *pmd; local 608 pmd = pmd_off_k(virt); 609 *pmd = __pmd(phys | prot); 610 flush_pmd_entry(pmd);
|
/linux-master/arch/m68k/mm/ |
H A D | motorola.c | 251 unsigned long pmd, last; local 264 pmd = pgd_page_vaddr(kernel_pg_dir[i]); 265 if (pmd > last) 266 last = pmd;
|
/linux-master/arch/parisc/kernel/ |
H A D | vmlinux.lds.S | 17 #define BSS_FIRST_SECTIONS *(.data..vm0.pmd) \
|
/linux-master/arch/x86/include/asm/ |
H A D | pgtable_64_types.h | 22 typedef struct { pmdval_t pmd; } pmd_t; member in struct:__anon217
|
/linux-master/arch/sparc/mm/ |
H A D | init_64.c | 1640 pmd_t *pmd; local 1671 pmd = pmd_offset(pud, addr); 1672 if (pmd_none(*pmd)) 1675 if (pmd_leaf(*pmd)) 1676 return pfn_valid(pmd_pfn(*pmd)); 1678 pte = pte_offset_kernel(pmd, addr); 1726 pmd_t *pmd) 1736 pmd_val(*pmd) = pte_val | _PAGE_PMD_HUGE; 1753 pmd_val(*pmd) = pte_val; 1757 pmd 1724 kernel_map_hugepmd(unsigned long vstart, unsigned long vend, pmd_t *pmd) argument 1791 pmd_t *pmd; local 2616 pmd_t *pmd; local 2959 update_mmu_cache_pmd(struct vm_area_struct *vma, unsigned long addr, pmd_t *pmd) argument [all...] |