/linux-master/arch/powerpc/include/asm/book3s/32/ |
H A D | pgtable.h | 124 * are an index to the second level table. The combined pgdir/pmd first 218 #define pmd_none(pmd) (!pmd_val(pmd)) 219 #define pmd_bad(pmd) (pmd_val(pmd) & _PMD_BAD) 220 #define pmd_present(pmd) (pmd_val(pmd) & _PMD_PRESENT_MASK) 343 #define pmd_pfn(pmd) (pmd_val(pmd) >> PAGE_SHIFT) 344 #define pmd_page(pmd) pfn_to_pag [all...] |
/linux-master/arch/x86/mm/ |
H A D | kasan_init_64.c | 38 static void __init kasan_populate_pmd(pmd_t *pmd, unsigned long addr, argument 43 if (pmd_none(*pmd)) { 50 if (p && pmd_set_huge(pmd, __pa(p), PAGE_KERNEL)) 56 pmd_populate_kernel(&init_mm, pmd, p); 59 pte = pte_offset_kernel(pmd, addr); 76 pmd_t *pmd; local 95 pmd = pmd_offset(pud, addr); 98 if (!pmd_leaf(*pmd)) 99 kasan_populate_pmd(pmd, addr, next, nid); 100 } while (pmd [all...] |
H A D | hugetlbpage.c | 23 * pmd_huge() returns 1 if @pmd is hugetlb related entry, that is normal 27 int pmd_huge(pmd_t pmd) argument 29 return !pmd_none(pmd) && 30 (pmd_val(pmd) & (_PAGE_PRESENT|_PAGE_PSE)) != _PAGE_PRESENT;
|
/linux-master/arch/arm/mm/ |
H A D | fault.c | 54 pmd_t *pmd; local 78 pmd = pmd_offset(pud, addr); 80 pr_cont(", *pmd=%08llx", (long long)pmd_val(*pmd)); 82 if (pmd_none(*pmd)) 85 if (pmd_bad(*pmd)) { 91 if (PageHighMem(pfn_to_page(pmd_val(*pmd) >> PAGE_SHIFT))) 94 pte = pte_offset_map(pmd, addr); 441 pmd_t *pmd, *pmd_k; local 470 pmd [all...] |
/linux-master/arch/x86/xen/ |
H A D | mmu_pv.c | 22 * pte/pmd/pgd, it converts the passed pfn into an mfn. Conversely, 23 * when reading the content back with __(pgd|pmd|pte)_val, it converts 95 pmdval_t xen_pmd_val(pmd_t pmd); 100 pmd_t xen_make_pmd(pmdval_t pmd); 403 __visible pmdval_t xen_pmd_val(pmd_t pmd) argument 405 return pte_mfn_to_pfn(pmd.pmd); 441 __visible pmd_t xen_make_pmd(pmdval_t pmd) argument 443 pmd = pte_pfn_to_mfn(pmd); 553 xen_pmd_walk(struct mm_struct *mm, pmd_t *pmd, void (*func)(struct mm_struct *mm, struct page *, enum pt_level), bool last, unsigned long limit) argument 576 pmd_t *pmd; local 1016 pmd_t *pmd = level2_kernel_pgt + pmd_index(vaddr); local 1056 xen_cleanmfnmap_pmd(pmd_t *pmd, bool unpin) argument 1851 pmd_t pmd; local 1896 pmd_t *pmd; local [all...] |
/linux-master/arch/x86/include/asm/ |
H A D | pgtable_types.h | 284 /* Extracts the PFN from a (pte|pmd|pud|pgd)val_t of a 4KB page */ 288 * Extracts the flags from a (pte|pmd|pud|pgd)val_t 396 return (pmd_t) { .pmd = val }; 399 static inline pmdval_t native_pmd_val(pmd_t pmd) argument 401 return pmd.pmd; 411 static inline pmdval_t native_pmd_val(pmd_t pmd) argument 413 return native_pgd_val(pmd.pud.p4d.pgd); 451 static inline pmdval_t pmd_pfn_mask(pmd_t pmd) argument 453 if (native_pmd_val(pmd) 459 pmd_flags_mask(pmd_t pmd) argument 464 pmd_flags(pmd_t pmd) argument [all...] |
H A D | pgtable-2level.h | 20 static inline void native_set_pmd(pmd_t *pmdp, pmd_t pmd) argument 22 *pmdp = pmd;
|
/linux-master/arch/m68k/include/asm/ |
H A D | mmu_context.h | 101 pmd_t *pmd; local 136 pmd = pmd_offset(pud, mmuar); 137 if (pmd_none(*pmd)) 140 pte = (mmuar >= PAGE_OFFSET) ? pte_offset_kernel(pmd, mmuar) 141 : pte_offset_map(pmd, mmuar);
|
/linux-master/arch/m68k/mm/ |
H A D | mcfmmu.c | 93 pmd_t *pmd; local 119 pmd = pmd_offset(pud, mmuar); 120 if (pmd_none(*pmd)) 123 pte = (KMAPAREA(mmuar)) ? pte_offset_kernel(pmd, mmuar) 124 : pte_offset_map(pmd, mmuar);
|
/linux-master/drivers/pmdomain/renesas/ |
H A D | rmobile-sysc.c | 303 struct device_node *np, *pmd; local 315 pmd = of_get_child_by_name(np, "pm-domains"); 316 if (!pmd) { 328 ret = rmobile_add_pm_domains(base, pmd, NULL); 329 of_node_put(pmd);
|
/linux-master/arch/nios2/mm/ |
H A D | fault.c | 233 pmd_t *pmd, *pmd_k; local 251 pmd = pmd_offset(pud, address); 255 set_pmd(pmd, *pmd_k);
|
/linux-master/arch/openrisc/mm/ |
H A D | fault.c | 298 pmd_t *pmd, *pmd_k; local 333 pmd = pmd_offset(pud, address); 339 set_pmd(pmd, *pmd_k);
|
/linux-master/arch/m68k/sun3x/ |
H A D | dvma.c | 98 pmd_t *pmd; local 101 if((pmd = pmd_alloc(&init_mm, pud, vaddr)) == NULL) { 115 if((pte = pte_alloc_kernel(pmd, vaddr)) == NULL) {
|
/linux-master/arch/openrisc/include/asm/ |
H A D | pgtable.h | 349 #define pmd_pfn(pmd) (pmd_val(pmd) >> PAGE_SHIFT) 350 #define pmd_page(pmd) (pfn_to_page(pmd_val(pmd) >> PAGE_SHIFT)) 352 static inline unsigned long pmd_page_vaddr(pmd_t pmd) argument 354 return ((unsigned long) __va(pmd_val(pmd) & PAGE_MASK));
|
/linux-master/mm/ |
H A D | hugetlb_vmemmap.c | 49 static int vmemmap_split_pmd(pmd_t *pmd, struct page *head, unsigned long start, argument 73 if (likely(pmd_leaf(*pmd))) { 82 /* Make pte visible before pmd. See comment in pmd_install(). */ 84 pmd_populate_kernel(&init_mm, pmd, pgtable); 95 static int vmemmap_pmd_entry(pmd_t *pmd, unsigned long addr, argument 107 head = pmd_leaf(*pmd) ? pmd_page(*pmd) : NULL; 124 pte_page(ptep_get(pte_offset_kernel(pmd, addr))); 133 return vmemmap_split_pmd(pmd, head, addr & PMD_MASK, vmemmap_walk);
|
H A D | khugepaged.c | 740 pmd_t *pmd, 753 pmd_ptl = pmd_lock(vma->vm_mm, pmd); 754 pmd_populate(vma->vm_mm, pmd, pmd_pgtable(orig_pmd)); 771 * @pmd: pointer to the new hugepage's PMD 780 pmd_t *pmd, 814 __collapse_huge_page_copy_failed(pte, pmd, orig_pmd, vma, 947 pmd_t **pmd) 951 *pmd = mm_find_pmd(mm, address); 952 if (!*pmd) 955 pmde = pmdp_get_lockless(*pmd); 739 __collapse_huge_page_copy_failed(pte_t *pte, pmd_t *pmd, pmd_t orig_pmd, struct vm_area_struct *vma, struct list_head *compound_pagelist) argument 778 __collapse_huge_page_copy(pte_t *pte, struct page *page, pmd_t *pmd, pmd_t orig_pmd, struct vm_area_struct *vma, unsigned long address, spinlock_t *ptl, struct list_head *compound_pagelist) argument 945 find_pmd_or_thp_or_none(struct mm_struct *mm, unsigned long address, pmd_t **pmd) argument 969 check_pmd_still_valid(struct mm_struct *mm, unsigned long address, pmd_t *pmd) argument 990 __collapse_huge_page_swapin(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long haddr, pmd_t *pmd, int referenced) argument 1092 pmd_t *pmd, _pmd; local 1256 pmd_t *pmd; local 1489 pmd_t *pmd, pgt_pmd; local 1695 pmd_t *pmd, pgt_pmd; local [all...] |
/linux-master/arch/riscv/mm/ |
H A D | pageattr.c | 55 static int pageattr_pmd_entry(pmd_t *pmd, unsigned long addr, argument 58 pmd_t val = pmdp_get(pmd); 62 set_pmd(pmd, val); 410 pmd_t *pmd; local 431 pmd = pmd_offset(pud, addr); 432 if (!pmd_present(pmdp_get(pmd))) 434 if (pmd_leaf(pmdp_get(pmd))) 437 pte = pte_offset_kernel(pmd, addr);
|
/linux-master/arch/sparc/include/asm/ |
H A D | page_32.h | 57 typedef struct { unsigned long pmd; } pmd_t; member in struct:__anon131 65 #define pmd_val(x) ((x).pmd)
|
/linux-master/arch/powerpc/mm/book3s64/ |
H A D | hash_hugepage.c | 36 pmd_t pmd = READ_ONCE(*pmdp); local 38 old_pmd = pmd_val(pmd); 158 * Hypervisor failure. Restore old pmd and return -1 183 * huge pmd.
|
/linux-master/mm/damon/ |
H A D | vaddr.c | 300 static int damon_mkold_pmd_entry(pmd_t *pmd, unsigned long addr, argument 307 if (pmd_trans_huge(pmdp_get(pmd))) { 308 ptl = pmd_lock(walk->mm, pmd); 309 pmde = pmdp_get(pmd); 317 damon_pmdp_mkold(pmd, walk->vma, addr); 324 pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); 437 static int damon_young_pmd_entry(pmd_t *pmd, unsigned long addr, argument 447 if (pmd_trans_huge(pmdp_get(pmd))) { 450 ptl = pmd_lock(walk->mm, pmd); 451 pmde = pmdp_get(pmd); [all...] |
/linux-master/include/linux/ |
H A D | mm.h | 530 pmd_t *pmd; /* Pointer to pmd entry matching member in struct:vm_fault 555 * is not NULL, otherwise pmd. 2373 unsigned long addr, pmd_t pmd); 2375 pmd_t pmd); 2806 int __pte_alloc(struct mm_struct *mm, pmd_t *pmd); 2807 int __pte_alloc_kernel(pmd_t *pmd); 2913 static inline spinlock_t *pte_lockptr(struct mm_struct *mm, pmd_t *pmd) argument 2915 return ptlock_ptr(page_ptdesc(pmd_page(*pmd))); 2938 static inline spinlock_t *pte_lockptr(struct mm_struct *mm, pmd_t *pmd) argument 2968 pte_offset_map(pmd_t *pmd, unsigned long addr) argument 2975 pte_offset_map_lock(struct mm_struct *mm, pmd_t *pmd, unsigned long addr, spinlock_t **ptlp) argument 3007 pmd_pgtable_page(pmd_t *pmd) argument 3013 pmd_ptdesc(pmd_t *pmd) argument 3018 pmd_lockptr(struct mm_struct *mm, pmd_t *pmd) argument 3043 pmd_lockptr(struct mm_struct *mm, pmd_t *pmd) argument 3055 pmd_lock(struct mm_struct *mm, pmd_t *pmd) argument [all...] |
/linux-master/drivers/edac/ |
H A D | xgene_edac.c | 515 u32 pmd; member in struct:xgene_edac_pmd_ctx 533 ctx->pmd * MAX_CPU_PER_PMD + cpu_idx, val, 573 ctx->pmd * MAX_CPU_PER_PMD + cpu_idx, val, 617 ctx->pmd * MAX_CPU_PER_PMD + cpu_idx, val, 677 ctx->pmd, val, val_hi, val_lo); 729 ctx->pmd, val, val_hi, val_lo); 741 if (!((PMD0_MERR_MASK << ctx->pmd) & pcp_hp_stat)) 791 PMD0_MERR_MASK << ctx->pmd); 794 PMD0_MERR_MASK << ctx->pmd); 870 snprintf(name, sizeof(name), "PMD%d", ctx->pmd); 881 xgene_edac_pmd_available(u32 efuse, int pmd) argument 893 u32 pmd; local 979 xgene_edac_pmd_remove(struct xgene_edac_pmd_ctx *pmd) argument 1814 struct xgene_edac_pmd_ctx *pmd; local 1968 struct xgene_edac_pmd_ctx *pmd; local [all...] |
/linux-master/arch/arm64/mm/ |
H A D | mmu.c | 206 pmd_t pmd = READ_ONCE(*pmdp); local 208 BUG_ON(pmd_sect(pmd)); 209 if (pmd_none(pmd)) { 218 pmd = READ_ONCE(*pmdp); 220 BUG_ON(pmd_bad(pmd)); 478 * We don't select ARCH_ENABLE_SPLIT_PMD_PTLOCK if pmd is 865 pmd_t *pmdp, pmd; local 870 pmd = READ_ONCE(*pmdp); 871 if (pmd_none(pmd)) 874 WARN_ON(!pmd_present(pmd)); 1012 pmd_t *pmdp, pmd; local 1235 pmd_t pmd; local [all...] |
/linux-master/arch/microblaze/mm/ |
H A D | pgtable.c | 200 pmd_t *pmd; local 208 pmd = pmd_offset(pud, addr & PAGE_MASK); 209 if (pmd_present(*pmd)) { 210 pte = pte_offset_kernel(pmd, addr & PAGE_MASK);
|
/linux-master/arch/powerpc/include/asm/book3s/64/ |
H A D | tlbflush-hash.h | 75 void flush_hash_table_pmd_range(struct mm_struct *mm, pmd_t *pmd, unsigned long addr);
|