/linux-master/arch/x86/include/asm/ |
H A D | pgtable_32.h | 56 * With PAE paging (PTRS_PER_PMD > 1), we allocate PTRS_PER_PGD == 4 pages for 59 #if PTRS_PER_PMD > 1 60 #define PAGE_TABLE_SIZE(pages) (((pages) / PTRS_PER_PMD) + PTRS_PER_PGD)
|
H A D | pgtable-3level_types.h | 45 #define PTRS_PER_PMD 512 macro
|
H A D | pgtable_64_types.h | 92 #define PTRS_PER_PMD 512 macro
|
/linux-master/arch/m68k/include/asm/ |
H A D | pgtable_mm.h | 60 #define PTRS_PER_PMD 1 macro 65 #define PTRS_PER_PMD 1 macro 69 #define PTRS_PER_PMD 128 macro
|
/linux-master/arch/sh/include/asm/ |
H A D | pgtable-3level.h | 26 #define PTRS_PER_PMD ((1 << PGDIR_SHIFT) / PMD_SIZE) macro
|
/linux-master/arch/um/include/asm/ |
H A D | pgtable-3level.h | 36 #define PTRS_PER_PMD 512 macro 39 #define PTRS_PER_PMD 1024 macro
|
/linux-master/arch/sh/mm/ |
H A D | pgtable.c | 27 PTRS_PER_PMD * (1<<PTE_MAGNITUDE),
|
/linux-master/arch/powerpc/include/asm/nohash/64/ |
H A D | pgtable-4k.h | 25 #define PTRS_PER_PMD (1 << PMD_INDEX_SIZE) macro
|
/linux-master/include/asm-generic/ |
H A D | pgtable-nopmd.h | 21 #define PTRS_PER_PMD 1 macro
|
/linux-master/arch/x86/kernel/ |
H A D | head64.c | 51 extern pmd_t early_dynamic_pgts[EARLY_DYNAMIC_PAGE_TABLES][PTRS_PER_PMD]; 123 early_snp_set_memory_shared(__pa(vaddr), __pa(vaddr), PTRS_PER_PMD); local 146 pmd_t (*early_pgts)[PTRS_PER_PMD] = RIP_REL_REF(early_dynamic_pgts); 236 pmd[idx % PTRS_PER_PMD] = pmd_entry + i * PMD_SIZE; 267 for (; i < PTRS_PER_PMD; i++) 344 memset(pmd_p, 0, sizeof(*pmd_p) * PTRS_PER_PMD);
|
/linux-master/arch/m68k/mm/ |
H A D | init.c | 111 for (j = 0; j < PTRS_PER_PMD; j++) {
|
/linux-master/arch/powerpc/mm/kasan/ |
H A D | init_book3s_64.c | 71 for (i = 0; i < PTRS_PER_PMD; i++)
|
/linux-master/arch/x86/virt/svm/ |
H A D | sev.c | 316 pfn_i = ALIGN_DOWN(pfn, PTRS_PER_PMD); 317 pfn_end = pfn_i + PTRS_PER_PMD; 429 (!IS_ALIGNED(pfn, PTRS_PER_PMD) || !pfn_valid(pfn + PTRS_PER_PMD - 1)))
|
/linux-master/arch/mips/include/asm/ |
H A D | pgtable-64.h | 31 * single 4K page, giving 512 (== PTRS_PER_PMD) 8 byte pointers to page 130 #define PTRS_PER_PMD ((PAGE_SIZE << PMD_TABLE_ORDER) / sizeof(pmd_t)) macro 144 min(PTRS_PER_PGD * PTRS_PER_PUD * PTRS_PER_PMD * PTRS_PER_PTE * PAGE_SIZE, \ 234 extern pmd_t invalid_pmd_table[PTRS_PER_PMD];
|
/linux-master/arch/powerpc/mm/book3s64/ |
H A D | hash_pgtable.c | 276 pgtable_slot = (pgtable_t *)pmdp + PTRS_PER_PMD; 294 pgtable_slot = (pgtable_t *)pmdp + PTRS_PER_PMD; 360 pgtable_slot = (pgtable_t *)pmdp + PTRS_PER_PMD;
|
/linux-master/arch/arm/mm/ |
H A D | idmap.c | 42 PTRS_PER_PMD * sizeof(pmd_t));
|
/linux-master/arch/sparc/include/asm/ |
H A D | pgtsrmmu.h | 21 #define SRMMU_PMD_TABLE_SIZE (PTRS_PER_PMD*4)
|
/linux-master/arch/arc/mm/ |
H A D | init.c | 191 BUILD_BUG_ON((PTRS_PER_PMD * sizeof(pmd_t)) > PAGE_SIZE);
|
/linux-master/arch/mips/mm/ |
H A D | pgtable-64.c | 52 end = p + PTRS_PER_PMD;
|
/linux-master/arch/x86/mm/ |
H A D | mem_encrypt_identity.c | 139 memset(pmd, 0, sizeof(*pmd) * PTRS_PER_PMD); 140 ppd->pgtable_area += sizeof(*pmd) * PTRS_PER_PMD; 272 entries += (DIV_ROUND_UP(len, PUD_SIZE) + 1) * sizeof(pmd_t) * PTRS_PER_PMD; 283 tables += DIV_ROUND_UP(entries, PUD_SIZE) * sizeof(pmd_t) * PTRS_PER_PMD;
|
/linux-master/arch/arm/include/asm/ |
H A D | pgtable-2level.h | 71 #define PTRS_PER_PMD 1 macro
|
/linux-master/arch/riscv/mm/ |
H A D | pgtable.c | 84 for (i = 0; i < PTRS_PER_PMD; i++) {
|
/linux-master/arch/loongarch/mm/ |
H A D | pgtable.c | 78 end = p + PTRS_PER_PMD;
|
/linux-master/arch/s390/mm/ |
H A D | pageattr.c | 71 mask = ~(PTRS_PER_PMD * sizeof(pmd_t) - 1); 228 for (i = 0; i < PTRS_PER_PMD; i++) { 235 update_page_count(PG_DIRECT_MAP_1M, PTRS_PER_PMD);
|
/linux-master/arch/x86/power/ |
H A D | hibernate_32.c | 101 for (pmd_idx = 0; pmd_idx < PTRS_PER_PMD; pmd++, pmd_idx++) {
|