/linux-master/arch/x86/mm/ |
H A D | hugetlbpage.c | 127 if (size == PMD_SIZE)
|
H A D | mem_encrypt_boot.S | 29 * - intermediate copy buffer (PMD_SIZE) 126 movq $PMD_SIZE, %r12
|
H A D | ident_map.c | 11 for (; addr < end; addr += PMD_SIZE) {
|
/linux-master/arch/csky/mm/ |
H A D | init.c | 185 vaddr += PMD_SIZE; 198 fixrange_init(vaddr, vaddr + PMD_SIZE, swapper_pg_dir);
|
/linux-master/arch/arm/mm/ |
H A D | kasan_init.c | 80 * Early shadow mappings are PMD_SIZE aligned, so if the 191 for (; start && start < end; start += PMD_SIZE) 280 * 2. PKMAP_BASE ~ PKMAP_BASE+PMD_SIZE's shadow and MODULES_VADDR 281 * ~ MODULES_END's shadow is in the same PMD_SIZE, so we can't 286 create_mapping((void *)PKMAP_BASE, (void *)(PKMAP_BASE + PMD_SIZE));
|
H A D | ioremap.c | 174 addr += PMD_SIZE; 207 addr += PMD_SIZE; 238 addr += PMD_SIZE;
|
/linux-master/arch/mips/include/asm/ |
H A D | pgtable-64.h | 50 #define PMD_SIZE (1UL << PMD_SHIFT) macro 51 #define PMD_MASK (~(PMD_SIZE-1))
|
/linux-master/mm/ |
H A D | page_vma_mapped.c | 278 step_forward(pvmw, PMD_SIZE); 295 if ((pvmw->address & (PMD_SIZE - PAGE_SIZE)) == 0) {
|
H A D | page_table_check.c | 171 page_table_check_clear(pmd_pfn(pmd), PMD_SIZE >> PAGE_SHIFT); 237 page_table_check_set(pmd_pfn(pmd), PMD_SIZE >> PAGE_SHIFT,
|
/linux-master/arch/x86/mm/pat/ |
H A D | set_memory.c | 249 return __pa_symbol(roundup(_brk_end, PMD_SIZE) - 1) >> PAGE_SHIFT; 1118 pfninc = PMD_SIZE >> PAGE_SHIFT; 1120 lpinc = PMD_SIZE; 1260 if (start & (PMD_SIZE - 1)) { 1261 unsigned long next_page = (start + PMD_SIZE) & PMD_MASK; 1273 while (end - start >= PMD_SIZE) { 1277 __unmap_pmd_range(pud, pmd, start, start + PMD_SIZE); 1279 start += PMD_SIZE; 1390 if (start & (PMD_SIZE - 1)) { 1392 unsigned long next_page = (start + PMD_SIZE) [all...] |
/linux-master/arch/x86/xen/ |
H A D | p2m.c | 400 (unsigned long)(p2m + pfn) + i * PMD_SIZE); 416 PMD_SIZE * PMDS_PER_MID_PAGE); 417 vm_area_register_early(&vm, PMD_SIZE * PMDS_PER_MID_PAGE); 484 vaddr = addr & ~(PMD_SIZE * PMDS_PER_MID_PAGE - 1); 513 vaddr += PMD_SIZE;
|
/linux-master/arch/riscv/mm/ |
H A D | kasan_init.c | 66 if (pmd_none(pmdp_get(pmdp)) && IS_ALIGNED(vaddr, PMD_SIZE) && 67 (next - vaddr) >= PMD_SIZE) { 68 phys_addr = memblock_phys_alloc(PMD_SIZE, PMD_SIZE); 71 memset(__va(phys_addr), KASAN_SHADOW_INIT, PMD_SIZE); local
|
/linux-master/arch/arm64/kvm/ |
H A D | mmu.c | 96 n += DIV_ROUND_UP(range, PMD_SIZE); 1278 * the stage2 PFN and IPA accordingly. Only PMD_SIZE THPs are currently 1295 if (fault_supports_stage2_huge_mapping(memslot, hva, PMD_SIZE)) { 1301 if (sz < PMD_SIZE) 1308 return PMD_SIZE; 1336 if ((hva & (PMD_SIZE - 1)) == (pa & (PMD_SIZE - 1)) && 1337 ALIGN_DOWN(hva, PMD_SIZE) >= vma->vm_start && 1338 ALIGN(hva, PMD_SIZE) <= vma->vm_end) 1455 if (fault_supports_stage2_huge_mapping(memslot, hva, PMD_SIZE)) [all...] |
/linux-master/arch/x86/kernel/ |
H A D | head64.c | 112 for (; vaddr < vaddr_end; vaddr += PMD_SIZE) { 233 for (i = 0; i < DIV_ROUND_UP(_end - _text, PMD_SIZE); i++) { 236 pmd[idx % PTRS_PER_PMD] = pmd_entry + i * PMD_SIZE;
|
H A D | setup_percpu.c | 142 * On 64bit, use PMD_SIZE for atom_size so that embedded 149 atom_size = PMD_SIZE;
|
/linux-master/arch/powerpc/mm/book3s64/ |
H A D | hash_tlb.c | 230 addr = ALIGN_DOWN(addr, PMD_SIZE);
|
/linux-master/arch/m68k/sun3x/ |
H A D | dvma.c | 121 end3 = (vaddr + (PMD_SIZE-1)) & PMD_MASK;
|
/linux-master/drivers/dax/hmem/ |
H A D | hmem.c | 31 mri->target_node, PMD_SIZE, flags);
|
/linux-master/arch/powerpc/include/asm/ |
H A D | pgtable.h | 193 * some of the restrictions. We don't check for PMD_SIZE because our 203 * With 4K page size and 2M PMD_SIZE, we can align 205 * starting from 128MB. Hence align things with PMD_SIZE. 208 return IS_ALIGNED(vmemmap_size, PMD_SIZE);
|
/linux-master/arch/riscv/kvm/ |
H A D | mmu.c | 562 WARN_ON(size != PAGE_SIZE && size != PMD_SIZE && size != PUD_SIZE); 580 WARN_ON(size != PAGE_SIZE && size != PMD_SIZE && size != PUD_SIZE); 629 if (vma_pagesize == PMD_SIZE || vma_pagesize == PUD_SIZE) 644 vma_pagesize != PMD_SIZE &&
|
/linux-master/arch/sh/mm/ |
H A D | init.c | 199 vaddr += PMD_SIZE; 331 end = (FIXADDR_TOP + PMD_SIZE - 1) & PMD_MASK;
|
/linux-master/arch/alpha/include/asm/ |
H A D | pgtable.h | 32 #define PMD_SIZE (1UL << PMD_SHIFT) macro 33 #define PMD_MASK (~(PMD_SIZE-1))
|
/linux-master/arch/riscv/include/asm/ |
H A D | pgtable-64.h | 43 #define PMD_SIZE (_AC(1, UL) << PMD_SHIFT) macro 44 #define PMD_MASK (~(PMD_SIZE - 1))
|
/linux-master/arch/sparc/include/asm/ |
H A D | pgtable_32.h | 15 #define PMD_SIZE (1UL << PMD_SHIFT) macro 16 #define PMD_MASK (~(PMD_SIZE-1))
|
/linux-master/include/linux/ |
H A D | mmu_notifier.h | 540 PMD_SIZE); \ 562 ___address + PMD_SIZE); \
|