/linux-master/arch/x86/include/asm/ |
H A D | pgtable_32_types.h | 12 # define PMD_SIZE (1UL << PMD_SHIFT) macro 13 # define PMD_MASK (~(PMD_SIZE - 1))
|
H A D | pgtable_32_areas.h | 36 #define LDT_END_ADDR (LDT_BASE_ADDR + PMD_SIZE)
|
H A D | pgtable_64_types.h | 99 #define PMD_SIZE (_AC(1, UL) << PMD_SHIFT) macro 100 #define PMD_MASK (~(PMD_SIZE - 1))
|
/linux-master/arch/riscv/include/asm/ |
H A D | crash_reserve.h | 5 #define CRASH_ALIGN PMD_SIZE
|
/linux-master/arch/sh/include/asm/ |
H A D | pgtable-3level.h | 23 #define PMD_SIZE (1UL << PMD_SHIFT) macro 24 #define PMD_MASK (~(PMD_SIZE-1)) 26 #define PTRS_PER_PMD ((1 << PGDIR_SHIFT) / PMD_SIZE)
|
/linux-master/arch/m68k/mm/ |
H A D | kmap.c | 50 #define IO_SIZE PMD_SIZE 85 virtaddr += PMD_SIZE; 86 size -= PMD_SIZE; 254 if (!(virtaddr & (PMD_SIZE-1))) 269 physaddr += PMD_SIZE; 270 virtaddr += PMD_SIZE; 271 size -= PMD_SIZE; 379 virtaddr += PMD_SIZE; 380 size -= PMD_SIZE;
|
/linux-master/arch/powerpc/include/asm/nohash/64/ |
H A D | pgtable-4k.h | 31 #define PMD_SIZE (1UL << PMD_SHIFT) macro 32 #define PMD_MASK (~(PMD_SIZE-1))
|
/linux-master/include/asm-generic/ |
H A D | pgtable-nopmd.h | 22 #define PMD_SIZE (1UL << PMD_SHIFT) macro 23 #define PMD_MASK (~(PMD_SIZE-1))
|
/linux-master/arch/x86/mm/ |
H A D | init.c | 169 * the 0-ISA_END_ADDRESS range and secondly for the initial PMD_SIZE mapping. 355 unsigned long start = round_down(mr[i].start, PMD_SIZE); 356 unsigned long end = round_up(mr[i].end, PMD_SIZE); 422 end_pfn = PFN_DOWN(PMD_SIZE); 424 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); 426 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); 436 start_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); 438 end_pfn = round_down(limit_pfn, PFN_DOWN(PMD_SIZE)); 441 if (end_pfn > round_down(limit_pfn, PFN_DOWN(PMD_SIZE))) 442 end_pfn = round_down(limit_pfn, PFN_DOWN(PMD_SIZE)); [all...] |
H A D | init_64.c | 378 for (; size; phys += PMD_SIZE, size -= PMD_SIZE) { 430 unsigned long end = roundup((unsigned long)_brk_end, PMD_SIZE) - 1; 441 for (; vaddr + PMD_SIZE - 1 < vaddr_end; pmd++, vaddr += PMD_SIZE) { 521 paddr_next = (paddr & PMD_MASK) + PMD_SIZE; 835 * from unused_pmd_start to next PMD_SIZE boundary. 847 ALIGN(unused_pmd_start, PMD_SIZE) - unused_pmd_start); 855 unsigned long start = ALIGN_DOWN(addr, PMD_SIZE); 864 return !memchr_inv((void *)start, PAGE_UNUSED, PMD_SIZE); [all...] |
/linux-master/arch/arm/include/asm/ |
H A D | highmem.h | 8 #define PKMAP_BASE (PAGE_OFFSET - PMD_SIZE)
|
H A D | kfence.h | 24 flush_tlb_kernel_range(addr, addr + PMD_SIZE);
|
H A D | pgtable-2level.h | 87 #define PMD_SIZE (1UL << PMD_SHIFT) macro 88 #define PMD_MASK (~(PMD_SIZE-1))
|
/linux-master/drivers/dax/ |
H A D | cxl.c | 21 PMD_SIZE, IORESOURCE_DAX_KMEM);
|
H A D | dax-private.h | 106 if (align == PMD_SIZE && has_transparent_hugepage())
|
/linux-master/arch/s390/mm/ |
H A D | vmem.c | 90 * from unused_sub_pmd_start to next PMD_SIZE boundary. 99 ALIGN(unused_sub_pmd_start, PMD_SIZE) - unused_sub_pmd_start); 122 if (likely(IS_ALIGNED(unused_sub_pmd_start, PMD_SIZE))) 132 unsigned long page = ALIGN_DOWN(start, PMD_SIZE); 140 if (!IS_ALIGNED(start, PMD_SIZE)) 147 if (!IS_ALIGNED(end, PMD_SIZE)) 154 unsigned long page = ALIGN_DOWN(start, PMD_SIZE); 158 return !memchr_inv((void *)page, PAGE_UNUSED, PMD_SIZE); 240 if (IS_ALIGNED(addr, PMD_SIZE) && 241 IS_ALIGNED(next, PMD_SIZE)) { [all...] |
/linux-master/arch/um/include/asm/ |
H A D | pgtable-3level.h | 27 #define PMD_SIZE (1UL << PMD_SHIFT) macro 28 #define PMD_MASK (~(PMD_SIZE-1))
|
/linux-master/arch/arm64/include/asm/ |
H A D | kernel-pgtable.h | 21 #if defined(PMD_SIZE) && PMD_SIZE <= MIN_KIMG_ALIGN
|
/linux-master/arch/riscv/mm/ |
H A D | init.c | 210 * Make sure we align the reservation on PMD_SIZE since we will 215 vmlinux_end = (vmlinux_end + PMD_SIZE - 1) & PMD_MASK; 461 if (sz == PMD_SIZE) { 680 !(pa & (PMD_SIZE - 1)) && !(va & (PMD_SIZE - 1)) && size >= PMD_SIZE) 681 return PMD_SIZE; 801 PMD_SIZE, PAGE_KERNEL_EXEC); 803 set_satp_mode_pmd + PMD_SIZE, 804 set_satp_mode_pmd + PMD_SIZE, [all...] |
H A D | tlbflush.c | 190 else if (stride_size >= PMD_SIZE) 191 stride_size = PMD_SIZE; 212 start, end - start, PMD_SIZE); local
|
/linux-master/arch/nios2/mm/ |
H A D | ioremap.c | 33 if (end > PMD_SIZE) 34 end = PMD_SIZE; 70 address = (address + PMD_SIZE) & PMD_MASK;
|
/linux-master/arch/parisc/kernel/ |
H A D | pci-dma.c | 85 if (end > PMD_SIZE) 86 end = PMD_SIZE; 120 vaddr = (vaddr + PMD_SIZE) & PMD_MASK; 121 orig_vaddr += PMD_SIZE; 170 if (end > PMD_SIZE) 171 end = PMD_SIZE; 210 vaddr = (vaddr + PMD_SIZE) & PMD_MASK; 211 orig_vaddr += PMD_SIZE;
|
/linux-master/arch/m68k/include/asm/ |
H A D | pgtable_mm.h | 39 #define PMD_SIZE (1UL << PMD_SHIFT) macro 40 #define PMD_MASK (~(PMD_SIZE-1))
|
/linux-master/arch/arc/include/asm/ |
H A D | processor.h | 90 #define VMALLOC_SIZE ((CONFIG_ARC_KVADDR_SIZE << 20) - PMD_SIZE * 4)
|
/linux-master/arch/x86/kernel/ |
H A D | vmlinux.lds.S | 70 #define ALIGN_ENTRY_TEXT_BEGIN . = ALIGN(PMD_SIZE); 71 #define ALIGN_ENTRY_TEXT_END . = ALIGN(PMD_SIZE); 82 . = ALIGN(PMD_SIZE); \ 87 . = ALIGN(PMD_SIZE); \
|