/linux-master/arch/x86/include/asm/ |
H A D | pgtable_32_types.h | 12 # define PMD_SIZE (1UL << PMD_SHIFT) macro 13 # define PMD_MASK (~(PMD_SIZE - 1))
|
H A D | pgtable_32_areas.h | 36 #define LDT_END_ADDR (LDT_BASE_ADDR + PMD_SIZE)
|
H A D | pgtable_64_types.h | 99 #define PMD_SIZE (_AC(1, UL) << PMD_SHIFT) macro 100 #define PMD_MASK (~(PMD_SIZE - 1))
|
/linux-master/arch/riscv/include/asm/ |
H A D | crash_reserve.h | 5 #define CRASH_ALIGN PMD_SIZE
|
/linux-master/arch/sh/include/asm/ |
H A D | pgtable-3level.h | 23 #define PMD_SIZE (1UL << PMD_SHIFT) macro 24 #define PMD_MASK (~(PMD_SIZE-1)) 26 #define PTRS_PER_PMD ((1 << PGDIR_SHIFT) / PMD_SIZE)
|
/linux-master/arch/m68k/mm/ |
H A D | kmap.c | 50 #define IO_SIZE PMD_SIZE 85 virtaddr += PMD_SIZE; 86 size -= PMD_SIZE; 254 if (!(virtaddr & (PMD_SIZE-1))) 269 physaddr += PMD_SIZE; 270 virtaddr += PMD_SIZE; 271 size -= PMD_SIZE; 379 virtaddr += PMD_SIZE; 380 size -= PMD_SIZE;
|
/linux-master/arch/powerpc/include/asm/nohash/64/ |
H A D | pgtable-4k.h | 31 #define PMD_SIZE (1UL << PMD_SHIFT) macro 32 #define PMD_MASK (~(PMD_SIZE-1))
|
/linux-master/include/asm-generic/ |
H A D | pgtable-nopmd.h | 22 #define PMD_SIZE (1UL << PMD_SHIFT) macro 23 #define PMD_MASK (~(PMD_SIZE-1))
|
/linux-master/arch/x86/mm/ |
H A D | init.c | 170 * the 0-ISA_END_ADDRESS range and secondly for the initial PMD_SIZE mapping. 352 unsigned long start = round_down(mr[i].start, PMD_SIZE); 353 unsigned long end = round_up(mr[i].end, PMD_SIZE); 419 end_pfn = PFN_DOWN(PMD_SIZE); 421 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); 423 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); 433 start_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); 435 end_pfn = round_down(limit_pfn, PFN_DOWN(PMD_SIZE)); 438 if (end_pfn > round_down(limit_pfn, PFN_DOWN(PMD_SIZE))) 439 end_pfn = round_down(limit_pfn, PFN_DOWN(PMD_SIZE)); [all...] |
H A D | init_64.c | 378 for (; size; phys += PMD_SIZE, size -= PMD_SIZE) { 430 unsigned long end = roundup((unsigned long)_brk_end, PMD_SIZE) - 1; 441 for (; vaddr + PMD_SIZE - 1 < vaddr_end; pmd++, vaddr += PMD_SIZE) { 521 paddr_next = (paddr & PMD_MASK) + PMD_SIZE; 835 * from unused_pmd_start to next PMD_SIZE boundary. 847 ALIGN(unused_pmd_start, PMD_SIZE) - unused_pmd_start); 855 unsigned long start = ALIGN_DOWN(addr, PMD_SIZE); 864 return !memchr_inv((void *)start, PAGE_UNUSED, PMD_SIZE); [all...] |
/linux-master/arch/arm/include/asm/ |
H A D | highmem.h | 8 #define PKMAP_BASE (PAGE_OFFSET - PMD_SIZE)
|
H A D | kfence.h | 24 flush_tlb_kernel_range(addr, addr + PMD_SIZE);
|
H A D | pgtable-2level.h | 87 #define PMD_SIZE (1UL << PMD_SHIFT) macro 88 #define PMD_MASK (~(PMD_SIZE-1))
|
/linux-master/drivers/dax/ |
H A D | cxl.c | 21 PMD_SIZE, IORESOURCE_DAX_KMEM);
|
/linux-master/arch/s390/mm/ |
H A D | vmem.c | 93 * from unused_sub_pmd_start to next PMD_SIZE boundary. 102 ALIGN(unused_sub_pmd_start, PMD_SIZE) - unused_sub_pmd_start); 125 if (likely(IS_ALIGNED(unused_sub_pmd_start, PMD_SIZE))) 135 unsigned long page = ALIGN_DOWN(start, PMD_SIZE); 143 if (!IS_ALIGNED(start, PMD_SIZE)) 150 if (!IS_ALIGNED(end, PMD_SIZE)) 157 unsigned long page = ALIGN_DOWN(start, PMD_SIZE); 161 return !memchr_inv((void *)page, PAGE_UNUSED, PMD_SIZE); 243 if (IS_ALIGNED(addr, PMD_SIZE) && 244 IS_ALIGNED(next, PMD_SIZE)) { [all...] |
/linux-master/arch/um/include/asm/ |
H A D | pgtable-3level.h | 27 #define PMD_SIZE (1UL << PMD_SHIFT) macro 28 #define PMD_MASK (~(PMD_SIZE-1))
|
/linux-master/arch/arm64/include/asm/ |
H A D | kernel-pgtable.h | 21 #if defined(PMD_SIZE) && PMD_SIZE <= MIN_KIMG_ALIGN
|
/linux-master/arch/arm64/mm/ |
H A D | hugetlbpage.c | 60 case PMD_SIZE: 95 *pgsize = PMD_SIZE; 114 case PMD_SIZE: 118 *pgsize = PMD_SIZE; 282 } else if (sz == PMD_SIZE) { 326 if (!(sz == PMD_SIZE || sz == CONT_PMD_SIZE) && 349 case PMD_SIZE: 350 return PUD_SIZE - PMD_SIZE; 352 return PMD_SIZE - CONT_PTE_SIZE; 369 } else if (pagesize != PUD_SIZE && pagesize != PMD_SIZE) { [all...] |
/linux-master/arch/riscv/mm/ |
H A D | init.c | 225 * Make sure we align the reservation on PMD_SIZE since we will 230 vmlinux_end = (vmlinux_end + PMD_SIZE - 1) & PMD_MASK; 476 if (sz == PMD_SIZE) { 695 !(pa & (PMD_SIZE - 1)) && !(va & (PMD_SIZE - 1)) && size >= PMD_SIZE) 696 return PMD_SIZE; 816 PMD_SIZE, PAGE_KERNEL_EXEC); 818 set_satp_mode_pmd + PMD_SIZE, 819 set_satp_mode_pmd + PMD_SIZE, [all...] |
H A D | tlbflush.c | 157 else if (stride_size >= PMD_SIZE) 158 stride_size = PMD_SIZE; 179 start, end - start, PMD_SIZE); local
|
/linux-master/arch/nios2/mm/ |
H A D | ioremap.c | 33 if (end > PMD_SIZE) 34 end = PMD_SIZE; 70 address = (address + PMD_SIZE) & PMD_MASK;
|
/linux-master/arch/x86/virt/svm/ |
H A D | sev.c | 168 if (IS_ALIGNED(pa, PMD_SIZE)) 187 pa = ALIGN_DOWN(pa, PMD_SIZE); 188 if (e820__mapped_any(pa, pa + PMD_SIZE, E820_TYPE_RAM)) { 190 e820__range_update(pa, PMD_SIZE, E820_TYPE_RAM, E820_TYPE_RESERVED); 191 e820__range_update_table(e820_table_kexec, pa, PMD_SIZE, E820_TYPE_RAM, E820_TYPE_RESERVED); 192 e820__range_update_table(e820_table_firmware, pa, PMD_SIZE, E820_TYPE_RAM, E820_TYPE_RESERVED);
|
/linux-master/arch/parisc/kernel/ |
H A D | pci-dma.c | 85 if (end > PMD_SIZE) 86 end = PMD_SIZE; 120 vaddr = (vaddr + PMD_SIZE) & PMD_MASK; 121 orig_vaddr += PMD_SIZE; 170 if (end > PMD_SIZE) 171 end = PMD_SIZE; 210 vaddr = (vaddr + PMD_SIZE) & PMD_MASK; 211 orig_vaddr += PMD_SIZE;
|
/linux-master/arch/m68k/include/asm/ |
H A D | pgtable_mm.h | 39 #define PMD_SIZE (1UL << PMD_SHIFT) macro 40 #define PMD_MASK (~(PMD_SIZE-1))
|
/linux-master/arch/arc/include/asm/ |
H A D | processor.h | 90 #define VMALLOC_SIZE ((CONFIG_ARC_KVADDR_SIZE << 20) - PMD_SIZE * 4)
|