/linux-master/arch/powerpc/mm/kasan/ |
H A D | init_32.c | 27 for (i = 0; i < PTRS_PER_PTE; i++, ptep++)
|
/linux-master/arch/x86/kernel/ |
H A D | espfix_64.c | 89 #define ESPFIX_PTE_CLONES (PTRS_PER_PTE/PTE_STRIDE)
|
/linux-master/arch/arm64/mm/ |
H A D | fixmap.c | 34 static pte_t bm_pte[NR_BM_PTE_TABLES][PTRS_PER_PTE] __page_aligned_bss;
|
/linux-master/arch/powerpc/mm/ |
H A D | pgtable_32.c | 45 addr += PGDIR_SIZE, ptep += PTRS_PER_PTE, pmdp++)
|
/linux-master/arch/loongarch/include/asm/ |
H A D | pgtable.h | 54 #define PTRS_PER_PTE (PAGE_SIZE >> 3) macro 96 min(PTRS_PER_PGD * PTRS_PER_PUD * PTRS_PER_PMD * PTRS_PER_PTE * PAGE_SIZE, (1UL << cpu_vabits)) - PMD_SIZE - VMEMMAP_SIZE - KFENCE_AREA_SIZE) 100 min(PTRS_PER_PGD * PTRS_PER_PUD * PTRS_PER_PMD * PTRS_PER_PTE * PAGE_SIZE, (1UL << cpu_vabits) / 2) - PMD_SIZE - VMEMMAP_SIZE - KFENCE_AREA_SIZE) 122 extern pte_t invalid_pte_table[PTRS_PER_PTE];
|
/linux-master/arch/sparc/include/asm/ |
H A D | pgtable_32.h | 47 #define PTRS_PER_PTE 64 macro 51 #define PTE_SIZE (PTRS_PER_PTE*4)
|
/linux-master/arch/loongarch/kvm/ |
H A D | mmu.c | 128 end = p + PTRS_PER_PTE; 767 for (i = 0; i < PTRS_PER_PTE; i++) { 776 kvm->stat.pages += PTRS_PER_PTE; 778 return child + (gfn & (PTRS_PER_PTE - 1)); 898 gfn = gfn & ~(PTRS_PER_PTE - 1); 899 pfn = pfn & ~(PTRS_PER_PTE - 1);
|
/linux-master/arch/parisc/mm/ |
H A D | init.c | 46 pte_t pg0[PT_INITIAL * PTRS_PER_PTE] __section(".data..vm0.pte") __attribute__ ((aligned(PAGE_SIZE))); 368 start_pte = ((start_vaddr >> PAGE_SHIFT) & (PTRS_PER_PTE - 1)); 397 for (tmp2 = start_pte; tmp2 < PTRS_PER_PTE; tmp2++, pg_table++) {
|
/linux-master/arch/riscv/kvm/ |
H A D | mmu.c | 48 mask = (PTRS_PER_PTE * (1UL << gstage_pgd_xbits)) - 1; 50 mask = PTRS_PER_PTE - 1; 255 for (i = 0; i < PTRS_PER_PTE; i++)
|
/linux-master/arch/riscv/mm/ |
H A D | kasan_init.c | 35 p = memblock_alloc(PTRS_PER_PTE * sizeof(pte_t), PAGE_SIZE); 329 for (i = 0; i < PTRS_PER_PTE; ++i) 525 for (i = 0; i < PTRS_PER_PTE; i++)
|
/linux-master/arch/s390/mm/ |
H A D | vmem.c | 66 unsigned long size = PTRS_PER_PTE * sizeof(pte_t); 75 memset64((u64 *)pte, _PAGE_INVALID, PTRS_PER_PTE); 214 for (i = 0; i < PTRS_PER_PTE; i++, pte++) {
|
/linux-master/arch/m68k/mm/ |
H A D | motorola.c | 241 last_pte_table += PTRS_PER_PTE; 354 for (i = 1; i < PTRS_PER_PTE; physaddr += PAGE_SIZE, i++)
|
/linux-master/arch/x86/mm/ |
H A D | kasan_init_64.c | 303 for (i = 0; i < PTRS_PER_PTE; i++) 441 for (i = 0; i < PTRS_PER_PTE; i++) {
|
/linux-master/arch/s390/boot/ |
H A D | vmem.c | 75 memset64((u64 *)kasan_early_shadow_pte, pte_val(pte_z), PTRS_PER_PTE); 234 memset64((u64 *)pte, _PAGE_INVALID, PTRS_PER_PTE);
|
/linux-master/arch/loongarch/mm/ |
H A D | init.c | 249 pte_t invalid_pte_table[PTRS_PER_PTE] __page_aligned_bss;
|
H A D | kasan_init.c | 286 for (i = 0; i < PTRS_PER_PTE; i++)
|
/linux-master/arch/arm/mm/ |
H A D | kasan_init.c | 292 for (i = 0; i < PTRS_PER_PTE; i++)
|
/linux-master/arch/arm64/include/asm/ |
H A D | pgtable-hwdef.h | 43 #define PTRS_PER_PTE (1 << (PAGE_SHIFT - 3)) macro
|
H A D | tlbflush.h | 354 #define MAX_DVM_OPS PTRS_PER_PTE
|
/linux-master/arch/nios2/include/asm/ |
H A D | pgtable.h | 56 #define PTRS_PER_PTE (PAGE_SIZE / sizeof(pte_t)) macro
|
/linux-master/arch/x86/xen/ |
H A D | p2m.c | 92 #define PMDS_PER_MID_PAGE (P2M_MID_PER_PAGE / PTRS_PER_PTE) 344 for (i = 0; i < PTRS_PER_PTE; i++) {
|
/linux-master/arch/openrisc/include/asm/ |
H A D | pgtable.h | 66 #define PTRS_PER_PTE (1UL << (PAGE_SHIFT-2)) macro
|
/linux-master/arch/xtensa/include/asm/ |
H A D | pgtable.h | 43 * One page (4 kB) of 1024 (PTRS_PER_PTE) PTEs with a special PTE 57 #define PTRS_PER_PTE 1024 macro
|
/linux-master/arch/alpha/include/asm/ |
H A D | pgtable.h | 44 #define PTRS_PER_PTE (1UL << (PAGE_SHIFT-3)) macro
|
/linux-master/arch/microblaze/include/asm/ |
H A D | pgtable.h | 97 #define PTRS_PER_PTE (1 << PTE_SHIFT) macro
|