/linux-master/drivers/xen/ |
H A D | mem-reservation.c | 51 (unsigned long)__va(pfn << PAGE_SHIFT), 75 (unsigned long)__va(pfn << PAGE_SHIFT),
|
/linux-master/arch/csky/include/asm/ |
H A D | page.h | 76 #define __va(x) ((void *)((unsigned long)(x) + PAGE_OFFSET - va_pa_offset)) macro 89 #define pfn_to_kaddr(x) __va(PFN_PHYS(x))
|
/linux-master/arch/hexagon/include/asm/ |
H A D | page.h | 81 * We need a __pa and a __va routine for kernel space. 86 #define __va(x) ((void *)((unsigned long)(x) - PHYS_OFFSET + PAGE_OFFSET)) macro 132 #define page_to_virt(page) __va(page_to_phys(page))
|
/linux-master/arch/x86/mm/ |
H A D | mem_encrypt_amd.c | 67 early_snp_set_memory_shared((unsigned long)__va(paddr), paddr, npages); 72 early_snp_set_memory_private((unsigned long)__va(paddr), paddr, npages); 194 __sme_early_map_unmap_mem(__va(cmdline_paddr), COMMAND_LINE_SIZE, false); 214 __sme_early_map_unmap_mem(__va(cmdline_paddr), COMMAND_LINE_SIZE, true); 342 clflush_cache_range(__va(pa), size); 354 early_snp_set_memory_shared((unsigned long)__va(pa), pa, 1); 366 early_snp_set_memory_private((unsigned long)__va(pa), pa, 1);
|
/linux-master/arch/m68k/include/asm/ |
H A D | motorola_pgtable.h | 106 #define __pte_page(pte) ((unsigned long)__va(pte_val(pte) & PAGE_MASK)) 107 #define pmd_page_vaddr(pmd) ((unsigned long)__va(pmd_val(pmd) & _TABLE_MASK)) 108 #define pud_pgtable(pud) ((pmd_t *)__va(pud_val(pud) & _TABLE_MASK)) 116 #define pte_page(pte) virt_to_page(__va(pte_val(pte))) 138 #define pud_page(pud) (mem_map + ((unsigned long)(__va(pud_val(pud)) - PAGE_OFFSET) >> PAGE_SHIFT))
|
/linux-master/arch/x86/realmode/ |
H A D | init.c | 137 __va(real_mode_header->trampoline_header); 160 trampoline_pgd = (u64 *) __va(real_mode_header->trampoline_pgd); 199 (unsigned long) __va(real_mode_header->text_start);
|
/linux-master/arch/sparc/include/asm/ |
H A D | page_64.h | 147 #define __va(x) ((void *)((unsigned long) (x) + PAGE_OFFSET)) macro 149 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) 156 #define phys_to_virt __va
|
H A D | pgtsrmmu.h | 107 #define __nocache_va(PADDR) (__va((unsigned long)PADDR) - (unsigned long)srmmu_nocache_pool + SRMMU_NOCACHE_VADDR) 108 #define __nocache_fix(VADDR) ((__typeof__(VADDR))__va(__nocache_pa(VADDR)))
|
H A D | page_32.h | 125 #define __va(x) ((void *)((unsigned long) (x) - phys_base + PAGE_OFFSET)) macro 128 #define phys_to_virt __va
|
/linux-master/arch/loongarch/mm/ |
H A D | kasan_init.c | 112 memcpy(__va(pte_phys), kasan_early_shadow_pte, sizeof(kasan_early_shadow_pte)); 113 pmd_populate_kernel(NULL, pmdp, (pte_t *)__va(pte_phys)); 125 memcpy(__va(pmd_phys), kasan_early_shadow_pmd, sizeof(kasan_early_shadow_pmd)); 126 pud_populate(&init_mm, pudp, (pmd_t *)__va(pmd_phys)); 138 memcpy(__va(pud_phys), kasan_early_shadow_pud, sizeof(kasan_early_shadow_pud)); 139 p4d_populate(&init_mm, p4dp, (pud_t *)__va(pud_phys));
|
/linux-master/arch/x86/kernel/ |
H A D | check.c | 129 memset(__va(start), 0, end - start); 149 unsigned long *addr = __va(scan_areas[i].addr);
|
/linux-master/arch/x86/platform/pvh/ |
H A D | enlighten.c | 52 ep = __va(pvh_start_info.memmap_paddr); 87 __va(pvh_start_info.modlist_paddr);
|
/linux-master/arch/riscv/mm/ |
H A D | init.c | 246 * In 64-bit, any use of __va/__pa before this point is wrong as we 269 high_memory = (void *)(__va(PFN_PHYS(max_low_pfn))); 362 return (pte_t *) __va(pa); 445 return (pmd_t *) __va(pa); 508 return (pud_t *)__va(pa); 546 return (p4d_t *)__va(pa); 626 #define alloc_pgd_next(__va) (pgtable_l5_enabled ? \ 627 pt_ops.alloc_p4d(__va) : (pgtable_l4_enabled ? \ 628 pt_ops.alloc_pud(__va) : pt_ops.alloc_pmd(__va))) [all...] |
/linux-master/arch/x86/include/asm/ |
H A D | realmode.h | 91 real_mode_header = (struct real_mode_header *) __va(mem);
|
/linux-master/arch/arm/common/ |
H A D | sharpsl_param.c | 28 #define param_start(x) __va(x)
|
/linux-master/arch/openrisc/include/asm/ |
H A D | page.h | 72 #define __va(x) ((void *)((unsigned long)(x) + PAGE_OFFSET)) macro
|
/linux-master/arch/x86/include/asm/uv/ |
H A D | uv_hub.h | 379 * Note: use the standard __pa() & __va() macros for converting 538 return __va(((unsigned long)pnode << m_val) | offset); 544 return __va((unsigned long)offset); 547 return __va(base << UV_GAM_RANGE_SHFT | offset); 565 return __va(UV_GLOBAL_MMR32_BASE | 585 return __va(UV_GLOBAL_MMR64_BASE | 615 return __va(UV_LOCAL_MMR_BASE | offset);
|
/linux-master/arch/mips/loongson64/ |
H A D | numa.c | 105 nd = __va(nd_pa); 183 high_memory = (void *) __va(get_num_physpages() << PAGE_SHIFT);
|
/linux-master/arch/mips/include/asm/ |
H A D | page.h | 155 * __pa()/__va() should be used only during mem init. 185 #define __va(x) ((void *)((unsigned long)(x) + PAGE_OFFSET - PHYS_OFFSET)) macro 211 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT)
|
/linux-master/arch/xtensa/include/asm/ |
H A D | page.h | 192 #define __va(x) \ macro 196 #define page_to_virt(page) __va(page_to_pfn(page) << PAGE_SHIFT)
|
/linux-master/arch/sh/include/asm/ |
H A D | page.h | 133 #define __va(x) (void *)___va((unsigned long)x) macro 149 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT)
|
/linux-master/arch/riscv/include/asm/ |
H A D | page.h | 173 #define __va(x) ((void *)__pa_to_va_nodebug((phys_addr_t)(x))) macro 179 #define pfn_to_virt(pfn) (__va(pfn_to_phys(pfn)))
|
/linux-master/arch/s390/include/asm/ |
H A D | physmem_info.h | 137 range && range->end; range = range->chain ? __va(range->chain) : NULL, \ 149 return __va(range->chain);
|
/linux-master/arch/csky/mm/ |
H A D | dma-mapping.c | 18 void *start = __va(page_to_phys(page));
|
/linux-master/include/xen/arm/ |
H A D | page.h | 84 #define gfn_to_virt(m) (__va(gfn_to_pfn(m) << XEN_PAGE_SHIFT))
|