Searched refs:__pa (Results 51 - 75 of 415) sorted by relevance

1234567891011>>

/linux-master/arch/arm/include/asm/
H A Dkfence.h14 unsigned long pfn = PFN_DOWN(__pa(addr));
/linux-master/arch/s390/kernel/
H A Dabs_lowcore.c11 unsigned long phys = __pa(lc);
/linux-master/drivers/pci/controller/
H A Dpci-versatile.c94 writel(__pa(PAGE_OFFSET) >> 28, PCI_SMAP(mem));
131 writel(__pa(PAGE_OFFSET), local_pci_cfg_base + PCI_BASE_ADDRESS_0);
132 writel(__pa(PAGE_OFFSET), local_pci_cfg_base + PCI_BASE_ADDRESS_1);
133 writel(__pa(PAGE_OFFSET), local_pci_cfg_base + PCI_BASE_ADDRESS_2);
/linux-master/arch/powerpc/include/asm/
H A Drtas-work-area.h82 return __pa(area->buf);
H A Dpage.h143 * existing equation for the __va()/__pa() translations:
177 * To make the cost of __va() / __pa() more light weight, we introduce
190 * __pa(x) = x + PHYSICAL_START - Effective KERNELBASE
194 * the other definitions for __va & __pa.
198 #define __pa(x) ((phys_addr_t)(unsigned long)(x) - VIRT_PHYS_OFFSET) macro
215 #define __pa(x) \ macro
223 #define __pa(x) ((unsigned long)(x) - PAGE_OFFSET + MEMORY_START) macro
230 return __pa(kaddr) >> PAGE_SHIFT;
/linux-master/arch/sparc/include/asm/
H A Dpage_64.h146 #define __pa(x) ((unsigned long)(x) - PAGE_OFFSET) macro
151 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr)>>PAGE_SHIFT)
153 #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
155 #define virt_to_phys __pa
/linux-master/arch/xtensa/include/asm/
H A Dpage.h187 #define __pa(x) ___pa((unsigned long)(x)) macro
189 #define __pa(x) \ macro
195 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
197 #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
/linux-master/arch/arm64/include/asm/
H A Dpgalloc.h34 __pud_populate(pudp, __pa(pmdp), pudval);
56 __p4d_populate(p4dp, __pa(pudp), p4dval);
85 __pgd_populate(pgdp, __pa(p4dp), pgdval);
130 __pmd_populate(pmdp, __pa(ptep), PMD_TYPE_TABLE | PMD_TABLE_UXN);
/linux-master/arch/s390/include/asm/
H A Dpgalloc.h105 set_pgd(pgd, __pgd(_REGION1_ENTRY | __pa(p4d)));
110 set_p4d(p4d, __p4d(_REGION2_ENTRY | __pa(pud)));
115 set_pud(pud, __pud(_REGION3_ENTRY | __pa(pmd)));
131 set_pmd(pmd, __pmd(_SEGMENT_ENTRY | __pa(pte)));
/linux-master/arch/x86/mm/
H A Dident_map.c55 set_pud(pud, __pud(__pa(pmd) | info->kernpg_flag));
91 set_p4d(p4d, __p4d(__pa(pud) | info->kernpg_flag));
135 set_pgd(pgd, __pgd(__pa(p4d) | info->kernpg_flag));
142 set_pgd(pgd, __pgd(__pa(pud) | info->kernpg_flag));
/linux-master/arch/powerpc/platforms/powernv/
H A Dopal-nvram.c36 rc = opal_read_nvram(__pa(buf), count, off);
59 rc = opal_write_nvram(__pa(buf), count, off);
/linux-master/arch/arm/mach-zynq/
H A Dplatsmp.c46 if (__pa(PAGE_OFFSET)) {
70 if (__pa(PAGE_OFFSET))
/linux-master/arch/powerpc/mm/kasan/
H A D8xx.c27 *new = pte_val(pte_mkhuge(pfn_pte(PHYS_PFN(__pa(block)), PAGE_KERNEL)));
63 pte_t pte = pfn_pte(PHYS_PFN(__pa(va)), PAGE_KERNEL);
H A Dinit_32.c24 phys_addr_t pa = __pa(kasan_early_shadow_page);
75 pte_t pte = pfn_pte(PHYS_PFN(__pa(va)), PAGE_KERNEL);
104 phys_addr_t pa = __pa(kasan_early_shadow_page);
/linux-master/arch/mips/dec/prom/
H A Dmemory.c111 end = __pa(&_text) - 0x00020000;
114 end = __pa(&_text);
/linux-master/arch/parisc/include/asm/
H A Dmmu_context.h58 mtctl(__pa(__ldcw_align(&pgd_lock->rlock.raw_lock)), 28);
60 mtctl(__pa(next->pgd), 25);
/linux-master/arch/openrisc/kernel/
H A Ddma.c39 for (cl = __pa(addr); cl < __pa(next); cl += cpuinfo->dcache_block_size)
/linux-master/arch/powerpc/platforms/chrp/
H A Dnvram.c35 __pa(nvram_buf), 1) != 0) || 1 != done)
57 __pa(nvram_buf), 1) != 0) || 1 != done)
/linux-master/arch/powerpc/include/asm/nohash/32/
H A Dhugetlb-8xx.h35 *hpdp = __hugepd(__pa(new) | _PMD_USER | _PMD_PRESENT | _PMD_PAGE_8M);
40 *hpdp = __hugepd(__pa(new) | _PMD_PRESENT | _PMD_PAGE_8M);
/linux-master/arch/arc/include/asm/
H A Dpage.h107 * __pa, __va, virt_to_page (ALERT: deprecated, don't use them)
113 #define __pa(vaddr) ((unsigned long)(vaddr)) macro
125 return __pa(kaddr) >> PAGE_SHIFT;
/linux-master/drivers/i2c/busses/
H A Di2c-opal.c90 req.buffer_ra = cpu_to_be64(__pa(msgs[0].buf));
101 req.buffer_ra = cpu_to_be64(__pa(msgs[1].buf));
126 req.buffer_ra = cpu_to_be64(__pa(&data->byte));
134 req.buffer_ra = cpu_to_be64(__pa(&data->byte));
146 req.buffer_ra = cpu_to_be64(__pa(local));
154 req.buffer_ra = cpu_to_be64(__pa(&data->block[1]));
/linux-master/arch/s390/mm/
H A Dpgalloc.c111 mm->context.asce = __pa(mm->pgd) | _ASCE_TABLE_LENGTH |
120 mm->context.asce = __pa(mm->pgd) | _ASCE_TABLE_LENGTH |
325 *ste = __pa(table) | _SEGMENT_ENTRY;
354 *rtte = __pa(table) | _REGION3_ENTRY;
382 *rste = __pa(table) | _REGION2_ENTRY;
410 *rfte = __pa(table) | _REGION1_ENTRY;
494 asce = __pa(table) | _ASCE_TYPE_SEGMENT | _ASCE_TABLE_LENGTH;
500 asce = __pa(table) | _ASCE_TYPE_REGION3 | _ASCE_TABLE_LENGTH;
506 asce = __pa(table) | _ASCE_TYPE_REGION2 | _ASCE_TABLE_LENGTH;
512 asce = __pa(tabl
[all...]
/linux-master/arch/riscv/mm/
H A Dinit.c232 * In 64-bit, any use of __va/__pa before this point is wrong as we
248 max_mapped_addr = __pa(~(ulong)0);
272 * early_init_fdt_reserve_self() since __pa() does
370 return __pa((pte_t *)ptdesc_address(ptdesc));
451 return __pa((pmd_t *)ptdesc_address(ptdesc));
516 return __pa(vaddr);
554 return __pa(vaddr);
615 #define get_pgd_next_virt(__pa) (pgtable_l5_enabled ? \
616 pt_ops.get_p4d_virt(__pa) : (pgd_next_t *)(pgtable_l4_enabled ? \
617 pt_ops.get_pud_virt(__pa)
[all...]
/linux-master/drivers/crypto/cavium/zip/
H A Dzip_inflate.c94 zip_cmd->s.out_ptr_addr.s.addr = __pa(zip_ops->output);
104 zip_cmd->s.inp_ptr_addr.s.addr = __pa((u8 *)zip_ops->input);
108 zip_cmd->s.res_ptr_addr.s.addr = __pa(result_ptr);
H A Dzip_deflate.c85 zip_cmd->s.inp_ptr_addr.s.addr = __pa(zip_ops->input);
91 zip_cmd->s.out_ptr_addr.s.addr = __pa(zip_ops->output);
97 zip_cmd->s.res_ptr_addr.s.addr = __pa(result_ptr);

Completed in 216 milliseconds

1234567891011>>