Lines Matching refs:pa
122 static status_t get_physical_page_tmap_internal(addr_t pa, addr_t *va, uint32 flags);
227 addr_t pa;
235 pa = PRE_TO_TA(pr[index]);
236 // pa == va when in TT
238 pd = (page_directory_entry *)pa;
244 pa = PDE_TO_TA(pd[index]);
245 pt = (page_table_entry *)pa;
252 pa = PIE_TO_TA(pi[index]);
253 pt = (page_table_entry *)pa;
517 map_tmap(vm_translation_map *map, addr_t va, addr_t pa, uint32 attributes)
526 TRACE(("map_tmap: entry pa 0x%lx va 0x%lx\n", pa, va));
628 put_page_table_entry_in_pgtable(&pt[pindex], pa, attributes,
854 TRACE(("query_tmap: returning pa 0x%lx for va 0x%lx\n", *_physical, va));
1064 map_iospace_chunk(addr_t va, addr_t pa, uint32 flags)
1070 pa &= ~(B_PAGE_SIZE - 1); // make sure it's page aligned
1076 for (i = 0; i < NUM_PAGEENT_PER_TBL; i++, pa += B_PAGE_SIZE) {
1078 pt[i].addr = TA_TO_PTEA(pa);
1101 get_physical_page_tmap_internal(addr_t pa, addr_t *va, uint32 flags)
1103 return generic_get_physical_page(pa, va, flags);
1285 // we also still have transparent translation enabled, va==pa.
1379 // first get pa for the indirect descriptor
1395 // pa of the page
1445 m68k_vm_translation_map_early_map(kernel_args *args, addr_t va, addr_t pa,
1454 TRACE(("early_tmap: entry pa 0x%lx va 0x%lx\n", pa, va));
1456 // everything much simpler here because pa = va
1507 put_page_table_entry_in_pgtable(&pt[index], pa, attributes,