/haiku/src/system/kernel/arch/x86/paging/64bit/ |
H A D | paging.h | 87 #define VADDR_TO_PML5E(va) (((va) & 0x01fffffffffff000L) / k64BitPML4TRange) 88 #define VADDR_TO_PML4E(va) (((va) % k64BitPML4TRange) / k64BitPDPTRange) 89 #define VADDR_TO_PDPTE(va) (((va) % k64BitPDPTRange) / k64BitPageDirectoryRange) 90 #define VADDR_TO_PDE(va) (((va) % k64BitPageDirectoryRange) / k64BitPageTableRange) 91 #define VADDR_TO_PTE(va) (((va) [all...] |
/haiku/headers/private/kernel/arch/arm64/ |
H A D | arch_pte.h | 120 #define pmap_l0_index(va) (((va) >> L0_SHIFT) & L0_ADDR_MASK) 121 #define pmap_l1_index(va) (((va) >> L1_SHIFT) & Ln_ADDR_MASK) 122 #define pmap_l2_index(va) (((va) >> L2_SHIFT) & Ln_ADDR_MASK) 123 #define pmap_l3_index(va) (((va) >> L3_SHIFT) & Ln_ADDR_MASK)
|
/haiku/src/system/kernel/arch/generic/ |
H A D | generic_vm_physical_page_mapper.cpp | 35 addr_t va; member in struct:paddr_chunk_descriptor 57 generic_get_physical_page(phys_addr_t pa, addr_t *va, uint32 flags) argument 67 if (paddr_desc[index].va != 0) { 72 *va = paddr_desc[index].va + pa % sIOSpaceChunkSize; 80 paddr_desc[index].va = first_free_vmapping * sIOSpaceChunkSize 82 *va = paddr_desc[index].va + pa % sIOSpaceChunkSize; 93 sMapIOSpaceChunk(paddr_desc[index].va, index * sIOSpaceChunkSize, 117 paddr_desc[index].va 135 generic_put_physical_page(addr_t va) argument [all...] |
H A D | generic_vm_physical_page_mapper.h | 24 status_t generic_get_physical_page(phys_addr_t pa, addr_t *va, uint32 flags); 25 status_t generic_put_physical_page(addr_t va);
|
/haiku/headers/private/kernel/arch/arm/ |
H A D | arm_mmu.h | 92 #define VADDR_TO_PDENT(va) ((va) >> 20) 93 #define VADDR_TO_PTENT(va) (((va) & 0xff000) >> 12) 94 #define VADDR_TO_PGOFF(va) ((va) & 0x0fff)
|
/haiku/src/system/kernel/arch/x86/paging/32bit/ |
H A D | paging.h | 19 #define VADDR_TO_PDENT(va) (((va) / B_PAGE_SIZE) / 1024) 20 #define VADDR_TO_PTENT(va) (((va) / B_PAGE_SIZE) % 1024)
|
H A D | X86VMTranslationMap32Bit.cpp | 140 X86VMTranslationMap32Bit::Map(addr_t va, phys_addr_t pa, uint32 attributes, argument 143 TRACE("map_tmap: entry pa 0x%lx va 0x%lx\n", pa, va); 147 dprintf("index is %d\n", va / B_PAGE_SIZE / 1024); 148 dprintf("final at 0x%x\n", &pgdir[va / B_PAGE_SIZE / 1024]); 149 dprintf("value is 0x%x\n", *(int *)&pgdir[va / B_PAGE_SIZE / 1024]); 150 dprintf("present bit is %d\n", pgdir[va / B_PAGE_SIZE / 1024].present); 151 dprintf("addr is %d\n", pgdir[va / B_PAGE_SIZE / 1024].addr); 156 uint32 index = VADDR_TO_PDENT(va); 192 index = VADDR_TO_PTENT(va); 610 Query(addr_t va, phys_addr_t *_physical, uint32 *_flags) argument 654 QueryInterrupt(addr_t va, phys_addr_t *_physical, uint32 *_flags) argument 769 ClearFlags(addr_t va, uint32 flags) argument [all...] |
/haiku/headers/libs/agg/ |
H A D | agg_span_gouraud_gray.h | 170 int vv, va; local 182 va = a.y(); 184 if(va < 0) va = 0; if(va > lim) va = lim; 186 span->a = (value_type)va; 217 va = a.y(); 219 if(va < 0) va [all...] |
H A D | agg_span_gouraud_rgba.h | 186 int vr, vg, vb, va; local 200 va = a.y(); 204 if(va < 0) va = 0; if(va > lim) va = lim; 208 span->a = (value_type)va; 247 va = a.y(); 251 if(va < 0) va [all...] |
/haiku/src/system/kernel/arch/m68k/paging/040/ |
H A D | paging.h | 75 #define VADDR_TO_PDENT(va) (((va) / B_PAGE_SIZE) / 1024) 76 #define VADDR_TO_PTENT(va) (((va) / B_PAGE_SIZE) % 1024)
|
/haiku/src/system/kernel/arch/arm64/ |
H A D | VMSAv8TranslationMap.cpp | 207 VMSAv8TranslationMap::MapRange(phys_addr_t ptPa, int level, addr_t va, phys_addr_t pa, size_t size, 221 uint64_t nextVa = va; 222 uint64_t end = va + size; 226 if ((va & entryMask) != 0) { 227 uint64_t aligned = (va & ~entryMask) + entrySize; 229 index = (va >> shift) & tableMask; 231 MapRange(table, level + 1, va, pa, aligned - va, action, attr, reservation); 238 phys_addr_t targetPa = pa + (nextVa - va); 291 table, level + 1, nextVa, pa + (nextVa - va), en [all...] |
H A D | arch_vm_translation_map.cpp | 127 map_page_early(phys_addr_t ptPa, int level, addr_t va, phys_addr_t pa, argument 135 int index = (va >> shift) & tableMask; 168 map_page_early(table, level + 1, va, pa, get_free_page, args); 174 arch_vm_translation_map_early_map(kernel_args* args, addr_t va, phys_addr_t pa, uint8 attributes, argument 179 ASSERT((va & ~va_mask) == ~va_mask); 183 va &= va_mask; 186 map_page_early(ptPa, level, va, pa, get_free_page, args); 193 arch_vm_translation_map_is_kernel_page_accessible(addr_t va, uint32 protection) argument 196 asm("at s1e1w, %0" : : "r"((uint64_t) va)); 199 asm("at s1e1r, %0" : : "r"((uint64_t) va)); [all...] |
H A D | VMSAv8TranslationMap.h | 90 bool ValidateVa(addr_t va); 95 void MapRange(phys_addr_t ptPa, int level, addr_t va, phys_addr_t pa, size_t size, 97 bool WalkTable(phys_addr_t ptPa, int level, addr_t va, phys_addr_t* pa, uint64_t* attr);
|
/haiku/src/system/kernel/arch/m68k/ |
H A D | arch_vm_translation_map_impl.cpp | 121 static status_t early_query(addr_t va, addr_t *out_physical); 122 static status_t get_physical_page_tmap_internal(addr_t pa, addr_t *va, uint32 flags); 123 static status_t put_physical_page_tmap_internal(addr_t va); 221 early_query(addr_t va, addr_t *_physicalAddress) argument 230 TRACE(("%s(%p,)\n", __FUNCTION__, va)); 232 index = VADDR_TO_PRENT(va); 236 // pa == va when in TT 240 index = VADDR_TO_PDENT(va); 247 index = VADDR_TO_PTENT(va); 259 // we should only be passed page va, bu 517 map_tmap(vm_translation_map *map, addr_t va, addr_t pa, uint32 attributes) argument 722 query_tmap_interrupt(vm_translation_map *map, addr_t va, addr_t *_physical, uint32 *_flags) argument 778 query_tmap(vm_translation_map *map, addr_t va, addr_t *_physical, uint32 *_flags) argument 948 clear_flags_tmap(vm_translation_map *map, addr_t va, uint32 flags) argument 1064 map_iospace_chunk(addr_t va, addr_t pa, uint32 flags) argument 1101 get_physical_page_tmap_internal(addr_t pa, addr_t *va, uint32 flags) argument 1108 put_physical_page_tmap_internal(addr_t va) argument 1445 m68k_vm_translation_map_early_map(kernel_args *args, addr_t va, addr_t pa, uint8 attributes, addr_t (*get_free_page)(kernel_args *)) argument [all...] |
H A D | arch_vm_translation_map.cpp | 146 arch_vm_translation_map_early_map(kernel_args *args, addr_t va, phys_addr_t pa, argument 149 TRACE("early_tmap: entry pa 0x%lx va 0x%lx\n", pa, va); 151 return gM68KPagingMethod->MapEarly(args, va, pa, attributes, get_free_page); 158 arch_vm_translation_map_early_query(addr_t va, addr_t *out_physical) 160 return get_vm_ops()->arch_vm_translation_map_early_query(va, out_physical);
|
/haiku/headers/private/kernel/arch/m68k/ |
H A D | arch_040_mmu.h | 183 #define VADDR_TO_PRENT(va) (((va) / B_PAGE_SIZE) / (64*128)) 184 #define VADDR_TO_PDENT(va) ((((va) / B_PAGE_SIZE) / 64) % 128) 185 #define VADDR_TO_PTENT(va) (((va) / B_PAGE_SIZE) % 64)
|
H A D | arch_030_mmu.h | 198 #define VADDR_TO_PRENT(va) (((va) / B_PAGE_SIZE) / (64*128)) 199 #define VADDR_TO_PDENT(va) ((((va) / B_PAGE_SIZE) / 64) % 128) 200 #define VADDR_TO_PTENT(va) (((va) / B_PAGE_SIZE) % 64)
|
H A D | stage2_priv.h | 28 void mmu_map_page(unsigned int vsid, unsigned long pa, unsigned long va);;
|
/haiku/src/system/kernel/arch/arm/ |
H A D | arch_vm_translation_map.cpp | 103 arch_vm_translation_map_early_map(kernel_args *args, addr_t va, phys_addr_t pa, argument 106 TRACE("early_tmap: entry pa 0x%lx va 0x%lx\n", pa, va); 108 return gARMPagingMethod->MapEarly(args, va, pa, attributes, get_free_page);
|
/haiku/headers/private/kernel/arch/ |
H A D | vm_translation_map.h | 29 status_t arch_vm_translation_map_early_map(struct kernel_args *args, addr_t va,
|
/haiku/src/system/kernel/arch/arm/paging/32bit/ |
H A D | ARMVMTranslationMap32Bit.cpp | 157 ARMVMTranslationMap32Bit::Map(addr_t va, phys_addr_t pa, uint32 attributes, argument 160 TRACE("map_tmap: entry pa 0x%lx va 0x%lx\n", pa, va); 164 dprintf("index is %d\n", va / B_PAGE_SIZE / 1024); 165 dprintf("final at 0x%x\n", &pgdir[va / B_PAGE_SIZE / 1024]); 166 dprintf("value is 0x%x\n", *(int *)&pgdir[va / B_PAGE_SIZE / 1024]); 167 dprintf("present bit is %d\n", pgdir[va / B_PAGE_SIZE / 1024].present); 168 dprintf("addr is %d\n", pgdir[va / B_PAGE_SIZE / 1024].addr); 173 uint32 index = VADDR_TO_PDENT(va); 190 (va < KERNEL_BAS 627 Query(addr_t va, phys_addr_t *_physical, uint32 *_flags) argument 664 QueryInterrupt(addr_t va, phys_addr_t *_physical, uint32 *_flags) argument 794 ClearFlags(addr_t va, uint32 flags) argument [all...] |
/haiku/headers/private/kernel/arch/ppc/ |
H A D | stage2_priv.h | 28 void mmu_map_page(unsigned int vsid, unsigned long pa, unsigned long va);;
|
/haiku/src/libs/compat/freebsd_network/compat/machine/x86_64/ |
H A D | cpufunc.h | 44 #define readb(va) (*(volatile uint8_t *) (va)) 45 #define readw(va) (*(volatile uint16_t *) (va)) 46 #define readl(va) (*(volatile uint32_t *) (va)) 47 #define readq(va) (*(volatile uint64_t *) (va)) 49 #define writeb(va, d) (*(volatile uint8_t *) (va) [all...] |
/haiku/src/system/kernel/arch/sparc/ |
H A D | arch_vm_translation_map.cpp | 79 arch_vm_translation_map_early_map(kernel_args *args, addr_t va, phys_addr_t pa, argument 82 TRACE("early_tmap: entry pa 0x%lx va 0x%lx\n", pa, va);
|
/haiku/headers/private/kernel/arch/x86/ |
H A D | arch_cpuasm.h | 61 #define invalidate_TLB(va) \ 62 __asm__("invlpg (%0)" : : "r" (va))
|