/fuchsia/zircon/kernel/vm/include/vm/ |
H A D | kstack.h | 17 vaddr_t base; 19 vaddr_t top; 27 vaddr_t unsafe_base;
|
H A D | arch_vm_aspace.h | 37 virtual zx_status_t Init(vaddr_t base, size_t size, uint mmu_flags) = 0; 44 virtual zx_status_t MapContiguous(vaddr_t vaddr, paddr_t paddr, size_t count, 50 virtual zx_status_t Map(vaddr_t vaddr, paddr_t* phys, size_t count, 54 virtual zx_status_t Unmap(vaddr_t vaddr, size_t count, size_t* unmapped) = 0; 57 virtual zx_status_t Protect(vaddr_t vaddr, size_t count, uint mmu_flags) = 0; 59 virtual zx_status_t Query(vaddr_t vaddr, paddr_t* paddr, uint* mmu_flags) = 0; 61 virtual vaddr_t PickSpot(vaddr_t base, uint prev_region_mmu_flags, 62 vaddr_t end, uint next_region_mmu_flags, 63 vaddr_t alig [all...] |
H A D | vm.h | 25 static inline bool is_kernel_address(vaddr_t va) { 26 return (va >= (vaddr_t)KERNEL_ASPACE_BASE && 27 va - (vaddr_t)KERNEL_ASPACE_BASE < (vaddr_t)KERNEL_ASPACE_SIZE); 33 static inline bool is_user_address(vaddr_t va) { 37 static inline bool is_user_address_range(vaddr_t va, size_t len) {
|
H A D | vm_address_region.h | 105 vaddr_t base() const { return base_; } 120 virtual zx_status_t PageFault(vaddr_t va, uint pf_flags) = 0; 123 vaddr_t GetKey() const { return base(); } 149 VmAddressRegionOrMapping(vaddr_t base, size_t size, uint32_t flags, 156 bool is_in_range(vaddr_t base, size_t size) const { 180 vaddr_t base_; 223 virtual fbl::RefPtr<VmAddressRegionOrMapping> FindRegion(vaddr_t addr); 229 virtual zx_status_t Unmap(vaddr_t base, size_t size); 234 zx_status_t UnmapAllowPartial(vaddr_t base, size_t size); 239 virtual zx_status_t Protect(vaddr_t bas [all...] |
H A D | vm_aspace.h | 60 vaddr_t base() const { return base_; } 125 fbl::RefPtr<VmAddressRegionOrMapping> FindRegion(vaddr_t va); 135 zx_status_t ReserveSpace(const char* name, size_t size, vaddr_t vaddr); 143 zx_status_t FreeRegion(vaddr_t va); 170 VmAspace(vaddr_t base, size_t size, uint32_t flags, const char* name); 184 zx_status_t PageFault(vaddr_t va, uint flags); 185 friend zx_status_t vmm_page_fault_handler(vaddr_t va, uint flags); 192 vaddr_t base_;
|
/fuchsia/zircon/kernel/arch/arm64/include/arch/ |
H A D | aspace.h | 22 zx_status_t Init(vaddr_t base, size_t size, uint mmu_flags) override; 27 zx_status_t Map(vaddr_t vaddr, paddr_t* phys, size_t count, uint mmu_flags, 29 zx_status_t MapContiguous(vaddr_t vaddr, paddr_t paddr, size_t count, 32 zx_status_t Unmap(vaddr_t vaddr, size_t count, size_t* unmapped) override; 34 zx_status_t Protect(vaddr_t vaddr, size_t count, uint mmu_flags) override; 36 zx_status_t Query(vaddr_t vaddr, paddr_t* paddr, uint* mmu_flags) override; 38 vaddr_t PickSpot(vaddr_t base, uint prev_region_mmu_flags, 39 vaddr_t end, uint next_region_mmu_flags, 40 vaddr_t alig [all...] |
H A D | arch_thread.h | 36 vaddr_t unsafe_sp; 39 vaddr_t sp;
|
H A D | arm64.h | 53 void arm64_context_switch(vaddr_t* old_sp, vaddr_t new_sp); 56 vaddr_t kstack, uint32_t spsr,
|
/fuchsia/zircon/kernel/arch/x86/include/arch/ |
H A D | arch_thread.h | 19 vaddr_t sp; 21 vaddr_t unsafe_sp; 23 vaddr_t fs_base; 24 vaddr_t gs_base;
|
H A D | aspace.h | 39 bool check_vaddr(vaddr_t vaddr) final; 61 bool check_vaddr(vaddr_t vaddr) final; 76 zx_status_t Init(vaddr_t base, size_t size, uint mmu_flags) override; 81 zx_status_t MapContiguous(vaddr_t vaddr, paddr_t paddr, size_t count, 83 zx_status_t Map(vaddr_t vaddr, paddr_t* phys, size_t count, uint mmu_flags, 85 zx_status_t Unmap(vaddr_t vaddr, size_t count, size_t* unmapped) override; 86 zx_status_t Protect(vaddr_t vaddr, size_t count, uint mmu_flags) override; 87 zx_status_t Query(vaddr_t vaddr, paddr_t* paddr, uint* mmu_flags) override; 89 vaddr_t PickSpot(vaddr_t bas [all...] |
/fuchsia/zircon/kernel/arch/arm64/include/arch/arm64/ |
H A D | periphmap.h | 23 vaddr_t periph_paddr_to_vaddr(paddr_t paddr);
|
/fuchsia/zircon/kernel/arch/arm64/ |
H A D | user_copy_c.cpp | 16 if (!is_user_address_range((vaddr_t)src, len)) { 25 if (!is_user_address_range((vaddr_t)dst, len)) {
|
H A D | mmu.cpp | 295 zx_status_t ArmArchVmAspace::Query(vaddr_t vaddr, paddr_t* paddr, uint* mmu_flags) { 300 zx_status_t ArmArchVmAspace::QueryLocked(vaddr_t vaddr, paddr_t* paddr, uint* mmu_flags) { 308 vaddr_t vaddr_rem; 324 vaddr_t kernel_base = ~0UL << MMU_KERNEL_SIZE_SHIFT; 349 vaddr_rem -= (vaddr_t)index << index_shift; 429 volatile pte_t* ArmArchVmAspace::GetPageTable(vaddr_t index, uint page_size_shift, 492 void ArmArchVmAspace::FlushTLBEntry(vaddr_t vaddr, bool terminal) { 507 ARM64_TLBI(vale1is, vaddr >> 12 | (vaddr_t)asid_ << 48); 509 ARM64_TLBI(vae1is, vaddr >> 12 | (vaddr_t)asid_ << 48); 515 ssize_t ArmArchVmAspace::UnmapPageTable(vaddr_t vadd [all...] |
H A D | thread.cpp | 43 void arch_thread_initialize(thread_t* t, vaddr_t entry_point) { 48 vaddr_t stack_top = t->stack.top; 66 t->arch.sp = (vaddr_t)frame;
|
/fuchsia/zircon/kernel/include/arch/ |
H A D | thread.h | 18 void arch_thread_initialize(struct thread *, vaddr_t entry_point);
|
/fuchsia/zircon/kernel/arch/x86/page_tables/include/arch/x86/page_tables/ |
H A D | page_tables.h | 36 vaddr_t addr() const { return encoded_addr() << PAGE_SIZE_SHIFT; } 54 void enqueue(vaddr_t v, PageTableLevel level, bool is_global_page, bool is_terminal); 85 zx_status_t MapPages(vaddr_t vaddr, paddr_t* phys, size_t count, 87 zx_status_t MapPagesContiguous(vaddr_t vaddr, paddr_t paddr, const size_t count, 89 zx_status_t UnmapPages(vaddr_t vaddr, const size_t count, size_t* unmapped); 90 zx_status_t ProtectPages(vaddr_t vaddr, size_t count, uint flags); 92 zx_status_t QueryVaddr(vaddr_t vaddr, paddr_t* paddr, uint* mmu_flags); 100 void Destroy(vaddr_t base, size_t size); 109 virtual bool check_vaddr(vaddr_t vaddr) = 0; 167 zx_status_t GetMapping(volatile pt_entry_t* table, vaddr_t vadd [all...] |
/fuchsia/zircon/kernel/vm/ |
H A D | vm_address_region.cpp | 24 VmAddressRegion::VmAddressRegion(VmAspace& aspace, vaddr_t base, size_t size, uint32_t vmar_flags) 35 VmAddressRegion::VmAddressRegion(VmAddressRegion& parent, vaddr_t base, size_t size, 133 vaddr_t new_base = -1; 270 vaddr_t base, size_t size, uint32_t vmar_flags, 350 fbl::RefPtr<VmAddressRegionOrMapping> VmAddressRegion::FindRegion(vaddr_t addr) { 358 fbl::RefPtr<VmAddressRegionOrMapping> VmAddressRegion::FindRegionLocked(vaddr_t addr) { 386 zx_status_t VmAddressRegion::PageFault(vaddr_t va, uint pf_flags) { 401 bool VmAddressRegion::IsRangeAvailableLocked(vaddr_t base, size_t size) { 413 vaddr_t prev_last_byte; 423 vaddr_t last_byt [all...] |
/fuchsia/zircon/kernel/dev/interrupt/arm_gic/v2/include/dev/interrupt/ |
H A D | arm_gicv2m.h | 34 void arm_gicv2m_init(const paddr_t* reg_frames, const vaddr_t* reg_frames_virt, uint reg_frame_count);
|
/fuchsia/zircon/kernel/include/sys/ |
H A D | types.h | 24 typedef uintptr_t vaddr_t; typedef
|
/fuchsia/zircon/kernel/arch/x86/ |
H A D | thread.cpp | 24 void arch_thread_initialize(thread_t* t, vaddr_t entry_point) { 26 vaddr_t stack_top = t->stack.top; 50 vaddr_t buf = ROUNDUP(((vaddr_t)t->arch.extended_register_buffer), 64); 51 __UNUSED size_t overhead = buf - (vaddr_t)t->arch.extended_register_buffer; 54 t->arch.extended_register_state = (vaddr_t*)buf; 58 t->arch.sp = (vaddr_t)frame;
|
/fuchsia/zircon/kernel/arch/x86/include/arch/x86/ |
H A D | mmu.h | 96 vaddr_t start_vaddr; 101 bool x86_is_vaddr_canonical(vaddr_t vaddr);
|
/fuchsia/zircon/kernel/dev/iommu/intel/ |
H A D | second_level_pt.h | 29 bool check_vaddr(vaddr_t vaddr) final; 46 vaddr_t valid_vaddr_mask_;
|
/fuchsia/zircon/kernel/dev/power/hisi/ |
H A D | power.c | 20 static vaddr_t sctrl_base; 21 static vaddr_t pmu_base;
|
/fuchsia/zircon/kernel/object/include/object/ |
H A D | vm_address_region_dispatcher.h | 41 zx_status_t Protect(vaddr_t base, size_t len, uint32_t flags); 43 zx_status_t Unmap(vaddr_t base, size_t len);
|
/fuchsia/zircon/kernel/lib/fbl/ |
H A D | arena_tests.cpp | 57 auto start = reinterpret_cast<vaddr_t>(arena.start()); 58 auto end = reinterpret_cast<vaddr_t>(arena.end()); 167 vaddr_t start, vaddr_t end, size_t* committed, size_t* uncommitted) { 198 auto start = reinterpret_cast<vaddr_t>(arena.start()); 199 auto end = reinterpret_cast<vaddr_t>(arena.end()); 210 auto obj = reinterpret_cast<vaddr_t>(arena.Alloc()); 264 vaddr_t control_start() const { 265 return reinterpret_cast<vaddr_t>(arena_.control_.start()); 268 vaddr_t control_en [all...] |