/fuchsia/zircon/kernel/vm/ |
H A D | vm_address_region_or_mapping.cpp | 73 bool VmAddressRegionOrMapping::is_valid_mapping_flags(uint arch_mmu_flags) { argument 74 if (!(flags_ & VMAR_FLAG_CAN_MAP_READ) && (arch_mmu_flags & ARCH_MMU_FLAG_PERM_READ)) { 77 if (!(flags_ & VMAR_FLAG_CAN_MAP_WRITE) && (arch_mmu_flags & ARCH_MMU_FLAG_PERM_WRITE)) { 80 if (!(flags_ & VMAR_FLAG_CAN_MAP_EXECUTE) && (arch_mmu_flags & ARCH_MMU_FLAG_PERM_EXECUTE)) {
|
H A D | vm_address_region.cpp | 76 uint64_t vmo_offset, uint arch_mmu_flags, 112 if ((arch_mmu_flags & ARCH_MMU_FLAG_CACHE_MASK) != 0 && 113 (arch_mmu_flags & ARCH_MMU_FLAG_CACHE_MASK) != cache_policy) { 115 "arch_mmu_flags %02x.\n", 116 name, cache_policy, arch_mmu_flags & ARCH_MMU_FLAG_CACHE_MASK); 118 arch_mmu_flags |= cache_policy; 145 vmo, vmo_offset, arch_mmu_flags, out); 151 zx_status_t status = AllocSpotLocked(size, align_pow2, arch_mmu_flags, &new_base); 160 (arch_mmu_flags & ARCH_MMU_FLAG_PERM_EXECUTE) && 168 fbl::move(vmo), vmo_offset, arch_mmu_flags)); 74 CreateSubVmarInternal(size_t offset, size_t size, uint8_t align_pow2, uint32_t vmar_flags, fbl::RefPtr<VmObject> vmo, uint64_t vmo_offset, uint arch_mmu_flags, const char* name, fbl::RefPtr<VmAddressRegionOrMapping>* out) argument 218 CreateVmMapping(size_t mapping_offset, size_t size, uint8_t align_pow2, uint32_t vmar_flags, fbl::RefPtr<VmObject> vmo, uint64_t vmo_offset, uint arch_mmu_flags, const char* name, fbl::RefPtr<VmMapping>* out) argument 269 OverwriteVmMapping( vaddr_t base, size_t size, uint32_t vmar_flags, fbl::RefPtr<VmObject> vmo, uint64_t vmo_offset, uint arch_mmu_flags, fbl::RefPtr<VmAddressRegionOrMapping>* out) argument 434 CheckGapLocked(const ChildList::iterator& prev, const ChildList::iterator& next, vaddr_t* pva, vaddr_t search_base, vaddr_t align, size_t region_size, size_t min_gap, uint arch_mmu_flags) argument 518 AllocSpotLocked(size_t size, uint8_t align_pow2, uint arch_mmu_flags, vaddr_t* spot) argument 855 LinearRegionAllocatorLocked(size_t size, uint8_t align_pow2, uint arch_mmu_flags, vaddr_t* spot) argument 927 NonCompactRandomizedRegionAllocatorLocked(size_t size, uint8_t align_pow2, uint arch_mmu_flags, vaddr_t* spot) argument 998 CompactRandomizedRegionAllocatorLocked(size_t size, uint8_t align_pow2, uint arch_mmu_flags, vaddr_t* spot) argument [all...] |
H A D | vm_aspace.cpp | 285 uint arch_mmu_flags) { 289 "ptr %p align %hhu vmm_flags %#x arch_mmu_flags %#x\n", 290 this, name, vmo.get(), offset, size, ptr ? *ptr : 0, align_pow2, vmm_flags, arch_mmu_flags); 292 DEBUG_ASSERT(!is_user() || !(arch_mmu_flags & ARCH_MMU_FLAG_PERM_USER)); 336 vmo, offset, arch_mmu_flags, name, &r); 388 uint arch_mmu_flags = 0; local 389 auto err = arch_aspace_.Query(vaddr, nullptr, &arch_mmu_flags); 392 arch_mmu_flags = ARCH_MMU_FLAG_CACHED | ARCH_MMU_FLAG_PERM_READ; 398 arch_mmu_flags); 402 paddr_t paddr, uint vmm_flags, uint arch_mmu_flags) { 283 MapObjectInternal(fbl::RefPtr<VmObject> vmo, const char* name, uint64_t offset, size_t size, void** ptr, uint8_t align_pow2, uint vmm_flags, uint arch_mmu_flags) argument 401 AllocPhysical(const char* name, size_t size, void** ptr, uint8_t align_pow2, paddr_t paddr, uint vmm_flags, uint arch_mmu_flags) argument 440 AllocContiguous(const char* name, size_t size, void** ptr, uint8_t align_pow2, uint vmm_flags, uint arch_mmu_flags) argument 468 Alloc(const char* name, size_t size, void** ptr, uint8_t align_pow2, uint vmm_flags, uint arch_mmu_flags) argument [all...] |
H A D | vm.cpp | 63 zx_status_t ProtectRegion(VmAspace* aspace, vaddr_t va, uint arch_mmu_flags) { argument 74 return vm_mapping->Protect(vm_mapping->base(), vm_mapping->size(), arch_mmu_flags); 128 uint arch_mmu_flags; member in struct:temp_region 134 .arch_mmu_flags = ARCH_MMU_FLAG_PERM_READ | ARCH_MMU_FLAG_PERM_EXECUTE, 140 .arch_mmu_flags = ARCH_MMU_FLAG_PERM_READ, 146 .arch_mmu_flags = ARCH_MMU_FLAG_PERM_READ | ARCH_MMU_FLAG_PERM_WRITE, 152 .arch_mmu_flags = ARCH_MMU_FLAG_PERM_READ | ARCH_MMU_FLAG_PERM_WRITE, 161 region->base, region->base + region->size, region->arch_mmu_flags, region->name); 165 status = ProtectRegion(aspace, region->base, region->arch_mmu_flags);
|
H A D | vm_mapping.cpp | 25 fbl::RefPtr<VmObject> vmo, uint64_t vmo_offset, uint arch_mmu_flags) 28 object_(fbl::move(vmo)), object_offset_(vmo_offset), arch_mmu_flags_(arch_mmu_flags) { 447 uint flags = mapping_->arch_mmu_flags(); 24 VmMapping(VmAddressRegion& parent, vaddr_t base, size_t size, uint32_t vmar_flags, fbl::RefPtr<VmObject> vmo, uint64_t vmo_offset, uint arch_mmu_flags) argument
|
/fuchsia/zircon/kernel/object/ |
H A D | vm_address_region_dispatcher.cpp | 29 zx_status_t split_syscall_flags(uint32_t flags, uint32_t* vmar_flags, uint* arch_mmu_flags) { argument 30 // Figure out arch_mmu_flags 45 // Mask out arch_mmu_flags options 87 *arch_mmu_flags |= mmu_flags; 135 uint arch_mmu_flags = 0; local 136 zx_status_t status = split_syscall_flags(flags, &vmar_flags, &arch_mmu_flags); 141 if (arch_mmu_flags != 0) { 178 // Split flags into vmar_flags and arch_mmu_flags 180 uint arch_mmu_flags = base_arch_mmu_flags_; local 181 zx_status_t status = split_syscall_flags(flags, &vmar_flags, &arch_mmu_flags); 215 uint arch_mmu_flags = base_arch_mmu_flags_; local [all...] |
H A D | diagnostics.cpp | 442 unsigned int arch_mmu_flags_to_vm_flags(unsigned int arch_mmu_flags) { argument 443 if (arch_mmu_flags & ARCH_MMU_FLAG_INVALID) { 447 if (arch_mmu_flags & ARCH_MMU_FLAG_PERM_READ) { 450 if (arch_mmu_flags & ARCH_MMU_FLAG_PERM_WRITE) { 453 if (arch_mmu_flags & ARCH_MMU_FLAG_PERM_EXECUTE) { 497 arch_mmu_flags_to_vm_flags(map->arch_mmu_flags());
|
/fuchsia/zircon/kernel/vm/include/vm/ |
H A D | vm_address_region.h | 152 // Check if the given *arch_mmu_flags* are allowed under this 154 bool is_valid_mapping_flags(uint arch_mmu_flags); 218 uint arch_mmu_flags, const char* name, 292 uint arch_mmu_flags, const char* name, 300 uint arch_mmu_flags, 325 size_t region_size, size_t min_gap, uint arch_mmu_flags); 328 zx_status_t AllocSpotLocked(size_t size, uint8_t align_pow2, uint arch_mmu_flags, vaddr_t* spot); 331 zx_status_t LinearRegionAllocatorLocked(size_t size, uint8_t align_pow2, uint arch_mmu_flags, 334 uint arch_mmu_flags, vaddr_t* spot); 336 uint arch_mmu_flags, vaddr_ 425 uint arch_mmu_flags() const { return arch_mmu_flags_; } function in class:final [all...] |
H A D | vm_aspace.h | 138 uint arch_mmu_flags); 140 uint vmm_flags, uint arch_mmu_flags); 142 uint vmm_flags, uint arch_mmu_flags); 149 uint arch_mmu_flags);
|
/fuchsia/zircon/kernel/lib/unittest/ |
H A D | user_memory.cpp | 38 constexpr uint arch_mmu_flags = ARCH_MMU_FLAG_PERM_READ | local 41 vmo, 0, arch_mmu_flags, "unittest", &mapping);
|
/fuchsia/zircon/kernel/lib/hypervisor/ |
H A D | guest_physical_address_space.cpp | 128 if (mapping->arch_mmu_flags() & ARCH_MMU_FLAG_PERM_WRITE) { 131 if (mapping->arch_mmu_flags() & ARCH_MMU_FLAG_PERM_EXECUTE) {
|
/fuchsia/zircon/kernel/lib/vdso/ |
H A D | vdso.cpp | 47 const uint arch_mmu_flags = ARCH_MMU_FLAG_PERM_READ | ARCH_MMU_FLAG_PERM_WRITE; local 50 fbl::move(vmo), page_offset, arch_mmu_flags, name, &mapping_);
|
/fuchsia/zircon/kernel/arch/x86/ |
H A D | perf_mon.cpp | 1354 uint32_t arch_mmu_flags = ARCH_MMU_FLAG_PERM_READ; 1358 0, arch_mmu_flags, name, 1405 const uint arch_mmu_flags = ARCH_MMU_FLAG_PERM_READ | ARCH_MMU_FLAG_PERM_WRITE; 1409 data->buffer_vmo, vmo_offset, arch_mmu_flags, name, 1352 uint32_t arch_mmu_flags = ARCH_MMU_FLAG_PERM_READ; local 1403 const uint arch_mmu_flags = ARCH_MMU_FLAG_PERM_READ | ARCH_MMU_FLAG_PERM_WRITE; local
|