Searched refs:arch_mmu_flags (Results 1 - 13 of 13) sorted by relevance

/fuchsia/zircon/kernel/vm/
H A Dvm_address_region_or_mapping.cpp73 bool VmAddressRegionOrMapping::is_valid_mapping_flags(uint arch_mmu_flags) { argument
74 if (!(flags_ & VMAR_FLAG_CAN_MAP_READ) && (arch_mmu_flags & ARCH_MMU_FLAG_PERM_READ)) {
77 if (!(flags_ & VMAR_FLAG_CAN_MAP_WRITE) && (arch_mmu_flags & ARCH_MMU_FLAG_PERM_WRITE)) {
80 if (!(flags_ & VMAR_FLAG_CAN_MAP_EXECUTE) && (arch_mmu_flags & ARCH_MMU_FLAG_PERM_EXECUTE)) {
H A Dvm_address_region.cpp76 uint64_t vmo_offset, uint arch_mmu_flags,
112 if ((arch_mmu_flags & ARCH_MMU_FLAG_CACHE_MASK) != 0 &&
113 (arch_mmu_flags & ARCH_MMU_FLAG_CACHE_MASK) != cache_policy) {
115 "arch_mmu_flags %02x.\n",
116 name, cache_policy, arch_mmu_flags & ARCH_MMU_FLAG_CACHE_MASK);
118 arch_mmu_flags |= cache_policy;
145 vmo, vmo_offset, arch_mmu_flags, out);
151 zx_status_t status = AllocSpotLocked(size, align_pow2, arch_mmu_flags, &new_base);
160 (arch_mmu_flags & ARCH_MMU_FLAG_PERM_EXECUTE) &&
168 fbl::move(vmo), vmo_offset, arch_mmu_flags));
74 CreateSubVmarInternal(size_t offset, size_t size, uint8_t align_pow2, uint32_t vmar_flags, fbl::RefPtr<VmObject> vmo, uint64_t vmo_offset, uint arch_mmu_flags, const char* name, fbl::RefPtr<VmAddressRegionOrMapping>* out) argument
218 CreateVmMapping(size_t mapping_offset, size_t size, uint8_t align_pow2, uint32_t vmar_flags, fbl::RefPtr<VmObject> vmo, uint64_t vmo_offset, uint arch_mmu_flags, const char* name, fbl::RefPtr<VmMapping>* out) argument
269 OverwriteVmMapping( vaddr_t base, size_t size, uint32_t vmar_flags, fbl::RefPtr<VmObject> vmo, uint64_t vmo_offset, uint arch_mmu_flags, fbl::RefPtr<VmAddressRegionOrMapping>* out) argument
434 CheckGapLocked(const ChildList::iterator& prev, const ChildList::iterator& next, vaddr_t* pva, vaddr_t search_base, vaddr_t align, size_t region_size, size_t min_gap, uint arch_mmu_flags) argument
518 AllocSpotLocked(size_t size, uint8_t align_pow2, uint arch_mmu_flags, vaddr_t* spot) argument
855 LinearRegionAllocatorLocked(size_t size, uint8_t align_pow2, uint arch_mmu_flags, vaddr_t* spot) argument
927 NonCompactRandomizedRegionAllocatorLocked(size_t size, uint8_t align_pow2, uint arch_mmu_flags, vaddr_t* spot) argument
998 CompactRandomizedRegionAllocatorLocked(size_t size, uint8_t align_pow2, uint arch_mmu_flags, vaddr_t* spot) argument
[all...]
H A Dvm_aspace.cpp285 uint arch_mmu_flags) {
289 "ptr %p align %hhu vmm_flags %#x arch_mmu_flags %#x\n",
290 this, name, vmo.get(), offset, size, ptr ? *ptr : 0, align_pow2, vmm_flags, arch_mmu_flags);
292 DEBUG_ASSERT(!is_user() || !(arch_mmu_flags & ARCH_MMU_FLAG_PERM_USER));
336 vmo, offset, arch_mmu_flags, name, &r);
388 uint arch_mmu_flags = 0; local
389 auto err = arch_aspace_.Query(vaddr, nullptr, &arch_mmu_flags);
392 arch_mmu_flags = ARCH_MMU_FLAG_CACHED | ARCH_MMU_FLAG_PERM_READ;
398 arch_mmu_flags);
402 paddr_t paddr, uint vmm_flags, uint arch_mmu_flags) {
283 MapObjectInternal(fbl::RefPtr<VmObject> vmo, const char* name, uint64_t offset, size_t size, void** ptr, uint8_t align_pow2, uint vmm_flags, uint arch_mmu_flags) argument
401 AllocPhysical(const char* name, size_t size, void** ptr, uint8_t align_pow2, paddr_t paddr, uint vmm_flags, uint arch_mmu_flags) argument
440 AllocContiguous(const char* name, size_t size, void** ptr, uint8_t align_pow2, uint vmm_flags, uint arch_mmu_flags) argument
468 Alloc(const char* name, size_t size, void** ptr, uint8_t align_pow2, uint vmm_flags, uint arch_mmu_flags) argument
[all...]
H A Dvm.cpp63 zx_status_t ProtectRegion(VmAspace* aspace, vaddr_t va, uint arch_mmu_flags) { argument
74 return vm_mapping->Protect(vm_mapping->base(), vm_mapping->size(), arch_mmu_flags);
128 uint arch_mmu_flags; member in struct:temp_region
134 .arch_mmu_flags = ARCH_MMU_FLAG_PERM_READ | ARCH_MMU_FLAG_PERM_EXECUTE,
140 .arch_mmu_flags = ARCH_MMU_FLAG_PERM_READ,
146 .arch_mmu_flags = ARCH_MMU_FLAG_PERM_READ | ARCH_MMU_FLAG_PERM_WRITE,
152 .arch_mmu_flags = ARCH_MMU_FLAG_PERM_READ | ARCH_MMU_FLAG_PERM_WRITE,
161 region->base, region->base + region->size, region->arch_mmu_flags, region->name);
165 status = ProtectRegion(aspace, region->base, region->arch_mmu_flags);
H A Dvm_mapping.cpp25 fbl::RefPtr<VmObject> vmo, uint64_t vmo_offset, uint arch_mmu_flags)
28 object_(fbl::move(vmo)), object_offset_(vmo_offset), arch_mmu_flags_(arch_mmu_flags) {
447 uint flags = mapping_->arch_mmu_flags();
24 VmMapping(VmAddressRegion& parent, vaddr_t base, size_t size, uint32_t vmar_flags, fbl::RefPtr<VmObject> vmo, uint64_t vmo_offset, uint arch_mmu_flags) argument
/fuchsia/zircon/kernel/object/
H A Dvm_address_region_dispatcher.cpp29 zx_status_t split_syscall_flags(uint32_t flags, uint32_t* vmar_flags, uint* arch_mmu_flags) { argument
30 // Figure out arch_mmu_flags
45 // Mask out arch_mmu_flags options
87 *arch_mmu_flags |= mmu_flags;
135 uint arch_mmu_flags = 0; local
136 zx_status_t status = split_syscall_flags(flags, &vmar_flags, &arch_mmu_flags);
141 if (arch_mmu_flags != 0) {
178 // Split flags into vmar_flags and arch_mmu_flags
180 uint arch_mmu_flags = base_arch_mmu_flags_; local
181 zx_status_t status = split_syscall_flags(flags, &vmar_flags, &arch_mmu_flags);
215 uint arch_mmu_flags = base_arch_mmu_flags_; local
[all...]
H A Ddiagnostics.cpp442 unsigned int arch_mmu_flags_to_vm_flags(unsigned int arch_mmu_flags) { argument
443 if (arch_mmu_flags & ARCH_MMU_FLAG_INVALID) {
447 if (arch_mmu_flags & ARCH_MMU_FLAG_PERM_READ) {
450 if (arch_mmu_flags & ARCH_MMU_FLAG_PERM_WRITE) {
453 if (arch_mmu_flags & ARCH_MMU_FLAG_PERM_EXECUTE) {
497 arch_mmu_flags_to_vm_flags(map->arch_mmu_flags());
/fuchsia/zircon/kernel/vm/include/vm/
H A Dvm_address_region.h152 // Check if the given *arch_mmu_flags* are allowed under this
154 bool is_valid_mapping_flags(uint arch_mmu_flags);
218 uint arch_mmu_flags, const char* name,
292 uint arch_mmu_flags, const char* name,
300 uint arch_mmu_flags,
325 size_t region_size, size_t min_gap, uint arch_mmu_flags);
328 zx_status_t AllocSpotLocked(size_t size, uint8_t align_pow2, uint arch_mmu_flags, vaddr_t* spot);
331 zx_status_t LinearRegionAllocatorLocked(size_t size, uint8_t align_pow2, uint arch_mmu_flags,
334 uint arch_mmu_flags, vaddr_t* spot);
336 uint arch_mmu_flags, vaddr_
425 uint arch_mmu_flags() const { return arch_mmu_flags_; } function in class:final
[all...]
H A Dvm_aspace.h138 uint arch_mmu_flags);
140 uint vmm_flags, uint arch_mmu_flags);
142 uint vmm_flags, uint arch_mmu_flags);
149 uint arch_mmu_flags);
/fuchsia/zircon/kernel/lib/unittest/
H A Duser_memory.cpp38 constexpr uint arch_mmu_flags = ARCH_MMU_FLAG_PERM_READ | local
41 vmo, 0, arch_mmu_flags, "unittest", &mapping);
/fuchsia/zircon/kernel/lib/hypervisor/
H A Dguest_physical_address_space.cpp128 if (mapping->arch_mmu_flags() & ARCH_MMU_FLAG_PERM_WRITE) {
131 if (mapping->arch_mmu_flags() & ARCH_MMU_FLAG_PERM_EXECUTE) {
/fuchsia/zircon/kernel/lib/vdso/
H A Dvdso.cpp47 const uint arch_mmu_flags = ARCH_MMU_FLAG_PERM_READ | ARCH_MMU_FLAG_PERM_WRITE; local
50 fbl::move(vmo), page_offset, arch_mmu_flags, name, &mapping_);
/fuchsia/zircon/kernel/arch/x86/
H A Dperf_mon.cpp1354 uint32_t arch_mmu_flags = ARCH_MMU_FLAG_PERM_READ;
1358 0, arch_mmu_flags, name,
1405 const uint arch_mmu_flags = ARCH_MMU_FLAG_PERM_READ | ARCH_MMU_FLAG_PERM_WRITE;
1409 data->buffer_vmo, vmo_offset, arch_mmu_flags, name,
1352 uint32_t arch_mmu_flags = ARCH_MMU_FLAG_PERM_READ; local
1403 const uint arch_mmu_flags = ARCH_MMU_FLAG_PERM_READ | ARCH_MMU_FLAG_PERM_WRITE; local

Completed in 162 milliseconds