Lines Matching defs:arch_mmu_flags

76                                                    uint64_t vmo_offset, uint arch_mmu_flags,
112 if ((arch_mmu_flags & ARCH_MMU_FLAG_CACHE_MASK) != 0 &&
113 (arch_mmu_flags & ARCH_MMU_FLAG_CACHE_MASK) != cache_policy) {
115 "arch_mmu_flags %02x.\n",
116 name, cache_policy, arch_mmu_flags & ARCH_MMU_FLAG_CACHE_MASK);
118 arch_mmu_flags |= cache_policy;
145 vmo, vmo_offset, arch_mmu_flags, out);
151 zx_status_t status = AllocSpotLocked(size, align_pow2, arch_mmu_flags, &new_base);
160 (arch_mmu_flags & ARCH_MMU_FLAG_PERM_EXECUTE) &&
168 fbl::move(vmo), vmo_offset, arch_mmu_flags));
220 uint64_t vmo_offset, uint arch_mmu_flags, const char* name,
230 // Validate that arch_mmu_flags does not contain any prohibited flags
231 if (!is_valid_mapping_flags(arch_mmu_flags)) {
247 if (arch_mmu_flags & ARCH_MMU_FLAG_PERM_READ) {
250 if (arch_mmu_flags & ARCH_MMU_FLAG_PERM_WRITE) {
253 if (arch_mmu_flags & ARCH_MMU_FLAG_PERM_EXECUTE) {
260 vmo_offset, arch_mmu_flags, name, &res);
272 uint arch_mmu_flags, fbl::RefPtr<VmAddressRegionOrMapping>* out) {
283 fbl::move(vmo), vmo_offset, arch_mmu_flags));
437 size_t region_size, size_t min_gap, uint arch_mmu_flags) {
492 ? prev->as_vm_mapping()->arch_mmu_flags()
496 ? next->as_vm_mapping()->arch_mmu_flags()
500 next_arch_mmu_flags, align, region_size, arch_mmu_flags);
518 zx_status_t VmAddressRegion::AllocSpotLocked(size_t size, uint8_t align_pow2, uint arch_mmu_flags,
529 return CompactRandomizedRegionAllocatorLocked(size, align_pow2, arch_mmu_flags, spot);
531 return NonCompactRandomizedRegionAllocatorLocked(size, align_pow2, arch_mmu_flags,
535 return LinearRegionAllocatorLocked(size, align_pow2, arch_mmu_flags, spot);
856 uint arch_mmu_flags, vaddr_t* spot) {
872 if (CheckGapLocked(before_iter, after_iter, spot, base, align, size, 0, arch_mmu_flags)) {
928 uint arch_mmu_flags,
988 arch_mmu_flags) &&
999 uint arch_mmu_flags,
1007 return NonCompactRandomizedRegionAllocatorLocked(size, align_pow2, arch_mmu_flags, spot);
1054 arch_mmu_flags) &&