Lines Matching refs:vaddr_t

24 VmAddressRegion::VmAddressRegion(VmAspace& aspace, vaddr_t base, size_t size, uint32_t vmar_flags)
35 VmAddressRegion::VmAddressRegion(VmAddressRegion& parent, vaddr_t base, size_t size,
133 vaddr_t new_base = -1;
270 vaddr_t base, size_t size, uint32_t vmar_flags,
350 fbl::RefPtr<VmAddressRegionOrMapping> VmAddressRegion::FindRegion(vaddr_t addr) {
358 fbl::RefPtr<VmAddressRegionOrMapping> VmAddressRegion::FindRegionLocked(vaddr_t addr) {
386 zx_status_t VmAddressRegion::PageFault(vaddr_t va, uint pf_flags) {
401 bool VmAddressRegion::IsRangeAvailableLocked(vaddr_t base, size_t size) {
413 vaddr_t prev_last_byte;
423 vaddr_t last_byte;
436 vaddr_t* pva, vaddr_t search_base, vaddr_t align,
440 vaddr_t gap_beg; // first byte of a gap
441 vaddr_t gap_end; // last byte of a gap
519 vaddr_t* spot) {
615 zx_status_t VmAddressRegion::Unmap(vaddr_t base, size_t size) {
632 zx_status_t VmAddressRegion::UnmapAllowPartial(vaddr_t base, size_t size) {
649 VmAddressRegion::ChildList::iterator VmAddressRegion::UpperBoundInternalLocked(vaddr_t base) {
662 zx_status_t VmAddressRegion::UnmapInternalLocked(vaddr_t base, size_t size,
682 const vaddr_t end_addr = base + size;
690 const vaddr_t itr_end = itr->base() + itr->size();
705 const vaddr_t curr_end = curr->base() + curr->size();
706 const vaddr_t unmap_base = fbl::max(curr->base(), base);
707 const vaddr_t unmap_end = fbl::min(curr_end, end_addr);
731 vaddr_t unmap_base = 0;
775 zx_status_t VmAddressRegion::Protect(vaddr_t base, size_t size, uint new_arch_mmu_flags) {
796 const vaddr_t end_addr = base + size;
809 vaddr_t last_mapped = begin->base();
836 const vaddr_t curr_end = itr->base() + itr->size();
837 const vaddr_t protect_base = fbl::max(itr->base(), base);
838 const vaddr_t protect_end = fbl::min(curr_end, end_addr);
856 uint arch_mmu_flags, vaddr_t* spot) {
859 const vaddr_t base = 0;
864 const vaddr_t align = 1UL << align_pow2;
873 if (*spot != static_cast<vaddr_t>(-1)) {
889 const vaddr_t align = 1UL << align_pow2;
894 vaddr_t prev_region_end = ROUNDUP(base_, align);
907 const vaddr_t end = base_ + size_;
929 vaddr_t* spot) {
934 const vaddr_t align = 1UL << align_pow2;
938 ForEachGap([align, align_pow2, size, &candidate_spaces](vaddr_t gap_base, size_t gap_len) -> bool {
956 vaddr_t alloc_spot = static_cast<vaddr_t>(-1);
957 ForEachGap([align_pow2, size, &alloc_spot, &selected_index](vaddr_t gap_base,
972 ASSERT(alloc_spot != static_cast<vaddr_t>(-1));
989 *spot != static_cast<vaddr_t>(-1)) {
1000 vaddr_t* spot) {
1004 const vaddr_t align = 1UL << align_pow2;
1027 vaddr_t chosen_base;
1032 vaddr_t base;
1044 vaddr_t base;
1055 *spot != static_cast<vaddr_t>(-1)) {