Lines Matching refs:mapping

136     // TODO(teisenbe): deal with error mapping on arch_mmu_protect fail
138 // If we're changing the whole mapping, just make the change.
148 // Create a new mapping for the right half (has old perms)
150 fbl::RefPtr<VmMapping> mapping(fbl::AdoptRef(
162 mapping->ActivateLocked();
168 // Create a new mapping for the right half (has new perms)
171 fbl::RefPtr<VmMapping> mapping(fbl::AdoptRef(
183 mapping->ActivateLocked();
241 // If we're unmapping everything, destroy this mapping
291 // We're unmapping from the center, so we need to split the mapping
299 fbl::RefPtr<VmMapping> mapping(fbl::AdoptRef(
315 mapping->ActivateLocked();
330 // mapping list with the VMO lock held before dropping this state to DEAD. The
355 // compute the intersection of the passed in vmo range and our mapping
394 VmMappingCoalescer(VmMapping* mapping, vaddr_t base);
397 // Add a page to the mapping run. If this fails, the VmMappingCoalescer is
434 VmMappingCoalescer::VmMappingCoalescer(VmMapping* mapping, vaddr_t base)
435 : mapping_(mapping), base_(base), count_(0), aborted_(false) {}
453 TRACEF("error %d mapping %zu pages starting at va %#" PRIxPTR "\n", ret, count_, base_);
502 // mapping it in
524 LTRACEF_LEVEL(2, "mapping pa %#" PRIxPTR " to va %#" PRIxPTR "\n", pa, va);
561 // The vDSO code mapping can never be unmapped, not even
690 TRACEF("failed to modify permissions on existing mapping\n");
699 // assert that we're not accidentally mapping the zero page writable
705 TRACEF("failed to remove old mapping before replacing\n");
721 LTRACEF("mapping pa %#" PRIxPTR " to va %#" PRIxPTR " is zero page %d\n",
724 // assert that we're not accidentally mapping the zero page writable
748 // function is for splitting one mapping object into several that will be backed