Lines Matching refs:VmObjectPaged

57     if (*out_size > VmObjectPaged::MAX_SIZE) {
66 VmObjectPaged::VmObjectPaged(
77 VmObjectPaged::~VmObjectPaged() {
95 zx_status_t VmObjectPaged::Create(uint32_t pmm_alloc_flags,
111 new (&ac) VmObjectPaged(options, pmm_alloc_flags, size, nullptr));
121 zx_status_t VmObjectPaged::CreateContiguous(uint32_t pmm_alloc_flags, uint64_t size,
132 new (&ac) VmObjectPaged(kContiguous, pmm_alloc_flags, size, nullptr));
158 VmObjectPaged* vmop = static_cast<VmObjectPaged*>(vmo.get());
187 zx_status_t VmObjectPaged::CreateFromROData(const void* data, size_t size, fbl::RefPtr<VmObject>* obj) {
228 auto vmo2 = static_cast<VmObjectPaged*>(vmo.get());
238 zx_status_t VmObjectPaged::CloneCOW(bool resizable, uint64_t offset, uint64_t size,
254 auto vmo = fbl::AdoptRef<VmObjectPaged>(
255 new (&ac) VmObjectPaged(options, pmm_alloc_flags_, size, fbl::WrapRefPtr(this)));
286 void VmObjectPaged::Dump(uint depth, bool verbose) {
319 size_t VmObjectPaged::AllocatedPagesInRange(uint64_t offset, uint64_t len) const {
339 zx_status_t VmObjectPaged::AddPage(vm_page_t* p, uint64_t offset) {
345 zx_status_t VmObjectPaged::AddPageLocked(vm_page_t* p, uint64_t offset) {
375 zx_status_t VmObjectPaged::GetPageLocked(uint64_t offset, uint pf_flags, list_node* free_list,
536 zx_status_t VmObjectPaged::CommitRange(uint64_t offset, uint64_t len, uint64_t* committed) {
623 zx_status_t VmObjectPaged::DecommitRange(uint64_t offset, uint64_t len, uint64_t* decommitted) {
681 zx_status_t VmObjectPaged::Pin(uint64_t offset, uint64_t len) {
688 zx_status_t VmObjectPaged::PinLocked(uint64_t offset, uint64_t len) {
732 void VmObjectPaged::Unpin(uint64_t offset, uint64_t len) {
737 void VmObjectPaged::UnpinLocked(uint64_t offset, uint64_t len) {
770 bool VmObjectPaged::AnyPagesPinnedLocked(uint64_t offset, size_t len) {
794 zx_status_t VmObjectPaged::ResizeLocked(uint64_t s) {
852 zx_status_t VmObjectPaged::Resize(uint64_t s) {
858 zx_status_t VmObjectPaged::SetParentOffsetLocked(uint64_t offset) {
882 zx_status_t VmObjectPaged::ReadWriteInternal(uint64_t offset, size_t len, bool write, T copyfunc) {
931 zx_status_t VmObjectPaged::Read(void* _ptr, uint64_t offset, size_t len) {
949 zx_status_t VmObjectPaged::Write(const void* _ptr, uint64_t offset, size_t len) {
967 zx_status_t VmObjectPaged::Lookup(uint64_t offset, uint64_t len, uint pf_flags,
1045 zx_status_t VmObjectPaged::ReadUser(user_out_ptr<void> ptr, uint64_t offset, size_t len) {
1056 zx_status_t VmObjectPaged::WriteUser(user_in_ptr<const void> ptr, uint64_t offset, size_t len) {
1067 zx_status_t VmObjectPaged::LookupUser(uint64_t offset, uint64_t len, user_inout_ptr<paddr_t> buffer,
1087 zx_status_t VmObjectPaged::InvalidateCache(const uint64_t offset, const uint64_t len) {
1091 zx_status_t VmObjectPaged::CleanCache(const uint64_t offset, const uint64_t len) {
1095 zx_status_t VmObjectPaged::CleanInvalidateCache(const uint64_t offset, const uint64_t len) {
1099 zx_status_t VmObjectPaged::SyncCache(const uint64_t offset, const uint64_t len) {
1103 zx_status_t VmObjectPaged::CacheOp(const uint64_t start_offset, const uint64_t len,
1166 zx_status_t VmObjectPaged::GetMappingCachePolicy(uint32_t* cache_policy) {
1174 zx_status_t VmObjectPaged::SetMappingCachePolicy(const uint32_t cache_policy) {
1205 void VmObjectPaged::RangeChangeUpdateFromParentLocked(const uint64_t offset, const uint64_t len) {