Lines Matching defs:pages
61 void *pages;
71 Slab *ConstructSlab(Slab *slab, void *pages, size_t byteCount,
100 void **pages, size_t byteCount, uint32_t flags);
101 static void FreePages(BaseCache *cache, void *pages);
109 static status_t AllocatePages(BaseCache *cache, area_id *id, void **pages,
179 BaseCache::Slab *_ConstructSlab(Slab *slab, void *pages, size_t tailSpace,
182 return fParent->ConstructSlab(slab, pages, SlabSize(tailSpace)
197 // slab at the end of the allocated pages. It uses aligned allocations to
229 void *pages;
232 // map objects to slabs we required this set of pages to be
234 if (Backend::AllocatePages(fParent, &id, &pages, _SlabSize(),
238 _SlabInPages(pages)->id = id;
240 return BaseCacheStrategy<Backend>::_ConstructSlab(_SlabInPages(pages),
241 pages, sizeof(Slab), _Linkage, this);
261 Slab *_SlabInPages(const void *pages) const
263 return (Slab *)(((uint8_t *)pages) + _SlabSize() - sizeof(Slab));
405 void *pages;
406 if (Backend::AllocatePages(fParent, &slab->id, &pages, byteCount,
412 if (_PrepareSlab(fParent, slab, pages, byteCount, flags) < B_OK) {
420 return BaseCacheStrategy<Backend>::_ConstructSlab(slab, pages, 0,
426 _ClearSlab(fParent, slab->pages, _SlabSize());
437 status_t _PrepareSlab(BaseCache *parent, Slab *slab, void *pages,
440 uint8_t *data = (uint8_t *)pages;
458 void _ClearSlab(BaseCache *parent, void *pages, size_t byteCount)
460 _ClearSlabRange(parent, (uint8_t *)pages,
461 ((uint8_t *)pages) + byteCount);