Lines Matching refs:BG
374 // FreeList - > BG -> BG -> BG
382 // Each BlockGroup(BG) will associate with unique group id and the free blocks
404 BatchGroup *BG = nullptr;
408 BG = reinterpret_cast<BatchGroup *>(
410 BG->Batches.clear();
416 BG = C->createGroup();
417 BG->Batches.clear();
423 BG->GroupId = GroupId;
424 BG->Batches.push_front(TB);
425 BG->PushedBlocks = 0;
426 BG->PushedBlocksAtLastCheckpoint = 0;
427 BG->MaxCachedPerBatch =
430 return BG;
433 auto InsertBlocks = [&](BatchGroup *BG, CompactPtrT *Array, u32 Size) {
434 SinglyLinkedList<TransferBatch> &Batches = BG->Batches;
439 DCHECK_GE(BG->MaxCachedPerBatch, CurBatch->getCount());
441 static_cast<u16>(BG->MaxCachedPerBatch - CurBatch->getCount());
448 UnusedSlots = BG->MaxCachedPerBatch;
456 BG->PushedBlocks += Size;
541 BatchGroup *BG = Sci->FreeList.front();
550 C->deallocate(SizeClassMap::BatchClassId, BG);
699 for (BatchGroup &BG : Sci->FreeList) {
701 BG.PushedBlocks - BG.PushedBlocksAtLastCheckpoint;
705 uptr AllocatedGroupSize = BG.GroupId == CurRegionGroupId
711 // TransferBatches are pushed in front of BG.Batches. The first one may
713 const uptr NumBlocks = (BG.Batches.size() - 1) * BG.MaxCachedPerBatch +
714 BG.Batches.front()->getCount();
724 BG.PushedBlocksAtLastCheckpoint = BG.PushedBlocks;
727 Context.markFreeBlocks(BG.Batches, DecompactPtr, Base);