Searched refs:pageIndex (Results 1 - 6 of 6) sorted by relevance
/haiku-fatelf/src/system/kernel/vm/ |
H A D | VMAnonymousCache.cpp | 177 ReadPage(VMAnonymousCache* cache, page_num_t pageIndex, argument 181 fPageIndex(pageIndex), 201 WritePage(VMAnonymousCache* cache, page_num_t pageIndex, argument 205 fPageIndex(pageIndex), 406 void SetTo(page_num_t pageIndex, swap_addr_t slotIndex, bool newSlot) argument 408 fPageIndex = pageIndex; 496 for (page_num_t pageIndex = (newSize + B_PAGE_SIZE - 1) >> PAGE_SHIFT; 497 pageIndex < oldPageCount && fAllocatedSwapSize > 0; pageIndex++) { 502 swap_addr_t blockIndex = pageIndex 582 page_num_t pageIndex = offset >> PAGE_SHIFT; local 596 off_t pageIndex = offset >> PAGE_SHIFT; local 628 off_t pageIndex = offset >> PAGE_SHIFT; local 722 page_num_t pageIndex = offset >> PAGE_SHIFT; local 883 off_t pageIndex = startPageIndex + i; local 928 off_t pageIndex = startPageIndex + i; local 951 _SwapBlockGetAddress(off_t pageIndex) argument 1130 off_t pageIndex = swapBlockPageIndex + i; local [all...] |
H A D | VMAnonymousCache.h | 76 void _SwapBlockBuild(off_t pageIndex, 78 void _SwapBlockFree(off_t pageIndex, uint32 count); 79 swap_addr_t _SwapBlockGetAddress(off_t pageIndex);
|
H A D | vm.cpp | 475 uint32 pageIndex = (pageAddress - area->Base()) / B_PAGE_SIZE; local 476 uint8& entry = area->page_protections[pageIndex / 2]; 477 if (pageIndex % 2 == 0) 490 uint32 pageIndex = (pageAddress - area->Base()) / B_PAGE_SIZE; 491 uint32 protection = area->page_protections[pageIndex / 2]; 492 if (pageIndex % 2 == 0)
|
/haiku-fatelf/src/system/kernel/ |
H A D | guarded_heap.cpp | 154 guarded_heap_page_protect(guarded_heap_area& area, size_t pageIndex, argument 160 addr_t address = area.base + pageIndex * B_PAGE_SIZE; 220 guarded_heap_free_page(guarded_heap_area& area, size_t pageIndex, argument 223 guarded_heap_page& page = area.pages[pageIndex]; 244 guarded_heap_page_protect(area, pageIndex, 0); 246 T(Free(area.heap, (void*)(area.base + pageIndex * B_PAGE_SIZE))); 287 size_t pageIndex = page - area.pages; local 288 if (pageIndex > area.page_count - pagesNeeded) 295 if ((area.pages[pageIndex + j].flags & GUARDED_HEAP_PAGE_FLAG_USED) 309 void* result = (void*)((area.base + pageIndex * B_PAGE_SIZ 464 size_t pageIndex = ((addr_t)address - area.base) / B_PAGE_SIZE; local 497 size_t pageIndex = guarded_heap_area_page_index_for(area, address); local 549 size_t pageIndex = guarded_heap_area_page_index_for(*area, address); local 604 size_t pageIndex = ((addr_t)address - area->base) / B_PAGE_SIZE; local [all...] |
/haiku-fatelf/src/system/libroot/posix/malloc_debug/ |
H A D | guarded_heap.cpp | 184 guarded_heap_page_protect(guarded_heap_area& area, size_t pageIndex, argument 187 addr_t address = area.base + pageIndex * B_PAGE_SIZE; 234 guarded_heap_free_page(guarded_heap_area& area, size_t pageIndex, argument 237 guarded_heap_page& page = area.pages[pageIndex]; 248 guarded_heap_page_protect(area, pageIndex, 0); 278 size_t pageIndex = page - area.pages; local 279 if (pageIndex > area.page_count - pagesNeeded) 286 if ((area.pages[pageIndex + j].flags & GUARDED_HEAP_PAGE_FLAG_USED) 297 void* result = (void*)((area.base + pageIndex * B_PAGE_SIZE 300 guarded_heap_page_allocate(area, pageIndex, pagesNeede 478 size_t pageIndex = ((addr_t)address - area.base) / B_PAGE_SIZE; local 511 size_t pageIndex = guarded_heap_area_page_index_for(area, address); local 603 size_t pageIndex = guarded_heap_area_page_index_for(*area, address); local 685 size_t pageIndex = ((addr_t)address - area->base) / B_PAGE_SIZE; local [all...] |
/haiku-fatelf/src/system/kernel/cache/ |
H A D | file_cache.cpp | 392 int32 pageIndex = 0; local 396 vm_page* page = pages[pageIndex++] = vm_page_allocate_page( 420 for (int32 i = 0; i < pageIndex; i++) { 431 for (int32 i = 0; i < pageIndex; i++) { 449 for (int32 i = pageIndex; i-- > 0;) { 507 int32 pageIndex = 0; local 520 vm_page* page = pages[pageIndex++] = vm_page_allocate_page( 622 for (int32 i = pageIndex; i-- > 0;) {
|
Completed in 98 milliseconds