Searched refs:pages (Results 1 - 25 of 25) sorted by relevance

/macosx-10.5.8/xnu-1228.15.4/iokit/Kernel/
H A DIOCopyMapper.cpp141 fNumZones -= 3; // correct for overshoot and minumum 16K pages allocation
204 ppnum_t IOCopyMapper::iovmAlloc(IOItemCount pages) argument
211 if (pages < kMinZoneSize)
212 pages = kMinZoneSize;
215 if (pages >= fMapperRegionSize/2)
217 panic("iovmAlloc 0x%lx", pages);
222 for (zone = 0, zoneSize = kMinZoneSize; pages > zoneSize; zone++)
251 DEBG("va: 0x%lx, %ld, ret %x next %x\n", (ret * kActivePerFree) + fBufferPage, pages, ret, next);
261 for (cnt = 0; cnt < pages; cnt++) {
266 fMapperRegionUsed += pages;
285 iovmFree(ppnum_t addr, IOItemCount pages) argument
[all...]
H A DIOCopyMapper.h41 #define kMinZoneSize 4 // Minimum Zone size in pages
42 #define kMaxNumZones (31 - 14) // 31 bit mapped in 16K super pages
77 virtual ppnum_t iovmAlloc(IOItemCount pages);
78 virtual void iovmFree(ppnum_t addr, IOItemCount pages);
H A DIOMapper.cpp188 ppnum_t IOMapperIOVMAlloc(unsigned pages) argument
193 return IOMapper::gSystem->iovmAlloc((IOItemCount) pages);
198 void IOMapperIOVMFree(ppnum_t addr, unsigned pages) argument
201 IOMapper::gSystem->iovmFree(addr, (IOItemCount) pages);
H A DIOMemoryDescriptor.cpp848 unsigned int dataSize = computeDataSize(/* pages */ 0, /* upls */ 1);
938 UInt32 pages = 0; local
945 pages += (atop_64(addr + len + PAGE_MASK) - atop_64(addr));
958 _pages = pages;
2620 // in the middle of the loop only map whole pages
3385 data->fSGCount = 0; // XXX gvdl: need to compute and pages
/macosx-10.5.8/xnu-1228.15.4/iokit/IOKit/
H A DIOMapper.h39 ppnum_t IOMapperIOVMAlloc(unsigned pages);
40 void IOMapperIOVMFree(ppnum_t addr, unsigned pages);
104 virtual ppnum_t iovmAlloc(IOItemCount pages) = 0;
105 virtual void iovmFree(ppnum_t addr, IOItemCount pages) = 0;
/macosx-10.5.8/AppleMacRiscPCI-3.4/
H A DAppleDART.cpp146 #define kMinZoneSize 4 // Minimum Zone size in pages
147 #define kMaxNumZones (31 - 14) // 31 bit mapped in 16K super pages
202 virtual ppnum_t iovmAlloc(IOItemCount pages);
203 virtual void iovmFree(ppnum_t addr, IOItemCount pages);
323 fNumZones -= 3; // correct for overshoot and minumum 16K pages allocation
406 ppnum_t AppleDART::iovmAlloc(IOItemCount pages) argument
413 pages += 1;
416 if (pages < kMinZoneSize)
417 pages = kMinZoneSize;
420 if (pages >
521 iovmFree(ppnum_t addr, IOItemCount pages) argument
[all...]
/macosx-10.5.8/xnu-1228.15.4/osfmk/i386/
H A Dhibernate_i386.c57 uint32_t pages, page_count; local
123 pages = dram_ranges[bank].last_page + 1 - dram_ranges[bank].first_page;
124 page_count += pages;
125 size += sizeof(hibernate_bitmap_t) + ((pages + 31) >> 5) * sizeof(uint32_t);
154 // mark pages not to be saved, but available for scratch usage during restore
163 // mark pages not to be saved and not for scratch usage during restore
/macosx-10.5.8/xnu-1228.15.4/osfmk/ppc/
H A Dhibernate_ppc.c62 uint32_t pages, page_count; local
72 pages = pmap_mem_regions[bank].mrEnd + 1 - pmap_mem_regions[bank].mrStart;
73 page_count += pages;
74 size += ((pages + 31) >> 5) * sizeof(uint32_t);
115 HIBLOG("removed hash, pca: %d pages\n", count);
120 // mark pages not to be saved and not for scratch usage during restore
H A Dvmachmon.c134 uint32_t pages = GV_HPAGES; /* Number of pages in the hash table */ local
139 if ((pages * sizeof(addr64_t)) + (pages * sizeof(vm_offset_t)) > freeSize) {
169 for (idx = 0; idx < pages; idx++) {
191 for (idx = 0; idx < pages; idx++) { /* De-allocate any pages we managed to allocate */
1147 ** This function maps a list of pages into various address spaces
1152 ** count - number of pages to release
1188 for(i = 0; i < cnt; i++) { /* Step and release all pages i
[all...]
H A Dskiplists.s51 * pages, which is 64GB of resident physical memory per pmap. Pmaps can be larger than
163 lhz r11,mpBSize(r9) ; get #pages/#segments in block/submap mapping
251 lhz r11,mpBSize(r9) ; get #pages/#segments in block/submap mapping
378 lhz r11,mpBSize(r3) ; get #pages/#segments in block/submap mapping
461 lhz r11,mpBSize(r3) ; get #pages/#segments in block/submap mapping
1070 lhz r27,mpBSize(r26) ; get #pages or #segments
H A Dsavearea_asm.s960 * freed pool pages is returned to our caller.
1013 * r8 = #pages to trim (>0)
1048 ; r8 = #pages to trim (>0)
1121 * r8 = #pages to trim (>0)
1152 ; r8 = #pages to trim (>0)
1299 * r8 = #pages to recover
1331 ; r8 = #pages more we still need to recover
1388 * r8 = #pages to recover
1414 ; r8 = #pages more we still need to recover
H A Dvmachmon_asm.s80 .long EXT(vmm_unmap_all_pages) ; Unmaps all pages from the VM space
100 .long EXT(vmm_map_list32) ; Maps a list of pages - supports 32-bit
102 .long EXT(vmm_unmap_list32) ; Unmaps a list of pages - supports 32-bit
132 .long EXT(vmm_map_list64) ; Map a list of pages into guest address spaces - supports 64-bit
134 .long EXT(vmm_unmap_list64) ; Unmap a list of pages from guest address spaces - supports 64-bit
H A Dlowmem_vectors.s3705 .long 0 ; 5094 Size of trace table (1 - 256 pages)
/macosx-10.5.8/xnu-1228.15.4/tools/tests/affinity/
H A Dsets.c36 * byte. By default the buffers are 1MB (256 pages) in size but this can be
134 " [-p P] Number of pages per buffer (256=1MB)]\n"
318 * Calculate number of buffers of size pages*4096 bytes
342 int pages = 256; /* 1MB */ local
374 pages = atoi(optarg);
401 auto_config(pages, &buffers, &sets);
414 if (pages < 256)
415 mutter(" %dkB bytes per buffer, ", pages * 4);
417 mutter(" %dMB bytes per buffer, ", pages / 256);
420 if (buffers * pages < 25
[all...]
H A Dpool.c131 " [-p P] Number of pages per buffer (256=1MB)]\n"
323 int pages = 256; /* 1MB */ local
353 pages = atoi(optarg);
396 if (pages < 256)
397 mutter(" %dkB bytes per buffer, ", pages * 4);
399 mutter(" %dMB bytes per buffer, ", pages / 256);
402 if (buffers * pages < 256)
403 mutter("(total %dkB)\n", buffers * pages * 4);
405 mutter("(total %dMB)\n", buffers * pages / 256);
438 isize = pages * 409
[all...]
/macosx-10.5.8/xnu-1228.15.4/osfmk/kern/
H A Dhibernate.c179 Bits zero in the bitmaps => needs to be saved. All pages default to be saved,
180 pages known to VM to not need saving are subtracted.
181 Wired pages to be saved are present in page_list_wired, pageable in page_list.
191 uint32_t pages = page_list->page_count; local
194 uint32_t count_wire = pages;
215 pages--;
229 pages--;
241 pages--;
357 hibernate_page_list_setall_machine(page_list, page_list_wired, &pages);
363 HIBLOG("pages
[all...]
/macosx-10.5.8/xnu-1228.15.4/osfmk/vm/
H A Dvm_purgeable.c42 int32_t token_new_pagecount = 0; /* count of pages that will
186 int64_t pages = purgeable_queues[i].new_pages += token_new_pagecount; local
187 assert(pages >= 0);
188 assert(pages <= TOKEN_COUNT_MAX);
189 purgeable_queues[i].new_pages=pages;
227 tokens[token].count, /* num pages on token
283 tokens[queue->token_q_head].count, /* num pages on new
285 token_new_pagecount, /* num pages waiting for
317 int64_t pages = purgeable_queues[i].new_pages += token_new_pagecount; local
318 assert(pages >
[all...]
H A Dvm_resident.c128 vm_page_t pages; member in struct:__anon969
207 * Resident pages that represent real memory
226 * real pages, for example to leave a page with
237 * Fictitious pages don't have a physical address,
245 * Guard pages are not accessible so they don't
262 * affinity to zf pages
277 unsigned int vm_page_purgeable_count = 0; /* # of pages purgeable now */
278 uint64_t vm_page_purged_count = 0; /* total count of purged pages */
305 * that pages are "uninteresting" and should be placed
493 * than the number of physical pages i
2882 vm_page_verify_contiguous( vm_page_t pages, unsigned int npages) argument
3446 vm_page_t pages; local
[all...]
H A Dvm_kern.c123 vm_page_t m, pages; local
163 kr = cpm_allocate(CAST_DOWN(vm_size_t, map_size), &pages, max_pnum, FALSE);
175 m = pages;
176 pages = NEXT_PAGE(m);
214 * KMA_NOPAGEWAIT don't wait for pages if unavailable
218 * grab pages from it... this also implies
261 * Guard pages:
263 * Guard pages are implemented as ficticious pages. By placing guard pages
[all...]
H A Dvm_pageout.h214 extern void vm_page_free_reserve(int pages);
229 extern boolean_t vm_pages_encrypted; /* are there encrypted pages ? */
H A Dvm_debug.c509 * Return an array of virtual pages that are mapped to a task.
514 __DEBUG_ONLY page_address_array_t *pages,
558 *pages = 0;
572 (vm_map_copy_t *)pages);
512 vm_mapped_pages_info( __DEBUG_ONLY vm_map_t map, __DEBUG_ONLY page_address_array_t *pages, __DEBUG_ONLY mach_msg_type_number_t *pages_count) argument
H A Dvm_map.c566 * individually; we guess at most one entry per eight pages in the
568 * or roughly 1900 entries (64K) for a 64M machine with 4K pages.
1092 * out pages (if we implement it
1395 * Force pages from the specified object to be entered into
2136 * old pages here...
2529 vm_page_t m, pages; local
2559 &pages, 0, TRUE)) != KERN_SUCCESS)
2572 * Insert pages into object.
2577 m = pages;
2578 pages
[all...]
H A Dvm_pageout.c116 #ifndef VM_PAGEOUT_BURST_ACTIVE_THROTTLE /* maximum iterations of the active queue to move pages to inactive */
129 #define VM_PAGEOUT_DEADLOCK_RELIEF 100 /* number of pages to move to break deadlock */
133 #define VM_PAGEOUT_INACTIVE_RELIEF 50 /* minimum number of pages to move to the inactive q */
167 * needs to be large enough to give pages on it a chance to be
169 * of active+inactive pages that should be inactive.
213 * only vm-privileged threads can allocate pages. vm-privilege
216 * operation by dipping into the reserved pool of pages.
225 * When we dequeue pages from the inactive list, they are
228 * processors are referencing pages faster than we can turn off
248 queue_head_t pgo_pending; /* laundry pages t
1993 vm_page_free_reserve( int pages) argument
[all...]
/macosx-10.5.8/xnu-1228.15.4/osfmk/default_pager/
H A Ddp_memory_object.c641 * Request must be on a page boundary and a multiple of pages.
661 /* merge with data return. It is the kernel that should decide what pages */
731 /* for some type of pager in the future the pages will have to be */
1040 default_pager_page_t * pages = 0; local
1087 actual = ps_vstruct_allocated_pages(entry, pages, potential);
1098 size = round_page(actual * sizeof * pages);
1103 pages = (default_pager_page_t *)addr;
1104 potential = size / sizeof * pages;
1111 pages[--potential].dpp_offset = 0;
H A Ddp_backing_store.c107 * 0 means no shift to pages, so == 1 page/cluster. 1 would mean
108 * 2 pages/cluster, 2 means 4 pages/cluster, and so on.
121 * async_num_queued: number of pages completed, but not yet
123 * async_requests_out: number of pages of requests not completed.
190 * Total pages free in system
836 /* remove all pages associated with the */
837 /* segment from the list of free pages */
839 /* segment pages will appear to be free */
1696 * new indirect map and copy the old direct pages int
3243 ps_vstruct_allocated_pages( vstruct_t vs, default_pager_page_t *pages, size_t pages_size) argument
[all...]

Completed in 217 milliseconds