Searched refs:PAGE_SIZE_64 (Results 1 - 25 of 32) sorted by relevance
12
/xnu-2782.1.97/osfmk/vm/ |
H A D | vm_map_store.h | 79 VMHE_start= VMHE_map->highest_entry_end + PAGE_SIZE_64; \ 82 VMHE_start = VMHE_map->highest_entry_end + PAGE_SIZE_64; \
|
H A D | vm_kern.c | 308 fill_start += PAGE_SIZE_64; 309 fill_size -= PAGE_SIZE_64; 319 fill_size -= PAGE_SIZE_64; 327 wired_page_count = (int) (fill_size / PAGE_SIZE_64); 328 assert(wired_page_count * PAGE_SIZE_64 == fill_size); 431 pg_offset += PAGE_SIZE_64; 439 for (pg_offset = fill_start; pg_offset < fill_start + fill_size; pg_offset += PAGE_SIZE_64) { 535 page_count = (int) (size / PAGE_SIZE_64); 560 pg_offset += PAGE_SIZE_64) { 637 pg_offset += PAGE_SIZE_64) { [all...] |
H A D | vm_init.c | 156 kmapoff_pgcnt * PAGE_SIZE_64, VM_FLAGS_ANYWHERE) != KERN_SUCCESS)
|
H A D | vm_apple_protect.c | 402 2 * PAGE_SIZE_64, 415 dst_vaddr = CAST_DOWN(vm_offset_t, kernel_mapping + PAGE_SIZE_64); 532 kernel_mapping + PAGE_SIZE_64, 589 (addr64_t) (kernel_mapping + (2 * PAGE_SIZE_64))); 666 kernel_mapping + (2 * PAGE_SIZE_64),
|
H A D | vm_compressor_backing_store.c | 305 vm_page_encrypt_counter += (size/PAGE_SIZE_64); 351 vm_page_decrypt_counter += (size/PAGE_SIZE_64); 929 retval = vm_swapfile_io(swf->swp_vp, file_offset, addr, (int)(size / PAGE_SIZE_64), SWAP_READ); 1063 error = vm_swapfile_io(swf->swp_vp, file_offset, addr, (int) (size / PAGE_SIZE_64), SWAP_WRITE); 1452 if (vm_swapfile_io(swf->swp_vp, f_offset, addr, (int)(c_size / PAGE_SIZE_64), SWAP_READ)) {
|
H A D | vm_swapfile_pager.c | 374 PAGE_SIZE_64, 429 (addr64_t) (kernel_mapping + PAGE_SIZE_64)); 465 kernel_mapping + PAGE_SIZE_64,
|
H A D | vm_object.c | 2586 #define CHUNK_SIZE (PAGES_IN_A_CHUNK * PAGE_SIZE_64) /* Size of a chunk in bytes */ 2604 for (p = (len) / PAGE_SIZE_64; p < PAGES_IN_A_CHUNK; p++) \ 2759 for(p = 0; size && CHUNK_NOT_COMPLETE(*chunk_state); p++, size -= PAGE_SIZE_64, offset += PAGE_SIZE_64) { 3134 cur_offset += PAGE_SIZE_64) { 3241 phys_addr += PAGE_SIZE_64) { 3289 start + PAGE_SIZE_64, 3324 start + PAGE_SIZE_64, 3462 src_offset += PAGE_SIZE_64, 3463 new_offset += PAGE_SIZE_64, siz [all...] |
H A D | vm_compressor.h | 255 (compressor_kvspace_used - (compressor_object->resident_page_count * PAGE_SIZE_64)) > compressor_kvwaste_limit) \
|
H A D | vm_fault.c | 278 #define ALIGNED(x) (((x) & (PAGE_SIZE_64 - 1)) == 0) 326 if (offset && last_alloc == offset - PAGE_SIZE_64) { 341 if (last_alloc && last_alloc == offset + PAGE_SIZE_64) { 357 if (offset && last_alloc == (offset - PAGE_SIZE_64)) { 366 } else if (last_alloc && last_alloc == (offset + PAGE_SIZE_64)) { 463 run_offset = 0 - PAGE_SIZE_64; 469 run_offset = PAGE_SIZE_64; 475 { vm_object_offset_t behind = vm_default_behind * PAGE_SIZE_64; 491 pg_offset = PAGE_SIZE_64; 497 pg_offset = 0 - PAGE_SIZE_64; [all...] |
H A D | vm_compressor.c | 246 return (((uint64_t)AVAILABLE_NON_COMPRESSED_MEMORY) * PAGE_SIZE_64); 253 return (c_segment_pages_compressed * PAGE_SIZE_64); 413 if (c_segment_pages_compressed_limit < (uint32_t)(max_mem / PAGE_SIZE_64)) 414 c_segment_pages_compressed_limit = (uint32_t)(max_mem / PAGE_SIZE_64); 1500 total_bytes = compressor_object->resident_page_count * PAGE_SIZE_64;
|
H A D | vm_pageout.c | 1184 (m)->offset + PAGE_SIZE_64, \ 5368 dst_offset += PAGE_SIZE_64; 5956 offset += PAGE_SIZE_64; 5957 new_offset += PAGE_SIZE_64; 6007 offset += PAGE_SIZE_64; 6655 target_offset += PAGE_SIZE_64; 7139 target_offset += PAGE_SIZE_64; 7699 dst_offset += PAGE_SIZE_64; 8022 dst_offset += PAGE_SIZE_64; 8528 map_size -= PAGE_SIZE_64, page_map_offse [all...] |
H A D | memory_object.c | 548 offset += PAGE_SIZE_64) { 611 next_offset = offset + PAGE_SIZE_64;
|
H A D | bsd_vm.c | 456 offset += PAGE_SIZE_64;
|
/xnu-2782.1.97/osfmk/mach/ |
H A D | vm_param.h | 85 #define PAGE_SIZE_64 (unsigned long long)PAGE_SIZE /* pagesize in addr units */ macro
|
/xnu-2782.1.97/bsd/vm/ |
H A D | vm_compressor_backing_file.c | 153 uint64_t io_size = npages * PAGE_SIZE_64;
|
H A D | dp_backing_file.c | 611 *pagesize_p = (vm_size_t)PAGE_SIZE_64;
|
/xnu-2782.1.97/osfmk/kern/ |
H A D | bsd_kern.c | 340 return((uint64_t)pmap_resident_count(map->pmap) * PAGE_SIZE_64); 348 return((uint64_t)pmap_compressed(map->pmap) * PAGE_SIZE_64); 356 return((uint64_t)pmap_resident_max(map->pmap) * PAGE_SIZE_64); 651 * PAGE_SIZE_64;
|
H A D | task.c | 1194 task_page_count = (get_task_phys_footprint(task) / PAGE_SIZE_64); 2574 * PAGE_SIZE_64; 2614 basic_info->resident_size *= PAGE_SIZE_64; 2618 basic_info->resident_size_max *= PAGE_SIZE_64;
|
/xnu-2782.1.97/bsd/vfs/ |
H A D | vfs_cluster.c | 1736 f_offset = (off_t)(r_addr * PAGE_SIZE_64); 1740 ubc_range_op(vp, f_offset, f_offset + PAGE_SIZE_64, UPL_ROP_PRESENT, &size_of_prefetch); 2696 zero_pg_index = (int)((zero_off - upl_f_offset) / PAGE_SIZE_64); 2815 zero_cnt1 = PAGE_SIZE_64 - (zero_off1 & PAGE_MASK_64); 2856 cl.b_addr = (daddr64_t)(upl_f_offset / PAGE_SIZE_64); 2887 upl_f_offset += PAGE_SIZE_64; 3092 cl.e_addr = (daddr64_t)((upl_f_offset + (off_t)upl_size) / PAGE_SIZE_64); 3548 extent.b_addr = uio->uio_offset / PAGE_SIZE_64; 3549 extent.e_addr = (last_request_offset - 1) / PAGE_SIZE_64; 3560 last_ioread_offset = (rap->cl_maxra * PAGE_SIZE_64) [all...] |
/xnu-2782.1.97/bsd/kern/ |
H A D | kern_memorystatus.c | 2100 *footprint = (uint32_t)(get_task_phys_footprint(task) / PAGE_SIZE_64); 2102 *max_footprint = (uint32_t)(get_task_phys_footprint_max(task) / PAGE_SIZE_64); 2105 *max_footprint_lifetime = (uint32_t)(get_task_resident_max(task) / PAGE_SIZE_64); 2108 *purgeable_pages = (uint32_t)(get_task_purgeable_size(task) / PAGE_SIZE_64);
|
H A D | kern_mman.c | 199 if (file_pos + user_size > (vm_object_offset_t)-PAGE_SIZE_64)
|
H A D | mach_loader.c | 865 #define APPLE_UNPROTECTED_HEADER_SIZE (3 * PAGE_SIZE_64)
|
/xnu-2782.1.97/osfmk/kdp/ml/i386/ |
H A D | kdp_x86_common.c | 465 vcur += PAGE_SIZE_64;
|
/xnu-2782.1.97/osfmk/i386/ |
H A D | pmap_x86_common.c | 979 cpte++, vaddr += PAGE_SIZE_64) { 1030 cpte++, vaddr += PAGE_SIZE_64) {
|
/xnu-2782.1.97/bsd/hfs/ |
H A D | hfs_readwrite.c | 604 inval_start = (fp->ff_size + (PAGE_SIZE_64 - 1)) & ~PAGE_MASK_64; 662 tail_off = (writelimit + (PAGE_SIZE_64 - 1)) & ~PAGE_MASK_64; 3418 zero_limit = (fp->ff_size + (PAGE_SIZE_64 - 1)) & ~PAGE_MASK_64;
|
Completed in 257 milliseconds
12