Searched refs:PAGE_SIZE_64 (Results 1 - 25 of 32) sorted by relevance

12

/xnu-2782.1.97/osfmk/vm/
H A Dvm_map_store.h79 VMHE_start= VMHE_map->highest_entry_end + PAGE_SIZE_64; \
82 VMHE_start = VMHE_map->highest_entry_end + PAGE_SIZE_64; \
H A Dvm_kern.c308 fill_start += PAGE_SIZE_64;
309 fill_size -= PAGE_SIZE_64;
319 fill_size -= PAGE_SIZE_64;
327 wired_page_count = (int) (fill_size / PAGE_SIZE_64);
328 assert(wired_page_count * PAGE_SIZE_64 == fill_size);
431 pg_offset += PAGE_SIZE_64;
439 for (pg_offset = fill_start; pg_offset < fill_start + fill_size; pg_offset += PAGE_SIZE_64) {
535 page_count = (int) (size / PAGE_SIZE_64);
560 pg_offset += PAGE_SIZE_64) {
637 pg_offset += PAGE_SIZE_64) {
[all...]
H A Dvm_init.c156 kmapoff_pgcnt * PAGE_SIZE_64, VM_FLAGS_ANYWHERE) != KERN_SUCCESS)
H A Dvm_apple_protect.c402 2 * PAGE_SIZE_64,
415 dst_vaddr = CAST_DOWN(vm_offset_t, kernel_mapping + PAGE_SIZE_64);
532 kernel_mapping + PAGE_SIZE_64,
589 (addr64_t) (kernel_mapping + (2 * PAGE_SIZE_64)));
666 kernel_mapping + (2 * PAGE_SIZE_64),
H A Dvm_compressor_backing_store.c305 vm_page_encrypt_counter += (size/PAGE_SIZE_64);
351 vm_page_decrypt_counter += (size/PAGE_SIZE_64);
929 retval = vm_swapfile_io(swf->swp_vp, file_offset, addr, (int)(size / PAGE_SIZE_64), SWAP_READ);
1063 error = vm_swapfile_io(swf->swp_vp, file_offset, addr, (int) (size / PAGE_SIZE_64), SWAP_WRITE);
1452 if (vm_swapfile_io(swf->swp_vp, f_offset, addr, (int)(c_size / PAGE_SIZE_64), SWAP_READ)) {
H A Dvm_swapfile_pager.c374 PAGE_SIZE_64,
429 (addr64_t) (kernel_mapping + PAGE_SIZE_64));
465 kernel_mapping + PAGE_SIZE_64,
H A Dvm_object.c2586 #define CHUNK_SIZE (PAGES_IN_A_CHUNK * PAGE_SIZE_64) /* Size of a chunk in bytes */
2604 for (p = (len) / PAGE_SIZE_64; p < PAGES_IN_A_CHUNK; p++) \
2759 for(p = 0; size && CHUNK_NOT_COMPLETE(*chunk_state); p++, size -= PAGE_SIZE_64, offset += PAGE_SIZE_64) {
3134 cur_offset += PAGE_SIZE_64) {
3241 phys_addr += PAGE_SIZE_64) {
3289 start + PAGE_SIZE_64,
3324 start + PAGE_SIZE_64,
3462 src_offset += PAGE_SIZE_64,
3463 new_offset += PAGE_SIZE_64, siz
[all...]
H A Dvm_compressor.h255 (compressor_kvspace_used - (compressor_object->resident_page_count * PAGE_SIZE_64)) > compressor_kvwaste_limit) \
H A Dvm_fault.c278 #define ALIGNED(x) (((x) & (PAGE_SIZE_64 - 1)) == 0)
326 if (offset && last_alloc == offset - PAGE_SIZE_64) {
341 if (last_alloc && last_alloc == offset + PAGE_SIZE_64) {
357 if (offset && last_alloc == (offset - PAGE_SIZE_64)) {
366 } else if (last_alloc && last_alloc == (offset + PAGE_SIZE_64)) {
463 run_offset = 0 - PAGE_SIZE_64;
469 run_offset = PAGE_SIZE_64;
475 { vm_object_offset_t behind = vm_default_behind * PAGE_SIZE_64;
491 pg_offset = PAGE_SIZE_64;
497 pg_offset = 0 - PAGE_SIZE_64;
[all...]
H A Dvm_compressor.c246 return (((uint64_t)AVAILABLE_NON_COMPRESSED_MEMORY) * PAGE_SIZE_64);
253 return (c_segment_pages_compressed * PAGE_SIZE_64);
413 if (c_segment_pages_compressed_limit < (uint32_t)(max_mem / PAGE_SIZE_64))
414 c_segment_pages_compressed_limit = (uint32_t)(max_mem / PAGE_SIZE_64);
1500 total_bytes = compressor_object->resident_page_count * PAGE_SIZE_64;
H A Dvm_pageout.c1184 (m)->offset + PAGE_SIZE_64, \
5368 dst_offset += PAGE_SIZE_64;
5956 offset += PAGE_SIZE_64;
5957 new_offset += PAGE_SIZE_64;
6007 offset += PAGE_SIZE_64;
6655 target_offset += PAGE_SIZE_64;
7139 target_offset += PAGE_SIZE_64;
7699 dst_offset += PAGE_SIZE_64;
8022 dst_offset += PAGE_SIZE_64;
8528 map_size -= PAGE_SIZE_64, page_map_offse
[all...]
H A Dmemory_object.c548 offset += PAGE_SIZE_64) {
611 next_offset = offset + PAGE_SIZE_64;
H A Dbsd_vm.c456 offset += PAGE_SIZE_64;
/xnu-2782.1.97/osfmk/mach/
H A Dvm_param.h85 #define PAGE_SIZE_64 (unsigned long long)PAGE_SIZE /* pagesize in addr units */ macro
/xnu-2782.1.97/bsd/vm/
H A Dvm_compressor_backing_file.c153 uint64_t io_size = npages * PAGE_SIZE_64;
H A Ddp_backing_file.c611 *pagesize_p = (vm_size_t)PAGE_SIZE_64;
/xnu-2782.1.97/osfmk/kern/
H A Dbsd_kern.c340 return((uint64_t)pmap_resident_count(map->pmap) * PAGE_SIZE_64);
348 return((uint64_t)pmap_compressed(map->pmap) * PAGE_SIZE_64);
356 return((uint64_t)pmap_resident_max(map->pmap) * PAGE_SIZE_64);
651 * PAGE_SIZE_64;
H A Dtask.c1194 task_page_count = (get_task_phys_footprint(task) / PAGE_SIZE_64);
2574 * PAGE_SIZE_64;
2614 basic_info->resident_size *= PAGE_SIZE_64;
2618 basic_info->resident_size_max *= PAGE_SIZE_64;
/xnu-2782.1.97/bsd/vfs/
H A Dvfs_cluster.c1736 f_offset = (off_t)(r_addr * PAGE_SIZE_64);
1740 ubc_range_op(vp, f_offset, f_offset + PAGE_SIZE_64, UPL_ROP_PRESENT, &size_of_prefetch);
2696 zero_pg_index = (int)((zero_off - upl_f_offset) / PAGE_SIZE_64);
2815 zero_cnt1 = PAGE_SIZE_64 - (zero_off1 & PAGE_MASK_64);
2856 cl.b_addr = (daddr64_t)(upl_f_offset / PAGE_SIZE_64);
2887 upl_f_offset += PAGE_SIZE_64;
3092 cl.e_addr = (daddr64_t)((upl_f_offset + (off_t)upl_size) / PAGE_SIZE_64);
3548 extent.b_addr = uio->uio_offset / PAGE_SIZE_64;
3549 extent.e_addr = (last_request_offset - 1) / PAGE_SIZE_64;
3560 last_ioread_offset = (rap->cl_maxra * PAGE_SIZE_64)
[all...]
/xnu-2782.1.97/bsd/kern/
H A Dkern_memorystatus.c2100 *footprint = (uint32_t)(get_task_phys_footprint(task) / PAGE_SIZE_64);
2102 *max_footprint = (uint32_t)(get_task_phys_footprint_max(task) / PAGE_SIZE_64);
2105 *max_footprint_lifetime = (uint32_t)(get_task_resident_max(task) / PAGE_SIZE_64);
2108 *purgeable_pages = (uint32_t)(get_task_purgeable_size(task) / PAGE_SIZE_64);
H A Dkern_mman.c199 if (file_pos + user_size > (vm_object_offset_t)-PAGE_SIZE_64)
H A Dmach_loader.c865 #define APPLE_UNPROTECTED_HEADER_SIZE (3 * PAGE_SIZE_64)
/xnu-2782.1.97/osfmk/kdp/ml/i386/
H A Dkdp_x86_common.c465 vcur += PAGE_SIZE_64;
/xnu-2782.1.97/osfmk/i386/
H A Dpmap_x86_common.c979 cpte++, vaddr += PAGE_SIZE_64) {
1030 cpte++, vaddr += PAGE_SIZE_64) {
/xnu-2782.1.97/bsd/hfs/
H A Dhfs_readwrite.c604 inval_start = (fp->ff_size + (PAGE_SIZE_64 - 1)) & ~PAGE_MASK_64;
662 tail_off = (writelimit + (PAGE_SIZE_64 - 1)) & ~PAGE_MASK_64;
3418 zero_limit = (fp->ff_size + (PAGE_SIZE_64 - 1)) & ~PAGE_MASK_64;

Completed in 344 milliseconds

12