/xnu-2422.115.4/osfmk/i386/vmx/ |
H A D | vmx_shims.c | 42 ret = kmem_alloc_kobject(kernel_map, (vm_offset_t *)&pptr, PAGE_SIZE); 44 bzero(pptr, PAGE_SIZE); 57 kmem_free(kernel_map, (vm_offset_t)va, PAGE_SIZE);
|
/xnu-2422.115.4/tools/tests/superpages/ |
H A D | measure_tlbs.c | 28 mach_vm_size_t size = SUPERPAGE_ROUND_UP(pages*PAGE_SIZE); /* allocate full superpages */ 40 sum += data[p*PAGE_SIZE]; 49 data[p*PAGE_SIZE] = run & 0xFF; 55 sum += data[p*PAGE_SIZE]; 62 for (run = 0; run < RUNS1/PAGE_SIZE; run++) { 63 for (i = 0; i<pages*PAGE_SIZE; i++) { 68 for (run = 0; run < RUNS1/PAGE_SIZE; run++) { 69 for (i = 0; i<pages*PAGE_SIZE; i++) { 83 p2 = p % (pages*PAGE_SIZE/NODE_SIZE); 92 p2 = p % (pages*PAGE_SIZE/NODE_SIZ [all...] |
H A D | testsp.c | 108 for (i=0; i<size/PAGE_SIZE; i++) { 109 if ((ret = check_r(addr+i*PAGE_SIZE, PAGE_SIZE, res))) { 138 for (i=0; i<size/PAGE_SIZE; i++) { 139 if ((ret = check_w(addr+i*PAGE_SIZE, PAGE_SIZE))) { 221 mach_vm_size_t size = 2*PAGE_SIZE; /* will be rounded up to some superpage size */ 288 mach_vm_size_t size = PAGE_SIZE; /* != 2 MB */ 313 kr = mach_vm_deallocate(mach_task_self(), addr + PAGE_SIZE, size); 334 for (i=0; i<SUPERPAGE_SIZE/PAGE_SIZE; [all...] |
/xnu-2422.115.4/osfmk/ipc/ |
H A D | ipc_table.c | 111 (index < num) && (size < PAGE_SIZE); 121 for (incrsize = PAGE_SIZE; index < num;) { 132 if (incrsize < (PAGE_SIZE << 3)) 181 if (size < PAGE_SIZE) 205 if (size < PAGE_SIZE)
|
/xnu-2422.115.4/osfmk/x86_64/ |
H A D | boot_pt.c | 49 [0] = ((uint64_t)(PAGE_SIZE) | PML4_PROT), 50 [KERNEL_PML4_INDEX] = ((uint64_t)(PAGE_SIZE) | PML4_PROT), 56 [0] = ((uint64_t)(2*PAGE_SIZE) | PDPT_PROT), 57 [1] = ((uint64_t)(3*PAGE_SIZE) | PDPT_PROT), 58 [2] = ((uint64_t)(4*PAGE_SIZE) | PDPT_PROT), 59 [3] = ((uint64_t)(5*PAGE_SIZE) | PDPT_PROT),
|
H A D | copyio.c | 300 lop = (unsigned int)(PAGE_SIZE - (snk64 & (PAGE_SIZE - 1))); /* Assume sink smallest */ 302 if (lop > (unsigned int)(PAGE_SIZE - (src64 & (PAGE_SIZE - 1)))) 303 lop = (unsigned int)(PAGE_SIZE - (src64 & (PAGE_SIZE - 1))); /* No, source is smaller */ 311 lop = (unsigned int)(PAGE_SIZE - (src64 & (PAGE_SIZE - 1))); 313 lop = (unsigned int)(PAGE_SIZE - (snk64 & (PAGE_SIZE [all...] |
H A D | lowmem_vectors.c | 76 lowglo lowGlo __attribute__ ((aligned(PAGE_SIZE))) = {
|
H A D | pmap.c | 271 ps = PAGE_SIZE; 637 _vm_object_allocate((vm_object_size_t)NPML4PGS * PAGE_SIZE, &kpml4obj_object_store); 640 _vm_object_allocate((vm_object_size_t)NPDPTPGS * PAGE_SIZE, &kpdptobj_object_store); 643 _vm_object_allocate((vm_object_size_t)NPDEPGS * PAGE_SIZE, &kptobj_object_store); 721 vaddr += PAGE_SIZE; 722 vsize -= PAGE_SIZE; 732 pmap_anchor_zone = zinit(PAGE_SIZE, task_max, PAGE_SIZE, "pagetable anchors"); 757 vaddr += PAGE_SIZE; 861 PMAP_ZINFO_PALLOC(kernel_pmap, bootstrap_wired_pages * PAGE_SIZE); [all...] |
/xnu-2422.115.4/libsyscall/mach/ |
H A D | mach_init.c | 75 vm_size_t vm_page_size = PAGE_SIZE; 90 *out_page_size = PAGE_SIZE;
|
/xnu-2422.115.4/bsd/vm/ |
H A D | vnode_pager.c | 366 for (offset = upl_offset; isize; isize -= PAGE_SIZE, offset += PAGE_SIZE) { 379 ubc_upl_abort_range(upl, offset, PAGE_SIZE, UPL_ABORT_FREE_ON_EMPTY); 385 ubc_upl_commit_range(upl, offset, PAGE_SIZE, UPL_COMMIT_FREE_ON_EMPTY); 387 f_offset += PAGE_SIZE; 397 base_index = upl_offset / PAGE_SIZE; 399 for (pg_index = (upl_offset + isize) / PAGE_SIZE; pg_index > base_index;) { 413 isize = ((pg_index + 1) - base_index) * PAGE_SIZE; 428 f_offset += PAGE_SIZE; 429 offset += PAGE_SIZE; [all...] |
/xnu-2422.115.4/bsd/dev/i386/ |
H A D | mem.c | 172 size= PAGE_SIZE; 178 c = min(PAGE_SIZE - o, uio_curriovlen(uio)); 180 kmem_free(kernel_map, where, PAGE_SIZE); 210 MALLOC(devzerobuf, caddr_t,PAGE_SIZE, M_TEMP, M_WAITOK); 211 bzero(devzerobuf, PAGE_SIZE); 217 c = min(uio_curriovlen(uio), PAGE_SIZE);
|
/xnu-2422.115.4/bsd/sys/ |
H A D | dtrace_ptss.h | 75 #define DTRACE_PTSS_ENTRIES_PER_PAGE (PAGE_SIZE / DTRACE_PTSS_SCRATCH_SPACE_PER_THREAD)
|
H A D | pipe.h | 81 #define SMALL_PIPE_SIZE PAGE_SIZE 92 #define PIPENPAGES (BIG_PIPE_SIZE / PAGE_SIZE + 1)
|
/xnu-2422.115.4/bsd/dev/dtrace/ |
H A D | dtrace_alloc.c | 90 if (PAGE_SIZE < 16*1024) 93 dtrace_alloc_max = PAGE_SIZE;
|
/xnu-2422.115.4/osfmk/vm/ |
H A D | vm_compressor_pager.c | 206 assert(pager_page_size == PAGE_SIZE); 405 if ((uint32_t)(offset/PAGE_SIZE) != (offset/PAGE_SIZE)) { 422 if (offset / PAGE_SIZE > pager->cpgr_num_slots) { 461 cur_offset += PAGE_SIZE) { 516 if ((uint32_t)(new_size/PAGE_SIZE) != (new_size/PAGE_SIZE)) { 531 pager->cpgr_num_slots = (uint32_t)(new_size/PAGE_SIZE); 586 page_num = (uint32_t)(offset/PAGE_SIZE); 587 if (page_num != (offset/PAGE_SIZE)) { [all...] |
H A D | vm_compressor.c | 80 #define UNPACK_C_SIZE(cs) ((cs->c_size == (PAGE_SIZE-1)) ? 4096 : cs->c_size) 81 #define PACK_C_SIZE(cs, size) (cs->c_size = ((size == PAGE_SIZE) ? PAGE_SIZE - 1 : size)) 149 #define C_SEGMENTS_PER_PAGE (PAGE_SIZE / sizeof(union c_segu)) 310 assert((C_SEGMENTS_PER_PAGE * sizeof(union c_segu)) == PAGE_SIZE); 358 c_segment_pages_compressed_limit = (uint32_t)((max_mem / PAGE_SIZE)) * vm_scale; 374 if ((c_segments_limit = c_segment_pages_compressed_limit / (C_SEG_BUFSIZE / PAGE_SIZE)) > C_SEG_MAX_LIMIT) 764 pages_populated = (round_page_32(C_SEG_OFFSET_TO_BYTES(c_seg->c_populated_offset))) / PAGE_SIZE; 779 kernel_memory_depopulate(kernel_map, (vm_offset_t) c_buffer, pages_populated * PAGE_SIZE, KMA_COMPRESSOR); local 831 c_seg->c_populated_offset = (c_offset + (C_SEG_BYTES_TO_OFFSET(PAGE_SIZE) 1047 PAGE_SIZE, local 2309 PAGE_SIZE, local 2555 kernel_memory_depopulate(kernel_map, (vm_offset_t) c_seg->c_store.c_buffer, pages_populated * PAGE_SIZE, KMA_COMPRESSOR); local 2627 pmap_ledger_debit(current_task()->map->pmap, task_ledgers.phys_footprint, PAGE_SIZE); local [all...] |
/xnu-2422.115.4/osfmk/mach/i386/ |
H A D | vm_param.h | 104 #define PAGE_SIZE I386_PGBYTES macro 106 #define PAGE_MASK (PAGE_SIZE - 1) 263 ((addr) < (VM_MAX_USER_PAGE_ADDRESS + PAGE_SIZE))
|
/xnu-2422.115.4/osfmk/i386/ |
H A D | gdt.c | 67 __attribute__((aligned(PAGE_SIZE))) = {
|
H A D | pmap_x86_common.c | 383 PMAP_UPDATE_TLBS(pmap, vaddr, vaddr + PAGE_SIZE); 606 pmap_ledger_credit(pmap, task_ledgers.wired_mem, PAGE_SIZE); 612 pmap_ledger_debit(pmap, task_ledgers.wired_mem, PAGE_SIZE); 668 PMAP_UPDATE_TLBS(pmap, vaddr, vaddr + PAGE_SIZE); 677 pmap_ledger_debit(pmap, task_ledgers.phys_mem, PAGE_SIZE); 678 pmap_ledger_debit(pmap, task_ledgers.phys_footprint, PAGE_SIZE); 697 PAGE_SIZE); 725 pmap_ledger_debit(pmap, task_ledgers.wired_mem, PAGE_SIZE); 816 pmap_ledger_credit(pmap, task_ledgers.phys_mem, PAGE_SIZE); 817 pmap_ledger_credit(pmap, task_ledgers.phys_footprint, PAGE_SIZE); [all...] |
H A D | phys.c | 100 bzero_phys((addr64_t)i386_ptob(pn), PAGE_SIZE); 115 assert(offset + len <= PAGE_SIZE); 140 assert((((uintptr_t)dst & PAGE_MASK) + dst_offset + len) <= PAGE_SIZE); 141 assert((((uintptr_t)src & PAGE_MASK) + src_offset + len) <= PAGE_SIZE); 162 assert((dst_offset + len) <= PAGE_SIZE); 180 assert((src_offset + len) <= PAGE_SIZE);
|
H A D | pmap.h | 128 #define NPTEPG (PAGE_SIZE/(sizeof (pt_entry_t))) 129 #define NPTDPG (PAGE_SIZE/(sizeof (pd_entry_t))) 133 #define NPDEPG (PAGE_SIZE/(sizeof (pd_entry_t))) 141 #define NPML4PG (PAGE_SIZE/(sizeof (pml4_entry_t))) 149 #define NPDPTPG (PAGE_SIZE/(sizeof (pdpt_entry_t))) 157 #define NPDPG (PAGE_SIZE/(sizeof (pd_entry_t))) 165 #define NPTPG (PAGE_SIZE/(sizeof (pt_entry_t))) 213 #define NPML4PGS (1ULL * (PAGE_SIZE/(sizeof (pml4_entry_t)))) 214 #define NPDPTPGS (NPML4PGS * (PAGE_SIZE/(sizeof (pdpt_entry_t)))) 215 #define NPDEPGS (NPDPTPGS * (PAGE_SIZE/(sizeo [all...] |
H A D | mp_desc.h | 87 uint8_t fstk[PAGE_SIZE] __attribute__ ((aligned (16)));
|
/xnu-2422.115.4/osfmk/kern/ |
H A D | gzalloc.c | 283 gzalloc_reserve += rounded_size + PAGE_SIZE; 284 gzalloc_reserve_size -= rounded_size + PAGE_SIZE; 289 &gzaddr, rounded_size + (1*PAGE_SIZE), 297 gzaddr += PAGE_SIZE; 352 saddr = (vm_offset_t) addr - PAGE_SIZE; 394 saddr + rounded_size + (1 * PAGE_SIZE), 427 free_addr + rounded_size + (1 * PAGE_SIZE),
|
/xnu-2422.115.4/bsd/vfs/ |
H A D | vfs_cluster.c | 124 #define MAX_VECTOR_UPL_SIZE (2 * MAX_UPL_SIZE) * PAGE_SIZE 236 #define MAX_IO_CONTIG_SIZE (MAX_UPL_SIZE * PAGE_SIZE) 238 #define MIN_DIRECT_WRITE_SIZE (4 * PAGE_SIZE) 320 max_io_size = min((segcnt * PAGE_SIZE), maxcnt); 322 if (max_io_size < (MAX_UPL_TRANSFER * PAGE_SIZE)) { 326 max_io_size = (MAX_UPL_TRANSFER * PAGE_SIZE); 329 * make sure the size specified is a multiple of PAGE_SIZE 469 if (VNOP_BLOCKMAP(vp, f_offset, PAGE_SIZE, &blkno, &io_size, NULL, VNODE_READ, NULL)) 671 cluster_zero(upl, zero_offset, PAGE_SIZE - (zero_offset & PAGE_MASK), real_bp); 706 commit_size = (pg_offset + transaction_size + (PAGE_SIZE [all...] |
/xnu-2422.115.4/libkern/kxld/tests/ |
H A D | kxld_array_test.c | 39 #define kNumStorageTestItems (u_int) (4 * PAGE_SIZE / sizeof(u_int)) 60 titems = PAGE_SIZE / sizeof(u_int); 86 titems = 2 * PAGE_SIZE / sizeof(u_int) + 100;
|