/xnu-2782.1.97/osfmk/i386/ |
H A D | io_map.c | 81 virtual_avail += round_page(size); 83 (void) pmap_map_bd(start, phys_addr, phys_addr + round_page(size), 88 (void) kmem_alloc_pageable(kernel_map, &start, round_page(size)); 89 (void) pmap_map(start, phys_addr, phys_addr + round_page(size),
|
/xnu-2782.1.97/libsyscall/mach/mach/ |
H A D | vm_page_size.h | 50 #define round_page(x) trunc_page((x) + (vm_page_size - 1)) macro
|
/xnu-2782.1.97/libsyscall/mach/ |
H A D | mach_msg.c | 391 request_alloc = (mach_msg_size_t)round_page(max_size + trailer_alloc); 396 reply_alloc = (mach_msg_size_t)round_page((options & MACH_SEND_TRAILER) ? 428 new_request_alloc = (mach_msg_size_t)round_page(bufRequest->Head.msgh_size + 521 reply_alloc = (mach_msg_size_t)round_page((options & MACH_SEND_TRAILER) ? 533 new_request_alloc = (mach_msg_size_t)round_page(max_size + trailer_alloc); 650 new_request_alloc = (mach_msg_size_t)round_page(bufRequest->Head.msgh_size + 700 reply_alloc = (mach_msg_size_t)round_page((options & MACH_SEND_TRAILER) ? 712 new_request_alloc = (mach_msg_size_t)round_page(max_size + trailer_alloc); 799 new_request_alloc = (mach_msg_size_t)round_page(bufRequest->Head.msgh_size +
|
/xnu-2782.1.97/iokit/Kernel/ |
H A D | IOBufferMemoryDescriptor.cpp | 163 capacity = round_page(capacity); 220 capacity = round_page(capacity); 255 vm_size_t size = round_page(capacity); 479 size = round_page(size); 495 debug_iomallocpageable_size -= round_page(size); 500 if (kInternalFlagPageSized & internalFlags) size = round_page(size);
|
H A D | IODataQueue.cpp | 92 allocSize = round_page(size + DATA_QUEUE_MEMORY_HEADER_SIZE); 135 IOFreeAligned(dataQueue, round_page(dataQueue->queueSize + DATA_QUEUE_MEMORY_HEADER_SIZE));
|
H A D | IOSharedDataQueue.cpp | 92 allocSize = round_page(size + DATA_QUEUE_MEMORY_HEADER_SIZE + DATA_QUEUE_MEMORY_APPENDIX_SIZE); 122 IOFreeAligned(dataQueue, round_page(getQueueSize() + DATA_QUEUE_MEMORY_HEADER_SIZE + DATA_QUEUE_MEMORY_APPENDIX_SIZE));
|
H A D | IOMapper.cpp | 233 size = round_page(size); 255 size = round_page(size);
|
H A D | IOLib.cpp | 443 address = round_page(address); 892 length = round_page(address + length) - trunc_page( address );
|
/xnu-2782.1.97/libkern/kxld/ |
H A D | kxld_seg.c | 198 seg->link_addr = round_page(hdrsize); 212 seg->vmsize = round_page(sect_offset) - seg->link_addr; 423 le->link_addr = round_page(seg->link_addr + seg->vmsize); 514 return round_page(size); 575 *data_offset = round_page(*data_offset); 761 seg->vmsize = round_page(maxaddr + maxsize - seg->base_addr); 831 seg->vmsize = round_page(size);
|
H A D | kxld_util.c | 162 size = round_page(size); 189 bytes_allocated += round_page(size); 201 size = round_page(size); 239 size = round_page(size); 259 bytes_freed += round_page(size);
|
H A D | kxld_array.c | 155 array->pool_capacity = round_page(array->itemsize * nitems); 225 needed_capacity = round_page(srcarray->nitems * srcarray->itemsize);
|
/xnu-2782.1.97/osfmk/kern/ |
H A D | gzalloc.c | 144 vm_size_t gzfcsz = round_page(sizeof(*z->gz.gzfc) * gzfc_size); 268 vm_offset_t rounded_size = round_page(zone->elem_size + GZHEADER_SIZE); 345 vm_offset_t rounded_size = round_page(zone->elem_size + GZHEADER_SIZE);
|
H A D | stack.c | 174 if (kernel_stack_size < round_page(kernel_stack_size)) 589 *residentp = *spacep = total * round_page(kernel_stack_size);
|
H A D | btlog.c | 118 buffersize_needed = round_page(buffersize_needed);
|
H A D | zalloc.c | 1691 alloc = round_page(alloc); 1692 max = round_page(max); 1863 alloc_size = round_page(z->elem_size); 2022 zdata_size = round_page(zdata_size); 2048 size = round_page(size); 2187 zone_max = zone_min + round_page(max_zonemap_size); 2257 vm_size_t second_level_size = round_page(zone_page_table_second_level_size * sizeof(struct zone_page_table_entry)); 2456 round_page(zone->elem_size); 3879 names_size = round_page(max_zones * sizeof *names); 3886 info_size = round_page(max_zone [all...] |
/xnu-2782.1.97/bsd/kern/ |
H A D | mach_fat.c | 120 size = round_page(end_of_archs);
|
H A D | kern_core.c | 356 foffset = round_page(header_size); /* offset into file */
|
H A D | kern_sysctl.c | 1351 arg_size = round_page(p->p_argslen); 1384 ret = kmem_alloc(kernel_map, ©_start, round_page(arg_size)); 1390 copy_end = round_page(copy_start + arg_size); 1396 round_page(arg_size)); 1410 round_page(arg_size));
|
/xnu-2782.1.97/osfmk/mach/i386/ |
H A D | vm_param.h | 249 #define round_i386_to_vm(p) (atop(round_page(i386_ptob(p))))
|
/xnu-2782.1.97/libkern/ |
H A D | OSKextLib.cpp | 343 kmem_free(kernel_map, (vm_offset_t)response, round_page(responseLength)); 347 kmem_free(kernel_map, (vm_offset_t)logData, round_page(logDataLength));
|
/xnu-2782.1.97/libkern/kxld/tests/ |
H A D | kxld_array_test.c | 120 titems = (u_int) (round_page(titems * sizeof(u_int)) / sizeof(u_int));
|
/xnu-2782.1.97/osfmk/mach/ |
H A D | vm_param.h | 130 #define round_page(x) (((vm_offset_t)(x) + PAGE_MASK) & ~((vm_offset_t)PAGE_MASK)) macro
|
/xnu-2782.1.97/bsd/dev/ |
H A D | unix_startup.c | 137 size = round_page(size);
|
/xnu-2782.1.97/osfmk/i386/AT386/ |
H A D | model_dep.c | 525 pmap_map_bd(vm_addr, phys_addr, phys_addr + round_page(vm_size), 608 pmap_map(vm_addr, phys_addr, phys_addr + round_page(vm_size),
|
/xnu-2782.1.97/libkern/c++/ |
H A D | OSKext.cpp | 429 u_long roundSize = round_page(size); 3095 kmem_result = kmem_alloc(kernel_map, (vm_offset_t *)&buffer, round_page(logInfoLength)); 3106 (round_page(logInfoLength) - logInfoLength)); 4772 new_kextsize = round_page(kmod_info->size - reloc_size); 4802 linkeditSeg->vmsize = round_page(linkeditSeg->vmsize - reloc_size); 5224 if (round_page(kmod_info->address + kmod_info->size) != 5225 round_page(linkedit->vmaddr + linkedit->vmsize)) 5232 linkeditsize = round_page(linkedit->vmsize); 5418 start = round_page(seg->vmaddr); 5601 address < round_page(se [all...] |