Lines Matching defs:round*
144 vm_size_t gzfcsz = round_page(sizeof(*z->gz.gzfc) * gzfc_size);
268 vm_offset_t rounded_size = round_page(zone->elem_size + GZHEADER_SIZE);
269 vm_offset_t residue = rounded_size - zone->elem_size;
277 if (gzalloc_reserve_size < rounded_size)
283 gzalloc_reserve += rounded_size + PAGE_SIZE;
284 gzalloc_reserve_size -= rounded_size + PAGE_SIZE;
285 OSAddAtomic64((SInt32) (rounded_size), &gzalloc_early_alloc);
289 &gzaddr, rounded_size + (1*PAGE_SIZE),
292 panic("gzalloc: kernel_memory_allocate for size 0x%llx failed with %d", (uint64_t)rounded_size, kr);
309 * data use. Since the element size may be "rounded up"
312 * size and the rounded size via visual inspection.
314 * prior to rounding by kalloc/IOMalloc etc.
318 memset((void *)gzaddr, gzalloc_fill_pattern, rounded_size);
327 zone->cur_size += rounded_size;
330 OSAddAtomic64((SInt32) rounded_size, &gzalloc_allocated);
331 OSAddAtomic64((SInt32) (rounded_size - zone->elem_size), &gzalloc_wasted);
345 vm_offset_t rounded_size = round_page(zone->elem_size + GZHEADER_SIZE);
346 vm_offset_t residue = rounded_size - zone->elem_size;
379 OSAddAtomic64((SInt32) (rounded_size), &gzalloc_early_free);
394 saddr + rounded_size + (1 * PAGE_SIZE),
418 zone->cur_size -= rounded_size;
427 free_addr + rounded_size + (1 * PAGE_SIZE),
432 OSAddAtomic64((SInt32)rounded_size, &gzalloc_freed);
433 OSAddAtomic64(-((SInt32) (rounded_size - zone->elem_size)), &gzalloc_wasted);