• Home
  • History
  • Annotate
  • Raw
  • Download
  • only in /macosx-10.5.8/xnu-1228.15.4/osfmk/vm/

Lines Matching refs:object

97 	register vm_object_t		object,
102 register vm_object_t object,
117 vm_object_t object;
138 * Allocate a new object (if necessary) and the reference we
143 object = kernel_object;
144 vm_object_reference(object);
146 object = vm_object_allocate(map_size);
151 vm_object_deallocate(object);
155 entry->object.vm_object = object;
156 entry->offset = offset = (object == kernel_object) ?
159 /* Take an extra object ref in case the map entry gets deleted */
160 vm_object_reference(object);
168 vm_object_deallocate(object);
173 vm_object_lock(object);
179 vm_page_insert(m, object, offset + i);
181 vm_object_unlock(object);
186 if (object == kernel_object) {
187 vm_object_lock(object);
188 vm_object_page_remove(object, offset, offset + map_size);
189 vm_object_unlock(object);
193 vm_object_deallocate(object);
196 vm_object_deallocate(object);
198 if (object == kernel_object)
230 vm_object_t object;
298 * Allocate a new object (if necessary). We must do this before
302 object = kernel_object;
303 vm_object_reference(object);
305 object = vm_object_allocate(map_size);
312 vm_object_deallocate(object);
316 entry->object.vm_object = object;
317 entry->offset = offset = (object == kernel_object) ?
320 vm_object_reference(object);
323 vm_object_lock(object);
333 mem = vm_page_alloc_guard(object, offset + i);
340 vm_object_unlock(object);
342 vm_object_lock(object);
355 mem = vm_page_alloclo(object, offset + i);
357 mem = vm_page_alloc(object, offset + i);
366 vm_object_unlock(object);
368 vm_object_lock(object);
381 mem = vm_page_alloc_guard(object, offset + i);
388 vm_object_unlock(object);
390 vm_object_lock(object);
394 vm_object_unlock(object);
399 vm_object_lock(object);
404 vm_object_deallocate(object);
405 if (object == kernel_object)
415 if (object == kernel_object)
416 vm_object_page_remove(object, offset, offset + i);
417 vm_object_unlock(object);
419 vm_object_deallocate(object);
459 vm_object_t object;
478 * Find the VM object backing the old region.
485 object = oldentry->object.vm_object;
488 * Increase the size of the object and
492 vm_object_reference(object);
493 /* by grabbing the object lock before unlocking the map */
496 vm_object_lock(object);
498 if (object->size != oldmapsize)
500 object->size = newmapsize;
501 vm_object_unlock(object);
504 /* object is still not mapped */
505 kmem_alloc_pages(object, vm_object_round_page(oldmapsize),
515 vm_object_lock(object);
518 if ((mem = vm_page_lookup(object, offset)) != VM_PAGE_NULL) {
524 object->size = oldmapsize;
525 vm_object_unlock(object);
526 vm_object_deallocate(object);
529 newentry->object.vm_object = object;
536 vm_object_reference(object);
542 vm_object_lock(object);
544 if ((mem = vm_page_lookup(object, offset)) != VM_PAGE_NULL) {
550 object->size = oldmapsize;
551 vm_object_unlock(object);
552 vm_object_deallocate(object);
555 vm_object_deallocate(object);
660 * Allocate new pages in an object.
665 register vm_object_t object,
672 vm_object_lock(object);
681 (mem = vm_page_alloc(object, offset))) {
682 vm_object_unlock(object);
684 vm_object_lock(object);
691 vm_object_unlock(object);
696 * Remap wired pages in an object into a new region.
697 * The object is assumed to be mapped into the kernel map or
702 register vm_object_t object,
723 vm_object_lock(object);
728 if ((mem = vm_page_lookup(object, offset)) == VM_PAGE_NULL)
737 vm_object_unlock(object);
752 ((unsigned int)(mem->object->wimg_bits))
793 * Need reference on submap object because it is internal
993 if (entry->object.vm_object == VM_OBJECT_NULL) {
1001 vm_map_lock(entry->object.sub_map);
1002 map = entry->object.sub_map;
1007 obj = entry->object.vm_object;