/macosx-10.5.8/xnu-1228.15.4/osfmk/vm/ |
H A D | vm_object.h | 63 * Virtual memory object module definitions. 97 * vm_object_t Virtual memory object. 131 * temporary object that 132 * shadows this object, for 139 vm_object_offset_t paging_offset; /* Offset into memory object */ 146 /* The memory object ports are 162 /* boolean_t */ pager_trusted:1,/* The pager for this object 168 * for this object (and rights 169 * to the memory object) after 186 /* boolean_t */ private:1, /* magic device_pager object, 358 vm_object_t object; /* back pointer */ member in struct:msync_req [all...] |
H A D | vm_object.c | 62 * Virtual memory object module. 101 * page of memory exists within exactly one object. 103 * An object is only deallocated when all "references" 106 * Associated with each object is a list of all resident 107 * memory pages belonging to that object; this list is 108 * maintained by the "vm_page" module, but locked by the object's 111 * Each object also records the memory object reference 113 * back data (the memory object, field "pager"), etc... 117 * memory object int 278 vm_object_t object; /* corresponding object */ member in struct:vm_object_hash_entry 375 _vm_object_allocate( vm_object_size_t size, vm_object_t object) argument 397 register vm_object_t object; local 645 vm_object_deallocate( register vm_object_t object) argument 989 register vm_object_t object = VM_OBJECT_NULL; local 1090 vm_object_terminate( register vm_object_t object) argument 1363 vm_object_reap( vm_object_t object) argument 1499 vm_object_reap_async( vm_object_t object) argument 1519 vm_object_t object, shadow_object; local 1642 vm_object_destroy( vm_object_t object, __unused kern_return_t reason) argument 1718 vm_object_deactivate_all_pages( register vm_object_t object) argument 1764 vm_object_deactivate_pages( vm_object_t object, vm_object_offset_t offset, vm_object_size_t size, boolean_t kill_page) argument 1891 vm_object_pmap_protect( register vm_object_t object, register vm_object_offset_t offset, vm_object_size_t size, pmap_t pmap, vm_map_offset_t pmap_start, vm_prot_t prot) argument 2287 vm_object_t object = *_object; local 2892 vm_object_shadow( vm_object_t *object, vm_object_offset_t *offset, vm_object_size_t length) argument 3057 register vm_object_t object; local 3262 vm_object_pager_create( register vm_object_t object) argument 3380 vm_object_remove( vm_object_t object) argument 3426 vm_object_do_collapse( vm_object_t object, vm_object_t backing_object) argument 3623 vm_object_do_bypass( vm_object_t object, vm_object_t backing_object) argument 3753 vm_object_collapse( register vm_object_t object, register vm_object_offset_t hint_offset, boolean_t can_bypass) argument 4159 vm_object_page_remove( register vm_object_t object, register vm_object_offset_t start, register vm_object_offset_t end) argument 4325 vm_object_page_map( vm_object_t object, vm_object_offset_t offset, vm_object_size_t size, vm_object_offset_t (*map_fn)(void *map_fn_data, vm_object_offset_t offset), void *map_fn_data) argument 4401 vm_object_cached( register vm_object_t object) argument 4445 vm_follow_object( vm_object_t object) argument 4478 vm_object_t object; local 4642 vm_object_find( vm_object_t object) argument 4693 vm_object_populate_with_private( vm_object_t object, vm_object_offset_t offset, ppnum_t phys_page, vm_size_t size) argument 4811 register vm_object_t object = VM_OBJECT_NULL; local 4876 vm_object_t object; local 4929 vm_object_t object; local 5007 vm_object_release_name( vm_object_t object, int flags) argument 5105 vm_object_lock_request( vm_object_t object, vm_object_offset_t offset, vm_object_size_t size, memory_object_return_t should_return, int flags, vm_prot_t prot) argument 5165 vm_object_purge(vm_object_t object) argument 5383 vm_object_purgable_control( vm_object_t object, vm_purgable_t control, int *state) argument 5554 vm_object_res_deallocate( vm_object_t object) argument 5597 vm_object_res_reference( vm_object_t object) argument 5630 vm_object_reference( register vm_object_t object) argument 5993 vm_object_cluster_size(vm_object_t object, vm_object_offset_t *start, vm_size_t *length, vm_object_fault_info_t fault_info) argument 6230 vm_object_page_op( vm_object_t object, vm_object_offset_t offset, int ops, ppnum_t *phys_entry, int *flags) argument 6391 vm_object_range_op( vm_object_t object, vm_object_offset_t offset_beg, vm_object_offset_t offset_end, int ops, int *range) argument 6467 vm_object_lock(vm_object_t object) argument 6477 vm_object_lock_try(vm_object_t object) argument 6487 vm_object_lock_shared(vm_object_t object) argument 6497 vm_object_lock_try_shared(vm_object_t object) argument [all...] |
H A D | vm_fault.c | 112 * an object at least this big. 145 extern void vm_fault_classify(vm_object_t object, 187 * The paging reference for "object" is released. 188 * "object" is unlocked. 190 * freed and the paging reference for the object 194 * "object" must be locked. 198 register vm_object_t object, 201 vm_object_paging_end(object); 202 vm_object_unlock(object); 205 object 197 vm_fault_cleanup( register vm_object_t object, register vm_page_t top_page) argument 255 vm_fault_is_sequential( vm_object_t object, vm_object_offset_t offset, vm_behavior_t behavior) argument 371 vm_fault_deactivate_behind( vm_object_t object, vm_object_offset_t offset, vm_behavior_t behavior) argument 455 vm_fault_check(vm_object_t object, vm_page_t m, vm_page_t first_m, boolean_t interruptible_state) argument 643 vm_object_t object; local 2235 vm_object_t object; /* Top-level object */ local 3466 vm_object_t object; local 3605 vm_object_t object; local 3773 vm_object_t object = page->object; local 3788 vm_object_t object; local 4131 vm_fault_classify(vm_object_t object, vm_object_offset_t offset, vm_prot_t fault_type) argument 4203 vm_object_t object; local 4285 vm_object_t object; local [all...] |
H A D | vm_kern.c | 97 register vm_object_t object, 102 register vm_object_t object, 117 vm_object_t object; local 138 * Allocate a new object (if necessary) and the reference we 143 object = kernel_object; 144 vm_object_reference(object); 146 object = vm_object_allocate(map_size); 151 vm_object_deallocate(object); 155 entry->object.vm_object = object; 230 vm_object_t object; local 459 vm_object_t object; local 664 kmem_alloc_pages( register vm_object_t object, register vm_object_offset_t offset, register vm_object_size_t size) argument 701 kmem_remap_pages( register vm_object_t object, register vm_object_offset_t offset, register vm_offset_t start, register vm_offset_t end, vm_prot_t protection) argument [all...] |
H A D | memory_object.c | 298 #define LIST_REQ_PAGEOUT_PAGES(object, data_cnt, action, po, ro, ioerr, iosync) \ 304 if ((pager = (object)->pager) != MEMORY_OBJECT_NULL) { \ 305 vm_object_paging_begin(object); \ 306 vm_object_unlock(object); \ 322 vm_object_lock(object); \ 323 vm_object_paging_end(object); \ 332 * memory object. For each page in the given range, 363 vm_object_t object; local 376 object = memory_object_control_to_vm_object(control); 377 if (object 424 vm_object_t object; local 447 vm_object_t object; local 481 vm_object_sync( vm_object_t object, vm_object_offset_t offset, vm_object_size_t size, boolean_t should_flush, boolean_t should_return, boolean_t should_iosync) argument 529 vm_object_update_extent( vm_object_t object, vm_object_offset_t offset, vm_object_offset_t offset_end, vm_object_offset_t *offset_resid, int *io_errno, boolean_t should_flush, memory_object_return_t should_return, boolean_t should_iosync, vm_prot_t prot) argument 679 vm_object_update( register vm_object_t object, register vm_object_offset_t offset, register vm_object_size_t size, register vm_object_offset_t *resid_offset, int *io_errno, memory_object_return_t should_return, int flags, vm_prot_t protection) argument 1045 vm_object_t object; local 1088 vm_object_set_attributes_common( vm_object_t object, boolean_t may_cache, memory_object_copy_strategy_t copy_strategy, boolean_t temporary, boolean_t silent_overwrite, boolean_t advisory_pageout) argument 1170 vm_object_t object; local 1319 vm_object_t object; local 1447 vm_object_t object; local 1591 vm_object_t object; local 1628 vm_object_t object; local 1648 vm_object_t object; local 1826 vm_object_t object; local 1853 vm_object_t object; local 1872 vm_object_t object; local 1891 vm_object_t object; local 1917 memory_object_control_allocate( vm_object_t object) argument 1931 memory_object_control_collapse( memory_object_control_t control, vm_object_t object) argument 2016 convert_memory_object_to_port( __unused memory_object_t object) argument [all...] |
H A D | vm_pageout.c | 414 * The object must be locked, and will be returned locked. 418 vm_object_t object) 423 * Deal with the deallocation (last reference) of a pageout object 425 * freeing pages in the original object. 428 assert(object->pageout); 429 shadow_object = object->shadow; 432 while (!queue_empty(&object->memq)) { 436 p = (vm_page_t) queue_first(&object->memq); 448 offset + object->shadow_offset); 490 * Revoke all access to the page. Since the object i 417 vm_pageout_object_terminate( vm_object_t object) argument 668 vm_object_t object; local 769 vm_object_t object = m->object; local 890 vm_object_t object; local 2081 vm_object_t object; local 2481 vm_object_t object; local 2590 vm_object_upl_request( vm_object_t object, vm_object_offset_t offset, upl_size_t size, upl_t *upl_ptr, upl_page_info_array_t user_page_list, unsigned int *page_list_count, int cntrl_flags) argument 3466 vm_object_super_upl_request( vm_object_t object, vm_object_offset_t offset, upl_size_t size, upl_size_t super_cluster, upl_t *upl, upl_page_info_t *user_page_list, unsigned int *page_list_count, int cntrl_flags) argument 3586 vm_object_t object; local 3742 vm_object_t object; local 3943 vm_object_t object; local 4427 vm_object_t object; local 4713 vm_object_iopl_request( vm_object_t object, vm_object_offset_t offset, upl_size_t size, upl_t *upl_ptr, upl_page_info_array_t user_page_list, unsigned int *page_list_count, int cntrl_flags) argument 5330 vm_paging_map_object( vm_map_offset_t *address, vm_page_t page, vm_object_t object, vm_object_offset_t offset, vm_map_size_t *size, vm_prot_t protection, boolean_t can_unlock_object) argument 5535 vm_paging_unmap_object( vm_object_t object, vm_map_offset_t start, vm_map_offset_t end) argument [all...] |
H A D | vm_pageout.h | 107 vm_object_t object); 123 #define upl_lock_init(object) mutex_init(&(object)->Lock, 0) 124 #define upl_lock(object) mutex_lock(&(object)->Lock) 125 #define upl_unlock(object) mutex_unlock(&(object)->Lock) 134 vm_object_t src_object; /* object derived from */ 168 vm_object_t object, 177 vm_object_t object, [all...] |
H A D | vm_map.c | 136 vm_object_t object, 261 vm_object_t object, 269 vm_object_t object); 375 * by copying VM object references from one map to 378 * to a VM object region exists in any map when this strategy 379 * is used -- this means that shadow object creation can be 382 * the same region of a vm object, and hence cannot delay creating 395 * Placeholder object for submap operations. This object is dropped 417 * This remaps the requested part of the object wit 1026 vm_object_t object = entry->object.vm_object; local 1087 vm_object_t object = entry->object.vm_object; local 1406 vm_map_pmap_enter( vm_map_t map, register vm_map_offset_t addr, register vm_map_offset_t end_addr, register vm_object_t object, vm_object_offset_t offset, vm_prot_t protection) argument 1504 vm_map_enter( vm_map_t map, vm_map_offset_t *address, vm_map_size_t size, vm_map_offset_t mask, int flags, vm_object_t object, vm_object_offset_t offset, boolean_t needs_copy, vm_prot_t cur_protection, vm_prot_t max_protection, vm_inherit_t inheritance) argument 2182 vm_object_t object; local 3069 register vm_object_t object; local 3655 vm_object_t object; local 4364 register vm_object_t object; local 4511 register vm_object_t object; local 6090 vm_object_t object; local 6562 vm_object_t object = copy->cpy_object; local 6706 register vm_object_t object; local 7417 vm_map_copyin_object( vm_object_t object, vm_object_offset_t offset, vm_object_size_t size, vm_map_copy_t *copy_result) argument 7446 vm_object_t object; local 7936 vm_map_lookup_locked( vm_map_t *var_map, vm_map_offset_t vaddr, vm_prot_t fault_type, int object_lock_type, vm_map_version_t *out_version, vm_object_t *object, vm_object_offset_t *offset, vm_prot_t *out_prot, boolean_t *wired, vm_object_fault_info_t fault_info, vm_map_t *real_map) argument 9164 vm_map_region_look_for_page( __unused vm_map_t map, __unused vm_map_offset_t va, vm_object_t object, vm_object_offset_t offset, int max_refcnt, int depth, vm_region_extended_info_t extended) argument 9273 vm_map_region_count_obj_refs( vm_map_entry_t entry, vm_object_t object) argument 9503 vm_object_t object; local 9932 vm_map_entry_insert( vm_map_t map, vm_map_entry_t insp_entry, vm_map_offset_t start, vm_map_offset_t end, vm_object_t object, vm_object_offset_t offset, boolean_t needs_copy, boolean_t is_shared, boolean_t in_transition, vm_prot_t cur_protection, vm_prot_t max_protection, vm_behavior_t behavior, vm_inherit_t inheritance, unsigned wired_count, boolean_t no_cache) argument 10019 vm_object_t object; local 10766 vm_object_t object; local 10843 vm_object_t object; local 11069 vm_object_t object; local 11388 vm_object_t object; local [all...] |
H A D | vm_user.c | 445 * If IPC isn't used, the caller must deal with the vm_map_copy_t object 487 * If IPC isn't used, the caller must deal with the vm_map_copy_t object 824 * Map some range of an object into an address space. 826 * The object can be one of several types of objects: 829 * or a range within a memory object 830 * a whole memory object 1663 /* Retrieve a upl for an object underlying an address range in a map */ 1728 vm_object_t object; local 1782 object = parent_entry->backing.object; 2559 vm_object_t object; local 2673 vm_object_t object; local 2725 vm_object_t object; local 2968 vm_object_t object; local 3063 vm_object_t object; local [all...] |
H A D | vm_purgeable.c | 327 * object is ripe to be purged. It is not purged immediately, because that 383 * grab any ripe object and purge it obsolete queue first. then, go through 386 * 1. Look at queue. Is there an object? 388 * No - check other queue. Is there an object? 487 /* Find an object that can be locked. Returns locked object. */ 497 vm_object_t object; local 498 for (object = (vm_object_t) queue_first(&queue->objq[group]); 499 !queue_end(&queue->objq[group], (queue_entry_t) object); 500 object 522 vm_object_t object = 0; local 591 vm_purgeable_object_add(vm_object_t object, purgeable_q_t queue, int group) argument 618 vm_purgeable_object_remove(vm_object_t object) argument [all...] |
H A D | vm_purgeable_internal.h | 104 /* the object purger. purges the next eligible object from memory. */ 107 /* insert purgeable object into queue */ 108 void vm_purgeable_object_add(vm_object_t object, purgeable_q_t queue, int group); 110 /* look for object. If found, remove from purgeable queue. */ 111 purgeable_q_t vm_purgeable_object_remove(vm_object_t object);
|
H A D | memory_object.h | 84 vm_object_t object); 89 vm_object_t object); 108 memory_object_t object);
|
H A D | vm_resident.c | 121 * (virtual memory object, offset) to page lookup, employs 198 * object structure, be sure to add initialization 369 * for the object/offset-to-page hash table headers. 390 m->object = VM_OBJECT_NULL; /* reset later */ 509 * Calculate object shift value for hashing algorithm: 512 * hash shifts the object left by 864 * Distributes the object/offset key pair among hash buckets. 868 #define vm_page_hash(object, offset) (\ 869 ( (natural_t)((uint32_t)object * vm_page_bucket_hash) + ((uint32_t)atop_64(offset) ^ vm_page_bucket_hash))\ 876 * Inserts the given mem entry into the object/objec 882 vm_page_insert( vm_page_t mem, vm_object_t object, vm_object_offset_t offset) argument 892 vm_page_insert_internal( vm_page_t mem, vm_object_t object, vm_object_offset_t offset, boolean_t queues_lock_held) argument 988 vm_page_replace( register vm_page_t mem, register vm_object_t object, register vm_object_offset_t offset) argument 1189 vm_page_lookup( register vm_object_t object, register vm_object_offset_t offset) argument 1968 vm_page_alloc( vm_object_t object, vm_object_offset_t offset) argument 1985 vm_page_alloclo( vm_object_t object, vm_object_offset_t offset) argument 2011 vm_page_alloc_guard( vm_object_t object, vm_object_offset_t offset) argument 3251 vm_object_t object; local [all...] |
H A D | vm_page.h | 142 * perform object/offset lookups 144 * A list of all pages for a given object, 150 * In addition, the structure contains the object 155 * object that the page belongs to (O) or by the lock on the page 163 queue_chain_t listq; /* all pages in same object (O) */ 166 vm_object_t object; /* which object am I in (O&P) */ member in struct:vm_page 167 vm_object_offset_t offset; /* offset into that object (O,P) */ 190 * by the "VM object" lock. 300 * object/offse [all...] |
H A D | bsd_vm.c | 138 memory_object_control_t control_handle; /* mem object control handle */ 337 vm_object_t object; local 348 object = memory_object_control_to_vm_object(control); 349 if (object == VM_OBJECT_NULL) { 352 assert(!object->internal); 354 vm_object_lock(object); 356 if (mark_dirty && object->copy != VM_OBJECT_NULL) { 363 vm_object_unlock(object); 377 if ((dst_page = vm_page_lookup(object, offset)) == VM_PAGE_NULL) 392 PAGE_SLEEP(object, dst_pag 877 vm_object_t object; local 1159 vm_object_t top_object, object; local [all...] |
H A D | vm_print.h | 47 vm_object_t object);
|
/macosx-10.5.8/xnu-1228.15.4/osfmk/ipc/ |
H A D | ipc_object.c | 103 * Take a reference to an object. 108 ipc_object_t object) 110 io_lock(object); 111 assert(object->io_references > 0); 112 io_reference(object); 113 io_unlock(object); 119 * Release a reference to an object. 124 ipc_object_t object) 126 io_lock(object); 127 assert(object 107 ipc_object_reference( ipc_object_t object) argument 123 ipc_object_release( ipc_object_t object) argument 154 ipc_object_t object; local 203 ipc_object_t object; local 331 ipc_object_t object; local 402 ipc_object_t object; local 563 ipc_object_copyin_from_kernel( ipc_object_t object, mach_msg_type_name_t msgt_name) argument 658 ipc_object_destroy( ipc_object_t object, mach_msg_type_name_t msgt_name) argument 701 ipc_object_copyout( ipc_space_t space, ipc_object_t object, mach_msg_type_name_t msgt_name, boolean_t overflow, mach_port_name_t *namep) argument 790 ipc_object_copyout_name( ipc_space_t space, ipc_object_t object, mach_msg_type_name_t msgt_name, boolean_t overflow, mach_port_name_t name) argument 867 ipc_object_copyout_dest( ipc_space_t space, ipc_object_t object, mach_msg_type_name_t msgt_name, mach_port_name_t *namep) argument 1028 io_free( unsigned int otype, ipc_object_t object) argument 1107 ipc_object_print( ipc_object_t object) argument [all...] |
H A D | ipc_object.h | 126 #define IO_BITS_KOTYPE 0x00000fff /* used by the object */ 128 #define IO_BITS_ACTIVE 0x80000000 /* is object alive? */ 156 ipc_object_t object); 228 /* Take a reference to an object */ 230 ipc_object_t object); 232 /* Release a reference to an object */ 234 ipc_object_t object); 236 /* Look up an object in a space */ 264 /* Allocate an object */ 273 /* Allocate an object, wit [all...] |
/macosx-10.5.8/xnu-1228.15.4/bsd/netat/ |
H A D | atp.h | 237 #define ATP_Q_ENTER(hdr, object, entry) { \ 239 (hdr).head->entry.prev = (object); \ 240 (object)->entry.next = (hdr).head; \ 242 (hdr).tail = (object); \ 243 (object)->entry.next = NULL; \ 245 (object)->entry.prev = NULL; \ 246 (hdr).head = (object); \ 249 #define ATP_Q_APPEND(hdr, object, entry) { \ 251 (hdr).tail->entry.next = (object); \ 252 (object) [all...] |
H A D | ddp_nbp.c | 143 p = (u_char*)&nbp->tuple[0].enu_entity; /* p -> object */ 207 char zone[35],object[35],type[35]; local 209 strlcpy(object,nbp_req.nve.object.str, sizeof(object)); 213 ("nbp_LKUP for:%s:%s@%s", object, type, zone)); 489 register at_nvestr_t *object, *type, *zone; local 507 object = (at_nvestr_t *)&tuple->enu_entity; 508 type = (at_nvestr_t *)(&object->str[object 622 char zone[35],object[35],type[35]; local 696 char zone[35],object[35],type[35]; local 776 char zone[35],object[35],type[35]; local 1408 char zone[35],object[35],type[35]; local [all...] |
/macosx-10.5.8/xnu-1228.15.4/libkern/c++/ |
H A D | OSUnserializeXML.y | 73 typedef struct object { 74 struct object *next; 75 struct object *free; 76 struct object *elements; 77 OSObject *object; 96 OSObject *parsedObject; // resultant object of parsed text 153 | object { STATE->parsedObject = $1->object; 154 $1->object = 0; 163 object label [all...] |
H A D | OSUnserialize.y | 65 typedef struct object { 66 struct object *next; 67 struct object *prev; 68 void *object; 98 // resultant object of parsed text 123 | object { parsedObject = (OSObject *)$1; YYACCEPT; } 127 object: dict { $$ = (object_t *)buildOSDictionary($1); } label 143 | object '@' NUMBER { $$ = $1; 159 pair: object '=' object ';' { [all...] |
H A D | OSUnserializeXML.cpp | 168 typedef struct object { struct 169 struct object *next; 170 struct object *free; 171 struct object *elements; 172 OSObject *object; member in struct:object 191 OSObject *parsedObject; // resultant object of parsed text 571 "'('", "')'", "'['", "']'", "$accept", "input", "object", "dict", 1509 { STATE->parsedObject = (yyvsp[(1) - (1)])->object; 1510 (yyvsp[(1) - (1)])->object = 0; 1562 (yyval)->object [all...] |
/macosx-10.5.8/xnu-1228.15.4/osfmk/mach/ |
H A D | mig.h | 151 /* MIG object runtime - not ready for public consumption */ 156 * MIG object runtime definitions 163 * - Dymanic port allocation on first "remoting" of an object. 164 * - Reference conversions from object to port and vice versa. 195 IMIGObject *object, 200 IMIGObject *object); 203 IMIGObject *object); 206 IMIGObject *object, 210 IMIGObject *object, 214 IMIGObject *object, [all...] |
/macosx-10.5.8/xnu-1228.15.4/osfmk/kern/ |
H A D | hibernate.c | 76 vm_object_t object = NULL; local 85 if (!vm_object_lock_try(m->object)) 88 object = m->object; 95 if (m->busy || !object->alive) 128 || (VM_PURGABLE_VOLATILE == object->purgable) 129 || (VM_PURGABLE_EMPTY == m->object->purgable); 133 if (object) 134 vm_object_unlock(object); 161 if (VM_PURGABLE_VOLATILE == m->object [all...] |