/macosx-10.5.8/xnu-1228.15.4/osfmk/mach/ |
H A D | vm_region.h | 71 vm_prot_t protection; member in struct:vm_region_basic_info_64 99 vm_prot_t protection; member in struct:vm_region_basic_info 134 vm_prot_t protection; member in struct:vm_region_extended_info 192 vm_prot_t protection; /* present access protection */ member in struct:vm_region_submap_info 218 vm_prot_t protection; /* present access protection */ member in struct:vm_region_submap_info_64 244 vm_prot_t protection; /* present access protection */ member in struct:vm_region_submap_short_info_64
|
H A D | shared_memory_server.h | 95 vm_prot_t protection; /* read/write/execute/COW/ZF */ member in struct:sf_mapping
|
/macosx-10.5.8/xnu-1228.15.4/osfmk/mach/i386/ |
H A D | vm_param.h | 195 #define PMAP_ENTER(pmap, virtual_address, page, protection, flags, wired) \ 199 vm_prot_t __prot__ = (protection); \
|
/macosx-10.5.8/xnu-1228.15.4/osfmk/vm/ |
H A D | vm_fault.h | 117 vm_prot_t *protection, /* Protection for mapping */
|
H A D | vm_user.c | 331 * Sets the protection of the specified range in the 359 * Sets the protection of the specified range in the 1495 info->protection = info64.protection; 1787 if((access != GET_MAP_MEM(parent_entry->protection)) && 1788 !(parent_entry->protection & VM_PROT_WRITE)) { 1792 SET_MAP_MEM(access, parent_entry->protection); 1795 SET_MAP_MEM(access, parent_entry->protection); 1798 SET_MAP_MEM(access, parent_entry->protection); 1801 SET_MAP_MEM(access, parent_entry->protection); [all...] |
H A D | vm_map.c | 138 vm_prot_t protection); 346 * Virtual memory maps provide for the mapping, protection, 485 map_entry->protection, 1357 new_entry->protection = VM_PROT_DEFAULT; 1412 vm_prot_t protection) 1444 kr = vm_fault_enter(m, map->pmap, addr, protection, 1796 entry->protection != cur_protection || 1888 (entry->protection == cur_protection) && 1926 * allocation into 4GB chunks if the max protection is NONE. That 2000 new_entry->protection, TRU 1406 vm_map_pmap_enter( vm_map_t map, register vm_map_offset_t addr, register vm_map_offset_t end_addr, register vm_object_t object, vm_object_offset_t offset, vm_prot_t protection) argument 10704 vm_map_check_protection(vm_map_t map, vm_map_offset_t start, vm_map_offset_t end, vm_prot_t protection) argument [all...] |
H A D | pmap.h | 361 #define PMAP_ENTER(pmap, virtual_address, page, protection, flags, wired) \ 376 (protection), \
|
H A D | vm_pageout.h | 239 vm_prot_t protection,
|
H A D | vm_kern.c | 106 vm_prot_t protection); 706 vm_prot_t protection) 751 PMAP_ENTER(kernel_pmap, map_start, mem, protection, 701 kmem_remap_pages( register vm_object_t object, register vm_object_offset_t offset, register vm_offset_t start, register vm_offset_t end, vm_prot_t protection) argument
|
H A D | vm_map.h | 173 vm_prot_t protection; /* access permissions */ member in struct:vm_named_entry 190 * and user-exported inheritance and protection information. 218 /* vm_prot_t */ protection:3, /* protection code */ member in struct:vm_map_entry 219 /* vm_prot_t */ max_protection:3,/* maximum protection */ 437 /* Find the VM object, offset, and protection for a given virtual address 811 /* Change protection */ 819 /* Check protection */ 824 vm_prot_t protection); 935 * usual form; it handles a copyin based on the current protection [all...] |
H A D | vm_protos.h | 54 vm_prot_t protection,
|
H A D | vm_debug.c | 177 region.vir_protection = entry->protection; 378 region.vir_protection = entry->protection;
|
H A D | memory_object.c | 226 * that to be done as part of the protection. 237 * (pmap_page_protect may not increase protection). 461 * same as memory_object_lock_request but page protection 687 vm_prot_t protection) 1021 should_flush, should_return, should_iosync, protection)) 1472 if((named_entry->protection & VM_PROT_READ) 1477 if((named_entry->protection & 679 vm_object_update( register vm_object_t object, register vm_object_offset_t offset, register vm_object_size_t size, register vm_object_offset_t *resid_offset, int *io_errno, memory_object_return_t should_return, int flags, vm_prot_t protection) argument
|
H A D | bsd_vm.c | 1088 pinfo->pri_protection = entry->protection; 1103 extended.protection = entry->protection;
|
H A D | vm_fault.c | 581 * in "protection". 623 vm_prot_t *protection, /* Protection for mapping */ 738 *protection &= ~VM_PROT_WRITE; 779 (integer_t)object, offset, fault_type, *protection, 0); 1645 *protection &= (~VM_PROT_WRITE); 1668 *protection &= ~VM_PROT_WRITE; 2322 * If the page is wired, we must fault for the current protection 3179 * Check whether the protection has changed or the object 3267 /* While we do not worry about execution protection in */ 3270 /* to execute, we return with a protection failur 616 vm_fault_page( vm_object_t first_object, vm_object_offset_t first_offset, vm_prot_t fault_type, boolean_t must_be_resident, vm_prot_t *protection, vm_page_t *result_page, vm_page_t *top_page, int *type_of_fault, kern_return_t *error_code, boolean_t no_zero_fill, boolean_t data_supply, vm_object_fault_info_t fault_info) argument [all...] |
H A D | vm_shared_region.c | 592 mem_entry->protection = VM_PROT_ALL; 1228 mem_entry->protection = VM_PROT_ALL;
|
H A D | vm_pageout.c | 3580 if (!(entry->protection & VM_PROT_WRITE)) { 5336 vm_prot_t protection, 5419 protection, 5463 protection, 5514 protection, 5330 vm_paging_map_object( vm_map_offset_t *address, vm_page_t page, vm_object_t object, vm_object_offset_t offset, vm_map_size_t *size, vm_prot_t protection, boolean_t can_unlock_object) argument
|
/macosx-10.5.8/xnu-1228.15.4/osfmk/kdp/ |
H A D | kdp_protocol.h | 211 vm_prot_t protection; member in struct:__anon798
|
H A D | kdp.c | 560 r->protection = VM_PROT_ALL; r++; rp->nregions++;
|
/macosx-10.5.8/xnu-1228.15.4/osfmk/ppc/ |
H A D | model_dep.c | 990 if (!(info.protection & VM_PROT_WRITE)) { 991 /* Save the original protection values for restoration later */ 992 reprotect = info.protection;
|
H A D | vmachmon_asm.s | 94 .long EXT(vmm_protect_page32) ; Sets protection values for a page - supports 32-bit 98 .long EXT(vmm_protect_execute32) ; Sets protection values for a page and launches VM - supports 32-bit 126 .long EXT(vmm_protect_page) ; Sets protection values for a page - supports 64-bit 282 mr r7,r6 ; Move protection bits 302 mr r7,r6 ; Move protection bits 524 rlwimi r15,r20,32+vmmKeyb-userProtKeybit,userProtKeybit,userProtKeybit ; Set the protection key 783 rlwinm r11,r11,0,userProtKeybit+1,userProtKeybit-1 ; Set back to normal protection key 849 rlwinm r9,r9,0,userProtKeybit+1,userProtKeybit-1 ; Set back to normal protection key 2095 rlwimi r7,r5,32+vmmKeyb-userProtKeybit,userProtKeybit,userProtKeybit ; Set the protection key 2280 rlwimi r7,r5,32+vmmKeyb-userProtKeybit,userProtKeybit,userProtKeybit ; Set the protection ke [all...] |
/macosx-10.5.8/xnu-1228.15.4/bsd/kern/ |
H A D | kern_core.c | 369 prot = vbr.protection;
|
/macosx-10.5.8/xnu-1228.15.4/osfmk/kdp/ml/i386/ |
H A D | kdp_vm.c | 422 prot = vbr.protection;
|
/macosx-10.5.8/xnu-1228.15.4/osfmk/kdp/ml/ppc/ |
H A D | kdp_vm.c | 443 prot = vbr.protection;
|
/macosx-10.5.8/xnu-1228.15.4/bsd/dev/dtrace/ |
H A D | dtrace_glue.c | 178 if (!(info.protection & VM_PROT_WRITE)) { 179 /* Save the original protection values for restoration later */ 180 reprotect = info.protection;
|