• Home
  • History
  • Annotate
  • Raw
  • Download
  • only in /macosx-10.10/xnu-2782.1.97/osfmk/vm/

Lines Matching defs:pmap

175 	pmap_t		pmap);
243 pmap_t pmap,
339 * specific pmap files since the default behavior varies according to architecture. The
732 pmap_t pmap,
761 result->pmap = pmap;
990 if(map->pmap)
991 pmap_destroy(map->pmap);
1396 * the pmap at the specified address if they are present.
1417 if(map->pmap == 0)
1455 kr = vm_fault_enter(m, map->pmap, addr, protection, protection,
1479 return pmap_is_empty(map->pmap, start, end);
1484 if (map->pmap == NULL) {
1491 phys_page = pmap_find_phys(map->pmap, offset);
2041 if (map->pmap == kernel_pmap) {
2173 * Turn off pmap accounting since
2206 if (use_pmap && submap->pmap == NULL) {
2207 ledger_t ledger = map->pmap->ledger;
2208 /* we need a sub pmap to nest... */
2209 submap->pmap = pmap_create(ledger, 0,
2211 if (submap->pmap == NULL) {
2215 if (use_pmap && submap->pmap != NULL) {
2216 kr = pmap_nest(map->pmap,
2217 submap->pmap,
2351 pmap_empty = FALSE; /* pmap won't be empty */
2637 * that uses a different pmap.
2641 * than just the submap's pmap.
3086 kr = pmap_enter_options(target_map->pmap,
3324 pmap_t pmap;
3404 * we can talk to the pmap module about which addresses
3448 pmap = vm_map_pmap(map);
3449 pmap_pageable(pmap, start, end, FALSE);
3452 * Enter each page into the pmap, to avoid faults.
3469 vm_fault_enter(m, pmap, va, VM_PROT_ALL, VM_PROT_WRITE,
3565 if (pmap_adjust_unnest_parameters(map->pmap, &start_unnest, &end_unnest)) {
3590 pmap_unnest(map->pmap,
3643 pmap_remove(map->pmap,
3744 pmap_remove(map->pmap,
3944 * that uses a different pmap.
3948 * than just the submap's pmap.
3956 if(submap->pmap == NULL) {
3957 ledger_t ledger = map->pmap->ledger;
3958 submap->pmap = pmap_create(ledger,
3960 if(submap->pmap == PMAP_NULL) {
3965 result = pmap_nest(map->pmap,
3966 (entry->object.sub_map)->pmap,
3975 pmap_remove(map->pmap, (addr64_t)start, (addr64_t)end);
4133 * we won't do it for real (in pmap). This is because
4152 pmap_protect(current->object.sub_map->pmap,
4157 pmap_protect(map->pmap,
4523 pmap_t pmap;
4554 pmap = entry->object.sub_map->pmap;
4556 /* submaps pmap address ranges line */
4563 pmap = map->pmap;
4636 pmap = map_pmap;
4648 user_wire, pmap, pmap_addr,
4887 &tmp_entry, map->pmap,
5115 pmap_t pmap;
5126 pmap = entry->object.sub_map->pmap;
5129 pmap = map->pmap;
5173 sub_start, sub_end, user_wire, pmap, pmap_addr);
5288 &tmp_entry, FALSE, map->pmap,
5399 * pmap entries pointing to its pages.
5455 pmap_remove(map->pmap,
5488 pmap_remove(map->pmap,
5602 if (map->pmap == kernel_pmap &&
5842 pmap_t pmap;
5851 pmap = sub_map->pmap;
5854 pmap = map->pmap;
5860 pmap, pmap_addr);
5865 map->pmap,
5874 map->pmap, tmp_entry.vme_start);
5929 * pmap. The original pmap has already been flushed
5932 * we don't have any pmap to flush, so let's just skip
5938 pmap_unnest(map->pmap,
5969 (map->pmap == kernel_pmap)) {
5981 pmap_remove_options(map->pmap,
5997 * All pmap mappings for this map entry must have been
6008 if (map->pmap == kernel_pmap &&
6431 pmap_t pmap,
6807 if((entry->use_pmap) && (pmap == NULL)) {
6813 entry->object.sub_map->pmap,
6815 } else if (pmap != NULL) {
6820 interruptible, pmap,
6828 dst_map->pmap,
6947 if(pmap)
6948 local_pmap = pmap;
6950 local_pmap = dst_map->pmap;
7542 __unused pmap_t pmap)
7732 pmap_unnest(dst_map->pmap,
7767 dst_map->pmap,
7788 * we could try to push pages into the pmap at this point, BUT
7793 * it also puts a lot of pressure on the pmap layer w/r to mapping structures
8426 pmap_pageable(dst_map->pmap,
8475 vm_fault_enter(m, dst_map->pmap, va, prot, prot,
8921 : src_map->pmap),
9481 result = pmap_nest(new_map->pmap,
9482 (old_entry->object.sub_map)->pmap,
9622 pmap_protect(old_map->pmap,
9680 /* Bill Angell pmap support goes here */
9682 pmap_copy(new_map->pmap, old_map->pmap, new_entry->vme_start,
9787 old_map->pmap->pm_task_map != TASK_MAP_32BIT
9872 old_map->pmap),
10215 PMAP_NULL : map->pmap,
11661 * everything else in the pmap module. If any non-volatile
11662 * information must be kept, the pmap module should handle
11689 ret = pmap_attribute(map->pmap, start, end-start,
12309 * "iokit_acct" and "reusable" in the pmap stats and ledgers.
12472 * object: "use_pmap" means "use pmap accounting" for footprint.
12609 * no pmap accounting for them.
12615 * must be accounted by pmap stats.
12747 PMAP_NULL : map->pmap),
13480 map->pmap == kernel_pmap) {
14268 if (map->pmap == NULL)
14271 pmap_disable_NX(map->pmap);
14436 * IOKit has mapped a region into this map; adjust the pmap's ledgers appropriately.
14443 pmap_t pmap = vm_map_pmap(map);
14445 ledger_credit(pmap->ledger, task_ledgers.iokit_mapped, bytes);
14446 ledger_credit(pmap->ledger, task_ledgers.phys_footprint, bytes);
14452 pmap_t pmap = vm_map_pmap(map);
14454 ledger_debit(pmap->ledger, task_ledgers.iokit_mapped, bytes);
14455 ledger_debit(pmap->ledger, task_ledgers.phys_footprint, bytes);
14535 /* Pull the dirty status from the pmap, since we cleared the
15046 volatile_pmap_count += pmap_query_resident(map->pmap,