Searched refs:pmap_enter (Results 1 - 10 of 10) sorted by relevance

/freebsd-current/sys/dev/xdma/
H A Dxdma_iommu.c90 pmap_enter(p, va, m, prot, prot | PMAP_ENTER_WIRED, 0);
/freebsd-current/sys/vm/
H A Dpmap.h100 * Flags for pmap_enter(). The bits in the low-order byte are reserved
130 int pmap_enter(pmap_t pmap, vm_offset_t va, vm_page_t m,
H A Dvm_fault.c411 if (pmap_enter(fs->map->pmap, vaddr, m_map, fs->prot, fs->fault_type |
573 rv = pmap_enter(fs->map->pmap, vaddr, m, fs->prot,
629 rv = pmap_enter(fs->map->pmap, vaddr, m, fs->prot, fs->fault_type |
633 * pmap_enter() may fail for a superpage mapping if additional
645 rv = pmap_enter(fs->map->pmap, vaddr + ptoa(i),
1058 * pmap_enter is complete.
1075 * pmap_enter() will replace an existing mapping in the current
1753 * because pmap_enter() may sleep. We don't put the page
1757 pmap_enter(fs.map->pmap, vaddr, fs.m, fs.prot,
1913 * If pmap_enter() ha
[all...]
H A Dvm_kern.c287 pmap_enter(kernel_pmap, addr + i, m, prot,
379 pmap_enter(kernel_pmap, tmp, m, VM_PROT_RW,
571 pmap_enter(kernel_pmap, addr + i, m, prot,
/freebsd-current/sys/i386/i386/
H A Dpmap_base.c675 pmap_enter(pmap_t pmap, vm_offset_t va, vm_page_t m, vm_prot_t prot, function
/freebsd-current/share/man/man9/
H A DMakefile270 pmap_enter.9 \
/freebsd-current/sys/riscv/riscv/
H A Dpmap.c297 * Internal flags for pmap_enter()'s helper functions.
2910 pmap_enter(pmap_t pmap, vm_offset_t va, vm_page_t m, vm_prot_t prot, function
2956 CTR2(KTR_PMAP, "pmap_enter: %.16lx -> %.16lx", va, pa);
2965 ("pmap_enter: va %#lx unaligned", va));
2966 KASSERT(m->psind > 0, ("pmap_enter: m->psind < psind"));
2984 CTR0(KTR_PMAP, "pmap_enter: mpte == NULL");
3000 panic("pmap_enter: l2 pte_m == NULL");
3016 panic("pmap_enter: l3 pte_m == NULL");
3055 ("pmap_enter: missing reference to page table page,"
3084 ("pmap_enter
[all...]
/freebsd-current/sys/arm/arm/
H A Dpmap-v6.c321 * Internal flags for pmap_enter()'s helper functions.
3839 pmap_enter(pmap_t pmap, vm_offset_t va, vm_page_t m, vm_prot_t prot, function
3853 ("%s: invalid to pmap_enter page table pages (va: 0x%x)", __func__,
4505 * but is *MUCH* faster than pmap_enter...
4630 * is set. QQQ: For more info, see comments in pmap_enter().
4753 * is set. QQQ: For more info, see comments in pmap_enter().
6476 * demotion. If not caught here, pmap_enter() can find a section
/freebsd-current/sys/arm64/arm64/
H A Dpmap.c445 * Internal flags for pmap_enter()'s helper functions.
5062 pmap_enter(pmap_t pmap, vm_offset_t va, vm_page_t m, vm_prot_t prot, function
5122 CTR2(KTR_PMAP, "pmap_enter: %.16lx -> %.16lx", va, pa);
5143 KASSERT((va & L2_OFFSET) == 0, ("pmap_enter: va unaligned"));
5144 KASSERT(m->psind > 0, ("pmap_enter: m->psind < psind"));
5188 CTR0(KTR_PMAP, "pmap_enter: mpte == NULL");
5194 panic("pmap_enter: missing L3 table for kernel va %#lx", va);
5224 ("pmap_enter: missing reference to page table page,"
5249 ("pmap_enter: unexpected pa update for %#lx", va));
5275 ("pmap_enter
[all...]
/freebsd-current/sys/amd64/amd64/
H A Dpmap.c1245 * Internal flags for pmap_enter()'s helper functions.
1747 * pages. (pmap_enter requires a PD page to exist for each KPML4E.)
7150 pmap_enter(pmap_t pmap, vm_offset_t va, vm_page_t m, vm_prot_t prot, function
7170 KASSERT(va <= VM_MAX_KERNEL_ADDRESS, ("pmap_enter: toobig"));
7172 ("pmap_enter: invalid to pmap_enter page table pages (va: 0x%lx)",
7175 ("pmap_enter: managed mapping within the clean submap"));
7179 ("pmap_enter: flags %u has reserved bits set", flags));
7187 ("pmap_enter: flags includes VM_PROT_WRITE but prot doesn't"));
7220 KASSERT((va & PDRMASK) == 0, ("pmap_enter
[all...]

Completed in 313 milliseconds