Lines Matching defs:pde
288 * for legacy, returns the address of the pde entry.
289 * for 64 bit, causes the pdpt page containing the pde entry to be mapped,
290 * then returns the mapped address of the pde entry in that page
295 pd_entry_t *pde;
297 pde = (&((m)->dirbase[(vm_offset_t)(v) >> PDESHIFT]));
301 pde = pmap64_pde(m, v);
303 return pde;
365 * maps in the pdpt page, if any, containing the pde entry requested
366 * and returns the address of the pde entry in that mapped page
429 * maps the pde page, if any, containing the pte in and returns
435 pd_entry_t *pde;
440 pde = pmap_pde(pmap,vaddr);
442 if (pde && ((*pde & INTEL_PTE_VALID))) {
443 if (*pde & INTEL_PTE_PS)
444 return pde;
453 newpf = *pde & PG_FRAME;
715 /* make sure G bit is on for high shared pde entry */
1147 pt_entry_t pde;
1156 pde = *pdep & PTMASK; /* page attributes from pde */
1157 pde |= INTEL_PTE_PS; /* make it a 2M entry */
1158 pde |= pte_phys; /* take page frame from pte */
1161 pde &= ~INTEL_PTE_RW;
1163 (void *)pdep, pde);
1164 pmap_store_pte(pdep, pde);
1375 /* map the high shared pde */
1394 /* allocate the vm_objs to hold the pdpt, pde and pte pages */
1414 pmap_expand_pdpt(p, (uint64_t)HIGH_MEM_BASE, PMAP_EXPAND_OPTIONS_NONE); /* need room for another pde entry */
1675 register pt_entry_t *pde;
1709 pde = pmap_pde(map, sva);
1710 if (pde && (*pde & INTEL_PTE_VALID)) {
2013 pmap_expand_pdpt(map, vaddr, PMAP_EXPAND_OPTIONS_NONE); /* need room for another pde entry */
2020 * Allocate a VM page for the pde entries.