Lines Matching refs:mapped
1289 /* Get vm_page_t for mapped pte. */
1430 * Insert pv_entry into pv_list for mapped page if part of managed
1533 /* Get vm_page_t for mapped pte. */
1620 * Insert pv_entry into pv_list for mapped page if part of managed
1750 * kernsize is the size of the kernel that is actually mapped.
2571 * mapped at the given virtual address start. Each subsequent page is
2572 * mapped at a virtual address that is offset from start by the same
2575 * m_start that can be mapped at a virtual address less than the given
2577 * is mapped; only those for which a resident page exists with the
2578 * corresponding offset from m_start are mapped.
3570 * Return a pointer to where it is mapped. This routine is intended to be used
3674 * 'Unmap' a range mapped by mmu_booke_mapdev().
4005 * Convert TLB TSIZE value to mapped region size.
4098 vm_size_t mapped, pgsz, base, mask;
4104 mapped = 0;
4108 while (mapped < size) {
4109 while (mapped < size && idx < KERNEL_REGION_MAX_TLB_ENTRIES) {
4110 while (pgsz > (size - mapped))
4113 mapped += pgsz;
4117 if (mapped < size) {
4120 mapped -= pgsz;
4125 mapped += pgsz;
4147 mapped = (va - base);
4149 printf("mapped size 0x%"PRIxPTR" (wasted space 0x%"PRIxPTR")\n",
4150 mapped, mapped - size);
4151 return (mapped);