Lines Matching refs:l3
76 pt_entry_t *l3 __diagused;
80 l3 = (pd_entry_t *)PHYS_TO_DMAP(l2e & ~ATTR_MASK);
82 KASSERT(l3[i] == 0, ("%s: l3 still mapped: %p %lx", __func__,
83 &l3[i], l3[i]));
242 pt_entry_t new_l2e, l2e, *l2, *l3;
278 l3 = (pt_entry_t *)PHYS_TO_DMAP(l2e & ~ATTR_MASK);
279 return (l3);
288 pd_entry_t l3e, *l3;
312 l3 = vmmpmap_l3_table(va);
313 if (l3 == NULL)
318 * Ensure no other threads can write to l3 between the KASSERT
323 KASSERT(atomic_load_64(&l3[pmap_l3_index(va)]) == 0,
326 atomic_store_64(&l3[pmap_l3_index(va)], l3e | pa);
343 pd_entry_t *l3, l3e, **l3_list;
390 l3 = (pd_entry_t *)PHYS_TO_DMAP(l2e & ~ATTR_MASK);
392 l3e = atomic_load_64(&l3[pmap_l3_index(va)]);
400 atomic_store_64(&l3[pmap_l3_index(va)], l3e);
402 l3_list[i] = &l3[pmap_l3_index(va)];
409 atomic_store_64(&l3[pmap_l3_index(va)], 0);