tlb.h (91782) | tlb.h (91783) |
---|---|
1/*- 2 * Copyright (c) 2001 Jake Burkholder. 3 * All rights reserved. 4 * 5 * Redistribution and use in source and binary forms, with or without 6 * modification, are permitted provided that the following conditions 7 * are met: 8 * 1. Redistributions of source code must retain the above copyright --- 9 unchanged lines hidden (view full) --- 18 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 19 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 20 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 21 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 22 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 23 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 24 * SUCH DAMAGE. 25 * | 1/*- 2 * Copyright (c) 2001 Jake Burkholder. 3 * All rights reserved. 4 * 5 * Redistribution and use in source and binary forms, with or without 6 * modification, are permitted provided that the following conditions 7 * are met: 8 * 1. Redistributions of source code must retain the above copyright --- 9 unchanged lines hidden (view full) --- 18 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 19 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 20 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 21 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 22 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 23 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 24 * SUCH DAMAGE. 25 * |
26 * $FreeBSD: head/sys/sparc64/include/tlb.h 91782 2002-03-07 05:25:15Z jake $ | 26 * $FreeBSD: head/sys/sparc64/include/tlb.h 91783 2002-03-07 06:01:40Z jake $ |
27 */ 28 29#ifndef _MACHINE_TLB_H_ 30#define _MACHINE_TLB_H_ 31 | 27 */ 28 29#ifndef _MACHINE_TLB_H_ 30#define _MACHINE_TLB_H_ 31 |
32#define TLB_SLOT_COUNT 64 | 32#define TLB_SLOT_COUNT 64 /* XXX */ |
33 34#define TLB_SLOT_TSB_KERNEL_MIN 62 /* XXX */ | 33 34#define TLB_SLOT_TSB_KERNEL_MIN 62 /* XXX */ |
35#define TLB_SLOT_KERNEL 63 | 35#define TLB_SLOT_KERNEL 63 /* XXX */ |
36 37#define TLB_DAR_SLOT_SHIFT (3) 38#define TLB_DAR_SLOT(slot) ((slot) << TLB_DAR_SLOT_SHIFT) 39 40#define TAR_VPN_SHIFT (13) 41#define TAR_CTX_MASK ((1 << TAR_VPN_SHIFT) - 1) 42 43#define TLB_TAR_VA(va) ((va) & ~TAR_CTX_MASK) --- 40 unchanged lines hidden (view full) --- 84#define MMU_SFSR_CT_SIZE (2) 85 86#define MMU_SFSR_W (1L << MMU_SFSR_W_SHIFT) 87 88extern int kernel_tlb_slots; 89extern struct tte *kernel_ttes; 90 91/* | 36 37#define TLB_DAR_SLOT_SHIFT (3) 38#define TLB_DAR_SLOT(slot) ((slot) << TLB_DAR_SLOT_SHIFT) 39 40#define TAR_VPN_SHIFT (13) 41#define TAR_CTX_MASK ((1 << TAR_VPN_SHIFT) - 1) 42 43#define TLB_TAR_VA(va) ((va) & ~TAR_CTX_MASK) --- 40 unchanged lines hidden (view full) --- 84#define MMU_SFSR_CT_SIZE (2) 85 86#define MMU_SFSR_W (1L << MMU_SFSR_W_SHIFT) 87 88extern int kernel_tlb_slots; 89extern struct tte *kernel_ttes; 90 91/* |
92 * Some tlb operations must be atomical, so no interrupt or trap can be allowed | 92 * Some tlb operations must be atomic, so no interrupt or trap can be allowed |
93 * while they are in progress. Traps should not happen, but interrupts need to 94 * be explicitely disabled. critical_enter() cannot be used here, since it only 95 * disables soft interrupts. | 93 * while they are in progress. Traps should not happen, but interrupts need to 94 * be explicitely disabled. critical_enter() cannot be used here, since it only 95 * disables soft interrupts. |
96 * XXX: is something like this needed elsewhere, too? | |
97 */ 98 99static __inline void | 96 */ 97 98static __inline void |
100tlb_dtlb_context_primary_demap(void) | 99tlb_context_demap(struct pmap *pm) |
101{ | 100{ |
102 stxa(TLB_DEMAP_PRIMARY | TLB_DEMAP_CONTEXT, ASI_DMMU_DEMAP, 0); 103 membar(Sync); 104} | 101 void *cookie; 102 u_long s; |
105 | 103 |
106static __inline void 107tlb_dtlb_page_demap(struct pmap *pm, vm_offset_t va) 108{ 109 u_int ctx; 110 111 ctx = pm->pm_context[PCPU_GET(cpuid)]; 112 if (ctx == TLB_CTX_KERNEL) { 113 stxa(TLB_DEMAP_VA(va) | TLB_DEMAP_NUCLEUS | TLB_DEMAP_PAGE, 114 ASI_DMMU_DEMAP, 0); | 104 cookie = ipi_tlb_context_demap(pm); 105 if (pm->pm_active & PCPU_GET(cpumask)) { 106 KASSERT(pm->pm_context[PCPU_GET(cpuid)] != -1, 107 ("tlb_context_demap: inactive pmap?")); 108 s = intr_disable(); 109 stxa(TLB_DEMAP_PRIMARY | TLB_DEMAP_CONTEXT, ASI_DMMU_DEMAP, 0); 110 stxa(TLB_DEMAP_PRIMARY | TLB_DEMAP_CONTEXT, ASI_IMMU_DEMAP, 0); |
115 membar(Sync); | 111 membar(Sync); |
116 } else if (ctx != -1) { 117 stxa(TLB_DEMAP_VA(va) | TLB_DEMAP_PRIMARY | TLB_DEMAP_PAGE, 118 ASI_DMMU_DEMAP, 0); 119 membar(Sync); | 112 intr_restore(s); |
120 } | 113 } |
114 ipi_wait(cookie); |
|
121} 122 123static __inline void | 115} 116 117static __inline void |
124tlb_dtlb_store(vm_offset_t va, u_long ctx, struct tte tte) | 118tlb_page_demap(u_int tlb, struct pmap *pm, vm_offset_t va) |
125{ | 119{ |
126 u_long pst; | 120 u_long flags; 121 void *cookie; 122 u_long s; |
127 | 123 |
128 pst = intr_disable(); 129 stxa(AA_DMMU_TAR, ASI_DMMU, 130 TLB_TAR_VA(va) | TLB_TAR_CTX(ctx)); 131 stxa(0, ASI_DTLB_DATA_IN_REG, tte.tte_data); 132 membar(Sync); 133 intr_restore(pst); 134} 135 136static __inline void 137tlb_dtlb_store_slot(vm_offset_t va, u_long ctx, struct tte tte, int slot) 138{ 139 u_long pst; 140 141 pst = intr_disable(); 142 stxa(AA_DMMU_TAR, ASI_DMMU, TLB_TAR_VA(va) | TLB_TAR_CTX(ctx)); 143 stxa(TLB_DAR_SLOT(slot), ASI_DTLB_DATA_ACCESS_REG, tte.tte_data); 144 membar(Sync); 145 intr_restore(pst); 146} 147 148static __inline void 149tlb_itlb_context_primary_demap(void) 150{ 151 stxa(TLB_DEMAP_PRIMARY | TLB_DEMAP_CONTEXT, ASI_IMMU_DEMAP, 0); 152 membar(Sync); 153} 154 155static __inline void 156tlb_itlb_page_demap(struct pmap *pm, vm_offset_t va) 157{ 158 u_int ctx; 159 160 ctx = pm->pm_context[PCPU_GET(cpuid)]; 161 if (ctx == TLB_CTX_KERNEL) { 162 stxa(TLB_DEMAP_VA(va) | TLB_DEMAP_NUCLEUS | TLB_DEMAP_PAGE, 163 ASI_IMMU_DEMAP, 0); 164 flush(KERNBASE); 165 } else if (ctx != -1) { 166 stxa(TLB_DEMAP_VA(va) | TLB_DEMAP_PRIMARY | TLB_DEMAP_PAGE, 167 ASI_IMMU_DEMAP, 0); 168 membar(Sync); | 124 cookie = ipi_tlb_page_demap(tlb, pm, va); 125 if (pm->pm_active & PCPU_GET(cpumask)) { 126 KASSERT(pm->pm_context[PCPU_GET(cpuid)] != -1, 127 ("tlb_page_demap: inactive pmap?")); 128 if (pm == kernel_pmap) 129 flags = TLB_DEMAP_NUCLEUS | TLB_DEMAP_PAGE; 130 else 131 flags = TLB_DEMAP_PRIMARY | TLB_DEMAP_PAGE; 132 133 s = intr_disable(); 134 if (tlb & TLB_DTLB) { 135 stxa(TLB_DEMAP_VA(va) | flags, ASI_DMMU_DEMAP, 0); 136 membar(Sync); 137 } 138 if (tlb & TLB_ITLB) { 139 stxa(TLB_DEMAP_VA(va) | flags, ASI_IMMU_DEMAP, 0); 140 membar(Sync); 141 } 142 intr_restore(s); |
169 } | 143 } |
144 ipi_wait(cookie); |
|
170} 171 172static __inline void | 145} 146 147static __inline void |
173tlb_itlb_store(vm_offset_t va, u_long ctx, struct tte tte) | 148tlb_range_demap(struct pmap *pm, vm_offset_t start, vm_offset_t end) |
174{ | 149{ |
175 u_long pst; | 150 vm_offset_t va; 151 void *cookie; 152 u_long flags; 153 u_long s; |
176 | 154 |
177 pst = intr_disable(); 178 stxa(AA_IMMU_TAR, ASI_IMMU, TLB_TAR_VA(va) | TLB_TAR_CTX(ctx)); 179 stxa(0, ASI_ITLB_DATA_IN_REG, tte.tte_data); 180 if (ctx == TLB_CTX_KERNEL) 181 flush(va); 182 else { 183 /* 184 * flush probably not needed and impossible here, no access to 185 * user page. 186 */ 187 membar(Sync); | 155 cookie = ipi_tlb_range_demap(pm, start, end); 156 if (pm->pm_active & PCPU_GET(cpumask)) { 157 KASSERT(pm->pm_context[PCPU_GET(cpuid)] != -1, 158 ("tlb_range_demap: inactive pmap?")); 159 if (pm == kernel_pmap) 160 flags = TLB_DEMAP_NUCLEUS | TLB_DEMAP_PAGE; 161 else 162 flags = TLB_DEMAP_PRIMARY | TLB_DEMAP_PAGE; 163 164 s = intr_disable(); 165 for (va = start; va < end; va += PAGE_SIZE) { 166 stxa(TLB_DEMAP_VA(va) | flags, ASI_DMMU_DEMAP, 0); 167 stxa(TLB_DEMAP_VA(va) | flags, ASI_IMMU_DEMAP, 0); 168 membar(Sync); 169 } 170 intr_restore(s); |
188 } | 171 } |
189 intr_restore(pst); | 172 ipi_wait(cookie); |
190} 191 | 173} 174 |
192static __inline void 193tlb_context_demap(struct pmap *pm) 194{ 195 u_int ctx; | 175#define tlb_tte_demap(tte, pm) \ 176 tlb_page_demap(TD_GET_TLB((tte).tte_data), pm, \ 177 TV_GET_VA((tte).tte_vpn)); |
196 | 178 |
197 ctx = pm->pm_context[PCPU_GET(cpuid)]; 198 if (ctx != -1) { 199 tlb_dtlb_context_primary_demap(); 200 tlb_itlb_context_primary_demap(); 201 } 202} 203 204static __inline void 205tlb_itlb_store_slot(vm_offset_t va, u_long ctx, struct tte tte, int slot) 206{ 207 u_long pst; 208 209 pst = intr_disable(); 210 stxa(AA_IMMU_TAR, ASI_IMMU, TLB_TAR_VA(va) | TLB_TAR_CTX(ctx)); 211 stxa(TLB_DAR_SLOT(slot), ASI_ITLB_DATA_ACCESS_REG, tte.tte_data); 212 flush(va); 213 intr_restore(pst); 214} 215 216static __inline void 217tlb_page_demap(u_int tlb, struct pmap *pm, vm_offset_t va) 218{ 219 if (tlb & TLB_DTLB) 220 tlb_dtlb_page_demap(pm, va); 221 if (tlb & TLB_ITLB) 222 tlb_itlb_page_demap(pm, va); 223} 224 225static __inline void 226tlb_range_demap(struct pmap *pm, vm_offset_t start, vm_offset_t end) 227{ 228 for (; start < end; start += PAGE_SIZE) 229 tlb_page_demap(TLB_DTLB | TLB_ITLB, pm, start); 230} 231 232static __inline void 233tlb_tte_demap(struct tte tte, struct pmap *pm) 234{ 235 tlb_page_demap(TD_GET_TLB(tte.tte_data), pm, TV_GET_VA(tte.tte_vpn)); 236} 237 238static __inline void 239tlb_store(u_int tlb, vm_offset_t va, u_long ctx, struct tte tte) 240{ 241 KASSERT(ctx != -1, ("tlb_store: invalid context")); 242 if (tlb & TLB_DTLB) 243 tlb_dtlb_store(va, ctx, tte); 244 if (tlb & TLB_ITLB) 245 tlb_itlb_store(va, ctx, tte); 246} 247 248static __inline void 249tlb_store_slot(u_int tlb, vm_offset_t va, u_long ctx, struct tte tte, int slot) 250{ 251 KASSERT(ctx != -1, ("tlb_store_slot: invalid context")); 252 if (tlb & TLB_DTLB) 253 tlb_dtlb_store_slot(va, ctx, tte, slot); 254 if (tlb & TLB_ITLB) 255 tlb_itlb_store_slot(va, ctx, tte, slot); 256} 257 | |
258#endif /* !_MACHINE_TLB_H_ */ | 179#endif /* !_MACHINE_TLB_H_ */ |