1/*- 2 * SPDX-License-Identifier: BSD-2-Clause 3 * 4 * Copyright (c) 2005 Peter Grehan 5 * Copyright (c) 2009 Nathan Whitehorn 6 * All rights reserved. 7 * 8 * Redistribution and use in source and binary forms, with or without 9 * modification, are permitted provided that the following conditions 10 * are met: 11 * 1. Redistributions of source code must retain the above copyright 12 * notice, this list of conditions and the following disclaimer. 13 * 2. Redistributions in binary form must reproduce the above copyright 14 * notice, this list of conditions and the following disclaimer in the 15 * documentation and/or other materials provided with the distribution. 16 * 17 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND 18 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 19 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 20 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE 21 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 22 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 23 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 24 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 25 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 26 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 27 * SUCH DAMAGE. 28 * 29 */ 30 31#include <sys/cdefs.h> 32/* 33 * Dispatch platform calls to the appropriate platform implementation 34 * through a previously registered kernel object. 35 */ 36 37#include <sys/param.h> 38#include <sys/kernel.h> 39#include <sys/lock.h> 40#include <sys/ktr.h> 41#include <sys/mutex.h> 42#include <sys/proc.h> 43#include <sys/systm.h> 44#include <sys/smp.h> 45#include <sys/sysctl.h> 46#include <sys/types.h> 47 48#include <vm/vm.h> 49#include <vm/vm_param.h> 50#include <vm/vm_page.h> 51#include <vm/vm_phys.h> 52 53#include <machine/cpu.h> 54#include <machine/md_var.h> 55#include <machine/ofw_machdep.h> 56#include <machine/platform.h> 57#include <machine/platformvar.h> 58#include <machine/smp.h> 59#include <machine/vmparam.h> 60 61#include "platform_if.h" 62 63static platform_def_t *plat_def_impl; 64static platform_t plat_obj; 65static struct kobj_ops plat_kernel_kops; 66static struct platform_kobj plat_kernel_obj; 67 68static char plat_name[64] = ""; 69SYSCTL_STRING(_hw, OID_AUTO, platform, CTLFLAG_RDTUN, 70 plat_name, 0, "Platform currently in use"); 71 72static struct mem_affinity mem_info[VM_PHYSSEG_MAX + 1]; 73static int vm_locality_table[MAXMEMDOM * MAXMEMDOM]; 74static struct mem_region pregions[PHYS_AVAIL_SZ]; 75static struct numa_mem_region numa_pregions[PHYS_AVAIL_SZ]; 76static struct mem_region aregions[PHYS_AVAIL_SZ]; 77static int nnumapregions, npregions, naregions; 78 79/* 80 * Memory region utilities: determine if two regions overlap, 81 * and merge two overlapping regions into one 82 */ 83static int 84memr_overlap(struct mem_region *r1, struct mem_region *r2) 85{ 86 if ((r1->mr_start + r1->mr_size) < r2->mr_start || 87 (r2->mr_start + r2->mr_size) < r1->mr_start) 88 return (FALSE); 89 90 return (TRUE); 91} 92 93static void 94memr_merge(struct mem_region *from, struct mem_region *to) 95{ 96 vm_offset_t end; 97 end = uqmax(to->mr_start + to->mr_size, from->mr_start + from->mr_size); 98 to->mr_start = uqmin(from->mr_start, to->mr_start); 99 to->mr_size = end - to->mr_start; 100} 101 102/* 103 * Quick sort callout for comparing memory regions. 104 */ 105static int 106mr_cmp(const void *a, const void *b) 107{ 108 const struct mem_region *regiona, *regionb; 109 110 regiona = a; 111 regionb = b; 112 if (regiona->mr_start < regionb->mr_start) 113 return (-1); 114 else if (regiona->mr_start > regionb->mr_start) 115 return (1); 116 else 117 return (0); 118} 119 120void 121numa_mem_regions(struct numa_mem_region **phys, int *physsz) 122{ 123 struct mem_affinity *mi; 124 int i, j, maxdom, ndomain, offset; 125 126 nnumapregions = 0; 127 PLATFORM_NUMA_MEM_REGIONS(plat_obj, numa_pregions, &nnumapregions); 128 129 if (physsz != NULL) 130 *physsz = nnumapregions; 131 if (phys != NULL) 132 *phys = numa_pregions; 133 if (physsz == NULL || phys == NULL) { 134 printf("unset value\n"); 135 return; 136 } 137 maxdom = 0; 138 for (i = 0; i < nnumapregions; i++) 139 if (numa_pregions[i].mr_domain > maxdom) 140 maxdom = numa_pregions[i].mr_domain; 141 142 mi = mem_info; 143 for (i = 0; i < nnumapregions; i++, mi++) { 144 mi->start = numa_pregions[i].mr_start; 145 mi->end = numa_pregions[i].mr_start + numa_pregions[i].mr_size; 146 mi->domain = numa_pregions[i].mr_domain; 147 } 148 offset = 0; 149 vm_locality_table[offset] = 10; 150 ndomain = maxdom + 1; 151 if (ndomain > 1) { 152 for (i = 0; i < ndomain; i++) { 153 for (j = 0; j < ndomain; j++) { 154 /* 155 * Not sure what these values should actually be 156 */ 157 if (i == j) 158 vm_locality_table[offset] = 10; 159 else 160 vm_locality_table[offset] = 21; 161 offset++; 162 } 163 } 164 } 165 vm_phys_register_domains(ndomain, mem_info, vm_locality_table); 166} 167 168void 169mem_regions(struct mem_region **phys, int *physsz, struct mem_region **avail, 170 int *availsz) 171{ 172 int i, j, still_merging; 173 174 if (npregions == 0) { 175 PLATFORM_MEM_REGIONS(plat_obj, pregions, &npregions, 176 aregions, &naregions); 177 qsort(pregions, npregions, sizeof(*pregions), mr_cmp); 178 qsort(aregions, naregions, sizeof(*aregions), mr_cmp); 179 180 /* Remove overlapping available regions */ 181 do { 182 still_merging = FALSE; 183 for (i = 0; i < naregions; i++) { 184 if (aregions[i].mr_size == 0) 185 continue; 186 for (j = i+1; j < naregions; j++) { 187 if (aregions[j].mr_size == 0) 188 continue; 189 if (!memr_overlap(&aregions[j], 190 &aregions[i])) 191 continue; 192 193 memr_merge(&aregions[j], &aregions[i]); 194 /* mark inactive */ 195 aregions[j].mr_size = 0; 196 still_merging = TRUE; 197 } 198 } 199 } while (still_merging == TRUE); 200 201 /* Collapse zero-length available regions */ 202 for (i = 0; i < naregions; i++) { 203 if (aregions[i].mr_size == 0) { 204 memcpy(&aregions[i], &aregions[i+1], 205 (naregions - i - 1)*sizeof(*aregions)); 206 naregions--; 207 i--; 208 } 209 } 210 } 211 212 if (phys != NULL) 213 *phys = pregions; 214 if (avail != NULL) 215 *avail = aregions; 216 if (physsz != NULL) 217 *physsz = npregions; 218 if (availsz != NULL) 219 *availsz = naregions; 220} 221 222int 223mem_valid(vm_offset_t addr, int len) 224{ 225 int i; 226 227 if (npregions == 0) { 228 struct mem_region *p, *a; 229 int na, np; 230 mem_regions(&p, &np, &a, &na); 231 } 232 233 for (i = 0; i < npregions; i++) 234 if ((addr >= pregions[i].mr_start) 235 && (addr + len <= pregions[i].mr_start + pregions[i].mr_size)) 236 return (0); 237 238 return (EFAULT); 239} 240 241vm_offset_t 242platform_real_maxaddr(void) 243{ 244 return (PLATFORM_REAL_MAXADDR(plat_obj)); 245} 246 247const char * 248installed_platform(void) 249{ 250 return (plat_def_impl->name); 251} 252 253u_long 254platform_timebase_freq(struct cpuref *cpu) 255{ 256 return (PLATFORM_TIMEBASE_FREQ(plat_obj, cpu)); 257} 258 259/* 260 * Put the current CPU, as last step in suspend, to sleep 261 */ 262void 263platform_sleep(void) 264{ 265 PLATFORM_SLEEP(plat_obj); 266} 267 268int 269platform_smp_first_cpu(struct cpuref *cpu) 270{ 271 return (PLATFORM_SMP_FIRST_CPU(plat_obj, cpu)); 272} 273 274int 275platform_smp_next_cpu(struct cpuref *cpu) 276{ 277 return (PLATFORM_SMP_NEXT_CPU(plat_obj, cpu)); 278} 279 280int 281platform_smp_get_bsp(struct cpuref *cpu) 282{ 283 return (PLATFORM_SMP_GET_BSP(plat_obj, cpu)); 284} 285 286int 287platform_smp_start_cpu(struct pcpu *cpu) 288{ 289 return (PLATFORM_SMP_START_CPU(plat_obj, cpu)); 290} 291 292void 293platform_smp_ap_init(void) 294{ 295 PLATFORM_SMP_AP_INIT(plat_obj); 296} 297 298void 299platform_smp_probe_threads(void) 300{ 301 PLATFORM_SMP_PROBE_THREADS(plat_obj); 302} 303 304#ifdef SMP 305struct cpu_group * 306cpu_topo(void) 307{ 308 return (PLATFORM_SMP_TOPO(plat_obj)); 309} 310#endif 311 312int 313platform_node_numa_domain(phandle_t node) 314{ 315 return (PLATFORM_NODE_NUMA_DOMAIN(plat_obj, node)); 316} 317 318/* 319 * Reset back to firmware. 320 */ 321void 322cpu_reset(void) 323{ 324 PLATFORM_RESET(plat_obj); 325} 326 327void platform_smp_timebase_sync(u_long tb, int ap) 328{ 329 330 PLATFORM_SMP_TIMEBASE_SYNC(plat_obj, tb, ap); 331} 332 333/* 334 * Platform install routines. Highest priority wins, using the same 335 * algorithm as bus attachment. 336 */ 337SET_DECLARE(platform_set, platform_def_t); 338 339void 340platform_probe_and_attach(void) 341{ 342 platform_def_t **platpp, *platp; 343 int prio, best_prio; 344 345 plat_obj = &plat_kernel_obj; 346 best_prio = 0; 347 348 /* 349 * Try to locate the best platform kobj 350 */ 351 SET_FOREACH(platpp, platform_set) { 352 platp = *platpp; 353 354 /* 355 * Take care of compiling the selected class, and 356 * then statically initialise the MMU object 357 */ 358 kobj_class_compile_static(platp, &plat_kernel_kops); 359 kobj_init_static((kobj_t)plat_obj, platp); 360 361 prio = PLATFORM_PROBE(plat_obj); 362 363 /* Check for errors */ 364 if (prio > 0) 365 continue; 366 367 /* 368 * Check if this module was specifically requested through 369 * the loader tunable we provide. 370 */ 371 if (strcmp(platp->name,plat_name) == 0) { 372 plat_def_impl = platp; 373 break; 374 } 375 376 /* Otherwise, see if it is better than our current best */ 377 if (plat_def_impl == NULL || prio > best_prio) { 378 best_prio = prio; 379 plat_def_impl = platp; 380 } 381 382 /* 383 * We can't free the KOBJ, since it is static. Reset the ops 384 * member of this class so that we can come back later. 385 */ 386 platp->ops = NULL; 387 } 388 389 if (plat_def_impl == NULL) 390 panic("No platform module found!"); 391 392 /* 393 * Recompile to make sure we ended with the 394 * correct one, and then attach. 395 */ 396 397 kobj_class_compile_static(plat_def_impl, &plat_kernel_kops); 398 kobj_init_static((kobj_t)plat_obj, plat_def_impl); 399 400 strlcpy(plat_name,plat_def_impl->name,sizeof(plat_name)); 401 402 PLATFORM_ATTACH(plat_obj); 403} 404