acpi_pci_layerscape_gen4.c revision 1.2
1/* $NetBSD: acpi_pci_layerscape_gen4.c,v 1.2 2020/02/02 16:44:25 jmcneill Exp $ */ 2 3/*- 4 * Copyright (c) 2020 The NetBSD Foundation, Inc. 5 * All rights reserved. 6 * 7 * This code is derived from software contributed to The NetBSD Foundation 8 * by Jared McNeill <jmcneill@invisible.ca>. 9 * 10 * Redistribution and use in source and binary forms, with or without 11 * modification, are permitted provided that the following conditions 12 * are met: 13 * 1. Redistributions of source code must retain the above copyright 14 * notice, this list of conditions and the following disclaimer. 15 * 2. Redistributions in binary form must reproduce the above copyright 16 * notice, this list of conditions and the following disclaimer in the 17 * documentation and/or other materials provided with the distribution. 18 * 19 * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS 20 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED 21 * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR 22 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS 23 * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR 24 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF 25 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS 26 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN 27 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) 28 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE 29 * POSSIBILITY OF SUCH DAMAGE. 30 */ 31 32/* 33 * NXP Layerscape PCIe Gen4 controller (not ECAM compliant) 34 */ 35 36#include <sys/cdefs.h> 37__KERNEL_RCSID(0, "$NetBSD: acpi_pci_layerscape_gen4.c,v 1.2 2020/02/02 16:44:25 jmcneill Exp $"); 38 39#include <sys/param.h> 40#include <sys/bus.h> 41#include <sys/device.h> 42#include <sys/intr.h> 43#include <sys/systm.h> 44#include <sys/kernel.h> 45#include <sys/extent.h> 46#include <sys/kmem.h> 47#include <sys/mutex.h> 48 49#include <machine/cpu.h> 50 51#include <dev/pci/pcireg.h> 52#include <dev/pci/pcivar.h> 53#include <dev/pci/pciconf.h> 54 55#include <dev/acpi/acpivar.h> 56#include <dev/acpi/acpi_pci.h> 57#include <dev/acpi/acpi_mcfg.h> 58 59#include <arm/acpi/acpi_pci_machdep.h> 60 61#define PAB_CTRL 0x808 62#define PAB_CTRL_PAGE_SEL __BITS(18,13) 63#define PAB_AXI_AMAP_PEX_WIN_L(x) (0xba8 + 0x10 * (x)) 64#define PAB_AXI_AMAP_PEX_WIN_H(x) (0xbac + 0x10 * (x)) 65#define INDIRECT_ADDR_BOUNDARY 0xc00 66 67#define LUT_BASE 0x80000 68#define LUT_GCR 0x28 69#define LUT_GCR_RRE __BIT(0) 70 71#define REG_TO_PAGE_INDEX(reg) (((reg) >> 10) & 0x3ff) 72#define REG_TO_PAGE_ADDR(reg) (((reg) & 0x3ff) | INDIRECT_ADDR_BOUNDARY) 73 74#define PAB_TARGET_BUS(b) ((b) << 24) 75#define PAB_TARGET_DEV(d) ((d) << 19) 76#define PAB_TARGET_FUNC(f) ((f) << 16) 77 78struct acpi_pci_layerscape_gen4 { 79 bus_space_tag_t bst; 80 bus_space_handle_t bsh; 81 bus_space_handle_t win_bsh; 82 uint8_t rev; 83 kmutex_t lock; 84}; 85 86static void 87acpi_pci_layerscape_gen4_ccsr_setpage(struct acpi_pci_layerscape_gen4 *pcie, u_int page_index) 88{ 89 uint32_t val; 90 91 val = bus_space_read_4(pcie->bst, pcie->bsh, PAB_CTRL); 92 val &= ~PAB_CTRL_PAGE_SEL; 93 val |= __SHIFTIN(page_index, PAB_CTRL_PAGE_SEL); 94 bus_space_write_4(pcie->bst, pcie->bsh, PAB_CTRL, val); 95} 96 97static uint32_t 98acpi_pci_layerscape_gen4_ccsr_read4(struct acpi_pci_layerscape_gen4 *pcie, bus_size_t reg) 99{ 100 const bool indirect = reg >= INDIRECT_ADDR_BOUNDARY; 101 const u_int page_index = indirect ? REG_TO_PAGE_INDEX(reg) : 0; 102 const bus_size_t page_addr = indirect ? REG_TO_PAGE_ADDR(reg) : reg; 103 104 acpi_pci_layerscape_gen4_ccsr_setpage(pcie, page_index); 105 return bus_space_read_4(pcie->bst, pcie->bsh, page_addr); 106} 107 108static void 109acpi_pci_layerscape_gen4_ccsr_write4(struct acpi_pci_layerscape_gen4 *pcie, 110 bus_size_t reg, pcireg_t data) 111{ 112 const bool indirect = reg >= INDIRECT_ADDR_BOUNDARY; 113 const u_int page_index = indirect ? REG_TO_PAGE_INDEX(reg) : 0; 114 const bus_size_t page_addr = indirect ? REG_TO_PAGE_ADDR(reg) : reg; 115 116 acpi_pci_layerscape_gen4_ccsr_setpage(pcie, page_index); 117 bus_space_write_4(pcie->bst, pcie->bsh, page_addr, data); 118} 119 120static void 121acpi_pci_layerscape_gen4_select_target(struct acpi_pci_layerscape_gen4 *pcie, 122 pci_chipset_tag_t pc, pcitag_t tag) 123{ 124 int b, d, f; 125 126 pci_decompose_tag(pc, tag, &b, &d, &f); 127 128 const uint32_t target = PAB_TARGET_BUS(b) | 129 PAB_TARGET_DEV(d) | PAB_TARGET_FUNC(f); 130 131 acpi_pci_layerscape_gen4_ccsr_write4(pcie, PAB_AXI_AMAP_PEX_WIN_L(0), target); 132 acpi_pci_layerscape_gen4_ccsr_write4(pcie, PAB_AXI_AMAP_PEX_WIN_H(0), 0); 133} 134 135static bool 136acpi_pci_layerscape_gen4_is_tag_okay(pci_chipset_tag_t pc, pcitag_t tag, int reg) 137{ 138 struct acpi_pci_context *ap = pc->pc_conf_v; 139 int b, d, f; 140 141 pci_decompose_tag(pc, tag, &b, &d, &f); 142 143 if (b <= ap->ap_bus + 1 && d > 0) 144 return false; 145 146 if (b != ap->ap_bus) 147 return acpimcfg_conf_valid(pc, tag, reg); 148 149 return true; 150} 151 152static int 153acpi_pci_layerscape_gen4_conf_read(pci_chipset_tag_t pc, pcitag_t tag, int reg, pcireg_t *data) 154{ 155 struct acpi_pci_context *ap = pc->pc_conf_v; 156 struct acpi_pci_layerscape_gen4 *pcie = ap->ap_conf_priv; 157 int b, d, f; 158 159 pci_decompose_tag(pc, tag, &b, &d, &f); 160 161 if (!acpi_pci_layerscape_gen4_is_tag_okay(pc, tag, reg)) { 162 *data = -1; 163 return EINVAL; 164 } 165 166 mutex_enter(&pcie->lock); 167 168 if (pcie->rev == 0x10 && reg == PCI_ID_REG) 169 bus_space_write_4(pcie->bst, pcie->bsh, LUT_BASE + LUT_GCR, 0); 170 171 if (b == ap->ap_bus) { 172 *data = acpi_pci_layerscape_gen4_ccsr_read4(pcie, reg); 173 } else { 174 acpi_pci_layerscape_gen4_select_target(pcie, pc, tag); 175 *data = bus_space_read_4(pcie->bst, pcie->win_bsh, reg); 176 } 177 178 if (pcie->rev == 0x10 && reg == PCI_ID_REG) 179 bus_space_write_4(pcie->bst, pcie->bsh, LUT_BASE + LUT_GCR, LUT_GCR_RRE); 180 181 mutex_exit(&pcie->lock); 182 183 return 0; 184} 185 186static int 187acpi_pci_layerscape_gen4_conf_write(pci_chipset_tag_t pc, pcitag_t tag, int reg, pcireg_t data) 188{ 189 struct acpi_pci_context *ap = pc->pc_conf_v; 190 struct acpi_pci_layerscape_gen4 *pcie = ap->ap_conf_priv; 191 int b, d, f; 192 193 pci_decompose_tag(pc, tag, &b, &d, &f); 194 195 if (!acpi_pci_layerscape_gen4_is_tag_okay(pc, tag, reg)) 196 return EINVAL; 197 198 mutex_enter(&pcie->lock); 199 200 if (b == ap->ap_bus) { 201 acpi_pci_layerscape_gen4_ccsr_write4(pcie, reg, data); 202 } else { 203 acpi_pci_layerscape_gen4_select_target(pcie, pc, tag); 204 bus_space_write_4(pcie->bst, pcie->win_bsh, reg, data); 205 } 206 207 mutex_exit(&pcie->lock); 208 209 return 0; 210} 211 212static UINT64 213acpi_pci_layerscape_win_base(ACPI_INTEGER seg) 214{ 215 ACPI_TABLE_MCFG *mcfg; 216 ACPI_MCFG_ALLOCATION *ama; 217 ACPI_STATUS rv; 218 uint32_t off; 219 int i; 220 221 rv = AcpiGetTable(ACPI_SIG_MCFG, 0, (ACPI_TABLE_HEADER **)&mcfg); 222 if (ACPI_FAILURE(rv)) 223 return 0; 224 225 off = sizeof(ACPI_TABLE_MCFG); 226 ama = ACPI_ADD_PTR(ACPI_MCFG_ALLOCATION, mcfg, off); 227 for (i = 0; off + sizeof(ACPI_MCFG_ALLOCATION) <= mcfg->Header.Length; i++) { 228 if (ama->PciSegment == seg) 229 return ama->Address; 230 off += sizeof(ACPI_MCFG_ALLOCATION); 231 ama = ACPI_ADD_PTR(ACPI_MCFG_ALLOCATION, mcfg, off); 232 } 233 234 return 0; /* not found */ 235} 236 237static ACPI_STATUS 238acpi_pci_layerscape_gen4_map(ACPI_HANDLE handle, UINT32 level, void *ctx, void **retval) 239{ 240 struct acpi_pci_context *ap = ctx; 241 struct acpi_resources res; 242 struct acpi_mem *mem; 243 struct acpi_pci_layerscape_gen4 *pcie; 244 bus_space_handle_t bsh; 245 ACPI_HANDLE parent; 246 ACPI_INTEGER seg; 247 ACPI_STATUS rv; 248 UINT64 win_base; 249 int error; 250 251 rv = AcpiGetParent(handle, &parent); 252 if (ACPI_FAILURE(rv)) 253 return rv; 254 rv = acpi_eval_integer(parent, "_SEG", &seg); 255 if (ACPI_FAILURE(rv)) 256 seg = 0; 257 if (ap->ap_seg != seg) 258 return AE_OK; 259 260 rv = acpi_resource_parse(ap->ap_dev, handle, "_CRS", &res, &acpi_resource_parse_ops_quiet); 261 if (ACPI_FAILURE(rv)) 262 return rv; 263 264 mem = acpi_res_mem(&res, 0); 265 if (mem == NULL) { 266 acpi_resource_cleanup(&res); 267 return AE_NOT_FOUND; 268 } 269 270 win_base = acpi_pci_layerscape_win_base(seg); 271 if (win_base == 0) { 272 aprint_error_dev(ap->ap_dev, "couldn't find MCFG entry for segment %ld\n", seg); 273 return AE_NOT_FOUND; 274 } 275 276 error = bus_space_map(ap->ap_bst, mem->ar_base, mem->ar_length, 277 _ARM_BUS_SPACE_MAP_STRONGLY_ORDERED, &bsh); 278 if (error != 0) 279 return AE_NO_MEMORY; 280 281 pcie = kmem_alloc(sizeof(*pcie), KM_SLEEP); 282 pcie->bst = ap->ap_bst; 283 pcie->bsh = bsh; 284 mutex_init(&pcie->lock, MUTEX_DEFAULT, IPL_HIGH); 285 286 error = bus_space_map(ap->ap_bst, win_base, PCI_EXTCONF_SIZE, 287 _ARM_BUS_SPACE_MAP_STRONGLY_ORDERED, &pcie->win_bsh); 288 if (error != 0) 289 return AE_NO_MEMORY; 290 291 const pcireg_t cr = bus_space_read_4(pcie->bst, pcie->bsh, PCI_CLASS_REG); 292 pcie->rev = PCI_REVISION(cr); 293 294 ap->ap_conf_read = acpi_pci_layerscape_gen4_conf_read; 295 ap->ap_conf_write = acpi_pci_layerscape_gen4_conf_write; 296 ap->ap_conf_priv = pcie; 297 298 aprint_verbose_dev(ap->ap_dev, 299 "PCIe segment %lu: Layerscape Gen4 rev. %#x found at %#lx-%#lx\n", 300 seg, pcie->rev, mem->ar_base, mem->ar_base + mem->ar_length - 1); 301 302 return AE_CTRL_TERMINATE; 303} 304 305void 306acpi_pci_layerscape_gen4_init(struct acpi_pci_context *ap) 307{ 308 ACPI_STATUS rv; 309 310 rv = AcpiGetDevices(__UNCONST("NXP0016"), acpi_pci_layerscape_gen4_map, ap, NULL); 311 if (ACPI_FAILURE(rv)) 312 return; 313} 314