1/* SPDX-License-Identifier: GPL-2.0-only */
2/*
3 * Support Intel uncore PerfMon discovery mechanism.
4 * Copyright(c) 2021 Intel Corporation.
5 */
6#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
7
8#include "uncore.h"
9#include "uncore_discovery.h"
10
11static struct rb_root discovery_tables = RB_ROOT;
12static int num_discovered_types[UNCORE_ACCESS_MAX];
13
14static bool has_generic_discovery_table(void)
15{
16	struct pci_dev *dev;
17	int dvsec;
18
19	dev = pci_get_device(PCI_VENDOR_ID_INTEL, UNCORE_DISCOVERY_TABLE_DEVICE, NULL);
20	if (!dev)
21		return false;
22
23	/* A discovery table device has the unique capability ID. */
24	dvsec = pci_find_next_ext_capability(dev, 0, UNCORE_EXT_CAP_ID_DISCOVERY);
25	pci_dev_put(dev);
26	if (dvsec)
27		return true;
28
29	return false;
30}
31
32static int logical_die_id;
33
34static int get_device_die_id(struct pci_dev *dev)
35{
36	int node = pcibus_to_node(dev->bus);
37
38	/*
39	 * If the NUMA info is not available, assume that the logical die id is
40	 * continuous in the order in which the discovery table devices are
41	 * detected.
42	 */
43	if (node < 0)
44		return logical_die_id++;
45
46	return uncore_device_to_die(dev);
47}
48
49#define __node_2_type(cur)	\
50	rb_entry((cur), struct intel_uncore_discovery_type, node)
51
52static inline int __type_cmp(const void *key, const struct rb_node *b)
53{
54	struct intel_uncore_discovery_type *type_b = __node_2_type(b);
55	const u16 *type_id = key;
56
57	if (type_b->type > *type_id)
58		return -1;
59	else if (type_b->type < *type_id)
60		return 1;
61
62	return 0;
63}
64
65static inline struct intel_uncore_discovery_type *
66search_uncore_discovery_type(u16 type_id)
67{
68	struct rb_node *node = rb_find(&type_id, &discovery_tables, __type_cmp);
69
70	return (node) ? __node_2_type(node) : NULL;
71}
72
73static inline bool __type_less(struct rb_node *a, const struct rb_node *b)
74{
75	return (__node_2_type(a)->type < __node_2_type(b)->type);
76}
77
78static struct intel_uncore_discovery_type *
79add_uncore_discovery_type(struct uncore_unit_discovery *unit)
80{
81	struct intel_uncore_discovery_type *type;
82
83	if (unit->access_type >= UNCORE_ACCESS_MAX) {
84		pr_warn("Unsupported access type %d\n", unit->access_type);
85		return NULL;
86	}
87
88	type = kzalloc(sizeof(struct intel_uncore_discovery_type), GFP_KERNEL);
89	if (!type)
90		return NULL;
91
92	type->box_ctrl_die = kcalloc(__uncore_max_dies, sizeof(u64), GFP_KERNEL);
93	if (!type->box_ctrl_die)
94		goto free_type;
95
96	type->access_type = unit->access_type;
97	num_discovered_types[type->access_type]++;
98	type->type = unit->box_type;
99
100	rb_add(&type->node, &discovery_tables, __type_less);
101
102	return type;
103
104free_type:
105	kfree(type);
106
107	return NULL;
108
109}
110
111static struct intel_uncore_discovery_type *
112get_uncore_discovery_type(struct uncore_unit_discovery *unit)
113{
114	struct intel_uncore_discovery_type *type;
115
116	type = search_uncore_discovery_type(unit->box_type);
117	if (type)
118		return type;
119
120	return add_uncore_discovery_type(unit);
121}
122
123static void
124uncore_insert_box_info(struct uncore_unit_discovery *unit,
125		       int die, bool parsed)
126{
127	struct intel_uncore_discovery_type *type;
128	unsigned int *ids;
129	u64 *box_offset;
130	int i;
131
132	if (!unit->ctl || !unit->ctl_offset || !unit->ctr_offset) {
133		pr_info("Invalid address is detected for uncore type %d box %d, "
134			"Disable the uncore unit.\n",
135			unit->box_type, unit->box_id);
136		return;
137	}
138
139	if (parsed) {
140		type = search_uncore_discovery_type(unit->box_type);
141		if (!type) {
142			pr_info("A spurious uncore type %d is detected, "
143				"Disable the uncore type.\n",
144				unit->box_type);
145			return;
146		}
147		/* Store the first box of each die */
148		if (!type->box_ctrl_die[die])
149			type->box_ctrl_die[die] = unit->ctl;
150		return;
151	}
152
153	type = get_uncore_discovery_type(unit);
154	if (!type)
155		return;
156
157	box_offset = kcalloc(type->num_boxes + 1, sizeof(u64), GFP_KERNEL);
158	if (!box_offset)
159		return;
160
161	ids = kcalloc(type->num_boxes + 1, sizeof(unsigned int), GFP_KERNEL);
162	if (!ids)
163		goto free_box_offset;
164
165	/* Store generic information for the first box */
166	if (!type->num_boxes) {
167		type->box_ctrl = unit->ctl;
168		type->box_ctrl_die[die] = unit->ctl;
169		type->num_counters = unit->num_regs;
170		type->counter_width = unit->bit_width;
171		type->ctl_offset = unit->ctl_offset;
172		type->ctr_offset = unit->ctr_offset;
173		*ids = unit->box_id;
174		goto end;
175	}
176
177	for (i = 0; i < type->num_boxes; i++) {
178		ids[i] = type->ids[i];
179		box_offset[i] = type->box_offset[i];
180
181		if (unit->box_id == ids[i]) {
182			pr_info("Duplicate uncore type %d box ID %d is detected, "
183				"Drop the duplicate uncore unit.\n",
184				unit->box_type, unit->box_id);
185			goto free_ids;
186		}
187	}
188	ids[i] = unit->box_id;
189	box_offset[i] = unit->ctl - type->box_ctrl;
190	kfree(type->ids);
191	kfree(type->box_offset);
192end:
193	type->ids = ids;
194	type->box_offset = box_offset;
195	type->num_boxes++;
196	return;
197
198free_ids:
199	kfree(ids);
200
201free_box_offset:
202	kfree(box_offset);
203
204}
205
206static bool
207uncore_ignore_unit(struct uncore_unit_discovery *unit, int *ignore)
208{
209	int i;
210
211	if (!ignore)
212		return false;
213
214	for (i = 0; ignore[i] != UNCORE_IGNORE_END ; i++) {
215		if (unit->box_type == ignore[i])
216			return true;
217	}
218
219	return false;
220}
221
222static int parse_discovery_table(struct pci_dev *dev, int die,
223				 u32 bar_offset, bool *parsed,
224				 int *ignore)
225{
226	struct uncore_global_discovery global;
227	struct uncore_unit_discovery unit;
228	void __iomem *io_addr;
229	resource_size_t addr;
230	unsigned long size;
231	u32 val;
232	int i;
233
234	pci_read_config_dword(dev, bar_offset, &val);
235
236	if (val & ~PCI_BASE_ADDRESS_MEM_MASK & ~PCI_BASE_ADDRESS_MEM_TYPE_64)
237		return -EINVAL;
238
239	addr = (resource_size_t)(val & PCI_BASE_ADDRESS_MEM_MASK);
240#ifdef CONFIG_PHYS_ADDR_T_64BIT
241	if ((val & PCI_BASE_ADDRESS_MEM_TYPE_MASK) == PCI_BASE_ADDRESS_MEM_TYPE_64) {
242		u32 val2;
243
244		pci_read_config_dword(dev, bar_offset + 4, &val2);
245		addr |= ((resource_size_t)val2) << 32;
246	}
247#endif
248	size = UNCORE_DISCOVERY_GLOBAL_MAP_SIZE;
249	io_addr = ioremap(addr, size);
250	if (!io_addr)
251		return -ENOMEM;
252
253	/* Read Global Discovery State */
254	memcpy_fromio(&global, io_addr, sizeof(struct uncore_global_discovery));
255	if (uncore_discovery_invalid_unit(global)) {
256		pr_info("Invalid Global Discovery State: 0x%llx 0x%llx 0x%llx\n",
257			global.table1, global.ctl, global.table3);
258		iounmap(io_addr);
259		return -EINVAL;
260	}
261	iounmap(io_addr);
262
263	size = (1 + global.max_units) * global.stride * 8;
264	io_addr = ioremap(addr, size);
265	if (!io_addr)
266		return -ENOMEM;
267
268	/* Parsing Unit Discovery State */
269	for (i = 0; i < global.max_units; i++) {
270		memcpy_fromio(&unit, io_addr + (i + 1) * (global.stride * 8),
271			      sizeof(struct uncore_unit_discovery));
272
273		if (uncore_discovery_invalid_unit(unit))
274			continue;
275
276		if (unit.access_type >= UNCORE_ACCESS_MAX)
277			continue;
278
279		if (uncore_ignore_unit(&unit, ignore))
280			continue;
281
282		uncore_insert_box_info(&unit, die, *parsed);
283	}
284
285	*parsed = true;
286	iounmap(io_addr);
287	return 0;
288}
289
290bool intel_uncore_has_discovery_tables(int *ignore)
291{
292	u32 device, val, entry_id, bar_offset;
293	int die, dvsec = 0, ret = true;
294	struct pci_dev *dev = NULL;
295	bool parsed = false;
296
297	if (has_generic_discovery_table())
298		device = UNCORE_DISCOVERY_TABLE_DEVICE;
299	else
300		device = PCI_ANY_ID;
301
302	/*
303	 * Start a new search and iterates through the list of
304	 * the discovery table devices.
305	 */
306	while ((dev = pci_get_device(PCI_VENDOR_ID_INTEL, device, dev)) != NULL) {
307		while ((dvsec = pci_find_next_ext_capability(dev, dvsec, UNCORE_EXT_CAP_ID_DISCOVERY))) {
308			pci_read_config_dword(dev, dvsec + UNCORE_DISCOVERY_DVSEC_OFFSET, &val);
309			entry_id = val & UNCORE_DISCOVERY_DVSEC_ID_MASK;
310			if (entry_id != UNCORE_DISCOVERY_DVSEC_ID_PMON)
311				continue;
312
313			pci_read_config_dword(dev, dvsec + UNCORE_DISCOVERY_DVSEC2_OFFSET, &val);
314
315			if (val & ~UNCORE_DISCOVERY_DVSEC2_BIR_MASK) {
316				ret = false;
317				goto err;
318			}
319			bar_offset = UNCORE_DISCOVERY_BIR_BASE +
320				     (val & UNCORE_DISCOVERY_DVSEC2_BIR_MASK) * UNCORE_DISCOVERY_BIR_STEP;
321
322			die = get_device_die_id(dev);
323			if (die < 0)
324				continue;
325
326			parse_discovery_table(dev, die, bar_offset, &parsed, ignore);
327		}
328	}
329
330	/* None of the discovery tables are available */
331	if (!parsed)
332		ret = false;
333err:
334	pci_dev_put(dev);
335
336	return ret;
337}
338
339void intel_uncore_clear_discovery_tables(void)
340{
341	struct intel_uncore_discovery_type *type, *next;
342
343	rbtree_postorder_for_each_entry_safe(type, next, &discovery_tables, node) {
344		kfree(type->box_ctrl_die);
345		kfree(type);
346	}
347}
348
349DEFINE_UNCORE_FORMAT_ATTR(event, event, "config:0-7");
350DEFINE_UNCORE_FORMAT_ATTR(umask, umask, "config:8-15");
351DEFINE_UNCORE_FORMAT_ATTR(edge, edge, "config:18");
352DEFINE_UNCORE_FORMAT_ATTR(inv, inv, "config:23");
353DEFINE_UNCORE_FORMAT_ATTR(thresh, thresh, "config:24-31");
354
355static struct attribute *generic_uncore_formats_attr[] = {
356	&format_attr_event.attr,
357	&format_attr_umask.attr,
358	&format_attr_edge.attr,
359	&format_attr_inv.attr,
360	&format_attr_thresh.attr,
361	NULL,
362};
363
364static const struct attribute_group generic_uncore_format_group = {
365	.name = "format",
366	.attrs = generic_uncore_formats_attr,
367};
368
369void intel_generic_uncore_msr_init_box(struct intel_uncore_box *box)
370{
371	wrmsrl(uncore_msr_box_ctl(box), GENERIC_PMON_BOX_CTL_INT);
372}
373
374void intel_generic_uncore_msr_disable_box(struct intel_uncore_box *box)
375{
376	wrmsrl(uncore_msr_box_ctl(box), GENERIC_PMON_BOX_CTL_FRZ);
377}
378
379void intel_generic_uncore_msr_enable_box(struct intel_uncore_box *box)
380{
381	wrmsrl(uncore_msr_box_ctl(box), 0);
382}
383
384static void intel_generic_uncore_msr_enable_event(struct intel_uncore_box *box,
385					    struct perf_event *event)
386{
387	struct hw_perf_event *hwc = &event->hw;
388
389	wrmsrl(hwc->config_base, hwc->config);
390}
391
392static void intel_generic_uncore_msr_disable_event(struct intel_uncore_box *box,
393					     struct perf_event *event)
394{
395	struct hw_perf_event *hwc = &event->hw;
396
397	wrmsrl(hwc->config_base, 0);
398}
399
400static struct intel_uncore_ops generic_uncore_msr_ops = {
401	.init_box		= intel_generic_uncore_msr_init_box,
402	.disable_box		= intel_generic_uncore_msr_disable_box,
403	.enable_box		= intel_generic_uncore_msr_enable_box,
404	.disable_event		= intel_generic_uncore_msr_disable_event,
405	.enable_event		= intel_generic_uncore_msr_enable_event,
406	.read_counter		= uncore_msr_read_counter,
407};
408
409void intel_generic_uncore_pci_init_box(struct intel_uncore_box *box)
410{
411	struct pci_dev *pdev = box->pci_dev;
412	int box_ctl = uncore_pci_box_ctl(box);
413
414	__set_bit(UNCORE_BOX_FLAG_CTL_OFFS8, &box->flags);
415	pci_write_config_dword(pdev, box_ctl, GENERIC_PMON_BOX_CTL_INT);
416}
417
418void intel_generic_uncore_pci_disable_box(struct intel_uncore_box *box)
419{
420	struct pci_dev *pdev = box->pci_dev;
421	int box_ctl = uncore_pci_box_ctl(box);
422
423	pci_write_config_dword(pdev, box_ctl, GENERIC_PMON_BOX_CTL_FRZ);
424}
425
426void intel_generic_uncore_pci_enable_box(struct intel_uncore_box *box)
427{
428	struct pci_dev *pdev = box->pci_dev;
429	int box_ctl = uncore_pci_box_ctl(box);
430
431	pci_write_config_dword(pdev, box_ctl, 0);
432}
433
434static void intel_generic_uncore_pci_enable_event(struct intel_uncore_box *box,
435					    struct perf_event *event)
436{
437	struct pci_dev *pdev = box->pci_dev;
438	struct hw_perf_event *hwc = &event->hw;
439
440	pci_write_config_dword(pdev, hwc->config_base, hwc->config);
441}
442
443void intel_generic_uncore_pci_disable_event(struct intel_uncore_box *box,
444					    struct perf_event *event)
445{
446	struct pci_dev *pdev = box->pci_dev;
447	struct hw_perf_event *hwc = &event->hw;
448
449	pci_write_config_dword(pdev, hwc->config_base, 0);
450}
451
452u64 intel_generic_uncore_pci_read_counter(struct intel_uncore_box *box,
453					  struct perf_event *event)
454{
455	struct pci_dev *pdev = box->pci_dev;
456	struct hw_perf_event *hwc = &event->hw;
457	u64 count = 0;
458
459	pci_read_config_dword(pdev, hwc->event_base, (u32 *)&count);
460	pci_read_config_dword(pdev, hwc->event_base + 4, (u32 *)&count + 1);
461
462	return count;
463}
464
465static struct intel_uncore_ops generic_uncore_pci_ops = {
466	.init_box	= intel_generic_uncore_pci_init_box,
467	.disable_box	= intel_generic_uncore_pci_disable_box,
468	.enable_box	= intel_generic_uncore_pci_enable_box,
469	.disable_event	= intel_generic_uncore_pci_disable_event,
470	.enable_event	= intel_generic_uncore_pci_enable_event,
471	.read_counter	= intel_generic_uncore_pci_read_counter,
472};
473
474#define UNCORE_GENERIC_MMIO_SIZE		0x4000
475
476static u64 generic_uncore_mmio_box_ctl(struct intel_uncore_box *box)
477{
478	struct intel_uncore_type *type = box->pmu->type;
479
480	if (!type->box_ctls || !type->box_ctls[box->dieid] || !type->mmio_offsets)
481		return 0;
482
483	return type->box_ctls[box->dieid] + type->mmio_offsets[box->pmu->pmu_idx];
484}
485
486void intel_generic_uncore_mmio_init_box(struct intel_uncore_box *box)
487{
488	u64 box_ctl = generic_uncore_mmio_box_ctl(box);
489	struct intel_uncore_type *type = box->pmu->type;
490	resource_size_t addr;
491
492	if (!box_ctl) {
493		pr_warn("Uncore type %d box %d: Invalid box control address.\n",
494			type->type_id, type->box_ids[box->pmu->pmu_idx]);
495		return;
496	}
497
498	addr = box_ctl;
499	box->io_addr = ioremap(addr, UNCORE_GENERIC_MMIO_SIZE);
500	if (!box->io_addr) {
501		pr_warn("Uncore type %d box %d: ioremap error for 0x%llx.\n",
502			type->type_id, type->box_ids[box->pmu->pmu_idx],
503			(unsigned long long)addr);
504		return;
505	}
506
507	writel(GENERIC_PMON_BOX_CTL_INT, box->io_addr);
508}
509
510void intel_generic_uncore_mmio_disable_box(struct intel_uncore_box *box)
511{
512	if (!box->io_addr)
513		return;
514
515	writel(GENERIC_PMON_BOX_CTL_FRZ, box->io_addr);
516}
517
518void intel_generic_uncore_mmio_enable_box(struct intel_uncore_box *box)
519{
520	if (!box->io_addr)
521		return;
522
523	writel(0, box->io_addr);
524}
525
526void intel_generic_uncore_mmio_enable_event(struct intel_uncore_box *box,
527					    struct perf_event *event)
528{
529	struct hw_perf_event *hwc = &event->hw;
530
531	if (!box->io_addr)
532		return;
533
534	writel(hwc->config, box->io_addr + hwc->config_base);
535}
536
537void intel_generic_uncore_mmio_disable_event(struct intel_uncore_box *box,
538					     struct perf_event *event)
539{
540	struct hw_perf_event *hwc = &event->hw;
541
542	if (!box->io_addr)
543		return;
544
545	writel(0, box->io_addr + hwc->config_base);
546}
547
548static struct intel_uncore_ops generic_uncore_mmio_ops = {
549	.init_box	= intel_generic_uncore_mmio_init_box,
550	.exit_box	= uncore_mmio_exit_box,
551	.disable_box	= intel_generic_uncore_mmio_disable_box,
552	.enable_box	= intel_generic_uncore_mmio_enable_box,
553	.disable_event	= intel_generic_uncore_mmio_disable_event,
554	.enable_event	= intel_generic_uncore_mmio_enable_event,
555	.read_counter	= uncore_mmio_read_counter,
556};
557
558static bool uncore_update_uncore_type(enum uncore_access_type type_id,
559				      struct intel_uncore_type *uncore,
560				      struct intel_uncore_discovery_type *type)
561{
562	uncore->type_id = type->type;
563	uncore->num_boxes = type->num_boxes;
564	uncore->num_counters = type->num_counters;
565	uncore->perf_ctr_bits = type->counter_width;
566	uncore->box_ids = type->ids;
567
568	switch (type_id) {
569	case UNCORE_ACCESS_MSR:
570		uncore->ops = &generic_uncore_msr_ops;
571		uncore->perf_ctr = (unsigned int)type->box_ctrl + type->ctr_offset;
572		uncore->event_ctl = (unsigned int)type->box_ctrl + type->ctl_offset;
573		uncore->box_ctl = (unsigned int)type->box_ctrl;
574		uncore->msr_offsets = type->box_offset;
575		break;
576	case UNCORE_ACCESS_PCI:
577		uncore->ops = &generic_uncore_pci_ops;
578		uncore->perf_ctr = (unsigned int)UNCORE_DISCOVERY_PCI_BOX_CTRL(type->box_ctrl) + type->ctr_offset;
579		uncore->event_ctl = (unsigned int)UNCORE_DISCOVERY_PCI_BOX_CTRL(type->box_ctrl) + type->ctl_offset;
580		uncore->box_ctl = (unsigned int)UNCORE_DISCOVERY_PCI_BOX_CTRL(type->box_ctrl);
581		uncore->box_ctls = type->box_ctrl_die;
582		uncore->pci_offsets = type->box_offset;
583		break;
584	case UNCORE_ACCESS_MMIO:
585		uncore->ops = &generic_uncore_mmio_ops;
586		uncore->perf_ctr = (unsigned int)type->ctr_offset;
587		uncore->event_ctl = (unsigned int)type->ctl_offset;
588		uncore->box_ctl = (unsigned int)type->box_ctrl;
589		uncore->box_ctls = type->box_ctrl_die;
590		uncore->mmio_offsets = type->box_offset;
591		uncore->mmio_map_size = UNCORE_GENERIC_MMIO_SIZE;
592		break;
593	default:
594		return false;
595	}
596
597	return true;
598}
599
600struct intel_uncore_type **
601intel_uncore_generic_init_uncores(enum uncore_access_type type_id, int num_extra)
602{
603	struct intel_uncore_discovery_type *type;
604	struct intel_uncore_type **uncores;
605	struct intel_uncore_type *uncore;
606	struct rb_node *node;
607	int i = 0;
608
609	uncores = kcalloc(num_discovered_types[type_id] + num_extra + 1,
610			  sizeof(struct intel_uncore_type *), GFP_KERNEL);
611	if (!uncores)
612		return empty_uncore;
613
614	for (node = rb_first(&discovery_tables); node; node = rb_next(node)) {
615		type = rb_entry(node, struct intel_uncore_discovery_type, node);
616		if (type->access_type != type_id)
617			continue;
618
619		uncore = kzalloc(sizeof(struct intel_uncore_type), GFP_KERNEL);
620		if (!uncore)
621			break;
622
623		uncore->event_mask = GENERIC_PMON_RAW_EVENT_MASK;
624		uncore->format_group = &generic_uncore_format_group;
625
626		if (!uncore_update_uncore_type(type_id, uncore, type)) {
627			kfree(uncore);
628			continue;
629		}
630		uncores[i++] = uncore;
631	}
632
633	return uncores;
634}
635
636void intel_uncore_generic_uncore_cpu_init(void)
637{
638	uncore_msr_uncores = intel_uncore_generic_init_uncores(UNCORE_ACCESS_MSR, 0);
639}
640
641int intel_uncore_generic_uncore_pci_init(void)
642{
643	uncore_pci_uncores = intel_uncore_generic_init_uncores(UNCORE_ACCESS_PCI, 0);
644
645	return 0;
646}
647
648void intel_uncore_generic_uncore_mmio_init(void)
649{
650	uncore_mmio_uncores = intel_uncore_generic_init_uncores(UNCORE_ACCESS_MMIO, 0);
651}
652