1// SPDX-License-Identifier: GPL-2.0
2/*
3 * Copyright(C) 2015 Linaro Limited. All rights reserved.
4 * Author: Mathieu Poirier <mathieu.poirier@linaro.org>
5 */
6
7#include <linux/pid_namespace.h>
8#include <linux/pm_runtime.h>
9#include <linux/sysfs.h>
10#include "coresight-etm4x.h"
11#include "coresight-priv.h"
12#include "coresight-syscfg.h"
13
14static int etm4_set_mode_exclude(struct etmv4_drvdata *drvdata, bool exclude)
15{
16	u8 idx;
17	struct etmv4_config *config = &drvdata->config;
18
19	idx = config->addr_idx;
20
21	/*
22	 * TRCACATRn.TYPE bit[1:0]: type of comparison
23	 * the trace unit performs
24	 */
25	if (FIELD_GET(TRCACATRn_TYPE_MASK, config->addr_acc[idx]) == TRCACATRn_TYPE_ADDR) {
26		if (idx % 2 != 0)
27			return -EINVAL;
28
29		/*
30		 * We are performing instruction address comparison. Set the
31		 * relevant bit of ViewInst Include/Exclude Control register
32		 * for corresponding address comparator pair.
33		 */
34		if (config->addr_type[idx] != ETM_ADDR_TYPE_RANGE ||
35		    config->addr_type[idx + 1] != ETM_ADDR_TYPE_RANGE)
36			return -EINVAL;
37
38		if (exclude == true) {
39			/*
40			 * Set exclude bit and unset the include bit
41			 * corresponding to comparator pair
42			 */
43			config->viiectlr |= BIT(idx / 2 + 16);
44			config->viiectlr &= ~BIT(idx / 2);
45		} else {
46			/*
47			 * Set include bit and unset exclude bit
48			 * corresponding to comparator pair
49			 */
50			config->viiectlr |= BIT(idx / 2);
51			config->viiectlr &= ~BIT(idx / 2 + 16);
52		}
53	}
54	return 0;
55}
56
57static ssize_t nr_pe_cmp_show(struct device *dev,
58			      struct device_attribute *attr,
59			      char *buf)
60{
61	unsigned long val;
62	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
63
64	val = drvdata->nr_pe_cmp;
65	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
66}
67static DEVICE_ATTR_RO(nr_pe_cmp);
68
69static ssize_t nr_addr_cmp_show(struct device *dev,
70				struct device_attribute *attr,
71				char *buf)
72{
73	unsigned long val;
74	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
75
76	val = drvdata->nr_addr_cmp;
77	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
78}
79static DEVICE_ATTR_RO(nr_addr_cmp);
80
81static ssize_t nr_cntr_show(struct device *dev,
82			    struct device_attribute *attr,
83			    char *buf)
84{
85	unsigned long val;
86	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
87
88	val = drvdata->nr_cntr;
89	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
90}
91static DEVICE_ATTR_RO(nr_cntr);
92
93static ssize_t nr_ext_inp_show(struct device *dev,
94			       struct device_attribute *attr,
95			       char *buf)
96{
97	unsigned long val;
98	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
99
100	val = drvdata->nr_ext_inp;
101	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
102}
103static DEVICE_ATTR_RO(nr_ext_inp);
104
105static ssize_t numcidc_show(struct device *dev,
106			    struct device_attribute *attr,
107			    char *buf)
108{
109	unsigned long val;
110	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
111
112	val = drvdata->numcidc;
113	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
114}
115static DEVICE_ATTR_RO(numcidc);
116
117static ssize_t numvmidc_show(struct device *dev,
118			     struct device_attribute *attr,
119			     char *buf)
120{
121	unsigned long val;
122	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
123
124	val = drvdata->numvmidc;
125	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
126}
127static DEVICE_ATTR_RO(numvmidc);
128
129static ssize_t nrseqstate_show(struct device *dev,
130			       struct device_attribute *attr,
131			       char *buf)
132{
133	unsigned long val;
134	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
135
136	val = drvdata->nrseqstate;
137	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
138}
139static DEVICE_ATTR_RO(nrseqstate);
140
141static ssize_t nr_resource_show(struct device *dev,
142				struct device_attribute *attr,
143				char *buf)
144{
145	unsigned long val;
146	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
147
148	val = drvdata->nr_resource;
149	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
150}
151static DEVICE_ATTR_RO(nr_resource);
152
153static ssize_t nr_ss_cmp_show(struct device *dev,
154			      struct device_attribute *attr,
155			      char *buf)
156{
157	unsigned long val;
158	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
159
160	val = drvdata->nr_ss_cmp;
161	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
162}
163static DEVICE_ATTR_RO(nr_ss_cmp);
164
165static ssize_t reset_store(struct device *dev,
166			   struct device_attribute *attr,
167			   const char *buf, size_t size)
168{
169	int i;
170	unsigned long val;
171	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
172	struct etmv4_config *config = &drvdata->config;
173
174	if (kstrtoul(buf, 16, &val))
175		return -EINVAL;
176
177	spin_lock(&drvdata->spinlock);
178	if (val)
179		config->mode = 0x0;
180
181	/* Disable data tracing: do not trace load and store data transfers */
182	config->mode &= ~(ETM_MODE_LOAD | ETM_MODE_STORE);
183	config->cfg &= ~(TRCCONFIGR_INSTP0_LOAD | TRCCONFIGR_INSTP0_STORE);
184
185	/* Disable data value and data address tracing */
186	config->mode &= ~(ETM_MODE_DATA_TRACE_ADDR |
187			   ETM_MODE_DATA_TRACE_VAL);
188	config->cfg &= ~(TRCCONFIGR_DA | TRCCONFIGR_DV);
189
190	/* Disable all events tracing */
191	config->eventctrl0 = 0x0;
192	config->eventctrl1 = 0x0;
193
194	/* Disable timestamp event */
195	config->ts_ctrl = 0x0;
196
197	/* Disable stalling */
198	config->stall_ctrl = 0x0;
199
200	/* Reset trace synchronization period  to 2^8 = 256 bytes*/
201	if (drvdata->syncpr == false)
202		config->syncfreq = 0x8;
203
204	/*
205	 * Enable ViewInst to trace everything with start-stop logic in
206	 * started state. ARM recommends start-stop logic is set before
207	 * each trace run.
208	 */
209	config->vinst_ctrl = FIELD_PREP(TRCVICTLR_EVENT_MASK, 0x01);
210	if (drvdata->nr_addr_cmp > 0) {
211		config->mode |= ETM_MODE_VIEWINST_STARTSTOP;
212		/* SSSTATUS, bit[9] */
213		config->vinst_ctrl |= TRCVICTLR_SSSTATUS;
214	}
215
216	/* No address range filtering for ViewInst */
217	config->viiectlr = 0x0;
218
219	/* No start-stop filtering for ViewInst */
220	config->vissctlr = 0x0;
221	config->vipcssctlr = 0x0;
222
223	/* Disable seq events */
224	for (i = 0; i < drvdata->nrseqstate-1; i++)
225		config->seq_ctrl[i] = 0x0;
226	config->seq_rst = 0x0;
227	config->seq_state = 0x0;
228
229	/* Disable external input events */
230	config->ext_inp = 0x0;
231
232	config->cntr_idx = 0x0;
233	for (i = 0; i < drvdata->nr_cntr; i++) {
234		config->cntrldvr[i] = 0x0;
235		config->cntr_ctrl[i] = 0x0;
236		config->cntr_val[i] = 0x0;
237	}
238
239	config->res_idx = 0x0;
240	for (i = 2; i < 2 * drvdata->nr_resource; i++)
241		config->res_ctrl[i] = 0x0;
242
243	config->ss_idx = 0x0;
244	for (i = 0; i < drvdata->nr_ss_cmp; i++) {
245		config->ss_ctrl[i] = 0x0;
246		config->ss_pe_cmp[i] = 0x0;
247	}
248
249	config->addr_idx = 0x0;
250	for (i = 0; i < drvdata->nr_addr_cmp * 2; i++) {
251		config->addr_val[i] = 0x0;
252		config->addr_acc[i] = 0x0;
253		config->addr_type[i] = ETM_ADDR_TYPE_NONE;
254	}
255
256	config->ctxid_idx = 0x0;
257	for (i = 0; i < drvdata->numcidc; i++)
258		config->ctxid_pid[i] = 0x0;
259
260	config->ctxid_mask0 = 0x0;
261	config->ctxid_mask1 = 0x0;
262
263	config->vmid_idx = 0x0;
264	for (i = 0; i < drvdata->numvmidc; i++)
265		config->vmid_val[i] = 0x0;
266	config->vmid_mask0 = 0x0;
267	config->vmid_mask1 = 0x0;
268
269	spin_unlock(&drvdata->spinlock);
270
271	/* for sysfs - only release trace id when resetting */
272	etm4_release_trace_id(drvdata);
273
274	cscfg_csdev_reset_feats(to_coresight_device(dev));
275
276	return size;
277}
278static DEVICE_ATTR_WO(reset);
279
280static ssize_t mode_show(struct device *dev,
281			 struct device_attribute *attr,
282			 char *buf)
283{
284	unsigned long val;
285	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
286	struct etmv4_config *config = &drvdata->config;
287
288	val = config->mode;
289	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
290}
291
292static ssize_t mode_store(struct device *dev,
293			  struct device_attribute *attr,
294			  const char *buf, size_t size)
295{
296	unsigned long val, mode;
297	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
298	struct etmv4_config *config = &drvdata->config;
299
300	if (kstrtoul(buf, 16, &val))
301		return -EINVAL;
302
303	spin_lock(&drvdata->spinlock);
304	config->mode = val & ETMv4_MODE_ALL;
305
306	if (drvdata->instrp0 == true) {
307		/* start by clearing instruction P0 field */
308		config->cfg  &= ~TRCCONFIGR_INSTP0_LOAD_STORE;
309		if (config->mode & ETM_MODE_LOAD)
310			/* 0b01 Trace load instructions as P0 instructions */
311			config->cfg  |= TRCCONFIGR_INSTP0_LOAD;
312		if (config->mode & ETM_MODE_STORE)
313			/* 0b10 Trace store instructions as P0 instructions */
314			config->cfg  |= TRCCONFIGR_INSTP0_STORE;
315		if (config->mode & ETM_MODE_LOAD_STORE)
316			/*
317			 * 0b11 Trace load and store instructions
318			 * as P0 instructions
319			 */
320			config->cfg  |= TRCCONFIGR_INSTP0_LOAD_STORE;
321	}
322
323	/* bit[3], Branch broadcast mode */
324	if ((config->mode & ETM_MODE_BB) && (drvdata->trcbb == true))
325		config->cfg |= TRCCONFIGR_BB;
326	else
327		config->cfg &= ~TRCCONFIGR_BB;
328
329	/* bit[4], Cycle counting instruction trace bit */
330	if ((config->mode & ETMv4_MODE_CYCACC) &&
331		(drvdata->trccci == true))
332		config->cfg |= TRCCONFIGR_CCI;
333	else
334		config->cfg &= ~TRCCONFIGR_CCI;
335
336	/* bit[6], Context ID tracing bit */
337	if ((config->mode & ETMv4_MODE_CTXID) && (drvdata->ctxid_size))
338		config->cfg |= TRCCONFIGR_CID;
339	else
340		config->cfg &= ~TRCCONFIGR_CID;
341
342	if ((config->mode & ETM_MODE_VMID) && (drvdata->vmid_size))
343		config->cfg |= TRCCONFIGR_VMID;
344	else
345		config->cfg &= ~TRCCONFIGR_VMID;
346
347	/* bits[10:8], Conditional instruction tracing bit */
348	mode = ETM_MODE_COND(config->mode);
349	if (drvdata->trccond == true) {
350		config->cfg &= ~TRCCONFIGR_COND_MASK;
351		config->cfg |= mode << __bf_shf(TRCCONFIGR_COND_MASK);
352	}
353
354	/* bit[11], Global timestamp tracing bit */
355	if ((config->mode & ETMv4_MODE_TIMESTAMP) && (drvdata->ts_size))
356		config->cfg |= TRCCONFIGR_TS;
357	else
358		config->cfg &= ~TRCCONFIGR_TS;
359
360	/* bit[12], Return stack enable bit */
361	if ((config->mode & ETM_MODE_RETURNSTACK) &&
362					(drvdata->retstack == true))
363		config->cfg |= TRCCONFIGR_RS;
364	else
365		config->cfg &= ~TRCCONFIGR_RS;
366
367	/* bits[14:13], Q element enable field */
368	mode = ETM_MODE_QELEM(config->mode);
369	/* start by clearing QE bits */
370	config->cfg &= ~(TRCCONFIGR_QE_W_COUNTS | TRCCONFIGR_QE_WO_COUNTS);
371	/*
372	 * if supported, Q elements with instruction counts are enabled.
373	 * Always set the low bit for any requested mode. Valid combos are
374	 * 0b00, 0b01 and 0b11.
375	 */
376	if (mode && drvdata->q_support)
377		config->cfg |= TRCCONFIGR_QE_W_COUNTS;
378	/*
379	 * if supported, Q elements with and without instruction
380	 * counts are enabled
381	 */
382	if ((mode & BIT(1)) && (drvdata->q_support & BIT(1)))
383		config->cfg |= TRCCONFIGR_QE_WO_COUNTS;
384
385	/* bit[11], AMBA Trace Bus (ATB) trigger enable bit */
386	if ((config->mode & ETM_MODE_ATB_TRIGGER) &&
387	    (drvdata->atbtrig == true))
388		config->eventctrl1 |= TRCEVENTCTL1R_ATB;
389	else
390		config->eventctrl1 &= ~TRCEVENTCTL1R_ATB;
391
392	/* bit[12], Low-power state behavior override bit */
393	if ((config->mode & ETM_MODE_LPOVERRIDE) &&
394	    (drvdata->lpoverride == true))
395		config->eventctrl1 |= TRCEVENTCTL1R_LPOVERRIDE;
396	else
397		config->eventctrl1 &= ~TRCEVENTCTL1R_LPOVERRIDE;
398
399	/* bit[8], Instruction stall bit */
400	if ((config->mode & ETM_MODE_ISTALL_EN) && (drvdata->stallctl == true))
401		config->stall_ctrl |= TRCSTALLCTLR_ISTALL;
402	else
403		config->stall_ctrl &= ~TRCSTALLCTLR_ISTALL;
404
405	/* bit[10], Prioritize instruction trace bit */
406	if (config->mode & ETM_MODE_INSTPRIO)
407		config->stall_ctrl |= TRCSTALLCTLR_INSTPRIORITY;
408	else
409		config->stall_ctrl &= ~TRCSTALLCTLR_INSTPRIORITY;
410
411	/* bit[13], Trace overflow prevention bit */
412	if ((config->mode & ETM_MODE_NOOVERFLOW) &&
413		(drvdata->nooverflow == true))
414		config->stall_ctrl |= TRCSTALLCTLR_NOOVERFLOW;
415	else
416		config->stall_ctrl &= ~TRCSTALLCTLR_NOOVERFLOW;
417
418	/* bit[9] Start/stop logic control bit */
419	if (config->mode & ETM_MODE_VIEWINST_STARTSTOP)
420		config->vinst_ctrl |= TRCVICTLR_SSSTATUS;
421	else
422		config->vinst_ctrl &= ~TRCVICTLR_SSSTATUS;
423
424	/* bit[10], Whether a trace unit must trace a Reset exception */
425	if (config->mode & ETM_MODE_TRACE_RESET)
426		config->vinst_ctrl |= TRCVICTLR_TRCRESET;
427	else
428		config->vinst_ctrl &= ~TRCVICTLR_TRCRESET;
429
430	/* bit[11], Whether a trace unit must trace a system error exception */
431	if ((config->mode & ETM_MODE_TRACE_ERR) &&
432		(drvdata->trc_error == true))
433		config->vinst_ctrl |= TRCVICTLR_TRCERR;
434	else
435		config->vinst_ctrl &= ~TRCVICTLR_TRCERR;
436
437	if (config->mode & (ETM_MODE_EXCL_KERN | ETM_MODE_EXCL_USER))
438		etm4_config_trace_mode(config);
439
440	spin_unlock(&drvdata->spinlock);
441
442	return size;
443}
444static DEVICE_ATTR_RW(mode);
445
446static ssize_t pe_show(struct device *dev,
447		       struct device_attribute *attr,
448		       char *buf)
449{
450	unsigned long val;
451	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
452	struct etmv4_config *config = &drvdata->config;
453
454	val = config->pe_sel;
455	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
456}
457
458static ssize_t pe_store(struct device *dev,
459			struct device_attribute *attr,
460			const char *buf, size_t size)
461{
462	unsigned long val;
463	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
464	struct etmv4_config *config = &drvdata->config;
465
466	if (kstrtoul(buf, 16, &val))
467		return -EINVAL;
468
469	spin_lock(&drvdata->spinlock);
470	if (val > drvdata->nr_pe) {
471		spin_unlock(&drvdata->spinlock);
472		return -EINVAL;
473	}
474
475	config->pe_sel = val;
476	spin_unlock(&drvdata->spinlock);
477	return size;
478}
479static DEVICE_ATTR_RW(pe);
480
481static ssize_t event_show(struct device *dev,
482			  struct device_attribute *attr,
483			  char *buf)
484{
485	unsigned long val;
486	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
487	struct etmv4_config *config = &drvdata->config;
488
489	val = config->eventctrl0;
490	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
491}
492
493static ssize_t event_store(struct device *dev,
494			   struct device_attribute *attr,
495			   const char *buf, size_t size)
496{
497	unsigned long val;
498	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
499	struct etmv4_config *config = &drvdata->config;
500
501	if (kstrtoul(buf, 16, &val))
502		return -EINVAL;
503
504	spin_lock(&drvdata->spinlock);
505	switch (drvdata->nr_event) {
506	case 0x0:
507		/* EVENT0, bits[7:0] */
508		config->eventctrl0 = val & 0xFF;
509		break;
510	case 0x1:
511		 /* EVENT1, bits[15:8] */
512		config->eventctrl0 = val & 0xFFFF;
513		break;
514	case 0x2:
515		/* EVENT2, bits[23:16] */
516		config->eventctrl0 = val & 0xFFFFFF;
517		break;
518	case 0x3:
519		/* EVENT3, bits[31:24] */
520		config->eventctrl0 = val;
521		break;
522	default:
523		break;
524	}
525	spin_unlock(&drvdata->spinlock);
526	return size;
527}
528static DEVICE_ATTR_RW(event);
529
530static ssize_t event_instren_show(struct device *dev,
531				  struct device_attribute *attr,
532				  char *buf)
533{
534	unsigned long val;
535	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
536	struct etmv4_config *config = &drvdata->config;
537
538	val = FIELD_GET(TRCEVENTCTL1R_INSTEN_MASK, config->eventctrl1);
539	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
540}
541
542static ssize_t event_instren_store(struct device *dev,
543				   struct device_attribute *attr,
544				   const char *buf, size_t size)
545{
546	unsigned long val;
547	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
548	struct etmv4_config *config = &drvdata->config;
549
550	if (kstrtoul(buf, 16, &val))
551		return -EINVAL;
552
553	spin_lock(&drvdata->spinlock);
554	/* start by clearing all instruction event enable bits */
555	config->eventctrl1 &= ~TRCEVENTCTL1R_INSTEN_MASK;
556	switch (drvdata->nr_event) {
557	case 0x0:
558		/* generate Event element for event 1 */
559		config->eventctrl1 |= val & TRCEVENTCTL1R_INSTEN_1;
560		break;
561	case 0x1:
562		/* generate Event element for event 1 and 2 */
563		config->eventctrl1 |= val & (TRCEVENTCTL1R_INSTEN_0 | TRCEVENTCTL1R_INSTEN_1);
564		break;
565	case 0x2:
566		/* generate Event element for event 1, 2 and 3 */
567		config->eventctrl1 |= val & (TRCEVENTCTL1R_INSTEN_0 |
568					     TRCEVENTCTL1R_INSTEN_1 |
569					     TRCEVENTCTL1R_INSTEN_2);
570		break;
571	case 0x3:
572		/* generate Event element for all 4 events */
573		config->eventctrl1 |= val & (TRCEVENTCTL1R_INSTEN_0 |
574					     TRCEVENTCTL1R_INSTEN_1 |
575					     TRCEVENTCTL1R_INSTEN_2 |
576					     TRCEVENTCTL1R_INSTEN_3);
577		break;
578	default:
579		break;
580	}
581	spin_unlock(&drvdata->spinlock);
582	return size;
583}
584static DEVICE_ATTR_RW(event_instren);
585
586static ssize_t event_ts_show(struct device *dev,
587			     struct device_attribute *attr,
588			     char *buf)
589{
590	unsigned long val;
591	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
592	struct etmv4_config *config = &drvdata->config;
593
594	val = config->ts_ctrl;
595	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
596}
597
598static ssize_t event_ts_store(struct device *dev,
599			      struct device_attribute *attr,
600			      const char *buf, size_t size)
601{
602	unsigned long val;
603	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
604	struct etmv4_config *config = &drvdata->config;
605
606	if (kstrtoul(buf, 16, &val))
607		return -EINVAL;
608	if (!drvdata->ts_size)
609		return -EINVAL;
610
611	config->ts_ctrl = val & ETMv4_EVENT_MASK;
612	return size;
613}
614static DEVICE_ATTR_RW(event_ts);
615
616static ssize_t syncfreq_show(struct device *dev,
617			     struct device_attribute *attr,
618			     char *buf)
619{
620	unsigned long val;
621	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
622	struct etmv4_config *config = &drvdata->config;
623
624	val = config->syncfreq;
625	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
626}
627
628static ssize_t syncfreq_store(struct device *dev,
629			      struct device_attribute *attr,
630			      const char *buf, size_t size)
631{
632	unsigned long val;
633	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
634	struct etmv4_config *config = &drvdata->config;
635
636	if (kstrtoul(buf, 16, &val))
637		return -EINVAL;
638	if (drvdata->syncpr == true)
639		return -EINVAL;
640
641	config->syncfreq = val & ETMv4_SYNC_MASK;
642	return size;
643}
644static DEVICE_ATTR_RW(syncfreq);
645
646static ssize_t cyc_threshold_show(struct device *dev,
647				  struct device_attribute *attr,
648				  char *buf)
649{
650	unsigned long val;
651	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
652	struct etmv4_config *config = &drvdata->config;
653
654	val = config->ccctlr;
655	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
656}
657
658static ssize_t cyc_threshold_store(struct device *dev,
659				   struct device_attribute *attr,
660				   const char *buf, size_t size)
661{
662	unsigned long val;
663	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
664	struct etmv4_config *config = &drvdata->config;
665
666	if (kstrtoul(buf, 16, &val))
667		return -EINVAL;
668
669	/* mask off max threshold before checking min value */
670	val &= ETM_CYC_THRESHOLD_MASK;
671	if (val < drvdata->ccitmin)
672		return -EINVAL;
673
674	config->ccctlr = val;
675	return size;
676}
677static DEVICE_ATTR_RW(cyc_threshold);
678
679static ssize_t bb_ctrl_show(struct device *dev,
680			    struct device_attribute *attr,
681			    char *buf)
682{
683	unsigned long val;
684	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
685	struct etmv4_config *config = &drvdata->config;
686
687	val = config->bb_ctrl;
688	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
689}
690
691static ssize_t bb_ctrl_store(struct device *dev,
692			     struct device_attribute *attr,
693			     const char *buf, size_t size)
694{
695	unsigned long val;
696	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
697	struct etmv4_config *config = &drvdata->config;
698
699	if (kstrtoul(buf, 16, &val))
700		return -EINVAL;
701	if (drvdata->trcbb == false)
702		return -EINVAL;
703	if (!drvdata->nr_addr_cmp)
704		return -EINVAL;
705
706	/*
707	 * Bit[8] controls include(1) / exclude(0), bits[0-7] select
708	 * individual range comparators. If include then at least 1
709	 * range must be selected.
710	 */
711	if ((val & TRCBBCTLR_MODE) && (FIELD_GET(TRCBBCTLR_RANGE_MASK, val) == 0))
712		return -EINVAL;
713
714	config->bb_ctrl = val & (TRCBBCTLR_MODE | TRCBBCTLR_RANGE_MASK);
715	return size;
716}
717static DEVICE_ATTR_RW(bb_ctrl);
718
719static ssize_t event_vinst_show(struct device *dev,
720				struct device_attribute *attr,
721				char *buf)
722{
723	unsigned long val;
724	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
725	struct etmv4_config *config = &drvdata->config;
726
727	val = FIELD_GET(TRCVICTLR_EVENT_MASK, config->vinst_ctrl);
728	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
729}
730
731static ssize_t event_vinst_store(struct device *dev,
732				 struct device_attribute *attr,
733				 const char *buf, size_t size)
734{
735	unsigned long val;
736	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
737	struct etmv4_config *config = &drvdata->config;
738
739	if (kstrtoul(buf, 16, &val))
740		return -EINVAL;
741
742	spin_lock(&drvdata->spinlock);
743	val &= TRCVICTLR_EVENT_MASK >> __bf_shf(TRCVICTLR_EVENT_MASK);
744	config->vinst_ctrl &= ~TRCVICTLR_EVENT_MASK;
745	config->vinst_ctrl |= FIELD_PREP(TRCVICTLR_EVENT_MASK, val);
746	spin_unlock(&drvdata->spinlock);
747	return size;
748}
749static DEVICE_ATTR_RW(event_vinst);
750
751static ssize_t s_exlevel_vinst_show(struct device *dev,
752				    struct device_attribute *attr,
753				    char *buf)
754{
755	unsigned long val;
756	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
757	struct etmv4_config *config = &drvdata->config;
758
759	val = FIELD_GET(TRCVICTLR_EXLEVEL_S_MASK, config->vinst_ctrl);
760	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
761}
762
763static ssize_t s_exlevel_vinst_store(struct device *dev,
764				     struct device_attribute *attr,
765				     const char *buf, size_t size)
766{
767	unsigned long val;
768	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
769	struct etmv4_config *config = &drvdata->config;
770
771	if (kstrtoul(buf, 16, &val))
772		return -EINVAL;
773
774	spin_lock(&drvdata->spinlock);
775	/* clear all EXLEVEL_S bits  */
776	config->vinst_ctrl &= ~TRCVICTLR_EXLEVEL_S_MASK;
777	/* enable instruction tracing for corresponding exception level */
778	val &= drvdata->s_ex_level;
779	config->vinst_ctrl |= val << __bf_shf(TRCVICTLR_EXLEVEL_S_MASK);
780	spin_unlock(&drvdata->spinlock);
781	return size;
782}
783static DEVICE_ATTR_RW(s_exlevel_vinst);
784
785static ssize_t ns_exlevel_vinst_show(struct device *dev,
786				     struct device_attribute *attr,
787				     char *buf)
788{
789	unsigned long val;
790	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
791	struct etmv4_config *config = &drvdata->config;
792
793	/* EXLEVEL_NS, bits[23:20] */
794	val = FIELD_GET(TRCVICTLR_EXLEVEL_NS_MASK, config->vinst_ctrl);
795	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
796}
797
798static ssize_t ns_exlevel_vinst_store(struct device *dev,
799				      struct device_attribute *attr,
800				      const char *buf, size_t size)
801{
802	unsigned long val;
803	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
804	struct etmv4_config *config = &drvdata->config;
805
806	if (kstrtoul(buf, 16, &val))
807		return -EINVAL;
808
809	spin_lock(&drvdata->spinlock);
810	/* clear EXLEVEL_NS bits  */
811	config->vinst_ctrl &= ~TRCVICTLR_EXLEVEL_NS_MASK;
812	/* enable instruction tracing for corresponding exception level */
813	val &= drvdata->ns_ex_level;
814	config->vinst_ctrl |= val << __bf_shf(TRCVICTLR_EXLEVEL_NS_MASK);
815	spin_unlock(&drvdata->spinlock);
816	return size;
817}
818static DEVICE_ATTR_RW(ns_exlevel_vinst);
819
820static ssize_t addr_idx_show(struct device *dev,
821			     struct device_attribute *attr,
822			     char *buf)
823{
824	unsigned long val;
825	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
826	struct etmv4_config *config = &drvdata->config;
827
828	val = config->addr_idx;
829	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
830}
831
832static ssize_t addr_idx_store(struct device *dev,
833			      struct device_attribute *attr,
834			      const char *buf, size_t size)
835{
836	unsigned long val;
837	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
838	struct etmv4_config *config = &drvdata->config;
839
840	if (kstrtoul(buf, 16, &val))
841		return -EINVAL;
842	if (val >= drvdata->nr_addr_cmp * 2)
843		return -EINVAL;
844
845	/*
846	 * Use spinlock to ensure index doesn't change while it gets
847	 * dereferenced multiple times within a spinlock block elsewhere.
848	 */
849	spin_lock(&drvdata->spinlock);
850	config->addr_idx = val;
851	spin_unlock(&drvdata->spinlock);
852	return size;
853}
854static DEVICE_ATTR_RW(addr_idx);
855
856static ssize_t addr_instdatatype_show(struct device *dev,
857				      struct device_attribute *attr,
858				      char *buf)
859{
860	ssize_t len;
861	u8 val, idx;
862	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
863	struct etmv4_config *config = &drvdata->config;
864
865	spin_lock(&drvdata->spinlock);
866	idx = config->addr_idx;
867	val = FIELD_GET(TRCACATRn_TYPE_MASK, config->addr_acc[idx]);
868	len = scnprintf(buf, PAGE_SIZE, "%s\n",
869			val == TRCACATRn_TYPE_ADDR ? "instr" :
870			(val == TRCACATRn_TYPE_DATA_LOAD_ADDR ? "data_load" :
871			(val == TRCACATRn_TYPE_DATA_STORE_ADDR ? "data_store" :
872			"data_load_store")));
873	spin_unlock(&drvdata->spinlock);
874	return len;
875}
876
877static ssize_t addr_instdatatype_store(struct device *dev,
878				       struct device_attribute *attr,
879				       const char *buf, size_t size)
880{
881	u8 idx;
882	char str[20] = "";
883	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
884	struct etmv4_config *config = &drvdata->config;
885
886	if (strlen(buf) >= 20)
887		return -EINVAL;
888	if (sscanf(buf, "%s", str) != 1)
889		return -EINVAL;
890
891	spin_lock(&drvdata->spinlock);
892	idx = config->addr_idx;
893	if (!strcmp(str, "instr"))
894		/* TYPE, bits[1:0] */
895		config->addr_acc[idx] &= ~TRCACATRn_TYPE_MASK;
896
897	spin_unlock(&drvdata->spinlock);
898	return size;
899}
900static DEVICE_ATTR_RW(addr_instdatatype);
901
902static ssize_t addr_single_show(struct device *dev,
903				struct device_attribute *attr,
904				char *buf)
905{
906	u8 idx;
907	unsigned long val;
908	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
909	struct etmv4_config *config = &drvdata->config;
910
911	idx = config->addr_idx;
912	spin_lock(&drvdata->spinlock);
913	if (!(config->addr_type[idx] == ETM_ADDR_TYPE_NONE ||
914	      config->addr_type[idx] == ETM_ADDR_TYPE_SINGLE)) {
915		spin_unlock(&drvdata->spinlock);
916		return -EPERM;
917	}
918	val = (unsigned long)config->addr_val[idx];
919	spin_unlock(&drvdata->spinlock);
920	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
921}
922
923static ssize_t addr_single_store(struct device *dev,
924				 struct device_attribute *attr,
925				 const char *buf, size_t size)
926{
927	u8 idx;
928	unsigned long val;
929	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
930	struct etmv4_config *config = &drvdata->config;
931
932	if (kstrtoul(buf, 16, &val))
933		return -EINVAL;
934
935	spin_lock(&drvdata->spinlock);
936	idx = config->addr_idx;
937	if (!(config->addr_type[idx] == ETM_ADDR_TYPE_NONE ||
938	      config->addr_type[idx] == ETM_ADDR_TYPE_SINGLE)) {
939		spin_unlock(&drvdata->spinlock);
940		return -EPERM;
941	}
942
943	config->addr_val[idx] = (u64)val;
944	config->addr_type[idx] = ETM_ADDR_TYPE_SINGLE;
945	spin_unlock(&drvdata->spinlock);
946	return size;
947}
948static DEVICE_ATTR_RW(addr_single);
949
950static ssize_t addr_range_show(struct device *dev,
951			       struct device_attribute *attr,
952			       char *buf)
953{
954	u8 idx;
955	unsigned long val1, val2;
956	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
957	struct etmv4_config *config = &drvdata->config;
958
959	spin_lock(&drvdata->spinlock);
960	idx = config->addr_idx;
961	if (idx % 2 != 0) {
962		spin_unlock(&drvdata->spinlock);
963		return -EPERM;
964	}
965	if (!((config->addr_type[idx] == ETM_ADDR_TYPE_NONE &&
966	       config->addr_type[idx + 1] == ETM_ADDR_TYPE_NONE) ||
967	      (config->addr_type[idx] == ETM_ADDR_TYPE_RANGE &&
968	       config->addr_type[idx + 1] == ETM_ADDR_TYPE_RANGE))) {
969		spin_unlock(&drvdata->spinlock);
970		return -EPERM;
971	}
972
973	val1 = (unsigned long)config->addr_val[idx];
974	val2 = (unsigned long)config->addr_val[idx + 1];
975	spin_unlock(&drvdata->spinlock);
976	return scnprintf(buf, PAGE_SIZE, "%#lx %#lx\n", val1, val2);
977}
978
979static ssize_t addr_range_store(struct device *dev,
980				struct device_attribute *attr,
981				const char *buf, size_t size)
982{
983	u8 idx;
984	unsigned long val1, val2;
985	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
986	struct etmv4_config *config = &drvdata->config;
987	int elements, exclude;
988
989	elements = sscanf(buf, "%lx %lx %x", &val1, &val2, &exclude);
990
991	/*  exclude is optional, but need at least two parameter */
992	if (elements < 2)
993		return -EINVAL;
994	/* lower address comparator cannot have a higher address value */
995	if (val1 > val2)
996		return -EINVAL;
997
998	spin_lock(&drvdata->spinlock);
999	idx = config->addr_idx;
1000	if (idx % 2 != 0) {
1001		spin_unlock(&drvdata->spinlock);
1002		return -EPERM;
1003	}
1004
1005	if (!((config->addr_type[idx] == ETM_ADDR_TYPE_NONE &&
1006	       config->addr_type[idx + 1] == ETM_ADDR_TYPE_NONE) ||
1007	      (config->addr_type[idx] == ETM_ADDR_TYPE_RANGE &&
1008	       config->addr_type[idx + 1] == ETM_ADDR_TYPE_RANGE))) {
1009		spin_unlock(&drvdata->spinlock);
1010		return -EPERM;
1011	}
1012
1013	config->addr_val[idx] = (u64)val1;
1014	config->addr_type[idx] = ETM_ADDR_TYPE_RANGE;
1015	config->addr_val[idx + 1] = (u64)val2;
1016	config->addr_type[idx + 1] = ETM_ADDR_TYPE_RANGE;
1017	/*
1018	 * Program include or exclude control bits for vinst or vdata
1019	 * whenever we change addr comparators to ETM_ADDR_TYPE_RANGE
1020	 * use supplied value, or default to bit set in 'mode'
1021	 */
1022	if (elements != 3)
1023		exclude = config->mode & ETM_MODE_EXCLUDE;
1024	etm4_set_mode_exclude(drvdata, exclude ? true : false);
1025
1026	spin_unlock(&drvdata->spinlock);
1027	return size;
1028}
1029static DEVICE_ATTR_RW(addr_range);
1030
1031static ssize_t addr_start_show(struct device *dev,
1032			       struct device_attribute *attr,
1033			       char *buf)
1034{
1035	u8 idx;
1036	unsigned long val;
1037	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1038	struct etmv4_config *config = &drvdata->config;
1039
1040	spin_lock(&drvdata->spinlock);
1041	idx = config->addr_idx;
1042
1043	if (!(config->addr_type[idx] == ETM_ADDR_TYPE_NONE ||
1044	      config->addr_type[idx] == ETM_ADDR_TYPE_START)) {
1045		spin_unlock(&drvdata->spinlock);
1046		return -EPERM;
1047	}
1048
1049	val = (unsigned long)config->addr_val[idx];
1050	spin_unlock(&drvdata->spinlock);
1051	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
1052}
1053
1054static ssize_t addr_start_store(struct device *dev,
1055				struct device_attribute *attr,
1056				const char *buf, size_t size)
1057{
1058	u8 idx;
1059	unsigned long val;
1060	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1061	struct etmv4_config *config = &drvdata->config;
1062
1063	if (kstrtoul(buf, 16, &val))
1064		return -EINVAL;
1065
1066	spin_lock(&drvdata->spinlock);
1067	idx = config->addr_idx;
1068	if (!drvdata->nr_addr_cmp) {
1069		spin_unlock(&drvdata->spinlock);
1070		return -EINVAL;
1071	}
1072	if (!(config->addr_type[idx] == ETM_ADDR_TYPE_NONE ||
1073	      config->addr_type[idx] == ETM_ADDR_TYPE_START)) {
1074		spin_unlock(&drvdata->spinlock);
1075		return -EPERM;
1076	}
1077
1078	config->addr_val[idx] = (u64)val;
1079	config->addr_type[idx] = ETM_ADDR_TYPE_START;
1080	config->vissctlr |= BIT(idx);
1081	spin_unlock(&drvdata->spinlock);
1082	return size;
1083}
1084static DEVICE_ATTR_RW(addr_start);
1085
1086static ssize_t addr_stop_show(struct device *dev,
1087			      struct device_attribute *attr,
1088			      char *buf)
1089{
1090	u8 idx;
1091	unsigned long val;
1092	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1093	struct etmv4_config *config = &drvdata->config;
1094
1095	spin_lock(&drvdata->spinlock);
1096	idx = config->addr_idx;
1097
1098	if (!(config->addr_type[idx] == ETM_ADDR_TYPE_NONE ||
1099	      config->addr_type[idx] == ETM_ADDR_TYPE_STOP)) {
1100		spin_unlock(&drvdata->spinlock);
1101		return -EPERM;
1102	}
1103
1104	val = (unsigned long)config->addr_val[idx];
1105	spin_unlock(&drvdata->spinlock);
1106	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
1107}
1108
1109static ssize_t addr_stop_store(struct device *dev,
1110			       struct device_attribute *attr,
1111			       const char *buf, size_t size)
1112{
1113	u8 idx;
1114	unsigned long val;
1115	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1116	struct etmv4_config *config = &drvdata->config;
1117
1118	if (kstrtoul(buf, 16, &val))
1119		return -EINVAL;
1120
1121	spin_lock(&drvdata->spinlock);
1122	idx = config->addr_idx;
1123	if (!drvdata->nr_addr_cmp) {
1124		spin_unlock(&drvdata->spinlock);
1125		return -EINVAL;
1126	}
1127	if (!(config->addr_type[idx] == ETM_ADDR_TYPE_NONE ||
1128	       config->addr_type[idx] == ETM_ADDR_TYPE_STOP)) {
1129		spin_unlock(&drvdata->spinlock);
1130		return -EPERM;
1131	}
1132
1133	config->addr_val[idx] = (u64)val;
1134	config->addr_type[idx] = ETM_ADDR_TYPE_STOP;
1135	config->vissctlr |= BIT(idx + 16);
1136	spin_unlock(&drvdata->spinlock);
1137	return size;
1138}
1139static DEVICE_ATTR_RW(addr_stop);
1140
1141static ssize_t addr_ctxtype_show(struct device *dev,
1142				 struct device_attribute *attr,
1143				 char *buf)
1144{
1145	ssize_t len;
1146	u8 idx, val;
1147	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1148	struct etmv4_config *config = &drvdata->config;
1149
1150	spin_lock(&drvdata->spinlock);
1151	idx = config->addr_idx;
1152	/* CONTEXTTYPE, bits[3:2] */
1153	val = FIELD_GET(TRCACATRn_CONTEXTTYPE_MASK, config->addr_acc[idx]);
1154	len = scnprintf(buf, PAGE_SIZE, "%s\n", val == ETM_CTX_NONE ? "none" :
1155			(val == ETM_CTX_CTXID ? "ctxid" :
1156			(val == ETM_CTX_VMID ? "vmid" : "all")));
1157	spin_unlock(&drvdata->spinlock);
1158	return len;
1159}
1160
1161static ssize_t addr_ctxtype_store(struct device *dev,
1162				  struct device_attribute *attr,
1163				  const char *buf, size_t size)
1164{
1165	u8 idx;
1166	char str[10] = "";
1167	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1168	struct etmv4_config *config = &drvdata->config;
1169
1170	if (strlen(buf) >= 10)
1171		return -EINVAL;
1172	if (sscanf(buf, "%s", str) != 1)
1173		return -EINVAL;
1174
1175	spin_lock(&drvdata->spinlock);
1176	idx = config->addr_idx;
1177	if (!strcmp(str, "none"))
1178		/* start by clearing context type bits */
1179		config->addr_acc[idx] &= ~TRCACATRn_CONTEXTTYPE_MASK;
1180	else if (!strcmp(str, "ctxid")) {
1181		/* 0b01 The trace unit performs a Context ID */
1182		if (drvdata->numcidc) {
1183			config->addr_acc[idx] |= TRCACATRn_CONTEXTTYPE_CTXID;
1184			config->addr_acc[idx] &= ~TRCACATRn_CONTEXTTYPE_VMID;
1185		}
1186	} else if (!strcmp(str, "vmid")) {
1187		/* 0b10 The trace unit performs a VMID */
1188		if (drvdata->numvmidc) {
1189			config->addr_acc[idx] &= ~TRCACATRn_CONTEXTTYPE_CTXID;
1190			config->addr_acc[idx] |= TRCACATRn_CONTEXTTYPE_VMID;
1191		}
1192	} else if (!strcmp(str, "all")) {
1193		/*
1194		 * 0b11 The trace unit performs a Context ID
1195		 * comparison and a VMID
1196		 */
1197		if (drvdata->numcidc)
1198			config->addr_acc[idx] |= TRCACATRn_CONTEXTTYPE_CTXID;
1199		if (drvdata->numvmidc)
1200			config->addr_acc[idx] |= TRCACATRn_CONTEXTTYPE_VMID;
1201	}
1202	spin_unlock(&drvdata->spinlock);
1203	return size;
1204}
1205static DEVICE_ATTR_RW(addr_ctxtype);
1206
1207static ssize_t addr_context_show(struct device *dev,
1208				 struct device_attribute *attr,
1209				 char *buf)
1210{
1211	u8 idx;
1212	unsigned long val;
1213	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1214	struct etmv4_config *config = &drvdata->config;
1215
1216	spin_lock(&drvdata->spinlock);
1217	idx = config->addr_idx;
1218	/* context ID comparator bits[6:4] */
1219	val = FIELD_GET(TRCACATRn_CONTEXT_MASK, config->addr_acc[idx]);
1220	spin_unlock(&drvdata->spinlock);
1221	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
1222}
1223
1224static ssize_t addr_context_store(struct device *dev,
1225				  struct device_attribute *attr,
1226				  const char *buf, size_t size)
1227{
1228	u8 idx;
1229	unsigned long val;
1230	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1231	struct etmv4_config *config = &drvdata->config;
1232
1233	if (kstrtoul(buf, 16, &val))
1234		return -EINVAL;
1235	if ((drvdata->numcidc <= 1) && (drvdata->numvmidc <= 1))
1236		return -EINVAL;
1237	if (val >=  (drvdata->numcidc >= drvdata->numvmidc ?
1238		     drvdata->numcidc : drvdata->numvmidc))
1239		return -EINVAL;
1240
1241	spin_lock(&drvdata->spinlock);
1242	idx = config->addr_idx;
1243	/* clear context ID comparator bits[6:4] */
1244	config->addr_acc[idx] &= ~TRCACATRn_CONTEXT_MASK;
1245	config->addr_acc[idx] |= val << __bf_shf(TRCACATRn_CONTEXT_MASK);
1246	spin_unlock(&drvdata->spinlock);
1247	return size;
1248}
1249static DEVICE_ATTR_RW(addr_context);
1250
1251static ssize_t addr_exlevel_s_ns_show(struct device *dev,
1252				      struct device_attribute *attr,
1253				      char *buf)
1254{
1255	u8 idx;
1256	unsigned long val;
1257	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1258	struct etmv4_config *config = &drvdata->config;
1259
1260	spin_lock(&drvdata->spinlock);
1261	idx = config->addr_idx;
1262	val = FIELD_GET(TRCACATRn_EXLEVEL_MASK, config->addr_acc[idx]);
1263	spin_unlock(&drvdata->spinlock);
1264	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
1265}
1266
1267static ssize_t addr_exlevel_s_ns_store(struct device *dev,
1268				       struct device_attribute *attr,
1269				       const char *buf, size_t size)
1270{
1271	u8 idx;
1272	unsigned long val;
1273	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1274	struct etmv4_config *config = &drvdata->config;
1275
1276	if (kstrtoul(buf, 0, &val))
1277		return -EINVAL;
1278
1279	if (val & ~(TRCACATRn_EXLEVEL_MASK >> __bf_shf(TRCACATRn_EXLEVEL_MASK)))
1280		return -EINVAL;
1281
1282	spin_lock(&drvdata->spinlock);
1283	idx = config->addr_idx;
1284	/* clear Exlevel_ns & Exlevel_s bits[14:12, 11:8], bit[15] is res0 */
1285	config->addr_acc[idx] &= ~TRCACATRn_EXLEVEL_MASK;
1286	config->addr_acc[idx] |= val << __bf_shf(TRCACATRn_EXLEVEL_MASK);
1287	spin_unlock(&drvdata->spinlock);
1288	return size;
1289}
1290static DEVICE_ATTR_RW(addr_exlevel_s_ns);
1291
1292static const char * const addr_type_names[] = {
1293	"unused",
1294	"single",
1295	"range",
1296	"start",
1297	"stop"
1298};
1299
1300static ssize_t addr_cmp_view_show(struct device *dev,
1301				  struct device_attribute *attr, char *buf)
1302{
1303	u8 idx, addr_type;
1304	unsigned long addr_v, addr_v2, addr_ctrl;
1305	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1306	struct etmv4_config *config = &drvdata->config;
1307	int size = 0;
1308	bool exclude = false;
1309
1310	spin_lock(&drvdata->spinlock);
1311	idx = config->addr_idx;
1312	addr_v = config->addr_val[idx];
1313	addr_ctrl = config->addr_acc[idx];
1314	addr_type = config->addr_type[idx];
1315	if (addr_type == ETM_ADDR_TYPE_RANGE) {
1316		if (idx & 0x1) {
1317			idx -= 1;
1318			addr_v2 = addr_v;
1319			addr_v = config->addr_val[idx];
1320		} else {
1321			addr_v2 = config->addr_val[idx + 1];
1322		}
1323		exclude = config->viiectlr & BIT(idx / 2 + 16);
1324	}
1325	spin_unlock(&drvdata->spinlock);
1326	if (addr_type) {
1327		size = scnprintf(buf, PAGE_SIZE, "addr_cmp[%i] %s %#lx", idx,
1328				 addr_type_names[addr_type], addr_v);
1329		if (addr_type == ETM_ADDR_TYPE_RANGE) {
1330			size += scnprintf(buf + size, PAGE_SIZE - size,
1331					  " %#lx %s", addr_v2,
1332					  exclude ? "exclude" : "include");
1333		}
1334		size += scnprintf(buf + size, PAGE_SIZE - size,
1335				  " ctrl(%#lx)\n", addr_ctrl);
1336	} else {
1337		size = scnprintf(buf, PAGE_SIZE, "addr_cmp[%i] unused\n", idx);
1338	}
1339	return size;
1340}
1341static DEVICE_ATTR_RO(addr_cmp_view);
1342
1343static ssize_t vinst_pe_cmp_start_stop_show(struct device *dev,
1344					    struct device_attribute *attr,
1345					    char *buf)
1346{
1347	unsigned long val;
1348	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1349	struct etmv4_config *config = &drvdata->config;
1350
1351	if (!drvdata->nr_pe_cmp)
1352		return -EINVAL;
1353	val = config->vipcssctlr;
1354	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
1355}
1356static ssize_t vinst_pe_cmp_start_stop_store(struct device *dev,
1357					     struct device_attribute *attr,
1358					     const char *buf, size_t size)
1359{
1360	unsigned long val;
1361	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1362	struct etmv4_config *config = &drvdata->config;
1363
1364	if (kstrtoul(buf, 16, &val))
1365		return -EINVAL;
1366	if (!drvdata->nr_pe_cmp)
1367		return -EINVAL;
1368
1369	spin_lock(&drvdata->spinlock);
1370	config->vipcssctlr = val;
1371	spin_unlock(&drvdata->spinlock);
1372	return size;
1373}
1374static DEVICE_ATTR_RW(vinst_pe_cmp_start_stop);
1375
1376static ssize_t seq_idx_show(struct device *dev,
1377			    struct device_attribute *attr,
1378			    char *buf)
1379{
1380	unsigned long val;
1381	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1382	struct etmv4_config *config = &drvdata->config;
1383
1384	val = config->seq_idx;
1385	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
1386}
1387
1388static ssize_t seq_idx_store(struct device *dev,
1389			     struct device_attribute *attr,
1390			     const char *buf, size_t size)
1391{
1392	unsigned long val;
1393	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1394	struct etmv4_config *config = &drvdata->config;
1395
1396	if (kstrtoul(buf, 16, &val))
1397		return -EINVAL;
1398	if (val >= drvdata->nrseqstate - 1)
1399		return -EINVAL;
1400
1401	/*
1402	 * Use spinlock to ensure index doesn't change while it gets
1403	 * dereferenced multiple times within a spinlock block elsewhere.
1404	 */
1405	spin_lock(&drvdata->spinlock);
1406	config->seq_idx = val;
1407	spin_unlock(&drvdata->spinlock);
1408	return size;
1409}
1410static DEVICE_ATTR_RW(seq_idx);
1411
1412static ssize_t seq_state_show(struct device *dev,
1413			      struct device_attribute *attr,
1414			      char *buf)
1415{
1416	unsigned long val;
1417	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1418	struct etmv4_config *config = &drvdata->config;
1419
1420	val = config->seq_state;
1421	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
1422}
1423
1424static ssize_t seq_state_store(struct device *dev,
1425			       struct device_attribute *attr,
1426			       const char *buf, size_t size)
1427{
1428	unsigned long val;
1429	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1430	struct etmv4_config *config = &drvdata->config;
1431
1432	if (kstrtoul(buf, 16, &val))
1433		return -EINVAL;
1434	if (val >= drvdata->nrseqstate)
1435		return -EINVAL;
1436
1437	config->seq_state = val;
1438	return size;
1439}
1440static DEVICE_ATTR_RW(seq_state);
1441
1442static ssize_t seq_event_show(struct device *dev,
1443			      struct device_attribute *attr,
1444			      char *buf)
1445{
1446	u8 idx;
1447	unsigned long val;
1448	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1449	struct etmv4_config *config = &drvdata->config;
1450
1451	spin_lock(&drvdata->spinlock);
1452	idx = config->seq_idx;
1453	val = config->seq_ctrl[idx];
1454	spin_unlock(&drvdata->spinlock);
1455	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
1456}
1457
1458static ssize_t seq_event_store(struct device *dev,
1459			       struct device_attribute *attr,
1460			       const char *buf, size_t size)
1461{
1462	u8 idx;
1463	unsigned long val;
1464	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1465	struct etmv4_config *config = &drvdata->config;
1466
1467	if (kstrtoul(buf, 16, &val))
1468		return -EINVAL;
1469
1470	spin_lock(&drvdata->spinlock);
1471	idx = config->seq_idx;
1472	/* Seq control has two masks B[15:8] F[7:0] */
1473	config->seq_ctrl[idx] = val & 0xFFFF;
1474	spin_unlock(&drvdata->spinlock);
1475	return size;
1476}
1477static DEVICE_ATTR_RW(seq_event);
1478
1479static ssize_t seq_reset_event_show(struct device *dev,
1480				    struct device_attribute *attr,
1481				    char *buf)
1482{
1483	unsigned long val;
1484	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1485	struct etmv4_config *config = &drvdata->config;
1486
1487	val = config->seq_rst;
1488	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
1489}
1490
1491static ssize_t seq_reset_event_store(struct device *dev,
1492				     struct device_attribute *attr,
1493				     const char *buf, size_t size)
1494{
1495	unsigned long val;
1496	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1497	struct etmv4_config *config = &drvdata->config;
1498
1499	if (kstrtoul(buf, 16, &val))
1500		return -EINVAL;
1501	if (!(drvdata->nrseqstate))
1502		return -EINVAL;
1503
1504	config->seq_rst = val & ETMv4_EVENT_MASK;
1505	return size;
1506}
1507static DEVICE_ATTR_RW(seq_reset_event);
1508
1509static ssize_t cntr_idx_show(struct device *dev,
1510			     struct device_attribute *attr,
1511			     char *buf)
1512{
1513	unsigned long val;
1514	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1515	struct etmv4_config *config = &drvdata->config;
1516
1517	val = config->cntr_idx;
1518	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
1519}
1520
1521static ssize_t cntr_idx_store(struct device *dev,
1522			      struct device_attribute *attr,
1523			      const char *buf, size_t size)
1524{
1525	unsigned long val;
1526	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1527	struct etmv4_config *config = &drvdata->config;
1528
1529	if (kstrtoul(buf, 16, &val))
1530		return -EINVAL;
1531	if (val >= drvdata->nr_cntr)
1532		return -EINVAL;
1533
1534	/*
1535	 * Use spinlock to ensure index doesn't change while it gets
1536	 * dereferenced multiple times within a spinlock block elsewhere.
1537	 */
1538	spin_lock(&drvdata->spinlock);
1539	config->cntr_idx = val;
1540	spin_unlock(&drvdata->spinlock);
1541	return size;
1542}
1543static DEVICE_ATTR_RW(cntr_idx);
1544
1545static ssize_t cntrldvr_show(struct device *dev,
1546			     struct device_attribute *attr,
1547			     char *buf)
1548{
1549	u8 idx;
1550	unsigned long val;
1551	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1552	struct etmv4_config *config = &drvdata->config;
1553
1554	spin_lock(&drvdata->spinlock);
1555	idx = config->cntr_idx;
1556	val = config->cntrldvr[idx];
1557	spin_unlock(&drvdata->spinlock);
1558	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
1559}
1560
1561static ssize_t cntrldvr_store(struct device *dev,
1562			      struct device_attribute *attr,
1563			      const char *buf, size_t size)
1564{
1565	u8 idx;
1566	unsigned long val;
1567	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1568	struct etmv4_config *config = &drvdata->config;
1569
1570	if (kstrtoul(buf, 16, &val))
1571		return -EINVAL;
1572	if (val > ETM_CNTR_MAX_VAL)
1573		return -EINVAL;
1574
1575	spin_lock(&drvdata->spinlock);
1576	idx = config->cntr_idx;
1577	config->cntrldvr[idx] = val;
1578	spin_unlock(&drvdata->spinlock);
1579	return size;
1580}
1581static DEVICE_ATTR_RW(cntrldvr);
1582
1583static ssize_t cntr_val_show(struct device *dev,
1584			     struct device_attribute *attr,
1585			     char *buf)
1586{
1587	u8 idx;
1588	unsigned long val;
1589	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1590	struct etmv4_config *config = &drvdata->config;
1591
1592	spin_lock(&drvdata->spinlock);
1593	idx = config->cntr_idx;
1594	val = config->cntr_val[idx];
1595	spin_unlock(&drvdata->spinlock);
1596	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
1597}
1598
1599static ssize_t cntr_val_store(struct device *dev,
1600			      struct device_attribute *attr,
1601			      const char *buf, size_t size)
1602{
1603	u8 idx;
1604	unsigned long val;
1605	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1606	struct etmv4_config *config = &drvdata->config;
1607
1608	if (kstrtoul(buf, 16, &val))
1609		return -EINVAL;
1610	if (val > ETM_CNTR_MAX_VAL)
1611		return -EINVAL;
1612
1613	spin_lock(&drvdata->spinlock);
1614	idx = config->cntr_idx;
1615	config->cntr_val[idx] = val;
1616	spin_unlock(&drvdata->spinlock);
1617	return size;
1618}
1619static DEVICE_ATTR_RW(cntr_val);
1620
1621static ssize_t cntr_ctrl_show(struct device *dev,
1622			      struct device_attribute *attr,
1623			      char *buf)
1624{
1625	u8 idx;
1626	unsigned long val;
1627	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1628	struct etmv4_config *config = &drvdata->config;
1629
1630	spin_lock(&drvdata->spinlock);
1631	idx = config->cntr_idx;
1632	val = config->cntr_ctrl[idx];
1633	spin_unlock(&drvdata->spinlock);
1634	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
1635}
1636
1637static ssize_t cntr_ctrl_store(struct device *dev,
1638			       struct device_attribute *attr,
1639			       const char *buf, size_t size)
1640{
1641	u8 idx;
1642	unsigned long val;
1643	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1644	struct etmv4_config *config = &drvdata->config;
1645
1646	if (kstrtoul(buf, 16, &val))
1647		return -EINVAL;
1648
1649	spin_lock(&drvdata->spinlock);
1650	idx = config->cntr_idx;
1651	config->cntr_ctrl[idx] = val;
1652	spin_unlock(&drvdata->spinlock);
1653	return size;
1654}
1655static DEVICE_ATTR_RW(cntr_ctrl);
1656
1657static ssize_t res_idx_show(struct device *dev,
1658			    struct device_attribute *attr,
1659			    char *buf)
1660{
1661	unsigned long val;
1662	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1663	struct etmv4_config *config = &drvdata->config;
1664
1665	val = config->res_idx;
1666	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
1667}
1668
1669static ssize_t res_idx_store(struct device *dev,
1670			     struct device_attribute *attr,
1671			     const char *buf, size_t size)
1672{
1673	unsigned long val;
1674	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1675	struct etmv4_config *config = &drvdata->config;
1676
1677	if (kstrtoul(buf, 16, &val))
1678		return -EINVAL;
1679	/*
1680	 * Resource selector pair 0 is always implemented and reserved,
1681	 * namely an idx with 0 and 1 is illegal.
1682	 */
1683	if ((val < 2) || (val >= 2 * drvdata->nr_resource))
1684		return -EINVAL;
1685
1686	/*
1687	 * Use spinlock to ensure index doesn't change while it gets
1688	 * dereferenced multiple times within a spinlock block elsewhere.
1689	 */
1690	spin_lock(&drvdata->spinlock);
1691	config->res_idx = val;
1692	spin_unlock(&drvdata->spinlock);
1693	return size;
1694}
1695static DEVICE_ATTR_RW(res_idx);
1696
1697static ssize_t res_ctrl_show(struct device *dev,
1698			     struct device_attribute *attr,
1699			     char *buf)
1700{
1701	u8 idx;
1702	unsigned long val;
1703	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1704	struct etmv4_config *config = &drvdata->config;
1705
1706	spin_lock(&drvdata->spinlock);
1707	idx = config->res_idx;
1708	val = config->res_ctrl[idx];
1709	spin_unlock(&drvdata->spinlock);
1710	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
1711}
1712
1713static ssize_t res_ctrl_store(struct device *dev,
1714			      struct device_attribute *attr,
1715			      const char *buf, size_t size)
1716{
1717	u8 idx;
1718	unsigned long val;
1719	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1720	struct etmv4_config *config = &drvdata->config;
1721
1722	if (kstrtoul(buf, 16, &val))
1723		return -EINVAL;
1724
1725	spin_lock(&drvdata->spinlock);
1726	idx = config->res_idx;
1727	/* For odd idx pair inversal bit is RES0 */
1728	if (idx % 2 != 0)
1729		/* PAIRINV, bit[21] */
1730		val &= ~TRCRSCTLRn_PAIRINV;
1731	config->res_ctrl[idx] = val & (TRCRSCTLRn_PAIRINV |
1732				       TRCRSCTLRn_INV |
1733				       TRCRSCTLRn_GROUP_MASK |
1734				       TRCRSCTLRn_SELECT_MASK);
1735	spin_unlock(&drvdata->spinlock);
1736	return size;
1737}
1738static DEVICE_ATTR_RW(res_ctrl);
1739
1740static ssize_t sshot_idx_show(struct device *dev,
1741			      struct device_attribute *attr, char *buf)
1742{
1743	unsigned long val;
1744	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1745	struct etmv4_config *config = &drvdata->config;
1746
1747	val = config->ss_idx;
1748	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
1749}
1750
1751static ssize_t sshot_idx_store(struct device *dev,
1752			       struct device_attribute *attr,
1753			       const char *buf, size_t size)
1754{
1755	unsigned long val;
1756	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1757	struct etmv4_config *config = &drvdata->config;
1758
1759	if (kstrtoul(buf, 16, &val))
1760		return -EINVAL;
1761	if (val >= drvdata->nr_ss_cmp)
1762		return -EINVAL;
1763
1764	spin_lock(&drvdata->spinlock);
1765	config->ss_idx = val;
1766	spin_unlock(&drvdata->spinlock);
1767	return size;
1768}
1769static DEVICE_ATTR_RW(sshot_idx);
1770
1771static ssize_t sshot_ctrl_show(struct device *dev,
1772			       struct device_attribute *attr,
1773			       char *buf)
1774{
1775	unsigned long val;
1776	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1777	struct etmv4_config *config = &drvdata->config;
1778
1779	spin_lock(&drvdata->spinlock);
1780	val = config->ss_ctrl[config->ss_idx];
1781	spin_unlock(&drvdata->spinlock);
1782	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
1783}
1784
1785static ssize_t sshot_ctrl_store(struct device *dev,
1786				struct device_attribute *attr,
1787				const char *buf, size_t size)
1788{
1789	u8 idx;
1790	unsigned long val;
1791	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1792	struct etmv4_config *config = &drvdata->config;
1793
1794	if (kstrtoul(buf, 16, &val))
1795		return -EINVAL;
1796
1797	spin_lock(&drvdata->spinlock);
1798	idx = config->ss_idx;
1799	config->ss_ctrl[idx] = FIELD_PREP(TRCSSCCRn_SAC_ARC_RST_MASK, val);
1800	/* must clear bit 31 in related status register on programming */
1801	config->ss_status[idx] &= ~TRCSSCSRn_STATUS;
1802	spin_unlock(&drvdata->spinlock);
1803	return size;
1804}
1805static DEVICE_ATTR_RW(sshot_ctrl);
1806
1807static ssize_t sshot_status_show(struct device *dev,
1808				 struct device_attribute *attr, char *buf)
1809{
1810	unsigned long val;
1811	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1812	struct etmv4_config *config = &drvdata->config;
1813
1814	spin_lock(&drvdata->spinlock);
1815	val = config->ss_status[config->ss_idx];
1816	spin_unlock(&drvdata->spinlock);
1817	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
1818}
1819static DEVICE_ATTR_RO(sshot_status);
1820
1821static ssize_t sshot_pe_ctrl_show(struct device *dev,
1822				  struct device_attribute *attr,
1823				  char *buf)
1824{
1825	unsigned long val;
1826	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1827	struct etmv4_config *config = &drvdata->config;
1828
1829	spin_lock(&drvdata->spinlock);
1830	val = config->ss_pe_cmp[config->ss_idx];
1831	spin_unlock(&drvdata->spinlock);
1832	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
1833}
1834
1835static ssize_t sshot_pe_ctrl_store(struct device *dev,
1836				   struct device_attribute *attr,
1837				   const char *buf, size_t size)
1838{
1839	u8 idx;
1840	unsigned long val;
1841	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1842	struct etmv4_config *config = &drvdata->config;
1843
1844	if (kstrtoul(buf, 16, &val))
1845		return -EINVAL;
1846
1847	spin_lock(&drvdata->spinlock);
1848	idx = config->ss_idx;
1849	config->ss_pe_cmp[idx] = FIELD_PREP(TRCSSPCICRn_PC_MASK, val);
1850	/* must clear bit 31 in related status register on programming */
1851	config->ss_status[idx] &= ~TRCSSCSRn_STATUS;
1852	spin_unlock(&drvdata->spinlock);
1853	return size;
1854}
1855static DEVICE_ATTR_RW(sshot_pe_ctrl);
1856
1857static ssize_t ctxid_idx_show(struct device *dev,
1858			      struct device_attribute *attr,
1859			      char *buf)
1860{
1861	unsigned long val;
1862	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1863	struct etmv4_config *config = &drvdata->config;
1864
1865	val = config->ctxid_idx;
1866	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
1867}
1868
1869static ssize_t ctxid_idx_store(struct device *dev,
1870			       struct device_attribute *attr,
1871			       const char *buf, size_t size)
1872{
1873	unsigned long val;
1874	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1875	struct etmv4_config *config = &drvdata->config;
1876
1877	if (kstrtoul(buf, 16, &val))
1878		return -EINVAL;
1879	if (val >= drvdata->numcidc)
1880		return -EINVAL;
1881
1882	/*
1883	 * Use spinlock to ensure index doesn't change while it gets
1884	 * dereferenced multiple times within a spinlock block elsewhere.
1885	 */
1886	spin_lock(&drvdata->spinlock);
1887	config->ctxid_idx = val;
1888	spin_unlock(&drvdata->spinlock);
1889	return size;
1890}
1891static DEVICE_ATTR_RW(ctxid_idx);
1892
1893static ssize_t ctxid_pid_show(struct device *dev,
1894			      struct device_attribute *attr,
1895			      char *buf)
1896{
1897	u8 idx;
1898	unsigned long val;
1899	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1900	struct etmv4_config *config = &drvdata->config;
1901
1902	/*
1903	 * Don't use contextID tracing if coming from a PID namespace.  See
1904	 * comment in ctxid_pid_store().
1905	 */
1906	if (task_active_pid_ns(current) != &init_pid_ns)
1907		return -EINVAL;
1908
1909	spin_lock(&drvdata->spinlock);
1910	idx = config->ctxid_idx;
1911	val = (unsigned long)config->ctxid_pid[idx];
1912	spin_unlock(&drvdata->spinlock);
1913	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
1914}
1915
1916static ssize_t ctxid_pid_store(struct device *dev,
1917			       struct device_attribute *attr,
1918			       const char *buf, size_t size)
1919{
1920	u8 idx;
1921	unsigned long pid;
1922	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1923	struct etmv4_config *config = &drvdata->config;
1924
1925	/*
1926	 * When contextID tracing is enabled the tracers will insert the
1927	 * value found in the contextID register in the trace stream.  But if
1928	 * a process is in a namespace the PID of that process as seen from the
1929	 * namespace won't be what the kernel sees, something that makes the
1930	 * feature confusing and can potentially leak kernel only information.
1931	 * As such refuse to use the feature if @current is not in the initial
1932	 * PID namespace.
1933	 */
1934	if (task_active_pid_ns(current) != &init_pid_ns)
1935		return -EINVAL;
1936
1937	/*
1938	 * only implemented when ctxid tracing is enabled, i.e. at least one
1939	 * ctxid comparator is implemented and ctxid is greater than 0 bits
1940	 * in length
1941	 */
1942	if (!drvdata->ctxid_size || !drvdata->numcidc)
1943		return -EINVAL;
1944	if (kstrtoul(buf, 16, &pid))
1945		return -EINVAL;
1946
1947	spin_lock(&drvdata->spinlock);
1948	idx = config->ctxid_idx;
1949	config->ctxid_pid[idx] = (u64)pid;
1950	spin_unlock(&drvdata->spinlock);
1951	return size;
1952}
1953static DEVICE_ATTR_RW(ctxid_pid);
1954
1955static ssize_t ctxid_masks_show(struct device *dev,
1956				struct device_attribute *attr,
1957				char *buf)
1958{
1959	unsigned long val1, val2;
1960	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1961	struct etmv4_config *config = &drvdata->config;
1962
1963	/*
1964	 * Don't use contextID tracing if coming from a PID namespace.  See
1965	 * comment in ctxid_pid_store().
1966	 */
1967	if (task_active_pid_ns(current) != &init_pid_ns)
1968		return -EINVAL;
1969
1970	spin_lock(&drvdata->spinlock);
1971	val1 = config->ctxid_mask0;
1972	val2 = config->ctxid_mask1;
1973	spin_unlock(&drvdata->spinlock);
1974	return scnprintf(buf, PAGE_SIZE, "%#lx %#lx\n", val1, val2);
1975}
1976
1977static ssize_t ctxid_masks_store(struct device *dev,
1978				struct device_attribute *attr,
1979				const char *buf, size_t size)
1980{
1981	u8 i, j, maskbyte;
1982	unsigned long val1, val2, mask;
1983	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
1984	struct etmv4_config *config = &drvdata->config;
1985	int nr_inputs;
1986
1987	/*
1988	 * Don't use contextID tracing if coming from a PID namespace.  See
1989	 * comment in ctxid_pid_store().
1990	 */
1991	if (task_active_pid_ns(current) != &init_pid_ns)
1992		return -EINVAL;
1993
1994	/*
1995	 * only implemented when ctxid tracing is enabled, i.e. at least one
1996	 * ctxid comparator is implemented and ctxid is greater than 0 bits
1997	 * in length
1998	 */
1999	if (!drvdata->ctxid_size || !drvdata->numcidc)
2000		return -EINVAL;
2001	/* one mask if <= 4 comparators, two for up to 8 */
2002	nr_inputs = sscanf(buf, "%lx %lx", &val1, &val2);
2003	if ((drvdata->numcidc > 4) && (nr_inputs != 2))
2004		return -EINVAL;
2005
2006	spin_lock(&drvdata->spinlock);
2007	/*
2008	 * each byte[0..3] controls mask value applied to ctxid
2009	 * comparator[0..3]
2010	 */
2011	switch (drvdata->numcidc) {
2012	case 0x1:
2013		/* COMP0, bits[7:0] */
2014		config->ctxid_mask0 = val1 & 0xFF;
2015		break;
2016	case 0x2:
2017		/* COMP1, bits[15:8] */
2018		config->ctxid_mask0 = val1 & 0xFFFF;
2019		break;
2020	case 0x3:
2021		/* COMP2, bits[23:16] */
2022		config->ctxid_mask0 = val1 & 0xFFFFFF;
2023		break;
2024	case 0x4:
2025		 /* COMP3, bits[31:24] */
2026		config->ctxid_mask0 = val1;
2027		break;
2028	case 0x5:
2029		/* COMP4, bits[7:0] */
2030		config->ctxid_mask0 = val1;
2031		config->ctxid_mask1 = val2 & 0xFF;
2032		break;
2033	case 0x6:
2034		/* COMP5, bits[15:8] */
2035		config->ctxid_mask0 = val1;
2036		config->ctxid_mask1 = val2 & 0xFFFF;
2037		break;
2038	case 0x7:
2039		/* COMP6, bits[23:16] */
2040		config->ctxid_mask0 = val1;
2041		config->ctxid_mask1 = val2 & 0xFFFFFF;
2042		break;
2043	case 0x8:
2044		/* COMP7, bits[31:24] */
2045		config->ctxid_mask0 = val1;
2046		config->ctxid_mask1 = val2;
2047		break;
2048	default:
2049		break;
2050	}
2051	/*
2052	 * If software sets a mask bit to 1, it must program relevant byte
2053	 * of ctxid comparator value 0x0, otherwise behavior is unpredictable.
2054	 * For example, if bit[3] of ctxid_mask0 is 1, we must clear bits[31:24]
2055	 * of ctxid comparator0 value (corresponding to byte 0) register.
2056	 */
2057	mask = config->ctxid_mask0;
2058	for (i = 0; i < drvdata->numcidc; i++) {
2059		/* mask value of corresponding ctxid comparator */
2060		maskbyte = mask & ETMv4_EVENT_MASK;
2061		/*
2062		 * each bit corresponds to a byte of respective ctxid comparator
2063		 * value register
2064		 */
2065		for (j = 0; j < 8; j++) {
2066			if (maskbyte & 1)
2067				config->ctxid_pid[i] &= ~(0xFFUL << (j * 8));
2068			maskbyte >>= 1;
2069		}
2070		/* Select the next ctxid comparator mask value */
2071		if (i == 3)
2072			/* ctxid comparators[4-7] */
2073			mask = config->ctxid_mask1;
2074		else
2075			mask >>= 0x8;
2076	}
2077
2078	spin_unlock(&drvdata->spinlock);
2079	return size;
2080}
2081static DEVICE_ATTR_RW(ctxid_masks);
2082
2083static ssize_t vmid_idx_show(struct device *dev,
2084			     struct device_attribute *attr,
2085			     char *buf)
2086{
2087	unsigned long val;
2088	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
2089	struct etmv4_config *config = &drvdata->config;
2090
2091	val = config->vmid_idx;
2092	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
2093}
2094
2095static ssize_t vmid_idx_store(struct device *dev,
2096			      struct device_attribute *attr,
2097			      const char *buf, size_t size)
2098{
2099	unsigned long val;
2100	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
2101	struct etmv4_config *config = &drvdata->config;
2102
2103	if (kstrtoul(buf, 16, &val))
2104		return -EINVAL;
2105	if (val >= drvdata->numvmidc)
2106		return -EINVAL;
2107
2108	/*
2109	 * Use spinlock to ensure index doesn't change while it gets
2110	 * dereferenced multiple times within a spinlock block elsewhere.
2111	 */
2112	spin_lock(&drvdata->spinlock);
2113	config->vmid_idx = val;
2114	spin_unlock(&drvdata->spinlock);
2115	return size;
2116}
2117static DEVICE_ATTR_RW(vmid_idx);
2118
2119static ssize_t vmid_val_show(struct device *dev,
2120			     struct device_attribute *attr,
2121			     char *buf)
2122{
2123	unsigned long val;
2124	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
2125	struct etmv4_config *config = &drvdata->config;
2126
2127	/*
2128	 * Don't use virtual contextID tracing if coming from a PID namespace.
2129	 * See comment in ctxid_pid_store().
2130	 */
2131	if (!task_is_in_init_pid_ns(current))
2132		return -EINVAL;
2133
2134	spin_lock(&drvdata->spinlock);
2135	val = (unsigned long)config->vmid_val[config->vmid_idx];
2136	spin_unlock(&drvdata->spinlock);
2137	return scnprintf(buf, PAGE_SIZE, "%#lx\n", val);
2138}
2139
2140static ssize_t vmid_val_store(struct device *dev,
2141			      struct device_attribute *attr,
2142			      const char *buf, size_t size)
2143{
2144	unsigned long val;
2145	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
2146	struct etmv4_config *config = &drvdata->config;
2147
2148	/*
2149	 * Don't use virtual contextID tracing if coming from a PID namespace.
2150	 * See comment in ctxid_pid_store().
2151	 */
2152	if (!task_is_in_init_pid_ns(current))
2153		return -EINVAL;
2154
2155	/*
2156	 * only implemented when vmid tracing is enabled, i.e. at least one
2157	 * vmid comparator is implemented and at least 8 bit vmid size
2158	 */
2159	if (!drvdata->vmid_size || !drvdata->numvmidc)
2160		return -EINVAL;
2161	if (kstrtoul(buf, 16, &val))
2162		return -EINVAL;
2163
2164	spin_lock(&drvdata->spinlock);
2165	config->vmid_val[config->vmid_idx] = (u64)val;
2166	spin_unlock(&drvdata->spinlock);
2167	return size;
2168}
2169static DEVICE_ATTR_RW(vmid_val);
2170
2171static ssize_t vmid_masks_show(struct device *dev,
2172			       struct device_attribute *attr, char *buf)
2173{
2174	unsigned long val1, val2;
2175	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
2176	struct etmv4_config *config = &drvdata->config;
2177
2178	/*
2179	 * Don't use virtual contextID tracing if coming from a PID namespace.
2180	 * See comment in ctxid_pid_store().
2181	 */
2182	if (!task_is_in_init_pid_ns(current))
2183		return -EINVAL;
2184
2185	spin_lock(&drvdata->spinlock);
2186	val1 = config->vmid_mask0;
2187	val2 = config->vmid_mask1;
2188	spin_unlock(&drvdata->spinlock);
2189	return scnprintf(buf, PAGE_SIZE, "%#lx %#lx\n", val1, val2);
2190}
2191
2192static ssize_t vmid_masks_store(struct device *dev,
2193				struct device_attribute *attr,
2194				const char *buf, size_t size)
2195{
2196	u8 i, j, maskbyte;
2197	unsigned long val1, val2, mask;
2198	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
2199	struct etmv4_config *config = &drvdata->config;
2200	int nr_inputs;
2201
2202	/*
2203	 * Don't use virtual contextID tracing if coming from a PID namespace.
2204	 * See comment in ctxid_pid_store().
2205	 */
2206	if (!task_is_in_init_pid_ns(current))
2207		return -EINVAL;
2208
2209	/*
2210	 * only implemented when vmid tracing is enabled, i.e. at least one
2211	 * vmid comparator is implemented and at least 8 bit vmid size
2212	 */
2213	if (!drvdata->vmid_size || !drvdata->numvmidc)
2214		return -EINVAL;
2215	/* one mask if <= 4 comparators, two for up to 8 */
2216	nr_inputs = sscanf(buf, "%lx %lx", &val1, &val2);
2217	if ((drvdata->numvmidc > 4) && (nr_inputs != 2))
2218		return -EINVAL;
2219
2220	spin_lock(&drvdata->spinlock);
2221
2222	/*
2223	 * each byte[0..3] controls mask value applied to vmid
2224	 * comparator[0..3]
2225	 */
2226	switch (drvdata->numvmidc) {
2227	case 0x1:
2228		/* COMP0, bits[7:0] */
2229		config->vmid_mask0 = val1 & 0xFF;
2230		break;
2231	case 0x2:
2232		/* COMP1, bits[15:8] */
2233		config->vmid_mask0 = val1 & 0xFFFF;
2234		break;
2235	case 0x3:
2236		/* COMP2, bits[23:16] */
2237		config->vmid_mask0 = val1 & 0xFFFFFF;
2238		break;
2239	case 0x4:
2240		/* COMP3, bits[31:24] */
2241		config->vmid_mask0 = val1;
2242		break;
2243	case 0x5:
2244		/* COMP4, bits[7:0] */
2245		config->vmid_mask0 = val1;
2246		config->vmid_mask1 = val2 & 0xFF;
2247		break;
2248	case 0x6:
2249		/* COMP5, bits[15:8] */
2250		config->vmid_mask0 = val1;
2251		config->vmid_mask1 = val2 & 0xFFFF;
2252		break;
2253	case 0x7:
2254		/* COMP6, bits[23:16] */
2255		config->vmid_mask0 = val1;
2256		config->vmid_mask1 = val2 & 0xFFFFFF;
2257		break;
2258	case 0x8:
2259		/* COMP7, bits[31:24] */
2260		config->vmid_mask0 = val1;
2261		config->vmid_mask1 = val2;
2262		break;
2263	default:
2264		break;
2265	}
2266
2267	/*
2268	 * If software sets a mask bit to 1, it must program relevant byte
2269	 * of vmid comparator value 0x0, otherwise behavior is unpredictable.
2270	 * For example, if bit[3] of vmid_mask0 is 1, we must clear bits[31:24]
2271	 * of vmid comparator0 value (corresponding to byte 0) register.
2272	 */
2273	mask = config->vmid_mask0;
2274	for (i = 0; i < drvdata->numvmidc; i++) {
2275		/* mask value of corresponding vmid comparator */
2276		maskbyte = mask & ETMv4_EVENT_MASK;
2277		/*
2278		 * each bit corresponds to a byte of respective vmid comparator
2279		 * value register
2280		 */
2281		for (j = 0; j < 8; j++) {
2282			if (maskbyte & 1)
2283				config->vmid_val[i] &= ~(0xFFUL << (j * 8));
2284			maskbyte >>= 1;
2285		}
2286		/* Select the next vmid comparator mask value */
2287		if (i == 3)
2288			/* vmid comparators[4-7] */
2289			mask = config->vmid_mask1;
2290		else
2291			mask >>= 0x8;
2292	}
2293	spin_unlock(&drvdata->spinlock);
2294	return size;
2295}
2296static DEVICE_ATTR_RW(vmid_masks);
2297
2298static ssize_t cpu_show(struct device *dev,
2299			struct device_attribute *attr, char *buf)
2300{
2301	int val;
2302	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
2303
2304	val = drvdata->cpu;
2305	return scnprintf(buf, PAGE_SIZE, "%d\n", val);
2306
2307}
2308static DEVICE_ATTR_RO(cpu);
2309
2310static ssize_t ts_source_show(struct device *dev,
2311			      struct device_attribute *attr,
2312			      char *buf)
2313{
2314	int val;
2315	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
2316
2317	if (!drvdata->trfcr) {
2318		val = -1;
2319		goto out;
2320	}
2321
2322	switch (drvdata->trfcr & TRFCR_ELx_TS_MASK) {
2323	case TRFCR_ELx_TS_VIRTUAL:
2324	case TRFCR_ELx_TS_GUEST_PHYSICAL:
2325	case TRFCR_ELx_TS_PHYSICAL:
2326		val = FIELD_GET(TRFCR_ELx_TS_MASK, drvdata->trfcr);
2327		break;
2328	default:
2329		val = -1;
2330		break;
2331	}
2332
2333out:
2334	return sysfs_emit(buf, "%d\n", val);
2335}
2336static DEVICE_ATTR_RO(ts_source);
2337
2338static struct attribute *coresight_etmv4_attrs[] = {
2339	&dev_attr_nr_pe_cmp.attr,
2340	&dev_attr_nr_addr_cmp.attr,
2341	&dev_attr_nr_cntr.attr,
2342	&dev_attr_nr_ext_inp.attr,
2343	&dev_attr_numcidc.attr,
2344	&dev_attr_numvmidc.attr,
2345	&dev_attr_nrseqstate.attr,
2346	&dev_attr_nr_resource.attr,
2347	&dev_attr_nr_ss_cmp.attr,
2348	&dev_attr_reset.attr,
2349	&dev_attr_mode.attr,
2350	&dev_attr_pe.attr,
2351	&dev_attr_event.attr,
2352	&dev_attr_event_instren.attr,
2353	&dev_attr_event_ts.attr,
2354	&dev_attr_syncfreq.attr,
2355	&dev_attr_cyc_threshold.attr,
2356	&dev_attr_bb_ctrl.attr,
2357	&dev_attr_event_vinst.attr,
2358	&dev_attr_s_exlevel_vinst.attr,
2359	&dev_attr_ns_exlevel_vinst.attr,
2360	&dev_attr_addr_idx.attr,
2361	&dev_attr_addr_instdatatype.attr,
2362	&dev_attr_addr_single.attr,
2363	&dev_attr_addr_range.attr,
2364	&dev_attr_addr_start.attr,
2365	&dev_attr_addr_stop.attr,
2366	&dev_attr_addr_ctxtype.attr,
2367	&dev_attr_addr_context.attr,
2368	&dev_attr_addr_exlevel_s_ns.attr,
2369	&dev_attr_addr_cmp_view.attr,
2370	&dev_attr_vinst_pe_cmp_start_stop.attr,
2371	&dev_attr_sshot_idx.attr,
2372	&dev_attr_sshot_ctrl.attr,
2373	&dev_attr_sshot_pe_ctrl.attr,
2374	&dev_attr_sshot_status.attr,
2375	&dev_attr_seq_idx.attr,
2376	&dev_attr_seq_state.attr,
2377	&dev_attr_seq_event.attr,
2378	&dev_attr_seq_reset_event.attr,
2379	&dev_attr_cntr_idx.attr,
2380	&dev_attr_cntrldvr.attr,
2381	&dev_attr_cntr_val.attr,
2382	&dev_attr_cntr_ctrl.attr,
2383	&dev_attr_res_idx.attr,
2384	&dev_attr_res_ctrl.attr,
2385	&dev_attr_ctxid_idx.attr,
2386	&dev_attr_ctxid_pid.attr,
2387	&dev_attr_ctxid_masks.attr,
2388	&dev_attr_vmid_idx.attr,
2389	&dev_attr_vmid_val.attr,
2390	&dev_attr_vmid_masks.attr,
2391	&dev_attr_cpu.attr,
2392	&dev_attr_ts_source.attr,
2393	NULL,
2394};
2395
2396/*
2397 * Trace ID allocated dynamically on enable - but also allocate on read
2398 * in case sysfs or perf read before enable to ensure consistent metadata
2399 * information for trace decode
2400 */
2401static ssize_t trctraceid_show(struct device *dev,
2402			       struct device_attribute *attr,
2403			       char *buf)
2404{
2405	int trace_id;
2406	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
2407
2408	trace_id = etm4_read_alloc_trace_id(drvdata);
2409	if (trace_id < 0)
2410		return trace_id;
2411
2412	return sysfs_emit(buf, "0x%x\n", trace_id);
2413}
2414
2415struct etmv4_reg {
2416	struct coresight_device *csdev;
2417	u32 offset;
2418	u32 data;
2419};
2420
2421static void do_smp_cross_read(void *data)
2422{
2423	struct etmv4_reg *reg = data;
2424
2425	reg->data = etm4x_relaxed_read32(&reg->csdev->access, reg->offset);
2426}
2427
2428static u32 etmv4_cross_read(const struct etmv4_drvdata *drvdata, u32 offset)
2429{
2430	struct etmv4_reg reg;
2431
2432	reg.offset = offset;
2433	reg.csdev = drvdata->csdev;
2434
2435	/*
2436	 * smp cross call ensures the CPU will be powered up before
2437	 * accessing the ETMv4 trace core registers
2438	 */
2439	smp_call_function_single(drvdata->cpu, do_smp_cross_read, &reg, 1);
2440	return reg.data;
2441}
2442
2443static inline u32 coresight_etm4x_attr_to_offset(struct device_attribute *attr)
2444{
2445	struct dev_ext_attribute *eattr;
2446
2447	eattr = container_of(attr, struct dev_ext_attribute, attr);
2448	return (u32)(unsigned long)eattr->var;
2449}
2450
2451static ssize_t coresight_etm4x_reg_show(struct device *dev,
2452					struct device_attribute *d_attr,
2453					char *buf)
2454{
2455	u32 val, offset;
2456	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
2457
2458	offset = coresight_etm4x_attr_to_offset(d_attr);
2459
2460	pm_runtime_get_sync(dev->parent);
2461	val = etmv4_cross_read(drvdata, offset);
2462	pm_runtime_put_sync(dev->parent);
2463
2464	return scnprintf(buf, PAGE_SIZE, "0x%x\n", val);
2465}
2466
2467static inline bool
2468etm4x_register_implemented(struct etmv4_drvdata *drvdata, u32 offset)
2469{
2470	switch (offset) {
2471	ETM_COMMON_SYSREG_LIST_CASES
2472		/*
2473		 * Common registers to ETE & ETM4x accessible via system
2474		 * instructions are always implemented.
2475		 */
2476		return true;
2477
2478	ETM4x_ONLY_SYSREG_LIST_CASES
2479		/*
2480		 * We only support etm4x and ete. So if the device is not
2481		 * ETE, it must be ETMv4x.
2482		 */
2483		return !etm4x_is_ete(drvdata);
2484
2485	ETM4x_MMAP_LIST_CASES
2486		/*
2487		 * Registers accessible only via memory-mapped registers
2488		 * must not be accessed via system instructions.
2489		 * We cannot access the drvdata->csdev here, as this
2490		 * function is called during the device creation, via
2491		 * coresight_register() and the csdev is not initialized
2492		 * until that is done. So rely on the drvdata->base to
2493		 * detect if we have a memory mapped access.
2494		 * Also ETE doesn't implement memory mapped access, thus
2495		 * it is sufficient to check that we are using mmio.
2496		 */
2497		return !!drvdata->base;
2498
2499	ETE_ONLY_SYSREG_LIST_CASES
2500		return etm4x_is_ete(drvdata);
2501	}
2502
2503	return false;
2504}
2505
2506/*
2507 * Hide the ETM4x registers that may not be available on the
2508 * hardware.
2509 * There are certain management registers unavailable via system
2510 * instructions. Make those sysfs attributes hidden on such
2511 * systems.
2512 */
2513static umode_t
2514coresight_etm4x_attr_reg_implemented(struct kobject *kobj,
2515				     struct attribute *attr, int unused)
2516{
2517	struct device *dev = kobj_to_dev(kobj);
2518	struct etmv4_drvdata *drvdata = dev_get_drvdata(dev->parent);
2519	struct device_attribute *d_attr;
2520	u32 offset;
2521
2522	d_attr = container_of(attr, struct device_attribute, attr);
2523	offset = coresight_etm4x_attr_to_offset(d_attr);
2524
2525	if (etm4x_register_implemented(drvdata, offset))
2526		return attr->mode;
2527	return 0;
2528}
2529
2530/*
2531 * Macro to set an RO ext attribute with offset and show function.
2532 * Offset is used in mgmt group to ensure only correct registers for
2533 * the ETM / ETE variant are visible.
2534 */
2535#define coresight_etm4x_reg_showfn(name, offset, showfn) (	\
2536	&((struct dev_ext_attribute[]) {			\
2537	   {							\
2538		__ATTR(name, 0444, showfn, NULL),		\
2539		(void *)(unsigned long)offset			\
2540	   }							\
2541	})[0].attr.attr						\
2542	)
2543
2544/* macro using the default coresight_etm4x_reg_show function */
2545#define coresight_etm4x_reg(name, offset)	\
2546	coresight_etm4x_reg_showfn(name, offset, coresight_etm4x_reg_show)
2547
2548static struct attribute *coresight_etmv4_mgmt_attrs[] = {
2549	coresight_etm4x_reg(trcpdcr, TRCPDCR),
2550	coresight_etm4x_reg(trcpdsr, TRCPDSR),
2551	coresight_etm4x_reg(trclsr, TRCLSR),
2552	coresight_etm4x_reg(trcauthstatus, TRCAUTHSTATUS),
2553	coresight_etm4x_reg(trcdevid, TRCDEVID),
2554	coresight_etm4x_reg(trcdevtype, TRCDEVTYPE),
2555	coresight_etm4x_reg(trcpidr0, TRCPIDR0),
2556	coresight_etm4x_reg(trcpidr1, TRCPIDR1),
2557	coresight_etm4x_reg(trcpidr2, TRCPIDR2),
2558	coresight_etm4x_reg(trcpidr3, TRCPIDR3),
2559	coresight_etm4x_reg(trcoslsr, TRCOSLSR),
2560	coresight_etm4x_reg(trcconfig, TRCCONFIGR),
2561	coresight_etm4x_reg_showfn(trctraceid, TRCTRACEIDR, trctraceid_show),
2562	coresight_etm4x_reg(trcdevarch, TRCDEVARCH),
2563	NULL,
2564};
2565
2566static struct attribute *coresight_etmv4_trcidr_attrs[] = {
2567	coresight_etm4x_reg(trcidr0, TRCIDR0),
2568	coresight_etm4x_reg(trcidr1, TRCIDR1),
2569	coresight_etm4x_reg(trcidr2, TRCIDR2),
2570	coresight_etm4x_reg(trcidr3, TRCIDR3),
2571	coresight_etm4x_reg(trcidr4, TRCIDR4),
2572	coresight_etm4x_reg(trcidr5, TRCIDR5),
2573	/* trcidr[6,7] are reserved */
2574	coresight_etm4x_reg(trcidr8, TRCIDR8),
2575	coresight_etm4x_reg(trcidr9, TRCIDR9),
2576	coresight_etm4x_reg(trcidr10, TRCIDR10),
2577	coresight_etm4x_reg(trcidr11, TRCIDR11),
2578	coresight_etm4x_reg(trcidr12, TRCIDR12),
2579	coresight_etm4x_reg(trcidr13, TRCIDR13),
2580	NULL,
2581};
2582
2583static const struct attribute_group coresight_etmv4_group = {
2584	.attrs = coresight_etmv4_attrs,
2585};
2586
2587static const struct attribute_group coresight_etmv4_mgmt_group = {
2588	.is_visible = coresight_etm4x_attr_reg_implemented,
2589	.attrs = coresight_etmv4_mgmt_attrs,
2590	.name = "mgmt",
2591};
2592
2593static const struct attribute_group coresight_etmv4_trcidr_group = {
2594	.attrs = coresight_etmv4_trcidr_attrs,
2595	.name = "trcidr",
2596};
2597
2598const struct attribute_group *coresight_etmv4_groups[] = {
2599	&coresight_etmv4_group,
2600	&coresight_etmv4_mgmt_group,
2601	&coresight_etmv4_trcidr_group,
2602	NULL,
2603};
2604