1/* SPDX-License-Identifier: GPL-2.0 */
2
3#if !defined(_TRACE_KVM_PR_H) || defined(TRACE_HEADER_MULTI_READ)
4#define _TRACE_KVM_PR_H
5
6#include <linux/tracepoint.h>
7#include "trace_book3s.h"
8
9#undef TRACE_SYSTEM
10#define TRACE_SYSTEM kvm_pr
11
12TRACE_EVENT(kvm_book3s_reenter,
13	TP_PROTO(int r, struct kvm_vcpu *vcpu),
14	TP_ARGS(r, vcpu),
15
16	TP_STRUCT__entry(
17		__field(	unsigned int,	r		)
18		__field(	unsigned long,	pc		)
19	),
20
21	TP_fast_assign(
22		__entry->r		= r;
23		__entry->pc		= kvmppc_get_pc(vcpu);
24	),
25
26	TP_printk("reentry r=%d | pc=0x%lx", __entry->r, __entry->pc)
27);
28
29#ifdef CONFIG_PPC_BOOK3S_64
30
31TRACE_EVENT(kvm_book3s_64_mmu_map,
32	TP_PROTO(int rflags, ulong hpteg, ulong va, kvm_pfn_t hpaddr,
33		 struct kvmppc_pte *orig_pte),
34	TP_ARGS(rflags, hpteg, va, hpaddr, orig_pte),
35
36	TP_STRUCT__entry(
37		__field(	unsigned char,		flag_w		)
38		__field(	unsigned char,		flag_x		)
39		__field(	unsigned long,		eaddr		)
40		__field(	unsigned long,		hpteg		)
41		__field(	unsigned long,		va		)
42		__field(	unsigned long long,	vpage		)
43		__field(	unsigned long,		hpaddr		)
44	),
45
46	TP_fast_assign(
47		__entry->flag_w	= ((rflags & HPTE_R_PP) == 3) ? '-' : 'w';
48		__entry->flag_x	= (rflags & HPTE_R_N) ? '-' : 'x';
49		__entry->eaddr	= orig_pte->eaddr;
50		__entry->hpteg	= hpteg;
51		__entry->va	= va;
52		__entry->vpage	= orig_pte->vpage;
53		__entry->hpaddr	= hpaddr;
54	),
55
56	TP_printk("KVM: %c%c Map 0x%lx: [%lx] 0x%lx (0x%llx) -> %lx",
57		  __entry->flag_w, __entry->flag_x, __entry->eaddr,
58		  __entry->hpteg, __entry->va, __entry->vpage, __entry->hpaddr)
59);
60
61#endif /* CONFIG_PPC_BOOK3S_64 */
62
63TRACE_EVENT(kvm_book3s_mmu_map,
64	TP_PROTO(struct hpte_cache *pte),
65	TP_ARGS(pte),
66
67	TP_STRUCT__entry(
68		__field(	u64,		host_vpn	)
69		__field(	u64,		pfn		)
70		__field(	ulong,		eaddr		)
71		__field(	u64,		vpage		)
72		__field(	ulong,		raddr		)
73		__field(	int,		flags		)
74	),
75
76	TP_fast_assign(
77		__entry->host_vpn	= pte->host_vpn;
78		__entry->pfn		= pte->pfn;
79		__entry->eaddr		= pte->pte.eaddr;
80		__entry->vpage		= pte->pte.vpage;
81		__entry->raddr		= pte->pte.raddr;
82		__entry->flags		= (pte->pte.may_read ? 0x4 : 0) |
83					  (pte->pte.may_write ? 0x2 : 0) |
84					  (pte->pte.may_execute ? 0x1 : 0);
85	),
86
87	TP_printk("Map: hvpn=%llx pfn=%llx ea=%lx vp=%llx ra=%lx [%x]",
88		  __entry->host_vpn, __entry->pfn, __entry->eaddr,
89		  __entry->vpage, __entry->raddr, __entry->flags)
90);
91
92TRACE_EVENT(kvm_book3s_mmu_invalidate,
93	TP_PROTO(struct hpte_cache *pte),
94	TP_ARGS(pte),
95
96	TP_STRUCT__entry(
97		__field(	u64,		host_vpn	)
98		__field(	u64,		pfn		)
99		__field(	ulong,		eaddr		)
100		__field(	u64,		vpage		)
101		__field(	ulong,		raddr		)
102		__field(	int,		flags		)
103	),
104
105	TP_fast_assign(
106		__entry->host_vpn	= pte->host_vpn;
107		__entry->pfn		= pte->pfn;
108		__entry->eaddr		= pte->pte.eaddr;
109		__entry->vpage		= pte->pte.vpage;
110		__entry->raddr		= pte->pte.raddr;
111		__entry->flags		= (pte->pte.may_read ? 0x4 : 0) |
112					  (pte->pte.may_write ? 0x2 : 0) |
113					  (pte->pte.may_execute ? 0x1 : 0);
114	),
115
116	TP_printk("Flush: hva=%llx pfn=%llx ea=%lx vp=%llx ra=%lx [%x]",
117		  __entry->host_vpn, __entry->pfn, __entry->eaddr,
118		  __entry->vpage, __entry->raddr, __entry->flags)
119);
120
121TRACE_EVENT(kvm_book3s_mmu_flush,
122	TP_PROTO(const char *type, struct kvm_vcpu *vcpu, unsigned long long p1,
123		 unsigned long long p2),
124	TP_ARGS(type, vcpu, p1, p2),
125
126	TP_STRUCT__entry(
127		__field(	int,			count		)
128		__field(	unsigned long long,	p1		)
129		__field(	unsigned long long,	p2		)
130		__field(	const char *,		type		)
131	),
132
133	TP_fast_assign(
134		__entry->count		= to_book3s(vcpu)->hpte_cache_count;
135		__entry->p1		= p1;
136		__entry->p2		= p2;
137		__entry->type		= type;
138	),
139
140	TP_printk("Flush %d %sPTEs: %llx - %llx",
141		  __entry->count, __entry->type, __entry->p1, __entry->p2)
142);
143
144TRACE_EVENT(kvm_book3s_slb_found,
145	TP_PROTO(unsigned long long gvsid, unsigned long long hvsid),
146	TP_ARGS(gvsid, hvsid),
147
148	TP_STRUCT__entry(
149		__field(	unsigned long long,	gvsid		)
150		__field(	unsigned long long,	hvsid		)
151	),
152
153	TP_fast_assign(
154		__entry->gvsid		= gvsid;
155		__entry->hvsid		= hvsid;
156	),
157
158	TP_printk("%llx -> %llx", __entry->gvsid, __entry->hvsid)
159);
160
161TRACE_EVENT(kvm_book3s_slb_fail,
162	TP_PROTO(u16 sid_map_mask, unsigned long long gvsid),
163	TP_ARGS(sid_map_mask, gvsid),
164
165	TP_STRUCT__entry(
166		__field(	unsigned short,		sid_map_mask	)
167		__field(	unsigned long long,	gvsid		)
168	),
169
170	TP_fast_assign(
171		__entry->sid_map_mask	= sid_map_mask;
172		__entry->gvsid		= gvsid;
173	),
174
175	TP_printk("%x/%x: %llx", __entry->sid_map_mask,
176		  SID_MAP_MASK - __entry->sid_map_mask, __entry->gvsid)
177);
178
179TRACE_EVENT(kvm_book3s_slb_map,
180	TP_PROTO(u16 sid_map_mask, unsigned long long gvsid,
181		 unsigned long long hvsid),
182	TP_ARGS(sid_map_mask, gvsid, hvsid),
183
184	TP_STRUCT__entry(
185		__field(	unsigned short,		sid_map_mask	)
186		__field(	unsigned long long,	guest_vsid	)
187		__field(	unsigned long long,	host_vsid	)
188	),
189
190	TP_fast_assign(
191		__entry->sid_map_mask	= sid_map_mask;
192		__entry->guest_vsid	= gvsid;
193		__entry->host_vsid	= hvsid;
194	),
195
196	TP_printk("%x: %llx -> %llx", __entry->sid_map_mask,
197		  __entry->guest_vsid, __entry->host_vsid)
198);
199
200TRACE_EVENT(kvm_book3s_slbmte,
201	TP_PROTO(u64 slb_vsid, u64 slb_esid),
202	TP_ARGS(slb_vsid, slb_esid),
203
204	TP_STRUCT__entry(
205		__field(	u64,	slb_vsid	)
206		__field(	u64,	slb_esid	)
207	),
208
209	TP_fast_assign(
210		__entry->slb_vsid	= slb_vsid;
211		__entry->slb_esid	= slb_esid;
212	),
213
214	TP_printk("%llx, %llx", __entry->slb_vsid, __entry->slb_esid)
215);
216
217TRACE_EVENT(kvm_exit,
218	TP_PROTO(unsigned int exit_nr, struct kvm_vcpu *vcpu),
219	TP_ARGS(exit_nr, vcpu),
220
221	TP_STRUCT__entry(
222		__field(	unsigned int,	exit_nr		)
223		__field(	unsigned long,	pc		)
224		__field(	unsigned long,	msr		)
225		__field(	unsigned long,	dar		)
226		__field(	unsigned long,	srr1		)
227		__field(	unsigned long,	last_inst	)
228	),
229
230	TP_fast_assign(
231		__entry->exit_nr	= exit_nr;
232		__entry->pc		= kvmppc_get_pc(vcpu);
233		__entry->dar		= kvmppc_get_fault_dar(vcpu);
234		__entry->msr		= kvmppc_get_msr(vcpu);
235		__entry->srr1		= vcpu->arch.shadow_srr1;
236		__entry->last_inst	= vcpu->arch.last_inst;
237	),
238
239	TP_printk("exit=%s"
240		" | pc=0x%lx"
241		" | msr=0x%lx"
242		" | dar=0x%lx"
243		" | srr1=0x%lx"
244		" | last_inst=0x%lx"
245		,
246		__print_symbolic(__entry->exit_nr, kvm_trace_symbol_exit),
247		__entry->pc,
248		__entry->msr,
249		__entry->dar,
250		__entry->srr1,
251		__entry->last_inst
252		)
253);
254
255#endif /* _TRACE_KVM_H */
256
257/* This part must be outside protection */
258
259#undef TRACE_INCLUDE_PATH
260#undef TRACE_INCLUDE_FILE
261
262#define TRACE_INCLUDE_PATH .
263#define TRACE_INCLUDE_FILE trace_pr
264
265#include <trace/define_trace.h>
266