1/* SPDX-License-Identifier: GPL-2.0 */
2/*
3 * AArch64 processor specific defines
4 *
5 * Copyright (C) 2018, Red Hat, Inc.
6 */
7#ifndef SELFTEST_KVM_PROCESSOR_H
8#define SELFTEST_KVM_PROCESSOR_H
9
10#include "kvm_util.h"
11#include <linux/stringify.h>
12#include <linux/types.h>
13#include <asm/sysreg.h>
14
15
16#define ARM64_CORE_REG(x) (KVM_REG_ARM64 | KVM_REG_SIZE_U64 | \
17			   KVM_REG_ARM_CORE | KVM_REG_ARM_CORE_REG(x))
18
19/*
20 * KVM_ARM64_SYS_REG(sys_reg_id): Helper macro to convert
21 * SYS_* register definitions in asm/sysreg.h to use in KVM
22 * calls such as vcpu_get_reg() and vcpu_set_reg().
23 */
24#define KVM_ARM64_SYS_REG(sys_reg_id)			\
25	ARM64_SYS_REG(sys_reg_Op0(sys_reg_id),		\
26			sys_reg_Op1(sys_reg_id),	\
27			sys_reg_CRn(sys_reg_id),	\
28			sys_reg_CRm(sys_reg_id),	\
29			sys_reg_Op2(sys_reg_id))
30
31/*
32 * Default MAIR
33 *                  index   attribute
34 * DEVICE_nGnRnE      0     0000:0000
35 * DEVICE_nGnRE       1     0000:0100
36 * DEVICE_GRE         2     0000:1100
37 * NORMAL_NC          3     0100:0100
38 * NORMAL             4     1111:1111
39 * NORMAL_WT          5     1011:1011
40 */
41
42/* Linux doesn't use these memory types, so let's define them. */
43#define MAIR_ATTR_DEVICE_GRE	UL(0x0c)
44#define MAIR_ATTR_NORMAL_WT	UL(0xbb)
45
46#define MT_DEVICE_nGnRnE	0
47#define MT_DEVICE_nGnRE		1
48#define MT_DEVICE_GRE		2
49#define MT_NORMAL_NC		3
50#define MT_NORMAL		4
51#define MT_NORMAL_WT		5
52
53#define DEFAULT_MAIR_EL1							\
54	(MAIR_ATTRIDX(MAIR_ATTR_DEVICE_nGnRnE, MT_DEVICE_nGnRnE) |		\
55	 MAIR_ATTRIDX(MAIR_ATTR_DEVICE_nGnRE, MT_DEVICE_nGnRE) |		\
56	 MAIR_ATTRIDX(MAIR_ATTR_DEVICE_GRE, MT_DEVICE_GRE) |			\
57	 MAIR_ATTRIDX(MAIR_ATTR_NORMAL_NC, MT_NORMAL_NC) |			\
58	 MAIR_ATTRIDX(MAIR_ATTR_NORMAL, MT_NORMAL) |				\
59	 MAIR_ATTRIDX(MAIR_ATTR_NORMAL_WT, MT_NORMAL_WT))
60
61#define MPIDR_HWID_BITMASK (0xff00fffffful)
62
63void aarch64_vcpu_setup(struct kvm_vcpu *vcpu, struct kvm_vcpu_init *init);
64struct kvm_vcpu *aarch64_vcpu_add(struct kvm_vm *vm, uint32_t vcpu_id,
65				  struct kvm_vcpu_init *init, void *guest_code);
66
67struct ex_regs {
68	u64 regs[31];
69	u64 sp;
70	u64 pc;
71	u64 pstate;
72};
73
74#define VECTOR_NUM	16
75
76enum {
77	VECTOR_SYNC_CURRENT_SP0,
78	VECTOR_IRQ_CURRENT_SP0,
79	VECTOR_FIQ_CURRENT_SP0,
80	VECTOR_ERROR_CURRENT_SP0,
81
82	VECTOR_SYNC_CURRENT,
83	VECTOR_IRQ_CURRENT,
84	VECTOR_FIQ_CURRENT,
85	VECTOR_ERROR_CURRENT,
86
87	VECTOR_SYNC_LOWER_64,
88	VECTOR_IRQ_LOWER_64,
89	VECTOR_FIQ_LOWER_64,
90	VECTOR_ERROR_LOWER_64,
91
92	VECTOR_SYNC_LOWER_32,
93	VECTOR_IRQ_LOWER_32,
94	VECTOR_FIQ_LOWER_32,
95	VECTOR_ERROR_LOWER_32,
96};
97
98#define VECTOR_IS_SYNC(v) ((v) == VECTOR_SYNC_CURRENT_SP0 || \
99			   (v) == VECTOR_SYNC_CURRENT     || \
100			   (v) == VECTOR_SYNC_LOWER_64    || \
101			   (v) == VECTOR_SYNC_LOWER_32)
102
103#define ESR_EC_NUM		64
104#define ESR_EC_SHIFT		26
105#define ESR_EC_MASK		(ESR_EC_NUM - 1)
106
107#define ESR_EC_UNKNOWN		0x0
108#define ESR_EC_SVC64		0x15
109#define ESR_EC_IABT		0x21
110#define ESR_EC_DABT		0x25
111#define ESR_EC_HW_BP_CURRENT	0x31
112#define ESR_EC_SSTEP_CURRENT	0x33
113#define ESR_EC_WP_CURRENT	0x35
114#define ESR_EC_BRK_INS		0x3c
115
116/* Access flag */
117#define PTE_AF			(1ULL << 10)
118
119/* Access flag update enable/disable */
120#define TCR_EL1_HA		(1ULL << 39)
121
122void aarch64_get_supported_page_sizes(uint32_t ipa, uint32_t *ipa4k,
123					uint32_t *ipa16k, uint32_t *ipa64k);
124
125void vm_init_descriptor_tables(struct kvm_vm *vm);
126void vcpu_init_descriptor_tables(struct kvm_vcpu *vcpu);
127
128typedef void(*handler_fn)(struct ex_regs *);
129void vm_install_exception_handler(struct kvm_vm *vm,
130		int vector, handler_fn handler);
131void vm_install_sync_handler(struct kvm_vm *vm,
132		int vector, int ec, handler_fn handler);
133
134uint64_t *virt_get_pte_hva(struct kvm_vm *vm, vm_vaddr_t gva);
135
136static inline void cpu_relax(void)
137{
138	asm volatile("yield" ::: "memory");
139}
140
141#define isb()		asm volatile("isb" : : : "memory")
142#define dsb(opt)	asm volatile("dsb " #opt : : : "memory")
143#define dmb(opt)	asm volatile("dmb " #opt : : : "memory")
144
145#define dma_wmb()	dmb(oshst)
146#define __iowmb()	dma_wmb()
147
148#define dma_rmb()	dmb(oshld)
149
150#define __iormb(v)							\
151({									\
152	unsigned long tmp;						\
153									\
154	dma_rmb();							\
155									\
156	/*								\
157	 * Courtesy of arch/arm64/include/asm/io.h:			\
158	 * Create a dummy control dependency from the IO read to any	\
159	 * later instructions. This ensures that a subsequent call	\
160	 * to udelay() will be ordered due to the ISB in __delay().	\
161	 */								\
162	asm volatile("eor	%0, %1, %1\n"				\
163		     "cbnz	%0, ."					\
164		     : "=r" (tmp) : "r" ((unsigned long)(v))		\
165		     : "memory");					\
166})
167
168static __always_inline void __raw_writel(u32 val, volatile void *addr)
169{
170	asm volatile("str %w0, [%1]" : : "rZ" (val), "r" (addr));
171}
172
173static __always_inline u32 __raw_readl(const volatile void *addr)
174{
175	u32 val;
176	asm volatile("ldr %w0, [%1]" : "=r" (val) : "r" (addr));
177	return val;
178}
179
180#define writel_relaxed(v,c)	((void)__raw_writel((__force u32)cpu_to_le32(v),(c)))
181#define readl_relaxed(c)	({ u32 __r = le32_to_cpu((__force __le32)__raw_readl(c)); __r; })
182
183#define writel(v,c)		({ __iowmb(); writel_relaxed((v),(c));})
184#define readl(c)		({ u32 __v = readl_relaxed(c); __iormb(__v); __v; })
185
186static inline void local_irq_enable(void)
187{
188	asm volatile("msr daifclr, #3" : : : "memory");
189}
190
191static inline void local_irq_disable(void)
192{
193	asm volatile("msr daifset, #3" : : : "memory");
194}
195
196/**
197 * struct arm_smccc_res - Result from SMC/HVC call
198 * @a0-a3 result values from registers 0 to 3
199 */
200struct arm_smccc_res {
201	unsigned long a0;
202	unsigned long a1;
203	unsigned long a2;
204	unsigned long a3;
205};
206
207/**
208 * smccc_hvc - Invoke a SMCCC function using the hvc conduit
209 * @function_id: the SMCCC function to be called
210 * @arg0-arg6: SMCCC function arguments, corresponding to registers x1-x7
211 * @res: pointer to write the return values from registers x0-x3
212 *
213 */
214void smccc_hvc(uint32_t function_id, uint64_t arg0, uint64_t arg1,
215	       uint64_t arg2, uint64_t arg3, uint64_t arg4, uint64_t arg5,
216	       uint64_t arg6, struct arm_smccc_res *res);
217
218/**
219 * smccc_smc - Invoke a SMCCC function using the smc conduit
220 * @function_id: the SMCCC function to be called
221 * @arg0-arg6: SMCCC function arguments, corresponding to registers x1-x7
222 * @res: pointer to write the return values from registers x0-x3
223 *
224 */
225void smccc_smc(uint32_t function_id, uint64_t arg0, uint64_t arg1,
226	       uint64_t arg2, uint64_t arg3, uint64_t arg4, uint64_t arg5,
227	       uint64_t arg6, struct arm_smccc_res *res);
228
229#endif /* SELFTEST_KVM_PROCESSOR_H */
230