1/* SPDX-License-Identifier: GPL-2.0-only */
2/*
3 * Copyright (C) 2012 ARM Ltd.
4 * Author: Marc Zyngier <marc.zyngier@arm.com>
5 */
6
7#ifndef __ASM__VIRT_H
8#define __ASM__VIRT_H
9
10/*
11 * The arm64 hcall implementation uses x0 to specify the hcall
12 * number. A value less than HVC_STUB_HCALL_NR indicates a special
13 * hcall, such as set vector. Any other value is handled in a
14 * hypervisor specific way.
15 *
16 * The hypercall is allowed to clobber any of the caller-saved
17 * registers (x0-x18), so it is advisable to use it through the
18 * indirection of a function call (as implemented in hyp-stub.S).
19 */
20
21/*
22 * HVC_SET_VECTORS - Set the value of the vbar_el2 register.
23 *
24 * @x1: Physical address of the new vector table.
25 */
26#define HVC_SET_VECTORS 0
27
28/*
29 * HVC_SOFT_RESTART - CPU soft reset, used by the cpu_soft_restart routine.
30 */
31#define HVC_SOFT_RESTART 1
32
33/*
34 * HVC_RESET_VECTORS - Restore the vectors to the original HYP stubs
35 */
36#define HVC_RESET_VECTORS 2
37
38/*
39 * HVC_FINALISE_EL2 - Upgrade the CPU from EL1 to EL2, if possible
40 */
41#define HVC_FINALISE_EL2	3
42
43/* Max number of HYP stub hypercalls */
44#define HVC_STUB_HCALL_NR 4
45
46/* Error returned when an invalid stub number is passed into x0 */
47#define HVC_STUB_ERR	0xbadca11
48
49#define BOOT_CPU_MODE_EL1	(0xe11)
50#define BOOT_CPU_MODE_EL2	(0xe12)
51
52/*
53 * Flags returned together with the boot mode, but not preserved in
54 * __boot_cpu_mode. Used by the idreg override code to work out the
55 * boot state.
56 */
57#define BOOT_CPU_FLAG_E2H	BIT_ULL(32)
58
59#ifndef __ASSEMBLY__
60
61#include <asm/ptrace.h>
62#include <asm/sections.h>
63#include <asm/sysreg.h>
64#include <asm/cpufeature.h>
65
66/*
67 * __boot_cpu_mode records what mode CPUs were booted in.
68 * A correctly-implemented bootloader must start all CPUs in the same mode:
69 * In this case, both 32bit halves of __boot_cpu_mode will contain the
70 * same value (either 0 if booted in EL1, BOOT_CPU_MODE_EL2 if booted in EL2).
71 *
72 * Should the bootloader fail to do this, the two values will be different.
73 * This allows the kernel to flag an error when the secondaries have come up.
74 */
75extern u32 __boot_cpu_mode[2];
76
77#define ARM64_VECTOR_TABLE_LEN	SZ_2K
78
79void __hyp_set_vectors(phys_addr_t phys_vector_base);
80void __hyp_reset_vectors(void);
81bool is_kvm_arm_initialised(void);
82
83DECLARE_STATIC_KEY_FALSE(kvm_protected_mode_initialized);
84
85/* Reports the availability of HYP mode */
86static inline bool is_hyp_mode_available(void)
87{
88	/*
89	 * If KVM protected mode is initialized, all CPUs must have been booted
90	 * in EL2. Avoid checking __boot_cpu_mode as CPUs now come up in EL1.
91	 */
92	if (IS_ENABLED(CONFIG_KVM) &&
93	    static_branch_likely(&kvm_protected_mode_initialized))
94		return true;
95
96	return (__boot_cpu_mode[0] == BOOT_CPU_MODE_EL2 &&
97		__boot_cpu_mode[1] == BOOT_CPU_MODE_EL2);
98}
99
100/* Check if the bootloader has booted CPUs in different modes */
101static inline bool is_hyp_mode_mismatched(void)
102{
103	/*
104	 * If KVM protected mode is initialized, all CPUs must have been booted
105	 * in EL2. Avoid checking __boot_cpu_mode as CPUs now come up in EL1.
106	 */
107	if (IS_ENABLED(CONFIG_KVM) &&
108	    static_branch_likely(&kvm_protected_mode_initialized))
109		return false;
110
111	return __boot_cpu_mode[0] != __boot_cpu_mode[1];
112}
113
114static __always_inline bool is_kernel_in_hyp_mode(void)
115{
116	BUILD_BUG_ON(__is_defined(__KVM_NVHE_HYPERVISOR__) ||
117		     __is_defined(__KVM_VHE_HYPERVISOR__));
118	return read_sysreg(CurrentEL) == CurrentEL_EL2;
119}
120
121static __always_inline bool has_vhe(void)
122{
123	/*
124	 * Code only run in VHE/NVHE hyp context can assume VHE is present or
125	 * absent. Otherwise fall back to caps.
126	 * This allows the compiler to discard VHE-specific code from the
127	 * nVHE object, reducing the number of external symbol references
128	 * needed to link.
129	 */
130	if (is_vhe_hyp_code())
131		return true;
132	else if (is_nvhe_hyp_code())
133		return false;
134	else
135		return cpus_have_final_cap(ARM64_HAS_VIRT_HOST_EXTN);
136}
137
138static __always_inline bool is_protected_kvm_enabled(void)
139{
140	if (is_vhe_hyp_code())
141		return false;
142	else
143		return cpus_have_final_cap(ARM64_KVM_PROTECTED_MODE);
144}
145
146static __always_inline bool has_hvhe(void)
147{
148	if (is_vhe_hyp_code())
149		return false;
150
151	return cpus_have_final_cap(ARM64_KVM_HVHE);
152}
153
154static inline bool is_hyp_nvhe(void)
155{
156	return is_hyp_mode_available() && !is_kernel_in_hyp_mode();
157}
158
159#endif /* __ASSEMBLY__ */
160
161#endif /* ! __ASM__VIRT_H */
162