1/* SPDX-License-Identifier: GPL-2.0 */ 2/* Copyright (C) 2021-2022 Intel Corporation */ 3#ifndef _ASM_X86_TDX_H 4#define _ASM_X86_TDX_H 5 6#include <linux/init.h> 7#include <linux/bits.h> 8 9#include <asm/errno.h> 10#include <asm/ptrace.h> 11#include <asm/trapnr.h> 12#include <asm/shared/tdx.h> 13 14/* 15 * SW-defined error codes. 16 * 17 * Bits 47:40 == 0xFF indicate Reserved status code class that never used by 18 * TDX module. 19 */ 20#define TDX_ERROR _BITUL(63) 21#define TDX_SW_ERROR (TDX_ERROR | GENMASK_ULL(47, 40)) 22#define TDX_SEAMCALL_VMFAILINVALID (TDX_SW_ERROR | _UL(0xFFFF0000)) 23 24#define TDX_SEAMCALL_GP (TDX_SW_ERROR | X86_TRAP_GP) 25#define TDX_SEAMCALL_UD (TDX_SW_ERROR | X86_TRAP_UD) 26 27/* 28 * TDX module SEAMCALL leaf function error codes 29 */ 30#define TDX_SUCCESS 0ULL 31#define TDX_RND_NO_ENTROPY 0x8000020300000000ULL 32 33#ifndef __ASSEMBLY__ 34 35#include <uapi/asm/mce.h> 36 37/* 38 * Used by the #VE exception handler to gather the #VE exception 39 * info from the TDX module. This is a software only structure 40 * and not part of the TDX module/VMM ABI. 41 */ 42struct ve_info { 43 u64 exit_reason; 44 u64 exit_qual; 45 /* Guest Linear (virtual) Address */ 46 u64 gla; 47 /* Guest Physical Address */ 48 u64 gpa; 49 u32 instr_len; 50 u32 instr_info; 51}; 52 53#ifdef CONFIG_INTEL_TDX_GUEST 54 55void __init tdx_early_init(void); 56 57void tdx_get_ve_info(struct ve_info *ve); 58 59bool tdx_handle_virt_exception(struct pt_regs *regs, struct ve_info *ve); 60 61void tdx_safe_halt(void); 62 63bool tdx_early_handle_ve(struct pt_regs *regs); 64 65int tdx_mcall_get_report0(u8 *reportdata, u8 *tdreport); 66 67u64 tdx_hcall_get_quote(u8 *buf, size_t size); 68 69#else 70 71static inline void tdx_early_init(void) { }; 72static inline void tdx_safe_halt(void) { }; 73 74static inline bool tdx_early_handle_ve(struct pt_regs *regs) { return false; } 75 76#endif /* CONFIG_INTEL_TDX_GUEST */ 77 78#if defined(CONFIG_KVM_GUEST) && defined(CONFIG_INTEL_TDX_GUEST) 79long tdx_kvm_hypercall(unsigned int nr, unsigned long p1, unsigned long p2, 80 unsigned long p3, unsigned long p4); 81#else 82static inline long tdx_kvm_hypercall(unsigned int nr, unsigned long p1, 83 unsigned long p2, unsigned long p3, 84 unsigned long p4) 85{ 86 return -ENODEV; 87} 88#endif /* CONFIG_INTEL_TDX_GUEST && CONFIG_KVM_GUEST */ 89 90#ifdef CONFIG_INTEL_TDX_HOST 91u64 __seamcall(u64 fn, struct tdx_module_args *args); 92u64 __seamcall_ret(u64 fn, struct tdx_module_args *args); 93u64 __seamcall_saved_ret(u64 fn, struct tdx_module_args *args); 94void tdx_init(void); 95 96#include <asm/archrandom.h> 97 98typedef u64 (*sc_func_t)(u64 fn, struct tdx_module_args *args); 99 100static inline u64 sc_retry(sc_func_t func, u64 fn, 101 struct tdx_module_args *args) 102{ 103 int retry = RDRAND_RETRY_LOOPS; 104 u64 ret; 105 106 do { 107 ret = func(fn, args); 108 } while (ret == TDX_RND_NO_ENTROPY && --retry); 109 110 return ret; 111} 112 113#define seamcall(_fn, _args) sc_retry(__seamcall, (_fn), (_args)) 114#define seamcall_ret(_fn, _args) sc_retry(__seamcall_ret, (_fn), (_args)) 115#define seamcall_saved_ret(_fn, _args) sc_retry(__seamcall_saved_ret, (_fn), (_args)) 116int tdx_cpu_enable(void); 117int tdx_enable(void); 118const char *tdx_dump_mce_info(struct mce *m); 119#else 120static inline void tdx_init(void) { } 121static inline int tdx_cpu_enable(void) { return -ENODEV; } 122static inline int tdx_enable(void) { return -ENODEV; } 123static inline const char *tdx_dump_mce_info(struct mce *m) { return NULL; } 124#endif /* CONFIG_INTEL_TDX_HOST */ 125 126#endif /* !__ASSEMBLY__ */ 127#endif /* _ASM_X86_TDX_H */ 128