// SPDX-License-Identifier: GPL-2.0 /* * Hyper-V HvCallSendSyntheticClusterIpi{,Ex} tests * * Copyright (C) 2022, Red Hat, Inc. * */ #include #include #include "kvm_util.h" #include "hyperv.h" #include "test_util.h" #include "vmx.h" #define RECEIVER_VCPU_ID_1 2 #define RECEIVER_VCPU_ID_2 65 #define IPI_VECTOR 0xfe static volatile uint64_t ipis_rcvd[RECEIVER_VCPU_ID_2 + 1]; struct hv_vpset { u64 format; u64 valid_bank_mask; u64 bank_contents[2]; }; enum HV_GENERIC_SET_FORMAT { HV_GENERIC_SET_SPARSE_4K, HV_GENERIC_SET_ALL, }; /* HvCallSendSyntheticClusterIpi hypercall */ struct hv_send_ipi { u32 vector; u32 reserved; u64 cpu_mask; }; /* HvCallSendSyntheticClusterIpiEx hypercall */ struct hv_send_ipi_ex { u32 vector; u32 reserved; struct hv_vpset vp_set; }; static inline void hv_init(vm_vaddr_t pgs_gpa) { wrmsr(HV_X64_MSR_GUEST_OS_ID, HYPERV_LINUX_OS_ID); wrmsr(HV_X64_MSR_HYPERCALL, pgs_gpa); } static void receiver_code(void *hcall_page, vm_vaddr_t pgs_gpa) { u32 vcpu_id; x2apic_enable(); hv_init(pgs_gpa); vcpu_id = rdmsr(HV_X64_MSR_VP_INDEX); /* Signal sender vCPU we're ready */ ipis_rcvd[vcpu_id] = (u64)-1; for (;;) asm volatile("sti; hlt; cli"); } static void guest_ipi_handler(struct ex_regs *regs) { u32 vcpu_id = rdmsr(HV_X64_MSR_VP_INDEX); ipis_rcvd[vcpu_id]++; wrmsr(HV_X64_MSR_EOI, 1); } static inline void nop_loop(void) { int i; for (i = 0; i < 100000000; i++) asm volatile("nop"); } static void sender_guest_code(void *hcall_page, vm_vaddr_t pgs_gpa) { struct hv_send_ipi *ipi = (struct hv_send_ipi *)hcall_page; struct hv_send_ipi_ex *ipi_ex = (struct hv_send_ipi_ex *)hcall_page; int stage = 1, ipis_expected[2] = {0}; hv_init(pgs_gpa); GUEST_SYNC(stage++); /* Wait for receiver vCPUs to come up */ while (!ipis_rcvd[RECEIVER_VCPU_ID_1] || !ipis_rcvd[RECEIVER_VCPU_ID_2]) nop_loop(); ipis_rcvd[RECEIVER_VCPU_ID_1] = ipis_rcvd[RECEIVER_VCPU_ID_2] = 0; /* 'Slow' HvCallSendSyntheticClusterIpi to RECEIVER_VCPU_ID_1 */ ipi->vector = IPI_VECTOR; ipi->cpu_mask = 1 << RECEIVER_VCPU_ID_1; hyperv_hypercall(HVCALL_SEND_IPI, pgs_gpa, pgs_gpa + 4096); nop_loop(); GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_1] == ++ipis_expected[0]); GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_2] == ipis_expected[1]); GUEST_SYNC(stage++); /* 'Fast' HvCallSendSyntheticClusterIpi to RECEIVER_VCPU_ID_1 */ hyperv_hypercall(HVCALL_SEND_IPI | HV_HYPERCALL_FAST_BIT, IPI_VECTOR, 1 << RECEIVER_VCPU_ID_1); nop_loop(); GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_1] == ++ipis_expected[0]); GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_2] == ipis_expected[1]); GUEST_SYNC(stage++); /* 'Slow' HvCallSendSyntheticClusterIpiEx to RECEIVER_VCPU_ID_1 */ memset(hcall_page, 0, 4096); ipi_ex->vector = IPI_VECTOR; ipi_ex->vp_set.format = HV_GENERIC_SET_SPARSE_4K; ipi_ex->vp_set.valid_bank_mask = 1 << 0; ipi_ex->vp_set.bank_contents[0] = BIT(RECEIVER_VCPU_ID_1); hyperv_hypercall(HVCALL_SEND_IPI_EX | (1 << HV_HYPERCALL_VARHEAD_OFFSET), pgs_gpa, pgs_gpa + 4096); nop_loop(); GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_1] == ++ipis_expected[0]); GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_2] == ipis_expected[1]); GUEST_SYNC(stage++); /* 'XMM Fast' HvCallSendSyntheticClusterIpiEx to RECEIVER_VCPU_ID_1 */ hyperv_write_xmm_input(&ipi_ex->vp_set.valid_bank_mask, 1); hyperv_hypercall(HVCALL_SEND_IPI_EX | HV_HYPERCALL_FAST_BIT | (1 << HV_HYPERCALL_VARHEAD_OFFSET), IPI_VECTOR, HV_GENERIC_SET_SPARSE_4K); nop_loop(); GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_1] == ++ipis_expected[0]); GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_2] == ipis_expected[1]); GUEST_SYNC(stage++); /* 'Slow' HvCallSendSyntheticClusterIpiEx to RECEIVER_VCPU_ID_2 */ memset(hcall_page, 0, 4096); ipi_ex->vector = IPI_VECTOR; ipi_ex->vp_set.format = HV_GENERIC_SET_SPARSE_4K; ipi_ex->vp_set.valid_bank_mask = 1 << 1; ipi_ex->vp_set.bank_contents[0] = BIT(RECEIVER_VCPU_ID_2 - 64); hyperv_hypercall(HVCALL_SEND_IPI_EX | (1 << HV_HYPERCALL_VARHEAD_OFFSET), pgs_gpa, pgs_gpa + 4096); nop_loop(); GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_1] == ipis_expected[0]); GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_2] == ++ipis_expected[1]); GUEST_SYNC(stage++); /* 'XMM Fast' HvCallSendSyntheticClusterIpiEx to RECEIVER_VCPU_ID_2 */ hyperv_write_xmm_input(&ipi_ex->vp_set.valid_bank_mask, 1); hyperv_hypercall(HVCALL_SEND_IPI_EX | HV_HYPERCALL_FAST_BIT | (1 << HV_HYPERCALL_VARHEAD_OFFSET), IPI_VECTOR, HV_GENERIC_SET_SPARSE_4K); nop_loop(); GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_1] == ipis_expected[0]); GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_2] == ++ipis_expected[1]); GUEST_SYNC(stage++); /* 'Slow' HvCallSendSyntheticClusterIpiEx to both RECEIVER_VCPU_ID_{1,2} */ memset(hcall_page, 0, 4096); ipi_ex->vector = IPI_VECTOR; ipi_ex->vp_set.format = HV_GENERIC_SET_SPARSE_4K; ipi_ex->vp_set.valid_bank_mask = 1 << 1 | 1; ipi_ex->vp_set.bank_contents[0] = BIT(RECEIVER_VCPU_ID_1); ipi_ex->vp_set.bank_contents[1] = BIT(RECEIVER_VCPU_ID_2 - 64); hyperv_hypercall(HVCALL_SEND_IPI_EX | (2 << HV_HYPERCALL_VARHEAD_OFFSET), pgs_gpa, pgs_gpa + 4096); nop_loop(); GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_1] == ++ipis_expected[0]); GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_2] == ++ipis_expected[1]); GUEST_SYNC(stage++); /* 'XMM Fast' HvCallSendSyntheticClusterIpiEx to both RECEIVER_VCPU_ID_{1, 2} */ hyperv_write_xmm_input(&ipi_ex->vp_set.valid_bank_mask, 2); hyperv_hypercall(HVCALL_SEND_IPI_EX | HV_HYPERCALL_FAST_BIT | (2 << HV_HYPERCALL_VARHEAD_OFFSET), IPI_VECTOR, HV_GENERIC_SET_SPARSE_4K); nop_loop(); GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_1] == ++ipis_expected[0]); GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_2] == ++ipis_expected[1]); GUEST_SYNC(stage++); /* 'Slow' HvCallSendSyntheticClusterIpiEx to HV_GENERIC_SET_ALL */ memset(hcall_page, 0, 4096); ipi_ex->vector = IPI_VECTOR; ipi_ex->vp_set.format = HV_GENERIC_SET_ALL; hyperv_hypercall(HVCALL_SEND_IPI_EX, pgs_gpa, pgs_gpa + 4096); nop_loop(); GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_1] == ++ipis_expected[0]); GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_2] == ++ipis_expected[1]); GUEST_SYNC(stage++); /* * 'XMM Fast' HvCallSendSyntheticClusterIpiEx to HV_GENERIC_SET_ALL. */ ipi_ex->vp_set.valid_bank_mask = 0; hyperv_write_xmm_input(&ipi_ex->vp_set.valid_bank_mask, 2); hyperv_hypercall(HVCALL_SEND_IPI_EX | HV_HYPERCALL_FAST_BIT, IPI_VECTOR, HV_GENERIC_SET_ALL); nop_loop(); GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_1] == ++ipis_expected[0]); GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_2] == ++ipis_expected[1]); GUEST_SYNC(stage++); GUEST_DONE(); } static void *vcpu_thread(void *arg) { struct kvm_vcpu *vcpu = (struct kvm_vcpu *)arg; int old, r; r = pthread_setcanceltype(PTHREAD_CANCEL_ASYNCHRONOUS, &old); TEST_ASSERT(!r, "pthread_setcanceltype failed on vcpu_id=%u with errno=%d", vcpu->id, r); vcpu_run(vcpu); TEST_FAIL("vCPU %u exited unexpectedly", vcpu->id); return NULL; } static void cancel_join_vcpu_thread(pthread_t thread, struct kvm_vcpu *vcpu) { void *retval; int r; r = pthread_cancel(thread); TEST_ASSERT(!r, "pthread_cancel on vcpu_id=%d failed with errno=%d", vcpu->id, r); r = pthread_join(thread, &retval); TEST_ASSERT(!r, "pthread_join on vcpu_id=%d failed with errno=%d", vcpu->id, r); TEST_ASSERT(retval == PTHREAD_CANCELED, "expected retval=%p, got %p", PTHREAD_CANCELED, retval); } int main(int argc, char *argv[]) { struct kvm_vm *vm; struct kvm_vcpu *vcpu[3]; vm_vaddr_t hcall_page; pthread_t threads[2]; int stage = 1, r; struct ucall uc; TEST_REQUIRE(kvm_has_cap(KVM_CAP_HYPERV_SEND_IPI)); vm = vm_create_with_one_vcpu(&vcpu[0], sender_guest_code); /* Hypercall input/output */ hcall_page = vm_vaddr_alloc_pages(vm, 2); memset(addr_gva2hva(vm, hcall_page), 0x0, 2 * getpagesize()); vcpu[1] = vm_vcpu_add(vm, RECEIVER_VCPU_ID_1, receiver_code); vcpu_args_set(vcpu[1], 2, hcall_page, addr_gva2gpa(vm, hcall_page)); vcpu_set_msr(vcpu[1], HV_X64_MSR_VP_INDEX, RECEIVER_VCPU_ID_1); vcpu_set_hv_cpuid(vcpu[1]); vcpu[2] = vm_vcpu_add(vm, RECEIVER_VCPU_ID_2, receiver_code); vcpu_args_set(vcpu[2], 2, hcall_page, addr_gva2gpa(vm, hcall_page)); vcpu_set_msr(vcpu[2], HV_X64_MSR_VP_INDEX, RECEIVER_VCPU_ID_2); vcpu_set_hv_cpuid(vcpu[2]); vm_install_exception_handler(vm, IPI_VECTOR, guest_ipi_handler); vcpu_args_set(vcpu[0], 2, hcall_page, addr_gva2gpa(vm, hcall_page)); vcpu_set_hv_cpuid(vcpu[0]); r = pthread_create(&threads[0], NULL, vcpu_thread, vcpu[1]); TEST_ASSERT(!r, "pthread_create failed errno=%d", r); r = pthread_create(&threads[1], NULL, vcpu_thread, vcpu[2]); TEST_ASSERT(!r, "pthread_create failed errno=%d", errno); while (true) { vcpu_run(vcpu[0]); TEST_ASSERT_KVM_EXIT_REASON(vcpu[0], KVM_EXIT_IO); switch (get_ucall(vcpu[0], &uc)) { case UCALL_SYNC: TEST_ASSERT(uc.args[1] == stage, "Unexpected stage: %ld (%d expected)", uc.args[1], stage); break; case UCALL_DONE: goto done; case UCALL_ABORT: REPORT_GUEST_ASSERT(uc); /* NOT REACHED */ default: TEST_FAIL("Unknown ucall %lu", uc.cmd); } stage++; } done: cancel_join_vcpu_thread(threads[0], vcpu[1]); cancel_join_vcpu_thread(threads[1], vcpu[2]); kvm_vm_free(vm); return r; }