1// SPDX-License-Identifier: GPL-2.0
2/*
3 * Copyright (C) 2021 Western Digital Corporation or its affiliates.
4 * Copyright (C) 2022 Ventana Micro Systems Inc.
5 *
6 * Authors:
7 *	Anup Patel <apatel@ventanamicro.com>
8 */
9
10#include <linux/bits.h>
11#include <linux/kvm_host.h>
12#include <linux/uaccess.h>
13#include <asm/kvm_aia_imsic.h>
14
15static void unlock_vcpus(struct kvm *kvm, int vcpu_lock_idx)
16{
17	struct kvm_vcpu *tmp_vcpu;
18
19	for (; vcpu_lock_idx >= 0; vcpu_lock_idx--) {
20		tmp_vcpu = kvm_get_vcpu(kvm, vcpu_lock_idx);
21		mutex_unlock(&tmp_vcpu->mutex);
22	}
23}
24
25static void unlock_all_vcpus(struct kvm *kvm)
26{
27	unlock_vcpus(kvm, atomic_read(&kvm->online_vcpus) - 1);
28}
29
30static bool lock_all_vcpus(struct kvm *kvm)
31{
32	struct kvm_vcpu *tmp_vcpu;
33	unsigned long c;
34
35	kvm_for_each_vcpu(c, tmp_vcpu, kvm) {
36		if (!mutex_trylock(&tmp_vcpu->mutex)) {
37			unlock_vcpus(kvm, c - 1);
38			return false;
39		}
40	}
41
42	return true;
43}
44
45static int aia_create(struct kvm_device *dev, u32 type)
46{
47	int ret;
48	unsigned long i;
49	struct kvm *kvm = dev->kvm;
50	struct kvm_vcpu *vcpu;
51
52	if (irqchip_in_kernel(kvm))
53		return -EEXIST;
54
55	ret = -EBUSY;
56	if (!lock_all_vcpus(kvm))
57		return ret;
58
59	kvm_for_each_vcpu(i, vcpu, kvm) {
60		if (vcpu->arch.ran_atleast_once)
61			goto out_unlock;
62	}
63	ret = 0;
64
65	kvm->arch.aia.in_kernel = true;
66
67out_unlock:
68	unlock_all_vcpus(kvm);
69	return ret;
70}
71
72static void aia_destroy(struct kvm_device *dev)
73{
74	kfree(dev);
75}
76
77static int aia_config(struct kvm *kvm, unsigned long type,
78		      u32 *nr, bool write)
79{
80	struct kvm_aia *aia = &kvm->arch.aia;
81
82	/* Writes can only be done before irqchip is initialized */
83	if (write && kvm_riscv_aia_initialized(kvm))
84		return -EBUSY;
85
86	switch (type) {
87	case KVM_DEV_RISCV_AIA_CONFIG_MODE:
88		if (write) {
89			switch (*nr) {
90			case KVM_DEV_RISCV_AIA_MODE_EMUL:
91				break;
92			case KVM_DEV_RISCV_AIA_MODE_HWACCEL:
93			case KVM_DEV_RISCV_AIA_MODE_AUTO:
94				/*
95				 * HW Acceleration and Auto modes only
96				 * supported on host with non-zero guest
97				 * external interrupts (i.e. non-zero
98				 * VS-level IMSIC pages).
99				 */
100				if (!kvm_riscv_aia_nr_hgei)
101					return -EINVAL;
102				break;
103			default:
104				return -EINVAL;
105			}
106			aia->mode = *nr;
107		} else
108			*nr = aia->mode;
109		break;
110	case KVM_DEV_RISCV_AIA_CONFIG_IDS:
111		if (write) {
112			if ((*nr < KVM_DEV_RISCV_AIA_IDS_MIN) ||
113			    (*nr >= KVM_DEV_RISCV_AIA_IDS_MAX) ||
114			    ((*nr & KVM_DEV_RISCV_AIA_IDS_MIN) !=
115			     KVM_DEV_RISCV_AIA_IDS_MIN) ||
116			    (kvm_riscv_aia_max_ids <= *nr))
117				return -EINVAL;
118			aia->nr_ids = *nr;
119		} else
120			*nr = aia->nr_ids;
121		break;
122	case KVM_DEV_RISCV_AIA_CONFIG_SRCS:
123		if (write) {
124			if ((*nr >= KVM_DEV_RISCV_AIA_SRCS_MAX) ||
125			    (*nr >= kvm_riscv_aia_max_ids))
126				return -EINVAL;
127			aia->nr_sources = *nr;
128		} else
129			*nr = aia->nr_sources;
130		break;
131	case KVM_DEV_RISCV_AIA_CONFIG_GROUP_BITS:
132		if (write) {
133			if (*nr >= KVM_DEV_RISCV_AIA_GROUP_BITS_MAX)
134				return -EINVAL;
135			aia->nr_group_bits = *nr;
136		} else
137			*nr = aia->nr_group_bits;
138		break;
139	case KVM_DEV_RISCV_AIA_CONFIG_GROUP_SHIFT:
140		if (write) {
141			if ((*nr < KVM_DEV_RISCV_AIA_GROUP_SHIFT_MIN) ||
142			    (*nr >= KVM_DEV_RISCV_AIA_GROUP_SHIFT_MAX))
143				return -EINVAL;
144			aia->nr_group_shift = *nr;
145		} else
146			*nr = aia->nr_group_shift;
147		break;
148	case KVM_DEV_RISCV_AIA_CONFIG_HART_BITS:
149		if (write) {
150			if (*nr >= KVM_DEV_RISCV_AIA_HART_BITS_MAX)
151				return -EINVAL;
152			aia->nr_hart_bits = *nr;
153		} else
154			*nr = aia->nr_hart_bits;
155		break;
156	case KVM_DEV_RISCV_AIA_CONFIG_GUEST_BITS:
157		if (write) {
158			if (*nr >= KVM_DEV_RISCV_AIA_GUEST_BITS_MAX)
159				return -EINVAL;
160			aia->nr_guest_bits = *nr;
161		} else
162			*nr = aia->nr_guest_bits;
163		break;
164	default:
165		return -ENXIO;
166	}
167
168	return 0;
169}
170
171static int aia_aplic_addr(struct kvm *kvm, u64 *addr, bool write)
172{
173	struct kvm_aia *aia = &kvm->arch.aia;
174
175	if (write) {
176		/* Writes can only be done before irqchip is initialized */
177		if (kvm_riscv_aia_initialized(kvm))
178			return -EBUSY;
179
180		if (*addr & (KVM_DEV_RISCV_APLIC_ALIGN - 1))
181			return -EINVAL;
182
183		aia->aplic_addr = *addr;
184	} else
185		*addr = aia->aplic_addr;
186
187	return 0;
188}
189
190static int aia_imsic_addr(struct kvm *kvm, u64 *addr,
191			  unsigned long vcpu_idx, bool write)
192{
193	struct kvm_vcpu *vcpu;
194	struct kvm_vcpu_aia *vcpu_aia;
195
196	vcpu = kvm_get_vcpu(kvm, vcpu_idx);
197	if (!vcpu)
198		return -EINVAL;
199	vcpu_aia = &vcpu->arch.aia_context;
200
201	if (write) {
202		/* Writes can only be done before irqchip is initialized */
203		if (kvm_riscv_aia_initialized(kvm))
204			return -EBUSY;
205
206		if (*addr & (KVM_DEV_RISCV_IMSIC_ALIGN - 1))
207			return -EINVAL;
208	}
209
210	mutex_lock(&vcpu->mutex);
211	if (write)
212		vcpu_aia->imsic_addr = *addr;
213	else
214		*addr = vcpu_aia->imsic_addr;
215	mutex_unlock(&vcpu->mutex);
216
217	return 0;
218}
219
220static gpa_t aia_imsic_ppn(struct kvm_aia *aia, gpa_t addr)
221{
222	u32 h, l;
223	gpa_t mask = 0;
224
225	h = aia->nr_hart_bits + aia->nr_guest_bits +
226	    IMSIC_MMIO_PAGE_SHIFT - 1;
227	mask = GENMASK_ULL(h, 0);
228
229	if (aia->nr_group_bits) {
230		h = aia->nr_group_bits + aia->nr_group_shift - 1;
231		l = aia->nr_group_shift;
232		mask |= GENMASK_ULL(h, l);
233	}
234
235	return (addr & ~mask) >> IMSIC_MMIO_PAGE_SHIFT;
236}
237
238static u32 aia_imsic_hart_index(struct kvm_aia *aia, gpa_t addr)
239{
240	u32 hart, group = 0;
241
242	hart = (addr >> (aia->nr_guest_bits + IMSIC_MMIO_PAGE_SHIFT)) &
243		GENMASK_ULL(aia->nr_hart_bits - 1, 0);
244	if (aia->nr_group_bits)
245		group = (addr >> aia->nr_group_shift) &
246			GENMASK_ULL(aia->nr_group_bits - 1, 0);
247
248	return (group << aia->nr_hart_bits) | hart;
249}
250
251static int aia_init(struct kvm *kvm)
252{
253	int ret, i;
254	unsigned long idx;
255	struct kvm_vcpu *vcpu;
256	struct kvm_vcpu_aia *vaia;
257	struct kvm_aia *aia = &kvm->arch.aia;
258	gpa_t base_ppn = KVM_RISCV_AIA_UNDEF_ADDR;
259
260	/* Irqchip can be initialized only once */
261	if (kvm_riscv_aia_initialized(kvm))
262		return -EBUSY;
263
264	/* We might be in the middle of creating a VCPU? */
265	if (kvm->created_vcpus != atomic_read(&kvm->online_vcpus))
266		return -EBUSY;
267
268	/* Number of sources should be less than or equals number of IDs */
269	if (aia->nr_ids < aia->nr_sources)
270		return -EINVAL;
271
272	/* APLIC base is required for non-zero number of sources */
273	if (aia->nr_sources && aia->aplic_addr == KVM_RISCV_AIA_UNDEF_ADDR)
274		return -EINVAL;
275
276	/* Initialize APLIC */
277	ret = kvm_riscv_aia_aplic_init(kvm);
278	if (ret)
279		return ret;
280
281	/* Iterate over each VCPU */
282	kvm_for_each_vcpu(idx, vcpu, kvm) {
283		vaia = &vcpu->arch.aia_context;
284
285		/* IMSIC base is required */
286		if (vaia->imsic_addr == KVM_RISCV_AIA_UNDEF_ADDR) {
287			ret = -EINVAL;
288			goto fail_cleanup_imsics;
289		}
290
291		/* All IMSICs should have matching base PPN */
292		if (base_ppn == KVM_RISCV_AIA_UNDEF_ADDR)
293			base_ppn = aia_imsic_ppn(aia, vaia->imsic_addr);
294		if (base_ppn != aia_imsic_ppn(aia, vaia->imsic_addr)) {
295			ret = -EINVAL;
296			goto fail_cleanup_imsics;
297		}
298
299		/* Update HART index of the IMSIC based on IMSIC base */
300		vaia->hart_index = aia_imsic_hart_index(aia,
301							vaia->imsic_addr);
302
303		/* Initialize IMSIC for this VCPU */
304		ret = kvm_riscv_vcpu_aia_imsic_init(vcpu);
305		if (ret)
306			goto fail_cleanup_imsics;
307	}
308
309	/* Set the initialized flag */
310	kvm->arch.aia.initialized = true;
311
312	return 0;
313
314fail_cleanup_imsics:
315	for (i = idx - 1; i >= 0; i--) {
316		vcpu = kvm_get_vcpu(kvm, i);
317		if (!vcpu)
318			continue;
319		kvm_riscv_vcpu_aia_imsic_cleanup(vcpu);
320	}
321	kvm_riscv_aia_aplic_cleanup(kvm);
322	return ret;
323}
324
325static int aia_set_attr(struct kvm_device *dev, struct kvm_device_attr *attr)
326{
327	u32 nr;
328	u64 addr;
329	int nr_vcpus, r = -ENXIO;
330	unsigned long v, type = (unsigned long)attr->attr;
331	void __user *uaddr = (void __user *)(long)attr->addr;
332
333	switch (attr->group) {
334	case KVM_DEV_RISCV_AIA_GRP_CONFIG:
335		if (copy_from_user(&nr, uaddr, sizeof(nr)))
336			return -EFAULT;
337
338		mutex_lock(&dev->kvm->lock);
339		r = aia_config(dev->kvm, type, &nr, true);
340		mutex_unlock(&dev->kvm->lock);
341
342		break;
343
344	case KVM_DEV_RISCV_AIA_GRP_ADDR:
345		if (copy_from_user(&addr, uaddr, sizeof(addr)))
346			return -EFAULT;
347
348		nr_vcpus = atomic_read(&dev->kvm->online_vcpus);
349		mutex_lock(&dev->kvm->lock);
350		if (type == KVM_DEV_RISCV_AIA_ADDR_APLIC)
351			r = aia_aplic_addr(dev->kvm, &addr, true);
352		else if (type < KVM_DEV_RISCV_AIA_ADDR_IMSIC(nr_vcpus))
353			r = aia_imsic_addr(dev->kvm, &addr,
354			    type - KVM_DEV_RISCV_AIA_ADDR_IMSIC(0), true);
355		mutex_unlock(&dev->kvm->lock);
356
357		break;
358
359	case KVM_DEV_RISCV_AIA_GRP_CTRL:
360		switch (type) {
361		case KVM_DEV_RISCV_AIA_CTRL_INIT:
362			mutex_lock(&dev->kvm->lock);
363			r = aia_init(dev->kvm);
364			mutex_unlock(&dev->kvm->lock);
365			break;
366		}
367
368		break;
369	case KVM_DEV_RISCV_AIA_GRP_APLIC:
370		if (copy_from_user(&nr, uaddr, sizeof(nr)))
371			return -EFAULT;
372
373		mutex_lock(&dev->kvm->lock);
374		r = kvm_riscv_aia_aplic_set_attr(dev->kvm, type, nr);
375		mutex_unlock(&dev->kvm->lock);
376
377		break;
378	case KVM_DEV_RISCV_AIA_GRP_IMSIC:
379		if (copy_from_user(&v, uaddr, sizeof(v)))
380			return -EFAULT;
381
382		mutex_lock(&dev->kvm->lock);
383		r = kvm_riscv_aia_imsic_rw_attr(dev->kvm, type, true, &v);
384		mutex_unlock(&dev->kvm->lock);
385
386		break;
387	}
388
389	return r;
390}
391
392static int aia_get_attr(struct kvm_device *dev, struct kvm_device_attr *attr)
393{
394	u32 nr;
395	u64 addr;
396	int nr_vcpus, r = -ENXIO;
397	void __user *uaddr = (void __user *)(long)attr->addr;
398	unsigned long v, type = (unsigned long)attr->attr;
399
400	switch (attr->group) {
401	case KVM_DEV_RISCV_AIA_GRP_CONFIG:
402		if (copy_from_user(&nr, uaddr, sizeof(nr)))
403			return -EFAULT;
404
405		mutex_lock(&dev->kvm->lock);
406		r = aia_config(dev->kvm, type, &nr, false);
407		mutex_unlock(&dev->kvm->lock);
408		if (r)
409			return r;
410
411		if (copy_to_user(uaddr, &nr, sizeof(nr)))
412			return -EFAULT;
413
414		break;
415	case KVM_DEV_RISCV_AIA_GRP_ADDR:
416		if (copy_from_user(&addr, uaddr, sizeof(addr)))
417			return -EFAULT;
418
419		nr_vcpus = atomic_read(&dev->kvm->online_vcpus);
420		mutex_lock(&dev->kvm->lock);
421		if (type == KVM_DEV_RISCV_AIA_ADDR_APLIC)
422			r = aia_aplic_addr(dev->kvm, &addr, false);
423		else if (type < KVM_DEV_RISCV_AIA_ADDR_IMSIC(nr_vcpus))
424			r = aia_imsic_addr(dev->kvm, &addr,
425			    type - KVM_DEV_RISCV_AIA_ADDR_IMSIC(0), false);
426		mutex_unlock(&dev->kvm->lock);
427		if (r)
428			return r;
429
430		if (copy_to_user(uaddr, &addr, sizeof(addr)))
431			return -EFAULT;
432
433		break;
434	case KVM_DEV_RISCV_AIA_GRP_APLIC:
435		if (copy_from_user(&nr, uaddr, sizeof(nr)))
436			return -EFAULT;
437
438		mutex_lock(&dev->kvm->lock);
439		r = kvm_riscv_aia_aplic_get_attr(dev->kvm, type, &nr);
440		mutex_unlock(&dev->kvm->lock);
441		if (r)
442			return r;
443
444		if (copy_to_user(uaddr, &nr, sizeof(nr)))
445			return -EFAULT;
446
447		break;
448	case KVM_DEV_RISCV_AIA_GRP_IMSIC:
449		if (copy_from_user(&v, uaddr, sizeof(v)))
450			return -EFAULT;
451
452		mutex_lock(&dev->kvm->lock);
453		r = kvm_riscv_aia_imsic_rw_attr(dev->kvm, type, false, &v);
454		mutex_unlock(&dev->kvm->lock);
455		if (r)
456			return r;
457
458		if (copy_to_user(uaddr, &v, sizeof(v)))
459			return -EFAULT;
460
461		break;
462	}
463
464	return r;
465}
466
467static int aia_has_attr(struct kvm_device *dev, struct kvm_device_attr *attr)
468{
469	int nr_vcpus;
470
471	switch (attr->group) {
472	case KVM_DEV_RISCV_AIA_GRP_CONFIG:
473		switch (attr->attr) {
474		case KVM_DEV_RISCV_AIA_CONFIG_MODE:
475		case KVM_DEV_RISCV_AIA_CONFIG_IDS:
476		case KVM_DEV_RISCV_AIA_CONFIG_SRCS:
477		case KVM_DEV_RISCV_AIA_CONFIG_GROUP_BITS:
478		case KVM_DEV_RISCV_AIA_CONFIG_GROUP_SHIFT:
479		case KVM_DEV_RISCV_AIA_CONFIG_HART_BITS:
480		case KVM_DEV_RISCV_AIA_CONFIG_GUEST_BITS:
481			return 0;
482		}
483		break;
484	case KVM_DEV_RISCV_AIA_GRP_ADDR:
485		nr_vcpus = atomic_read(&dev->kvm->online_vcpus);
486		if (attr->attr == KVM_DEV_RISCV_AIA_ADDR_APLIC)
487			return 0;
488		else if (attr->attr < KVM_DEV_RISCV_AIA_ADDR_IMSIC(nr_vcpus))
489			return 0;
490		break;
491	case KVM_DEV_RISCV_AIA_GRP_CTRL:
492		switch (attr->attr) {
493		case KVM_DEV_RISCV_AIA_CTRL_INIT:
494			return 0;
495		}
496		break;
497	case KVM_DEV_RISCV_AIA_GRP_APLIC:
498		return kvm_riscv_aia_aplic_has_attr(dev->kvm, attr->attr);
499	case KVM_DEV_RISCV_AIA_GRP_IMSIC:
500		return kvm_riscv_aia_imsic_has_attr(dev->kvm, attr->attr);
501	}
502
503	return -ENXIO;
504}
505
506struct kvm_device_ops kvm_riscv_aia_device_ops = {
507	.name = "kvm-riscv-aia",
508	.create = aia_create,
509	.destroy = aia_destroy,
510	.set_attr = aia_set_attr,
511	.get_attr = aia_get_attr,
512	.has_attr = aia_has_attr,
513};
514
515int kvm_riscv_vcpu_aia_update(struct kvm_vcpu *vcpu)
516{
517	/* Proceed only if AIA was initialized successfully */
518	if (!kvm_riscv_aia_initialized(vcpu->kvm))
519		return 1;
520
521	/* Update the IMSIC HW state before entering guest mode */
522	return kvm_riscv_vcpu_aia_imsic_update(vcpu);
523}
524
525void kvm_riscv_vcpu_aia_reset(struct kvm_vcpu *vcpu)
526{
527	struct kvm_vcpu_aia_csr *csr = &vcpu->arch.aia_context.guest_csr;
528	struct kvm_vcpu_aia_csr *reset_csr =
529				&vcpu->arch.aia_context.guest_reset_csr;
530
531	if (!kvm_riscv_aia_available())
532		return;
533	memcpy(csr, reset_csr, sizeof(*csr));
534
535	/* Proceed only if AIA was initialized successfully */
536	if (!kvm_riscv_aia_initialized(vcpu->kvm))
537		return;
538
539	/* Reset the IMSIC context */
540	kvm_riscv_vcpu_aia_imsic_reset(vcpu);
541}
542
543int kvm_riscv_vcpu_aia_init(struct kvm_vcpu *vcpu)
544{
545	struct kvm_vcpu_aia *vaia = &vcpu->arch.aia_context;
546
547	if (!kvm_riscv_aia_available())
548		return 0;
549
550	/*
551	 * We don't do any memory allocations over here because these
552	 * will be done after AIA device is initialized by the user-space.
553	 *
554	 * Refer, aia_init() implementation for more details.
555	 */
556
557	/* Initialize default values in AIA vcpu context */
558	vaia->imsic_addr = KVM_RISCV_AIA_UNDEF_ADDR;
559	vaia->hart_index = vcpu->vcpu_idx;
560
561	return 0;
562}
563
564void kvm_riscv_vcpu_aia_deinit(struct kvm_vcpu *vcpu)
565{
566	/* Proceed only if AIA was initialized successfully */
567	if (!kvm_riscv_aia_initialized(vcpu->kvm))
568		return;
569
570	/* Cleanup IMSIC context */
571	kvm_riscv_vcpu_aia_imsic_cleanup(vcpu);
572}
573
574int kvm_riscv_aia_inject_msi_by_id(struct kvm *kvm, u32 hart_index,
575				   u32 guest_index, u32 iid)
576{
577	unsigned long idx;
578	struct kvm_vcpu *vcpu;
579
580	/* Proceed only if AIA was initialized successfully */
581	if (!kvm_riscv_aia_initialized(kvm))
582		return -EBUSY;
583
584	/* Inject MSI to matching VCPU */
585	kvm_for_each_vcpu(idx, vcpu, kvm) {
586		if (vcpu->arch.aia_context.hart_index == hart_index)
587			return kvm_riscv_vcpu_aia_imsic_inject(vcpu,
588							       guest_index,
589							       0, iid);
590	}
591
592	return 0;
593}
594
595int kvm_riscv_aia_inject_msi(struct kvm *kvm, struct kvm_msi *msi)
596{
597	gpa_t tppn, ippn;
598	unsigned long idx;
599	struct kvm_vcpu *vcpu;
600	u32 g, toff, iid = msi->data;
601	struct kvm_aia *aia = &kvm->arch.aia;
602	gpa_t target = (((gpa_t)msi->address_hi) << 32) | msi->address_lo;
603
604	/* Proceed only if AIA was initialized successfully */
605	if (!kvm_riscv_aia_initialized(kvm))
606		return -EBUSY;
607
608	/* Convert target address to target PPN */
609	tppn = target >> IMSIC_MMIO_PAGE_SHIFT;
610
611	/* Extract and clear Guest ID from target PPN */
612	g = tppn & (BIT(aia->nr_guest_bits) - 1);
613	tppn &= ~((gpa_t)(BIT(aia->nr_guest_bits) - 1));
614
615	/* Inject MSI to matching VCPU */
616	kvm_for_each_vcpu(idx, vcpu, kvm) {
617		ippn = vcpu->arch.aia_context.imsic_addr >>
618					IMSIC_MMIO_PAGE_SHIFT;
619		if (ippn == tppn) {
620			toff = target & (IMSIC_MMIO_PAGE_SZ - 1);
621			return kvm_riscv_vcpu_aia_imsic_inject(vcpu, g,
622							       toff, iid);
623		}
624	}
625
626	return 0;
627}
628
629int kvm_riscv_aia_inject_irq(struct kvm *kvm, unsigned int irq, bool level)
630{
631	/* Proceed only if AIA was initialized successfully */
632	if (!kvm_riscv_aia_initialized(kvm))
633		return -EBUSY;
634
635	/* Inject interrupt level change in APLIC */
636	return kvm_riscv_aia_aplic_inject(kvm, irq, level);
637}
638
639void kvm_riscv_aia_init_vm(struct kvm *kvm)
640{
641	struct kvm_aia *aia = &kvm->arch.aia;
642
643	if (!kvm_riscv_aia_available())
644		return;
645
646	/*
647	 * We don't do any memory allocations over here because these
648	 * will be done after AIA device is initialized by the user-space.
649	 *
650	 * Refer, aia_init() implementation for more details.
651	 */
652
653	/* Initialize default values in AIA global context */
654	aia->mode = (kvm_riscv_aia_nr_hgei) ?
655		KVM_DEV_RISCV_AIA_MODE_AUTO : KVM_DEV_RISCV_AIA_MODE_EMUL;
656	aia->nr_ids = kvm_riscv_aia_max_ids - 1;
657	aia->nr_sources = 0;
658	aia->nr_group_bits = 0;
659	aia->nr_group_shift = KVM_DEV_RISCV_AIA_GROUP_SHIFT_MIN;
660	aia->nr_hart_bits = 0;
661	aia->nr_guest_bits = 0;
662	aia->aplic_addr = KVM_RISCV_AIA_UNDEF_ADDR;
663}
664
665void kvm_riscv_aia_destroy_vm(struct kvm *kvm)
666{
667	/* Proceed only if AIA was initialized successfully */
668	if (!kvm_riscv_aia_initialized(kvm))
669		return;
670
671	/* Cleanup APLIC context */
672	kvm_riscv_aia_aplic_cleanup(kvm);
673}
674