1// SPDX-License-Identifier: GPL-2.0
2/*
3 * Copyright (C) 2021 Western Digital Corporation or its affiliates.
4 * Copyright (C) 2022 Ventana Micro Systems Inc.
5 *
6 * Authors:
7 *	Anup Patel <apatel@ventanamicro.com>
8 */
9
10#include <linux/kvm_host.h>
11#include <linux/math.h>
12#include <linux/spinlock.h>
13#include <linux/swab.h>
14#include <kvm/iodev.h>
15#include <asm/kvm_aia_aplic.h>
16
17struct aplic_irq {
18	raw_spinlock_t lock;
19	u32 sourcecfg;
20	u32 state;
21#define APLIC_IRQ_STATE_PENDING		BIT(0)
22#define APLIC_IRQ_STATE_ENABLED		BIT(1)
23#define APLIC_IRQ_STATE_ENPEND		(APLIC_IRQ_STATE_PENDING | \
24					 APLIC_IRQ_STATE_ENABLED)
25#define APLIC_IRQ_STATE_INPUT		BIT(8)
26	u32 target;
27};
28
29struct aplic {
30	struct kvm_io_device iodev;
31
32	u32 domaincfg;
33	u32 genmsi;
34
35	u32 nr_irqs;
36	u32 nr_words;
37	struct aplic_irq *irqs;
38};
39
40static u32 aplic_read_sourcecfg(struct aplic *aplic, u32 irq)
41{
42	u32 ret;
43	unsigned long flags;
44	struct aplic_irq *irqd;
45
46	if (!irq || aplic->nr_irqs <= irq)
47		return 0;
48	irqd = &aplic->irqs[irq];
49
50	raw_spin_lock_irqsave(&irqd->lock, flags);
51	ret = irqd->sourcecfg;
52	raw_spin_unlock_irqrestore(&irqd->lock, flags);
53
54	return ret;
55}
56
57static void aplic_write_sourcecfg(struct aplic *aplic, u32 irq, u32 val)
58{
59	unsigned long flags;
60	struct aplic_irq *irqd;
61
62	if (!irq || aplic->nr_irqs <= irq)
63		return;
64	irqd = &aplic->irqs[irq];
65
66	if (val & APLIC_SOURCECFG_D)
67		val = 0;
68	else
69		val &= APLIC_SOURCECFG_SM_MASK;
70
71	raw_spin_lock_irqsave(&irqd->lock, flags);
72	irqd->sourcecfg = val;
73	raw_spin_unlock_irqrestore(&irqd->lock, flags);
74}
75
76static u32 aplic_read_target(struct aplic *aplic, u32 irq)
77{
78	u32 ret;
79	unsigned long flags;
80	struct aplic_irq *irqd;
81
82	if (!irq || aplic->nr_irqs <= irq)
83		return 0;
84	irqd = &aplic->irqs[irq];
85
86	raw_spin_lock_irqsave(&irqd->lock, flags);
87	ret = irqd->target;
88	raw_spin_unlock_irqrestore(&irqd->lock, flags);
89
90	return ret;
91}
92
93static void aplic_write_target(struct aplic *aplic, u32 irq, u32 val)
94{
95	unsigned long flags;
96	struct aplic_irq *irqd;
97
98	if (!irq || aplic->nr_irqs <= irq)
99		return;
100	irqd = &aplic->irqs[irq];
101
102	val &= APLIC_TARGET_EIID_MASK |
103	       (APLIC_TARGET_HART_IDX_MASK << APLIC_TARGET_HART_IDX_SHIFT) |
104	       (APLIC_TARGET_GUEST_IDX_MASK << APLIC_TARGET_GUEST_IDX_SHIFT);
105
106	raw_spin_lock_irqsave(&irqd->lock, flags);
107	irqd->target = val;
108	raw_spin_unlock_irqrestore(&irqd->lock, flags);
109}
110
111static bool aplic_read_pending(struct aplic *aplic, u32 irq)
112{
113	bool ret;
114	unsigned long flags;
115	struct aplic_irq *irqd;
116
117	if (!irq || aplic->nr_irqs <= irq)
118		return false;
119	irqd = &aplic->irqs[irq];
120
121	raw_spin_lock_irqsave(&irqd->lock, flags);
122	ret = (irqd->state & APLIC_IRQ_STATE_PENDING) ? true : false;
123	raw_spin_unlock_irqrestore(&irqd->lock, flags);
124
125	return ret;
126}
127
128static void aplic_write_pending(struct aplic *aplic, u32 irq, bool pending)
129{
130	unsigned long flags, sm;
131	struct aplic_irq *irqd;
132
133	if (!irq || aplic->nr_irqs <= irq)
134		return;
135	irqd = &aplic->irqs[irq];
136
137	raw_spin_lock_irqsave(&irqd->lock, flags);
138
139	sm = irqd->sourcecfg & APLIC_SOURCECFG_SM_MASK;
140	if (sm == APLIC_SOURCECFG_SM_INACTIVE)
141		goto skip_write_pending;
142
143	if (sm == APLIC_SOURCECFG_SM_LEVEL_HIGH ||
144	    sm == APLIC_SOURCECFG_SM_LEVEL_LOW) {
145		if (!pending)
146			goto skip_write_pending;
147		if ((irqd->state & APLIC_IRQ_STATE_INPUT) &&
148		    sm == APLIC_SOURCECFG_SM_LEVEL_LOW)
149			goto skip_write_pending;
150		if (!(irqd->state & APLIC_IRQ_STATE_INPUT) &&
151		    sm == APLIC_SOURCECFG_SM_LEVEL_HIGH)
152			goto skip_write_pending;
153	}
154
155	if (pending)
156		irqd->state |= APLIC_IRQ_STATE_PENDING;
157	else
158		irqd->state &= ~APLIC_IRQ_STATE_PENDING;
159
160skip_write_pending:
161	raw_spin_unlock_irqrestore(&irqd->lock, flags);
162}
163
164static bool aplic_read_enabled(struct aplic *aplic, u32 irq)
165{
166	bool ret;
167	unsigned long flags;
168	struct aplic_irq *irqd;
169
170	if (!irq || aplic->nr_irqs <= irq)
171		return false;
172	irqd = &aplic->irqs[irq];
173
174	raw_spin_lock_irqsave(&irqd->lock, flags);
175	ret = (irqd->state & APLIC_IRQ_STATE_ENABLED) ? true : false;
176	raw_spin_unlock_irqrestore(&irqd->lock, flags);
177
178	return ret;
179}
180
181static void aplic_write_enabled(struct aplic *aplic, u32 irq, bool enabled)
182{
183	unsigned long flags;
184	struct aplic_irq *irqd;
185
186	if (!irq || aplic->nr_irqs <= irq)
187		return;
188	irqd = &aplic->irqs[irq];
189
190	raw_spin_lock_irqsave(&irqd->lock, flags);
191	if (enabled)
192		irqd->state |= APLIC_IRQ_STATE_ENABLED;
193	else
194		irqd->state &= ~APLIC_IRQ_STATE_ENABLED;
195	raw_spin_unlock_irqrestore(&irqd->lock, flags);
196}
197
198static bool aplic_read_input(struct aplic *aplic, u32 irq)
199{
200	u32 sourcecfg, sm, raw_input, irq_inverted;
201	struct aplic_irq *irqd;
202	unsigned long flags;
203	bool ret = false;
204
205	if (!irq || aplic->nr_irqs <= irq)
206		return false;
207	irqd = &aplic->irqs[irq];
208
209	raw_spin_lock_irqsave(&irqd->lock, flags);
210
211	sourcecfg = irqd->sourcecfg;
212	if (sourcecfg & APLIC_SOURCECFG_D)
213		goto skip;
214
215	sm = sourcecfg & APLIC_SOURCECFG_SM_MASK;
216	if (sm == APLIC_SOURCECFG_SM_INACTIVE)
217		goto skip;
218
219	raw_input = (irqd->state & APLIC_IRQ_STATE_INPUT) ? 1 : 0;
220	irq_inverted = (sm == APLIC_SOURCECFG_SM_LEVEL_LOW ||
221			sm == APLIC_SOURCECFG_SM_EDGE_FALL) ? 1 : 0;
222	ret = !!(raw_input ^ irq_inverted);
223
224skip:
225	raw_spin_unlock_irqrestore(&irqd->lock, flags);
226
227	return ret;
228}
229
230static void aplic_inject_msi(struct kvm *kvm, u32 irq, u32 target)
231{
232	u32 hart_idx, guest_idx, eiid;
233
234	hart_idx = target >> APLIC_TARGET_HART_IDX_SHIFT;
235	hart_idx &= APLIC_TARGET_HART_IDX_MASK;
236	guest_idx = target >> APLIC_TARGET_GUEST_IDX_SHIFT;
237	guest_idx &= APLIC_TARGET_GUEST_IDX_MASK;
238	eiid = target & APLIC_TARGET_EIID_MASK;
239	kvm_riscv_aia_inject_msi_by_id(kvm, hart_idx, guest_idx, eiid);
240}
241
242static void aplic_update_irq_range(struct kvm *kvm, u32 first, u32 last)
243{
244	bool inject;
245	u32 irq, target;
246	unsigned long flags;
247	struct aplic_irq *irqd;
248	struct aplic *aplic = kvm->arch.aia.aplic_state;
249
250	if (!(aplic->domaincfg & APLIC_DOMAINCFG_IE))
251		return;
252
253	for (irq = first; irq <= last; irq++) {
254		if (!irq || aplic->nr_irqs <= irq)
255			continue;
256		irqd = &aplic->irqs[irq];
257
258		raw_spin_lock_irqsave(&irqd->lock, flags);
259
260		inject = false;
261		target = irqd->target;
262		if ((irqd->state & APLIC_IRQ_STATE_ENPEND) ==
263		    APLIC_IRQ_STATE_ENPEND) {
264			irqd->state &= ~APLIC_IRQ_STATE_PENDING;
265			inject = true;
266		}
267
268		raw_spin_unlock_irqrestore(&irqd->lock, flags);
269
270		if (inject)
271			aplic_inject_msi(kvm, irq, target);
272	}
273}
274
275int kvm_riscv_aia_aplic_inject(struct kvm *kvm, u32 source, bool level)
276{
277	u32 target;
278	bool inject = false, ie;
279	unsigned long flags;
280	struct aplic_irq *irqd;
281	struct aplic *aplic = kvm->arch.aia.aplic_state;
282
283	if (!aplic || !source || (aplic->nr_irqs <= source))
284		return -ENODEV;
285	irqd = &aplic->irqs[source];
286	ie = (aplic->domaincfg & APLIC_DOMAINCFG_IE) ? true : false;
287
288	raw_spin_lock_irqsave(&irqd->lock, flags);
289
290	if (irqd->sourcecfg & APLIC_SOURCECFG_D)
291		goto skip_unlock;
292
293	switch (irqd->sourcecfg & APLIC_SOURCECFG_SM_MASK) {
294	case APLIC_SOURCECFG_SM_EDGE_RISE:
295		if (level && !(irqd->state & APLIC_IRQ_STATE_INPUT) &&
296		    !(irqd->state & APLIC_IRQ_STATE_PENDING))
297			irqd->state |= APLIC_IRQ_STATE_PENDING;
298		break;
299	case APLIC_SOURCECFG_SM_EDGE_FALL:
300		if (!level && (irqd->state & APLIC_IRQ_STATE_INPUT) &&
301		    !(irqd->state & APLIC_IRQ_STATE_PENDING))
302			irqd->state |= APLIC_IRQ_STATE_PENDING;
303		break;
304	case APLIC_SOURCECFG_SM_LEVEL_HIGH:
305		if (level && !(irqd->state & APLIC_IRQ_STATE_PENDING))
306			irqd->state |= APLIC_IRQ_STATE_PENDING;
307		break;
308	case APLIC_SOURCECFG_SM_LEVEL_LOW:
309		if (!level && !(irqd->state & APLIC_IRQ_STATE_PENDING))
310			irqd->state |= APLIC_IRQ_STATE_PENDING;
311		break;
312	}
313
314	if (level)
315		irqd->state |= APLIC_IRQ_STATE_INPUT;
316	else
317		irqd->state &= ~APLIC_IRQ_STATE_INPUT;
318
319	target = irqd->target;
320	if (ie && ((irqd->state & APLIC_IRQ_STATE_ENPEND) ==
321		   APLIC_IRQ_STATE_ENPEND)) {
322		irqd->state &= ~APLIC_IRQ_STATE_PENDING;
323		inject = true;
324	}
325
326skip_unlock:
327	raw_spin_unlock_irqrestore(&irqd->lock, flags);
328
329	if (inject)
330		aplic_inject_msi(kvm, source, target);
331
332	return 0;
333}
334
335static u32 aplic_read_input_word(struct aplic *aplic, u32 word)
336{
337	u32 i, ret = 0;
338
339	for (i = 0; i < 32; i++)
340		ret |= aplic_read_input(aplic, word * 32 + i) ? BIT(i) : 0;
341
342	return ret;
343}
344
345static u32 aplic_read_pending_word(struct aplic *aplic, u32 word)
346{
347	u32 i, ret = 0;
348
349	for (i = 0; i < 32; i++)
350		ret |= aplic_read_pending(aplic, word * 32 + i) ? BIT(i) : 0;
351
352	return ret;
353}
354
355static void aplic_write_pending_word(struct aplic *aplic, u32 word,
356				     u32 val, bool pending)
357{
358	u32 i;
359
360	for (i = 0; i < 32; i++) {
361		if (val & BIT(i))
362			aplic_write_pending(aplic, word * 32 + i, pending);
363	}
364}
365
366static u32 aplic_read_enabled_word(struct aplic *aplic, u32 word)
367{
368	u32 i, ret = 0;
369
370	for (i = 0; i < 32; i++)
371		ret |= aplic_read_enabled(aplic, word * 32 + i) ? BIT(i) : 0;
372
373	return ret;
374}
375
376static void aplic_write_enabled_word(struct aplic *aplic, u32 word,
377				     u32 val, bool enabled)
378{
379	u32 i;
380
381	for (i = 0; i < 32; i++) {
382		if (val & BIT(i))
383			aplic_write_enabled(aplic, word * 32 + i, enabled);
384	}
385}
386
387static int aplic_mmio_read_offset(struct kvm *kvm, gpa_t off, u32 *val32)
388{
389	u32 i;
390	struct aplic *aplic = kvm->arch.aia.aplic_state;
391
392	if ((off & 0x3) != 0)
393		return -EOPNOTSUPP;
394
395	if (off == APLIC_DOMAINCFG) {
396		*val32 = APLIC_DOMAINCFG_RDONLY |
397			 aplic->domaincfg | APLIC_DOMAINCFG_DM;
398	} else if ((off >= APLIC_SOURCECFG_BASE) &&
399		 (off < (APLIC_SOURCECFG_BASE + (aplic->nr_irqs - 1) * 4))) {
400		i = ((off - APLIC_SOURCECFG_BASE) >> 2) + 1;
401		*val32 = aplic_read_sourcecfg(aplic, i);
402	} else if ((off >= APLIC_SETIP_BASE) &&
403		   (off < (APLIC_SETIP_BASE + aplic->nr_words * 4))) {
404		i = (off - APLIC_SETIP_BASE) >> 2;
405		*val32 = aplic_read_pending_word(aplic, i);
406	} else if (off == APLIC_SETIPNUM) {
407		*val32 = 0;
408	} else if ((off >= APLIC_CLRIP_BASE) &&
409		   (off < (APLIC_CLRIP_BASE + aplic->nr_words * 4))) {
410		i = (off - APLIC_CLRIP_BASE) >> 2;
411		*val32 = aplic_read_input_word(aplic, i);
412	} else if (off == APLIC_CLRIPNUM) {
413		*val32 = 0;
414	} else if ((off >= APLIC_SETIE_BASE) &&
415		   (off < (APLIC_SETIE_BASE + aplic->nr_words * 4))) {
416		i = (off - APLIC_SETIE_BASE) >> 2;
417		*val32 = aplic_read_enabled_word(aplic, i);
418	} else if (off == APLIC_SETIENUM) {
419		*val32 = 0;
420	} else if ((off >= APLIC_CLRIE_BASE) &&
421		   (off < (APLIC_CLRIE_BASE + aplic->nr_words * 4))) {
422		*val32 = 0;
423	} else if (off == APLIC_CLRIENUM) {
424		*val32 = 0;
425	} else if (off == APLIC_SETIPNUM_LE) {
426		*val32 = 0;
427	} else if (off == APLIC_SETIPNUM_BE) {
428		*val32 = 0;
429	} else if (off == APLIC_GENMSI) {
430		*val32 = aplic->genmsi;
431	} else if ((off >= APLIC_TARGET_BASE) &&
432		   (off < (APLIC_TARGET_BASE + (aplic->nr_irqs - 1) * 4))) {
433		i = ((off - APLIC_TARGET_BASE) >> 2) + 1;
434		*val32 = aplic_read_target(aplic, i);
435	} else
436		return -ENODEV;
437
438	return 0;
439}
440
441static int aplic_mmio_read(struct kvm_vcpu *vcpu, struct kvm_io_device *dev,
442			   gpa_t addr, int len, void *val)
443{
444	if (len != 4)
445		return -EOPNOTSUPP;
446
447	return aplic_mmio_read_offset(vcpu->kvm,
448				      addr - vcpu->kvm->arch.aia.aplic_addr,
449				      val);
450}
451
452static int aplic_mmio_write_offset(struct kvm *kvm, gpa_t off, u32 val32)
453{
454	u32 i;
455	struct aplic *aplic = kvm->arch.aia.aplic_state;
456
457	if ((off & 0x3) != 0)
458		return -EOPNOTSUPP;
459
460	if (off == APLIC_DOMAINCFG) {
461		/* Only IE bit writeable */
462		aplic->domaincfg = val32 & APLIC_DOMAINCFG_IE;
463	} else if ((off >= APLIC_SOURCECFG_BASE) &&
464		 (off < (APLIC_SOURCECFG_BASE + (aplic->nr_irqs - 1) * 4))) {
465		i = ((off - APLIC_SOURCECFG_BASE) >> 2) + 1;
466		aplic_write_sourcecfg(aplic, i, val32);
467	} else if ((off >= APLIC_SETIP_BASE) &&
468		   (off < (APLIC_SETIP_BASE + aplic->nr_words * 4))) {
469		i = (off - APLIC_SETIP_BASE) >> 2;
470		aplic_write_pending_word(aplic, i, val32, true);
471	} else if (off == APLIC_SETIPNUM) {
472		aplic_write_pending(aplic, val32, true);
473	} else if ((off >= APLIC_CLRIP_BASE) &&
474		   (off < (APLIC_CLRIP_BASE + aplic->nr_words * 4))) {
475		i = (off - APLIC_CLRIP_BASE) >> 2;
476		aplic_write_pending_word(aplic, i, val32, false);
477	} else if (off == APLIC_CLRIPNUM) {
478		aplic_write_pending(aplic, val32, false);
479	} else if ((off >= APLIC_SETIE_BASE) &&
480		   (off < (APLIC_SETIE_BASE + aplic->nr_words * 4))) {
481		i = (off - APLIC_SETIE_BASE) >> 2;
482		aplic_write_enabled_word(aplic, i, val32, true);
483	} else if (off == APLIC_SETIENUM) {
484		aplic_write_enabled(aplic, val32, true);
485	} else if ((off >= APLIC_CLRIE_BASE) &&
486		   (off < (APLIC_CLRIE_BASE + aplic->nr_words * 4))) {
487		i = (off - APLIC_CLRIE_BASE) >> 2;
488		aplic_write_enabled_word(aplic, i, val32, false);
489	} else if (off == APLIC_CLRIENUM) {
490		aplic_write_enabled(aplic, val32, false);
491	} else if (off == APLIC_SETIPNUM_LE) {
492		aplic_write_pending(aplic, val32, true);
493	} else if (off == APLIC_SETIPNUM_BE) {
494		aplic_write_pending(aplic, __swab32(val32), true);
495	} else if (off == APLIC_GENMSI) {
496		aplic->genmsi = val32 & ~(APLIC_TARGET_GUEST_IDX_MASK <<
497					  APLIC_TARGET_GUEST_IDX_SHIFT);
498		kvm_riscv_aia_inject_msi_by_id(kvm,
499				val32 >> APLIC_TARGET_HART_IDX_SHIFT, 0,
500				val32 & APLIC_TARGET_EIID_MASK);
501	} else if ((off >= APLIC_TARGET_BASE) &&
502		   (off < (APLIC_TARGET_BASE + (aplic->nr_irqs - 1) * 4))) {
503		i = ((off - APLIC_TARGET_BASE) >> 2) + 1;
504		aplic_write_target(aplic, i, val32);
505	} else
506		return -ENODEV;
507
508	aplic_update_irq_range(kvm, 1, aplic->nr_irqs - 1);
509
510	return 0;
511}
512
513static int aplic_mmio_write(struct kvm_vcpu *vcpu, struct kvm_io_device *dev,
514			    gpa_t addr, int len, const void *val)
515{
516	if (len != 4)
517		return -EOPNOTSUPP;
518
519	return aplic_mmio_write_offset(vcpu->kvm,
520				       addr - vcpu->kvm->arch.aia.aplic_addr,
521				       *((const u32 *)val));
522}
523
524static struct kvm_io_device_ops aplic_iodoev_ops = {
525	.read = aplic_mmio_read,
526	.write = aplic_mmio_write,
527};
528
529int kvm_riscv_aia_aplic_set_attr(struct kvm *kvm, unsigned long type, u32 v)
530{
531	int rc;
532
533	if (!kvm->arch.aia.aplic_state)
534		return -ENODEV;
535
536	rc = aplic_mmio_write_offset(kvm, type, v);
537	if (rc)
538		return rc;
539
540	return 0;
541}
542
543int kvm_riscv_aia_aplic_get_attr(struct kvm *kvm, unsigned long type, u32 *v)
544{
545	int rc;
546
547	if (!kvm->arch.aia.aplic_state)
548		return -ENODEV;
549
550	rc = aplic_mmio_read_offset(kvm, type, v);
551	if (rc)
552		return rc;
553
554	return 0;
555}
556
557int kvm_riscv_aia_aplic_has_attr(struct kvm *kvm, unsigned long type)
558{
559	int rc;
560	u32 val;
561
562	if (!kvm->arch.aia.aplic_state)
563		return -ENODEV;
564
565	rc = aplic_mmio_read_offset(kvm, type, &val);
566	if (rc)
567		return rc;
568
569	return 0;
570}
571
572int kvm_riscv_aia_aplic_init(struct kvm *kvm)
573{
574	int i, ret = 0;
575	struct aplic *aplic;
576
577	/* Do nothing if we have zero sources */
578	if (!kvm->arch.aia.nr_sources)
579		return 0;
580
581	/* Allocate APLIC global state */
582	aplic = kzalloc(sizeof(*aplic), GFP_KERNEL);
583	if (!aplic)
584		return -ENOMEM;
585	kvm->arch.aia.aplic_state = aplic;
586
587	/* Setup APLIC IRQs */
588	aplic->nr_irqs = kvm->arch.aia.nr_sources + 1;
589	aplic->nr_words = DIV_ROUND_UP(aplic->nr_irqs, 32);
590	aplic->irqs = kcalloc(aplic->nr_irqs,
591			      sizeof(*aplic->irqs), GFP_KERNEL);
592	if (!aplic->irqs) {
593		ret = -ENOMEM;
594		goto fail_free_aplic;
595	}
596	for (i = 0; i < aplic->nr_irqs; i++)
597		raw_spin_lock_init(&aplic->irqs[i].lock);
598
599	/* Setup IO device */
600	kvm_iodevice_init(&aplic->iodev, &aplic_iodoev_ops);
601	mutex_lock(&kvm->slots_lock);
602	ret = kvm_io_bus_register_dev(kvm, KVM_MMIO_BUS,
603				      kvm->arch.aia.aplic_addr,
604				      KVM_DEV_RISCV_APLIC_SIZE,
605				      &aplic->iodev);
606	mutex_unlock(&kvm->slots_lock);
607	if (ret)
608		goto fail_free_aplic_irqs;
609
610	/* Setup default IRQ routing */
611	ret = kvm_riscv_setup_default_irq_routing(kvm, aplic->nr_irqs);
612	if (ret)
613		goto fail_unreg_iodev;
614
615	return 0;
616
617fail_unreg_iodev:
618	mutex_lock(&kvm->slots_lock);
619	kvm_io_bus_unregister_dev(kvm, KVM_MMIO_BUS, &aplic->iodev);
620	mutex_unlock(&kvm->slots_lock);
621fail_free_aplic_irqs:
622	kfree(aplic->irqs);
623fail_free_aplic:
624	kvm->arch.aia.aplic_state = NULL;
625	kfree(aplic);
626	return ret;
627}
628
629void kvm_riscv_aia_aplic_cleanup(struct kvm *kvm)
630{
631	struct aplic *aplic = kvm->arch.aia.aplic_state;
632
633	if (!aplic)
634		return;
635
636	mutex_lock(&kvm->slots_lock);
637	kvm_io_bus_unregister_dev(kvm, KVM_MMIO_BUS, &aplic->iodev);
638	mutex_unlock(&kvm->slots_lock);
639
640	kfree(aplic->irqs);
641
642	kvm->arch.aia.aplic_state = NULL;
643	kfree(aplic);
644}
645