Lines Matching refs:r9

537 	ld	r9, VCORE_KVM(r5)	/* pointer to struct kvm */
560 lwz r7,KVM_LPID(r9)
561 ld r6,KVM_SDR1(r9)
570 mr r3, r9 /* kvm pointer */
849 ld r9,VCPU_SLB_V(r6)
850 slbmte r9,r8
895 li r9, KVM_GUEST_MODE_GUEST_HV
896 stb r9, HSTATE_IN_GUEST(r13)
920 ld r9, VCPU_GPR(R9)(r4)
963 mr r9, r4
991 cmpwi r9,KVM_GUEST_MODE_HOST_HV
993 li r9, KVM_GUEST_MODE_HOST_HV
994 stb r9, HSTATE_IN_GUEST(r13)
996 ld r9, HSTATE_KVM_VCPU(r13)
1000 std r0, VCPU_GPR(R0)(r9)
1001 std r1, VCPU_GPR(R1)(r9)
1002 std r2, VCPU_GPR(R2)(r9)
1003 std r3, VCPU_GPR(R3)(r9)
1004 std r4, VCPU_GPR(R4)(r9)
1005 std r5, VCPU_GPR(R5)(r9)
1006 std r6, VCPU_GPR(R6)(r9)
1007 std r7, VCPU_GPR(R7)(r9)
1008 std r8, VCPU_GPR(R8)(r9)
1010 std r0, VCPU_GPR(R9)(r9)
1011 std r10, VCPU_GPR(R10)(r9)
1012 std r11, VCPU_GPR(R11)(r9)
1014 std r3, VCPU_GPR(R12)(r9)
1017 std r4, VCPU_CR(r9)
1020 std r3, VCPU_CFAR(r9)
1024 std r4, VCPU_PPR(r9)
1033 std r10, VCPU_SRR0(r9)
1034 std r11, VCPU_SRR1(r9)
1042 1: std r10, VCPU_PC(r9)
1043 std r11, VCPU_MSR(r9)
1047 std r3, VCPU_GPR(R13)(r9)
1048 std r4, VCPU_LR(r9)
1050 stw r12,VCPU_TRAP(r9)
1062 addi r3, r9, VCPU_TB_RMINTR
1063 mr r4, r9
1065 ld r5, VCPU_GPR(R5)(r9)
1066 ld r6, VCPU_GPR(R6)(r9)
1067 ld r7, VCPU_GPR(R7)(r9)
1068 ld r8, VCPU_GPR(R8)(r9)
1074 std r3,VCPU_LAST_INST(r9)
1078 11: std r3,VCPU_HEIR(r9)
1083 std r3, VCPU_CTR(r9)
1084 std r4, VCPU_XER(r9)
1089 std r3, VCPU_DAR(r9)
1090 stw r4, VCPU_DSISR(r9)
1095 std r3, VCPU_FAULT_DAR(r9)
1096 stw r4, VCPU_FAULT_DSISR(r9)
1106 mr r4,r9
1125 std r3, VCPU_HFSCR(r9)
1138 guest_exit_cont: /* r9 = vcpu, r12 = trap, r13 = paca */
1141 addi r3, r9, VCPU_TB_RMEXIT
1142 mr r4, r9
1155 lwz r0,VCPU_SLB_NR(r9) /* number of entries in SLB */
1158 addi r7,r9,VCPU_SLB
1175 stw r5,VCPU_SLB_MAX(r9)
1201 std r5,VCPU_DEC_EXPIRES(r9)
1207 ld r9, HSTATE_KVM_VCPU(r13)
1211 stw r0, VCPU_CPU(r9)
1212 stw r0, VCPU_THREAD_CPU(r9)
1216 stw r6,VCPU_CTRL(r9)
1227 ld r7,VCPU_PURR(r9)
1228 ld r8,VCPU_SPURR(r9)
1229 std r5,VCPU_PURR(r9)
1230 std r6,VCPU_SPURR(r9)
1252 std r5, VCPU_IAMR(r9)
1253 stw r6, VCPU_PSPB(r9)
1254 std r7, VCPU_FSCR(r9)
1257 std r5, VCPU_IC(r9)
1258 std r7, VCPU_TAR(r9)
1260 std r8, VCPU_EBBHR(r9)
1265 std r5, VCPU_EBBRR(r9)
1266 std r6, VCPU_BESCR(r9)
1267 stw r7, VCPU_GUEST_PID(r9)
1268 std r8, VCPU_WORT(r9)
1273 std r5, VCPU_TCSCR(r9)
1274 std r6, VCPU_ACOP(r9)
1275 std r7, VCPU_CSIGR(r9)
1276 std r8, VCPU_TACR(r9)
1301 std r5,VCPU_AMR(r9)
1302 std r6,VCPU_UAMOR(r9)
1311 std r8, VCPU_DSCR(r9)
1315 std r14, VCPU_GPR(R14)(r9)
1316 std r15, VCPU_GPR(R15)(r9)
1317 std r16, VCPU_GPR(R16)(r9)
1318 std r17, VCPU_GPR(R17)(r9)
1319 std r18, VCPU_GPR(R18)(r9)
1320 std r19, VCPU_GPR(R19)(r9)
1321 std r20, VCPU_GPR(R20)(r9)
1322 std r21, VCPU_GPR(R21)(r9)
1323 std r22, VCPU_GPR(R22)(r9)
1324 std r23, VCPU_GPR(R23)(r9)
1325 std r24, VCPU_GPR(R24)(r9)
1326 std r25, VCPU_GPR(R25)(r9)
1327 std r26, VCPU_GPR(R26)(r9)
1328 std r27, VCPU_GPR(R27)(r9)
1329 std r28, VCPU_GPR(R28)(r9)
1330 std r29, VCPU_GPR(R29)(r9)
1331 std r30, VCPU_GPR(R30)(r9)
1332 std r31, VCPU_GPR(R31)(r9)
1339 std r3, VCPU_SPRG0(r9)
1340 std r4, VCPU_SPRG1(r9)
1341 std r5, VCPU_SPRG2(r9)
1342 std r6, VCPU_SPRG3(r9)
1345 mr r3, r9
1355 mr r3, r9
1360 ld r9, HSTATE_KVM_VCPU(r13)
1365 ld r8, VCPU_VPA(r9) /* do they have a VPA? */
1373 stb r3, VCPU_VPA_DIRTY(r9)
1377 mr r3, r9
1382 ld r9, HSTATE_KVM_VCPU(r13)
1553 ld r9, HSTATE_KVM_VCPU(r13)
1588 stw r12, VCPU_TRAP(r9)
1600 mr r4, r9
1623 4: std r4, VCPU_FAULT_DAR(r9)
1624 stw r6, VCPU_FAULT_DSISR(r9)
1627 mr r3, r9 /* vcpu pointer */
1630 ld r9, HSTATE_KVM_VCPU(r13)
1631 ld r10, VCPU_PC(r9)
1632 ld r11, VCPU_MSR(r9)
1642 ld r4, VCPU_FAULT_DAR(r9)
1652 6: ld r7, VCPU_CTR(r9)
1653 ld r8, VCPU_XER(r9)
1656 mr r4, r9
1659 3: ld r5, VCPU_KVM(r9) /* not relocated, use VRMA */
1679 std r8, VCPU_LAST_INST(r9)
1701 mr r3, r9 /* vcpu pointer */
1706 ld r9, HSTATE_KVM_VCPU(r13)
1707 ld r10, VCPU_PC(r9)
1708 ld r11, VCPU_MSR(r9)
1724 3: ld r6, VCPU_KVM(r9) /* not relocated, use VRMA */
1735 * r9 = vcpu, r10 = pc, r11 = msr, r12 = trap, r13 = paca
1738 ld r3,VCPU_GPR(R3)(r9)
1746 ld r4, VCPU_KVM(r9)
1762 mr r3,r9 /* get vcpu pointer */
1763 ld r4,VCPU_GPR(R4)(r9)
1778 mr r4,r9
1786 ld r9, HSTATE_KVM_VCPU(r13)
2306 mr r9, r4
2322 ld r9, HSTATE_KVM_VCPU(r13)
2327 mr r3, r9 /* get vcpu pointer */
2331 ld r9, HSTATE_KVM_VCPU(r13)
2339 * r9 points to the vcpu on entry
2346 ld r9, HSTATE_KVM_VCPU(r13)
2481 mfmsr r9
2482 ori r8,r9,MSR_FP
2604 ld r9, HSTATE_KVM_VCPU(r13)
2607 std r5, VCPU_TFHAR(r9)
2608 std r6, VCPU_TFIAR(r9)
2683 * r9 is saved in HSTATE_SCRATCH2(r13)
2693 mr r9, r1
2697 std r9, 0(r1)
2699 std r9, GPR1(r1)
2721 ld r9, HSTATE_SCRATCH2(r13)
2753 * r9 has a vcpu pointer (in)
2759 ld r11, VCPU_INTR_MSR(r9)
2790 lwz r9, VCPU_PMC + 20(r4)
2796 mtspr SPRN_PMC6, r9
2813 ld r9, VCPU_MMCRS(r4)
2816 mtspr SPRN_MMCRS, r9
2845 lwz r9, HSTATE_PMC6(r13)
2851 mtspr SPRN_PMC6, r9
2863 ld r9, HSTATE_SIER(r13)
2865 mtspr SPRN_SIER, r9
2880 mr r9, r3
2917 std r3, VCPU_MMCR(r9) /* if not, set saved MMCR0 to FC */
2922 std r4, VCPU_MMCR(r9)
2923 std r5, VCPU_MMCR + 8(r9)
2924 std r6, VCPU_MMCRA(r9)
2926 std r10, VCPU_MMCR + 16(r9)
2928 std r7, VCPU_SIAR(r9)
2929 std r8, VCPU_SDAR(r9)
2936 stw r3, VCPU_PMC(r9)
2937 stw r4, VCPU_PMC + 4(r9)
2938 stw r5, VCPU_PMC + 8(r9)
2939 stw r6, VCPU_PMC + 12(r9)
2940 stw r7, VCPU_PMC + 16(r9)
2941 stw r8, VCPU_PMC + 20(r9)
2944 std r5, VCPU_SIER(r9)
2948 stw r6, VCPU_PMC + 24(r9)
2949 stw r7, VCPU_PMC + 28(r9)
2950 std r8, VCPU_MMCRS(r9)