Lines Matching refs:r9
96 lwz r9, HSTATE_PMC6(r13)
102 mtspr SPRN_PMC6, r9
114 ld r9, HSTATE_SIER(r13)
116 mtspr SPRN_SIER, r9
439 addi r9, r5, VCORE_ENTRY_EXIT
440 21: lwarx r3, 0, r9
444 stwcx. r3, 0, r9
448 ld r9,VCORE_KVM(r5) /* pointer to struct kvm */
451 ld r6,KVM_SDR1(r9)
452 lwz r7,KVM_LPID(r9)
465 add r6,r6,r9
534 ld r9,VCPU_SLB_V(r6)
535 slbmte r9,r8
627 ld r9, VCPU_TAR_TM(r4)
632 mtspr SPRN_TAR, r9
698 lwz r9, VCPU_PMC + 20(r4)
704 mtspr SPRN_PMC6, r9
719 ld r9, VCPU_MMCR + 32(r4)
724 mtspr SPRN_MMCRS, r9
913 mr r9, r4
931 li r9, KVM_GUEST_MODE_GUEST_HV
932 stb r9, HSTATE_IN_GUEST(r13)
962 ld r9, VCPU_GPR(R9)(r4)
991 mr r9, r4
1016 std r9, HSTATE_SCRATCH2(r13)
1018 lbz r9, HSTATE_IN_GUEST(r13)
1019 cmpwi r9, KVM_GUEST_MODE_HOST_HV
1022 cmpwi r9, KVM_GUEST_MODE_GUEST
1023 ld r9, HSTATE_SCRATCH2(r13)
1027 li r9, KVM_GUEST_MODE_HOST_HV
1028 stb r9, HSTATE_IN_GUEST(r13)
1030 ld r9, HSTATE_KVM_VCPU(r13)
1034 std r0, VCPU_GPR(R0)(r9)
1035 std r1, VCPU_GPR(R1)(r9)
1036 std r2, VCPU_GPR(R2)(r9)
1037 std r3, VCPU_GPR(R3)(r9)
1038 std r4, VCPU_GPR(R4)(r9)
1039 std r5, VCPU_GPR(R5)(r9)
1040 std r6, VCPU_GPR(R6)(r9)
1041 std r7, VCPU_GPR(R7)(r9)
1042 std r8, VCPU_GPR(R8)(r9)
1044 std r0, VCPU_GPR(R9)(r9)
1045 std r10, VCPU_GPR(R10)(r9)
1046 std r11, VCPU_GPR(R11)(r9)
1049 std r3, VCPU_GPR(R12)(r9)
1050 stw r4, VCPU_CR(r9)
1053 std r3, VCPU_CFAR(r9)
1057 std r4, VCPU_PPR(r9)
1066 std r10, VCPU_SRR0(r9)
1067 std r11, VCPU_SRR1(r9)
1073 1: std r10, VCPU_PC(r9)
1074 std r11, VCPU_MSR(r9)
1078 std r3, VCPU_GPR(R13)(r9)
1079 std r4, VCPU_LR(r9)
1081 stw r12,VCPU_TRAP(r9)
1084 addi r3, r9, VCPU_TB_RMINTR
1085 mr r4, r9
1087 ld r5, VCPU_GPR(R5)(r9)
1088 ld r6, VCPU_GPR(R6)(r9)
1089 ld r7, VCPU_GPR(R7)(r9)
1090 ld r8, VCPU_GPR(R8)(r9)
1096 stw r3,VCPU_LAST_INST(r9)
1100 11: stw r3,VCPU_HEIR(r9)
1105 std r3, VCPU_CTR(r9)
1106 stw r4, VCPU_XER(r9)
1119 mr r4,r9
1149 mr r4, r9
1156 std r6, VCPU_DAR(r9)
1157 stw r7, VCPU_DSISR(r9)
1161 std r6, VCPU_FAULT_DAR(r9)
1162 stw r7, VCPU_FAULT_DSISR(r9)
1169 addi r3, r9, VCPU_TB_RMEXIT
1170 mr r4, r9
1178 ld r9, HSTATE_KVM_VCPU(r13)
1179 lwz r12, VCPU_TRAP(r9)
1183 stw r6,VCPU_CTRL(r9)
1190 lwz r0,VCPU_SLB_NR(r9) /* number of entries in SLB */
1193 addi r7,r9,VCPU_SLB
1206 stw r5,VCPU_SLB_MAX(r9)
1213 ld r7,VCPU_PURR(r9)
1214 ld r8,VCPU_SPURR(r9)
1215 std r5,VCPU_PURR(r9)
1216 std r6,VCPU_SPURR(r9)
1240 std r5,VCPU_DEC_EXPIRES(r9)
1249 std r5, VCPU_IAMR(r9)
1250 stw r6, VCPU_PSPB(r9)
1251 std r7, VCPU_FSCR(r9)
1255 std r5, VCPU_IC(r9)
1256 std r6, VCPU_VTB(r9)
1257 std r7, VCPU_TAR(r9)
1259 std r8, VCPU_EBBHR(r9)
1264 std r5, VCPU_EBBRR(r9)
1265 std r6, VCPU_BESCR(r9)
1266 std r7, VCPU_CSIGR(r9)
1267 std r8, VCPU_TACR(r9)
1272 std r5, VCPU_TCSCR(r9)
1273 std r6, VCPU_ACOP(r9)
1274 stw r7, VCPU_GUEST_PID(r9)
1275 std r8, VCPU_WORT(r9)
1295 std r5,VCPU_AMR(r9)
1296 std r6,VCPU_UAMOR(r9)
1303 std r8, VCPU_DSCR(r9)
1307 std r14, VCPU_GPR(R14)(r9)
1308 std r15, VCPU_GPR(R15)(r9)
1309 std r16, VCPU_GPR(R16)(r9)
1310 std r17, VCPU_GPR(R17)(r9)
1311 std r18, VCPU_GPR(R18)(r9)
1312 std r19, VCPU_GPR(R19)(r9)
1313 std r20, VCPU_GPR(R20)(r9)
1314 std r21, VCPU_GPR(R21)(r9)
1315 std r22, VCPU_GPR(R22)(r9)
1316 std r23, VCPU_GPR(R23)(r9)
1317 std r24, VCPU_GPR(R24)(r9)
1318 std r25, VCPU_GPR(R25)(r9)
1319 std r26, VCPU_GPR(R26)(r9)
1320 std r27, VCPU_GPR(R27)(r9)
1321 std r28, VCPU_GPR(R28)(r9)
1322 std r29, VCPU_GPR(R29)(r9)
1323 std r30, VCPU_GPR(R30)(r9)
1324 std r31, VCPU_GPR(R31)(r9)
1331 std r3, VCPU_SPRG0(r9)
1332 std r4, VCPU_SPRG1(r9)
1333 std r5, VCPU_SPRG2(r9)
1334 std r6, VCPU_SPRG3(r9)
1337 mr r3, r9
1350 ld r5, VCPU_MSR(r9)
1366 std r9, PACATMSCRATCH(r13)
1367 ld r9, HSTATE_KVM_VCPU(r13)
1370 std r29, VCPU_GPRS_TM(29)(r9)
1371 std r30, VCPU_GPRS_TM(30)(r9)
1372 std r31, VCPU_GPRS_TM(31)(r9)
1385 std reg, VCPU_GPRS_TM(reg)(r9)
1391 std r4, VCPU_GPRS_TM(13)(r9)
1394 std r4, VCPU_GPRS_TM(9)(r9)
1405 std r31, VCPU_PPR_TM(r9)
1406 std r30, VCPU_DSCR_TM(r9)
1412 std r5, VCPU_LR_TM(r9)
1413 stw r6, VCPU_CR_TM(r9)
1414 std r7, VCPU_CTR_TM(r9)
1415 std r8, VCPU_AMR_TM(r9)
1416 std r10, VCPU_TAR_TM(r9)
1419 lwz r12, VCPU_TRAP(r9)
1422 addi r3, r9, VCPU_FPRS_TM
1424 addi r3, r9, VCPU_VRS_TM
1427 stw r6, VCPU_VRSAVE_TM(r9)
1438 std r5, VCPU_TFHAR(r9)
1439 std r6, VCPU_TFIAR(r9)
1440 std r7, VCPU_TEXASR(r9)
1445 ld r8, VCPU_VPA(r9) /* do they have a VPA? */
1453 stb r3, VCPU_VPA_DIRTY(r9)
1494 std r3, VCPU_MMCR(r9) /* if not, set saved MMCR0 to FC */
1499 std r4, VCPU_MMCR(r9)
1500 std r5, VCPU_MMCR + 8(r9)
1501 std r6, VCPU_MMCR + 16(r9)
1503 std r10, VCPU_MMCR + 24(r9)
1505 std r7, VCPU_SIAR(r9)
1506 std r8, VCPU_SDAR(r9)
1513 stw r3, VCPU_PMC(r9)
1514 stw r4, VCPU_PMC + 4(r9)
1515 stw r5, VCPU_PMC + 8(r9)
1516 stw r6, VCPU_PMC + 12(r9)
1517 stw r7, VCPU_PMC + 16(r9)
1518 stw r8, VCPU_PMC + 20(r9)
1524 std r5, VCPU_SIER(r9)
1525 stw r6, VCPU_PMC + 24(r9)
1526 stw r7, VCPU_PMC + 28(r9)
1527 std r8, VCPU_MMCR + 32(r9)
1665 4: std r4, VCPU_FAULT_DAR(r9)
1666 stw r6, VCPU_FAULT_DSISR(r9)
1669 mr r3, r9 /* vcpu pointer */
1672 ld r9, HSTATE_KVM_VCPU(r13)
1673 ld r10, VCPU_PC(r9)
1674 ld r11, VCPU_MSR(r9)
1684 ld r4, VCPU_FAULT_DAR(r9)
1693 6: ld r7, VCPU_CTR(r9)
1694 lwz r8, VCPU_XER(r9)
1697 mr r4, r9
1700 3: ld r5, VCPU_KVM(r9) /* not relocated, use VRMA */
1720 stw r8, VCPU_LAST_INST(r9)
1741 mr r3, r9 /* vcpu pointer */
1746 ld r9, HSTATE_KVM_VCPU(r13)
1747 ld r10, VCPU_PC(r9)
1748 ld r11, VCPU_MSR(r9)
1763 3: ld r6, VCPU_KVM(r9) /* not relocated, use VRMA */
1777 ld r3,VCPU_GPR(R3)(r9)
1785 ld r4, VCPU_KVM(r9)
1801 mr r3,r9 /* get vcpu pointer */
1802 ld r4,VCPU_GPR(R4)(r9)
1817 mr r4,r9
1825 ld r9, HSTATE_KVM_VCPU(r13)
2268 mr r9, r4
2290 ld r9, HSTATE_KVM_VCPU(r13)
2295 mr r3, r9 /* get vcpu pointer */
2299 ld r9, HSTATE_KVM_VCPU(r13)
2312 ld r10, VCPU_PC(r9)
2313 ld r11, VCPU_MSR(r9)
2317 ld r11, VCPU_MSR(r9)
2490 mfmsr r9
2491 ori r8,r9,MSR_FP
2535 ld r11, VCPU_INTR_MSR(r9)