Lines Matching refs:r9
96 lwz r9, HSTATE_PMC6(r13)
102 mtspr SPRN_PMC6, r9
114 ld r9, HSTATE_SIER(r13)
116 mtspr SPRN_SIER, r9
524 addi r9, r5, VCORE_ENTRY_EXIT
525 21: lwarx r3, 0, r9
529 stwcx. r3, 0, r9
533 ld r9,VCORE_KVM(r5) /* pointer to struct kvm */
536 ld r6,KVM_SDR1(r9)
537 lwz r7,KVM_LPID(r9)
550 add r6,r6,r9
619 ld r9,VCPU_SLB_V(r6)
620 slbmte r9,r8
712 ld r9, VCPU_TAR_TM(r4)
717 mtspr SPRN_TAR, r9
783 lwz r9, VCPU_PMC + 20(r4)
789 mtspr SPRN_PMC6, r9
804 ld r9, VCPU_MMCR + 32(r4)
809 mtspr SPRN_MMCRS, r9
1001 mr r9, r4
1019 li r9, KVM_GUEST_MODE_GUEST_HV
1020 stb r9, HSTATE_IN_GUEST(r13)
1050 ld r9, VCPU_GPR(R9)(r4)
1083 mr r9, r4
1108 std r9, HSTATE_SCRATCH2(r13)
1110 lbz r9, HSTATE_IN_GUEST(r13)
1111 cmpwi r9, KVM_GUEST_MODE_HOST_HV
1114 cmpwi r9, KVM_GUEST_MODE_GUEST
1115 ld r9, HSTATE_SCRATCH2(r13)
1119 li r9, KVM_GUEST_MODE_HOST_HV
1120 stb r9, HSTATE_IN_GUEST(r13)
1122 ld r9, HSTATE_KVM_VCPU(r13)
1126 std r0, VCPU_GPR(R0)(r9)
1127 std r1, VCPU_GPR(R1)(r9)
1128 std r2, VCPU_GPR(R2)(r9)
1129 std r3, VCPU_GPR(R3)(r9)
1130 std r4, VCPU_GPR(R4)(r9)
1131 std r5, VCPU_GPR(R5)(r9)
1132 std r6, VCPU_GPR(R6)(r9)
1133 std r7, VCPU_GPR(R7)(r9)
1134 std r8, VCPU_GPR(R8)(r9)
1136 std r0, VCPU_GPR(R9)(r9)
1137 std r10, VCPU_GPR(R10)(r9)
1138 std r11, VCPU_GPR(R11)(r9)
1141 std r3, VCPU_GPR(R12)(r9)
1142 stw r4, VCPU_CR(r9)
1145 std r3, VCPU_CFAR(r9)
1149 std r4, VCPU_PPR(r9)
1158 std r10, VCPU_SRR0(r9)
1159 std r11, VCPU_SRR1(r9)
1165 1: std r10, VCPU_PC(r9)
1166 std r11, VCPU_MSR(r9)
1170 std r3, VCPU_GPR(R13)(r9)
1171 std r4, VCPU_LR(r9)
1173 stw r12,VCPU_TRAP(r9)
1176 addi r3, r9, VCPU_TB_RMINTR
1177 mr r4, r9
1179 ld r5, VCPU_GPR(R5)(r9)
1180 ld r6, VCPU_GPR(R6)(r9)
1181 ld r7, VCPU_GPR(R7)(r9)
1182 ld r8, VCPU_GPR(R8)(r9)
1188 stw r3,VCPU_LAST_INST(r9)
1192 11: stw r3,VCPU_HEIR(r9)
1197 std r3, VCPU_CTR(r9)
1198 std r4, VCPU_XER(r9)
1211 mr r4,r9
1241 mr r4, r9
1248 std r6, VCPU_DAR(r9)
1249 stw r7, VCPU_DSISR(r9)
1253 std r6, VCPU_FAULT_DAR(r9)
1254 stw r7, VCPU_FAULT_DSISR(r9)
1261 addi r3, r9, VCPU_TB_RMEXIT
1262 mr r4, r9
1270 ld r9, HSTATE_KVM_VCPU(r13)
1271 lwz r12, VCPU_TRAP(r9)
1275 stw r0, VCPU_CPU(r9)
1276 stw r0, VCPU_THREAD_CPU(r9)
1280 stw r6,VCPU_CTRL(r9)
1287 lwz r0,VCPU_SLB_NR(r9) /* number of entries in SLB */
1290 addi r7,r9,VCPU_SLB
1303 stw r5,VCPU_SLB_MAX(r9)
1310 ld r7,VCPU_PURR(r9)
1311 ld r8,VCPU_SPURR(r9)
1312 std r5,VCPU_PURR(r9)
1313 std r6,VCPU_SPURR(r9)
1337 std r5,VCPU_DEC_EXPIRES(r9)
1346 std r5, VCPU_IAMR(r9)
1347 stw r6, VCPU_PSPB(r9)
1348 std r7, VCPU_FSCR(r9)
1352 std r5, VCPU_IC(r9)
1353 std r6, VCPU_VTB(r9)
1354 std r7, VCPU_TAR(r9)
1356 std r8, VCPU_EBBHR(r9)
1361 std r5, VCPU_EBBRR(r9)
1362 std r6, VCPU_BESCR(r9)
1363 std r7, VCPU_CSIGR(r9)
1364 std r8, VCPU_TACR(r9)
1369 std r5, VCPU_TCSCR(r9)
1370 std r6, VCPU_ACOP(r9)
1371 stw r7, VCPU_GUEST_PID(r9)
1372 std r8, VCPU_WORT(r9)
1392 std r5,VCPU_AMR(r9)
1393 std r6,VCPU_UAMOR(r9)
1400 std r8, VCPU_DSCR(r9)
1404 std r14, VCPU_GPR(R14)(r9)
1405 std r15, VCPU_GPR(R15)(r9)
1406 std r16, VCPU_GPR(R16)(r9)
1407 std r17, VCPU_GPR(R17)(r9)
1408 std r18, VCPU_GPR(R18)(r9)
1409 std r19, VCPU_GPR(R19)(r9)
1410 std r20, VCPU_GPR(R20)(r9)
1411 std r21, VCPU_GPR(R21)(r9)
1412 std r22, VCPU_GPR(R22)(r9)
1413 std r23, VCPU_GPR(R23)(r9)
1414 std r24, VCPU_GPR(R24)(r9)
1415 std r25, VCPU_GPR(R25)(r9)
1416 std r26, VCPU_GPR(R26)(r9)
1417 std r27, VCPU_GPR(R27)(r9)
1418 std r28, VCPU_GPR(R28)(r9)
1419 std r29, VCPU_GPR(R29)(r9)
1420 std r30, VCPU_GPR(R30)(r9)
1421 std r31, VCPU_GPR(R31)(r9)
1428 std r3, VCPU_SPRG0(r9)
1429 std r4, VCPU_SPRG1(r9)
1430 std r5, VCPU_SPRG2(r9)
1431 std r6, VCPU_SPRG3(r9)
1434 mr r3, r9
1447 ld r5, VCPU_MSR(r9)
1463 std r9, PACATMSCRATCH(r13)
1464 ld r9, HSTATE_KVM_VCPU(r13)
1467 std r29, VCPU_GPRS_TM(29)(r9)
1468 std r30, VCPU_GPRS_TM(30)(r9)
1469 std r31, VCPU_GPRS_TM(31)(r9)
1482 std reg, VCPU_GPRS_TM(reg)(r9)
1488 std r4, VCPU_GPRS_TM(13)(r9)
1491 std r4, VCPU_GPRS_TM(9)(r9)
1502 std r31, VCPU_PPR_TM(r9)
1503 std r30, VCPU_DSCR_TM(r9)
1509 std r5, VCPU_LR_TM(r9)
1510 stw r6, VCPU_CR_TM(r9)
1511 std r7, VCPU_CTR_TM(r9)
1512 std r8, VCPU_AMR_TM(r9)
1513 std r10, VCPU_TAR_TM(r9)
1516 lwz r12, VCPU_TRAP(r9)
1519 addi r3, r9, VCPU_FPRS_TM
1521 addi r3, r9, VCPU_VRS_TM
1524 stw r6, VCPU_VRSAVE_TM(r9)
1535 std r5, VCPU_TFHAR(r9)
1536 std r6, VCPU_TFIAR(r9)
1537 std r7, VCPU_TEXASR(r9)
1542 ld r8, VCPU_VPA(r9) /* do they have a VPA? */
1550 stb r3, VCPU_VPA_DIRTY(r9)
1591 std r3, VCPU_MMCR(r9) /* if not, set saved MMCR0 to FC */
1596 std r4, VCPU_MMCR(r9)
1597 std r5, VCPU_MMCR + 8(r9)
1598 std r6, VCPU_MMCR + 16(r9)
1600 std r10, VCPU_MMCR + 24(r9)
1602 std r7, VCPU_SIAR(r9)
1603 std r8, VCPU_SDAR(r9)
1610 stw r3, VCPU_PMC(r9)
1611 stw r4, VCPU_PMC + 4(r9)
1612 stw r5, VCPU_PMC + 8(r9)
1613 stw r6, VCPU_PMC + 12(r9)
1614 stw r7, VCPU_PMC + 16(r9)
1615 stw r8, VCPU_PMC + 20(r9)
1621 std r5, VCPU_SIER(r9)
1622 stw r6, VCPU_PMC + 24(r9)
1623 stw r7, VCPU_PMC + 28(r9)
1624 std r8, VCPU_MMCR + 32(r9)
1768 4: std r4, VCPU_FAULT_DAR(r9)
1769 stw r6, VCPU_FAULT_DSISR(r9)
1772 mr r3, r9 /* vcpu pointer */
1775 ld r9, HSTATE_KVM_VCPU(r13)
1776 ld r10, VCPU_PC(r9)
1777 ld r11, VCPU_MSR(r9)
1787 ld r4, VCPU_FAULT_DAR(r9)
1797 6: ld r7, VCPU_CTR(r9)
1798 ld r8, VCPU_XER(r9)
1801 mr r4, r9
1804 3: ld r5, VCPU_KVM(r9) /* not relocated, use VRMA */
1824 stw r8, VCPU_LAST_INST(r9)
1846 mr r3, r9 /* vcpu pointer */
1851 ld r9, HSTATE_KVM_VCPU(r13)
1852 ld r10, VCPU_PC(r9)
1853 ld r11, VCPU_MSR(r9)
1869 3: ld r6, VCPU_KVM(r9) /* not relocated, use VRMA */
1883 ld r3,VCPU_GPR(R3)(r9)
1891 ld r4, VCPU_KVM(r9)
1907 mr r3,r9 /* get vcpu pointer */
1908 ld r4,VCPU_GPR(R4)(r9)
1923 mr r4,r9
1931 ld r9, HSTATE_KVM_VCPU(r13)
2374 mr r9, r4
2396 ld r9, HSTATE_KVM_VCPU(r13)
2401 mr r3, r9 /* get vcpu pointer */
2404 ld r9, HSTATE_KVM_VCPU(r13)
2420 ld r11, VCPU_MSR(r9)
2423 ld r10, VCPU_PC(r9)
2605 mfmsr r9
2606 ori r8,r9,MSR_FP
2650 ld r11, VCPU_INTR_MSR(r9)