Lines Matching refs:r4
63 ld r4, HSTATE_KVM_VCPU(r13)
82 lbz r4, LPPACA_PMCINUSE(r3)
83 cmpwi r4, 0
87 andi. r4, r3, MMCR0_PMAO_SYNC | MMCR0_PMAO
88 cmpwi r4, MMCR0_PMAO
92 lwz r4, HSTATE_PMC2(r13)
98 mtspr SPRN_PMC2, r4
104 ld r4, HSTATE_MMCR1(r13)
108 mtspr SPRN_MMCR1, r4
127 mftb r4
128 subf r4, r4, r3
129 mtspr SPRN_DEC, r4
258 ld r4, HSTATE_KVM_VCPU(r13)
259 cmpdi r4, 0
263 addi r3, r4, VCPU_TB_RMENTRY
270 ld r4, HSTATE_KVM_VCPU(r13)
271 cmpdi r4, 0
273 addi r3, r4, VCPU_TB_RMEXIT
340 lbz r4, HSTATE_PTID(r13)
341 cmpwi r4, 0
360 ld r4, HSTATE_KVM_VCPU(r13)
400 mfspr r4, SPRN_LPCR
401 rlwimi r4, r3, 0, LPCR_PECE0 | LPCR_PECE1
402 mtspr SPRN_LPCR, r4
446 lhz r4, PACAPACAINDEX(r13)
447 clrldi r4, r4, 61 /* micro-threading => P8 => 8 threads/core */
448 addi r4, r4, KVM_SPLIT_NAPPED
449 stbx r0, r3, r4
456 mfspr r4, SPRN_LPCR
457 rlwimi r4, r3, 4, (LPCR_PECEDP | LPCR_PECEDH | LPCR_PECE0 | LPCR_PECE1)
458 mtspr SPRN_LPCR, r4
469 stbx r0, r3, r4
502 cmpdi r4, 0
504 addi r3, r4, VCPU_TB_RMENTRY
608 10: cmpdi r4, 0
613 lwz r5,VCPU_SLB_MAX(r4)
617 addi r6,r4,VCPU_SLB
625 ld r3, VCPU_VPA(r4)
633 stb r6, VCPU_VPA_DIRTY(r4)
641 ld r7,VCPU_PURR(r4)
642 ld r8,VCPU_SPURR(r4)
649 lwz r5,VCPU_DABRX(r4)
650 ld r6,VCPU_DABR(r4)
674 ld r5, VCPU_TFHAR(r4)
675 ld r6, VCPU_TFIAR(r4)
676 ld r7, VCPU_TEXASR(r4)
681 ld r5, VCPU_MSR(r4)
699 mr r31, r4
704 mr r4, r31
705 lwz r7, VCPU_VRSAVE_TM(r4)
708 ld r5, VCPU_LR_TM(r4)
709 lwz r6, VCPU_CR_TM(r4)
710 ld r7, VCPU_CTR_TM(r4)
711 ld r8, VCPU_AMR_TM(r4)
712 ld r9, VCPU_TAR_TM(r4)
724 ld r29, VCPU_DSCR_TM(r4)
725 ld r30, VCPU_PPR_TM(r4)
756 ld r4, HSTATE_KVM_VCPU(r13)
773 ld r3, VCPU_MMCR(r4)
778 lwz r3, VCPU_PMC(r4) /* always load up guest PMU registers */
779 lwz r5, VCPU_PMC + 4(r4) /* to prevent information leak */
780 lwz r6, VCPU_PMC + 8(r4)
781 lwz r7, VCPU_PMC + 12(r4)
782 lwz r8, VCPU_PMC + 16(r4)
783 lwz r9, VCPU_PMC + 20(r4)
790 ld r3, VCPU_MMCR(r4)
791 ld r5, VCPU_MMCR + 8(r4)
792 ld r6, VCPU_MMCR + 16(r4)
793 ld r7, VCPU_SIAR(r4)
794 ld r8, VCPU_SDAR(r4)
800 ld r5, VCPU_MMCR + 24(r4)
801 ld r6, VCPU_SIER(r4)
802 lwz r7, VCPU_PMC + 24(r4)
803 lwz r8, VCPU_PMC + 28(r4)
804 ld r9, VCPU_MMCR + 32(r4)
817 ld r14, VCPU_GPR(R14)(r4)
818 ld r15, VCPU_GPR(R15)(r4)
819 ld r16, VCPU_GPR(R16)(r4)
820 ld r17, VCPU_GPR(R17)(r4)
821 ld r18, VCPU_GPR(R18)(r4)
822 ld r19, VCPU_GPR(R19)(r4)
823 ld r20, VCPU_GPR(R20)(r4)
824 ld r21, VCPU_GPR(R21)(r4)
825 ld r22, VCPU_GPR(R22)(r4)
826 ld r23, VCPU_GPR(R23)(r4)
827 ld r24, VCPU_GPR(R24)(r4)
828 ld r25, VCPU_GPR(R25)(r4)
829 ld r26, VCPU_GPR(R26)(r4)
830 ld r27, VCPU_GPR(R27)(r4)
831 ld r28, VCPU_GPR(R28)(r4)
832 ld r29, VCPU_GPR(R29)(r4)
833 ld r30, VCPU_GPR(R30)(r4)
834 ld r31, VCPU_GPR(R31)(r4)
837 ld r5, VCPU_DSCR(r4)
851 ld r5, VCPU_IAMR(r4)
852 lwz r6, VCPU_PSPB(r4)
853 ld r7, VCPU_FSCR(r4)
857 ld r5, VCPU_DAWR(r4)
858 ld r6, VCPU_DAWRX(r4)
859 ld r7, VCPU_CIABR(r4)
860 ld r8, VCPU_TAR(r4)
865 ld r5, VCPU_IC(r4)
866 ld r6, VCPU_VTB(r4)
869 ld r8, VCPU_EBBHR(r4)
871 ld r5, VCPU_EBBRR(r4)
872 ld r6, VCPU_BESCR(r4)
873 ld r7, VCPU_CSIGR(r4)
874 ld r8, VCPU_TACR(r4)
879 ld r5, VCPU_TCSCR(r4)
880 ld r6, VCPU_ACOP(r4)
881 lwz r7, VCPU_GUEST_PID(r4)
882 ld r8, VCPU_WORT(r4)
892 ld r8,VCPU_DEC_EXPIRES(r4)
900 stw r3,VCPU_DEC(r4)
902 ld r5, VCPU_SPRG0(r4)
903 ld r6, VCPU_SPRG1(r4)
904 ld r7, VCPU_SPRG2(r4)
905 ld r8, VCPU_SPRG3(r4)
912 ld r5, VCPU_DAR(r4)
913 lwz r6, VCPU_DSISR(r4)
918 ld r5,VCPU_AMR(r4)
919 ld r6,VCPU_UAMOR(r4)
926 lwz r5,VCPU_CTRL(r4)
960 ld r6, VCPU_CTR(r4)
961 ld r7, VCPU_XER(r4)
967 ld r10, VCPU_PC(r4)
968 ld r11, VCPU_MSR(r4)
969 ld r6, VCPU_SRR0(r4)
970 ld r7, VCPU_SRR1(r4)
981 ld r0, VCPU_PENDING_EXC(r4)
1001 mr r9, r4
1014 stb r0,VCPU_CEDED(r4) /* cancel cede */
1024 addi r3, r4, VCPU_TB_GUEST
1031 ld r5, VCPU_CFAR(r4)
1035 ld r0, VCPU_PPR(r4)
1038 ld r5, VCPU_LR(r4)
1039 lwz r6, VCPU_CR(r4)
1043 ld r1, VCPU_GPR(R1)(r4)
1044 ld r2, VCPU_GPR(R2)(r4)
1045 ld r3, VCPU_GPR(R3)(r4)
1046 ld r5, VCPU_GPR(R5)(r4)
1047 ld r6, VCPU_GPR(R6)(r4)
1048 ld r7, VCPU_GPR(R7)(r4)
1049 ld r8, VCPU_GPR(R8)(r4)
1050 ld r9, VCPU_GPR(R9)(r4)
1051 ld r10, VCPU_GPR(R10)(r4)
1052 ld r11, VCPU_GPR(R11)(r4)
1053 ld r12, VCPU_GPR(R12)(r4)
1054 ld r13, VCPU_GPR(R13)(r4)
1059 ld r0, VCPU_GPR(R0)(r4)
1060 ld r4, VCPU_GPR(R4)(r4)
1067 cmpdi r4, 0
1069 stw r12, VCPU_TRAP(r4)
1071 addi r3, r4, VCPU_TB_RMEXIT
1082 12: stw r12, VCPU_TRAP(r4)
1083 mr r9, r4
1085 addi r3, r4, VCPU_TB_RMEXIT
1130 std r4, VCPU_GPR(R4)(r9)
1140 lwz r4, HSTATE_SCRATCH1(r13)
1142 stw r4, VCPU_CR(r9)
1148 ld r4, HSTATE_PPR(r13)
1149 std r4, VCPU_PPR(r9)
1169 mflr r4
1171 std r4, VCPU_LR(r9)
1177 mr r4, r9
1196 mfxer r4
1198 std r4, VCPU_XER(r9)
1211 mr r4,r9
1241 mr r4, r9
1262 mr r4, r9
1322 ld r4,HSTATE_SPURR(r13)
1324 add r4,r4,r6
1326 mtspr SPRN_SPURR,r4
1335 ld r4,VCORE_TB_OFFSET(r3)
1336 subf r5,r4,r5
1425 mfspr r4, SPRN_SPRG1
1429 std r4, VCPU_SPRG1(r9)
1487 GET_SCRATCH0(r4)
1488 std r4, VCPU_GPRS_TM(13)(r9)
1490 ld r4, PACATMSCRATCH(r13)
1491 std r4, VCPU_GPRS_TM(9)(r9)
1545 li r4, LPPACA_YIELDCOUNT
1546 LWZX_BE r3, r8, r4
1548 STWX_BE r3, r8, r4
1580 mfspr r4, SPRN_MMCR0 /* save MMCR0 */
1596 std r4, VCPU_MMCR(r9)
1605 mfspr r4, SPRN_PMC2
1611 stw r4, VCPU_PMC + 4(r9)
1625 lis r4, 0x8000
1626 mtspr SPRN_MMCRS, r4
1643 ld r4,VCORE_KVM(r5) /* pointer to struct kvm */
1668 ld r6,KVM_HOST_SDR1(r4)
1669 lwz r7,KVM_HOST_LPID(r4)
1713 16: ld r8,KVM_HOST_LPCR(r4)
1733 ld r4, HSTATE_KVM_VCPU(r13)
1734 cmpdi r4, 0
1757 mfspr r4, SPRN_HDAR
1764 clrrdi r0, r4, 28
1768 4: std r4, VCPU_FAULT_DAR(r9)
1787 ld r4, VCPU_FAULT_DAR(r9)
1791 7: mtspr SPRN_DAR, r4
1801 mr r4, r9
1818 ori r4, r3, MSR_DR /* Enable paging for data */
1819 mtmsrd r4
1847 mr r4, r10
1891 ld r4, VCPU_KVM(r9)
1894 add r4, r4, r0
1895 ld r0, KVM_ENABLED_HCALLS(r4)
1896 rlwinm r4, r3, 32-2, 0x3f /* r4 = (r3 / 4) & 0x3f */
1897 srd r0, r0, r4
1901 LOAD_REG_ADDR(r4, hcall_real_table)
1902 lwax r3,r3,r4
1905 add r12,r3,r4
1908 ld r4,VCPU_GPR(R4)(r9)
1912 ld r4,HSTATE_KVM_VCPU(r13)
1913 std r3,VCPU_GPR(R3)(r4)
1914 ld r10,VCPU_PC(r4)
1915 ld r11,VCPU_MSR(r4)
1923 mr r4,r9
2156 std r4,VCPU_DABR(r3)
2160 1: mtspr SPRN_DABR,r4
2162 cmpd r4, r5
2169 2: rlwimi r5, r4, 5, DAWRX_DR | DAWRX_DW
2170 rlwimi r5, r4, 2, DAWRX_WT
2171 clrrdi r4, r4, 3
2172 std r4, VCPU_DAWR(r3)
2174 mtspr SPRN_DAWR, r4
2205 31: lwarx r4,0,r6
2206 or r4,r4,r0
2207 cmpw r4,r8
2209 stwcx. r4,0,r6
2254 mfspr r4, SPRN_HDEC
2256 cmpw r3, r4
2258 mtspr SPRN_DEC, r4
2263 ld r4, HSTATE_KVM_VCPU(r13)
2267 std r3, VCPU_DEC_EXPIRES(r4)
2270 ld r4, HSTATE_KVM_VCPU(r13)
2271 addi r3, r4, VCPU_TB_CEDE
2307 33: mr r4, r3
2314 ld r4, HSTATE_KVM_VCPU(r13)
2320 addi r3, r4, VCPU_TB_RMINTR
2328 ld r3, VCPU_DEC_EXPIRES(r4)
2337 ld r14, VCPU_GPR(R14)(r4)
2338 ld r15, VCPU_GPR(R15)(r4)
2339 ld r16, VCPU_GPR(R16)(r4)
2340 ld r17, VCPU_GPR(R17)(r4)
2341 ld r18, VCPU_GPR(R18)(r4)
2342 ld r19, VCPU_GPR(R19)(r4)
2343 ld r20, VCPU_GPR(R20)(r4)
2344 ld r21, VCPU_GPR(R21)(r4)
2345 ld r22, VCPU_GPR(R22)(r4)
2346 ld r23, VCPU_GPR(R23)(r4)
2347 ld r24, VCPU_GPR(R24)(r4)
2348 ld r25, VCPU_GPR(R25)(r4)
2349 ld r26, VCPU_GPR(R26)(r4)
2350 ld r27, VCPU_GPR(R27)(r4)
2351 ld r28, VCPU_GPR(R28)(r4)
2352 ld r29, VCPU_GPR(R29)(r4)
2353 ld r30, VCPU_GPR(R30)(r4)
2354 ld r31, VCPU_GPR(R31)(r4)
2373 stw r12, VCPU_TRAP(r4)
2374 mr r9, r4
2604 mr r31,r4
2618 addi r3,r4,VCPU_FPRS
2629 mr r4,r31
2688 std r3, VCPU_CUR_ACTIVITY(r4)
2689 std r5, VCPU_ACTIVITY_START(r4)
2702 4: ld r5, VCPU_CUR_ACTIVITY(r4)
2703 ld r6, VCPU_ACTIVITY_START(r4)
2704 std r3, VCPU_CUR_ACTIVITY(r4)
2707 std r7, VCPU_ACTIVITY_START(r4)