vval 1419 arch/arm/kernel/perf_event_v7.c u32 vval, fval; vval 1440 arch/arm/kernel/perf_event_v7.c venum_pre_pmresr(&vval, &fval); vval 1446 arch/arm/kernel/perf_event_v7.c venum_post_pmresr(vval, fval); vval 1475 arch/arm/kernel/perf_event_v7.c u32 vval, fval; vval 1481 arch/arm/kernel/perf_event_v7.c venum_pre_pmresr(&vval, &fval); vval 1485 arch/arm/kernel/perf_event_v7.c venum_post_pmresr(vval, fval); vval 1557 arch/arm/kernel/perf_event_v7.c u32 vval, fval; vval 1568 arch/arm/kernel/perf_event_v7.c venum_pre_pmresr(&vval, &fval); vval 1570 arch/arm/kernel/perf_event_v7.c venum_post_pmresr(vval, fval); vval 1766 arch/arm/kernel/perf_event_v7.c u32 vval, fval; vval 1789 arch/arm/kernel/perf_event_v7.c venum_pre_pmresr(&vval, &fval); vval 1795 arch/arm/kernel/perf_event_v7.c venum_post_pmresr(vval, fval); vval 1808 arch/arm/kernel/perf_event_v7.c u32 vval, fval; vval 1814 arch/arm/kernel/perf_event_v7.c venum_pre_pmresr(&vval, &fval); vval 1818 arch/arm/kernel/perf_event_v7.c venum_post_pmresr(vval, fval); vval 1890 arch/arm/kernel/perf_event_v7.c u32 vval, fval; vval 1902 arch/arm/kernel/perf_event_v7.c venum_pre_pmresr(&vval, &fval); vval 1904 arch/arm/kernel/perf_event_v7.c venum_post_pmresr(vval, fval); vval 252 arch/powerpc/include/asm/kvm_ppc.h vector128 vval; vval 1779 arch/powerpc/kvm/book3s_hv.c val->vval = vcpu->arch.vr_tm.vr[i-32]; vval 2009 arch/powerpc/kvm/book3s_hv.c vcpu->arch.vr_tm.vr[i-32] = val->vval; vval 1561 arch/powerpc/kvm/book3s_pr.c val->vval = vcpu->arch.vr_tm.vr[i-32]; vval 1661 arch/powerpc/kvm/book3s_pr.c vcpu->arch.vr_tm.vr[i-32] = val->vval; vval 902 arch/powerpc/kvm/powerpc.c val.vval = VCPU_VSX_VR(vcpu, index - 32); vval 904 arch/powerpc/kvm/powerpc.c VCPU_VSX_VR(vcpu, index - 32) = val.vval; vval 917 arch/powerpc/kvm/powerpc.c val.vval = VCPU_VSX_VR(vcpu, index - 32); vval 920 arch/powerpc/kvm/powerpc.c VCPU_VSX_VR(vcpu, index - 32) = val.vval; vval 938 arch/powerpc/kvm/powerpc.c VCPU_VSX_VR(vcpu, index - 32) = val.vval; vval 959 arch/powerpc/kvm/powerpc.c val.vval = VCPU_VSX_VR(vcpu, index - 32); vval 961 arch/powerpc/kvm/powerpc.c VCPU_VSX_VR(vcpu, index - 32) = val.vval; vval 1026 arch/powerpc/kvm/powerpc.c val.vval = VCPU_VSX_VR(vcpu, index); vval 1028 arch/powerpc/kvm/powerpc.c VCPU_VSX_VR(vcpu, index) = val.vval; vval 1042 arch/powerpc/kvm/powerpc.c val.vval = VCPU_VSX_VR(vcpu, index); vval 1044 arch/powerpc/kvm/powerpc.c VCPU_VSX_VR(vcpu, index) = val.vval; vval 1058 arch/powerpc/kvm/powerpc.c val.vval = VCPU_VSX_VR(vcpu, index); vval 1060 arch/powerpc/kvm/powerpc.c VCPU_VSX_VR(vcpu, index) = val.vval; vval 1074 arch/powerpc/kvm/powerpc.c val.vval = VCPU_VSX_VR(vcpu, index); vval 1076 arch/powerpc/kvm/powerpc.c VCPU_VSX_VR(vcpu, index) = val.vval; vval 1393 arch/powerpc/kvm/powerpc.c reg.vval = VCPU_VSX_VR(vcpu, rs - 32); vval 1413 arch/powerpc/kvm/powerpc.c reg.vval = VCPU_VSX_VR(vcpu, rs - 32); vval 1528 arch/powerpc/kvm/powerpc.c reg.vval = VCPU_VSX_VR(vcpu, index); vval 1546 arch/powerpc/kvm/powerpc.c reg.vval = VCPU_VSX_VR(vcpu, index); vval 1564 arch/powerpc/kvm/powerpc.c reg.vval = VCPU_VSX_VR(vcpu, index); vval 1582 arch/powerpc/kvm/powerpc.c reg.vval = VCPU_VSX_VR(vcpu, index); vval 1691 arch/powerpc/kvm/powerpc.c val.vval = vcpu->arch.vr.vr[reg->id - KVM_REG_PPC_VR0]; vval 1742 arch/powerpc/kvm/powerpc.c vcpu->arch.vr.vr[reg->id - KVM_REG_PPC_VR0] = val.vval; vval 184 arch/xtensa/include/asm/atomic.h unsigned int vval; \ vval 193 arch/xtensa/include/asm/atomic.h : "=&a" (vval) \ vval 202 arch/xtensa/include/asm/atomic.h unsigned int vval; \ vval 211 arch/xtensa/include/asm/atomic.h : "=&a" (vval) \ vval 216 arch/xtensa/include/asm/atomic.h return vval; \ vval 222 arch/xtensa/include/asm/atomic.h unsigned int tmp, vval; \ vval 231 arch/xtensa/include/asm/atomic.h : "=&a" (vval), "=&a" (tmp) \ vval 236 arch/xtensa/include/asm/atomic.h return vval; \