VCPU_SREG_CS 785 arch/x86/kvm/emulate.c struct segmented_address addr = { .seg = VCPU_SREG_CS, VCPU_SREG_CS 883 arch/x86/kvm/emulate.c struct segmented_address addr = { .seg = VCPU_SREG_CS, VCPU_SREG_CS 1679 arch/x86/kvm/emulate.c if (seg == VCPU_SREG_CS || seg == VCPU_SREG_TR) VCPU_SREG_CS 1733 arch/x86/kvm/emulate.c case VCPU_SREG_CS: VCPU_SREG_CS 2093 arch/x86/kvm/emulate.c ctxt->src.val = get_segment_selector(ctxt, VCPU_SREG_CS); VCPU_SREG_CS 2116 arch/x86/kvm/emulate.c rc = load_segment_descriptor(ctxt, cs, VCPU_SREG_CS); VCPU_SREG_CS 2186 arch/x86/kvm/emulate.c rc = load_segment_descriptor(ctxt, (u16)cs, VCPU_SREG_CS); VCPU_SREG_CS 2231 arch/x86/kvm/emulate.c rc = __load_segment_descriptor(ctxt, sel, VCPU_SREG_CS, cpl, VCPU_SREG_CS 2313 arch/x86/kvm/emulate.c rc = __load_segment_descriptor(ctxt, (u16)cs, VCPU_SREG_CS, cpl, VCPU_SREG_CS 2650 arch/x86/kvm/emulate.c ctxt->ops->set_segment(ctxt, 0, &cs_desc, 0, VCPU_SREG_CS); VCPU_SREG_CS 2816 arch/x86/kvm/emulate.c ops->set_segment(ctxt, cs_sel, &cs, 0, VCPU_SREG_CS); VCPU_SREG_CS 2884 arch/x86/kvm/emulate.c ops->set_segment(ctxt, cs_sel, &cs, 0, VCPU_SREG_CS); VCPU_SREG_CS 2947 arch/x86/kvm/emulate.c ops->set_segment(ctxt, cs_sel, &cs, 0, VCPU_SREG_CS); VCPU_SREG_CS 3065 arch/x86/kvm/emulate.c tss->cs = get_segment_selector(ctxt, VCPU_SREG_CS); VCPU_SREG_CS 3094 arch/x86/kvm/emulate.c set_segment_selector(ctxt, tss->cs, VCPU_SREG_CS); VCPU_SREG_CS 3112 arch/x86/kvm/emulate.c ret = __load_segment_descriptor(ctxt, tss->cs, VCPU_SREG_CS, cpl, VCPU_SREG_CS 3179 arch/x86/kvm/emulate.c tss->cs = get_segment_selector(ctxt, VCPU_SREG_CS); VCPU_SREG_CS 3214 arch/x86/kvm/emulate.c set_segment_selector(ctxt, tss->cs, VCPU_SREG_CS); VCPU_SREG_CS 3245 arch/x86/kvm/emulate.c ret = __load_segment_descriptor(ctxt, tss->cs, VCPU_SREG_CS, cpl, VCPU_SREG_CS 3539 arch/x86/kvm/emulate.c ops->get_segment(ctxt, &old_cs, &old_desc, NULL, VCPU_SREG_CS); VCPU_SREG_CS 3542 arch/x86/kvm/emulate.c rc = __load_segment_descriptor(ctxt, sel, VCPU_SREG_CS, cpl, VCPU_SREG_CS 3566 arch/x86/kvm/emulate.c ops->set_segment(ctxt, old_cs, &old_desc, 0, VCPU_SREG_CS); VCPU_SREG_CS 3769 arch/x86/kvm/emulate.c if (ctxt->modrm_reg == VCPU_SREG_CS || ctxt->modrm_reg > VCPU_SREG_GS) VCPU_SREG_CS 5154 arch/x86/kvm/emulate.c op->val = VCPU_SREG_CS; VCPU_SREG_CS 5213 arch/x86/kvm/emulate.c ctxt->ops->get_segment(ctxt, &dummy, &desc, NULL, VCPU_SREG_CS); VCPU_SREG_CS 5258 arch/x86/kvm/emulate.c ctxt->seg_override = VCPU_SREG_CS; VCPU_SREG_CS 2438 arch/x86/kvm/svm.c case VCPU_SREG_CS: return &save->cs; VCPU_SREG_CS 747 arch/x86/kvm/trace.h __entry->csbase = kvm_x86_ops->get_segment_base(vcpu, VCPU_SREG_CS); VCPU_SREG_CS 3895 arch/x86/kvm/vmx/nested.c vmx_set_segment(vcpu, &seg, VCPU_SREG_CS); VCPU_SREG_CS 2635 arch/x86/kvm/vmx/vmx.c if (seg == VCPU_SREG_CS || seg == VCPU_SREG_SS) VCPU_SREG_CS 2657 arch/x86/kvm/vmx/vmx.c vmx_get_segment(vcpu, &vmx->rmode.segs[VCPU_SREG_CS], VCPU_SREG_CS); VCPU_SREG_CS 2675 arch/x86/kvm/vmx/vmx.c fix_pmode_seg(vcpu, VCPU_SREG_CS, &vmx->rmode.segs[VCPU_SREG_CS]); VCPU_SREG_CS 2689 arch/x86/kvm/vmx/vmx.c if (seg == VCPU_SREG_CS) VCPU_SREG_CS 2728 arch/x86/kvm/vmx/vmx.c vmx_get_segment(vcpu, &vmx->rmode.segs[VCPU_SREG_CS], VCPU_SREG_CS); VCPU_SREG_CS 2756 arch/x86/kvm/vmx/vmx.c fix_rmode_seg(VCPU_SREG_CS, &vmx->rmode.segs[VCPU_SREG_CS]); VCPU_SREG_CS 3200 arch/x86/kvm/vmx/vmx.c u32 ar = vmx_read_guest_seg_ar(to_vmx(vcpu), VCPU_SREG_CS); VCPU_SREG_CS 3237 arch/x86/kvm/vmx/vmx.c if (seg == VCPU_SREG_CS) VCPU_SREG_CS 3256 arch/x86/kvm/vmx/vmx.c vmx_get_segment(vcpu, &cs, VCPU_SREG_CS); VCPU_SREG_CS 3366 arch/x86/kvm/vmx/vmx.c vmx_get_segment(vcpu, &cs, VCPU_SREG_CS); VCPU_SREG_CS 3385 arch/x86/kvm/vmx/vmx.c if (!rmode_segment_valid(vcpu, VCPU_SREG_CS)) VCPU_SREG_CS 3511 arch/x86/kvm/vmx/vmx.c if (seg == VCPU_SREG_CS) VCPU_SREG_CS 4281 arch/x86/kvm/vmx/vmx.c seg_setup(VCPU_SREG_CS); VCPU_SREG_CS 7920 arch/x86/kvm/x86.c kvm_set_segment(vcpu, &cs, VCPU_SREG_CS); VCPU_SREG_CS 8716 arch/x86/kvm/x86.c kvm_get_segment(vcpu, &cs, VCPU_SREG_CS); VCPU_SREG_CS 8726 arch/x86/kvm/x86.c kvm_get_segment(vcpu, &sregs->cs, VCPU_SREG_CS); VCPU_SREG_CS 8927 arch/x86/kvm/x86.c kvm_set_segment(vcpu, &sregs->cs, VCPU_SREG_CS); VCPU_SREG_CS 9003 arch/x86/kvm/x86.c get_segment_base(vcpu, VCPU_SREG_CS); VCPU_SREG_CS 9289 arch/x86/kvm/x86.c kvm_get_segment(vcpu, &cs, VCPU_SREG_CS); VCPU_SREG_CS 9292 arch/x86/kvm/x86.c kvm_set_segment(vcpu, &cs, VCPU_SREG_CS); VCPU_SREG_CS 10026 arch/x86/kvm/x86.c return (u32)(get_segment_base(vcpu, VCPU_SREG_CS) +