VCPU_SREG_CS      785 arch/x86/kvm/emulate.c 	struct segmented_address addr = { .seg = VCPU_SREG_CS,
VCPU_SREG_CS      883 arch/x86/kvm/emulate.c 	struct segmented_address addr = { .seg = VCPU_SREG_CS,
VCPU_SREG_CS     1679 arch/x86/kvm/emulate.c 		if (seg == VCPU_SREG_CS || seg == VCPU_SREG_TR)
VCPU_SREG_CS     1733 arch/x86/kvm/emulate.c 	case VCPU_SREG_CS:
VCPU_SREG_CS     2093 arch/x86/kvm/emulate.c 	ctxt->src.val = get_segment_selector(ctxt, VCPU_SREG_CS);
VCPU_SREG_CS     2116 arch/x86/kvm/emulate.c 	rc = load_segment_descriptor(ctxt, cs, VCPU_SREG_CS);
VCPU_SREG_CS     2186 arch/x86/kvm/emulate.c 	rc = load_segment_descriptor(ctxt, (u16)cs, VCPU_SREG_CS);
VCPU_SREG_CS     2231 arch/x86/kvm/emulate.c 	rc = __load_segment_descriptor(ctxt, sel, VCPU_SREG_CS, cpl,
VCPU_SREG_CS     2313 arch/x86/kvm/emulate.c 	rc = __load_segment_descriptor(ctxt, (u16)cs, VCPU_SREG_CS, cpl,
VCPU_SREG_CS     2650 arch/x86/kvm/emulate.c 		ctxt->ops->set_segment(ctxt, 0, &cs_desc, 0, VCPU_SREG_CS);
VCPU_SREG_CS     2816 arch/x86/kvm/emulate.c 	ops->set_segment(ctxt, cs_sel, &cs, 0, VCPU_SREG_CS);
VCPU_SREG_CS     2884 arch/x86/kvm/emulate.c 	ops->set_segment(ctxt, cs_sel, &cs, 0, VCPU_SREG_CS);
VCPU_SREG_CS     2947 arch/x86/kvm/emulate.c 	ops->set_segment(ctxt, cs_sel, &cs, 0, VCPU_SREG_CS);
VCPU_SREG_CS     3065 arch/x86/kvm/emulate.c 	tss->cs = get_segment_selector(ctxt, VCPU_SREG_CS);
VCPU_SREG_CS     3094 arch/x86/kvm/emulate.c 	set_segment_selector(ctxt, tss->cs, VCPU_SREG_CS);
VCPU_SREG_CS     3112 arch/x86/kvm/emulate.c 	ret = __load_segment_descriptor(ctxt, tss->cs, VCPU_SREG_CS, cpl,
VCPU_SREG_CS     3179 arch/x86/kvm/emulate.c 	tss->cs = get_segment_selector(ctxt, VCPU_SREG_CS);
VCPU_SREG_CS     3214 arch/x86/kvm/emulate.c 	set_segment_selector(ctxt, tss->cs, VCPU_SREG_CS);
VCPU_SREG_CS     3245 arch/x86/kvm/emulate.c 	ret = __load_segment_descriptor(ctxt, tss->cs, VCPU_SREG_CS, cpl,
VCPU_SREG_CS     3539 arch/x86/kvm/emulate.c 	ops->get_segment(ctxt, &old_cs, &old_desc, NULL, VCPU_SREG_CS);
VCPU_SREG_CS     3542 arch/x86/kvm/emulate.c 	rc = __load_segment_descriptor(ctxt, sel, VCPU_SREG_CS, cpl,
VCPU_SREG_CS     3566 arch/x86/kvm/emulate.c 	ops->set_segment(ctxt, old_cs, &old_desc, 0, VCPU_SREG_CS);
VCPU_SREG_CS     3769 arch/x86/kvm/emulate.c 	if (ctxt->modrm_reg == VCPU_SREG_CS || ctxt->modrm_reg > VCPU_SREG_GS)
VCPU_SREG_CS     5154 arch/x86/kvm/emulate.c 		op->val = VCPU_SREG_CS;
VCPU_SREG_CS     5213 arch/x86/kvm/emulate.c 		ctxt->ops->get_segment(ctxt, &dummy, &desc, NULL, VCPU_SREG_CS);
VCPU_SREG_CS     5258 arch/x86/kvm/emulate.c 			ctxt->seg_override = VCPU_SREG_CS;
VCPU_SREG_CS     2438 arch/x86/kvm/svm.c 	case VCPU_SREG_CS: return &save->cs;
VCPU_SREG_CS      747 arch/x86/kvm/trace.h 		__entry->csbase = kvm_x86_ops->get_segment_base(vcpu, VCPU_SREG_CS);
VCPU_SREG_CS     3895 arch/x86/kvm/vmx/nested.c 	vmx_set_segment(vcpu, &seg, VCPU_SREG_CS);
VCPU_SREG_CS     2635 arch/x86/kvm/vmx/vmx.c 		if (seg == VCPU_SREG_CS || seg == VCPU_SREG_SS)
VCPU_SREG_CS     2657 arch/x86/kvm/vmx/vmx.c 	vmx_get_segment(vcpu, &vmx->rmode.segs[VCPU_SREG_CS], VCPU_SREG_CS);
VCPU_SREG_CS     2675 arch/x86/kvm/vmx/vmx.c 	fix_pmode_seg(vcpu, VCPU_SREG_CS, &vmx->rmode.segs[VCPU_SREG_CS]);
VCPU_SREG_CS     2689 arch/x86/kvm/vmx/vmx.c 	if (seg == VCPU_SREG_CS)
VCPU_SREG_CS     2728 arch/x86/kvm/vmx/vmx.c 	vmx_get_segment(vcpu, &vmx->rmode.segs[VCPU_SREG_CS], VCPU_SREG_CS);
VCPU_SREG_CS     2756 arch/x86/kvm/vmx/vmx.c 	fix_rmode_seg(VCPU_SREG_CS, &vmx->rmode.segs[VCPU_SREG_CS]);
VCPU_SREG_CS     3200 arch/x86/kvm/vmx/vmx.c 	u32 ar = vmx_read_guest_seg_ar(to_vmx(vcpu), VCPU_SREG_CS);
VCPU_SREG_CS     3237 arch/x86/kvm/vmx/vmx.c 	if (seg == VCPU_SREG_CS)
VCPU_SREG_CS     3256 arch/x86/kvm/vmx/vmx.c 	vmx_get_segment(vcpu, &cs, VCPU_SREG_CS);
VCPU_SREG_CS     3366 arch/x86/kvm/vmx/vmx.c 	vmx_get_segment(vcpu, &cs, VCPU_SREG_CS);
VCPU_SREG_CS     3385 arch/x86/kvm/vmx/vmx.c 		if (!rmode_segment_valid(vcpu, VCPU_SREG_CS))
VCPU_SREG_CS     3511 arch/x86/kvm/vmx/vmx.c 	if (seg == VCPU_SREG_CS)
VCPU_SREG_CS     4281 arch/x86/kvm/vmx/vmx.c 	seg_setup(VCPU_SREG_CS);
VCPU_SREG_CS     7920 arch/x86/kvm/x86.c 	kvm_set_segment(vcpu, &cs, VCPU_SREG_CS);
VCPU_SREG_CS     8716 arch/x86/kvm/x86.c 	kvm_get_segment(vcpu, &cs, VCPU_SREG_CS);
VCPU_SREG_CS     8726 arch/x86/kvm/x86.c 	kvm_get_segment(vcpu, &sregs->cs, VCPU_SREG_CS);
VCPU_SREG_CS     8927 arch/x86/kvm/x86.c 	kvm_set_segment(vcpu, &sregs->cs, VCPU_SREG_CS);
VCPU_SREG_CS     9003 arch/x86/kvm/x86.c 			get_segment_base(vcpu, VCPU_SREG_CS);
VCPU_SREG_CS     9289 arch/x86/kvm/x86.c 	kvm_get_segment(vcpu, &cs, VCPU_SREG_CS);
VCPU_SREG_CS     9292 arch/x86/kvm/x86.c 	kvm_set_segment(vcpu, &cs, VCPU_SREG_CS);
VCPU_SREG_CS     10026 arch/x86/kvm/x86.c 	return (u32)(get_segment_base(vcpu, VCPU_SREG_CS) +