VCPU_SREG_FS      603 arch/x86/kvm/emulate.c 	if (ctxt->mode == X86EMUL_MODE_PROT64 && seg < VCPU_SREG_FS)
VCPU_SREG_FS     3182 arch/x86/kvm/emulate.c 	tss->fs = get_segment_selector(ctxt, VCPU_SREG_FS);
VCPU_SREG_FS     3217 arch/x86/kvm/emulate.c 	set_segment_selector(ctxt, tss->fs, VCPU_SREG_FS);
VCPU_SREG_FS     3257 arch/x86/kvm/emulate.c 	ret = __load_segment_descriptor(ctxt, tss->fs, VCPU_SREG_FS, cpl,
VCPU_SREG_FS     5166 arch/x86/kvm/emulate.c 		op->val = VCPU_SREG_FS;
VCPU_SREG_FS     5270 arch/x86/kvm/emulate.c 			ctxt->seg_override = VCPU_SREG_FS;
VCPU_SREG_FS     2441 arch/x86/kvm/svm.c 	case VCPU_SREG_FS: return &save->fs;
VCPU_SREG_FS     2500 arch/x86/kvm/svm.c 	case VCPU_SREG_FS:
VCPU_SREG_FS     3913 arch/x86/kvm/vmx/nested.c 	vmx_set_segment(vcpu, &seg, VCPU_SREG_FS);
VCPU_SREG_FS     4268 arch/x86/kvm/vmx/nested.c 		if (seg_reg == VCPU_SREG_FS || seg_reg == VCPU_SREG_GS)
VCPU_SREG_FS     2654 arch/x86/kvm/vmx/vmx.c 	vmx_get_segment(vcpu, &vmx->rmode.segs[VCPU_SREG_FS], VCPU_SREG_FS);
VCPU_SREG_FS     2679 arch/x86/kvm/vmx/vmx.c 	fix_pmode_seg(vcpu, VCPU_SREG_FS, &vmx->rmode.segs[VCPU_SREG_FS]);
VCPU_SREG_FS     2725 arch/x86/kvm/vmx/vmx.c 	vmx_get_segment(vcpu, &vmx->rmode.segs[VCPU_SREG_FS], VCPU_SREG_FS);
VCPU_SREG_FS     2760 arch/x86/kvm/vmx/vmx.c 	fix_rmode_seg(VCPU_SREG_FS, &vmx->rmode.segs[VCPU_SREG_FS]);
VCPU_SREG_FS     3393 arch/x86/kvm/vmx/vmx.c 		if (!rmode_segment_valid(vcpu, VCPU_SREG_FS))
VCPU_SREG_FS     3409 arch/x86/kvm/vmx/vmx.c 		if (!data_segment_valid(vcpu, VCPU_SREG_FS))
VCPU_SREG_FS     4287 arch/x86/kvm/vmx/vmx.c 	seg_setup(VCPU_SREG_FS);
VCPU_SREG_FS     7923 arch/x86/kvm/x86.c 	kvm_set_segment(vcpu, &ds, VCPU_SREG_FS);
VCPU_SREG_FS     8729 arch/x86/kvm/x86.c 	kvm_get_segment(vcpu, &sregs->fs, VCPU_SREG_FS);
VCPU_SREG_FS     8930 arch/x86/kvm/x86.c 	kvm_set_segment(vcpu, &sregs->fs, VCPU_SREG_FS);