VCPU_SREG_FS 603 arch/x86/kvm/emulate.c if (ctxt->mode == X86EMUL_MODE_PROT64 && seg < VCPU_SREG_FS) VCPU_SREG_FS 3182 arch/x86/kvm/emulate.c tss->fs = get_segment_selector(ctxt, VCPU_SREG_FS); VCPU_SREG_FS 3217 arch/x86/kvm/emulate.c set_segment_selector(ctxt, tss->fs, VCPU_SREG_FS); VCPU_SREG_FS 3257 arch/x86/kvm/emulate.c ret = __load_segment_descriptor(ctxt, tss->fs, VCPU_SREG_FS, cpl, VCPU_SREG_FS 5166 arch/x86/kvm/emulate.c op->val = VCPU_SREG_FS; VCPU_SREG_FS 5270 arch/x86/kvm/emulate.c ctxt->seg_override = VCPU_SREG_FS; VCPU_SREG_FS 2441 arch/x86/kvm/svm.c case VCPU_SREG_FS: return &save->fs; VCPU_SREG_FS 2500 arch/x86/kvm/svm.c case VCPU_SREG_FS: VCPU_SREG_FS 3913 arch/x86/kvm/vmx/nested.c vmx_set_segment(vcpu, &seg, VCPU_SREG_FS); VCPU_SREG_FS 4268 arch/x86/kvm/vmx/nested.c if (seg_reg == VCPU_SREG_FS || seg_reg == VCPU_SREG_GS) VCPU_SREG_FS 2654 arch/x86/kvm/vmx/vmx.c vmx_get_segment(vcpu, &vmx->rmode.segs[VCPU_SREG_FS], VCPU_SREG_FS); VCPU_SREG_FS 2679 arch/x86/kvm/vmx/vmx.c fix_pmode_seg(vcpu, VCPU_SREG_FS, &vmx->rmode.segs[VCPU_SREG_FS]); VCPU_SREG_FS 2725 arch/x86/kvm/vmx/vmx.c vmx_get_segment(vcpu, &vmx->rmode.segs[VCPU_SREG_FS], VCPU_SREG_FS); VCPU_SREG_FS 2760 arch/x86/kvm/vmx/vmx.c fix_rmode_seg(VCPU_SREG_FS, &vmx->rmode.segs[VCPU_SREG_FS]); VCPU_SREG_FS 3393 arch/x86/kvm/vmx/vmx.c if (!rmode_segment_valid(vcpu, VCPU_SREG_FS)) VCPU_SREG_FS 3409 arch/x86/kvm/vmx/vmx.c if (!data_segment_valid(vcpu, VCPU_SREG_FS)) VCPU_SREG_FS 4287 arch/x86/kvm/vmx/vmx.c seg_setup(VCPU_SREG_FS); VCPU_SREG_FS 7923 arch/x86/kvm/x86.c kvm_set_segment(vcpu, &ds, VCPU_SREG_FS); VCPU_SREG_FS 8729 arch/x86/kvm/x86.c kvm_get_segment(vcpu, &sregs->fs, VCPU_SREG_FS); VCPU_SREG_FS 8930 arch/x86/kvm/x86.c kvm_set_segment(vcpu, &sregs->fs, VCPU_SREG_FS);