cr4 26 arch/sh/include/cpu-sh5/cpu/registers.h #define INTEVT cr4 cr4 550 arch/x86/include/asm/kvm_host.h unsigned long cr4; cr4 1049 arch/x86/include/asm/kvm_host.h int (*set_cr4)(struct kvm_vcpu *vcpu, unsigned long cr4); cr4 1384 arch/x86/include/asm/kvm_host.h int kvm_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4); cr4 49 arch/x86/include/asm/realmode.h u32 cr4; cr4 22 arch/x86/include/asm/suspend_32.h unsigned long cr0, cr2, cr3, cr4; cr4 37 arch/x86/include/asm/suspend_64.h unsigned long cr0, cr2, cr3, cr4; cr4 177 arch/x86/include/asm/svm.h u64 cr4; cr4 220 arch/x86/include/asm/tlbflush.h unsigned long cr4; cr4 282 arch/x86/include/asm/tlbflush.h this_cpu_write(cpu_tlbstate.cr4, __read_cr4()); cr4 285 arch/x86/include/asm/tlbflush.h static inline void __cr4_set(unsigned long cr4) cr4 288 arch/x86/include/asm/tlbflush.h this_cpu_write(cpu_tlbstate.cr4, cr4); cr4 289 arch/x86/include/asm/tlbflush.h __write_cr4(cr4); cr4 295 arch/x86/include/asm/tlbflush.h unsigned long cr4; cr4 297 arch/x86/include/asm/tlbflush.h cr4 = this_cpu_read(cpu_tlbstate.cr4); cr4 298 arch/x86/include/asm/tlbflush.h if ((cr4 | mask) != cr4) cr4 299 arch/x86/include/asm/tlbflush.h __cr4_set(cr4 | mask); cr4 305 arch/x86/include/asm/tlbflush.h unsigned long cr4; cr4 307 arch/x86/include/asm/tlbflush.h cr4 = this_cpu_read(cpu_tlbstate.cr4); cr4 308 arch/x86/include/asm/tlbflush.h if ((cr4 & ~mask) != cr4) cr4 309 arch/x86/include/asm/tlbflush.h __cr4_set(cr4 & ~mask); cr4 334 arch/x86/include/asm/tlbflush.h unsigned long cr4; cr4 336 arch/x86/include/asm/tlbflush.h cr4 = this_cpu_read(cpu_tlbstate.cr4); cr4 337 arch/x86/include/asm/tlbflush.h __cr4_set(cr4 ^ mask); cr4 343 arch/x86/include/asm/tlbflush.h return this_cpu_read(cpu_tlbstate.cr4); cr4 422 arch/x86/include/asm/tlbflush.h unsigned long cr4, flags; cr4 442 arch/x86/include/asm/tlbflush.h cr4 = this_cpu_read(cpu_tlbstate.cr4); cr4 444 arch/x86/include/asm/tlbflush.h native_write_cr4(cr4 ^ X86_CR4_PGE); cr4 446 arch/x86/include/asm/tlbflush.h native_write_cr4(cr4); cr4 154 arch/x86/include/uapi/asm/kvm.h __u64 cr0, cr2, cr3, cr4, cr8; cr4 79 arch/x86/kernel/asm-offsets_64.c ENTRY(cr4); cr4 413 arch/x86/kernel/cpu/common.c unsigned long cr4 = __read_cr4(); cr4 416 arch/x86/kernel/cpu/common.c cr4 |= X86_CR4_PCIDE; cr4 418 arch/x86/kernel/cpu/common.c cr4 |= cr4_pinned_bits; cr4 420 arch/x86/kernel/cpu/common.c __write_cr4(cr4); cr4 423 arch/x86/kernel/cpu/common.c this_cpu_write(cpu_tlbstate.cr4, cr4); cr4 436 arch/x86/kernel/cpu/common.c cr4_pinned_bits = this_cpu_read(cpu_tlbstate.cr4) & mask; cr4 135 arch/x86/kernel/cpu/mtrr/cyrix.c static u32 cr4, ccr3; cr4 143 arch/x86/kernel/cpu/mtrr/cyrix.c cr4 = __read_cr4(); cr4 144 arch/x86/kernel/cpu/mtrr/cyrix.c __write_cr4(cr4 & ~X86_CR4_PGE); cr4 176 arch/x86/kernel/cpu/mtrr/cyrix.c __write_cr4(cr4); cr4 720 arch/x86/kernel/cpu/mtrr/generic.c static unsigned long cr4; cr4 758 arch/x86/kernel/cpu/mtrr/generic.c cr4 = __read_cr4(); cr4 759 arch/x86/kernel/cpu/mtrr/generic.c __write_cr4(cr4 & ~X86_CR4_PGE); cr4 791 arch/x86/kernel/cpu/mtrr/generic.c __write_cr4(cr4); cr4 63 arch/x86/kernel/process_32.c unsigned long cr0 = 0L, cr2 = 0L, cr3 = 0L, cr4 = 0L; cr4 87 arch/x86/kernel/process_32.c cr4 = __read_cr4(); cr4 89 arch/x86/kernel/process_32.c cr0, cr2, cr3, cr4); cr4 69 arch/x86/kernel/process_64.c unsigned long cr0 = 0L, cr2 = 0L, cr3 = 0L, cr4 = 0L, fs, gs, shadowgs; cr4 115 arch/x86/kernel/process_64.c cr4 = __read_cr4(); cr4 122 arch/x86/kernel/process_64.c cr4); cr4 2447 arch/x86/kvm/emulate.c u64 cr0, u64 cr3, u64 cr4) cr4 2454 arch/x86/kvm/emulate.c if (cr4 & X86_CR4_PCIDE) { cr4 2468 arch/x86/kvm/emulate.c bad = ctxt->ops->set_cr(ctxt, 4, cr4 & ~X86_CR4_PCIDE); cr4 2476 arch/x86/kvm/emulate.c if (cr4 & X86_CR4_PCIDE) { cr4 2477 arch/x86/kvm/emulate.c bad = ctxt->ops->set_cr(ctxt, 4, cr4); cr4 2497 arch/x86/kvm/emulate.c u32 val, cr0, cr3, cr4; cr4 2539 arch/x86/kvm/emulate.c cr4 = GET_SMSTATE(u32, smstate, 0x7f14); cr4 2543 arch/x86/kvm/emulate.c return rsm_enter_protected_mode(ctxt, cr0, cr3, cr4); cr4 2552 arch/x86/kvm/emulate.c u64 val, cr0, cr3, cr4; cr4 2570 arch/x86/kvm/emulate.c cr4 = GET_SMSTATE(u64, smstate, 0x7f48); cr4 2597 arch/x86/kvm/emulate.c r = rsm_enter_protected_mode(ctxt, cr0, cr3, cr4); cr4 2613 arch/x86/kvm/emulate.c unsigned long cr0, cr4, efer; cr4 2642 arch/x86/kvm/emulate.c cr4 = ctxt->ops->get_cr(ctxt, 4); cr4 2643 arch/x86/kvm/emulate.c if (cr4 & X86_CR4_PCIDE) cr4 2644 arch/x86/kvm/emulate.c ctxt->ops->set_cr(ctxt, 4, cr4 & ~X86_CR4_PCIDE); cr4 2660 arch/x86/kvm/emulate.c cr4 = ctxt->ops->get_cr(ctxt, 4); cr4 2661 arch/x86/kvm/emulate.c if (cr4 & X86_CR4_PAE) cr4 2662 arch/x86/kvm/emulate.c ctxt->ops->set_cr(ctxt, 4, cr4 & ~X86_CR4_PAE); cr4 4252 arch/x86/kvm/emulate.c u64 cr4; cr4 4257 arch/x86/kvm/emulate.c cr4 = ctxt->ops->get_cr(ctxt, 4); cr4 4261 arch/x86/kvm/emulate.c !(cr4 & X86_CR4_PAE)) cr4 4317 arch/x86/kvm/emulate.c u64 cr4; cr4 4322 arch/x86/kvm/emulate.c cr4 = ctxt->ops->get_cr(ctxt, 4); cr4 4323 arch/x86/kvm/emulate.c if ((cr4 & X86_CR4_DE) && (dr == 4 || dr == 5)) cr4 4375 arch/x86/kvm/emulate.c u64 cr4 = ctxt->ops->get_cr(ctxt, 4); cr4 4377 arch/x86/kvm/emulate.c if (cr4 & X86_CR4_TSD && ctxt->ops->cpl(ctxt)) cr4 4385 arch/x86/kvm/emulate.c u64 cr4 = ctxt->ops->get_cr(ctxt, 4); cr4 4395 arch/x86/kvm/emulate.c if ((!(cr4 & X86_CR4_PCE) && ctxt->ops->cpl(ctxt)) || cr4 107 arch/x86/kvm/kvm_cache_regs.h return vcpu->arch.cr4 & mask; cr4 1638 arch/x86/kvm/svm.c save->cr4 = X86_CR4_PAE; cr4 1650 arch/x86/kvm/svm.c save->cr4 = 0; cr4 2633 arch/x86/kvm/svm.c static int svm_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4) cr4 2636 arch/x86/kvm/svm.c unsigned long old_cr4 = to_svm(vcpu)->vmcb->save.cr4; cr4 2638 arch/x86/kvm/svm.c if (cr4 & X86_CR4_VMXE) cr4 2641 arch/x86/kvm/svm.c if (npt_enabled && ((old_cr4 ^ cr4) & X86_CR4_PGE)) cr4 2644 arch/x86/kvm/svm.c vcpu->arch.cr4 = cr4; cr4 2646 arch/x86/kvm/svm.c cr4 |= X86_CR4_PAE; cr4 2647 arch/x86/kvm/svm.c cr4 |= host_cr4_mce; cr4 2648 arch/x86/kvm/svm.c to_svm(vcpu)->vmcb->save.cr4 = cr4; cr4 3391 arch/x86/kvm/svm.c nested_vmcb->save.cr4 = svm->vcpu.arch.cr4; cr4 3460 arch/x86/kvm/svm.c svm_set_cr4(&svm->vcpu, hsave->save.cr4); cr4 3564 arch/x86/kvm/svm.c svm_set_cr4(&svm->vcpu, nested_vmcb->save.cr4); cr4 3705 arch/x86/kvm/svm.c hsave->save.cr4 = svm->vcpu.arch.cr4; cr4 4942 arch/x86/kvm/svm.c "cr3:", save->cr3, "cr4:", save->cr4); cr4 7166 arch/x86/kvm/svm.c unsigned long cr4 = kvm_read_cr4(vcpu); cr4 7167 arch/x86/kvm/svm.c bool smep = cr4 & X86_CR4_SMEP; cr4 7168 arch/x86/kvm/svm.c bool smap = cr4 & X86_CR4_SMAP; cr4 2826 arch/x86/kvm/vmx/nested.c unsigned long cr3, cr4; cr4 2855 arch/x86/kvm/vmx/nested.c cr4 = cr4_read_shadow(); cr4 2856 arch/x86/kvm/vmx/nested.c if (unlikely(cr4 != vmx->loaded_vmcs->host_state.cr4)) { cr4 2857 arch/x86/kvm/vmx/nested.c vmcs_writel(HOST_CR4, cr4); cr4 2858 arch/x86/kvm/vmx/nested.c vmx->loaded_vmcs->host_state.cr4 = cr4; cr4 34 arch/x86/kvm/vmx/vmcs.h unsigned long cr4; /* May not match real cr4 */ cr4 2845 arch/x86/kvm/vmx/vmx.c vcpu->arch.cr4 &= ~cr4_guest_owned_bits; cr4 2846 arch/x86/kvm/vmx/vmx.c vcpu->arch.cr4 |= vmcs_readl(GUEST_CR4) & cr4_guest_owned_bits; cr4 3004 arch/x86/kvm/vmx/vmx.c int vmx_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4) cr4 3014 arch/x86/kvm/vmx/vmx.c hw_cr4 = (cr4_read_shadow() & X86_CR4_MCE) | (cr4 & ~X86_CR4_MCE); cr4 3023 arch/x86/kvm/vmx/vmx.c if (cr4 & X86_CR4_UMIP) { cr4 3032 arch/x86/kvm/vmx/vmx.c if (cr4 & X86_CR4_VMXE) { cr4 3044 arch/x86/kvm/vmx/vmx.c if (vmx->nested.vmxon && !nested_cr4_valid(vcpu, cr4)) cr4 3047 arch/x86/kvm/vmx/vmx.c vcpu->arch.cr4 = cr4; cr4 3054 arch/x86/kvm/vmx/vmx.c } else if (!(cr4 & X86_CR4_PAE)) { cr4 3074 arch/x86/kvm/vmx/vmx.c vmcs_writel(CR4_READ_SHADOW, cr4); cr4 3861 arch/x86/kvm/vmx/vmx.c unsigned long cr0, cr3, cr4; cr4 3876 arch/x86/kvm/vmx/vmx.c cr4 = cr4_read_shadow(); cr4 3877 arch/x86/kvm/vmx/vmx.c vmcs_writel(HOST_CR4, cr4); /* 22.2.3, 22.2.5 */ cr4 3878 arch/x86/kvm/vmx/vmx.c vmx->loaded_vmcs->host_state.cr4 = cr4; cr4 4797 arch/x86/kvm/vmx/vmx.c WARN_ON(!(vcpu->arch.cr4 & X86_CR4_UMIP)); cr4 5685 arch/x86/kvm/vmx/vmx.c unsigned long cr4; cr4 5698 arch/x86/kvm/vmx/vmx.c cr4 = vmcs_readl(GUEST_CR4); cr4 5709 arch/x86/kvm/vmx/vmx.c cr4, vmcs_readl(CR4_READ_SHADOW), vmcs_readl(CR4_GUEST_HOST_MASK)); cr4 5712 arch/x86/kvm/vmx/vmx.c (cr4 & X86_CR4_PAE) && !(efer & EFER_LMA)) cr4 6486 arch/x86/kvm/vmx/vmx.c unsigned long cr3, cr4; cr4 6517 arch/x86/kvm/vmx/vmx.c cr4 = cr4_read_shadow(); cr4 6518 arch/x86/kvm/vmx/vmx.c if (unlikely(cr4 != vmx->loaded_vmcs->host_state.cr4)) { cr4 6519 arch/x86/kvm/vmx/vmx.c vmcs_writel(HOST_CR4, cr4); cr4 6520 arch/x86/kvm/vmx/vmx.c vmx->loaded_vmcs->host_state.cr4 = cr4; cr4 324 arch/x86/kvm/vmx/vmx.h int vmx_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4); cr4 942 arch/x86/kvm/x86.c static int kvm_valid_cr4(struct kvm_vcpu *vcpu, unsigned long cr4) cr4 944 arch/x86/kvm/x86.c if (cr4 & cr4_reserved_bits) cr4 947 arch/x86/kvm/x86.c if (!guest_cpuid_has(vcpu, X86_FEATURE_XSAVE) && (cr4 & X86_CR4_OSXSAVE)) cr4 950 arch/x86/kvm/x86.c if (!guest_cpuid_has(vcpu, X86_FEATURE_SMEP) && (cr4 & X86_CR4_SMEP)) cr4 953 arch/x86/kvm/x86.c if (!guest_cpuid_has(vcpu, X86_FEATURE_SMAP) && (cr4 & X86_CR4_SMAP)) cr4 956 arch/x86/kvm/x86.c if (!guest_cpuid_has(vcpu, X86_FEATURE_FSGSBASE) && (cr4 & X86_CR4_FSGSBASE)) cr4 959 arch/x86/kvm/x86.c if (!guest_cpuid_has(vcpu, X86_FEATURE_PKU) && (cr4 & X86_CR4_PKE)) cr4 962 arch/x86/kvm/x86.c if (!guest_cpuid_has(vcpu, X86_FEATURE_LA57) && (cr4 & X86_CR4_LA57)) cr4 965 arch/x86/kvm/x86.c if (!guest_cpuid_has(vcpu, X86_FEATURE_UMIP) && (cr4 & X86_CR4_UMIP)) cr4 971 arch/x86/kvm/x86.c int kvm_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4) cr4 977 arch/x86/kvm/x86.c if (kvm_valid_cr4(vcpu, cr4)) cr4 981 arch/x86/kvm/x86.c if (!(cr4 & X86_CR4_PAE)) cr4 983 arch/x86/kvm/x86.c } else if (is_paging(vcpu) && (cr4 & X86_CR4_PAE) cr4 984 arch/x86/kvm/x86.c && ((cr4 ^ old_cr4) & pdptr_bits) cr4 989 arch/x86/kvm/x86.c if ((cr4 & X86_CR4_PCIDE) && !(old_cr4 & X86_CR4_PCIDE)) { cr4 998 arch/x86/kvm/x86.c if (kvm_x86_ops->set_cr4(vcpu, cr4)) cr4 1001 arch/x86/kvm/x86.c if (((cr4 ^ old_cr4) & pdptr_bits) || cr4 1002 arch/x86/kvm/x86.c (!(cr4 & X86_CR4_PCIDE) && (old_cr4 & X86_CR4_PCIDE))) cr4 1005 arch/x86/kvm/x86.c if ((cr4 ^ old_cr4) & (X86_CR4_OSXSAVE | X86_CR4_PKE)) cr4 8746 arch/x86/kvm/x86.c sregs->cr4 = kvm_read_cr4(vcpu); cr4 8849 arch/x86/kvm/x86.c if (!(sregs->cr4 & X86_CR4_PAE) cr4 8861 arch/x86/kvm/x86.c return kvm_valid_cr4(vcpu, sregs->cr4); cr4 8902 arch/x86/kvm/x86.c mmu_reset_needed |= kvm_read_cr4(vcpu) != sregs->cr4; cr4 8903 arch/x86/kvm/x86.c cpuid_update_needed |= ((kvm_read_cr4(vcpu) ^ sregs->cr4) & cr4 8905 arch/x86/kvm/x86.c kvm_x86_ops->set_cr4(vcpu, sregs->cr4); cr4 957 arch/x86/mm/init.c .cr4 = ~0UL, /* fail hard if we screw up cr4 shadow initialization */ cr4 125 arch/x86/power/cpu.c ctxt->cr4 = __read_cr4(); cr4 203 arch/x86/power/cpu.c if (ctxt->cr4) cr4 204 arch/x86/power/cpu.c __write_cr4(ctxt->cr4); cr4 208 arch/x86/power/cpu.c __write_cr4(ctxt->cr4); cr4 101 arch/x86/realmode/init.c trampoline_cr4_features = &trampoline_header->cr4; cr4 882 arch/x86/xen/enlighten_pv.c static void xen_write_cr4(unsigned long cr4) cr4 884 arch/x86/xen/enlighten_pv.c cr4 &= ~(X86_CR4_PGE | X86_CR4_PSE | X86_CR4_PCE); cr4 886 arch/x86/xen/enlighten_pv.c native_write_cr4(cr4); cr4 281 drivers/misc/lkdtm/bugs.c unsigned long cr4; cr4 284 drivers/misc/lkdtm/bugs.c cr4 = native_read_cr4(); cr4 286 drivers/misc/lkdtm/bugs.c if ((cr4 & X86_CR4_SMEP) != X86_CR4_SMEP) { cr4 290 drivers/misc/lkdtm/bugs.c cr4 &= ~(X86_CR4_SMEP); cr4 293 drivers/misc/lkdtm/bugs.c native_write_cr4(cr4); cr4 294 drivers/misc/lkdtm/bugs.c if (cr4 == native_read_cr4()) { cr4 296 drivers/misc/lkdtm/bugs.c cr4 |= X86_CR4_SMEP; cr4 298 drivers/misc/lkdtm/bugs.c native_write_cr4(cr4); cr4 327 drivers/misc/lkdtm/bugs.c direct_write_cr4(cr4); cr4 332 drivers/misc/lkdtm/bugs.c cr4 |= X86_CR4_SMEP; cr4 334 drivers/misc/lkdtm/bugs.c native_write_cr4(cr4); cr4 641 drivers/net/wireless/broadcom/brcm80211/brcmfmac/chip.c static u32 brcmf_chip_tcm_ramsize(struct brcmf_core_priv *cr4) cr4 651 drivers/net/wireless/broadcom/brcm80211/brcmfmac/chip.c corecap = brcmf_chip_core_read32(cr4, ARMCR4_CAP); cr4 658 drivers/net/wireless/broadcom/brcm80211/brcmfmac/chip.c brcmf_chip_core_write32(cr4, ARMCR4_BANKIDX, idx); cr4 659 drivers/net/wireless/broadcom/brcm80211/brcmfmac/chip.c bxinfo = brcmf_chip_core_read32(cr4, ARMCR4_BANKINFO); cr4 959 drivers/parport/parport_pc.c int cr1, cr4, cra, cr23, cr26, cr27; cr4 973 drivers/parport/parport_pc.c cr4 = inb(io + 1); cr4 988 drivers/parport/parport_pc.c cr1, cr4, cra, cr23, cr26, cr27); cr4 1005 drivers/parport/parport_pc.c : modes[cr4 & 0x03], cr4 1006 drivers/parport/parport_pc.c (cr4 & 0x40) ? "1.7" : "1.9"); cr4 1633 drivers/tty/serial/fsl_lpuart.c unsigned char cr1, old_cr1, old_cr2, cr3, cr4, bdh, modem; cr4 1641 drivers/tty/serial/fsl_lpuart.c cr4 = readb(sport->port.membase + UARTCR4); cr4 1760 drivers/tty/serial/fsl_lpuart.c cr4 &= ~UARTCR4_BRFA_MASK; cr4 1762 drivers/tty/serial/fsl_lpuart.c writeb(cr4 | brfa, sport->port.membase + UARTCR4); cr4 6364 drivers/video/fbdev/sis/init301.c unsigned char cr4, cr14, cr5, cr15; cr4 6366 drivers/video/fbdev/sis/init301.c cr4 = SiS_Pr->CCRT1CRTC[4]; cr4 6371 drivers/video/fbdev/sis/init301.c cr4 = SiS_Pr->SiS_CRT1Table[CRT1Index].CR[4]; cr4 6376 drivers/video/fbdev/sis/init301.c tempbx = ((cr4 | ((cr14 & 0xC0) << 2)) - 3) << 3; /* (VGAHRS-3)*8 */ cr4 42 include/xen/interface/hvm/hvm_vcpu.h uint32_t cr4; cr4 105 include/xen/interface/hvm/hvm_vcpu.h uint64_t cr4; cr4 154 tools/arch/x86/include/uapi/asm/kvm.h __u64 cr0, cr2, cr3, cr4, cr8; cr4 209 tools/testing/selftests/kvm/include/x86_64/processor.h uint64_t cr4; cr4 212 tools/testing/selftests/kvm/include/x86_64/processor.h : /* output */ [cr4]"=r"(cr4)); cr4 213 tools/testing/selftests/kvm/include/x86_64/processor.h return cr4; cr4 216 tools/testing/selftests/kvm/lib/x86_64/processor.c sregs->cr0, sregs->cr2, sregs->cr3, sregs->cr4); cr4 626 tools/testing/selftests/kvm/lib/x86_64/processor.c sregs.cr4 |= X86_CR4_PAE | X86_CR4_OSFXSR; cr4 138 tools/testing/selftests/kvm/lib/x86_64/vmx.c unsigned long cr4; cr4 150 tools/testing/selftests/kvm/lib/x86_64/vmx.c __asm__ __volatile__("mov %%cr4, %0" : "=r"(cr4) : : "memory"); cr4 151 tools/testing/selftests/kvm/lib/x86_64/vmx.c cr4 &= rdmsr(MSR_IA32_VMX_CR4_FIXED1); cr4 152 tools/testing/selftests/kvm/lib/x86_64/vmx.c cr4 |= rdmsr(MSR_IA32_VMX_CR4_FIXED0); cr4 154 tools/testing/selftests/kvm/lib/x86_64/vmx.c cr4 |= X86_CR4_VMXE; cr4 155 tools/testing/selftests/kvm/lib/x86_64/vmx.c __asm__ __volatile__("mov %0, %%cr4" : : "r"(cr4) : "memory"); cr4 30 tools/testing/selftests/kvm/x86_64/cr4_cpuid_sync_test.c uint64_t cr4; cr4 38 tools/testing/selftests/kvm/x86_64/cr4_cpuid_sync_test.c cr4 = get_cr4(); cr4 40 tools/testing/selftests/kvm/x86_64/cr4_cpuid_sync_test.c return (!!(ecx & X86_FEATURE_OSXSAVE)) == (!!(cr4 & X86_CR4_OSXSAVE)); cr4 45 tools/testing/selftests/kvm/x86_64/cr4_cpuid_sync_test.c uint64_t cr4; cr4 48 tools/testing/selftests/kvm/x86_64/cr4_cpuid_sync_test.c cr4 = get_cr4(); cr4 49 tools/testing/selftests/kvm/x86_64/cr4_cpuid_sync_test.c cr4 |= X86_CR4_OSXSAVE; cr4 50 tools/testing/selftests/kvm/x86_64/cr4_cpuid_sync_test.c set_cr4(cr4); cr4 100 tools/testing/selftests/kvm/x86_64/cr4_cpuid_sync_test.c sregs.cr4 &= ~X86_CR4_OSXSAVE;