cr4                26 arch/sh/include/cpu-sh5/cpu/registers.h #define INTEVT	cr4
cr4               550 arch/x86/include/asm/kvm_host.h 	unsigned long cr4;
cr4              1049 arch/x86/include/asm/kvm_host.h 	int (*set_cr4)(struct kvm_vcpu *vcpu, unsigned long cr4);
cr4              1384 arch/x86/include/asm/kvm_host.h int kvm_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4);
cr4                49 arch/x86/include/asm/realmode.h 	u32 cr4;
cr4                22 arch/x86/include/asm/suspend_32.h 	unsigned long cr0, cr2, cr3, cr4;
cr4                37 arch/x86/include/asm/suspend_64.h 	unsigned long cr0, cr2, cr3, cr4;
cr4               177 arch/x86/include/asm/svm.h 	u64 cr4;
cr4               220 arch/x86/include/asm/tlbflush.h 	unsigned long cr4;
cr4               282 arch/x86/include/asm/tlbflush.h 	this_cpu_write(cpu_tlbstate.cr4, __read_cr4());
cr4               285 arch/x86/include/asm/tlbflush.h static inline void __cr4_set(unsigned long cr4)
cr4               288 arch/x86/include/asm/tlbflush.h 	this_cpu_write(cpu_tlbstate.cr4, cr4);
cr4               289 arch/x86/include/asm/tlbflush.h 	__write_cr4(cr4);
cr4               295 arch/x86/include/asm/tlbflush.h 	unsigned long cr4;
cr4               297 arch/x86/include/asm/tlbflush.h 	cr4 = this_cpu_read(cpu_tlbstate.cr4);
cr4               298 arch/x86/include/asm/tlbflush.h 	if ((cr4 | mask) != cr4)
cr4               299 arch/x86/include/asm/tlbflush.h 		__cr4_set(cr4 | mask);
cr4               305 arch/x86/include/asm/tlbflush.h 	unsigned long cr4;
cr4               307 arch/x86/include/asm/tlbflush.h 	cr4 = this_cpu_read(cpu_tlbstate.cr4);
cr4               308 arch/x86/include/asm/tlbflush.h 	if ((cr4 & ~mask) != cr4)
cr4               309 arch/x86/include/asm/tlbflush.h 		__cr4_set(cr4 & ~mask);
cr4               334 arch/x86/include/asm/tlbflush.h 	unsigned long cr4;
cr4               336 arch/x86/include/asm/tlbflush.h 	cr4 = this_cpu_read(cpu_tlbstate.cr4);
cr4               337 arch/x86/include/asm/tlbflush.h 	__cr4_set(cr4 ^ mask);
cr4               343 arch/x86/include/asm/tlbflush.h 	return this_cpu_read(cpu_tlbstate.cr4);
cr4               422 arch/x86/include/asm/tlbflush.h 	unsigned long cr4, flags;
cr4               442 arch/x86/include/asm/tlbflush.h 	cr4 = this_cpu_read(cpu_tlbstate.cr4);
cr4               444 arch/x86/include/asm/tlbflush.h 	native_write_cr4(cr4 ^ X86_CR4_PGE);
cr4               446 arch/x86/include/asm/tlbflush.h 	native_write_cr4(cr4);
cr4               154 arch/x86/include/uapi/asm/kvm.h 	__u64 cr0, cr2, cr3, cr4, cr8;
cr4                79 arch/x86/kernel/asm-offsets_64.c 	ENTRY(cr4);
cr4               413 arch/x86/kernel/cpu/common.c 	unsigned long cr4 = __read_cr4();
cr4               416 arch/x86/kernel/cpu/common.c 		cr4 |= X86_CR4_PCIDE;
cr4               418 arch/x86/kernel/cpu/common.c 		cr4 |= cr4_pinned_bits;
cr4               420 arch/x86/kernel/cpu/common.c 	__write_cr4(cr4);
cr4               423 arch/x86/kernel/cpu/common.c 	this_cpu_write(cpu_tlbstate.cr4, cr4);
cr4               436 arch/x86/kernel/cpu/common.c 	cr4_pinned_bits = this_cpu_read(cpu_tlbstate.cr4) & mask;
cr4               135 arch/x86/kernel/cpu/mtrr/cyrix.c static u32 cr4, ccr3;
cr4               143 arch/x86/kernel/cpu/mtrr/cyrix.c 		cr4 = __read_cr4();
cr4               144 arch/x86/kernel/cpu/mtrr/cyrix.c 		__write_cr4(cr4 & ~X86_CR4_PGE);
cr4               176 arch/x86/kernel/cpu/mtrr/cyrix.c 		__write_cr4(cr4);
cr4               720 arch/x86/kernel/cpu/mtrr/generic.c static unsigned long cr4;
cr4               758 arch/x86/kernel/cpu/mtrr/generic.c 		cr4 = __read_cr4();
cr4               759 arch/x86/kernel/cpu/mtrr/generic.c 		__write_cr4(cr4 & ~X86_CR4_PGE);
cr4               791 arch/x86/kernel/cpu/mtrr/generic.c 		__write_cr4(cr4);
cr4                63 arch/x86/kernel/process_32.c 	unsigned long cr0 = 0L, cr2 = 0L, cr3 = 0L, cr4 = 0L;
cr4                87 arch/x86/kernel/process_32.c 	cr4 = __read_cr4();
cr4                89 arch/x86/kernel/process_32.c 			cr0, cr2, cr3, cr4);
cr4                69 arch/x86/kernel/process_64.c 	unsigned long cr0 = 0L, cr2 = 0L, cr3 = 0L, cr4 = 0L, fs, gs, shadowgs;
cr4               115 arch/x86/kernel/process_64.c 	cr4 = __read_cr4();
cr4               122 arch/x86/kernel/process_64.c 			cr4);
cr4              2447 arch/x86/kvm/emulate.c 				    u64 cr0, u64 cr3, u64 cr4)
cr4              2454 arch/x86/kvm/emulate.c 	if (cr4 & X86_CR4_PCIDE) {
cr4              2468 arch/x86/kvm/emulate.c 	bad = ctxt->ops->set_cr(ctxt, 4, cr4 & ~X86_CR4_PCIDE);
cr4              2476 arch/x86/kvm/emulate.c 	if (cr4 & X86_CR4_PCIDE) {
cr4              2477 arch/x86/kvm/emulate.c 		bad = ctxt->ops->set_cr(ctxt, 4, cr4);
cr4              2497 arch/x86/kvm/emulate.c 	u32 val, cr0, cr3, cr4;
cr4              2539 arch/x86/kvm/emulate.c 	cr4 = GET_SMSTATE(u32, smstate, 0x7f14);
cr4              2543 arch/x86/kvm/emulate.c 	return rsm_enter_protected_mode(ctxt, cr0, cr3, cr4);
cr4              2552 arch/x86/kvm/emulate.c 	u64 val, cr0, cr3, cr4;
cr4              2570 arch/x86/kvm/emulate.c 	cr4 =                       GET_SMSTATE(u64, smstate, 0x7f48);
cr4              2597 arch/x86/kvm/emulate.c 	r = rsm_enter_protected_mode(ctxt, cr0, cr3, cr4);
cr4              2613 arch/x86/kvm/emulate.c 	unsigned long cr0, cr4, efer;
cr4              2642 arch/x86/kvm/emulate.c 		cr4 = ctxt->ops->get_cr(ctxt, 4);
cr4              2643 arch/x86/kvm/emulate.c 		if (cr4 & X86_CR4_PCIDE)
cr4              2644 arch/x86/kvm/emulate.c 			ctxt->ops->set_cr(ctxt, 4, cr4 & ~X86_CR4_PCIDE);
cr4              2660 arch/x86/kvm/emulate.c 		cr4 = ctxt->ops->get_cr(ctxt, 4);
cr4              2661 arch/x86/kvm/emulate.c 		if (cr4 & X86_CR4_PAE)
cr4              2662 arch/x86/kvm/emulate.c 			ctxt->ops->set_cr(ctxt, 4, cr4 & ~X86_CR4_PAE);
cr4              4252 arch/x86/kvm/emulate.c 		u64 cr4;
cr4              4257 arch/x86/kvm/emulate.c 		cr4 = ctxt->ops->get_cr(ctxt, 4);
cr4              4261 arch/x86/kvm/emulate.c 		    !(cr4 & X86_CR4_PAE))
cr4              4317 arch/x86/kvm/emulate.c 	u64 cr4;
cr4              4322 arch/x86/kvm/emulate.c 	cr4 = ctxt->ops->get_cr(ctxt, 4);
cr4              4323 arch/x86/kvm/emulate.c 	if ((cr4 & X86_CR4_DE) && (dr == 4 || dr == 5))
cr4              4375 arch/x86/kvm/emulate.c 	u64 cr4 = ctxt->ops->get_cr(ctxt, 4);
cr4              4377 arch/x86/kvm/emulate.c 	if (cr4 & X86_CR4_TSD && ctxt->ops->cpl(ctxt))
cr4              4385 arch/x86/kvm/emulate.c 	u64 cr4 = ctxt->ops->get_cr(ctxt, 4);
cr4              4395 arch/x86/kvm/emulate.c 	if ((!(cr4 & X86_CR4_PCE) && ctxt->ops->cpl(ctxt)) ||
cr4               107 arch/x86/kvm/kvm_cache_regs.h 	return vcpu->arch.cr4 & mask;
cr4              1638 arch/x86/kvm/svm.c 	save->cr4 = X86_CR4_PAE;
cr4              1650 arch/x86/kvm/svm.c 		save->cr4 = 0;
cr4              2633 arch/x86/kvm/svm.c static int svm_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4)
cr4              2636 arch/x86/kvm/svm.c 	unsigned long old_cr4 = to_svm(vcpu)->vmcb->save.cr4;
cr4              2638 arch/x86/kvm/svm.c 	if (cr4 & X86_CR4_VMXE)
cr4              2641 arch/x86/kvm/svm.c 	if (npt_enabled && ((old_cr4 ^ cr4) & X86_CR4_PGE))
cr4              2644 arch/x86/kvm/svm.c 	vcpu->arch.cr4 = cr4;
cr4              2646 arch/x86/kvm/svm.c 		cr4 |= X86_CR4_PAE;
cr4              2647 arch/x86/kvm/svm.c 	cr4 |= host_cr4_mce;
cr4              2648 arch/x86/kvm/svm.c 	to_svm(vcpu)->vmcb->save.cr4 = cr4;
cr4              3391 arch/x86/kvm/svm.c 	nested_vmcb->save.cr4    = svm->vcpu.arch.cr4;
cr4              3460 arch/x86/kvm/svm.c 	svm_set_cr4(&svm->vcpu, hsave->save.cr4);
cr4              3564 arch/x86/kvm/svm.c 	svm_set_cr4(&svm->vcpu, nested_vmcb->save.cr4);
cr4              3705 arch/x86/kvm/svm.c 	hsave->save.cr4    = svm->vcpu.arch.cr4;
cr4              4942 arch/x86/kvm/svm.c 	       "cr3:", save->cr3, "cr4:", save->cr4);
cr4              7166 arch/x86/kvm/svm.c 	unsigned long cr4 = kvm_read_cr4(vcpu);
cr4              7167 arch/x86/kvm/svm.c 	bool smep = cr4 & X86_CR4_SMEP;
cr4              7168 arch/x86/kvm/svm.c 	bool smap = cr4 & X86_CR4_SMAP;
cr4              2826 arch/x86/kvm/vmx/nested.c 	unsigned long cr3, cr4;
cr4              2855 arch/x86/kvm/vmx/nested.c 	cr4 = cr4_read_shadow();
cr4              2856 arch/x86/kvm/vmx/nested.c 	if (unlikely(cr4 != vmx->loaded_vmcs->host_state.cr4)) {
cr4              2857 arch/x86/kvm/vmx/nested.c 		vmcs_writel(HOST_CR4, cr4);
cr4              2858 arch/x86/kvm/vmx/nested.c 		vmx->loaded_vmcs->host_state.cr4 = cr4;
cr4                34 arch/x86/kvm/vmx/vmcs.h 	unsigned long cr4;	/* May not match real cr4 */
cr4              2845 arch/x86/kvm/vmx/vmx.c 	vcpu->arch.cr4 &= ~cr4_guest_owned_bits;
cr4              2846 arch/x86/kvm/vmx/vmx.c 	vcpu->arch.cr4 |= vmcs_readl(GUEST_CR4) & cr4_guest_owned_bits;
cr4              3004 arch/x86/kvm/vmx/vmx.c int vmx_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4)
cr4              3014 arch/x86/kvm/vmx/vmx.c 	hw_cr4 = (cr4_read_shadow() & X86_CR4_MCE) | (cr4 & ~X86_CR4_MCE);
cr4              3023 arch/x86/kvm/vmx/vmx.c 		if (cr4 & X86_CR4_UMIP) {
cr4              3032 arch/x86/kvm/vmx/vmx.c 	if (cr4 & X86_CR4_VMXE) {
cr4              3044 arch/x86/kvm/vmx/vmx.c 	if (vmx->nested.vmxon && !nested_cr4_valid(vcpu, cr4))
cr4              3047 arch/x86/kvm/vmx/vmx.c 	vcpu->arch.cr4 = cr4;
cr4              3054 arch/x86/kvm/vmx/vmx.c 			} else if (!(cr4 & X86_CR4_PAE)) {
cr4              3074 arch/x86/kvm/vmx/vmx.c 	vmcs_writel(CR4_READ_SHADOW, cr4);
cr4              3861 arch/x86/kvm/vmx/vmx.c 	unsigned long cr0, cr3, cr4;
cr4              3876 arch/x86/kvm/vmx/vmx.c 	cr4 = cr4_read_shadow();
cr4              3877 arch/x86/kvm/vmx/vmx.c 	vmcs_writel(HOST_CR4, cr4);			/* 22.2.3, 22.2.5 */
cr4              3878 arch/x86/kvm/vmx/vmx.c 	vmx->loaded_vmcs->host_state.cr4 = cr4;
cr4              4797 arch/x86/kvm/vmx/vmx.c 	WARN_ON(!(vcpu->arch.cr4 & X86_CR4_UMIP));
cr4              5685 arch/x86/kvm/vmx/vmx.c 	unsigned long cr4;
cr4              5698 arch/x86/kvm/vmx/vmx.c 	cr4 = vmcs_readl(GUEST_CR4);
cr4              5709 arch/x86/kvm/vmx/vmx.c 	       cr4, vmcs_readl(CR4_READ_SHADOW), vmcs_readl(CR4_GUEST_HOST_MASK));
cr4              5712 arch/x86/kvm/vmx/vmx.c 	    (cr4 & X86_CR4_PAE) && !(efer & EFER_LMA))
cr4              6486 arch/x86/kvm/vmx/vmx.c 	unsigned long cr3, cr4;
cr4              6517 arch/x86/kvm/vmx/vmx.c 	cr4 = cr4_read_shadow();
cr4              6518 arch/x86/kvm/vmx/vmx.c 	if (unlikely(cr4 != vmx->loaded_vmcs->host_state.cr4)) {
cr4              6519 arch/x86/kvm/vmx/vmx.c 		vmcs_writel(HOST_CR4, cr4);
cr4              6520 arch/x86/kvm/vmx/vmx.c 		vmx->loaded_vmcs->host_state.cr4 = cr4;
cr4               324 arch/x86/kvm/vmx/vmx.h int vmx_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4);
cr4               942 arch/x86/kvm/x86.c static int kvm_valid_cr4(struct kvm_vcpu *vcpu, unsigned long cr4)
cr4               944 arch/x86/kvm/x86.c 	if (cr4 & cr4_reserved_bits)
cr4               947 arch/x86/kvm/x86.c 	if (!guest_cpuid_has(vcpu, X86_FEATURE_XSAVE) && (cr4 & X86_CR4_OSXSAVE))
cr4               950 arch/x86/kvm/x86.c 	if (!guest_cpuid_has(vcpu, X86_FEATURE_SMEP) && (cr4 & X86_CR4_SMEP))
cr4               953 arch/x86/kvm/x86.c 	if (!guest_cpuid_has(vcpu, X86_FEATURE_SMAP) && (cr4 & X86_CR4_SMAP))
cr4               956 arch/x86/kvm/x86.c 	if (!guest_cpuid_has(vcpu, X86_FEATURE_FSGSBASE) && (cr4 & X86_CR4_FSGSBASE))
cr4               959 arch/x86/kvm/x86.c 	if (!guest_cpuid_has(vcpu, X86_FEATURE_PKU) && (cr4 & X86_CR4_PKE))
cr4               962 arch/x86/kvm/x86.c 	if (!guest_cpuid_has(vcpu, X86_FEATURE_LA57) && (cr4 & X86_CR4_LA57))
cr4               965 arch/x86/kvm/x86.c 	if (!guest_cpuid_has(vcpu, X86_FEATURE_UMIP) && (cr4 & X86_CR4_UMIP))
cr4               971 arch/x86/kvm/x86.c int kvm_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4)
cr4               977 arch/x86/kvm/x86.c 	if (kvm_valid_cr4(vcpu, cr4))
cr4               981 arch/x86/kvm/x86.c 		if (!(cr4 & X86_CR4_PAE))
cr4               983 arch/x86/kvm/x86.c 	} else if (is_paging(vcpu) && (cr4 & X86_CR4_PAE)
cr4               984 arch/x86/kvm/x86.c 		   && ((cr4 ^ old_cr4) & pdptr_bits)
cr4               989 arch/x86/kvm/x86.c 	if ((cr4 & X86_CR4_PCIDE) && !(old_cr4 & X86_CR4_PCIDE)) {
cr4               998 arch/x86/kvm/x86.c 	if (kvm_x86_ops->set_cr4(vcpu, cr4))
cr4              1001 arch/x86/kvm/x86.c 	if (((cr4 ^ old_cr4) & pdptr_bits) ||
cr4              1002 arch/x86/kvm/x86.c 	    (!(cr4 & X86_CR4_PCIDE) && (old_cr4 & X86_CR4_PCIDE)))
cr4              1005 arch/x86/kvm/x86.c 	if ((cr4 ^ old_cr4) & (X86_CR4_OSXSAVE | X86_CR4_PKE))
cr4              8746 arch/x86/kvm/x86.c 	sregs->cr4 = kvm_read_cr4(vcpu);
cr4              8849 arch/x86/kvm/x86.c 		if (!(sregs->cr4 & X86_CR4_PAE)
cr4              8861 arch/x86/kvm/x86.c 	return kvm_valid_cr4(vcpu, sregs->cr4);
cr4              8902 arch/x86/kvm/x86.c 	mmu_reset_needed |= kvm_read_cr4(vcpu) != sregs->cr4;
cr4              8903 arch/x86/kvm/x86.c 	cpuid_update_needed |= ((kvm_read_cr4(vcpu) ^ sregs->cr4) &
cr4              8905 arch/x86/kvm/x86.c 	kvm_x86_ops->set_cr4(vcpu, sregs->cr4);
cr4               957 arch/x86/mm/init.c 	.cr4 = ~0UL,	/* fail hard if we screw up cr4 shadow initialization */
cr4               125 arch/x86/power/cpu.c 	ctxt->cr4 = __read_cr4();
cr4               203 arch/x86/power/cpu.c 	if (ctxt->cr4)
cr4               204 arch/x86/power/cpu.c 		__write_cr4(ctxt->cr4);
cr4               208 arch/x86/power/cpu.c 	__write_cr4(ctxt->cr4);
cr4               101 arch/x86/realmode/init.c 	trampoline_cr4_features = &trampoline_header->cr4;
cr4               882 arch/x86/xen/enlighten_pv.c static void xen_write_cr4(unsigned long cr4)
cr4               884 arch/x86/xen/enlighten_pv.c 	cr4 &= ~(X86_CR4_PGE | X86_CR4_PSE | X86_CR4_PCE);
cr4               886 arch/x86/xen/enlighten_pv.c 	native_write_cr4(cr4);
cr4               281 drivers/misc/lkdtm/bugs.c 	unsigned long cr4;
cr4               284 drivers/misc/lkdtm/bugs.c 	cr4 = native_read_cr4();
cr4               286 drivers/misc/lkdtm/bugs.c 	if ((cr4 & X86_CR4_SMEP) != X86_CR4_SMEP) {
cr4               290 drivers/misc/lkdtm/bugs.c 	cr4 &= ~(X86_CR4_SMEP);
cr4               293 drivers/misc/lkdtm/bugs.c 	native_write_cr4(cr4);
cr4               294 drivers/misc/lkdtm/bugs.c 	if (cr4 == native_read_cr4()) {
cr4               296 drivers/misc/lkdtm/bugs.c 		cr4 |= X86_CR4_SMEP;
cr4               298 drivers/misc/lkdtm/bugs.c 		native_write_cr4(cr4);
cr4               327 drivers/misc/lkdtm/bugs.c 	direct_write_cr4(cr4);
cr4               332 drivers/misc/lkdtm/bugs.c 		cr4 |= X86_CR4_SMEP;
cr4               334 drivers/misc/lkdtm/bugs.c 		native_write_cr4(cr4);
cr4               641 drivers/net/wireless/broadcom/brcm80211/brcmfmac/chip.c static u32 brcmf_chip_tcm_ramsize(struct brcmf_core_priv *cr4)
cr4               651 drivers/net/wireless/broadcom/brcm80211/brcmfmac/chip.c 	corecap = brcmf_chip_core_read32(cr4, ARMCR4_CAP);
cr4               658 drivers/net/wireless/broadcom/brcm80211/brcmfmac/chip.c 		brcmf_chip_core_write32(cr4, ARMCR4_BANKIDX, idx);
cr4               659 drivers/net/wireless/broadcom/brcm80211/brcmfmac/chip.c 		bxinfo = brcmf_chip_core_read32(cr4, ARMCR4_BANKINFO);
cr4               959 drivers/parport/parport_pc.c 	int cr1, cr4, cra, cr23, cr26, cr27;
cr4               973 drivers/parport/parport_pc.c 	cr4 = inb(io + 1);
cr4               988 drivers/parport/parport_pc.c 			cr1, cr4, cra, cr23, cr26, cr27);
cr4              1005 drivers/parport/parport_pc.c 					      : modes[cr4 & 0x03],
cr4              1006 drivers/parport/parport_pc.c 				(cr4 & 0x40) ? "1.7" : "1.9");
cr4              1633 drivers/tty/serial/fsl_lpuart.c 	unsigned char cr1, old_cr1, old_cr2, cr3, cr4, bdh, modem;
cr4              1641 drivers/tty/serial/fsl_lpuart.c 	cr4 = readb(sport->port.membase + UARTCR4);
cr4              1760 drivers/tty/serial/fsl_lpuart.c 	cr4 &= ~UARTCR4_BRFA_MASK;
cr4              1762 drivers/tty/serial/fsl_lpuart.c 	writeb(cr4 | brfa, sport->port.membase + UARTCR4);
cr4              6364 drivers/video/fbdev/sis/init301.c 	   unsigned char cr4, cr14, cr5, cr15;
cr4              6366 drivers/video/fbdev/sis/init301.c 	      cr4  = SiS_Pr->CCRT1CRTC[4];
cr4              6371 drivers/video/fbdev/sis/init301.c 	      cr4  = SiS_Pr->SiS_CRT1Table[CRT1Index].CR[4];
cr4              6376 drivers/video/fbdev/sis/init301.c 	   tempbx = ((cr4 | ((cr14 & 0xC0) << 2)) - 3) << 3; 		    /* (VGAHRS-3)*8 */
cr4                42 include/xen/interface/hvm/hvm_vcpu.h     uint32_t cr4;
cr4               105 include/xen/interface/hvm/hvm_vcpu.h     uint64_t cr4;
cr4               154 tools/arch/x86/include/uapi/asm/kvm.h 	__u64 cr0, cr2, cr3, cr4, cr8;
cr4               209 tools/testing/selftests/kvm/include/x86_64/processor.h 	uint64_t cr4;
cr4               212 tools/testing/selftests/kvm/include/x86_64/processor.h 			     : /* output */ [cr4]"=r"(cr4));
cr4               213 tools/testing/selftests/kvm/include/x86_64/processor.h 	return cr4;
cr4               216 tools/testing/selftests/kvm/lib/x86_64/processor.c 		sregs->cr0, sregs->cr2, sregs->cr3, sregs->cr4);
cr4               626 tools/testing/selftests/kvm/lib/x86_64/processor.c 		sregs.cr4 |= X86_CR4_PAE | X86_CR4_OSFXSR;
cr4               138 tools/testing/selftests/kvm/lib/x86_64/vmx.c 	unsigned long cr4;
cr4               150 tools/testing/selftests/kvm/lib/x86_64/vmx.c 	__asm__ __volatile__("mov %%cr4, %0" : "=r"(cr4) : : "memory");
cr4               151 tools/testing/selftests/kvm/lib/x86_64/vmx.c 	cr4 &= rdmsr(MSR_IA32_VMX_CR4_FIXED1);
cr4               152 tools/testing/selftests/kvm/lib/x86_64/vmx.c 	cr4 |= rdmsr(MSR_IA32_VMX_CR4_FIXED0);
cr4               154 tools/testing/selftests/kvm/lib/x86_64/vmx.c 	cr4 |= X86_CR4_VMXE;
cr4               155 tools/testing/selftests/kvm/lib/x86_64/vmx.c 	__asm__ __volatile__("mov %0, %%cr4" : : "r"(cr4) : "memory");
cr4                30 tools/testing/selftests/kvm/x86_64/cr4_cpuid_sync_test.c 	uint64_t cr4;
cr4                38 tools/testing/selftests/kvm/x86_64/cr4_cpuid_sync_test.c 	cr4 = get_cr4();
cr4                40 tools/testing/selftests/kvm/x86_64/cr4_cpuid_sync_test.c 	return (!!(ecx & X86_FEATURE_OSXSAVE)) == (!!(cr4 & X86_CR4_OSXSAVE));
cr4                45 tools/testing/selftests/kvm/x86_64/cr4_cpuid_sync_test.c 	uint64_t cr4;
cr4                48 tools/testing/selftests/kvm/x86_64/cr4_cpuid_sync_test.c 	cr4 = get_cr4();
cr4                49 tools/testing/selftests/kvm/x86_64/cr4_cpuid_sync_test.c 	cr4 |= X86_CR4_OSXSAVE;
cr4                50 tools/testing/selftests/kvm/x86_64/cr4_cpuid_sync_test.c 	set_cr4(cr4);
cr4               100 tools/testing/selftests/kvm/x86_64/cr4_cpuid_sync_test.c 			sregs.cr4 &= ~X86_CR4_OSXSAVE;