Lines Matching refs:context
3547 struct kvm_mmu *context) in nonpaging_init_context() argument
3549 context->page_fault = nonpaging_page_fault; in nonpaging_init_context()
3550 context->gva_to_gpa = nonpaging_gva_to_gpa; in nonpaging_init_context()
3551 context->sync_page = nonpaging_sync_page; in nonpaging_init_context()
3552 context->invlpg = nonpaging_invlpg; in nonpaging_init_context()
3553 context->update_pte = nonpaging_update_pte; in nonpaging_init_context()
3554 context->root_level = 0; in nonpaging_init_context()
3555 context->shadow_root_level = PT32E_ROOT_LEVEL; in nonpaging_init_context()
3556 context->root_hpa = INVALID_PAGE; in nonpaging_init_context()
3557 context->direct_map = true; in nonpaging_init_context()
3558 context->nx = false; in nonpaging_init_context()
3700 struct kvm_mmu *context) in reset_rsvds_bits_mask() argument
3702 __reset_rsvds_bits_mask(vcpu, &context->guest_rsvd_check, in reset_rsvds_bits_mask()
3703 cpuid_maxphyaddr(vcpu), context->root_level, in reset_rsvds_bits_mask()
3704 context->nx, guest_cpuid_has_gbpages(vcpu), in reset_rsvds_bits_mask()
3743 struct kvm_mmu *context, bool execonly) in reset_rsvds_bits_mask_ept() argument
3745 __reset_rsvds_bits_mask_ept(&context->guest_rsvd_check, in reset_rsvds_bits_mask_ept()
3755 reset_shadow_zero_bits_mask(struct kvm_vcpu *vcpu, struct kvm_mmu *context) in reset_shadow_zero_bits_mask() argument
3757 bool uses_nx = context->nx || context->base_role.smep_andnot_wp; in reset_shadow_zero_bits_mask()
3763 __reset_rsvds_bits_mask(vcpu, &context->shadow_zero_check, in reset_shadow_zero_bits_mask()
3765 context->shadow_root_level, uses_nx, in reset_shadow_zero_bits_mask()
3783 struct kvm_mmu *context) in reset_tdp_shadow_zero_bits_mask() argument
3786 __reset_rsvds_bits_mask(vcpu, &context->shadow_zero_check, in reset_tdp_shadow_zero_bits_mask()
3788 context->shadow_root_level, false, in reset_tdp_shadow_zero_bits_mask()
3791 __reset_rsvds_bits_mask_ept(&context->shadow_zero_check, in reset_tdp_shadow_zero_bits_mask()
3803 struct kvm_mmu *context, bool execonly) in reset_ept_shadow_zero_bits_mask() argument
3805 __reset_rsvds_bits_mask_ept(&context->shadow_zero_check, in reset_ept_shadow_zero_bits_mask()
3893 struct kvm_mmu *context, in paging64_init_context_common() argument
3896 context->nx = is_nx(vcpu); in paging64_init_context_common()
3897 context->root_level = level; in paging64_init_context_common()
3899 reset_rsvds_bits_mask(vcpu, context); in paging64_init_context_common()
3900 update_permission_bitmask(vcpu, context, false); in paging64_init_context_common()
3901 update_last_pte_bitmap(vcpu, context); in paging64_init_context_common()
3904 context->page_fault = paging64_page_fault; in paging64_init_context_common()
3905 context->gva_to_gpa = paging64_gva_to_gpa; in paging64_init_context_common()
3906 context->sync_page = paging64_sync_page; in paging64_init_context_common()
3907 context->invlpg = paging64_invlpg; in paging64_init_context_common()
3908 context->update_pte = paging64_update_pte; in paging64_init_context_common()
3909 context->shadow_root_level = level; in paging64_init_context_common()
3910 context->root_hpa = INVALID_PAGE; in paging64_init_context_common()
3911 context->direct_map = false; in paging64_init_context_common()
3915 struct kvm_mmu *context) in paging64_init_context() argument
3917 paging64_init_context_common(vcpu, context, PT64_ROOT_LEVEL); in paging64_init_context()
3921 struct kvm_mmu *context) in paging32_init_context() argument
3923 context->nx = false; in paging32_init_context()
3924 context->root_level = PT32_ROOT_LEVEL; in paging32_init_context()
3926 reset_rsvds_bits_mask(vcpu, context); in paging32_init_context()
3927 update_permission_bitmask(vcpu, context, false); in paging32_init_context()
3928 update_last_pte_bitmap(vcpu, context); in paging32_init_context()
3930 context->page_fault = paging32_page_fault; in paging32_init_context()
3931 context->gva_to_gpa = paging32_gva_to_gpa; in paging32_init_context()
3932 context->sync_page = paging32_sync_page; in paging32_init_context()
3933 context->invlpg = paging32_invlpg; in paging32_init_context()
3934 context->update_pte = paging32_update_pte; in paging32_init_context()
3935 context->shadow_root_level = PT32E_ROOT_LEVEL; in paging32_init_context()
3936 context->root_hpa = INVALID_PAGE; in paging32_init_context()
3937 context->direct_map = false; in paging32_init_context()
3941 struct kvm_mmu *context) in paging32E_init_context() argument
3943 paging64_init_context_common(vcpu, context, PT32E_ROOT_LEVEL); in paging32E_init_context()
3948 struct kvm_mmu *context = &vcpu->arch.mmu; in init_kvm_tdp_mmu() local
3950 context->base_role.word = 0; in init_kvm_tdp_mmu()
3951 context->base_role.smm = is_smm(vcpu); in init_kvm_tdp_mmu()
3952 context->page_fault = tdp_page_fault; in init_kvm_tdp_mmu()
3953 context->sync_page = nonpaging_sync_page; in init_kvm_tdp_mmu()
3954 context->invlpg = nonpaging_invlpg; in init_kvm_tdp_mmu()
3955 context->update_pte = nonpaging_update_pte; in init_kvm_tdp_mmu()
3956 context->shadow_root_level = kvm_x86_ops->get_tdp_level(); in init_kvm_tdp_mmu()
3957 context->root_hpa = INVALID_PAGE; in init_kvm_tdp_mmu()
3958 context->direct_map = true; in init_kvm_tdp_mmu()
3959 context->set_cr3 = kvm_x86_ops->set_tdp_cr3; in init_kvm_tdp_mmu()
3960 context->get_cr3 = get_cr3; in init_kvm_tdp_mmu()
3961 context->get_pdptr = kvm_pdptr_read; in init_kvm_tdp_mmu()
3962 context->inject_page_fault = kvm_inject_page_fault; in init_kvm_tdp_mmu()
3965 context->nx = false; in init_kvm_tdp_mmu()
3966 context->gva_to_gpa = nonpaging_gva_to_gpa; in init_kvm_tdp_mmu()
3967 context->root_level = 0; in init_kvm_tdp_mmu()
3969 context->nx = is_nx(vcpu); in init_kvm_tdp_mmu()
3970 context->root_level = PT64_ROOT_LEVEL; in init_kvm_tdp_mmu()
3971 reset_rsvds_bits_mask(vcpu, context); in init_kvm_tdp_mmu()
3972 context->gva_to_gpa = paging64_gva_to_gpa; in init_kvm_tdp_mmu()
3974 context->nx = is_nx(vcpu); in init_kvm_tdp_mmu()
3975 context->root_level = PT32E_ROOT_LEVEL; in init_kvm_tdp_mmu()
3976 reset_rsvds_bits_mask(vcpu, context); in init_kvm_tdp_mmu()
3977 context->gva_to_gpa = paging64_gva_to_gpa; in init_kvm_tdp_mmu()
3979 context->nx = false; in init_kvm_tdp_mmu()
3980 context->root_level = PT32_ROOT_LEVEL; in init_kvm_tdp_mmu()
3981 reset_rsvds_bits_mask(vcpu, context); in init_kvm_tdp_mmu()
3982 context->gva_to_gpa = paging32_gva_to_gpa; in init_kvm_tdp_mmu()
3985 update_permission_bitmask(vcpu, context, false); in init_kvm_tdp_mmu()
3986 update_last_pte_bitmap(vcpu, context); in init_kvm_tdp_mmu()
3987 reset_tdp_shadow_zero_bits_mask(vcpu, context); in init_kvm_tdp_mmu()
3994 struct kvm_mmu *context = &vcpu->arch.mmu; in kvm_init_shadow_mmu() local
3996 MMU_WARN_ON(VALID_PAGE(context->root_hpa)); in kvm_init_shadow_mmu()
3999 nonpaging_init_context(vcpu, context); in kvm_init_shadow_mmu()
4001 paging64_init_context(vcpu, context); in kvm_init_shadow_mmu()
4003 paging32E_init_context(vcpu, context); in kvm_init_shadow_mmu()
4005 paging32_init_context(vcpu, context); in kvm_init_shadow_mmu()
4007 context->base_role.nxe = is_nx(vcpu); in kvm_init_shadow_mmu()
4008 context->base_role.cr4_pae = !!is_pae(vcpu); in kvm_init_shadow_mmu()
4009 context->base_role.cr0_wp = is_write_protection(vcpu); in kvm_init_shadow_mmu()
4010 context->base_role.smep_andnot_wp in kvm_init_shadow_mmu()
4012 context->base_role.smap_andnot_wp in kvm_init_shadow_mmu()
4014 context->base_role.smm = is_smm(vcpu); in kvm_init_shadow_mmu()
4015 reset_shadow_zero_bits_mask(vcpu, context); in kvm_init_shadow_mmu()
4021 struct kvm_mmu *context = &vcpu->arch.mmu; in kvm_init_shadow_ept_mmu() local
4023 MMU_WARN_ON(VALID_PAGE(context->root_hpa)); in kvm_init_shadow_ept_mmu()
4025 context->shadow_root_level = kvm_x86_ops->get_tdp_level(); in kvm_init_shadow_ept_mmu()
4027 context->nx = true; in kvm_init_shadow_ept_mmu()
4028 context->page_fault = ept_page_fault; in kvm_init_shadow_ept_mmu()
4029 context->gva_to_gpa = ept_gva_to_gpa; in kvm_init_shadow_ept_mmu()
4030 context->sync_page = ept_sync_page; in kvm_init_shadow_ept_mmu()
4031 context->invlpg = ept_invlpg; in kvm_init_shadow_ept_mmu()
4032 context->update_pte = ept_update_pte; in kvm_init_shadow_ept_mmu()
4033 context->root_level = context->shadow_root_level; in kvm_init_shadow_ept_mmu()
4034 context->root_hpa = INVALID_PAGE; in kvm_init_shadow_ept_mmu()
4035 context->direct_map = false; in kvm_init_shadow_ept_mmu()
4037 update_permission_bitmask(vcpu, context, true); in kvm_init_shadow_ept_mmu()
4038 reset_rsvds_bits_mask_ept(vcpu, context, execonly); in kvm_init_shadow_ept_mmu()
4039 reset_ept_shadow_zero_bits_mask(vcpu, context, execonly); in kvm_init_shadow_ept_mmu()
4045 struct kvm_mmu *context = &vcpu->arch.mmu; in init_kvm_softmmu() local
4048 context->set_cr3 = kvm_x86_ops->set_cr3; in init_kvm_softmmu()
4049 context->get_cr3 = get_cr3; in init_kvm_softmmu()
4050 context->get_pdptr = kvm_pdptr_read; in init_kvm_softmmu()
4051 context->inject_page_fault = kvm_inject_page_fault; in init_kvm_softmmu()