Lines Matching refs:u64
227 u64 *spt;
260 u64 rsvd_bits_mask[2][4];
261 u64 bad_mt_xwr;
272 u64 (*get_pdptr)(struct kvm_vcpu *vcpu, int index);
285 u64 *spte, const void *pte);
299 u64 *pae_root;
300 u64 *lm_root;
320 u64 pdptrs[4]; /* pae */
331 u64 counter;
332 u64 eventsel;
341 u64 fixed_ctr_ctrl;
342 u64 global_ctrl;
343 u64 global_status;
344 u64 global_ovf_ctrl;
345 u64 counter_bitmask[2];
346 u64 global_ctrl_mask;
347 u64 reserved_bits;
352 u64 reprogram_pmi;
364 u64 base;
365 u64 mask;
372 u64 deftype;
379 u64 hv_vapic;
400 u64 efer;
401 u64 apic_base;
403 u64 eoi_exit_bitmap[4];
407 u64 ia32_misc_enable_msr;
408 u64 smbase;
410 u64 ia32_xss;
443 u64 xcr0;
444 u64 guest_supported_xcr0;
489 u64 msr_val;
490 u64 last_steal;
491 u64 accum_steal;
496 u64 last_guest_tsc;
497 u64 last_host_tsc;
498 u64 tsc_offset_adjustment;
499 u64 this_tsc_nsec;
500 u64 this_tsc_write;
501 u64 this_tsc_generation;
508 u64 tsc_scaling_ratio;
516 u64 pat;
525 u64 mcg_cap;
526 u64 mcg_status;
527 u64 mcg_ctl;
528 u64 *mce_banks;
531 u64 mmio_gva;
534 u64 mmio_gen;
552 u64 msr_val;
559 u64 length;
560 u64 status;
564 u64 msr_val;
617 u64 hv_guest_os_id;
618 u64 hv_hypercall;
619 u64 hv_tsc_page;
622 u64 hv_crash_param[HV_X64_MSR_CRASH_PARAMS];
623 u64 hv_crash_ctl;
664 u64 last_tsc_nsec;
665 u64 last_tsc_write;
667 u64 cur_tsc_nsec;
668 u64 cur_tsc_write;
669 u64 cur_tsc_offset;
670 u64 cur_tsc_generation;
675 u64 master_kernel_ns;
694 u64 disabled_quirks;
746 u64 data;
784 u64 (*get_segment_base)(struct kvm_vcpu *vcpu, int seg);
797 void (*set_efer)(struct kvm_vcpu *vcpu, u64 efer);
802 u64 (*get_dr6)(struct kvm_vcpu *vcpu);
844 u64 (*get_mt_mask)(struct kvm_vcpu *vcpu, gfn_t gfn, bool is_mmio);
856 u64 (*read_tsc_offset)(struct kvm_vcpu *vcpu);
857 void (*write_tsc_offset)(struct kvm_vcpu *vcpu, u64 offset);
859 u64 (*read_l1_tsc)(struct kvm_vcpu *vcpu, u64 host_tsc);
861 void (*get_exit_info)(struct kvm_vcpu *vcpu, u64 *info1, u64 *info2);
931 void kvm_mmu_set_mask_ptes(u64 user_mask, u64 accessed_mask,
932 u64 dirty_mask, u64 nx_mask, u64 x_mask);
973 u64 vcpu_tsc_khz(struct kvm_vcpu *vcpu);
982 extern u64 kvm_max_tsc_scaling_ratio;
1004 void kvm_enable_efer_bits(u64);
1005 bool kvm_valid_efer(struct kvm_vcpu *vcpu, u64 efer);
1033 int kvm_set_xcr(struct kvm_vcpu *vcpu, u32 index, u64 xcr);
1127 u64 value; in read_msr()
1144 static inline u64 get_canonical(u64 la) in get_canonical()
1149 static inline bool is_noncanonical_address(u64 la) in is_noncanonical_address()
1226 int kvm_set_shared_msr(unsigned index, u64 val, u64 mask);
1228 u64 kvm_scale_tsc(struct kvm_vcpu *vcpu, u64 tsc);
1229 u64 kvm_read_l1_tsc(struct kvm_vcpu *vcpu, u64 host_tsc);