Lines Matching defs:vmx
450 static inline void vmx_segment_cache_clear(struct vcpu_vmx *vmx) in vmx_segment_cache_clear()
589 struct vmx_uret_msr *vmx_find_uret_msr(struct vcpu_vmx *vmx, u32 msr) in vmx_find_uret_msr()
599 static int vmx_set_guest_uret_msr(struct vcpu_vmx *vmx, in vmx_set_guest_uret_msr()
667 static bool vmx_segment_cache_test_set(struct vcpu_vmx *vmx, unsigned seg, in vmx_segment_cache_test_set()
682 static u16 vmx_read_guest_seg_selector(struct vcpu_vmx *vmx, unsigned seg) in vmx_read_guest_seg_selector()
691 static ulong vmx_read_guest_seg_base(struct vcpu_vmx *vmx, unsigned seg) in vmx_read_guest_seg_base()
700 static u32 vmx_read_guest_seg_limit(struct vcpu_vmx *vmx, unsigned seg) in vmx_read_guest_seg_limit()
709 static u32 vmx_read_guest_seg_ar(struct vcpu_vmx *vmx, unsigned seg) in vmx_read_guest_seg_ar()
772 static bool msr_write_intercepted(struct vcpu_vmx *vmx, u32 msr) in msr_write_intercepted()
781 static void clear_atomic_switch_msr_special(struct vcpu_vmx *vmx, in clear_atomic_switch_msr_special()
799 static void clear_atomic_switch_msr(struct vcpu_vmx *vmx, unsigned msr) in clear_atomic_switch_msr()
839 static void add_atomic_switch_msr_special(struct vcpu_vmx *vmx, in add_atomic_switch_msr_special()
851 static void add_atomic_switch_msr(struct vcpu_vmx *vmx, unsigned msr, in add_atomic_switch_msr()
917 static bool update_transition_efer(struct vcpu_vmx *vmx) in update_transition_efer()
999 static inline bool pt_can_write_msr(struct vcpu_vmx *vmx) in pt_can_write_msr()
1039 static void pt_guest_enter(struct vcpu_vmx *vmx) in pt_guest_enter()
1056 static void pt_guest_exit(struct vcpu_vmx *vmx) in pt_guest_exit()
1103 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_prepare_switch_to_guest() local
1175 static void vmx_prepare_switch_to_host(struct vcpu_vmx *vmx) in vmx_prepare_switch_to_host()
1215 static u64 vmx_read_guest_kernel_gs_base(struct vcpu_vmx *vmx) in vmx_read_guest_kernel_gs_base()
1224 static void vmx_write_guest_kernel_gs_base(struct vcpu_vmx *vmx, u64 data) in vmx_write_guest_kernel_gs_base()
1237 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_vcpu_load_vmcs() local
1303 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_vcpu_load() local
1326 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_get_rflags() local
1344 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_set_rflags() local
1402 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_rtit_ctl_check() local
1559 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_update_emulated_instruction() local
1600 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_queue_exception() local
1635 static void vmx_setup_uret_msr(struct vcpu_vmx *vmx, unsigned int msr, in vmx_setup_uret_msr()
1653 static void vmx_setup_uret_msrs(struct vcpu_vmx *vmx) in vmx_setup_uret_msrs()
1762 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_get_msr() local
1940 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_set_msr() local
2730 struct vcpu_vmx *vmx = to_vmx(vcpu); in enter_pmode() local
2801 struct vcpu_vmx *vmx = to_vmx(vcpu); in enter_rmode() local
2847 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_set_efer() local
2896 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_flush_tlb_all() local
2996 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_set_cr0() local
3145 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_set_cr4() local
3208 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_get_segment() local
3255 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_get_cpl() local
3287 struct vcpu_vmx *vmx = to_vmx(vcpu); in __vmx_set_segment() local
3692 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_disable_intercept_for_msr() local
3737 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_enable_intercept_for_msr() local
3787 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_update_msr_bitmap_x2apic() local
3826 struct vcpu_vmx *vmx = to_vmx(vcpu); in pt_update_intercept_for_msr() local
3842 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_guest_apic_has_interrupt() local
3862 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_msr_filter_changed() local
3924 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_deliver_nested_posted_interrupt() local
3950 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_deliver_posted_interrupt() local
3979 void vmx_set_constant_host_state(struct vcpu_vmx *vmx) in vmx_set_constant_host_state()
4036 void set_cr4_guest_host_mask(struct vcpu_vmx *vmx) in set_cr4_guest_host_mask()
4050 static u32 vmx_pin_based_exec_ctrl(struct vcpu_vmx *vmx) in vmx_pin_based_exec_ctrl()
4092 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_refresh_apicv_exec_ctrl() local
4109 static u32 vmx_exec_control(struct vcpu_vmx *vmx) in vmx_exec_control()
4141 vmx_adjust_secondary_exec_control(struct vcpu_vmx *vmx, u32 *exec_control, in vmx_adjust_secondary_exec_control()
4172 #define vmx_adjust_sec_exec_control(vmx, exec_control, name, feat_name, ctrl_name, exiting) \ argument
4185 #define vmx_adjust_sec_exec_feature(vmx, exec_control, lname, uname) \ argument
4188 #define vmx_adjust_sec_exec_exiting(vmx, exec_control, lname, uname) \ argument
4191 static u32 vmx_secondary_exec_control(struct vcpu_vmx *vmx) in vmx_secondary_exec_control()
4282 static void init_vmcs(struct vcpu_vmx *vmx) in init_vmcs()
4388 struct vcpu_vmx *vmx = to_vmx(vcpu); in __vmx_vcpu_reset() local
4415 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_vcpu_reset() local
4488 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_inject_irq() local
4516 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_inject_nmi() local
4547 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_get_nmi_mask() local
4561 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_set_nmi_mask() local
4739 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_exception_nmi() local
5195 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_task_switch() local
5341 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_invalid_guest_state() local
5383 struct vcpu_vmx *vmx = to_vmx(vcpu); in grow_ple_window() local
5399 struct vcpu_vmx *vmx = to_vmx(vcpu); in shrink_ple_window() local
5495 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_fastpath_preemption_timer() local
5612 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_get_exit_info() local
5630 static void vmx_destroy_pml_buffer(struct vcpu_vmx *vmx) in vmx_destroy_pml_buffer()
5640 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_flush_pml_buffer() local
5697 struct vcpu_vmx *vmx = to_vmx(vcpu); in dump_vmcs() local
5867 struct vcpu_vmx *vmx = to_vmx(vcpu); in __vmx_handle_exit() local
6149 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_set_virtual_apic_mode() local
6279 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_sync_pir_to_irr() local
6336 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_apicv_post_state_restore() local
6352 static void handle_exception_nmi_irqoff(struct vcpu_vmx *vmx) in handle_exception_nmi_irqoff()
6383 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_handle_exit_irqoff() local
6418 static void vmx_recover_nmi_blocking(struct vcpu_vmx *vmx) in vmx_recover_nmi_blocking()
6512 static void vmx_complete_interrupts(struct vcpu_vmx *vmx) in vmx_complete_interrupts()
6529 static void atomic_switch_perf_msrs(struct vcpu_vmx *vmx) in atomic_switch_perf_msrs()
6549 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_update_hv_timer() local
6573 void noinstr vmx_update_host_rsp(struct vcpu_vmx *vmx, unsigned long host_rsp) in vmx_update_host_rsp()
6594 struct vcpu_vmx *vmx) in vmx_vcpu_enter_exit()
6617 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_vcpu_run() local
6803 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_free_vcpu() local
6815 struct vcpu_vmx *vmx; in vmx_create_vcpu() local
7018 static void vmcs_set_secondary_exec_control(struct vcpu_vmx *vmx, u32 new_ctl) in vmcs_set_secondary_exec_control()
7043 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_cr_fixed1_bits_update() local
7083 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_entry_exit_ctls_update() local
7100 struct vcpu_vmx *vmx = to_vmx(vcpu); in update_intel_pt_cfg() local
7169 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_vcpu_after_set_cpuid() local
7377 struct vcpu_vmx *vmx; in vmx_set_hv_timer() local
7428 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_update_cpu_dirty_logging() local
7485 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_enter_smm() local
7499 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_leave_smm() local