Searched refs:ghc (Results 1 – 6 of 6) sorted by relevance
| /linux/virt/kvm/ |
| A D | kvm_main.c | 2891 ghc->hva = KVM_HVA_ERR_BAD; in __kvm_gfn_to_hva_cache_init() 2901 ghc->hva = gfn_to_hva_many(ghc->memslot, start_gfn, in __kvm_gfn_to_hva_cache_init() 2909 ghc->hva += offset; in __kvm_gfn_to_hva_cache_init() 2911 ghc->memslot = NULL; in __kvm_gfn_to_hva_cache_init() 2913 ghc->gpa = gpa; in __kvm_gfn_to_hva_cache_init() 2914 ghc->len = len; in __kvm_gfn_to_hva_cache_init() 2932 gpa_t gpa = ghc->gpa + offset; in kvm_write_guest_offset_cached() 2938 if (__kvm_gfn_to_hva_cache_init(slots, ghc, ghc->gpa, ghc->len)) in kvm_write_guest_offset_cached() 2945 if (unlikely(!ghc->memslot)) in kvm_write_guest_offset_cached() 2976 if (__kvm_gfn_to_hva_cache_init(slots, ghc, ghc->gpa, ghc->len)) in kvm_read_guest_offset_cached() [all …]
|
| /linux/arch/x86/kvm/ |
| A D | xen.c | 200 struct gfn_to_hva_cache *ghc = &v->arch.xen.vcpu_info_cache; in __kvm_xen_has_interrupt() local 217 if (likely(slots->generation == ghc->generation && in __kvm_xen_has_interrupt() 218 !kvm_is_error_hva(ghc->hva) && ghc->memslot)) { in __kvm_xen_has_interrupt() 221 err = __get_user(rc, (u8 __user *)ghc->hva + offset); in __kvm_xen_has_interrupt() 240 kvm_read_guest_offset_cached(v->kvm, ghc, &rc, offset, in __kvm_xen_has_interrupt()
|
| A D | lapic.c | 2862 struct gfn_to_hva_cache *ghc = &vcpu->arch.pv_eoi.data; in kvm_lapic_set_pv_eoi() local 2870 if (addr == ghc->gpa && len <= ghc->len) in kvm_lapic_set_pv_eoi() 2871 new_len = ghc->len; in kvm_lapic_set_pv_eoi() 2875 ret = kvm_gfn_to_hva_cache_init(vcpu->kvm, ghc, addr, new_len); in kvm_lapic_set_pv_eoi()
|
| A D | x86.c | 3287 struct gfn_to_hva_cache *ghc = &vcpu->arch.st.cache; in record_steal_time() local 3306 if (unlikely(slots->generation != ghc->generation || in record_steal_time() 3307 kvm_is_error_hva(ghc->hva) || !ghc->memslot)) { in record_steal_time() 3314 kvm_is_error_hva(ghc->hva) || !ghc->memslot) in record_steal_time() 3318 st = (struct kvm_steal_time __user *)ghc->hva; in record_steal_time() 3380 mark_page_dirty_in_slot(vcpu->kvm, ghc->memslot, gpa_to_gfn(ghc->gpa)); in record_steal_time() 4421 struct gfn_to_hva_cache *ghc = &vcpu->arch.st.cache; in kvm_steal_time_set_preempted() local 4438 if (unlikely(slots->generation != ghc->generation || in kvm_steal_time_set_preempted() 4439 kvm_is_error_hva(ghc->hva) || !ghc->memslot)) in kvm_steal_time_set_preempted() 4442 st = (struct kvm_steal_time __user *)ghc->hva; in kvm_steal_time_set_preempted() [all …]
|
| /linux/include/linux/ |
| A D | kvm_host.h | 881 int kvm_read_guest_cached(struct kvm *kvm, struct gfn_to_hva_cache *ghc, 883 int kvm_read_guest_offset_cached(struct kvm *kvm, struct gfn_to_hva_cache *ghc, 890 int kvm_write_guest_cached(struct kvm *kvm, struct gfn_to_hva_cache *ghc, 892 int kvm_write_guest_offset_cached(struct kvm *kvm, struct gfn_to_hva_cache *ghc, 895 int kvm_gfn_to_hva_cache_init(struct kvm *kvm, struct gfn_to_hva_cache *ghc,
|
| /linux/arch/x86/kvm/vmx/ |
| A D | nested.c | 674 struct gfn_to_hva_cache *ghc = &vmx->nested.shadow_vmcs12_cache; in nested_cache_shadow_vmcs12() local 680 if (ghc->gpa != vmcs12->vmcs_link_pointer && in nested_cache_shadow_vmcs12() 681 kvm_gfn_to_hva_cache_init(vcpu->kvm, ghc, in nested_cache_shadow_vmcs12() 693 struct gfn_to_hva_cache *ghc = &vmx->nested.shadow_vmcs12_cache; in nested_flush_cached_shadow_vmcs12() local 699 if (ghc->gpa != vmcs12->vmcs_link_pointer && in nested_flush_cached_shadow_vmcs12() 700 kvm_gfn_to_hva_cache_init(vcpu->kvm, ghc, in nested_flush_cached_shadow_vmcs12() 2937 if (ghc->gpa != vmcs12->vmcs_link_pointer && in nested_vmx_check_vmcs_link_ptr() 2938 CC(kvm_gfn_to_hva_cache_init(vcpu->kvm, ghc, in nested_vmx_check_vmcs_link_ptr() 2942 if (CC(kvm_read_guest_offset_cached(vcpu->kvm, ghc, &hdr, in nested_vmx_check_vmcs_link_ptr() 5287 struct gfn_to_hva_cache *ghc = &vmx->nested.vmcs12_cache; in handle_vmptrld() local [all …]
|
Completed in 56 milliseconds