/xen/xen/arch/x86/mm/shadow/ |
A D | common.c | 230 mfn_x(gmfn)); in oos_audit_hash_is_present() 240 ASSERT(mfn_valid(gmfn)); in _sh_resync_l1() 256 ASSERT(mfn_valid(gmfn)); in sh_remove_write_access_from_sl1p() 362 mfn_x(gmfn)); in oos_fixup_add() 478 oos[idx] = gmfn; in oos_hash_add() 514 mfn_x(gmfn)); in oos_hash_remove() 539 mfn_x(gmfn)); in oos_snapshot_lookup() 570 mfn_x(gmfn)); in sh_resync() 667 pg = mfn_to_page(gmfn); in sh_unsync() 683 oos_hash_add(v, gmfn); in sh_unsync() [all …]
|
A D | pv.c | 32 sh_write_guest_entry(struct vcpu *v, intpte_t *p, intpte_t new, mfn_t gmfn) in sh_write_guest_entry() argument 39 sh_validate_guest_entry(v, gmfn, p, sizeof(new)); in sh_write_guest_entry() 53 intpte_t new, mfn_t gmfn) in sh_cmpxchg_guest_entry() argument 61 sh_validate_guest_entry(v, gmfn, p, sizeof(new)); in sh_cmpxchg_guest_entry()
|
A D | multi.c | 1848 mfn_t gmfn, sl4mfn; in sh_destroy_l4_shadow() local 1855 gmfn = backpointer(sp); in sh_destroy_l4_shadow() 1878 mfn_t gmfn, sl3mfn; in sh_destroy_l3_shadow() local 1885 gmfn = backpointer(sp); in sh_destroy_l3_shadow() 1909 mfn_t gmfn, sl2mfn; in sh_destroy_l2_shadow() local 1921 gmfn = backpointer(sp); in sh_destroy_l2_shadow() 2519 mfn_t gmfn; in sh_prefetch() local 2825 if ( mfn_valid(gmfn) && mfn_is_out_of_sync(gmfn) ) in sh_page_fault() 3925 mfn_t gmfn; in sh_update_cr3() local 4476 mfn_t smfn, gmfn; in sh_pagetable_dying() local [all …]
|
A D | private.h | 312 static inline int mfn_is_out_of_sync(mfn_t gmfn) in mfn_is_out_of_sync() argument 314 return page_is_out_of_sync(mfn_to_page(gmfn)); in mfn_is_out_of_sync() 323 static inline int mfn_oos_may_write(mfn_t gmfn) in mfn_oos_may_write() argument 325 return page_oos_may_write(mfn_to_page(gmfn)); in mfn_oos_may_write() 391 mfn_t gmfn); 393 intpte_t new, mfn_t gmfn); 406 int sh_unsync(struct vcpu *v, mfn_t gmfn); 409 void sh_resync(struct domain *d, mfn_t gmfn); 477 sh_mfn_is_a_page_table(mfn_t gmfn) in sh_mfn_is_a_page_table() argument 479 struct page_info *page = mfn_to_page(gmfn); in sh_mfn_is_a_page_table() [all …]
|
A D | hvm.c | 553 static inline void check_for_early_unshadow(struct vcpu *v, mfn_t gmfn) in check_for_early_unshadow() argument 567 && v->arch.paging.shadow.last_emulated_mfn_for_unshadow == mfn_x(gmfn) ) ) in check_for_early_unshadow() 568 && sh_mfn_is_a_page_table(gmfn) in check_for_early_unshadow() 570 !(mfn_to_page(gmfn)->shadow_flags in check_for_early_unshadow() 574 sh_remove_shadows(d, gmfn, 1, 0 /* Fast, can fail to unshadow */ ); in check_for_early_unshadow() 577 v->arch.paging.shadow.last_emulated_mfn_for_unshadow = mfn_x(gmfn); in check_for_early_unshadow() 582 static void validate_guest_pt_write(struct vcpu *v, mfn_t gmfn, in validate_guest_pt_write() argument 590 rc = sh_validate_guest_entry(v, gmfn, entry, size); in validate_guest_pt_write() 605 sh_remove_shadows(d, gmfn, 0, 0); in validate_guest_pt_write()
|
A D | multi.h | 107 (struct vcpu *v, mfn_t gmfn, mfn_t snpmfn); 111 (struct vcpu*v, mfn_t gmfn); 115 (struct domain *d, mfn_t gmfn, mfn_t smfn, unsigned long off);
|
/xen/xen/include/asm-x86/ |
A D | paging.h | 100 intpte_t new, mfn_t gmfn); 103 mfn_t gmfn); 109 unsigned long vaddr, mfn_t gmfn); 170 void paging_mark_dirty(struct domain *d, mfn_t gmfn); 176 int paging_mfn_is_dirty(struct domain *d, mfn_t gmfn); 317 struct vcpu *v, intpte_t *p, intpte_t new, mfn_t gmfn) in paging_write_guest_entry() argument 322 gmfn); in paging_write_guest_entry() 335 struct vcpu *v, intpte_t *p, intpte_t *old, intpte_t new, mfn_t gmfn) in paging_cmpxchg_guest_entry() argument 340 new, gmfn); in paging_cmpxchg_guest_entry()
|
A D | shadow.h | 83 void sh_remove_shadows(struct domain *d, mfn_t gmfn, int fast, int all); 110 static inline void sh_remove_shadows(struct domain *d, mfn_t gmfn, in sh_remove_shadows() argument 256 static inline void shadow_remove_all_shadows(struct domain *d, mfn_t gmfn) in shadow_remove_all_shadows() argument 259 sh_remove_shadows(d, gmfn, 0 /* Be thorough */, 1 /* Must succeed */); in shadow_remove_all_shadows()
|
/xen/xen/common/ |
A D | memory.c | 303 put_gfn(d, gmfn); in guest_remove_page() 324 put_gfn(d, gmfn); in guest_remove_page() 336 mfn = gfn_to_mfn(d, _gfn(gmfn)); in guest_remove_page() 341 put_gfn(d, gmfn); in guest_remove_page() 344 d->domain_id, gmfn); in guest_remove_page() 374 put_gfn(d, gmfn); in guest_remove_page() 410 put_gfn(d, gmfn); in guest_remove_page() 423 xen_pfn_t gmfn; in decrease_reservation() local 450 t.gfn = gmfn; in decrease_reservation() 522 xen_pfn_t gpfn, gmfn; in memory_exchange() local [all …]
|
A D | grant_table.c | 2038 xen_pfn_t gmfn = gfn_x(gnttab_shared_gfn(d, gt, i)); in gnttab_setup_table() local 2041 BUG_ON(SHARED_M2P(gmfn)); in gnttab_setup_table() 2043 if ( __copy_to_guest_offset(op.frame_list, i, &gmfn, 1) ) in gnttab_setup_table() 2884 rc = get_paged_frame(ptr->u.gmfn, &buf->mfn, &buf->page, in gnttab_copy_claim_buf() 2888 "source frame %"PRI_xen_pfn" invalid\n", ptr->u.gmfn); in gnttab_copy_claim_buf() 2890 buf->ptr.u.gmfn = ptr->u.gmfn; in gnttab_copy_claim_buf() 2924 return p->u.gmfn == b->ptr.u.gmfn; in gnttab_copy_buf_valid() 3222 uint64_t gmfn; in gnttab_get_status_frames() local 3276 gmfn = gfn_x(gnttab_status_gfn(d, gt, i)); in gnttab_get_status_frames() 3277 if ( copy_to_guest_offset(op.frame_list, i, &gmfn, 1) ) in gnttab_get_status_frames()
|
/xen/xen/drivers/passthrough/vtd/ |
A D | utils.c | 82 static u32 get_level_index(unsigned long gmfn, int level) in get_level_index() argument 85 gmfn = gmfn >> LEVEL_STRIDE; in get_level_index() 87 return gmfn & LEVEL_MASK; in get_level_index() 90 void print_vtd_entries(struct vtd_iommu *iommu, int bus, int devfn, u64 gmfn) in print_vtd_entries() argument 100 PCI_SLOT(devfn), PCI_FUNC(devfn), gmfn); in print_vtd_entries() 159 l_index = get_level_index(gmfn, level); in print_vtd_entries()
|
A D | extern.h | 34 void print_vtd_entries(struct vtd_iommu *iommu, int bus, int devfn, u64 gmfn);
|
/xen/tools/libxc/ |
A D | xc_core.c | 818 uint64_t gmfn; in xc_domain_dumpcore_via_callback() local 836 gmfn = p2m[i]; in xc_domain_dumpcore_via_callback() 838 gmfn = ((uint64_t *)p2m)[i]; in xc_domain_dumpcore_via_callback() 839 if ( gmfn == INVALID_PFN ) in xc_domain_dumpcore_via_callback() 844 gmfn = ((uint32_t *)p2m)[i]; in xc_domain_dumpcore_via_callback() 845 if ( gmfn == (uint32_t)INVALID_PFN ) in xc_domain_dumpcore_via_callback() 850 p2m_array[j].gmfn = gmfn; in xc_domain_dumpcore_via_callback() 857 gmfn = i; in xc_domain_dumpcore_via_callback() 862 xch, domid, PAGE_SIZE, PROT_READ, gmfn); in xc_domain_dumpcore_via_callback() 899 p2m_array[j].gmfn = XC_CORE_INVALID_GMFN; in xc_domain_dumpcore_via_callback()
|
A D | xc_dom_boot.c | 56 domctl.u.hypercall_init.gmfn = xc_dom_p2m(dom, pfn); in setup_hypercall_page() 230 xen_pfn_t gmfn; in xc_dom_gnttab_setup() local 242 gmfn = *gmfnp; in xc_dom_gnttab_setup() 254 return gmfn; in xc_dom_gnttab_setup()
|
A D | xc_core.h | 111 uint64_t gmfn; member
|
/xen/xen/arch/x86/hvm/viridian/ |
A D | viridian.c | 245 unsigned long gmfn = d->arch.hvm.viridian->hypercall_gpa.pfn; in enable_hypercall_page() local 246 struct page_info *page = get_page_from_gfn(d, gmfn, NULL, P2M_ALLOC); in enable_hypercall_page() 254 gmfn, mfn_x(page ? page_to_mfn(page) : INVALID_MFN)); in enable_hypercall_page() 722 unsigned long gmfn = vp->msr.pfn; in viridian_map_guest_page() local 728 page = get_page_from_gfn(d, gmfn, NULL, P2M_ALLOC); in viridian_map_guest_page() 750 gmfn, mfn_x(page ? page_to_mfn(page) : INVALID_MFN)); in viridian_map_guest_page()
|
/xen/docs/misc/ |
A D | dump-core-format.txt | 79 uint64_t gmfn; 85 gmfn: machine physical frame number 90 (pfn, gmfn) and the corresponding page has zero. There might 91 exist invalid (pfn, gmfn)'s at the end part of this array. 241 Invalid pfn/gmfn.
|
/xen/xen/arch/x86/mm/ |
A D | paging.c | 338 void paging_mark_dirty(struct domain *d, mfn_t gmfn) in paging_mark_dirty() argument 342 if ( !paging_mode_log_dirty(d) || !mfn_valid(gmfn) || in paging_mark_dirty() 343 page_get_owner(mfn_to_page(gmfn)) != d ) in paging_mark_dirty() 347 pfn = _pfn(get_gpfn_from_mfn(mfn_x(gmfn))); in paging_mark_dirty() 354 int paging_mfn_is_dirty(struct domain *d, mfn_t gmfn) in paging_mfn_is_dirty() argument 365 pfn = _pfn(get_gpfn_from_mfn(mfn_x(gmfn))); in paging_mfn_is_dirty()
|
/xen/xen/include/public/ |
A D | physdev.h | 65 xen_pfn_t gmfn; member
|
A D | grant_table.h | 462 xen_pfn_t gmfn; member
|
A D | domctl.h | 462 uint64_aligned_t gmfn; /* GMFN to be initialised */ member
|
/xen/xen/include/xen/ |
A D | mm.h | 617 int __must_check guest_remove_page(struct domain *d, unsigned long gmfn); 633 int prepare_ring_for_helper(struct domain *d, unsigned long gmfn,
|
/xen/xen/arch/x86/ |
A D | traps.c | 912 unsigned long gmfn = val >> PAGE_SHIFT; in guest_wrmsr_xen() local 925 page = get_page_from_gfn(d, gmfn, &t, P2M_ALLOC); in guest_wrmsr_xen() 934 p2m_mem_paging_populate(d, _gfn(gmfn)); in guest_wrmsr_xen() 940 gmfn, mfn_x(page ? page_to_mfn(page) : INVALID_MFN), base); in guest_wrmsr_xen()
|
A D | domctl.c | 465 unsigned long gmfn = domctl->u.hypercall_init.gmfn; in arch_do_domctl() local 469 page = get_page_from_gfn(d, gmfn, NULL, P2M_ALLOC); in arch_do_domctl()
|
/xen/tools/xentrace/ |
A D | xenalyze.c | 1584 unsigned long long gmfn; member 2565 p->gmfn, in interval_cr3_schedule_ordered_output() 3906 if(gmfn) { in cr3_switch() 3912 if(p->gmfn == gmfn) in cr3_switch() 3923 p->gmfn = gmfn; in cr3_switch() 3947 p->gmfn); in cr3_switch() 3979 if(cr3->gmfn == gmfn) in cr3_prealloc_unpin() 3987 gmfn); in cr3_prealloc_unpin() 4772 cur->gmfn); in hvm_handler_process() 5674 e->gmfn = r.gmfn; [all …]
|