/xen/xen/arch/x86/mm/hap/ |
A D | guest_walk.c | 64 P2M_ALLOC | P2M_UNSHARE); in hap_p2m_ga_to_gfn() 106 P2M_ALLOC | P2M_UNSHARE); in hap_p2m_ga_to_gfn()
|
A D | nested_ept.c | 165 lxp = map_domain_gfn(p2m, base_gfn, &lxmfn, P2M_ALLOC, &rc); in nept_walk_tables()
|
/xen/xen/arch/x86/pv/ |
A D | mm.c | 113 page = get_page_from_gfn(currd, l1e_get_pfn(gl1e), NULL, P2M_ALLOC); in pv_map_ldt_shadow_page()
|
A D | descriptor-tables.c | 201 page = get_page_from_gfn(currd, gfn_x(gfn), NULL, P2M_ALLOC); in do_update_descriptor()
|
A D | emul-priv-op.c | 787 page = get_page_from_gfn(currd, gfn, NULL, P2M_ALLOC); in write_cr()
|
/xen/xen/arch/arm/ |
A D | guestcopy.c | 40 page = get_page_from_gfn(info.gpa.d, paddr_to_pfn(addr), &p2mt, P2M_ALLOC); in translate_get_page()
|
A D | mm.c | 1461 page = get_page_from_gfn(od, idx, &p2mt, P2M_ALLOC); in xenmem_add_to_physmap_one()
|
/xen/xen/include/asm-x86/ |
A D | p2m.h | 78 #define P2M_ALLOC (1u<<0) /* Populate PoD and paged-out entries */ macro 448 #define get_gfn(d, g, t) get_gfn_type((d), (g), (t), P2M_ALLOC) 451 P2M_ALLOC | P2M_UNSHARE)
|
A D | guest_pt.h | 43 #define get_gfn(d, g, t) get_gfn_type((d), gfn_x(g), (t), P2M_ALLOC)
|
/xen/xen/include/asm-arm/ |
A D | p2m.h | 311 #define P2M_ALLOC (1u<<0) /* Populate PoD and paged-out entries */ macro
|
/xen/xen/arch/x86/mm/ |
A D | p2m-pt.c | 768 if ( q & P2M_ALLOC ) in p2m_pt_get_entry() 811 if ( q & P2M_ALLOC ) { in p2m_pt_get_entry() 850 if ( q & P2M_ALLOC ) { in p2m_pt_get_entry()
|
A D | guest_walk.c | 102 p2m_query_t qt = P2M_ALLOC | P2M_UNSHARE; in guest_walk_tables()
|
A D | p2m.c | 495 q |= P2M_ALLOC; in __get_gfn_type_access() 512 if ( (q & P2M_ALLOC) && p2m_is_hole(*t) && in __get_gfn_type_access() 533 if ( q & P2M_ALLOC ) in __get_gfn_type_access() 1972 P2M_ALLOC | P2M_UNSHARE, &page_order, 0); in altp2m_get_effective_entry() 2593 page = get_page_from_gfn(fdom, fgfn, &p2mt, P2M_ALLOC); in p2m_add_foreign()
|
A D | mem_access.c | 309 P2M_ALLOC, NULL, false); in set_mem_access()
|
A D | p2m-ept.c | 921 if ( !(q & P2M_ALLOC) ) in ept_get_entry() 946 if ( !(q & P2M_ALLOC) ) in ept_get_entry()
|
/xen/xen/arch/x86/hvm/viridian/ |
A D | viridian.c | 246 struct page_info *page = get_page_from_gfn(d, gmfn, NULL, P2M_ALLOC); in enable_hypercall_page() 728 page = get_page_from_gfn(d, gmfn, NULL, P2M_ALLOC); in viridian_map_guest_page()
|
/xen/xen/arch/x86/hvm/ |
A D | domain.c | 303 NULL, P2M_ALLOC); in arch_set_info_hvm_guest()
|
/xen/xen/common/ |
A D | event_fifo.c | 407 p = get_page_from_gfn(d, gfn, NULL, P2M_ALLOC); in map_guest_page()
|
A D | memory.c | 1432 page = get_page_from_gfn(d, xrfp.gpfn, NULL, P2M_ALLOC); in do_memory_op() 1699 p2m_query_t q = readonly ? P2M_ALLOC : P2M_UNSHARE; in check_get_page_from_gfn()
|
/xen/xen/arch/x86/ |
A D | mm.c | 2174 P2M_ALLOC | P2M_UNSHARE : P2M_ALLOC; in mod_l1_entry() 3469 page = get_page_from_gfn(pg_owner, op.arg1.mfn, NULL, P2M_ALLOC); in do_mmuext_op() 3535 page = get_page_from_gfn(pg_owner, op.arg1.mfn, NULL, P2M_ALLOC); in do_mmuext_op() 3761 page = get_page_from_gfn(pg_owner, op.arg1.mfn, &p2mt, P2M_ALLOC); in do_mmuext_op() 3790 P2M_ALLOC); in do_mmuext_op() 3806 P2M_ALLOC); in do_mmuext_op() 4008 page = get_page_from_gfn(pt_owner, gmfn, &p2mt, P2M_ALLOC); in do_mmu_update()
|
A D | physdev.c | 232 page = get_page_from_gfn(current->domain, info.gmfn, NULL, P2M_ALLOC); in do_physdev_op()
|
A D | domctl.c | 395 page = get_page_from_gfn(d, gfn, &t, P2M_ALLOC); in arch_do_domctl() 469 page = get_page_from_gfn(d, gmfn, NULL, P2M_ALLOC); in arch_do_domctl()
|
/xen/xen/arch/x86/mm/shadow/ |
A D | hvm.c | 419 page = get_page_from_gfn(v->domain, gfn, &p2mt, P2M_ALLOC); in emulate_gva_to_mfn()
|
/xen/xen/arch/x86/cpu/ |
A D | vpmu.c | 631 page = get_page_from_gfn(d, gfn, NULL, P2M_ALLOC); in pvpmu_init()
|
/xen/xen/arch/x86/hvm/vmx/ |
A D | vvmx.c | 725 apic_pg = get_page_from_gfn(v->domain, apic_gpfn, &p2mt, P2M_ALLOC); in nvmx_update_apic_access_address() 746 vapic_pg = get_page_from_gfn(v->domain, vapic_gpfn, &p2mt, P2M_ALLOC); in nvmx_update_virtual_apic_address()
|