| /xen/xen/drivers/passthrough/arm/ |
| A D | iommu_helpers.c | 60 return guest_physmap_add_entry(d, _gfn(dfn_x(dfn)), _mfn(dfn_x(dfn)), 0, t); in arm_iommu_map_page() 74 return guest_physmap_remove_page(d, _gfn(dfn_x(dfn)), _mfn(dfn_x(dfn)), 0); in arm_iommu_unmap_page()
|
| /xen/xen/include/asm-x86/ |
| A D | guest_pt.h | 78 { return _gfn(gl1e.l1 >> PAGE_SHIFT); } in guest_l1e_get_gfn() 80 { return _gfn(gl2e.l2 >> PAGE_SHIFT); } in guest_l2e_get_gfn() 151 { return _gfn(l1e_get_paddr(gl1e) >> PAGE_SHIFT); } in guest_l1e_get_gfn() 153 { return _gfn(l2e_get_paddr(gl2e) >> PAGE_SHIFT); } in guest_l2e_get_gfn() 155 { return _gfn(l3e_get_paddr(gl3e) >> PAGE_SHIFT); } in guest_l3e_get_gfn() 158 { return _gfn(l4e_get_paddr(gl4e) >> PAGE_SHIFT); } in guest_l4e_get_gfn() 374 ? _gfn(gw->el1e >> PAGE_SHIFT) in guest_walk_to_gfn()
|
| A D | grant_table.h | 45 VALID_M2P(gpfn_) ? _gfn(gpfn_) : INVALID_GFN; \
|
| A D | p2m.h | 495 return p2m_get_page_from_gfn(p2m_get_hostp2m(d), _gfn(gfn), t, NULL, q); in get_page_from_gfn() 508 return _gfn(get_gpfn_from_mfn(mfn_x(mfn))); in mfn_to_gfn() 510 return _gfn(mfn_x(mfn)); in mfn_to_gfn() 549 rval-> dest ## _gfn = source ## gfn; \ in get_two_gfns()
|
| /xen/xen/arch/x86/hvm/ |
| A D | grant_table.c | 43 _gfn(addr >> PAGE_SHIFT), in create_grant_p2m_mapping() 71 if ( guest_physmap_remove_page(d, _gfn(gfn), frame, PAGE_ORDER_4K) ) in replace_grant_p2m_mapping()
|
| A D | dom0_build.c | 83 rc = map ? map_mmio_regions(d, _gfn(pfn), nr_pages, _mfn(pfn)) in modify_identity_mmio() 84 : unmap_mmio_regions(d, _gfn(pfn), nr_pages, _mfn(pfn)); in modify_identity_mmio() 177 rc = guest_physmap_add_page(d, _gfn(start), page_to_mfn(page), in pvh_populate_memory_range() 319 ident_pt = map_domain_gfn(p2m_get_hostp2m(d), _gfn(PFN_DOWN(gaddr)), in pvh_setup_vmx_realmode_helpers()
|
| A D | dm.c | 281 p2m_mem_paging_populate(d, _gfn(pfn)); in set_mem_type() 505 rc = p2m_finish_type_change(d, _gfn(first_gfn), 256); in dm_op()
|
| /xen/xen/arch/x86/mm/hap/ |
| A D | nested_ept.c | 157 gfn_t base_gfn = _gfn(nhvm_vcpu_p2m_base(v) >> PAGE_SHIFT); in nept_walk_tables() 183 gfn_t start = _gfn(gw->lxe[lvl].mfn); in nept_walk_tables() 185 start = _gfn((gfn_x(start) & ~gfn_lvl_mask) + in nept_walk_tables() 193 base_gfn = _gfn(gw->lxe[lvl].mfn); in nept_walk_tables()
|
| A D | guest_walk.c | 62 top_gfn = _gfn(cr3 >> PAGE_SHIFT); in hap_p2m_ga_to_gfn()
|
| A D | nested_hap.c | 118 rc = p2m_set_entry(p2m, _gfn(gfn), mfn, page_order, p2mt, p2ma); in nestedhap_fix_p2m()
|
| /xen/xen/common/ |
| A D | mem_access.c | 69 rc = p2m_set_mem_access(d, _gfn(mao.pfn), mao.nr, start_iter, in mem_access_memop() 102 rc = p2m_get_mem_access(d, _gfn(mao.pfn), &access, 0); in mem_access_memop()
|
| A D | memory.c | 320 rc = guest_physmap_remove_page(d, _gfn(gmfn), mfn, 0); in guest_remove_page() 326 p2m_mem_paging_drop_page(d, _gfn(gmfn), p2mt); in guest_remove_page() 336 mfn = gfn_to_mfn(d, _gfn(gmfn)); in guest_remove_page() 391 rc = guest_physmap_remove_page(d, _gfn(gmfn), mfn, 0); in guest_remove_page() 459 p2m_pod_decrease_reservation(a->domain, _gfn(gmfn), in decrease_reservation() 656 mfn = gfn_to_mfn(d, _gfn(gmfn + k)); in memory_exchange() 716 if ( guest_physmap_remove_page(d, _gfn(gfn), mfn, 0) ) in memory_exchange() 761 guest_physmap_add_page(d, _gfn(gpfn), mfn, in memory_exchange() 846 xatp->idx, _gfn(xatp->gpfn)); in xenmem_add_to_physmap() 938 idx, _gfn(gpfn)); in xenmem_add_to_physmap_batch() [all …]
|
| A D | grant_table.c | 468 rc = check_get_page_from_gfn(rd, _gfn(gfn), readonly, &p2mt, page); in get_paged_frame() 1087 act_set_gfn(act, _gfn(gfn)); in map_grant_ref() 2208 mfn = gfn_to_mfn(d, _gfn(gop.mfn)); in gnttab_transfer() 2232 rc = guest_physmap_remove_page(d, _gfn(gop.mfn), mfn, 0); in gnttab_transfer() 2367 guest_physmap_add_page(e, _gfn(sha->frame), mfn, 0); in gnttab_transfer() 2669 act_set_gfn(act, _gfn(gfn)); in acquire_grant_for_copy() 2680 act_set_gfn(act, _gfn(sha2->full_page.frame)); in acquire_grant_for_copy() 2691 act_set_gfn(act, _gfn(sha2->sub_page.frame)); in acquire_grant_for_copy() 3916 *gfn = _gfn(sha1->frame); in mem_sharing_gref_to_gfn() 3924 *gfn = _gfn(sha2->sub_page.frame); in mem_sharing_gref_to_gfn() [all …]
|
| /xen/xen/arch/x86/mm/ |
| A D | mem_paging.c | 181 gfn_t gfn = _gfn(rsp->u.mem_access.gfn); in p2m_mem_paging_resume() 466 rc = nominate(d, _gfn(mpo.gfn)); in mem_paging_memop() 470 rc = evict(d, _gfn(mpo.gfn)); in mem_paging_memop() 474 rc = prepare(d, _gfn(mpo.gfn), mpo.buffer); in mem_paging_memop()
|
| A D | mem_sharing.c | 1112 get_two_gfns(sd, _gfn(sgfn), &smfn_type, NULL, &smfn, in add_to_physmap() 1147 ret = p2m_set_entry(p2m, _gfn(cgfn), smfn, PAGE_ORDER_4K, in add_to_physmap() 1356 mfn = p2m->get_entry(p2m, _gfn(gfn), &t, &a, 0, NULL, NULL); in relinquish_shared_pages() 1404 rc = nominate_page(d, _gfn(start), 0, false, &sh); in range_share() 1410 rc = nominate_page(cd, _gfn(start), 0, false, &ch); in range_share() 1417 rc = share_pages(d, _gfn(start), sh, cd, _gfn(start), ch); in range_share() 1705 old_gfn = _gfn(get_gpfn_from_mfn(mfn_x(old_mfn))); in copy_special_pages() 1706 new_gfn = _gfn(get_gpfn_from_mfn(mfn_x(new_mfn))); in copy_special_pages() 1937 sgfn = _gfn(mso.u.share.source_gfn); in mem_sharing_memop() 1953 cgfn = _gfn(mso.u.share.client_gfn); in mem_sharing_memop() [all …]
|
| A D | p2m-pod.c | 581 p2m_pod_zero_check_superpage(p2m, _gfn(gfn_x(gfn) & ~(SUPERPAGE_PAGES - 1))) ) in p2m_pod_decrease_reservation() 1023 if ( gfn_eq(p2m->pod.reclaim_single, _gfn(0)) ) in p2m_pod_emergency_sweep() 1039 (void)p2m->get_entry(p2m, _gfn(i), &t, &a, 0, NULL, NULL); in p2m_pod_emergency_sweep() 1042 gfns[j] = _gfn(i); in p2m_pod_emergency_sweep() 1066 p2m->pod.reclaim_single = _gfn(i ? i - 1 : i); in p2m_pod_emergency_sweep() 1084 gfn_t gfn = _gfn(mrp->list[idx]); in pod_eager_reclaim() 1090 gfn = _gfn(gfn_x(gfn) & ~POD_LAST_SUPERPAGE); in pod_eager_reclaim() 1127 gfn_t gfn_aligned = _gfn((gfn_x(gfn) >> order) << order); in p2m_pod_demand_populate() 1275 gfn_t gfn = _gfn(gfn_l); in guest_physmap_mark_populate_on_demand()
|
| A D | guest_walk.c | 223 start = _gfn((gfn_x(start) & ~GUEST_L3_GFN_MASK) + in guest_walk_tables() 318 gfn_t start = _gfn(unfold_pse36(gw->l2e.l2) >> PAGE_SHIFT); in guest_walk_tables() 340 start = _gfn((gfn_x(start) & ~GUEST_L2_GFN_MASK) + in guest_walk_tables()
|
| A D | mem_access.c | 98 if ( _p2m_get_mem_access(p2m, _gfn(data->gfn), &access) == 0 ) in p2m_mem_access_emulate_check() 395 rc = set_mem_access(d, p2m, ap2m, a, _gfn(gfn_l)); in p2m_set_mem_access() 462 rc = set_mem_access(d, p2m, ap2m, a, _gfn(gfn_l)); in p2m_set_mem_access_multi()
|
| A D | p2m.c | 491 gfn_t gfn = _gfn(gfn_l); in __get_gfn_type_access() 978 if ( !gfn_eq(ogfn, _gfn(INVALID_M2P_ENTRY)) && in guest_physmap_add_entry() 1033 gfn_t gfn = _gfn(gfn_l); in p2m_change_type_one() 1272 gfn_t gfn = _gfn(gfn_l); in set_typed_p2m_entry() 1346 gfn_t gfn = _gfn(gfn_l); in set_identity_p2m_entry() 1395 gfn_t gfn = _gfn(gfn_l); in clear_mmio_p2m_entry() 1437 gfn_t gfn = _gfn(gfn_l); in clear_identity_p2m_entry() 1475 gfn_t gfn = _gfn(gfn_l); in set_shared_p2m_entry() 1982 gfn_t gfn_aligned = _gfn(gfn_x(gfn) & mask); in altp2m_get_effective_entry() 2081 gfn = _gfn(gfn_l & mask); in p2m_altp2m_get_or_propagate() [all …]
|
| /xen/xen/include/xen/ |
| A D | mm.h | 102 #define INVALID_GFN _gfn(~0UL) 111 #define _gfn macro 114 #undef _gfn 120 return _gfn(gfn_x(gfn) + i); in gfn_add() 125 return _gfn(max(gfn_x(x), gfn_x(y))); in gfn_max() 130 return _gfn(min(gfn_x(x), gfn_x(y))); in gfn_min()
|
| /xen/xen/arch/arm/ |
| A D | domctl.c | 32 info.gfn = _gfn(vuart_op->gfn); in handle_vuart_init() 55 gfn_t s = _gfn(domctl->u.cacheflush.start_pfn); in arch_do_domctl()
|
| /xen/xen/include/asm-arm/ |
| A D | p2m.h | 329 return p2m_get_page_from_gfn(d, _gfn(gfn), t); in get_page_from_gfn() 390 gfn = _gfn(gfn_x(gfn) & ~((1UL << order) - 1)); in gfn_next_boundary()
|
| /xen/xen/arch/x86/mm/shadow/ |
| A D | types.h | 337 return _gfn(MASK_EXTR(sl1e.l1, SH_L1E_MMIO_GFN_MASK)); in sh_l1e_mmio_get_gfn() 348 #define sh_l1e_mmio(_gfn, _flags) shadow_l1e_empty() argument
|
| /xen/xen/arch/x86/ |
| A D | debug.c | 36 *gfn = _gfn(paging_gva_to_gfn(dp->vcpu[0], vaddr, &pfec)); in dbg_hvm_va2mfn()
|
| /xen/xen/drivers/vpci/ |
| A D | header.c | 55 rc = map->map ? map_mmio_regions(map->d, _gfn(s), size, _mfn(s)) in map_range() 56 : unmap_mmio_regions(map->d, _gfn(s), size, _mfn(s)); in map_range()
|