| /xen/xen/arch/arm/platforms/ |
| A D | midway.c | 30 BUILD_BUG_ON((MW_SREG_PWR_REQ & PAGE_MASK) != in midway_reset() 31 (MW_SREG_A15_PWR_CTRL & PAGE_MASK)); in midway_reset() 33 pmu = ioremap_nocache(MW_SREG_PWR_REQ & PAGE_MASK, PAGE_SIZE); in midway_reset() 40 writel(MW_PWR_HARD_RESET, pmu + (MW_SREG_PWR_REQ & ~PAGE_MASK)); in midway_reset() 41 writel(1, pmu + (MW_SREG_A15_PWR_CTRL & ~PAGE_MASK)); in midway_reset()
|
| A D | sunxi.c | 60 wdt = ioremap_nocache(wdt_start & PAGE_MASK, PAGE_SIZE); in sunxi_map_watchdog() 70 return wdt + (wdt_start & ~PAGE_MASK); in sunxi_map_watchdog()
|
| /xen/xen/include/asm-x86/ |
| A D | page.h | 11 #define PAGE_MASK (~(PAGE_SIZE-1)) macro 13 #define PAGE_OFFSET(ptr) ((unsigned long)(ptr) & ~PAGE_MASK) 83 ((paddr_t)(((x).l1 & (PADDR_MASK&PAGE_MASK)))) 85 ((paddr_t)(((x).l2 & (PADDR_MASK&PAGE_MASK)))) 87 ((paddr_t)(((x).l3 & (PADDR_MASK&PAGE_MASK)))) 89 ((paddr_t)(((x).l4 & (PADDR_MASK&PAGE_MASK)))) 134 ASSERT((pa & ~(PADDR_MASK & PAGE_MASK)) == 0); in l1e_from_paddr() 139 ASSERT((pa & ~(PADDR_MASK & PAGE_MASK)) == 0); in l2e_from_paddr() 144 ASSERT((pa & ~(PADDR_MASK & PAGE_MASK)) == 0); in l3e_from_paddr() 149 ASSERT((pa & ~(PADDR_MASK & PAGE_MASK)) == 0); in l4e_from_paddr() [all …]
|
| A D | fixmap.h | 83 #define __virt_to_fix(x) ((FIXADDR_TOP - ((x)&PAGE_MASK)) >> PAGE_SHIFT)
|
| A D | x86-defns.h | 49 #define X86_CR3_ADDR_MASK (PAGE_MASK & PADDR_MASK)
|
| /xen/misc/coverity/ |
| A D | model.c | 38 #define PAGE_MASK (~(PAGE_SIZE-1)) macro 95 if ( ptr & ~PAGE_MASK ) in map_domain_page() 107 unsigned long ptr = (unsigned long)va & PAGE_MASK; in unmap_domain_page()
|
| /xen/xen/include/xen/ |
| A D | pfn.h | 9 #define round_pgup(p) (((p) + (PAGE_SIZE - 1)) & PAGE_MASK) 10 #define round_pgdown(p) ((p) & PAGE_MASK)
|
| A D | vmap.h | 32 vunmap((void *)(addr & PAGE_MASK)); in iounmap()
|
| /xen/xen/drivers/acpi/apei/ |
| A D | apei-io.c | 83 pg = ((((paddr + size -1) & PAGE_MASK) in apei_range_map() 84 - (paddr & PAGE_MASK)) >> PAGE_SHIFT) + 1; in apei_range_map() 127 map->paddr = paddr & PAGE_MASK; in apei_pre_map() 128 map->size = (((paddr + size + PAGE_SIZE -1) & PAGE_MASK) in apei_pre_map() 129 - (paddr & PAGE_MASK)); in apei_pre_map()
|
| /xen/xen/include/asm-arm/ |
| A D | page.h | 15 #define PAGE_OFFSET(ptr) ((vaddr_t)(ptr) & ~PAGE_MASK) 282 *paddr = (par & PADDR_MASK & PAGE_MASK) | ((unsigned long) va & ~PAGE_MASK); in gva_to_ipa() 291 #define PAGE_ALIGN(x) (((x) + PAGE_SIZE - 1) & PAGE_MASK)
|
| A D | mm.h | 246 return (par & PADDR_MASK & PAGE_MASK) | (va & ~PAGE_MASK); in __virt_to_maddr() 279 *pa = (par & PADDR_MASK & PAGE_MASK) | ((unsigned long) va & ~PAGE_MASK); in gvirt_to_maddr()
|
| A D | early_printk.h | 18 (FIXMAP_ADDR(FIXMAP_CONSOLE) + (CONFIG_EARLY_UART_BASE_ADDRESS & ~PAGE_MASK))
|
| A D | config.h | 181 #define PAGE_MASK (~(PAGE_SIZE-1)) macro
|
| /xen/xen/arch/x86/x86_64/ |
| A D | kexec_reloc.S | 113 andq $PAGE_MASK, %rdi 119 andq $PAGE_MASK, %rbx 128 andq $PAGE_MASK, %rsi
|
| /xen/xen/arch/x86/mm/shadow/ |
| A D | hvm.c | 487 if ( likely(((vaddr + bytes - 1) & PAGE_MASK) == (vaddr & PAGE_MASK)) ) in sh_emulate_map_dest() 491 map = map_domain_page(sh_ctxt->mfn[0]) + (vaddr & ~PAGE_MASK); in sh_emulate_map_dest() 506 v, (vaddr + bytes - 1) & PAGE_MASK, sh_ctxt); in sh_emulate_map_dest() 528 map += (vaddr & ~PAGE_MASK); in sh_emulate_map_dest() 669 b1 = PAGE_SIZE - (((unsigned long)addr) & ~PAGE_MASK); in sh_emulate_unmap_dest() 686 vunmap((void *)((unsigned long)addr & PAGE_MASK)); in sh_emulate_unmap_dest()
|
| /xen/xen/arch/x86/hvm/ |
| A D | emulate.c | 230 unsigned int off = p.addr & ~PAGE_MASK; in hvmemul_do_io() 694 return mapping + (linear & ~PAGE_MASK); in hvmemul_map_linear_addr() 932 if ( (gpa & ~PAGE_MASK) + size > PAGE_SIZE ) in hvmemul_phys_mmio_access() 1046 vio->mmio_gla = gla & PAGE_MASK; in latch_linear_to_phys() 1058 unsigned long offset = gla & ~PAGE_MASK; in hvmemul_linear_mmio_access() 1140 return (vio->mmio_gla == (addr & PAGE_MASK) && in known_gla() 1149 unsigned int offset = addr & ~PAGE_MASK; in linear_read() 1204 unsigned int offset = addr & ~PAGE_MASK; in linear_write() 1869 bytes = PAGE_SIZE - (daddr & ~PAGE_MASK); in hvmemul_rep_movs() 2006 bytes = PAGE_SIZE - (addr & ~PAGE_MASK); in hvmemul_rep_stos() [all …]
|
| A D | monitor.c | 238 paddr_t gpa = (gfn_to_gaddr(gfn) | (gla & ~PAGE_MASK)); in hvm_monitor_check_p2m() 298 req.u.mem_access.offset = gpa & ~PAGE_MASK; in hvm_monitor_check_p2m()
|
| /xen/xen/common/ |
| A D | kimage.c | 128 if ( (mstart & ~PAGE_MASK) || (mend & ~PAGE_MASK) ) in do_kimage_alloc() 463 return kimage_add_entry(image, (destination & PAGE_MASK) | IND_DESTINATION); in kimage_set_destination() 469 return kimage_add_entry(image, (maddr & PAGE_MASK) | IND_SOURCE); in kimage_add_page() 557 destination = entry & PAGE_MASK; in kimage_dst_used() 641 *old = (addr & ~PAGE_MASK) | IND_SOURCE; in kimage_alloc_page() 785 for ( addr = segment->dest_maddr & PAGE_MASK; in kimage_load_segment()
|
| /xen/xen/arch/x86/guest/hyperv/ |
| A D | tlb.c | 49 gva_list[n] = cur & PAGE_MASK; in fill_gva_list() 57 gva_list[n] |= ~PAGE_MASK; in fill_gva_list()
|
| /xen/tools/libs/call/ |
| A D | private.h | 17 #define PAGE_MASK (~(PAGE_SIZE-1)) macro
|
| /xen/xen/arch/x86/pv/ |
| A D | grant_table.c | 81 pl1e = map_domain_page(gl1mfn) + (addr & ~PAGE_MASK); in create_grant_pv_mapping() 237 pl1e = map_domain_page(gl1mfn) + (addr & ~PAGE_MASK); in replace_grant_pv_mapping()
|
| A D | ro-page-fault.c | 75 if ( unlikely(((addr ^ ptwr_ctxt->cr2) & PAGE_MASK) || in ptwr_emulated_update() 167 pl1e = map_domain_page(mfn) + (addr & ~PAGE_MASK); in ptwr_emulated_update()
|
| /xen/tools/libs/foreignmemory/ |
| A D | private.h | 18 #define PAGE_MASK (~(PAGE_SIZE-1)) macro
|
| /xen/xen/arch/arm/ |
| A D | cpuerrata.c | 69 dst_remapped += (vaddr_t)dst & ~PAGE_MASK; in copy_hyp_vect_bpi() 79 vunmap((void *)((vaddr_t)dst_remapped & PAGE_MASK)); in copy_hyp_vect_bpi()
|
| /xen/xen/arch/x86/mm/hap/ |
| A D | guest_walk.c | 92 top_map += (cr3 & ~(PAGE_MASK | 31)); in hap_p2m_ga_to_gfn()
|