| /xen/xen/arch/x86/mm/ |
| A D | p2m-pod.c | 79 p = mfn_to_page(mfn_add(mfn, i)); in p2m_pod_cache_add() 99 clear_domain_page(mfn_add(page_to_mfn(page), i)); in p2m_pod_cache_add() 747 else if ( type != type0 || !mfn_eq(mfn, mfn_add(mfn0, i)) ) in p2m_pod_zero_check_superpage() 763 map = map_domain_page(mfn_add(mfn0, i)); in p2m_pod_zero_check_superpage() 790 mfn = mfn_add(mfn0, i); in p2m_pod_zero_check_superpage() 801 map = map_domain_page(mfn_add(mfn0, i)); in p2m_pod_zero_check_superpage()
|
| A D | p2m.c | 649 mfn = mfn_add(mfn, 1ul << order); in p2m_set_entry() 967 page_get_owner(mfn_to_page(mfn_add(mfn, i))) == dom_cow ) in guest_physmap_add_entry() 971 mfn_x(mfn_add(mfn, i)), d->domain_id); in guest_physmap_add_entry() 975 if ( page_get_owner(mfn_to_page(mfn_add(mfn, i))) != d ) in guest_physmap_add_entry() 977 ogfn = mfn_to_gfn(d, mfn_add(mfn, i)); in guest_physmap_add_entry() 984 mfn_x(mfn_add(mfn, i)), gfn_x(ogfn), in guest_physmap_add_entry() 992 if ( mfn_eq(omfn, mfn_add(mfn, i)) && in guest_physmap_add_entry() 1011 set_gpfn_from_mfn(mfn_x(mfn_add(mfn, i)), in guest_physmap_add_entry() 1298 ASSERT(mfn_valid(mfn_add(omfn, i))); in set_typed_p2m_entry() 1912 mfn_add(mfn, i), order); in map_mmio_regions() [all …]
|
| /xen/xen/arch/arm/ |
| A D | mm.c | 255 mapping = map_domain_page(mfn_add(root_mfn, root_table)); in dump_pt_walk() 527 return mfn_add(lpae_get_mfn(map[slot]), offset); in domain_page_map_to_mfn() 892 unsigned long nr_pdxs = mfn_to_pdx(mfn_add(maddr_to_mfn(pe), -1)) - in setup_frametable_mappings() 916 clear_page(mfn_to_virt(mfn_add(second_base, i))); in setup_frametable_mappings() 917 pte = mfn_to_xen_entry(mfn_add(second_base, i), MT_NORMAL); in setup_frametable_mappings() 1278 mfn = mfn_add(mfn, 1); in xen_pt_update()
|
| A D | p2m.c | 423 mfn = mfn_add(mfn, gfn_x(gfn) & ((1UL << level_orders[level]) - 1)); in p2m_get_entry() 830 lpae_set_mfn(pte, mfn_add(mfn, i << level_order)); in p2m_split_superpage() 1118 smfn = mfn_add(smfn, (1 << order)); in p2m_set_entry() 1707 mfn = mfn_add(mfn, 1); in p2m_cache_flush_range()
|
| A D | setup.c | 696 if ( s < mfn_to_maddr(mfn_add(xenheap_mfn_start, xenheap_pages)) in setup_mm() 700 n = mfn_to_maddr(mfn_add(xenheap_mfn_start, xenheap_pages)); in setup_mm()
|
| /xen/xen/common/ |
| A D | xenoprof.c | 191 struct page_info *page = mfn_to_page(mfn_add(mfn, i)); in share_xenoprof_page_with_guest() 196 d->domain_id, mfn_x(mfn_add(mfn, i)), page->count_info); in share_xenoprof_page_with_guest() 203 share_xen_page_with_guest(mfn_to_page(mfn_add(mfn, i)), d, SHARE_rw); in share_xenoprof_page_with_guest() 216 struct page_info *page = mfn_to_page(mfn_add(mfn, i)); in unshare_xenoprof_page_with_guest()
|
| A D | sysctl.c | 213 mfn = mfn_add(mfn, 1) ) in do_sysctl()
|
| A D | page_alloc.c | 1806 unsigned long e = mfn_x(mfn_add(page_to_mfn(pg + nr_pages - 1), 1)); in init_heap_pages() 2132 if ( ps && !is_xen_heap_mfn(mfn_add(maddr_to_mfn(ps), -1)) ) in init_xenheap_pages()
|
| A D | memory.c | 225 mfn = mfn_add(mfn, 1) ) in populate_physmap()
|
| /xen/xen/drivers/passthrough/ |
| A D | iommu.c | 259 mfn_add(mfn, i), flags, flush_flags); in iommu_map() 268 mfn_x(mfn_add(mfn, i)), rc); in iommu_map()
|
| /xen/xen/arch/x86/x86_64/ |
| A D | mmconfig_64.c | 128 mfn_add(maddr_to_mfn(cfg->address), in mcfg_ioremap()
|
| A D | mm.c | 186 struct page_info *page = mfn_to_page(mfn_add(m2p_start_mfn, i)); in share_hotadd_m2p_table() 188 if ( hotadd_mem_valid(mfn_x(mfn_add(m2p_start_mfn, i)), info) ) in share_hotadd_m2p_table() 208 struct page_info *page = mfn_to_page(mfn_add(m2p_start_mfn, i)); in share_hotadd_m2p_table() 210 if ( hotadd_mem_valid(mfn_x(mfn_add(m2p_start_mfn, i)), info) ) in share_hotadd_m2p_table()
|
| /xen/xen/include/xen/ |
| A D | mm.h | 80 static inline mfn_t mfn_add(mfn_t mfn, unsigned long i) in mfn_add() function
|
| /xen/xen/arch/x86/hvm/ |
| A D | mtrr.c | 842 if ( is_special_page(mfn_to_page(mfn_add(mfn, i))) ) in epte_get_entry_emt()
|
| /xen/xen/arch/x86/ |
| A D | mm.c | 5298 mfn = mfn_add(mfn, 1UL << (L3_PAGETABLE_SHIFT - PAGE_SHIFT)); in map_pages_to_xen() 5325 mfn = mfn_add(mfn, i); in map_pages_to_xen() 5395 mfn = mfn_add(mfn, 1UL << PAGETABLE_ORDER); in map_pages_to_xen() 5426 mfn = mfn_add(mfn, i); in map_pages_to_xen() 5472 mfn = mfn_add(mfn, 1UL); in map_pages_to_xen()
|
| /xen/xen/arch/x86/mm/shadow/ |
| A D | common.c | 3050 omfn = mfn_add(omfn, 1); in sh_unshadow_for_p2m_change()
|