| /xen/xen/arch/x86/ |
| A D | debug.c | 32 mfn_t mfn; in dbg_hvm_va2mfn() local 42 mfn = INVALID_MFN; in dbg_hvm_va2mfn() 50 return mfn; in dbg_hvm_va2mfn() 81 mfn = l4e_get_mfn(l4e); in dbg_pv_va2mfn() 88 mfn = l3e_get_mfn(l3e); in dbg_pv_va2mfn() 94 l2t = map_domain_page(mfn); in dbg_pv_va2mfn() 97 mfn = l2e_get_mfn(l2e); in dbg_pv_va2mfn() 102 l1t = map_domain_page(mfn); in dbg_pv_va2mfn() 105 mfn = l1e_get_mfn(l1e); in dbg_pv_va2mfn() 107 return mfn_valid(mfn) ? mfn : INVALID_MFN; in dbg_pv_va2mfn() [all …]
|
| A D | domain_page.c | 70 void *map_domain_page(mfn_t mfn) in map_domain_page() argument 81 return mfn_to_virt(mfn_x(mfn)); in map_domain_page() 86 return mfn_to_virt(mfn_x(mfn)); in map_domain_page() 91 return mfn_to_virt(mfn_x(mfn)); in map_domain_page() 98 if ( hashent->mfn == mfn_x(mfn) ) in map_domain_page() 148 hashent->mfn = ~0UL; in map_domain_page() 203 ASSERT(hashent->mfn == mfn); in unmap_domain_page() 213 hashent->mfn); in unmap_domain_page() 220 hashent->mfn = mfn; in unmap_domain_page() 305 void *map_domain_page_global(mfn_t mfn) in map_domain_page_global() argument [all …]
|
| A D | mm.c | 217 mfn_t mfn; in init_frametable_chunk() local 889 if ( !iomem_access_permitted(pg_owner, mfn, mfn) ) in get_page_from_l1e() 902 !iomem_access_permitted(l1e_owner, mfn, mfn) ) in get_page_from_l1e() 949 .mfn = mfn }); in get_page_from_l1e() 1047 mfn, get_gpfn_from_mfn(mfn), in get_page_from_l1e() 1058 mfn, get_gpfn_from_mfn(mfn), in get_page_from_l1e() 4586 if ( mfn >= s && mfn <= e && mfn < low ) in handle_iomem_range() 5204 mfn_t mfn, in map_pages_to_xen() argument 5325 mfn = mfn_add(mfn, i); in map_pages_to_xen() 5426 mfn = mfn_add(mfn, i); in map_pages_to_xen() [all …]
|
| A D | tboot.c | 184 unsigned long mfn; in update_pagetable_mac() local 186 for ( mfn = 0; mfn < max_page; mfn++ ) in update_pagetable_mac() 190 if ( !mfn_valid(_mfn(mfn)) ) in update_pagetable_mac() 252 static int mfn_in_guarded_stack(unsigned long mfn) in mfn_in_guarded_stack() argument 263 if ( mfn == virt_to_mfn(p) ) in mfn_in_guarded_stack() 273 unsigned long mfn; in tboot_gen_xenheap_integrity() local 278 for ( mfn = 0; mfn < max_page; mfn++ ) in tboot_gen_xenheap_integrity() 282 if ( !mfn_valid(_mfn(mfn)) ) in tboot_gen_xenheap_integrity() 284 if ( is_xen_fixed_mfn(_mfn(mfn)) ) in tboot_gen_xenheap_integrity() 296 if ( mfn_in_guarded_stack(mfn) ) in tboot_gen_xenheap_integrity() [all …]
|
| /xen/tools/misc/ |
| A D | xen-hptool.c | 38 unsigned long mfn; in hp_mem_online_func() local 46 sscanf(argv[0], "%lx", &mfn); in hp_mem_online_func() 49 ret = xc_mark_page_online(xch, mfn, mfn, &status); in hp_mem_online_func() 70 unsigned long mfn; in hp_mem_query_func() local 78 sscanf(argv[0], "%lx", &mfn); in hp_mem_query_func() 80 ret = xc_query_page_offline_status(xch, mfn, mfn, &status); in hp_mem_query_func() 151 unsigned long mfn; in hp_mem_offline_func() local 159 sscanf(argv[0], "%lx", &mfn); in hp_mem_offline_func() 161 ret = xc_mark_page_offline(xch, mfn, mfn, &status); in hp_mem_offline_func() 229 mfn, domid); in hp_mem_offline_func() [all …]
|
| A D | xen-mfndump.c | 102 xen_pfn_t mfn; in dump_p2m_func() local 105 mfn = ((uint64_t*)minfo.p2m_table)[i]; in dump_p2m_func() 108 mfn = ((uint32_t*)minfo.p2m_table)[i]; in dump_p2m_func() 111 mfn = ~0UL; in dump_p2m_func() 157 xen_pfn_t pfn, mfn, *m2p_table; in dump_ptes_func() local 165 mfn = strtoul(argv[1], NULL, 16); in dump_ptes_func() 184 if ( rc || (mfn > max_mfn) || in dump_ptes_func() 192 pfn = m2p_table[mfn]; in dump_ptes_func() 265 xen_pfn_t mfn; in lookup_pte_func() local 273 mfn = strtoul(argv[1], NULL, 16); in lookup_pte_func() [all …]
|
| /xen/tools/libxc/ |
| A D | xc_sr_common_x86_pv.c | 7 assert(mfn <= ctx->x86.pv.max_mfn); in mfn_to_pfn() 8 return ctx->x86.pv.m2p[mfn]; in mfn_to_pfn() 13 return ((mfn <= ctx->x86.pv.max_mfn) && in mfn_in_pseudophysmap() 26 if ( (mfn != ~0UL) && (mfn <= ctx->x86.pv.max_mfn) ) in dump_bad_pseudophysmap_entry() 28 pfn = ctx->x86.pv.m2p[mfn]; in dump_bad_pseudophysmap_entry() 30 mfn, pfn, ctx->x86.pv.max_pfn); in dump_bad_pseudophysmap_entry() 54 uint64_t mfn = _mfn; in mfn_to_cr3() local 57 return mfn << 12; in mfn_to_cr3() 61 if ( mfn & 0xffffffff00000000UL ) in mfn_to_cr3() 64 return (uint32_t)((mfn << 12) | (mfn >> 20)); in mfn_to_cr3() [all …]
|
| A D | xc_offline_page.c | 357 xen_pfn_t mfn) in clear_pte() argument 386 rc = xc_query_page_offline_status(xch, mfn, mfn, &status); in is_page_exchangable() 391 mfn, status); in is_page_exchangable() 446 *mfn0 = entries[0].mfn; in xc_map_m2p() 500 gpfn = m2p_table[mfn]; in xc_exchange_page() 564 mops.arg1.mfn = mfn; in xc_exchange_page() 571 mops.arg1.mfn = mfn; in xc_exchange_page() 577 PROT_READ, mfn); in xc_exchange_page() 604 1, 0, &mfn, in xc_exchange_page() 628 rc = xc_query_page_offline_status(xch, mfn, mfn, in xc_exchange_page() [all …]
|
| A D | xc_sr_common_x86_pv.h | 18 xen_pfn_t mfn_to_pfn(struct xc_sr_context *ctx, xen_pfn_t mfn); 23 bool mfn_in_pseudophysmap(struct xc_sr_context *ctx, xen_pfn_t mfn); 28 void dump_bad_pseudophysmap_entry(struct xc_sr_context *ctx, xen_pfn_t mfn); 44 uint64_t mfn_to_cr3(struct xc_sr_context *ctx, xen_pfn_t mfn); 68 static inline uint64_t merge_pte(uint64_t pte, xen_pfn_t mfn) in merge_pte() argument 70 return (pte & ~PTE_FRAME_MASK) | ((uint64_t)mfn << PAGE_SHIFT); in merge_pte()
|
| A D | xc_sr_restore_x86_pv.c | 137 xen_pfn_t pfn, mfn; in process_start_info() local 157 mfn = pfn_to_mfn(ctx, pfn); in process_start_info() 182 mfn = pfn_to_mfn(ctx, pfn); in process_start_info() 203 mfn = pfn_to_mfn(ctx, pfn); in process_start_info() 211 ctx->restore.console_gfn = mfn; in process_start_info() 212 SET_FIELD(guest_start_info, console.domU.mfn, mfn, ctx->x86.pv.width); in process_start_info() 240 xen_pfn_t pfn, mfn; in process_vcpu_basic() local 285 mfn = pfn_to_mfn(ctx, pfn); in process_vcpu_basic() 315 mfn = pfn_to_mfn(ctx, pfn); in process_vcpu_basic() 956 ((uint64_t *)ctx->x86.pv.p2m)[pfn] = mfn == INVALID_MFN ? ~0ULL : mfn; in x86_pv_set_gfn() [all …]
|
| /xen/xen/arch/x86/mm/ |
| A D | p2m-pt.c | 169 mfn_t mfn; in p2m_next_level() local 256 ASSERT(rc && mfn_valid(mfn)); in p2m_next_level() 495 ASSERT(mfn_valid(mfn) || mfn_eq(mfn, INVALID_MFN)); in check_entry() 497 ASSERT(mfn_valid(mfn)); in check_entry() 537 u64 gfn, mfn; in p2m_pt_set_entry() member 543 t.mfn = mfn_x(mfn); in p2m_pt_set_entry() 704 mfn_t mfn; in p2m_pt_get_entry() local 799 l2e = map_domain_page(mfn); in p2m_pt_get_entry() 832 mfn = l2e_get_mfn(*l2e); in p2m_pt_get_entry() 837 l1e = map_domain_page(mfn); in p2m_pt_get_entry() [all …]
|
| A D | mem_paging.c | 100 mfn_t mfn; in p2m_mem_paging_populate() local 176 mfn_t mfn; in p2m_mem_paging_resume() local 225 mfn_t mfn; in nominate() local 233 if ( !mfn_valid(mfn) ) in nominate() 241 if ( is_iomem_page(mfn) ) in nominate() 245 page = mfn_to_page(mfn); in nominate() 284 mfn_t mfn; in evict() local 300 page = mfn_to_page(mfn); in evict() 351 mfn_t mfn; in prepare() local 368 if ( !mfn_valid(mfn) ) in prepare() [all …]
|
| A D | p2m.c | 490 mfn_t mfn; in __get_gfn_type_access() local 537 return mfn; in __get_gfn_type_access() 559 mfn_t mfn; in p2m_get_page_from_gfn() local 649 mfn = mfn_add(mfn, 1ul << order); in p2m_set_entry() 1034 mfn_t mfn; in p2m_change_type_one() local 1347 mfn_t mfn; in set_identity_p2m_entry() local 1438 mfn_t mfn; in clear_identity_p2m_entry() local 1529 mfn_t mfn; in p2m_flush_table_locked() local 2313 mfn_t mfn; in p2m_change_altp2m_gfn() local 2693 mfn_t mfn; in p2m_set_suppress_ve_multi() local [all …]
|
| A D | p2m-ept.c | 206 mfn_t mfn; in ept_set_middle_entry() local 220 ept_entry->mfn = mfn_x(mfn); in ept_set_middle_entry() 503 unsigned long mfn = ept->mfn; in resolve_misconfig() local 507 if ( !mfn ) in resolve_misconfig() 588 mfn = e.mfn; in resolve_misconfig() 625 mfn = e.mfn; in resolve_misconfig() 807 new_entry.mfn = mfn_x(mfn); in ept_set_entry() 810 if ( ept_entry->mfn == new_entry.mfn && in ept_set_entry() 966 mfn = _mfn(ept_entry->mfn); in ept_get_entry() 1042 unsigned long mfn = p2m->ept.mfn; in ept_change_entry_type_global() local [all …]
|
| A D | paging.c | 75 if ( mfn_valid(mfn) ) in paging_new_log_dirty_leaf() 78 return mfn; in paging_new_log_dirty_leaf() 85 if ( mfn_valid(mfn) ) in paging_new_log_dirty_node() 93 return mfn; in paging_new_log_dirty_node() 297 mfn = l4[i4]; in paging_mark_pfn_dirty() 298 if ( !mfn_valid(mfn) ) in paging_mark_pfn_dirty() 301 if ( !mfn_valid(mfn) ) in paging_mark_pfn_dirty() 305 mfn = l3[i3]; in paging_mark_pfn_dirty() 306 if ( !mfn_valid(mfn) ) in paging_mark_pfn_dirty() 309 if ( !mfn_valid(mfn) ) in paging_mark_pfn_dirty() [all …]
|
| /xen/xen/include/asm-x86/ |
| A D | mm.h | 303 #define is_xen_heap_mfn(mfn) \ argument 304 (mfn_valid(mfn) && is_xen_heap_page(mfn_to_page(mfn))) 367 bool fill_ro_mpt(mfn_t mfn); 368 void zap_ro_mpt(mfn_t mfn); 370 bool is_iomem_page(mfn_t mfn); 373 unsigned long mfn; member 519 #define get_gpfn_from_mfn(mfn) (machine_to_phys_mapping[(mfn)]) argument 524 : (mfn) ) 579 void make_cr3(struct vcpu *v, mfn_t mfn); 588 void free_xen_pagetable_new(mfn_t mfn); [all …]
|
| /xen/xen/include/asm-arm/ |
| A D | mm.h | 146 #define is_xen_heap_mfn(mfn) ({ \ argument 147 unsigned long mfn_ = mfn_x(mfn); \ 153 #define is_xen_heap_mfn(mfn) \ argument 154 (mfn_valid(mfn) && is_xen_heap_page(mfn_to_page(mfn))) 159 (mfn_to_maddr(mfn) <= virt_to_maddr((vaddr_t)_end - 1))) 221 (frame_table + (mfn_to_pdx(mfn) - frametable_base_pdx)) 235 #define mfn_to_maddr(mfn) pfn_to_paddr(mfn_x(mfn)) argument 289 #define __mfn_to_virt(mfn) (maddr_to_virt((paddr_t)(mfn) << PAGE_SHIFT)) argument 297 #define mfn_to_virt(mfn) __mfn_to_virt(mfn) argument 330 #define set_gpfn_from_mfn(mfn, pfn) do { (void) (mfn), (void)(pfn); } while (0) argument [all …]
|
| A D | p2m.h | 279 mfn_t mfn, 285 mfn_t mfn); 290 mfn_t mfn); 294 mfn_t mfn, 301 mfn_t mfn, in guest_physmap_add_page() argument 320 mfn_t mfn; in get_page_from_gfn() local 340 mfn = _mfn(gfn); in get_page_from_gfn() 341 page = mfn_to_page(mfn); in get_page_from_gfn() 343 if ( !mfn_valid(mfn) || !get_page(page, d) ) in get_page_from_gfn() 355 bool is_iomem_page(mfn_t mfn); [all …]
|
| /xen/xen/include/xen/ |
| A D | domain_page.h | 17 void clear_domain_page(mfn_t mfn); 26 void *map_domain_page(mfn_t mfn); 44 void *map_domain_page_global(mfn_t mfn); 56 #define map_domain_page(mfn) __mfn_to_virt(mfn_x(mfn)) argument 61 static inline void *map_domain_page_global(mfn_t mfn) in map_domain_page_global() argument 63 return mfn_to_virt(mfn_x(mfn)); in map_domain_page_global()
|
| A D | grant_table.h | 58 mfn_t *mfn); 60 mfn_t *mfn); 62 mfn_t *mfn); 91 gfn_t gfn, mfn_t *mfn) in gnttab_map_frame() argument 97 mfn_t *mfn) in gnttab_get_shared_frame() argument 103 mfn_t *mfn) in gnttab_get_status_frame() argument
|
| /xen/xen/arch/x86/mm/shadow/ |
| A D | hvm.c | 403 mfn_t mfn; in emulate_gva_to_mfn() local 441 mfn = page_to_mfn(page); in emulate_gva_to_mfn() 442 ASSERT(mfn_valid(mfn)); in emulate_gva_to_mfn() 446 return mfn; in emulate_gva_to_mfn() 473 if ( !mfn_valid(sh_ctxt->mfn[0]) ) in sh_emulate_map_dest() 475 switch ( mfn_x(sh_ctxt->mfn[0]) ) in sh_emulate_map_dest() 490 sh_ctxt->mfn[1] = INVALID_MFN; in sh_emulate_map_dest() 507 if ( !mfn_valid(sh_ctxt->mfn[1]) ) in sh_emulate_map_dest() 521 map = vmap(sh_ctxt->mfn, 2); in sh_emulate_map_dest() 619 ASSERT(mfn_valid(sh_ctxt->mfn[0])); in sh_emulate_unmap_dest() [all …]
|
| /xen/xen/arch/x86/x86_64/ |
| A D | traps.c | 197 l4t = map_domain_page(_mfn(mfn)); in show_page_walk() 200 mfn = l4e_get_pfn(l4e); in show_page_walk() 206 !mfn_valid(_mfn(mfn)) ) in show_page_walk() 209 l3t = map_domain_page(_mfn(mfn)); in show_page_walk() 212 mfn = l3e_get_pfn(l3e); in show_page_walk() 220 !mfn_valid(_mfn(mfn)) ) in show_page_walk() 223 l2t = map_domain_page(_mfn(mfn)); in show_page_walk() 226 mfn = l2e_get_pfn(l2e); in show_page_walk() 234 !mfn_valid(_mfn(mfn)) ) in show_page_walk() 237 l1t = map_domain_page(_mfn(mfn)); in show_page_walk() [all …]
|
| /xen/xen/arch/x86/cpu/mcheck/ |
| A D | mcaction.c | 9 mfn_t mfn, uint32_t status) in mci_action_add_pageoffline() argument 25 rec->action_info.page_retire.mfn = mfn_x(mfn); in mci_action_add_pageoffline() 45 mfn_t mfn; in mc_memerr_dhandler() local 58 mfn = maddr_to_mfn(bank->mc_addr); in mc_memerr_dhandler() 59 if ( offline_page(mfn, 1, &status) ) in mc_memerr_dhandler() 63 mfn_x(mfn)); in mc_memerr_dhandler() 67 mci_action_add_pageoffline(binfo->bank, binfo->mi, mfn, status); in mc_memerr_dhandler() 94 if ( unmmap_broken_page(d, mfn, gfn) ) in mc_memerr_dhandler() 97 mfn_x(mfn), d->domain_id); in mc_memerr_dhandler()
|
| /xen/xen/arch/arm/ |
| A D | mm.c | 50 #define mfn_to_virt(mfn) __mfn_to_virt(mfn_x(mfn)) argument 357 lpae_set_mfn(e, mfn); in mfn_to_xen_entry() 417 return vmap(&mfn, 1); in map_domain_page_global() 426 void *map_domain_page(mfn_t mfn) in map_domain_page() argument 816 unsigned long mfn, end_mfn; in setup_xenheap_mappings() local 843 while ( mfn < end_mfn ) in setup_xenheap_mappings() 1039 mfn_t mfn; in xen_pt_next_level() local 1278 mfn = mfn_add(mfn, 1); in xen_pt_update() 1414 mfn_t mfn = INVALID_MFN; in xenmem_add_to_physmap_one() local 1621 bool is_iomem_page(mfn_t mfn) in is_iomem_page() argument [all …]
|
| A D | p2m.c | 310 mfn_t mfn; in p2m_next_level() local 423 mfn = mfn_add(mfn, gfn_x(gfn) & ((1UL << level_orders[level]) - 1)); in p2m_get_entry() 436 return mfn; in p2m_get_entry() 441 mfn_t mfn; in p2m_lookup() local 448 return mfn; in p2m_lookup() 625 lpae_set_mfn(e, mfn); in mfn_to_p2m_entry() 745 mfn_t mfn; in p2m_free_entry() local 1354 if ( !(nr && iomem_access_permitted(d, mfn_x(mfn), mfn_x(mfn) + nr - 1)) ) in map_dev_mmio_region() 1361 mfn_x(mfn), mfn_x(mfn) + nr - 1, d->domain_id); in map_dev_mmio_region() 1707 mfn = mfn_add(mfn, 1); in p2m_cache_flush_range() [all …]
|