/xen/xen/arch/x86/hvm/ |
A D | grant_table.c | 44 frame, PAGE_ORDER_4K, p2mt); in create_grant_p2m_mapping() 71 if ( guest_physmap_remove_page(d, _gfn(gfn), frame, PAGE_ORDER_4K) ) in replace_grant_p2m_mapping()
|
A D | dom0_build.c | 114 { .align = PFN_DOWN(KB(4)), .order = PAGE_ORDER_4K }, in pvh_populate_memory_range() 378 start = ROUNDUP(entry->addr, PAGE_SIZE << PAGE_ORDER_4K); in pvh_setup_e820() 380 ~((PAGE_SIZE << PAGE_ORDER_4K) - 1); in pvh_setup_e820()
|
A D | hvm.c | 1880 int res = p2m_set_entry(p2m, _gfn(gfn), mfn, PAGE_ORDER_4K, in hvm_hap_nested_page_fault()
|
/xen/xen/arch/x86/mm/ |
A D | mem_paging.c | 128 rc = p2m_set_entry(p2m, gfn, mfn, PAGE_ORDER_4K, p2m_ram_paging_in, a); in p2m_mem_paging_populate() 191 int rc = p2m_set_entry(p2m, gfn, mfn, PAGE_ORDER_4K, in p2m_mem_paging_resume() 254 ret = p2m_set_entry(p2m, gfn, mfn, PAGE_ORDER_4K, p2m_ram_paging_out, a); in nominate() 316 ret = p2m_set_entry(p2m, gfn, INVALID_MFN, PAGE_ORDER_4K, in evict() 413 ret = p2m_set_entry(p2m, gfn, mfn, PAGE_ORDER_4K, in prepare()
|
A D | p2m-pod.c | 121 case PAGE_ORDER_4K: in p2m_pod_cache_add() 147 else if ( order == PAGE_ORDER_4K && page_list_empty(&p2m->pod.single) ) in p2m_pod_cache_get() 175 case PAGE_ORDER_4K: in p2m_pod_cache_get() 214 order = PAGE_ORDER_4K; in p2m_pod_set_cache_target() 222 order = PAGE_ORDER_4K; in p2m_pod_set_cache_target() 257 order = PAGE_ORDER_4K; in p2m_pod_set_cache_target() 488 p2m_pod_cache_add(p2m, p, PAGE_ORDER_4K); in p2m_pod_offline_or_broken_replace() 916 if ( p2m_set_entry(p2m, gfns[i], INVALID_MFN, PAGE_ORDER_4K, in p2m_pod_zero_check() 930 if ( p2m_set_entry(p2m, gfns[i], mfns[i], PAGE_ORDER_4K, in p2m_pod_zero_check() 972 if ( p2m_set_entry(p2m, gfns[i], mfns[i], PAGE_ORDER_4K, in p2m_pod_zero_check() [all …]
|
A D | mem_access.c | 170 rc = p2m->set_entry(p2m, gfn, mfn, PAGE_ORDER_4K, p2mt, p2m_access_rw, -1); in p2m_mem_access_check() 178 rc = p2m->set_entry(p2m, gfn, mfn, PAGE_ORDER_4K, in p2m_mem_access_check() 205 rc = p2m->set_entry(p2m, gfn, mfn, PAGE_ORDER_4K, in p2m_mem_access_check() 283 return ap2m->set_entry(ap2m, gfn, mfn, PAGE_ORDER_4K, t, a, -1); in p2m_set_altp2m_mem_access() 310 rc = p2m->set_entry(p2m, gfn, mfn, PAGE_ORDER_4K, t, a, -1); in set_mem_access()
|
A D | p2m.c | 638 hap_has_2mb) ? PAGE_ORDER_2M : PAGE_ORDER_4K; in p2m_set_entry() 1045 ? p2m_set_entry(p2m, gfn, mfn, PAGE_ORDER_4K, nt, in p2m_change_type_one() 1332 if ( order > PAGE_ORDER_4K && in set_mmio_p2m_entry() 1335 return PAGE_ORDER_4K + 1; in set_mmio_p2m_entry() 1364 ret = p2m_set_entry(p2m, gfn, _mfn(gfn_l), PAGE_ORDER_4K, in set_identity_p2m_entry() 1446 return iommu_legacy_unmap(d, _dfn(gfn_l), PAGE_ORDER_4K); in clear_identity_p2m_entry() 1454 ret = p2m_set_entry(p2m, gfn, INVALID_MFN, PAGE_ORDER_4K, in clear_identity_p2m_entry() 1873 return PAGE_ORDER_4K; in mmio_order() 1887 return PAGE_ORDER_4K; in mmio_order() 1979 if ( prepopulate && page_order != PAGE_ORDER_4K ) in altp2m_get_effective_entry() [all …]
|
A D | mem_sharing.c | 1147 ret = p2m_set_entry(p2m, _gfn(cgfn), smfn, PAGE_ORDER_4K, in add_to_physmap() 1366 set_rc = p2m->set_entry(p2m, _gfn(gfn), INVALID_MFN, PAGE_ORDER_4K, in relinquish_shared_pages() 1544 return p2m->set_entry(p2m, gfn, new_mfn, PAGE_ORDER_4K, p2m_ram_rw, in mem_sharing_fork_page() 1605 PAGE_ORDER_4K, p2m_ram_rw, in copy_vcpu_settings() 1692 rc = p2m->set_entry(p2m, _gfn(value), new_mfn, PAGE_ORDER_4K, in copy_special_pages() 1713 rc = p2m->set_entry(p2m, new_gfn, INVALID_MFN, PAGE_ORDER_4K, in copy_special_pages() 1722 rc = p2m->set_entry(p2m, old_gfn, new_mfn, PAGE_ORDER_4K, in copy_special_pages() 1831 rc = p2m->set_entry(p2m, gfn, INVALID_MFN, PAGE_ORDER_4K, in mem_sharing_fork_reset()
|
A D | p2m-pt.c | 138 if ( page_order == PAGE_ORDER_4K in p2m_free_entry() 487 ASSERT(order > PAGE_ORDER_4K); in check_entry() 610 if ( page_order == PAGE_ORDER_4K ) in p2m_pt_set_entry() 851 if ( p2m_pod_demand_populate(p2m, gfn_, PAGE_ORDER_4K) ) in p2m_pt_get_entry()
|
A D | p2m-ept.c | 954 if ( !p2m_pod_demand_populate(p2m, gfn_, PAGE_ORDER_4K) ) in ept_get_entry()
|
/xen/xen/drivers/passthrough/x86/ |
A D | iommu.c | 245 rc = iommu_map(d, _dfn(pfn), _mfn(pfn), PAGE_ORDER_4K, in arch_iommu_hwdom_init()
|
/xen/xen/arch/x86/x86_64/ |
A D | mm.c | 1449 PAGE_ORDER_4K, in memory_add() 1457 PAGE_ORDER_4K) ) in memory_add()
|
/xen/xen/include/asm-x86/ |
A D | page.h | 15 #define PAGE_ORDER_4K 0 macro
|
/xen/xen/arch/x86/mm/hap/ |
A D | hap.c | 836 *page_order = PAGE_ORDER_4K; in hap_p2m_ga_to_gfn_real_mode()
|
/xen/xen/arch/x86/ |
A D | mm.c | 2493 int rc2 = iommu_legacy_unmap(d, _dfn(mfn), PAGE_ORDER_4K); in cleanup_page_mappings() 3020 rc = iommu_legacy_unmap(d, _dfn(mfn_x(mfn)), PAGE_ORDER_4K); in _get_page_type() 3022 rc = iommu_legacy_map(d, _dfn(mfn_x(mfn)), mfn, PAGE_ORDER_4K, in _get_page_type() 4659 rc = guest_physmap_remove_page(d, gpfn, prev_mfn, PAGE_ORDER_4K); in xenmem_add_to_physmap_one() 4679 rc = guest_physmap_remove_page(d, _gfn(old_gpfn), mfn, PAGE_ORDER_4K); in xenmem_add_to_physmap_one() 4683 rc = guest_physmap_add_page(d, gpfn, mfn, PAGE_ORDER_4K); in xenmem_add_to_physmap_one()
|
A D | domain.c | 780 ret = guest_physmap_remove_page(d, gfn, mfn, PAGE_ORDER_4K); in arch_domain_soft_reset() 791 PAGE_ORDER_4K); in arch_domain_soft_reset()
|
/xen/xen/common/ |
A D | memory.c | 332 rc = clear_mmio_p2m_entry(d, gmfn, mfn, PAGE_ORDER_4K); in guest_remove_page()
|
/xen/xen/arch/x86/hvm/vmx/ |
A D | vmx.c | 3077 PAGE_ORDER_4K); in vmx_alloc_vlapic_mapping()
|