Home
last modified time | relevance | path

Searched refs:PAGE_ORDER_4K (Results 1 – 18 of 18) sorted by relevance

/xen/xen/arch/x86/hvm/
A Dgrant_table.c44 frame, PAGE_ORDER_4K, p2mt); in create_grant_p2m_mapping()
71 if ( guest_physmap_remove_page(d, _gfn(gfn), frame, PAGE_ORDER_4K) ) in replace_grant_p2m_mapping()
A Ddom0_build.c114 { .align = PFN_DOWN(KB(4)), .order = PAGE_ORDER_4K }, in pvh_populate_memory_range()
378 start = ROUNDUP(entry->addr, PAGE_SIZE << PAGE_ORDER_4K); in pvh_setup_e820()
380 ~((PAGE_SIZE << PAGE_ORDER_4K) - 1); in pvh_setup_e820()
A Dhvm.c1880 int res = p2m_set_entry(p2m, _gfn(gfn), mfn, PAGE_ORDER_4K, in hvm_hap_nested_page_fault()
/xen/xen/arch/x86/mm/
A Dmem_paging.c128 rc = p2m_set_entry(p2m, gfn, mfn, PAGE_ORDER_4K, p2m_ram_paging_in, a); in p2m_mem_paging_populate()
191 int rc = p2m_set_entry(p2m, gfn, mfn, PAGE_ORDER_4K, in p2m_mem_paging_resume()
254 ret = p2m_set_entry(p2m, gfn, mfn, PAGE_ORDER_4K, p2m_ram_paging_out, a); in nominate()
316 ret = p2m_set_entry(p2m, gfn, INVALID_MFN, PAGE_ORDER_4K, in evict()
413 ret = p2m_set_entry(p2m, gfn, mfn, PAGE_ORDER_4K, in prepare()
A Dp2m-pod.c121 case PAGE_ORDER_4K: in p2m_pod_cache_add()
147 else if ( order == PAGE_ORDER_4K && page_list_empty(&p2m->pod.single) ) in p2m_pod_cache_get()
175 case PAGE_ORDER_4K: in p2m_pod_cache_get()
214 order = PAGE_ORDER_4K; in p2m_pod_set_cache_target()
222 order = PAGE_ORDER_4K; in p2m_pod_set_cache_target()
257 order = PAGE_ORDER_4K; in p2m_pod_set_cache_target()
488 p2m_pod_cache_add(p2m, p, PAGE_ORDER_4K); in p2m_pod_offline_or_broken_replace()
916 if ( p2m_set_entry(p2m, gfns[i], INVALID_MFN, PAGE_ORDER_4K, in p2m_pod_zero_check()
930 if ( p2m_set_entry(p2m, gfns[i], mfns[i], PAGE_ORDER_4K, in p2m_pod_zero_check()
972 if ( p2m_set_entry(p2m, gfns[i], mfns[i], PAGE_ORDER_4K, in p2m_pod_zero_check()
[all …]
A Dmem_access.c170 rc = p2m->set_entry(p2m, gfn, mfn, PAGE_ORDER_4K, p2mt, p2m_access_rw, -1); in p2m_mem_access_check()
178 rc = p2m->set_entry(p2m, gfn, mfn, PAGE_ORDER_4K, in p2m_mem_access_check()
205 rc = p2m->set_entry(p2m, gfn, mfn, PAGE_ORDER_4K, in p2m_mem_access_check()
283 return ap2m->set_entry(ap2m, gfn, mfn, PAGE_ORDER_4K, t, a, -1); in p2m_set_altp2m_mem_access()
310 rc = p2m->set_entry(p2m, gfn, mfn, PAGE_ORDER_4K, t, a, -1); in set_mem_access()
A Dp2m.c638 hap_has_2mb) ? PAGE_ORDER_2M : PAGE_ORDER_4K; in p2m_set_entry()
1045 ? p2m_set_entry(p2m, gfn, mfn, PAGE_ORDER_4K, nt, in p2m_change_type_one()
1332 if ( order > PAGE_ORDER_4K && in set_mmio_p2m_entry()
1335 return PAGE_ORDER_4K + 1; in set_mmio_p2m_entry()
1364 ret = p2m_set_entry(p2m, gfn, _mfn(gfn_l), PAGE_ORDER_4K, in set_identity_p2m_entry()
1446 return iommu_legacy_unmap(d, _dfn(gfn_l), PAGE_ORDER_4K); in clear_identity_p2m_entry()
1454 ret = p2m_set_entry(p2m, gfn, INVALID_MFN, PAGE_ORDER_4K, in clear_identity_p2m_entry()
1873 return PAGE_ORDER_4K; in mmio_order()
1887 return PAGE_ORDER_4K; in mmio_order()
1979 if ( prepopulate && page_order != PAGE_ORDER_4K ) in altp2m_get_effective_entry()
[all …]
A Dmem_sharing.c1147 ret = p2m_set_entry(p2m, _gfn(cgfn), smfn, PAGE_ORDER_4K, in add_to_physmap()
1366 set_rc = p2m->set_entry(p2m, _gfn(gfn), INVALID_MFN, PAGE_ORDER_4K, in relinquish_shared_pages()
1544 return p2m->set_entry(p2m, gfn, new_mfn, PAGE_ORDER_4K, p2m_ram_rw, in mem_sharing_fork_page()
1605 PAGE_ORDER_4K, p2m_ram_rw, in copy_vcpu_settings()
1692 rc = p2m->set_entry(p2m, _gfn(value), new_mfn, PAGE_ORDER_4K, in copy_special_pages()
1713 rc = p2m->set_entry(p2m, new_gfn, INVALID_MFN, PAGE_ORDER_4K, in copy_special_pages()
1722 rc = p2m->set_entry(p2m, old_gfn, new_mfn, PAGE_ORDER_4K, in copy_special_pages()
1831 rc = p2m->set_entry(p2m, gfn, INVALID_MFN, PAGE_ORDER_4K, in mem_sharing_fork_reset()
A Dp2m-pt.c138 if ( page_order == PAGE_ORDER_4K in p2m_free_entry()
487 ASSERT(order > PAGE_ORDER_4K); in check_entry()
610 if ( page_order == PAGE_ORDER_4K ) in p2m_pt_set_entry()
851 if ( p2m_pod_demand_populate(p2m, gfn_, PAGE_ORDER_4K) ) in p2m_pt_get_entry()
A Dp2m-ept.c954 if ( !p2m_pod_demand_populate(p2m, gfn_, PAGE_ORDER_4K) ) in ept_get_entry()
/xen/xen/drivers/passthrough/x86/
A Diommu.c245 rc = iommu_map(d, _dfn(pfn), _mfn(pfn), PAGE_ORDER_4K, in arch_iommu_hwdom_init()
/xen/xen/arch/x86/x86_64/
A Dmm.c1449 PAGE_ORDER_4K, in memory_add()
1457 PAGE_ORDER_4K) ) in memory_add()
/xen/xen/include/asm-x86/
A Dpage.h15 #define PAGE_ORDER_4K 0 macro
/xen/xen/arch/x86/mm/hap/
A Dhap.c836 *page_order = PAGE_ORDER_4K; in hap_p2m_ga_to_gfn_real_mode()
/xen/xen/arch/x86/
A Dmm.c2493 int rc2 = iommu_legacy_unmap(d, _dfn(mfn), PAGE_ORDER_4K); in cleanup_page_mappings()
3020 rc = iommu_legacy_unmap(d, _dfn(mfn_x(mfn)), PAGE_ORDER_4K); in _get_page_type()
3022 rc = iommu_legacy_map(d, _dfn(mfn_x(mfn)), mfn, PAGE_ORDER_4K, in _get_page_type()
4659 rc = guest_physmap_remove_page(d, gpfn, prev_mfn, PAGE_ORDER_4K); in xenmem_add_to_physmap_one()
4679 rc = guest_physmap_remove_page(d, _gfn(old_gpfn), mfn, PAGE_ORDER_4K); in xenmem_add_to_physmap_one()
4683 rc = guest_physmap_add_page(d, gpfn, mfn, PAGE_ORDER_4K); in xenmem_add_to_physmap_one()
A Ddomain.c780 ret = guest_physmap_remove_page(d, gfn, mfn, PAGE_ORDER_4K); in arch_domain_soft_reset()
791 PAGE_ORDER_4K); in arch_domain_soft_reset()
/xen/xen/common/
A Dmemory.c332 rc = clear_mmio_p2m_entry(d, gmfn, mfn, PAGE_ORDER_4K); in guest_remove_page()
/xen/xen/arch/x86/hvm/vmx/
A Dvmx.c3077 PAGE_ORDER_4K); in vmx_alloc_vlapic_mapping()

Completed in 58 milliseconds