/xen/xen/arch/x86/mm/ |
A D | mem_paging.c | 61 if ( p2mt != p2m_ram_paging_out ) in p2m_mem_paging_drop_page() 98 p2m_type_t p2mt; in p2m_mem_paging_populate() local 122 if ( p2mt == p2m_ram_paging_out || p2mt == p2m_ram_paged ) in p2m_mem_paging_populate() 141 else if ( p2mt != p2m_ram_paging_out && p2mt != p2m_ram_paged ) in p2m_mem_paging_populate() 150 req.u.mem_paging.p2mt = p2mt; in p2m_mem_paging_populate() 174 p2m_type_t p2mt; in p2m_mem_paging_resume() local 223 p2m_type_t p2mt; in nominate() local 237 if ( !p2m_is_pageable(p2mt) ) in nominate() 282 p2m_type_t p2mt; in evict() local 349 p2m_type_t p2mt; in prepare() local [all …]
|
A D | altp2m.c | 62 p2m_type_t p2mt; in altp2m_vcpu_enable_ve() local 70 rc = check_get_page_from_gfn(d, gfn, false, &p2mt, &pg); in altp2m_vcpu_enable_ve() 80 if ( !p2m_is_pageable(p2mt) || is_special_page(pg) ) in altp2m_vcpu_enable_ve()
|
A D | p2m-pt.c | 517 unsigned int iommu_pte_flags = p2m_get_iommu_flags(p2mt, mfn); in p2m_pt_set_entry() 538 int p2mt; in p2m_pt_set_entry() member 544 t.p2mt = p2mt; in p2m_pt_set_entry() 589 check_entry(mfn, p2mt, p2m_flags_to_type(flags), page_order); in p2m_pt_set_entry() 590 l3e_content = mfn_valid(mfn) || p2m_allows_invalid_mfn(p2mt) in p2m_pt_set_entry() 592 p2m_type_to_flags(p2m, p2mt, mfn, 2)) in p2m_pt_set_entry() 626 if ( mfn_valid(mfn) || p2m_allows_invalid_mfn(p2mt) ) in p2m_pt_set_entry() 661 check_entry(mfn, p2mt, p2m_flags_to_type(flags), page_order); in p2m_pt_set_entry() 662 l2e_content = mfn_valid(mfn) || p2m_allows_invalid_mfn(p2mt) in p2m_pt_set_entry() 664 p2m_type_to_flags(p2m, p2mt, mfn, 1)) in p2m_pt_set_entry() [all …]
|
A D | mem_access.c | 152 p2m_type_t p2mt; in p2m_mem_access_check() local 166 mfn = p2m->get_entry(p2m, gfn, &p2mt, &p2ma, 0, NULL, NULL); in p2m_mem_access_check() 170 rc = p2m->set_entry(p2m, gfn, mfn, PAGE_ORDER_4K, p2mt, p2m_access_rw, -1); in p2m_mem_access_check() 179 p2mt, p2m_access_rwx, -1); in p2m_mem_access_check() 199 mfn = p2m->get_entry(p2m, gfn, &p2mt, &p2ma, 0, NULL, NULL); in p2m_mem_access_check() 206 p2mt, p2m_access_rwx, -1); in p2m_mem_access_check()
|
A D | p2m.c | 1344 p2m_type_t p2mt; in set_identity_p2m_entry() local 1363 if ( p2mt == p2m_invalid || p2mt == p2m_mmio_dm ) in set_identity_p2m_entry() 1435 p2m_type_t p2mt; in clear_identity_p2m_entry() local 1452 if ( p2mt == p2m_mmio_direct && mfn_x(mfn) == gfn_l ) in clear_identity_p2m_entry() 1822 p2m_type_t p2mt; in map_domain_gfn() local 1832 page = p2m_get_page_from_gfn(p2m, gfn, &p2mt, NULL, q); in map_domain_gfn() 1833 if ( p2m_is_paging(p2mt) ) in map_domain_gfn() 1842 if ( p2m_is_shared(p2mt) ) in map_domain_gfn() 2062 *p2mt = ap2mt; in p2m_altp2m_get_or_propagate() 2558 p2m_type_t p2mt, p2mt_prev; in p2m_add_foreign() local [all …]
|
A D | mem_sharing.c | 605 .u.mem_sharing.p2mt = p2m_ram_shared, in mem_sharing_notify_enomem() 785 p2m_type_t p2mt; in debug_gfn() local 789 mfn = get_gfn_query(d, gfn_x(gfn), &p2mt); in debug_gfn() 825 p2m_type_t p2mt; in nominate_page() local 841 if ( p2m_is_shared(p2mt) ) in nominate_page() 854 if ( !p2m_is_sharable(p2mt) ) in nominate_page() 1214 p2m_type_t p2mt; in __mem_sharing_unshare_page() local 1220 mfn = get_gfn(d, gfn, &p2mt); in __mem_sharing_unshare_page() 1223 if ( !p2m_is_shared(p2mt) ) in __mem_sharing_unshare_page() 1479 p2m_type_t p2mt; in mem_sharing_fork_page() local [all …]
|
A D | p2m-ept.c | 669 unsigned int order, p2m_type_t p2mt, p2m_access_t p2ma, in ept_set_entry() argument 679 bool_t direct_mmio = (p2mt == p2m_mmio_direct); in ept_set_entry() 683 unsigned int iommu_flags = p2m_get_iommu_flags(p2mt, mfn); in ept_set_entry() 711 ASSERT(!p2m_is_foreign(p2mt) || target == 0); in ept_set_entry() 790 if ( mfn_valid(mfn) || p2m_allows_invalid_mfn(p2mt) ) in ept_set_entry() 802 new_entry.sa_p2mt = p2mt; in ept_set_entry() 831 if ( p2mt != p2m_invalid && in ept_set_entry() 867 ret = p2m_altp2m_propagate_change(d, _gfn(gfn), mfn, order, p2mt, p2ma); in ept_set_entry()
|
/xen/xen/arch/x86/mm/hap/ |
A D | nested_hap.c | 100 unsigned int page_order, p2m_type_t p2mt, p2m_access_t p2ma) in nestedhap_fix_p2m() argument 118 rc = p2m_set_entry(p2m, _gfn(gfn), mfn, page_order, p2mt, p2ma); in nestedhap_fix_p2m() 151 p2m_type_t *p2mt, p2m_access_t *p2ma, in nestedhap_walk_L0_p2m() argument 159 mfn = get_gfn_type_access(p2m, L1_gpa >> PAGE_SHIFT, p2mt, p2ma, in nestedhap_walk_L0_p2m() 163 if ( *p2mt == p2m_mmio_direct ) in nestedhap_walk_L0_p2m() 166 if ( *p2mt == p2m_mmio_dm ) in nestedhap_walk_L0_p2m() 170 if ( access_w && p2m_is_readonly(*p2mt) ) in nestedhap_walk_L0_p2m() 173 if ( p2m_is_paging(*p2mt) || p2m_is_shared(*p2mt) || !p2m_is_ram(*p2mt) ) in nestedhap_walk_L0_p2m()
|
A D | guest_walk.c | 56 p2m_type_t p2mt; in hap_p2m_ga_to_gfn() local 63 top_page = p2m_get_page_from_gfn(p2m, top_gfn, &p2mt, NULL, in hap_p2m_ga_to_gfn() 65 if ( p2m_is_paging(p2mt) ) in hap_p2m_ga_to_gfn() 74 if ( p2m_is_shared(p2mt) ) in hap_p2m_ga_to_gfn() 105 page = p2m_get_page_from_gfn(p2m, gfn, &p2mt, NULL, in hap_p2m_ga_to_gfn() 109 if ( p2m_is_paging(p2mt) ) in hap_p2m_ga_to_gfn() 116 if ( p2m_is_shared(p2mt) ) in hap_p2m_ga_to_gfn()
|
/xen/xen/arch/x86/hvm/ |
A D | grant_table.c | 32 p2m_type_t p2mt; in create_grant_p2m_mapping() local 39 p2mt = p2m_grant_map_ro; in create_grant_p2m_mapping() 41 p2mt = p2m_grant_map_rw; in create_grant_p2m_mapping() 44 frame, PAGE_ORDER_4K, p2mt); in create_grant_p2m_mapping()
|
A D | emulate.c | 291 p2m_type_t p2mt = p2m_invalid; in hvmemul_do_io() local 390 p2m_type_t p2mt; in hvmemul_acquire_page() local 410 if ( p2m_is_mmio(p2mt) ) in hvmemul_acquire_page() 608 p2m_type_t p2mt; in hvmemul_map_linear_addr() local 667 ASSERT(p2mt == p2m_ram_logdirty || !p2m_is_readonly(p2mt)); in hvmemul_map_linear_addr() 1708 p2m_type_t p2mt; in hvmemul_rep_ins() local 1726 if ( p2mt == p2m_mmio_direct || p2mt == p2m_mmio_dm ) in hvmemul_rep_ins() 1786 p2m_type_t p2mt; in hvmemul_rep_outs() local 1807 if ( p2mt == p2m_mmio_direct || p2mt == p2m_mmio_dm ) in hvmemul_rep_outs() 1998 p2m_type_t p2mt; in hvmemul_rep_stos() local [all …]
|
A D | hvm.c | 1739 p2m_type_t p2mt; in hvm_hap_nested_page_fault() local 1926 (p2m_is_discard_write(p2mt) || (p2mt == p2m_ioreq_server))) ) in hvm_hap_nested_page_fault() 1935 if ( p2m_is_paged(p2mt) || (p2mt == p2m_ram_paging_out) ) in hvm_hap_nested_page_fault() 1950 if ( p2m_is_ram(p2mt) ) in hvm_hap_nested_page_fault() 2649 p2m_type_t p2mt; in _hvm_map_guest_frame() local 3193 p2m_type_t p2mt; in hvm_translate_get_page() local 3238 if ( p2m_is_paging(p2mt) ) in hvm_translate_get_page() 3244 if ( p2m_is_shared(p2mt) ) in hvm_translate_get_page() 3249 if ( p2m_is_grant(p2mt) ) in hvm_translate_get_page() 3259 *p2mt_p = p2mt; in hvm_translate_get_page() [all …]
|
A D | mtrr.c | 366 p2m_type_t p2mt; in get_pat_flags() local 367 get_gfn_query_unlocked(d, paddr_to_pfn(gpaddr), &p2mt); in get_pat_flags() 368 if (p2m_is_ram(p2mt)) in get_pat_flags()
|
/xen/xen/arch/x86/mm/shadow/ |
A D | multi.c | 507 if ( (!p2m_is_valid(p2mt) && !p2m_is_grant(p2mt)) in _sh_propagate() 2088 p2m_type_t p2mt; in validate_gl4e() local 2147 p2m_type_t p2mt; in validate_gl3e() local 2181 p2m_type_t p2mt; in validate_gl2e() local 2523 p2m_type_t p2mt; in sh_prefetch() local 2784 p2m_type_t p2mt; in sh_page_fault() local 3039 ((!p2m_is_valid(p2mt) && !p2m_is_grant(p2mt)) || in sh_page_fault() 4447 p2m_type_t p2mt; in sh_pagetable_dying() local 4524 p2m_type_t p2mt; in sh_pagetable_dying() local 4624 p2m_type_t p2mt; in sh_audit_l1_table() local [all …]
|
A D | hvm.c | 404 p2m_type_t p2mt; in emulate_gva_to_mfn() local 419 page = get_page_from_gfn(v->domain, gfn, &p2mt, P2M_ALLOC); in emulate_gva_to_mfn() 426 if ( p2mt == p2m_ioreq_server ) in emulate_gva_to_mfn() 431 if ( p2m_is_discard_write(p2mt) ) in emulate_gva_to_mfn() 436 if ( !p2m_is_ram(p2mt) ) in emulate_gva_to_mfn()
|
/xen/xen/common/ |
A D | memory.c | 293 p2m_type_t p2mt; in guest_remove_page() local 301 if ( unlikely(p2mt == p2m_invalid) || unlikely(p2mt == p2m_mmio_dm) ) in guest_remove_page() 330 if ( p2mt == p2m_mmio_direct ) in guest_remove_page() 350 if ( p2m_is_shared(p2mt) ) in guest_remove_page() 365 ASSERT(!p2m_is_shared(p2mt)); in guest_remove_page() 375 if ( !p2m_is_paging(p2mt) ) in guest_remove_page() 645 p2m_type_t p2mt; in memory_exchange() local 1700 p2m_type_t p2mt; in check_get_page_from_gfn() local 1706 if ( p2m_is_paging(p2mt) ) in check_get_page_from_gfn() 1728 *p2mt_p = p2mt; in check_get_page_from_gfn() [all …]
|
/xen/xen/arch/arm/ |
A D | guestcopy.c | 33 p2m_type_t p2mt; in translate_get_page() local 40 page = get_page_from_gfn(info.gpa.d, paddr_to_pfn(addr), &p2mt, P2M_ALLOC); in translate_get_page() 45 if ( !p2m_is_ram(p2mt) ) in translate_get_page()
|
A D | domain_build.c | 52 p2m_type_t p2mt; member 1247 mr_data->p2mt); in map_range_to_domain() 1260 addr, addr + len, mr_data->p2mt); in map_range_to_domain() 1272 p2m_type_t p2mt) in map_device_children() argument 1274 struct map_range_data mr_data = { .d = d, .p2mt = p2mt }; in map_device_children() 1360 p2m_type_t p2mt) in handle_device() argument 1406 struct map_range_data mr_data = { .d = d, .p2mt = p2mt }; in handle_device() 1420 res = map_device_children(d, dev, p2mt); in handle_device() 1429 p2m_type_t p2mt) in handle_node() argument 1535 res = handle_device(d, node, p2mt); in handle_node() [all …]
|
A D | p2m.c | 455 p2m_type_t p2mt; in p2m_get_page_from_gfn() local 456 mfn_t mfn = p2m_lookup(d, gfn, &p2mt); in p2m_get_page_from_gfn() 459 *t = p2mt; in p2m_get_page_from_gfn() 461 if ( !p2m_is_any_ram(p2mt) ) in p2m_get_page_from_gfn() 473 if ( p2m_is_foreign(p2mt) ) in p2m_get_page_from_gfn() 1318 p2m_type_t p2mt) in map_regions_p2mt() argument 1320 return p2m_insert_mapping(d, gfn, nr, mfn, p2mt); in map_regions_p2mt()
|
A D | mm.c | 1440 p2m_type_t p2mt; in xenmem_add_to_physmap_one() local 1461 page = get_page_from_gfn(od, idx, &p2mt, P2M_ALLOC); in xenmem_add_to_physmap_one() 1468 if ( p2m_is_ram(p2mt) ) in xenmem_add_to_physmap_one() 1469 t = (p2mt == p2m_ram_rw) ? p2m_map_foreign_rw : p2m_map_foreign_ro; in xenmem_add_to_physmap_one()
|
/xen/xen/include/asm-x86/ |
A D | p2m.h | 244 p2m_type_t p2mt, 249 p2m_type_t *p2mt, 735 void p2m_mem_paging_drop_page(struct domain *d, gfn_t gfn, p2m_type_t p2mt); 752 unsigned int page_order, p2m_type_t p2mt, 870 mfn_t *mfn, p2m_type_t *p2mt, 893 p2m_type_t p2mt, p2m_access_t p2ma); 906 static inline unsigned int p2m_get_iommu_flags(p2m_type_t p2mt, mfn_t mfn) in p2m_get_iommu_flags() argument 910 switch( p2mt ) in p2m_get_iommu_flags()
|
/xen/xen/include/public/ |
A D | vm_event.h | 344 uint32_t p2mt; member 350 uint32_t p2mt; member
|
/xen/xen/drivers/passthrough/amd/ |
A D | iommu_guest.c | 109 p2m_type_t p2mt; in guest_iommu_get_table_mfn() local 114 mfn = mfn_x(get_gfn(d, gfn + idx, &p2mt)); in guest_iommu_get_table_mfn() 315 p2m_type_t p2mt; in do_completion_wait() local 341 vaddr = map_domain_page(get_gfn(d, gfn ,&p2mt)); in do_completion_wait() 366 p2m_type_t p2mt; in do_invalidate_dte() local 401 gcr3_mfn = mfn_x(get_gfn(d, gcr3_gfn, &p2mt)); in do_invalidate_dte()
|
/xen/xen/arch/x86/hvm/svm/ |
A D | svm.c | 1674 p2m_type_t p2mt = p2m_invalid; in svm_do_nested_pgfault() local 1704 uint32_t p2mt; in svm_do_nested_pgfault() member 1708 mfn = __get_gfn_type_access(p2m, gfn, &p2mt, &p2ma, 0, NULL, 0); in svm_do_nested_pgfault() 1713 _d.p2mt = p2mt; in svm_do_nested_pgfault() 1733 mfn = __get_gfn_type_access(p2m, gfn, &p2mt, &p2ma, 0, NULL, 0); in svm_do_nested_pgfault() 1737 gpa, mfn_x(mfn), p2mt); in svm_do_nested_pgfault() 2225 p2m_type_t p2mt; in nsvm_get_nvmcb_page() local 2235 &p2mt, P2M_ALLOC | P2M_UNSHARE); in nsvm_get_nvmcb_page() 2239 if ( !p2m_is_ram(p2mt) || p2m_is_readonly(p2mt) ) in nsvm_get_nvmcb_page()
|
/xen/xen/arch/x86/ |
A D | mm.c | 2171 p2m_type_t p2mt; in mod_l1_entry() local 2178 if ( p2m_is_paged(p2mt) ) in mod_l1_entry() 2186 if ( p2mt == p2m_ram_paging_in && !page ) in mod_l1_entry() 3442 p2m_type_t p2mt; in do_mmuext_op() local 3984 p2m_type_t p2mt; in do_mmu_update() local 4014 if ( p2m_is_paged(p2mt) ) in do_mmu_update() 4611 p2m_type_t p2mt; in xenmem_add_to_physmap_one() local 4626 p2m_type_t p2mt; in xenmem_add_to_physmap_one() local 4629 mfn = get_gfn_unshare(d, gfn, &p2mt); in xenmem_add_to_physmap_one() 4631 if ( p2m_is_shared(p2mt) ) in xenmem_add_to_physmap_one() [all …]
|