/xen/xen/arch/x86/ |
A D | machine_kexec.c | 37 l3_pgentry_t *l3 = NULL; in machine_kexec_add_page() local 63 l3 = __map_domain_page(l3_page); in machine_kexec_add_page() 64 l3 += l3_table_offset(vaddr); in machine_kexec_add_page() 65 if ( !(l3e_get_flags(*l3) & _PAGE_PRESENT) ) in machine_kexec_add_page() 70 l3e_write(l3, l3e_from_page(l2_page, __PAGE_HYPERVISOR)); in machine_kexec_add_page() 73 l2_page = l3e_get_page(*l3); in machine_kexec_add_page() 97 if ( l3 ) in machine_kexec_add_page() 98 unmap_domain_page(l3); in machine_kexec_add_page()
|
A D | sysctl.c | 246 (psr_cmt->l3.features & PSR_CMT_L3_OCCUPANCY); in arch_do_sysctl() 252 sysctl->u.psr_cmt_op.u.data = psr_cmt->l3.upscaling_factor; in arch_do_sysctl() 275 sysctl->u.psr_cmt_op.u.data = psr_cmt->l3.features; in arch_do_sysctl()
|
A D | psr.c | 643 psr_cmt->l3.upscaling_factor = ebx; in init_psr_cmt() 644 psr_cmt->l3.rmid_max = ecx; in init_psr_cmt() 645 psr_cmt->l3.features = edx; in init_psr_cmt() 648 psr_cmt->rmid_max = min(psr_cmt->rmid_max, psr_cmt->l3.rmid_max); in init_psr_cmt()
|
A D | mm.c | 1164 define_get_linear_pagetable(l3); 2349 rc = UPDATE_ENTRY(l3, pl3e, ol3e, nl3e, mfn, vcpu, preserve_ad); in mod_l3_entry() 2359 if ( unlikely(!UPDATE_ENTRY(l3, pl3e, ol3e, nl3e, mfn, vcpu, in mod_l3_entry() 2368 else if ( unlikely(!UPDATE_ENTRY(l3, pl3e, ol3e, nl3e, mfn, vcpu, in mod_l3_entry()
|
A D | domain.c | 2182 PROGRESS(l3): in domain_relinquish_resources()
|
/xen/xen/arch/x86/cpu/ |
A D | intel_cacheinfo.c | 121 unsigned int trace = 0, l1i = 0, l1d = 0, l2 = 0, l3 = 0; /* Cache sizes */ in init_intel_cacheinfo() local 216 l3 += cache_table[k].size; in init_intel_cacheinfo() 243 l3 = new_l3; in init_intel_cacheinfo() 260 if (l3) in init_intel_cacheinfo() 261 printk("CPU: L3 cache: %dK\n", l3); in init_intel_cacheinfo() 264 c->x86_cache_size = l3 ? l3 : (l2 ? l2 : (l1i+l1d)); in init_intel_cacheinfo()
|
/xen/xen/arch/x86/mm/ |
A D | paging.c | 112 mfn_t *l4, *l3, *l2; in paging_free_log_dirty_bitmap() local 173 unmap_domain_page(l3); in paging_free_log_dirty_bitmap() 263 mfn_t mfn, *l4, *l3, *l2; in paging_mark_pfn_dirty() local 304 l3 = map_domain_page(mfn); in paging_mark_pfn_dirty() 305 mfn = l3[i3]; in paging_mark_pfn_dirty() 308 unmap_domain_page(l3); in paging_mark_pfn_dirty() 357 mfn_t mfn, *l4, *l3, *l2; in paging_mfn_is_dirty() local 382 unmap_domain_page(l3); in paging_mfn_is_dirty() 478 l2 = ((l3 && mfn_valid(l3[i3])) ? in paging_log_dirty_op() 520 if ( l3 ) in paging_log_dirty_op() [all …]
|
A D | guest_walk.c | 498 if ( set_ad_bits(&l3p[guest_l3_table_offset(va)].l3, &gw->l3e.l3, in guest_walk_tables()
|
A D | p2m-pt.c | 594 entry_content.l1 = l3e_content.l3; in p2m_pt_set_entry()
|
/xen/xen/xsm/flask/ss/ |
A D | mls_types.h | 51 #define mls_level_between(l1, l2, l3) \ argument 52 (mls_level_dom((l1), (l2)) && mls_level_dom((l3), (l1)))
|
/xen/xen/include/asm-x86/ |
A D | page.h | 62 #define l3e_get_intpte(x) ((x).l3) 71 ((unsigned long)(((x).l3 & (PADDR_MASK&PAGE_MASK)) >> PAGE_SHIFT)) 87 ((paddr_t)(((x).l3 & (PADDR_MASK&PAGE_MASK)))) 100 #define l3e_get_flags(x) (get_pte_flags((x).l3)) 106 #define l3e_get_pkey(x) get_pte_pkey((x).l3) 169 #define l3e_add_flags(x, flags) ((x).l3 |= put_pte_flags(flags)) 175 #define l3e_remove_flags(x, flags) ((x).l3 &= ~put_pte_flags(flags)) 187 ( !!(((x).l3 ^ (y).l3) & ((PADDR_MASK&PAGE_MASK)|put_pte_flags(flags))) )
|
A D | guest_pt.h | 319 return ((l3e.l3 & (guest_rsvd_bits(v) | GUEST_L3_PAGETABLE_RSVD | in guest_l3e_rsvd_bits() 321 ((l3e.l3 & _PAGE_PSE) && in guest_l3e_rsvd_bits() 322 (l3e.l3 & SUPERPAGE_RSVD(GUEST_L3_PAGETABLE_SHIFT)))); in guest_l3e_rsvd_bits() 444 gw->l3e.l3, mfn_x(gw->l3mfn)); in print_gw() 446 gprintk(XENLOG_INFO, " l3e=%" PRI_gpte "\n", gw->l3e.l3); in print_gw()
|
A D | psr.h | 56 struct psr_cmt_l3 l3; member
|
A D | shadow.h | 226 return pv_l1tf_check_pte(d, 3, l3e.l3); in pv_l1tf_check_l3e()
|
/xen/xen/include/asm-x86/x86_64/ |
A D | page.h | 92 typedef struct { intpte_t l3; } l3_pgentry_t; member
|
/xen/tools/libfsimage/xfs/ |
A D | xfs.h | 206 xfs_uint32_t l0, l1, l2, l3; member
|
A D | fsys_xfs.c | 135 (((xfs_fsblock_t)le32 (r->l3)) >> 21); in xt_start() 149 return le32(r->l3) & mask32lo(21); in xt_len()
|
/xen/xen/arch/x86/mm/shadow/ |
A D | multi.c | 247 mismatch |= (gw->l3e.l3 != l3p[guest_l3_table_offset(va)].l3); in shadow_check_gwalk() 250 mismatch |= (gw->l3e.l3 != in shadow_check_gwalk() 251 v->arch.paging.shadow.gl3e[guest_l3_table_offset(va)].l3); in shadow_check_gwalk() 734 _sh_propagate(v, gl3e.l3, sl2mfn, sl3e, 3, ft, p2m_ram_rw); in l3e_propagate_from_guest() 954 if ( old_sl3e.l3 == new_sl3e.l3 ) return 0; /* Nothing to do */ in shadow_set_l3e()
|
/xen/docs/features/ |
A D | intel_psr_cat_cdp.pandoc | 68 `-l3`: Show cbm for L3 cache. 84 `-l3`: Specify cbm for L3 cache.
|