/xen/xen/arch/x86/ |
A D | msr.c | 210 if ( !(cp->x86_vendor & X86_VENDOR_INTEL) || in guest_rdmsr() 211 !(boot_cpu_data.x86_vendor & X86_VENDOR_INTEL) ) in guest_rdmsr() 230 if ( !(cp->x86_vendor & (X86_VENDOR_INTEL | X86_VENDOR_AMD)) || in guest_rdmsr() 231 !(boot_cpu_data.x86_vendor & in guest_rdmsr() 268 if ( !(cp->x86_vendor & (X86_VENDOR_INTEL | X86_VENDOR_CENTAUR)) ) in guest_rdmsr() 411 if ( d->arch.cpuid->x86_vendor != X86_VENDOR_INTEL || val != 0 ) in guest_wrmsr() 421 d->arch.cpuid->x86_vendor != X86_VENDOR_AMD ) in guest_wrmsr() 433 d->arch.cpuid->x86_vendor != X86_VENDOR_INTEL ) in guest_wrmsr() 503 if ( !(cp->x86_vendor & (X86_VENDOR_INTEL | X86_VENDOR_CENTAUR)) ) in guest_wrmsr()
|
A D | spec_ctrl.c | 202 if ( boot_cpu_data.x86_vendor & (X86_VENDOR_AMD | X86_VENDOR_HYGON) ) in xpti_init_default() 436 if ( boot_cpu_data.x86_vendor == X86_VENDOR_INTEL && !cpu_has_hypervisor && in check_smt_enabled() 457 if ( boot_cpu_data.x86_vendor & (X86_VENDOR_AMD | X86_VENDOR_HYGON) ) in retpoline_safe() 460 if ( boot_cpu_data.x86_vendor != X86_VENDOR_INTEL || in retpoline_safe() 564 if ( boot_cpu_data.x86_vendor != X86_VENDOR_INTEL || in should_use_eager_fpu() 641 if ( boot_cpu_data.x86_vendor == X86_VENDOR_INTEL && in l1tf_calculations() 784 if ( boot_cpu_data.x86_vendor != X86_VENDOR_INTEL || in mds_calculations()
|
A D | cpuid.c | 251 switch ( p->x86_vendor ) in recalculate_misc() 305 ASSERT(p->x86_vendor == boot_cpu_data.x86_vendor); in calculate_raw_policy() 356 if ( boot_cpu_data.x86_vendor == X86_VENDOR_INTEL && in guest_common_default_feature_adjustments() 460 if ( (boot_cpu_data.x86_vendor & (X86_VENDOR_AMD | X86_VENDOR_HYGON)) && in calculate_hvm_max_policy() 570 p->x86_vendor = x86_cpuid_lookup_vendor( in recalculate_cpuid_policy() 576 ((p->x86_vendor & (X86_VENDOR_AMD | in recalculate_cpuid_policy() 619 if ( !(boot_cpu_data.x86_vendor & (X86_VENDOR_AMD | X86_VENDOR_HYGON)) ) in recalculate_cpuid_policy() 1004 if ( boot_cpu_data.x86_vendor != X86_VENDOR_INTEL || in guest_cpuid() 1059 if ( p->x86_vendor == X86_VENDOR_INTEL && in guest_cpuid()
|
A D | nmi.c | 224 switch (boot_cpu_data.x86_vendor) { in disable_lapic_nmi_watchdog() 397 switch (boot_cpu_data.x86_vendor) { in setup_apic_nmi_watchdog()
|
A D | domctl.c | 75 switch ( boot_cpu_data.x86_vendor ) in domain_cpu_policy_changed() 128 if ( boot_cpu_data.x86_vendor == X86_VENDOR_AMD ) in domain_cpu_policy_changed() 143 if ( boot_cpu_data.x86_vendor & in domain_cpu_policy_changed() 156 if ( boot_cpu_data.x86_vendor == X86_VENDOR_INTEL ) in domain_cpu_policy_changed() 180 if ( !(p->x86_vendor & (X86_VENDOR_AMD | X86_VENDOR_HYGON)) ) in domain_cpu_policy_changed() 183 switch ( boot_cpu_data.x86_vendor ) in domain_cpu_policy_changed() 213 if ( boot_cpu_data.x86_vendor == X86_VENDOR_INTEL && in domain_cpu_policy_changed()
|
A D | alternative.c | 104 switch ( boot_cpu_data.x86_vendor ) in arch_init_ideal_nops()
|
A D | apic.c | 96 if (boot_cpu_data.x86_vendor == X86_VENDOR_AMD && in modern_apic() 101 if (boot_cpu_data.x86_vendor == X86_VENDOR_HYGON) in modern_apic() 463 if ((boot_cpu_data.x86_vendor == X86_VENDOR_INTEL) && (boot_cpu_data.x86 == 15)) in init_bsp_APIC()
|
/xen/xen/arch/x86/cpu/ |
A D | common.c | 303 c->x86_vendor = x86_cpuid_lookup_vendor(ebx, ecx, edx); in early_cpu_init() 304 switch (c->x86_vendor) { in early_cpu_init() 351 if (!(c->x86_vendor & (X86_VENDOR_AMD | X86_VENDOR_HYGON))) in early_cpu_init() 367 c->x86_vendor = x86_cpuid_lookup_vendor(ebx, ecx, edx); in generic_identify() 368 if (boot_cpu_data.x86_vendor != c->x86_vendor) in generic_identify() 370 smp_processor_id(), c->x86_vendor, in generic_identify() 371 boot_cpu_data.x86_vendor); in generic_identify() 442 c->x86_vendor = X86_VENDOR_UNKNOWN; in identify_cpu() 497 "%02x/%02x", c->x86_vendor, c->x86_model); in identify_cpu() 714 vendor = x86_cpuid_vendor_to_str(c->x86_vendor); in print_cpu_info() [all …]
|
A D | intel_cacheinfo.c | 172 c->x86_vendor != X86_VENDOR_SHANGHAI) in init_intel_cacheinfo()
|
A D | amd.c | 374 if (cpu->x86_vendor != X86_VENDOR_AMD) in cpu_has_amd_erratum() 466 if (!((boot_cpu_data.x86_vendor == X86_VENDOR_AMD) && in check_syscfg_dram_mod_en()
|
/xen/xen/arch/x86/pv/ |
A D | emul-priv-op.c | 241 boot_cpu_data.x86_vendor == X86_VENDOR_AMD && in pci_cfg_ok() 891 if ( boot_cpu_data.x86_vendor != X86_VENDOR_AMD ) in read_msr() 899 if ( boot_cpu_data.x86_vendor != X86_VENDOR_AMD || in read_msr() 924 if ( boot_cpu_data.x86_vendor == X86_VENDOR_INTEL ) in read_msr() 930 if ( vpmu_msr || (boot_cpu_data.x86_vendor & in read_msr() 1011 if ( !(boot_cpu_data.x86_vendor & in write_msr() 1030 if ( boot_cpu_data.x86_vendor != X86_VENDOR_AMD || in write_msr() 1059 if ( !(boot_cpu_data.x86_vendor & in write_msr() 1069 if ( boot_cpu_data.x86_vendor != X86_VENDOR_INTEL ) in write_msr() 1079 if ( boot_cpu_data.x86_vendor == X86_VENDOR_INTEL ) in write_msr() [all …]
|
/xen/xen/arch/x86/cpu/mcheck/ |
A D | mce_amd.c | 175 BUG_ON(c->x86_vendor != X86_VENDOR_AMD); in mcequirk_lookup_amd_quirkdata() 291 if ( ci->x86_vendor != X86_VENDOR_HYGON ) in amd_mcheck_init() 342 return ci->x86_vendor == X86_VENDOR_HYGON ? in amd_mcheck_init()
|
A D | vmce.c | 60 if ( boot_cpu_data.x86_vendor == X86_VENDOR_INTEL ) in vmce_init_vcpu() 78 if ( boot_cpu_data.x86_vendor == X86_VENDOR_INTEL ) in vmce_restore_vcpu() 151 switch ( boot_cpu_data.x86_vendor ) in bank_mce_rdmsr() 282 switch ( boot_cpu_data.x86_vendor ) in bank_mce_wrmsr()
|
A D | non-fatal.c | 103 switch (c->x86_vendor) { in init_nonfatal_mce_checker()
|
A D | mce.c | 369 if ( boot_cpu_data.x86_vendor == X86_VENDOR_INTEL ) in mcheck_mca_logout() 600 c->x86_vendor == X86_VENDOR_INTEL && c->x86_model < 0x1a; in mce_firstbank() 798 switch ( c->x86_vendor ) in mcheck_init() 894 m.cpuvendor = boot_cpu_data.x86_vendor; in x86_mcinfo_apei_save() 996 xcp->mc_vendor = c->x86_vendor; in do_mc_get_cpu_info() 1158 if ( c->x86_vendor == X86_VENDOR_AMD ) in x86_mc_msrinject_verify() 1193 if ( c->x86_vendor != X86_VENDOR_AMD ) in x86_mc_msrinject_verify() 1201 if ( c->x86_vendor & (X86_VENDOR_AMD | X86_VENDOR_HYGON) ) in x86_mc_msrinject_verify()
|
A D | amd_nonfatal.c | 206 if (!(c->x86_vendor & (X86_VENDOR_AMD | X86_VENDOR_HYGON))) in amd_nonfatal_mcheck_init()
|
A D | mcaction.c | 110 (boot_cpu_data.x86_vendor == X86_VENDOR_INTEL && in mc_memerr_dhandler()
|
A D | mce.h | 167 switch (boot_cpu_data.x86_vendor) { in mce_vendor_bank_msr()
|
/xen/tools/fuzz/cpu-policy/ |
A D | afl-policy-fuzzer.c | 41 data_end = offsetof(typeof(*cp), x86_vendor) + sizeof(cp->x86_vendor); in check_cpuid()
|
/xen/xen/arch/x86/acpi/cpufreq/ |
A D | cpufreq.c | 63 if (cpu->x86_vendor != X86_VENDOR_INTEL || in check_est_cpu() 598 if (c->x86_vendor == X86_VENDOR_INTEL && c->cpuid_level >= 6) in acpi_cpufreq_cpu_init() 644 (boot_cpu_data.x86_vendor == X86_VENDOR_INTEL)) in cpufreq_driver_init() 647 (boot_cpu_data.x86_vendor & in cpufreq_driver_init() 660 if ( boot_cpu_data.x86_vendor & in cpufreq_cpu_init()
|
/xen/xen/arch/x86/cpu/microcode/ |
A D | core.c | 161 if ( boot_cpu_data.x86_vendor == X86_VENDOR_AMD ) in microcode_scan_module() 163 else if ( boot_cpu_data.x86_vendor == X86_VENDOR_INTEL ) in microcode_scan_module() 774 switch ( c->x86_vendor ) in early_microcode_init()
|
/xen/xen/arch/x86/acpi/ |
A D | cpu_idle.c | 153 if ( c->x86_vendor != X86_VENDOR_INTEL || c->x86 != 6 ) in do_get_hw_residencies() 909 else if ( (current_cpu_data.x86_vendor & in acpi_dead_idle() 1036 else if ( (c->x86_vendor == X86_VENDOR_INTEL) || in acpi_processor_power_init_bm_check() 1037 ((c->x86_vendor == X86_VENDOR_AMD) && (c->x86 == 0x15)) ) in acpi_processor_power_init_bm_check() 1053 if ( c->x86_vendor == X86_VENDOR_INTEL && in acpi_processor_power_init_bm_check() 1414 if ( boot_cpu_data.x86_vendor != X86_VENDOR_HYGON ) in amd_cpuidle_init() 1639 if ( (boot_cpu_data.x86_vendor & in cpu_callback()
|
/xen/xen/arch/x86/cpu/mtrr/ |
A D | generic.c | 221 if ((boot_cpu_data.x86_vendor == X86_VENDOR_AMD && in print_mtrr_state() 223 boot_cpu_data.x86_vendor == X86_VENDOR_HYGON) { in print_mtrr_state()
|
/xen/xen/include/xen/lib/x86/ |
A D | cpuid.h | 278 uint8_t x86_vendor; member
|
/xen/xen/arch/x86/oprofile/ |
A D | nmi_int.c | 378 __u8 vendor = current_cpu_data.x86_vendor; in nmi_init()
|