[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [PATCH] x86: drop Xeon Phi support
Do as was decided in Lisbon. Reportedly Xen hasn't been working very well on those processors anyway. Signed-off-by: Jan Beulich <jbeulich@xxxxxxxx> --- One file I left untouched is the test harness'es predicates.c: Those tests are imo fine to retain. Plus of course the dependencies in gen-cpuid.py also want leaving in place imo (unless we were to remove the respective 4 lines from the public header). --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -11,6 +11,8 @@ The format is based on [Keep a Changelog ### Added ### Removed + - On x86: + - Support for running on Xeon Phi processors. ## [4.19.0](https://xenbits.xenproject.org/gitweb/?p=xen.git;a=shortlog;h=RELEASE-4.19.0) - 2024-07-29 --- a/SUPPORT.md +++ b/SUPPORT.md @@ -30,6 +30,7 @@ if the related features are marked as su ### x86-64 Status: Supported + Status, Xeon Phi: Not supported. ### ARM v7 + Virtualization Extensions --- a/tools/tests/x86_emulator/Makefile +++ b/tools/tests/x86_emulator/Makefile @@ -16,7 +16,7 @@ vpath %.c $(XEN_ROOT)/xen/lib/x86 CFLAGS += $(CFLAGS_xeninclude) -SIMD := 3dnow sse sse2 sse4 avx avx2 xop avx512f avx512bw avx512dq avx512er avx512vbmi avx512fp16 +SIMD := 3dnow sse sse2 sse4 avx avx2 xop avx512f avx512bw avx512dq avx512vbmi avx512fp16 FMA := fma4 fma SG := avx2-sg avx512f-sg AES := ssse3-aes avx-aes avx2-vaes avx512bw-vaes @@ -80,9 +80,6 @@ avx512bw-flts := avx512dq-vecs := $(avx512f-vecs) avx512dq-ints := $(avx512f-ints) avx512dq-flts := $(avx512f-flts) -avx512er-vecs := 64 -avx512er-ints := -avx512er-flts := 4 8 avx512vbmi-vecs := $(avx512bw-vecs) avx512vbmi-ints := $(avx512bw-ints) avx512vbmi-flts := $(avx512bw-flts) --- a/tools/tests/x86_emulator/evex-disp8.c +++ b/tools/tests/x86_emulator/evex-disp8.c @@ -524,37 +524,6 @@ static const struct test avx512dq_512[] INSN(inserti32x8, 66, 0f3a, 3a, el_8, d, vl), }; -static const struct test avx512er_512[] = { - INSN(exp2, 66, 0f38, c8, vl, sd, vl), - INSN(rcp28, 66, 0f38, ca, vl, sd, vl), - INSN(rcp28, 66, 0f38, cb, el, sd, el), - INSN(rsqrt28, 66, 0f38, cc, vl, sd, vl), - INSN(rsqrt28, 66, 0f38, cd, el, sd, el), -}; - -static const struct test avx512pf_512[] = { - INSNX(gatherpf0d, 66, 0f38, c6, 1, vl, sd, el), - INSNX(gatherpf0q, 66, 0f38, c7, 1, vl, sd, el), - INSNX(gatherpf1d, 66, 0f38, c6, 2, vl, sd, el), - INSNX(gatherpf1q, 66, 0f38, c7, 2, vl, sd, el), - INSNX(scatterpf0d, 66, 0f38, c6, 5, vl, sd, el), - INSNX(scatterpf0q, 66, 0f38, c7, 5, vl, sd, el), - INSNX(scatterpf1d, 66, 0f38, c6, 6, vl, sd, el), - INSNX(scatterpf1q, 66, 0f38, c7, 6, vl, sd, el), -}; - -static const struct test avx512_4fmaps_512[] = { - INSN(4fmaddps, f2, 0f38, 9a, el_4, d, vl), - INSN(4fmaddss, f2, 0f38, 9b, el_4, d, vl), - INSN(4fnmaddps, f2, 0f38, aa, el_4, d, vl), - INSN(4fnmaddss, f2, 0f38, ab, el_4, d, vl), -}; - -static const struct test avx512_4vnniw_512[] = { - INSN(p4dpwssd, f2, 0f38, 52, el_4, d, vl), - INSN(p4dpwssds, f2, 0f38, 53, el_4, d, vl), -}; - static const struct test avx512_bf16_all[] = { INSN(cvtne2ps2bf16, f2, 0f38, 72, vl, d, vl), INSN(cvtneps2bf16, f3, 0f38, 72, vl, d, vl), @@ -1110,11 +1079,6 @@ void evex_disp8_test(void *instr, struct RUN(avx512dq, 128); RUN(avx512dq, no128); RUN(avx512dq, 512); - RUN(avx512er, 512); -#define cpu_has_avx512pf cpu_has_avx512f - RUN(avx512pf, 512); - RUN(avx512_4fmaps, 512); - RUN(avx512_4vnniw, 512); RUN(avx512_bf16, all); RUN(avx512_bitalg, all); RUN(avx512_ifma, all); --- a/tools/tests/x86_emulator/test_x86_emulator.c +++ b/tools/tests/x86_emulator/test_x86_emulator.c @@ -39,7 +39,6 @@ asm ( ".pushsection .test, \"ax\", @prog #include "avx512bw-vpclmulqdq.h" #include "avx512bw-gf.h" #include "avx512dq.h" -#include "avx512er.h" #include "avx512vbmi.h" #include "avx512vbmi2-vpclmulqdq.h" #include "avx512fp16.h" @@ -147,11 +146,6 @@ static bool simd_check_avx512dq_vl(void) return cpu_has_avx512dq && cpu_has_avx512vl; } -static bool simd_check_avx512er(void) -{ - return cpu_has_avx512er; -} - static bool simd_check_avx512bw(void) { return cpu_has_avx512bw; @@ -507,10 +501,6 @@ static const struct { AVX512VL(DQ+VL u64x2, avx512dq, 16u8), AVX512VL(DQ+VL s64x4, avx512dq, 32i8), AVX512VL(DQ+VL u64x4, avx512dq, 32u8), - SIMD(AVX512ER f32 scalar,avx512er, f4), - SIMD(AVX512ER f32x16, avx512er, 64f4), - SIMD(AVX512ER f64 scalar,avx512er, f8), - SIMD(AVX512ER f64x8, avx512er, 64f8), SIMD(AVX512_VBMI s8x64, avx512vbmi, 64i1), SIMD(AVX512_VBMI u8x64, avx512vbmi, 64u1), SIMD(AVX512_VBMI s16x32, avx512vbmi, 64i2), @@ -4983,81 +4973,6 @@ int main(int argc, char **argv) } #endif - printf("%-40s", "Testing v4fmaddps 32(%ecx),%zmm4,%zmm4{%k5}..."); - if ( stack_exec && cpu_has_avx512_4fmaps ) - { - decl_insn(v4fmaddps); - static const struct { - float f[16]; - } in = {{ - 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16 - }}, out = {{ - 1 + 1 * 9 + 2 * 10 + 3 * 11 + 4 * 12, - 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, - 16 + 16 * 9 + 17 * 10 + 18 * 11 + 19 * 12 - }}; - - asm volatile ( "vmovups %1, %%zmm4\n\t" - "vbroadcastss %%xmm4, %%zmm7\n\t" - "vaddps %%zmm4, %%zmm7, %%zmm5\n\t" - "vaddps %%zmm5, %%zmm7, %%zmm6\n\t" - "vaddps %%zmm6, %%zmm7, %%zmm7\n\t" - "kmovw %2, %%k5\n" - put_insn(v4fmaddps, - "v4fmaddps 32(%0), %%zmm4, %%zmm4%{%%k5%}") - :: "c" (NULL), "m" (in), "rmk" (0x8001) ); - - set_insn(v4fmaddps); - regs.ecx = (unsigned long)∈ - rc = x86_emulate(&ctxt, &emulops); - if ( rc != X86EMUL_OKAY || !check_eip(v4fmaddps) ) - goto fail; - - asm ( "vcmpeqps %1, %%zmm4, %%k0\n\t" - "kmovw %%k0, %0" : "=g" (rc) : "m" (out) ); - if ( rc != 0xffff ) - goto fail; - printf("okay\n"); - } - else - printf("skipped\n"); - - printf("%-40s", "Testing v4fnmaddss 16(%edx),%zmm4,%zmm4{%k3}..."); - if ( stack_exec && cpu_has_avx512_4fmaps ) - { - decl_insn(v4fnmaddss); - static const struct { - float f[16]; - } in = {{ - 1, 2, 3, 4, 5, 6, 7, 8 - }}, out = {{ - 1 - 1 * 5 - 2 * 6 - 3 * 7 - 4 * 8, 2, 3, 4 - }}; - - asm volatile ( "vmovups %1, %%xmm4\n\t" - "vaddss %%xmm4, %%xmm4, %%xmm5\n\t" - "vaddss %%xmm5, %%xmm4, %%xmm6\n\t" - "vaddss %%xmm6, %%xmm4, %%xmm7\n\t" - "kmovw %2, %%k3\n" - put_insn(v4fnmaddss, - "v4fnmaddss 16(%0), %%xmm4, %%xmm4%{%%k3%}") - :: "d" (NULL), "m" (in), "rmk" (1) ); - - set_insn(v4fnmaddss); - regs.edx = (unsigned long)∈ - rc = x86_emulate(&ctxt, &emulops); - if ( rc != X86EMUL_OKAY || !check_eip(v4fnmaddss) ) - goto fail; - - asm ( "vcmpeqps %1, %%zmm4, %%k0\n\t" - "kmovw %%k0, %0" : "=g" (rc) : "m" (out) ); - if ( rc != 0xffff ) - goto fail; - printf("okay\n"); - } - else - printf("skipped\n"); - if ( stack_exec && cpu_has_avx512_bf16 ) { decl_insn(vcvtne2ps2bf16); --- a/tools/tests/x86_emulator/x86-emulate.c +++ b/tools/tests/x86_emulator/x86-emulate.c @@ -84,7 +84,6 @@ bool emul_test_init(void) cp.basic.movbe = true; cp.feat.invpcid = true; cp.feat.adx = true; - cp.feat.avx512pf = cp.feat.avx512f; cp.feat.rdpid = true; cp.feat.wrmsrns = true; cp.extd.clzero = true; --- a/tools/tests/x86_emulator/x86-emulate.h +++ b/tools/tests/x86_emulator/x86-emulate.h @@ -156,7 +156,6 @@ void wrpkru(unsigned int val); #define cpu_has_avx512f (cp.feat.avx512f && xcr0_mask(0xe6)) #define cpu_has_avx512dq (cp.feat.avx512dq && xcr0_mask(0xe6)) #define cpu_has_avx512_ifma (cp.feat.avx512_ifma && xcr0_mask(0xe6)) -#define cpu_has_avx512er (cp.feat.avx512er && xcr0_mask(0xe6)) #define cpu_has_avx512cd (cp.feat.avx512cd && xcr0_mask(0xe6)) #define cpu_has_sha cp.feat.sha #define cpu_has_avx512bw (cp.feat.avx512bw && xcr0_mask(0xe6)) @@ -171,8 +170,6 @@ void wrpkru(unsigned int val); #define cpu_has_avx512_vpopcntdq (cp.feat.avx512_vpopcntdq && xcr0_mask(0xe6)) #define cpu_has_movdiri cp.feat.movdiri #define cpu_has_movdir64b cp.feat.movdir64b -#define cpu_has_avx512_4vnniw (cp.feat.avx512_4vnniw && xcr0_mask(0xe6)) -#define cpu_has_avx512_4fmaps (cp.feat.avx512_4fmaps && xcr0_mask(0xe6)) #define cpu_has_avx512_vp2intersect (cp.feat.avx512_vp2intersect && xcr0_mask(0xe6)) #define cpu_has_serialize cp.feat.serialize #define cpu_has_avx512_fp16 (cp.feat.avx512_fp16 && xcr0_mask(0xe6)) --- a/xen/arch/x86/acpi/cpu_idle.c +++ b/xen/arch/x86/acpi/cpu_idle.c @@ -73,7 +73,6 @@ #define GET_CC3_RES(val) GET_HW_RES_IN_NS(0x3FC, val) #define GET_CC6_RES(val) GET_HW_RES_IN_NS(0x3FD, val) #define GET_CC7_RES(val) GET_HW_RES_IN_NS(0x3FE, val) /* SNB onwards */ -#define PHI_CC6_RES(val) GET_HW_RES_IN_NS(0x3FF, val) /* Xeon Phi only */ static void cf_check lapic_timer_nop(void) { } void (*__read_mostly lapic_timer_off)(void); @@ -222,18 +221,6 @@ static void cf_check do_get_hw_residenci GET_CC6_RES(hw_res->cc6); GET_CC7_RES(hw_res->cc7); break; - /* Xeon Phi Knights Landing */ - case 0x57: - /* Xeon Phi Knights Mill */ - case 0x85: - GET_CC3_RES(hw_res->mc0); /* abusing GET_CC3_RES */ - GET_CC6_RES(hw_res->mc6); /* abusing GET_CC6_RES */ - GET_PC2_RES(hw_res->pc2); - GET_PC3_RES(hw_res->pc3); - GET_PC6_RES(hw_res->pc6); - GET_PC7_RES(hw_res->pc7); - PHI_CC6_RES(hw_res->cc6); - break; /* various Atoms */ case 0x27: GET_PC3_RES(hw_res->pc2); /* abusing GET_PC3_RES */ --- a/xen/arch/x86/cpu/mcheck/mce_intel.c +++ b/xen/arch/x86/cpu/mcheck/mce_intel.c @@ -875,10 +875,8 @@ static void intel_init_ppin(const struct case 0x4f: /* Broadwell X */ case 0x55: /* Skylake X */ case 0x56: /* Broadwell Xeon D */ - case 0x57: /* Knights Landing */ case 0x6a: /* Icelake X */ case 0x6c: /* Icelake D */ - case 0x85: /* Knights Mill */ case 0x8f: /* Sapphire Rapids X */ if ( (c != &boot_cpu_data && !ppin_msr) || --- a/xen/arch/x86/cpu/mwait-idle.c +++ b/xen/arch/x86/cpu/mwait-idle.c @@ -778,22 +778,6 @@ static const struct cpuidle_state avn_cs {} }; -static const struct cpuidle_state knl_cstates[] = { - { - .name = "C1", - .flags = MWAIT2flg(0x00), - .exit_latency = 1, - .target_residency = 2, - }, - { - .name = "C6", - .flags = MWAIT2flg(0x10) | CPUIDLE_FLAG_TLB_FLUSHED, - .exit_latency = 120, - .target_residency = 500, - }, - {} -}; - static struct cpuidle_state __read_mostly bxt_cstates[] = { { .name = "C1", @@ -1121,10 +1105,6 @@ static const struct idle_cpu idle_cpu_av .c1e_promotion = C1E_PROMOTION_DISABLE, }; -static const struct idle_cpu idle_cpu_knl = { - .state_table = knl_cstates, -}; - static const struct idle_cpu idle_cpu_bxt = { .state_table = bxt_cstates, .c1e_promotion = C1E_PROMOTION_DISABLE, @@ -1181,8 +1161,6 @@ static const struct x86_cpu_id intel_idl ICPU(ALDERLAKE, adl), ICPU(ALDERLAKE_L, adl_l), ICPU(SAPPHIRERAPIDS_X, spr), - ICPU(XEON_PHI_KNL, knl), - ICPU(XEON_PHI_KNM, knl), ICPU(ATOM_GOLDMONT, bxt), ICPU(ATOM_GOLDMONT_PLUS, bxt), ICPU(ATOM_GOLDMONT_D, dnv), --- a/xen/arch/x86/hvm/vmx/vmx.c +++ b/xen/arch/x86/hvm/vmx/vmx.c @@ -533,10 +533,6 @@ static const struct lbr_info *__init get return at_lbr; /* Silvermont */ case 0x37: case 0x4a: case 0x4d: case 0x5a: case 0x5d: - /* Xeon Phi Knights Landing */ - case 0x57: - /* Xeon Phi Knights Mill */ - case 0x85: /* Airmont */ case 0x4c: return sm_lbr; @@ -2984,13 +2980,6 @@ static bool __init has_if_pschange_mc(vo case 0x7a: /* Gemini Lake */ case 0x86: /* Jacobsville */ - /* - * Knights processors are not vulnerable. - */ - case 0x57: /* Knights Landing */ - case 0x85: /* Knights Mill */ - return false; - default: printk("Unrecognised CPU model %#x - assuming vulnerable to IF_PSCHANGE_MC\n", boot_cpu_data.x86_model); --- a/xen/arch/x86/include/asm/cpufeature.h +++ b/xen/arch/x86/include/asm/cpufeature.h @@ -140,7 +140,6 @@ static inline bool boot_cpu_has(unsigned #define cpu_has_avx512_ifma boot_cpu_has(X86_FEATURE_AVX512_IFMA) #define cpu_has_clflushopt boot_cpu_has(X86_FEATURE_CLFLUSHOPT) #define cpu_has_clwb boot_cpu_has(X86_FEATURE_CLWB) -#define cpu_has_avx512er boot_cpu_has(X86_FEATURE_AVX512ER) #define cpu_has_avx512cd boot_cpu_has(X86_FEATURE_AVX512CD) #define cpu_has_proc_trace boot_cpu_has(X86_FEATURE_PROC_TRACE) #define cpu_has_sha boot_cpu_has(X86_FEATURE_SHA) @@ -174,8 +173,6 @@ static inline bool boot_cpu_has(unsigned #define cpu_has_auto_ibrs boot_cpu_has(X86_FEATURE_AUTO_IBRS) /* CPUID level 0x00000007:0.edx */ -#define cpu_has_avx512_4vnniw boot_cpu_has(X86_FEATURE_AVX512_4VNNIW) -#define cpu_has_avx512_4fmaps boot_cpu_has(X86_FEATURE_AVX512_4FMAPS) #define cpu_has_avx512_vp2intersect boot_cpu_has(X86_FEATURE_AVX512_VP2INTERSECT) #define cpu_has_srbds_ctrl boot_cpu_has(X86_FEATURE_SRBDS_CTRL) #define cpu_has_md_clear boot_cpu_has(X86_FEATURE_MD_CLEAR) --- a/xen/arch/x86/include/asm/intel-family.h +++ b/xen/arch/x86/include/asm/intel-family.h @@ -13,8 +13,8 @@ * INTEL_FAM6{OPTFAMILY}_{MICROARCH}{OPTDIFF} * where: * OPTFAMILY Describes the family of CPUs that this belongs to. Default - * is assumed to be "_CORE" (and should be omitted). Other values - * currently in use are _ATOM and _XEON_PHI + * is assumed to be "_CORE" (and should be omitted). The other + * value currently in use is _ATOM. * MICROARCH Is the code name for the micro-architecture for this core. * N.B. Not the platform name. * OPTDIFF If needed, a short string to differentiate by market segment. @@ -164,11 +164,6 @@ #define INTEL_FAM6_ATOM_DARKMONT_X 0xDD /* Clearwater Forest */ -/* Xeon Phi */ - -#define INTEL_FAM6_XEON_PHI_KNL 0x57 /* Knights Landing */ -#define INTEL_FAM6_XEON_PHI_KNM 0x85 /* Knights Mill */ - /* Family 5 */ #define INTEL_FAM5_QUARK_X1000 0x09 /* Quark X1000 SoC */ --- a/xen/arch/x86/spec_ctrl.c +++ b/xen/arch/x86/spec_ctrl.c @@ -887,11 +887,9 @@ static bool __init retpoline_calculation case 0x4d: /* Avaton / Rangely (Silvermont) */ case 0x4c: /* Cherrytrail / Brasswell */ case 0x4a: /* Merrifield */ - case 0x57: /* Knights Landing */ case 0x5a: /* Moorefield */ case 0x5c: /* Goldmont */ case 0x5f: /* Denverton */ - case 0x85: /* Knights Mill */ safe = true; break; @@ -1005,13 +1003,6 @@ static bool __init should_use_eager_fpu( case 0x7a: /* Gemini Lake */ return false; - /* - * Knights processors are not vulnerable. - */ - case 0x57: /* Knights Landing */ - case 0x85: /* Knights Mill */ - return false; - default: printk("Unrecognised CPU model %#x - assuming vulnerable to LazyFPU\n", boot_cpu_data.x86_model); @@ -1259,13 +1250,6 @@ static __init void l1tf_calculations(voi case 0x7a: /* Gemini Lake */ break; - /* - * Knights processors are not vulnerable. - */ - case 0x57: /* Knights Landing */ - case 0x85: /* Knights Mill */ - break; - default: /* Defer printk() until we've accounted for RDCL_NO. */ hit_default = true; @@ -1416,13 +1400,6 @@ static __init void mds_calculations(void case 0x65: /* SoFIA LTE AOSP */ case 0x6e: /* Cougar Mountain */ case 0x75: /* Lightning Mountain */ - /* - * Knights processors (which are based on the Silvermont/Airmont - * microarchitecture) are similarly only affected by the Store Buffer - * aspect. - */ - case 0x57: /* Knights Landing */ - case 0x85: /* Knights Mill */ cpu_has_bug_msbds_only = true; break; --- a/xen/arch/x86/x86_emulate/private.h +++ b/xen/arch/x86/x86_emulate/private.h @@ -564,8 +564,6 @@ amd_like(const struct x86_emulate_ctxt * #define vcpu_has_avx512_ifma() (ctxt->cpuid->feat.avx512_ifma) #define vcpu_has_clflushopt() (ctxt->cpuid->feat.clflushopt) #define vcpu_has_clwb() (ctxt->cpuid->feat.clwb) -#define vcpu_has_avx512pf() (ctxt->cpuid->feat.avx512pf) -#define vcpu_has_avx512er() (ctxt->cpuid->feat.avx512er) #define vcpu_has_avx512cd() (ctxt->cpuid->feat.avx512cd) #define vcpu_has_sha() (ctxt->cpuid->feat.sha) #define vcpu_has_avx512bw() (ctxt->cpuid->feat.avx512bw) @@ -582,8 +580,6 @@ amd_like(const struct x86_emulate_ctxt * #define vcpu_has_movdiri() (ctxt->cpuid->feat.movdiri) #define vcpu_has_movdir64b() (ctxt->cpuid->feat.movdir64b) #define vcpu_has_enqcmd() (ctxt->cpuid->feat.enqcmd) -#define vcpu_has_avx512_4vnniw() (ctxt->cpuid->feat.avx512_4vnniw) -#define vcpu_has_avx512_4fmaps() (ctxt->cpuid->feat.avx512_4fmaps) #define vcpu_has_avx512_vp2intersect() (ctxt->cpuid->feat.avx512_vp2intersect) #define vcpu_has_serialize() (ctxt->cpuid->feat.serialize) #define vcpu_has_tsxldtrk() (ctxt->cpuid->feat.tsxldtrk) --- a/xen/arch/x86/x86_emulate/x86_emulate.c +++ b/xen/arch/x86/x86_emulate/x86_emulate.c @@ -6029,7 +6029,6 @@ x86_emulate( case X86EMUL_OPC_EVEX_66(0x0f38, 0xbd): /* vfnmadd231s{s,d} xmm/mem,xmm,xmm{k} */ case X86EMUL_OPC_EVEX_66(0x0f38, 0xbf): /* vfnmsub231s{s,d} xmm/mem,xmm,xmm{k} */ host_and_vcpu_must_have(avx512f); - simd_zmm_scalar_sae: generate_exception_if(ea.type != OP_REG && evex.brs, X86_EXC_UD); if ( !evex.brs ) avx512_vlen_check(true); @@ -6123,15 +6122,6 @@ x86_emulate( avx512_vlen_check(true); goto simd_zmm; - case X86EMUL_OPC_EVEX_F2(0x0f38, 0x52): /* vp4dpwssd m128,zmm+3,zmm{k} */ - case X86EMUL_OPC_EVEX_F2(0x0f38, 0x53): /* vp4dpwssds m128,zmm+3,zmm{k} */ - host_and_vcpu_must_have(avx512_4vnniw); - generate_exception_if((ea.type != OP_MEM || evex.w || evex.brs || - evex.lr != 2), - X86_EXC_UD); - op_mask = op_mask & 0xffff ? 0xf : 0; - goto simd_zmm; - case X86EMUL_OPC_EVEX_66(0x0f38, 0x8f): /* vpshufbitqmb [xyz]mm/mem,[xyz]mm,k{k} */ generate_exception_if(evex.w || !evex.r || !evex.R || evex.z, X86_EXC_UD); /* fall through */ @@ -6601,24 +6591,6 @@ x86_emulate( host_and_vcpu_must_have(fma); goto simd_0f_ymm; - case X86EMUL_OPC_EVEX_F2(0x0f38, 0x9a): /* v4fmaddps m128,zmm+3,zmm{k} */ - case X86EMUL_OPC_EVEX_F2(0x0f38, 0xaa): /* v4fnmaddps m128,zmm+3,zmm{k} */ - host_and_vcpu_must_have(avx512_4fmaps); - generate_exception_if((ea.type != OP_MEM || evex.w || evex.brs || - evex.lr != 2), - X86_EXC_UD); - op_mask = op_mask & 0xffff ? 0xf : 0; - goto simd_zmm; - - case X86EMUL_OPC_EVEX_F2(0x0f38, 0x9b): /* v4fmaddss m128,xmm+3,xmm{k} */ - case X86EMUL_OPC_EVEX_F2(0x0f38, 0xab): /* v4fnmaddss m128,xmm+3,xmm{k} */ - host_and_vcpu_must_have(avx512_4fmaps); - generate_exception_if((ea.type != OP_MEM || evex.w || evex.brs || - evex.lr == 3), - X86_EXC_UD); - op_mask = op_mask & 1 ? 0xf : 0; - goto simd_zmm; - case X86EMUL_OPC_EVEX_66(0x0f38, 0xa0): /* vpscatterd{d,q} [xyz]mm,mem{k} */ case X86EMUL_OPC_EVEX_66(0x0f38, 0xa1): /* vpscatterq{d,q} [xyz]mm,mem{k} */ case X86EMUL_OPC_EVEX_66(0x0f38, 0xa2): /* vscatterdp{s,d} [xyz]mm,mem{k} */ @@ -6759,97 +6731,6 @@ x86_emulate( generate_exception_if(!evex.w, X86_EXC_UD); goto avx512f_no_sae; - case X86EMUL_OPC_EVEX_66(0x0f38, 0xc6): - case X86EMUL_OPC_EVEX_66(0x0f38, 0xc7): - { -#ifndef __XEN__ - typeof(evex) *pevex; - union { - int32_t dw[16]; - int64_t qw[8]; - } index; -#endif - - ASSERT(ea.type == OP_MEM); - generate_exception_if((!cpu_has_avx512f || !evex.opmsk || evex.brs || - evex.z || evex.reg != 0xf || evex.lr != 2), - X86_EXC_UD); - - switch ( modrm_reg & 7 ) - { - case 1: /* vgatherpf0{d,q}p{s,d} mem{k} */ - case 2: /* vgatherpf1{d,q}p{s,d} mem{k} */ - case 5: /* vscatterpf0{d,q}p{s,d} mem{k} */ - case 6: /* vscatterpf1{d,q}p{s,d} mem{k} */ - vcpu_must_have(avx512pf); - break; - default: - generate_exception(X86_EXC_UD); - } - - get_fpu(X86EMUL_FPU_zmm); - -#ifndef __XEN__ - /* - * For the test harness perform zero byte memory accesses, such that - * in particular correct Disp8 scaling can be verified. - */ - fail_if((modrm_reg & 4) && !ops->write); - - /* Read index register. */ - opc = init_evex(stub); - pevex = copy_EVEX(opc, evex); - pevex->opcx = vex_0f; - /* vmovdqu{32,64} */ - opc[0] = 0x7f; - pevex->pfx = vex_f3; - pevex->w = b & 1; - /* Use (%rax) as destination and sib_index as source. */ - pevex->b = 1; - opc[1] = (state->sib_index & 7) << 3; - pevex->r = !mode_64bit() || !(state->sib_index & 0x08); - pevex->R = !mode_64bit() || !(state->sib_index & 0x10); - pevex->RX = 1; - opc[2] = 0xc3; - - invoke_stub("", "", "=m" (index) : "a" (&index)); - put_stub(stub); - - /* Clear untouched parts of the mask value. */ - n = 1 << (4 - ((b & 1) | evex.w)); - op_mask &= (1 << n) - 1; - - for ( i = 0; rc == X86EMUL_OKAY && op_mask; ++i ) - { - long idx = b & 1 ? index.qw[i] : index.dw[i]; - - if ( !(op_mask & (1 << i)) ) - continue; - - rc = (modrm_reg & 4 - ? ops->write - : ops->read)(ea.mem.seg, - truncate_ea(ea.mem.off + - idx * (1 << state->sib_scale)), - NULL, 0, ctxt); - if ( rc == X86EMUL_EXCEPTION ) - { - /* Squash memory access related exceptions. */ - x86_emul_reset_event(ctxt); - rc = X86EMUL_OKAY; - } - - op_mask &= ~(1 << i); - } - - if ( rc != X86EMUL_OKAY ) - goto done; -#endif - - state->simd_size = simd_none; - break; - } - case X86EMUL_OPC(0x0f38, 0xc8): /* sha1nexte xmm/m128,xmm */ case X86EMUL_OPC(0x0f38, 0xc9): /* sha1msg1 xmm/m128,xmm */ case X86EMUL_OPC(0x0f38, 0xca): /* sha1msg2 xmm/m128,xmm */ @@ -6860,19 +6741,6 @@ x86_emulate( op_bytes = 16; goto simd_0f38_common; - case X86EMUL_OPC_EVEX_66(0x0f38, 0xc8): /* vexp2p{s,d} zmm/mem,zmm{k} */ - case X86EMUL_OPC_EVEX_66(0x0f38, 0xca): /* vrcp28p{s,d} zmm/mem,zmm{k} */ - case X86EMUL_OPC_EVEX_66(0x0f38, 0xcc): /* vrsqrt28p{s,d} zmm/mem,zmm{k} */ - host_and_vcpu_must_have(avx512er); - generate_exception_if((ea.type != OP_REG || !evex.brs) && evex.lr != 2, - X86_EXC_UD); - goto simd_zmm; - - case X86EMUL_OPC_EVEX_66(0x0f38, 0xcb): /* vrcp28s{s,d} xmm/mem,xmm,xmm{k} */ - case X86EMUL_OPC_EVEX_66(0x0f38, 0xcd): /* vrsqrt28s{s,d} xmm/mem,xmm,xmm{k} */ - host_and_vcpu_must_have(avx512er); - goto simd_zmm_scalar_sae; - case X86EMUL_OPC_VEX_F2(0x0f38, 0xcb): /* vsha512rnds2 xmm,ymm,ymm */ case X86EMUL_OPC_VEX_F2(0x0f38, 0xcc): /* vsha512msg1 xmm,ymm */ case X86EMUL_OPC_VEX_F2(0x0f38, 0xcd): /* vsha512msg2 ymm,ymm */ --- a/xen/include/public/arch-x86/cpufeatureset.h +++ b/xen/include/public/arch-x86/cpufeatureset.h @@ -208,8 +208,8 @@ XEN_CPUFEATURE(AVX512_IFMA, 5*32+21) / XEN_CPUFEATURE(CLFLUSHOPT, 5*32+23) /*A CLFLUSHOPT instruction */ XEN_CPUFEATURE(CLWB, 5*32+24) /*!A CLWB instruction */ XEN_CPUFEATURE(PROC_TRACE, 5*32+25) /* Processor Trace */ -XEN_CPUFEATURE(AVX512PF, 5*32+26) /*A AVX-512 Prefetch Instructions */ -XEN_CPUFEATURE(AVX512ER, 5*32+27) /*A AVX-512 Exponent & Reciprocal Instrs */ +XEN_CPUFEATURE(AVX512PF, 5*32+26) /* Xeon Phi AVX-512 Prefetch Instructions */ +XEN_CPUFEATURE(AVX512ER, 5*32+27) /* Xeon Phi AVX-512 Exponent & Reciprocal Instrs */ XEN_CPUFEATURE(AVX512CD, 5*32+28) /*A AVX-512 Conflict Detection Instrs */ XEN_CPUFEATURE(SHA, 5*32+29) /*A SHA1 & SHA256 instructions */ XEN_CPUFEATURE(AVX512BW, 5*32+30) /*A AVX-512 Byte and Word Instructions */ @@ -264,8 +264,8 @@ XEN_CPUFEATURE(BTC_NO, 8*32+29) / XEN_CPUFEATURE(IBPB_RET, 8*32+30) /*A IBPB clears RSB/RAS too. */ /* Intel-defined CPU features, CPUID level 0x00000007:0.edx, word 9 */ -XEN_CPUFEATURE(AVX512_4VNNIW, 9*32+ 2) /*A AVX512 Neural Network Instructions */ -XEN_CPUFEATURE(AVX512_4FMAPS, 9*32+ 3) /*A AVX512 Multiply Accumulation Single Precision */ +XEN_CPUFEATURE(AVX512_4VNNIW, 9*32+ 2) /* Xeon Phi AVX512 Neural Network Instructions */ +XEN_CPUFEATURE(AVX512_4FMAPS, 9*32+ 3) /* Xeon Phi AVX512 Multiply Accumulation Single Precision */ XEN_CPUFEATURE(FSRM, 9*32+ 4) /*A Fast Short REP MOVS */ XEN_CPUFEATURE(UINTR, 9*32+ 5) /* User-mode Interrupts */ XEN_CPUFEATURE(AVX512_VP2INTERSECT, 9*32+8) /*a VP2INTERSECT{D,Q} insns */
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |