[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [PATCH] x86: drop Xeon Phi support
On Tue, 2024-07-30 at 13:07 +0200, Jan Beulich wrote: > Do as was decided in Lisbon. Reportedly Xen hasn't been working very > well on those processors anyway. > > Signed-off-by: Jan Beulich <jbeulich@xxxxxxxx> > --- > One file I left untouched is the test harness'es predicates.c: Those > tests are imo fine to retain. Plus of course the dependencies in > gen-cpuid.py also want leaving in place imo (unless we were to remove > the respective 4 lines from the public header). > > --- a/CHANGELOG.md > +++ b/CHANGELOG.md > @@ -11,6 +11,8 @@ The format is based on [Keep a Changelog > ### Added > > ### Removed > + - On x86: > + - Support for running on Xeon Phi processors. Acked-By: Oleksii Kurochko <oleksii.kurochko@xxxxxxxxx> ~ Oleksii > > ## > [4.19.0](https://xenbits.xenproject.org/gitweb/?p=xen.git;a=shortlog; > h=RELEASE-4.19.0) - 2024-07-29 > > --- a/SUPPORT.md > +++ b/SUPPORT.md > @@ -30,6 +30,7 @@ if the related features are marked as su > ### x86-64 > > Status: Supported > + Status, Xeon Phi: Not supported. > > ### ARM v7 + Virtualization Extensions > > --- a/tools/tests/x86_emulator/Makefile > +++ b/tools/tests/x86_emulator/Makefile > @@ -16,7 +16,7 @@ vpath %.c $(XEN_ROOT)/xen/lib/x86 > > CFLAGS += $(CFLAGS_xeninclude) > > -SIMD := 3dnow sse sse2 sse4 avx avx2 xop avx512f avx512bw avx512dq > avx512er avx512vbmi avx512fp16 > +SIMD := 3dnow sse sse2 sse4 avx avx2 xop avx512f avx512bw avx512dq > avx512vbmi avx512fp16 > FMA := fma4 fma > SG := avx2-sg avx512f-sg > AES := ssse3-aes avx-aes avx2-vaes avx512bw-vaes > @@ -80,9 +80,6 @@ avx512bw-flts := > avx512dq-vecs := $(avx512f-vecs) > avx512dq-ints := $(avx512f-ints) > avx512dq-flts := $(avx512f-flts) > -avx512er-vecs := 64 > -avx512er-ints := > -avx512er-flts := 4 8 > avx512vbmi-vecs := $(avx512bw-vecs) > avx512vbmi-ints := $(avx512bw-ints) > avx512vbmi-flts := $(avx512bw-flts) > --- a/tools/tests/x86_emulator/evex-disp8.c > +++ b/tools/tests/x86_emulator/evex-disp8.c > @@ -524,37 +524,6 @@ static const struct test avx512dq_512[] > INSN(inserti32x8, 66, 0f3a, 3a, el_8, d, vl), > }; > > -static const struct test avx512er_512[] = { > - INSN(exp2, 66, 0f38, c8, vl, sd, vl), > - INSN(rcp28, 66, 0f38, ca, vl, sd, vl), > - INSN(rcp28, 66, 0f38, cb, el, sd, el), > - INSN(rsqrt28, 66, 0f38, cc, vl, sd, vl), > - INSN(rsqrt28, 66, 0f38, cd, el, sd, el), > -}; > - > -static const struct test avx512pf_512[] = { > - INSNX(gatherpf0d, 66, 0f38, c6, 1, vl, sd, el), > - INSNX(gatherpf0q, 66, 0f38, c7, 1, vl, sd, el), > - INSNX(gatherpf1d, 66, 0f38, c6, 2, vl, sd, el), > - INSNX(gatherpf1q, 66, 0f38, c7, 2, vl, sd, el), > - INSNX(scatterpf0d, 66, 0f38, c6, 5, vl, sd, el), > - INSNX(scatterpf0q, 66, 0f38, c7, 5, vl, sd, el), > - INSNX(scatterpf1d, 66, 0f38, c6, 6, vl, sd, el), > - INSNX(scatterpf1q, 66, 0f38, c7, 6, vl, sd, el), > -}; > - > -static const struct test avx512_4fmaps_512[] = { > - INSN(4fmaddps, f2, 0f38, 9a, el_4, d, vl), > - INSN(4fmaddss, f2, 0f38, 9b, el_4, d, vl), > - INSN(4fnmaddps, f2, 0f38, aa, el_4, d, vl), > - INSN(4fnmaddss, f2, 0f38, ab, el_4, d, vl), > -}; > - > -static const struct test avx512_4vnniw_512[] = { > - INSN(p4dpwssd, f2, 0f38, 52, el_4, d, vl), > - INSN(p4dpwssds, f2, 0f38, 53, el_4, d, vl), > -}; > - > static const struct test avx512_bf16_all[] = { > INSN(cvtne2ps2bf16, f2, 0f38, 72, vl, d, vl), > INSN(cvtneps2bf16, f3, 0f38, 72, vl, d, vl), > @@ -1110,11 +1079,6 @@ void evex_disp8_test(void *instr, struct > RUN(avx512dq, 128); > RUN(avx512dq, no128); > RUN(avx512dq, 512); > - RUN(avx512er, 512); > -#define cpu_has_avx512pf cpu_has_avx512f > - RUN(avx512pf, 512); > - RUN(avx512_4fmaps, 512); > - RUN(avx512_4vnniw, 512); > RUN(avx512_bf16, all); > RUN(avx512_bitalg, all); > RUN(avx512_ifma, all); > --- a/tools/tests/x86_emulator/test_x86_emulator.c > +++ b/tools/tests/x86_emulator/test_x86_emulator.c > @@ -39,7 +39,6 @@ asm ( ".pushsection .test, \"ax\", @prog > #include "avx512bw-vpclmulqdq.h" > #include "avx512bw-gf.h" > #include "avx512dq.h" > -#include "avx512er.h" > #include "avx512vbmi.h" > #include "avx512vbmi2-vpclmulqdq.h" > #include "avx512fp16.h" > @@ -147,11 +146,6 @@ static bool simd_check_avx512dq_vl(void) > return cpu_has_avx512dq && cpu_has_avx512vl; > } > > -static bool simd_check_avx512er(void) > -{ > - return cpu_has_avx512er; > -} > - > static bool simd_check_avx512bw(void) > { > return cpu_has_avx512bw; > @@ -507,10 +501,6 @@ static const struct { > AVX512VL(DQ+VL u64x2, avx512dq, 16u8), > AVX512VL(DQ+VL s64x4, avx512dq, 32i8), > AVX512VL(DQ+VL u64x4, avx512dq, 32u8), > - SIMD(AVX512ER f32 scalar,avx512er, f4), > - SIMD(AVX512ER f32x16, avx512er, 64f4), > - SIMD(AVX512ER f64 scalar,avx512er, f8), > - SIMD(AVX512ER f64x8, avx512er, 64f8), > SIMD(AVX512_VBMI s8x64, avx512vbmi, 64i1), > SIMD(AVX512_VBMI u8x64, avx512vbmi, 64u1), > SIMD(AVX512_VBMI s16x32, avx512vbmi, 64i2), > @@ -4983,81 +4973,6 @@ int main(int argc, char **argv) > } > #endif > > - printf("%-40s", "Testing v4fmaddps > 32(%ecx),%zmm4,%zmm4{%k5}..."); > - if ( stack_exec && cpu_has_avx512_4fmaps ) > - { > - decl_insn(v4fmaddps); > - static const struct { > - float f[16]; > - } in = {{ > - 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16 > - }}, out = {{ > - 1 + 1 * 9 + 2 * 10 + 3 * 11 + 4 * 12, > - 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, > - 16 + 16 * 9 + 17 * 10 + 18 * 11 + 19 * 12 > - }}; > - > - asm volatile ( "vmovups %1, %%zmm4\n\t" > - "vbroadcastss %%xmm4, %%zmm7\n\t" > - "vaddps %%zmm4, %%zmm7, %%zmm5\n\t" > - "vaddps %%zmm5, %%zmm7, %%zmm6\n\t" > - "vaddps %%zmm6, %%zmm7, %%zmm7\n\t" > - "kmovw %2, %%k5\n" > - put_insn(v4fmaddps, > - "v4fmaddps 32(%0), %%zmm4, > %%zmm4%{%%k5%}") > - :: "c" (NULL), "m" (in), "rmk" (0x8001) ); > - > - set_insn(v4fmaddps); > - regs.ecx = (unsigned long)∈ > - rc = x86_emulate(&ctxt, &emulops); > - if ( rc != X86EMUL_OKAY || !check_eip(v4fmaddps) ) > - goto fail; > - > - asm ( "vcmpeqps %1, %%zmm4, %%k0\n\t" > - "kmovw %%k0, %0" : "=g" (rc) : "m" (out) ); > - if ( rc != 0xffff ) > - goto fail; > - printf("okay\n"); > - } > - else > - printf("skipped\n"); > - > - printf("%-40s", "Testing v4fnmaddss > 16(%edx),%zmm4,%zmm4{%k3}..."); > - if ( stack_exec && cpu_has_avx512_4fmaps ) > - { > - decl_insn(v4fnmaddss); > - static const struct { > - float f[16]; > - } in = {{ > - 1, 2, 3, 4, 5, 6, 7, 8 > - }}, out = {{ > - 1 - 1 * 5 - 2 * 6 - 3 * 7 - 4 * 8, 2, 3, 4 > - }}; > - > - asm volatile ( "vmovups %1, %%xmm4\n\t" > - "vaddss %%xmm4, %%xmm4, %%xmm5\n\t" > - "vaddss %%xmm5, %%xmm4, %%xmm6\n\t" > - "vaddss %%xmm6, %%xmm4, %%xmm7\n\t" > - "kmovw %2, %%k3\n" > - put_insn(v4fnmaddss, > - "v4fnmaddss 16(%0), %%xmm4, > %%xmm4%{%%k3%}") > - :: "d" (NULL), "m" (in), "rmk" (1) ); > - > - set_insn(v4fnmaddss); > - regs.edx = (unsigned long)∈ > - rc = x86_emulate(&ctxt, &emulops); > - if ( rc != X86EMUL_OKAY || !check_eip(v4fnmaddss) ) > - goto fail; > - > - asm ( "vcmpeqps %1, %%zmm4, %%k0\n\t" > - "kmovw %%k0, %0" : "=g" (rc) : "m" (out) ); > - if ( rc != 0xffff ) > - goto fail; > - printf("okay\n"); > - } > - else > - printf("skipped\n"); > - > if ( stack_exec && cpu_has_avx512_bf16 ) > { > decl_insn(vcvtne2ps2bf16); > --- a/tools/tests/x86_emulator/x86-emulate.c > +++ b/tools/tests/x86_emulator/x86-emulate.c > @@ -84,7 +84,6 @@ bool emul_test_init(void) > cp.basic.movbe = true; > cp.feat.invpcid = true; > cp.feat.adx = true; > - cp.feat.avx512pf = cp.feat.avx512f; > cp.feat.rdpid = true; > cp.feat.wrmsrns = true; > cp.extd.clzero = true; > --- a/tools/tests/x86_emulator/x86-emulate.h > +++ b/tools/tests/x86_emulator/x86-emulate.h > @@ -156,7 +156,6 @@ void wrpkru(unsigned int val); > #define cpu_has_avx512f (cp.feat.avx512f && xcr0_mask(0xe6)) > #define cpu_has_avx512dq (cp.feat.avx512dq && xcr0_mask(0xe6)) > #define cpu_has_avx512_ifma (cp.feat.avx512_ifma && xcr0_mask(0xe6)) > -#define cpu_has_avx512er (cp.feat.avx512er && xcr0_mask(0xe6)) > #define cpu_has_avx512cd (cp.feat.avx512cd && xcr0_mask(0xe6)) > #define cpu_has_sha cp.feat.sha > #define cpu_has_avx512bw (cp.feat.avx512bw && xcr0_mask(0xe6)) > @@ -171,8 +170,6 @@ void wrpkru(unsigned int val); > #define cpu_has_avx512_vpopcntdq (cp.feat.avx512_vpopcntdq && > xcr0_mask(0xe6)) > #define cpu_has_movdiri cp.feat.movdiri > #define cpu_has_movdir64b cp.feat.movdir64b > -#define cpu_has_avx512_4vnniw (cp.feat.avx512_4vnniw && > xcr0_mask(0xe6)) > -#define cpu_has_avx512_4fmaps (cp.feat.avx512_4fmaps && > xcr0_mask(0xe6)) > #define cpu_has_avx512_vp2intersect (cp.feat.avx512_vp2intersect && > xcr0_mask(0xe6)) > #define cpu_has_serialize cp.feat.serialize > #define cpu_has_avx512_fp16 (cp.feat.avx512_fp16 && xcr0_mask(0xe6)) > --- a/xen/arch/x86/acpi/cpu_idle.c > +++ b/xen/arch/x86/acpi/cpu_idle.c > @@ -73,7 +73,6 @@ > #define GET_CC3_RES(val) GET_HW_RES_IN_NS(0x3FC, val) > #define GET_CC6_RES(val) GET_HW_RES_IN_NS(0x3FD, val) > #define GET_CC7_RES(val) GET_HW_RES_IN_NS(0x3FE, val) /* SNB > onwards */ > -#define PHI_CC6_RES(val) GET_HW_RES_IN_NS(0x3FF, val) /* Xeon Phi > only */ > > static void cf_check lapic_timer_nop(void) { } > void (*__read_mostly lapic_timer_off)(void); > @@ -222,18 +221,6 @@ static void cf_check do_get_hw_residenci > GET_CC6_RES(hw_res->cc6); > GET_CC7_RES(hw_res->cc7); > break; > - /* Xeon Phi Knights Landing */ > - case 0x57: > - /* Xeon Phi Knights Mill */ > - case 0x85: > - GET_CC3_RES(hw_res->mc0); /* abusing GET_CC3_RES */ > - GET_CC6_RES(hw_res->mc6); /* abusing GET_CC6_RES */ > - GET_PC2_RES(hw_res->pc2); > - GET_PC3_RES(hw_res->pc3); > - GET_PC6_RES(hw_res->pc6); > - GET_PC7_RES(hw_res->pc7); > - PHI_CC6_RES(hw_res->cc6); > - break; > /* various Atoms */ > case 0x27: > GET_PC3_RES(hw_res->pc2); /* abusing GET_PC3_RES */ > --- a/xen/arch/x86/cpu/mcheck/mce_intel.c > +++ b/xen/arch/x86/cpu/mcheck/mce_intel.c > @@ -875,10 +875,8 @@ static void intel_init_ppin(const struct > case 0x4f: /* Broadwell X */ > case 0x55: /* Skylake X */ > case 0x56: /* Broadwell Xeon D */ > - case 0x57: /* Knights Landing */ > case 0x6a: /* Icelake X */ > case 0x6c: /* Icelake D */ > - case 0x85: /* Knights Mill */ > case 0x8f: /* Sapphire Rapids X */ > > if ( (c != &boot_cpu_data && !ppin_msr) || > --- a/xen/arch/x86/cpu/mwait-idle.c > +++ b/xen/arch/x86/cpu/mwait-idle.c > @@ -778,22 +778,6 @@ static const struct cpuidle_state avn_cs > {} > }; > > -static const struct cpuidle_state knl_cstates[] = { > - { > - .name = "C1", > - .flags = MWAIT2flg(0x00), > - .exit_latency = 1, > - .target_residency = 2, > - }, > - { > - .name = "C6", > - .flags = MWAIT2flg(0x10) | CPUIDLE_FLAG_TLB_FLUSHED, > - .exit_latency = 120, > - .target_residency = 500, > - }, > - {} > -}; > - > static struct cpuidle_state __read_mostly bxt_cstates[] = { > { > .name = "C1", > @@ -1121,10 +1105,6 @@ static const struct idle_cpu idle_cpu_av > .c1e_promotion = C1E_PROMOTION_DISABLE, > }; > > -static const struct idle_cpu idle_cpu_knl = { > - .state_table = knl_cstates, > -}; > - > static const struct idle_cpu idle_cpu_bxt = { > .state_table = bxt_cstates, > .c1e_promotion = C1E_PROMOTION_DISABLE, > @@ -1181,8 +1161,6 @@ static const struct x86_cpu_id intel_idl > ICPU(ALDERLAKE, adl), > ICPU(ALDERLAKE_L, adl_l), > ICPU(SAPPHIRERAPIDS_X, spr), > - ICPU(XEON_PHI_KNL, knl), > - ICPU(XEON_PHI_KNM, knl), > ICPU(ATOM_GOLDMONT, bxt), > ICPU(ATOM_GOLDMONT_PLUS, bxt), > ICPU(ATOM_GOLDMONT_D, dnv), > --- a/xen/arch/x86/hvm/vmx/vmx.c > +++ b/xen/arch/x86/hvm/vmx/vmx.c > @@ -533,10 +533,6 @@ static const struct lbr_info *__init get > return at_lbr; > /* Silvermont */ > case 0x37: case 0x4a: case 0x4d: case 0x5a: case 0x5d: > - /* Xeon Phi Knights Landing */ > - case 0x57: > - /* Xeon Phi Knights Mill */ > - case 0x85: > /* Airmont */ > case 0x4c: > return sm_lbr; > @@ -2984,13 +2980,6 @@ static bool __init has_if_pschange_mc(vo > case 0x7a: /* Gemini Lake */ > case 0x86: /* Jacobsville */ > > - /* > - * Knights processors are not vulnerable. > - */ > - case 0x57: /* Knights Landing */ > - case 0x85: /* Knights Mill */ > - return false; > - > default: > printk("Unrecognised CPU model %#x - assuming vulnerable to > IF_PSCHANGE_MC\n", > boot_cpu_data.x86_model); > --- a/xen/arch/x86/include/asm/cpufeature.h > +++ b/xen/arch/x86/include/asm/cpufeature.h > @@ -140,7 +140,6 @@ static inline bool boot_cpu_has(unsigned > #define cpu_has_avx512_ifma > boot_cpu_has(X86_FEATURE_AVX512_IFMA) > #define cpu_has_clflushopt boot_cpu_has(X86_FEATURE_CLFLUSHOPT) > #define cpu_has_clwb boot_cpu_has(X86_FEATURE_CLWB) > -#define cpu_has_avx512er boot_cpu_has(X86_FEATURE_AVX512ER) > #define cpu_has_avx512cd boot_cpu_has(X86_FEATURE_AVX512CD) > #define cpu_has_proc_trace boot_cpu_has(X86_FEATURE_PROC_TRACE) > #define cpu_has_sha boot_cpu_has(X86_FEATURE_SHA) > @@ -174,8 +173,6 @@ static inline bool boot_cpu_has(unsigned > #define cpu_has_auto_ibrs boot_cpu_has(X86_FEATURE_AUTO_IBRS) > > /* CPUID level 0x00000007:0.edx */ > -#define cpu_has_avx512_4vnniw > boot_cpu_has(X86_FEATURE_AVX512_4VNNIW) > -#define cpu_has_avx512_4fmaps > boot_cpu_has(X86_FEATURE_AVX512_4FMAPS) > #define cpu_has_avx512_vp2intersect > boot_cpu_has(X86_FEATURE_AVX512_VP2INTERSECT) > #define cpu_has_srbds_ctrl boot_cpu_has(X86_FEATURE_SRBDS_CTRL) > #define cpu_has_md_clear boot_cpu_has(X86_FEATURE_MD_CLEAR) > --- a/xen/arch/x86/include/asm/intel-family.h > +++ b/xen/arch/x86/include/asm/intel-family.h > @@ -13,8 +13,8 @@ > * INTEL_FAM6{OPTFAMILY}_{MICROARCH}{OPTDIFF} > * where: > * OPTFAMILY Describes the family of CPUs that this belongs to. > Default > - * is assumed to be "_CORE" (and should be omitted). > Other values > - * currently in use are _ATOM and _XEON_PHI > + * is assumed to be "_CORE" (and should be omitted). > The other > + * value currently in use is _ATOM. > * MICROARCH Is the code name for the micro-architecture for this > core. > * N.B. Not the platform name. > * OPTDIFF If needed, a short string to differentiate by market > segment. > @@ -164,11 +164,6 @@ > > #define INTEL_FAM6_ATOM_DARKMONT_X 0xDD /* Clearwater Forest */ > > -/* Xeon Phi */ > - > -#define INTEL_FAM6_XEON_PHI_KNL 0x57 /* Knights > Landing */ > -#define INTEL_FAM6_XEON_PHI_KNM 0x85 /* Knights Mill > */ > - > /* Family 5 */ > #define INTEL_FAM5_QUARK_X1000 0x09 /* Quark X1000 SoC */ > > --- a/xen/arch/x86/spec_ctrl.c > +++ b/xen/arch/x86/spec_ctrl.c > @@ -887,11 +887,9 @@ static bool __init retpoline_calculation > case 0x4d: /* Avaton / Rangely (Silvermont) */ > case 0x4c: /* Cherrytrail / Brasswell */ > case 0x4a: /* Merrifield */ > - case 0x57: /* Knights Landing */ > case 0x5a: /* Moorefield */ > case 0x5c: /* Goldmont */ > case 0x5f: /* Denverton */ > - case 0x85: /* Knights Mill */ > safe = true; > break; > > @@ -1005,13 +1003,6 @@ static bool __init should_use_eager_fpu( > case 0x7a: /* Gemini Lake */ > return false; > > - /* > - * Knights processors are not vulnerable. > - */ > - case 0x57: /* Knights Landing */ > - case 0x85: /* Knights Mill */ > - return false; > - > default: > printk("Unrecognised CPU model %#x - assuming vulnerable to > LazyFPU\n", > boot_cpu_data.x86_model); > @@ -1259,13 +1250,6 @@ static __init void l1tf_calculations(voi > case 0x7a: /* Gemini Lake */ > break; > > - /* > - * Knights processors are not vulnerable. > - */ > - case 0x57: /* Knights Landing */ > - case 0x85: /* Knights Mill */ > - break; > - > default: > /* Defer printk() until we've accounted for RDCL_NO. */ > hit_default = true; > @@ -1416,13 +1400,6 @@ static __init void mds_calculations(void > case 0x65: /* SoFIA LTE AOSP */ > case 0x6e: /* Cougar Mountain */ > case 0x75: /* Lightning Mountain */ > - /* > - * Knights processors (which are based on the > Silvermont/Airmont > - * microarchitecture) are similarly only affected by the > Store Buffer > - * aspect. > - */ > - case 0x57: /* Knights Landing */ > - case 0x85: /* Knights Mill */ > cpu_has_bug_msbds_only = true; > break; > > --- a/xen/arch/x86/x86_emulate/private.h > +++ b/xen/arch/x86/x86_emulate/private.h > @@ -564,8 +564,6 @@ amd_like(const struct x86_emulate_ctxt * > #define vcpu_has_avx512_ifma() (ctxt->cpuid->feat.avx512_ifma) > #define vcpu_has_clflushopt() (ctxt->cpuid->feat.clflushopt) > #define vcpu_has_clwb() (ctxt->cpuid->feat.clwb) > -#define vcpu_has_avx512pf() (ctxt->cpuid->feat.avx512pf) > -#define vcpu_has_avx512er() (ctxt->cpuid->feat.avx512er) > #define vcpu_has_avx512cd() (ctxt->cpuid->feat.avx512cd) > #define vcpu_has_sha() (ctxt->cpuid->feat.sha) > #define vcpu_has_avx512bw() (ctxt->cpuid->feat.avx512bw) > @@ -582,8 +580,6 @@ amd_like(const struct x86_emulate_ctxt * > #define vcpu_has_movdiri() (ctxt->cpuid->feat.movdiri) > #define vcpu_has_movdir64b() (ctxt->cpuid->feat.movdir64b) > #define vcpu_has_enqcmd() (ctxt->cpuid->feat.enqcmd) > -#define vcpu_has_avx512_4vnniw() (ctxt->cpuid->feat.avx512_4vnniw) > -#define vcpu_has_avx512_4fmaps() (ctxt->cpuid->feat.avx512_4fmaps) > #define vcpu_has_avx512_vp2intersect() (ctxt->cpuid- > >feat.avx512_vp2intersect) > #define vcpu_has_serialize() (ctxt->cpuid->feat.serialize) > #define vcpu_has_tsxldtrk() (ctxt->cpuid->feat.tsxldtrk) > --- a/xen/arch/x86/x86_emulate/x86_emulate.c > +++ b/xen/arch/x86/x86_emulate/x86_emulate.c > @@ -6029,7 +6029,6 @@ x86_emulate( > case X86EMUL_OPC_EVEX_66(0x0f38, 0xbd): /* vfnmadd231s{s,d} > xmm/mem,xmm,xmm{k} */ > case X86EMUL_OPC_EVEX_66(0x0f38, 0xbf): /* vfnmsub231s{s,d} > xmm/mem,xmm,xmm{k} */ > host_and_vcpu_must_have(avx512f); > - simd_zmm_scalar_sae: > generate_exception_if(ea.type != OP_REG && evex.brs, > X86_EXC_UD); > if ( !evex.brs ) > avx512_vlen_check(true); > @@ -6123,15 +6122,6 @@ x86_emulate( > avx512_vlen_check(true); > goto simd_zmm; > > - case X86EMUL_OPC_EVEX_F2(0x0f38, 0x52): /* vp4dpwssd > m128,zmm+3,zmm{k} */ > - case X86EMUL_OPC_EVEX_F2(0x0f38, 0x53): /* vp4dpwssds > m128,zmm+3,zmm{k} */ > - host_and_vcpu_must_have(avx512_4vnniw); > - generate_exception_if((ea.type != OP_MEM || evex.w || > evex.brs || > - evex.lr != 2), > - X86_EXC_UD); > - op_mask = op_mask & 0xffff ? 0xf : 0; > - goto simd_zmm; > - > case X86EMUL_OPC_EVEX_66(0x0f38, 0x8f): /* vpshufbitqmb > [xyz]mm/mem,[xyz]mm,k{k} */ > generate_exception_if(evex.w || !evex.r || !evex.R || > evex.z, X86_EXC_UD); > /* fall through */ > @@ -6601,24 +6591,6 @@ x86_emulate( > host_and_vcpu_must_have(fma); > goto simd_0f_ymm; > > - case X86EMUL_OPC_EVEX_F2(0x0f38, 0x9a): /* v4fmaddps > m128,zmm+3,zmm{k} */ > - case X86EMUL_OPC_EVEX_F2(0x0f38, 0xaa): /* v4fnmaddps > m128,zmm+3,zmm{k} */ > - host_and_vcpu_must_have(avx512_4fmaps); > - generate_exception_if((ea.type != OP_MEM || evex.w || > evex.brs || > - evex.lr != 2), > - X86_EXC_UD); > - op_mask = op_mask & 0xffff ? 0xf : 0; > - goto simd_zmm; > - > - case X86EMUL_OPC_EVEX_F2(0x0f38, 0x9b): /* v4fmaddss > m128,xmm+3,xmm{k} */ > - case X86EMUL_OPC_EVEX_F2(0x0f38, 0xab): /* v4fnmaddss > m128,xmm+3,xmm{k} */ > - host_and_vcpu_must_have(avx512_4fmaps); > - generate_exception_if((ea.type != OP_MEM || evex.w || > evex.brs || > - evex.lr == 3), > - X86_EXC_UD); > - op_mask = op_mask & 1 ? 0xf : 0; > - goto simd_zmm; > - > case X86EMUL_OPC_EVEX_66(0x0f38, 0xa0): /* vpscatterd{d,q} > [xyz]mm,mem{k} */ > case X86EMUL_OPC_EVEX_66(0x0f38, 0xa1): /* vpscatterq{d,q} > [xyz]mm,mem{k} */ > case X86EMUL_OPC_EVEX_66(0x0f38, 0xa2): /* vscatterdp{s,d} > [xyz]mm,mem{k} */ > @@ -6759,97 +6731,6 @@ x86_emulate( > generate_exception_if(!evex.w, X86_EXC_UD); > goto avx512f_no_sae; > > - case X86EMUL_OPC_EVEX_66(0x0f38, 0xc6): > - case X86EMUL_OPC_EVEX_66(0x0f38, 0xc7): > - { > -#ifndef __XEN__ > - typeof(evex) *pevex; > - union { > - int32_t dw[16]; > - int64_t qw[8]; > - } index; > -#endif > - > - ASSERT(ea.type == OP_MEM); > - generate_exception_if((!cpu_has_avx512f || !evex.opmsk || > evex.brs || > - evex.z || evex.reg != 0xf || evex.lr > != 2), > - X86_EXC_UD); > - > - switch ( modrm_reg & 7 ) > - { > - case 1: /* vgatherpf0{d,q}p{s,d} mem{k} */ > - case 2: /* vgatherpf1{d,q}p{s,d} mem{k} */ > - case 5: /* vscatterpf0{d,q}p{s,d} mem{k} */ > - case 6: /* vscatterpf1{d,q}p{s,d} mem{k} */ > - vcpu_must_have(avx512pf); > - break; > - default: > - generate_exception(X86_EXC_UD); > - } > - > - get_fpu(X86EMUL_FPU_zmm); > - > -#ifndef __XEN__ > - /* > - * For the test harness perform zero byte memory accesses, > such that > - * in particular correct Disp8 scaling can be verified. > - */ > - fail_if((modrm_reg & 4) && !ops->write); > - > - /* Read index register. */ > - opc = init_evex(stub); > - pevex = copy_EVEX(opc, evex); > - pevex->opcx = vex_0f; > - /* vmovdqu{32,64} */ > - opc[0] = 0x7f; > - pevex->pfx = vex_f3; > - pevex->w = b & 1; > - /* Use (%rax) as destination and sib_index as source. */ > - pevex->b = 1; > - opc[1] = (state->sib_index & 7) << 3; > - pevex->r = !mode_64bit() || !(state->sib_index & 0x08); > - pevex->R = !mode_64bit() || !(state->sib_index & 0x10); > - pevex->RX = 1; > - opc[2] = 0xc3; > - > - invoke_stub("", "", "=m" (index) : "a" (&index)); > - put_stub(stub); > - > - /* Clear untouched parts of the mask value. */ > - n = 1 << (4 - ((b & 1) | evex.w)); > - op_mask &= (1 << n) - 1; > - > - for ( i = 0; rc == X86EMUL_OKAY && op_mask; ++i ) > - { > - long idx = b & 1 ? index.qw[i] : index.dw[i]; > - > - if ( !(op_mask & (1 << i)) ) > - continue; > - > - rc = (modrm_reg & 4 > - ? ops->write > - : ops->read)(ea.mem.seg, > - truncate_ea(ea.mem.off + > - idx * (1 << state- > >sib_scale)), > - NULL, 0, ctxt); > - if ( rc == X86EMUL_EXCEPTION ) > - { > - /* Squash memory access related exceptions. */ > - x86_emul_reset_event(ctxt); > - rc = X86EMUL_OKAY; > - } > - > - op_mask &= ~(1 << i); > - } > - > - if ( rc != X86EMUL_OKAY ) > - goto done; > -#endif > - > - state->simd_size = simd_none; > - break; > - } > - > case X86EMUL_OPC(0x0f38, 0xc8): /* sha1nexte xmm/m128,xmm */ > case X86EMUL_OPC(0x0f38, 0xc9): /* sha1msg1 xmm/m128,xmm */ > case X86EMUL_OPC(0x0f38, 0xca): /* sha1msg2 xmm/m128,xmm */ > @@ -6860,19 +6741,6 @@ x86_emulate( > op_bytes = 16; > goto simd_0f38_common; > > - case X86EMUL_OPC_EVEX_66(0x0f38, 0xc8): /* vexp2p{s,d} > zmm/mem,zmm{k} */ > - case X86EMUL_OPC_EVEX_66(0x0f38, 0xca): /* vrcp28p{s,d} > zmm/mem,zmm{k} */ > - case X86EMUL_OPC_EVEX_66(0x0f38, 0xcc): /* vrsqrt28p{s,d} > zmm/mem,zmm{k} */ > - host_and_vcpu_must_have(avx512er); > - generate_exception_if((ea.type != OP_REG || !evex.brs) && > evex.lr != 2, > - X86_EXC_UD); > - goto simd_zmm; > - > - case X86EMUL_OPC_EVEX_66(0x0f38, 0xcb): /* vrcp28s{s,d} > xmm/mem,xmm,xmm{k} */ > - case X86EMUL_OPC_EVEX_66(0x0f38, 0xcd): /* vrsqrt28s{s,d} > xmm/mem,xmm,xmm{k} */ > - host_and_vcpu_must_have(avx512er); > - goto simd_zmm_scalar_sae; > - > case X86EMUL_OPC_VEX_F2(0x0f38, 0xcb): /* vsha512rnds2 > xmm,ymm,ymm */ > case X86EMUL_OPC_VEX_F2(0x0f38, 0xcc): /* vsha512msg1 xmm,ymm */ > case X86EMUL_OPC_VEX_F2(0x0f38, 0xcd): /* vsha512msg2 ymm,ymm */ > --- a/xen/include/public/arch-x86/cpufeatureset.h > +++ b/xen/include/public/arch-x86/cpufeatureset.h > @@ -208,8 +208,8 @@ XEN_CPUFEATURE(AVX512_IFMA, 5*32+21) / > XEN_CPUFEATURE(CLFLUSHOPT, 5*32+23) /*A CLFLUSHOPT instruction > */ > XEN_CPUFEATURE(CLWB, 5*32+24) /*!A CLWB instruction */ > XEN_CPUFEATURE(PROC_TRACE, 5*32+25) /* Processor Trace */ > -XEN_CPUFEATURE(AVX512PF, 5*32+26) /*A AVX-512 Prefetch > Instructions */ > -XEN_CPUFEATURE(AVX512ER, 5*32+27) /*A AVX-512 Exponent & > Reciprocal Instrs */ > +XEN_CPUFEATURE(AVX512PF, 5*32+26) /* Xeon Phi AVX-512 > Prefetch Instructions */ > +XEN_CPUFEATURE(AVX512ER, 5*32+27) /* Xeon Phi AVX-512 > Exponent & Reciprocal Instrs */ > XEN_CPUFEATURE(AVX512CD, 5*32+28) /*A AVX-512 Conflict > Detection Instrs */ > XEN_CPUFEATURE(SHA, 5*32+29) /*A SHA1 & SHA256 > instructions */ > XEN_CPUFEATURE(AVX512BW, 5*32+30) /*A AVX-512 Byte and Word > Instructions */ > @@ -264,8 +264,8 @@ XEN_CPUFEATURE(BTC_NO, 8*32+29) / > XEN_CPUFEATURE(IBPB_RET, 8*32+30) /*A IBPB clears RSB/RAS too. > */ > > /* Intel-defined CPU features, CPUID level 0x00000007:0.edx, word 9 > */ > -XEN_CPUFEATURE(AVX512_4VNNIW, 9*32+ 2) /*A AVX512 Neural Network > Instructions */ > -XEN_CPUFEATURE(AVX512_4FMAPS, 9*32+ 3) /*A AVX512 Multiply > Accumulation Single Precision */ > +XEN_CPUFEATURE(AVX512_4VNNIW, 9*32+ 2) /* Xeon Phi AVX512 Neural > Network Instructions */ > +XEN_CPUFEATURE(AVX512_4FMAPS, 9*32+ 3) /* Xeon Phi AVX512 Multiply > Accumulation Single Precision */ > XEN_CPUFEATURE(FSRM, 9*32+ 4) /*A Fast Short REP MOVS */ > XEN_CPUFEATURE(UINTR, 9*32+ 5) /* User-mode Interrupts */ > XEN_CPUFEATURE(AVX512_VP2INTERSECT, 9*32+8) /*a VP2INTERSECT{D,Q} > insns */
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |