### Added
### Removed
+ - On x86:
+ - Support for running on Xeon Phi processors.
## [4.19.0](https://xenbits.xenproject.org/gitweb/?p=xen.git;a=shortlog;h=RELEASE-4.19.0) - 2024-07-29
### x86-64
Status: Supported
+ Status, Xeon Phi: Not supported.
### ARM v7 + Virtualization Extensions
CFLAGS += $(CFLAGS_xeninclude)
-SIMD := 3dnow sse sse2 sse4 avx avx2 xop avx512f avx512bw avx512dq avx512er avx512vbmi avx512fp16
+SIMD := 3dnow sse sse2 sse4 avx avx2 xop avx512f avx512bw avx512dq avx512vbmi avx512fp16
FMA := fma4 fma
SG := avx2-sg avx512f-sg
AES := ssse3-aes avx-aes avx2-vaes avx512bw-vaes
avx512dq-vecs := $(avx512f-vecs)
avx512dq-ints := $(avx512f-ints)
avx512dq-flts := $(avx512f-flts)
-avx512er-vecs := 64
-avx512er-ints :=
-avx512er-flts := 4 8
avx512vbmi-vecs := $(avx512bw-vecs)
avx512vbmi-ints := $(avx512bw-ints)
avx512vbmi-flts := $(avx512bw-flts)
INSN(inserti32x8, 66, 0f3a, 3a, el_8, d, vl),
};
-static const struct test avx512er_512[] = {
- INSN(exp2, 66, 0f38, c8, vl, sd, vl),
- INSN(rcp28, 66, 0f38, ca, vl, sd, vl),
- INSN(rcp28, 66, 0f38, cb, el, sd, el),
- INSN(rsqrt28, 66, 0f38, cc, vl, sd, vl),
- INSN(rsqrt28, 66, 0f38, cd, el, sd, el),
-};
-
-static const struct test avx512pf_512[] = {
- INSNX(gatherpf0d, 66, 0f38, c6, 1, vl, sd, el),
- INSNX(gatherpf0q, 66, 0f38, c7, 1, vl, sd, el),
- INSNX(gatherpf1d, 66, 0f38, c6, 2, vl, sd, el),
- INSNX(gatherpf1q, 66, 0f38, c7, 2, vl, sd, el),
- INSNX(scatterpf0d, 66, 0f38, c6, 5, vl, sd, el),
- INSNX(scatterpf0q, 66, 0f38, c7, 5, vl, sd, el),
- INSNX(scatterpf1d, 66, 0f38, c6, 6, vl, sd, el),
- INSNX(scatterpf1q, 66, 0f38, c7, 6, vl, sd, el),
-};
-
-static const struct test avx512_4fmaps_512[] = {
- INSN(4fmaddps, f2, 0f38, 9a, el_4, d, vl),
- INSN(4fmaddss, f2, 0f38, 9b, el_4, d, vl),
- INSN(4fnmaddps, f2, 0f38, aa, el_4, d, vl),
- INSN(4fnmaddss, f2, 0f38, ab, el_4, d, vl),
-};
-
-static const struct test avx512_4vnniw_512[] = {
- INSN(p4dpwssd, f2, 0f38, 52, el_4, d, vl),
- INSN(p4dpwssds, f2, 0f38, 53, el_4, d, vl),
-};
-
static const struct test avx512_bf16_all[] = {
INSN(cvtne2ps2bf16, f2, 0f38, 72, vl, d, vl),
INSN(cvtneps2bf16, f3, 0f38, 72, vl, d, vl),
RUN(avx512dq, 128);
RUN(avx512dq, no128);
RUN(avx512dq, 512);
- RUN(avx512er, 512);
-#define cpu_has_avx512pf cpu_has_avx512f
- RUN(avx512pf, 512);
- RUN(avx512_4fmaps, 512);
- RUN(avx512_4vnniw, 512);
RUN(avx512_bf16, all);
RUN(avx512_bitalg, all);
RUN(avx512_ifma, all);
#include "avx512bw-vpclmulqdq.h"
#include "avx512bw-gf.h"
#include "avx512dq.h"
-#include "avx512er.h"
#include "avx512vbmi.h"
#include "avx512vbmi2-vpclmulqdq.h"
#include "avx512fp16.h"
return cpu_has_avx512dq && cpu_has_avx512vl;
}
-static bool simd_check_avx512er(void)
-{
- return cpu_has_avx512er;
-}
-
static bool simd_check_avx512bw(void)
{
return cpu_has_avx512bw;
AVX512VL(DQ+VL u64x2, avx512dq, 16u8),
AVX512VL(DQ+VL s64x4, avx512dq, 32i8),
AVX512VL(DQ+VL u64x4, avx512dq, 32u8),
- SIMD(AVX512ER f32 scalar,avx512er, f4),
- SIMD(AVX512ER f32x16, avx512er, 64f4),
- SIMD(AVX512ER f64 scalar,avx512er, f8),
- SIMD(AVX512ER f64x8, avx512er, 64f8),
SIMD(AVX512_VBMI s8x64, avx512vbmi, 64i1),
SIMD(AVX512_VBMI u8x64, avx512vbmi, 64u1),
SIMD(AVX512_VBMI s16x32, avx512vbmi, 64i2),
}
#endif
- printf("%-40s", "Testing v4fmaddps 32(%ecx),%zmm4,%zmm4{%k5}...");
- if ( stack_exec && cpu_has_avx512_4fmaps )
- {
- decl_insn(v4fmaddps);
- static const struct {
- float f[16];
- } in = {{
- 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16
- }}, out = {{
- 1 + 1 * 9 + 2 * 10 + 3 * 11 + 4 * 12,
- 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15,
- 16 + 16 * 9 + 17 * 10 + 18 * 11 + 19 * 12
- }};
-
- asm volatile ( "vmovups %1, %%zmm4\n\t"
- "vbroadcastss %%xmm4, %%zmm7\n\t"
- "vaddps %%zmm4, %%zmm7, %%zmm5\n\t"
- "vaddps %%zmm5, %%zmm7, %%zmm6\n\t"
- "vaddps %%zmm6, %%zmm7, %%zmm7\n\t"
- "kmovw %2, %%k5\n"
- put_insn(v4fmaddps,
- "v4fmaddps 32(%0), %%zmm4, %%zmm4%{%%k5%}")
- :: "c" (NULL), "m" (in), "rmk" (0x8001) );
-
- set_insn(v4fmaddps);
- regs.ecx = (unsigned long)∈
- rc = x86_emulate(&ctxt, &emulops);
- if ( rc != X86EMUL_OKAY || !check_eip(v4fmaddps) )
- goto fail;
-
- asm ( "vcmpeqps %1, %%zmm4, %%k0\n\t"
- "kmovw %%k0, %0" : "=g" (rc) : "m" (out) );
- if ( rc != 0xffff )
- goto fail;
- printf("okay\n");
- }
- else
- printf("skipped\n");
-
- printf("%-40s", "Testing v4fnmaddss 16(%edx),%zmm4,%zmm4{%k3}...");
- if ( stack_exec && cpu_has_avx512_4fmaps )
- {
- decl_insn(v4fnmaddss);
- static const struct {
- float f[16];
- } in = {{
- 1, 2, 3, 4, 5, 6, 7, 8
- }}, out = {{
- 1 - 1 * 5 - 2 * 6 - 3 * 7 - 4 * 8, 2, 3, 4
- }};
-
- asm volatile ( "vmovups %1, %%xmm4\n\t"
- "vaddss %%xmm4, %%xmm4, %%xmm5\n\t"
- "vaddss %%xmm5, %%xmm4, %%xmm6\n\t"
- "vaddss %%xmm6, %%xmm4, %%xmm7\n\t"
- "kmovw %2, %%k3\n"
- put_insn(v4fnmaddss,
- "v4fnmaddss 16(%0), %%xmm4, %%xmm4%{%%k3%}")
- :: "d" (NULL), "m" (in), "rmk" (1) );
-
- set_insn(v4fnmaddss);
- regs.edx = (unsigned long)∈
- rc = x86_emulate(&ctxt, &emulops);
- if ( rc != X86EMUL_OKAY || !check_eip(v4fnmaddss) )
- goto fail;
-
- asm ( "vcmpeqps %1, %%zmm4, %%k0\n\t"
- "kmovw %%k0, %0" : "=g" (rc) : "m" (out) );
- if ( rc != 0xffff )
- goto fail;
- printf("okay\n");
- }
- else
- printf("skipped\n");
-
if ( stack_exec && cpu_has_avx512_bf16 )
{
decl_insn(vcvtne2ps2bf16);
cp.basic.movbe = true;
cp.feat.invpcid = true;
cp.feat.adx = true;
- cp.feat.avx512pf = cp.feat.avx512f;
cp.feat.rdpid = true;
cp.feat.wrmsrns = true;
cp.extd.clzero = true;
#define cpu_has_avx512f (cp.feat.avx512f && xcr0_mask(0xe6))
#define cpu_has_avx512dq (cp.feat.avx512dq && xcr0_mask(0xe6))
#define cpu_has_avx512_ifma (cp.feat.avx512_ifma && xcr0_mask(0xe6))
-#define cpu_has_avx512er (cp.feat.avx512er && xcr0_mask(0xe6))
#define cpu_has_avx512cd (cp.feat.avx512cd && xcr0_mask(0xe6))
#define cpu_has_sha cp.feat.sha
#define cpu_has_avx512bw (cp.feat.avx512bw && xcr0_mask(0xe6))
#define cpu_has_avx512_vpopcntdq (cp.feat.avx512_vpopcntdq && xcr0_mask(0xe6))
#define cpu_has_movdiri cp.feat.movdiri
#define cpu_has_movdir64b cp.feat.movdir64b
-#define cpu_has_avx512_4vnniw (cp.feat.avx512_4vnniw && xcr0_mask(0xe6))
-#define cpu_has_avx512_4fmaps (cp.feat.avx512_4fmaps && xcr0_mask(0xe6))
#define cpu_has_avx512_vp2intersect (cp.feat.avx512_vp2intersect && xcr0_mask(0xe6))
#define cpu_has_serialize cp.feat.serialize
#define cpu_has_avx512_fp16 (cp.feat.avx512_fp16 && xcr0_mask(0xe6))
#define GET_CC3_RES(val) GET_HW_RES_IN_NS(0x3FC, val)
#define GET_CC6_RES(val) GET_HW_RES_IN_NS(0x3FD, val)
#define GET_CC7_RES(val) GET_HW_RES_IN_NS(0x3FE, val) /* SNB onwards */
-#define PHI_CC6_RES(val) GET_HW_RES_IN_NS(0x3FF, val) /* Xeon Phi only */
static void cf_check lapic_timer_nop(void) { }
void (*__read_mostly lapic_timer_off)(void);
GET_CC6_RES(hw_res->cc6);
GET_CC7_RES(hw_res->cc7);
break;
- /* Xeon Phi Knights Landing */
- case 0x57:
- /* Xeon Phi Knights Mill */
- case 0x85:
- GET_CC3_RES(hw_res->mc0); /* abusing GET_CC3_RES */
- GET_CC6_RES(hw_res->mc6); /* abusing GET_CC6_RES */
- GET_PC2_RES(hw_res->pc2);
- GET_PC3_RES(hw_res->pc3);
- GET_PC6_RES(hw_res->pc6);
- GET_PC7_RES(hw_res->pc7);
- PHI_CC6_RES(hw_res->cc6);
- break;
/* various Atoms */
case 0x27:
GET_PC3_RES(hw_res->pc2); /* abusing GET_PC3_RES */
case 0x4f: /* Broadwell X */
case 0x55: /* Skylake X */
case 0x56: /* Broadwell Xeon D */
- case 0x57: /* Knights Landing */
case 0x6a: /* Icelake X */
case 0x6c: /* Icelake D */
- case 0x85: /* Knights Mill */
case 0x8f: /* Sapphire Rapids X */
if ( (c != &boot_cpu_data && !ppin_msr) ||
{}
};
-static const struct cpuidle_state knl_cstates[] = {
- {
- .name = "C1",
- .flags = MWAIT2flg(0x00),
- .exit_latency = 1,
- .target_residency = 2,
- },
- {
- .name = "C6",
- .flags = MWAIT2flg(0x10) | CPUIDLE_FLAG_TLB_FLUSHED,
- .exit_latency = 120,
- .target_residency = 500,
- },
- {}
-};
-
static struct cpuidle_state __read_mostly bxt_cstates[] = {
{
.name = "C1",
.c1e_promotion = C1E_PROMOTION_DISABLE,
};
-static const struct idle_cpu idle_cpu_knl = {
- .state_table = knl_cstates,
-};
-
static const struct idle_cpu idle_cpu_bxt = {
.state_table = bxt_cstates,
.c1e_promotion = C1E_PROMOTION_DISABLE,
ICPU(ALDERLAKE, adl),
ICPU(ALDERLAKE_L, adl_l),
ICPU(SAPPHIRERAPIDS_X, spr),
- ICPU(XEON_PHI_KNL, knl),
- ICPU(XEON_PHI_KNM, knl),
ICPU(ATOM_GOLDMONT, bxt),
ICPU(ATOM_GOLDMONT_PLUS, bxt),
ICPU(ATOM_GOLDMONT_D, dnv),
return at_lbr;
/* Silvermont */
case 0x37: case 0x4a: case 0x4d: case 0x5a: case 0x5d:
- /* Xeon Phi Knights Landing */
- case 0x57:
- /* Xeon Phi Knights Mill */
- case 0x85:
/* Airmont */
case 0x4c:
return sm_lbr;
case 0x7a: /* Gemini Lake */
case 0x86: /* Jacobsville */
- /*
- * Knights processors are not vulnerable.
- */
- case 0x57: /* Knights Landing */
- case 0x85: /* Knights Mill */
- return false;
-
default:
printk("Unrecognised CPU model %#x - assuming vulnerable to IF_PSCHANGE_MC\n",
boot_cpu_data.x86_model);
#define cpu_has_avx512_ifma boot_cpu_has(X86_FEATURE_AVX512_IFMA)
#define cpu_has_clflushopt boot_cpu_has(X86_FEATURE_CLFLUSHOPT)
#define cpu_has_clwb boot_cpu_has(X86_FEATURE_CLWB)
-#define cpu_has_avx512er boot_cpu_has(X86_FEATURE_AVX512ER)
#define cpu_has_avx512cd boot_cpu_has(X86_FEATURE_AVX512CD)
#define cpu_has_proc_trace boot_cpu_has(X86_FEATURE_PROC_TRACE)
#define cpu_has_sha boot_cpu_has(X86_FEATURE_SHA)
#define cpu_has_auto_ibrs boot_cpu_has(X86_FEATURE_AUTO_IBRS)
/* CPUID level 0x00000007:0.edx */
-#define cpu_has_avx512_4vnniw boot_cpu_has(X86_FEATURE_AVX512_4VNNIW)
-#define cpu_has_avx512_4fmaps boot_cpu_has(X86_FEATURE_AVX512_4FMAPS)
#define cpu_has_avx512_vp2intersect boot_cpu_has(X86_FEATURE_AVX512_VP2INTERSECT)
#define cpu_has_srbds_ctrl boot_cpu_has(X86_FEATURE_SRBDS_CTRL)
#define cpu_has_md_clear boot_cpu_has(X86_FEATURE_MD_CLEAR)
* INTEL_FAM6{OPTFAMILY}_{MICROARCH}{OPTDIFF}
* where:
* OPTFAMILY Describes the family of CPUs that this belongs to. Default
- * is assumed to be "_CORE" (and should be omitted). Other values
- * currently in use are _ATOM and _XEON_PHI
+ * is assumed to be "_CORE" (and should be omitted). The other
+ * value currently in use is _ATOM.
* MICROARCH Is the code name for the micro-architecture for this core.
* N.B. Not the platform name.
* OPTDIFF If needed, a short string to differentiate by market segment.
#define INTEL_FAM6_ATOM_DARKMONT_X 0xDD /* Clearwater Forest */
-/* Xeon Phi */
-
-#define INTEL_FAM6_XEON_PHI_KNL 0x57 /* Knights Landing */
-#define INTEL_FAM6_XEON_PHI_KNM 0x85 /* Knights Mill */
-
/* Family 5 */
#define INTEL_FAM5_QUARK_X1000 0x09 /* Quark X1000 SoC */
case 0x4d: /* Avaton / Rangely (Silvermont) */
case 0x4c: /* Cherrytrail / Brasswell */
case 0x4a: /* Merrifield */
- case 0x57: /* Knights Landing */
case 0x5a: /* Moorefield */
case 0x5c: /* Goldmont */
case 0x5f: /* Denverton */
- case 0x85: /* Knights Mill */
safe = true;
break;
case 0x7a: /* Gemini Lake */
return false;
- /*
- * Knights processors are not vulnerable.
- */
- case 0x57: /* Knights Landing */
- case 0x85: /* Knights Mill */
- return false;
-
default:
printk("Unrecognised CPU model %#x - assuming vulnerable to LazyFPU\n",
boot_cpu_data.x86_model);
case 0x7a: /* Gemini Lake */
break;
- /*
- * Knights processors are not vulnerable.
- */
- case 0x57: /* Knights Landing */
- case 0x85: /* Knights Mill */
- break;
-
default:
/* Defer printk() until we've accounted for RDCL_NO. */
hit_default = true;
case 0x65: /* SoFIA LTE AOSP */
case 0x6e: /* Cougar Mountain */
case 0x75: /* Lightning Mountain */
- /*
- * Knights processors (which are based on the Silvermont/Airmont
- * microarchitecture) are similarly only affected by the Store Buffer
- * aspect.
- */
- case 0x57: /* Knights Landing */
- case 0x85: /* Knights Mill */
cpu_has_bug_msbds_only = true;
break;
#define vcpu_has_avx512_ifma() (ctxt->cpuid->feat.avx512_ifma)
#define vcpu_has_clflushopt() (ctxt->cpuid->feat.clflushopt)
#define vcpu_has_clwb() (ctxt->cpuid->feat.clwb)
-#define vcpu_has_avx512pf() (ctxt->cpuid->feat.avx512pf)
-#define vcpu_has_avx512er() (ctxt->cpuid->feat.avx512er)
#define vcpu_has_avx512cd() (ctxt->cpuid->feat.avx512cd)
#define vcpu_has_sha() (ctxt->cpuid->feat.sha)
#define vcpu_has_avx512bw() (ctxt->cpuid->feat.avx512bw)
#define vcpu_has_movdiri() (ctxt->cpuid->feat.movdiri)
#define vcpu_has_movdir64b() (ctxt->cpuid->feat.movdir64b)
#define vcpu_has_enqcmd() (ctxt->cpuid->feat.enqcmd)
-#define vcpu_has_avx512_4vnniw() (ctxt->cpuid->feat.avx512_4vnniw)
-#define vcpu_has_avx512_4fmaps() (ctxt->cpuid->feat.avx512_4fmaps)
#define vcpu_has_avx512_vp2intersect() (ctxt->cpuid->feat.avx512_vp2intersect)
#define vcpu_has_serialize() (ctxt->cpuid->feat.serialize)
#define vcpu_has_tsxldtrk() (ctxt->cpuid->feat.tsxldtrk)
case X86EMUL_OPC_EVEX_66(0x0f38, 0xbd): /* vfnmadd231s{s,d} xmm/mem,xmm,xmm{k} */
case X86EMUL_OPC_EVEX_66(0x0f38, 0xbf): /* vfnmsub231s{s,d} xmm/mem,xmm,xmm{k} */
host_and_vcpu_must_have(avx512f);
- simd_zmm_scalar_sae:
generate_exception_if(ea.type != OP_REG && evex.brs, X86_EXC_UD);
if ( !evex.brs )
avx512_vlen_check(true);
avx512_vlen_check(true);
goto simd_zmm;
- case X86EMUL_OPC_EVEX_F2(0x0f38, 0x52): /* vp4dpwssd m128,zmm+3,zmm{k} */
- case X86EMUL_OPC_EVEX_F2(0x0f38, 0x53): /* vp4dpwssds m128,zmm+3,zmm{k} */
- host_and_vcpu_must_have(avx512_4vnniw);
- generate_exception_if((ea.type != OP_MEM || evex.w || evex.brs ||
- evex.lr != 2),
- X86_EXC_UD);
- op_mask = op_mask & 0xffff ? 0xf : 0;
- goto simd_zmm;
-
case X86EMUL_OPC_EVEX_66(0x0f38, 0x8f): /* vpshufbitqmb [xyz]mm/mem,[xyz]mm,k{k} */
generate_exception_if(evex.w || !evex.r || !evex.R || evex.z, X86_EXC_UD);
/* fall through */
host_and_vcpu_must_have(fma);
goto simd_0f_ymm;
- case X86EMUL_OPC_EVEX_F2(0x0f38, 0x9a): /* v4fmaddps m128,zmm+3,zmm{k} */
- case X86EMUL_OPC_EVEX_F2(0x0f38, 0xaa): /* v4fnmaddps m128,zmm+3,zmm{k} */
- host_and_vcpu_must_have(avx512_4fmaps);
- generate_exception_if((ea.type != OP_MEM || evex.w || evex.brs ||
- evex.lr != 2),
- X86_EXC_UD);
- op_mask = op_mask & 0xffff ? 0xf : 0;
- goto simd_zmm;
-
- case X86EMUL_OPC_EVEX_F2(0x0f38, 0x9b): /* v4fmaddss m128,xmm+3,xmm{k} */
- case X86EMUL_OPC_EVEX_F2(0x0f38, 0xab): /* v4fnmaddss m128,xmm+3,xmm{k} */
- host_and_vcpu_must_have(avx512_4fmaps);
- generate_exception_if((ea.type != OP_MEM || evex.w || evex.brs ||
- evex.lr == 3),
- X86_EXC_UD);
- op_mask = op_mask & 1 ? 0xf : 0;
- goto simd_zmm;
-
case X86EMUL_OPC_EVEX_66(0x0f38, 0xa0): /* vpscatterd{d,q} [xyz]mm,mem{k} */
case X86EMUL_OPC_EVEX_66(0x0f38, 0xa1): /* vpscatterq{d,q} [xyz]mm,mem{k} */
case X86EMUL_OPC_EVEX_66(0x0f38, 0xa2): /* vscatterdp{s,d} [xyz]mm,mem{k} */
generate_exception_if(!evex.w, X86_EXC_UD);
goto avx512f_no_sae;
- case X86EMUL_OPC_EVEX_66(0x0f38, 0xc6):
- case X86EMUL_OPC_EVEX_66(0x0f38, 0xc7):
- {
-#ifndef __XEN__
- typeof(evex) *pevex;
- union {
- int32_t dw[16];
- int64_t qw[8];
- } index;
-#endif
-
- ASSERT(ea.type == OP_MEM);
- generate_exception_if((!cpu_has_avx512f || !evex.opmsk || evex.brs ||
- evex.z || evex.reg != 0xf || evex.lr != 2),
- X86_EXC_UD);
-
- switch ( modrm_reg & 7 )
- {
- case 1: /* vgatherpf0{d,q}p{s,d} mem{k} */
- case 2: /* vgatherpf1{d,q}p{s,d} mem{k} */
- case 5: /* vscatterpf0{d,q}p{s,d} mem{k} */
- case 6: /* vscatterpf1{d,q}p{s,d} mem{k} */
- vcpu_must_have(avx512pf);
- break;
- default:
- generate_exception(X86_EXC_UD);
- }
-
- get_fpu(X86EMUL_FPU_zmm);
-
-#ifndef __XEN__
- /*
- * For the test harness perform zero byte memory accesses, such that
- * in particular correct Disp8 scaling can be verified.
- */
- fail_if((modrm_reg & 4) && !ops->write);
-
- /* Read index register. */
- opc = init_evex(stub);
- pevex = copy_EVEX(opc, evex);
- pevex->opcx = vex_0f;
- /* vmovdqu{32,64} */
- opc[0] = 0x7f;
- pevex->pfx = vex_f3;
- pevex->w = b & 1;
- /* Use (%rax) as destination and sib_index as source. */
- pevex->b = 1;
- opc[1] = (state->sib_index & 7) << 3;
- pevex->r = !mode_64bit() || !(state->sib_index & 0x08);
- pevex->R = !mode_64bit() || !(state->sib_index & 0x10);
- pevex->RX = 1;
- opc[2] = 0xc3;
-
- invoke_stub("", "", "=m" (index) : "a" (&index));
- put_stub(stub);
-
- /* Clear untouched parts of the mask value. */
- n = 1 << (4 - ((b & 1) | evex.w));
- op_mask &= (1 << n) - 1;
-
- for ( i = 0; rc == X86EMUL_OKAY && op_mask; ++i )
- {
- long idx = b & 1 ? index.qw[i] : index.dw[i];
-
- if ( !(op_mask & (1 << i)) )
- continue;
-
- rc = (modrm_reg & 4
- ? ops->write
- : ops->read)(ea.mem.seg,
- truncate_ea(ea.mem.off +
- idx * (1 << state->sib_scale)),
- NULL, 0, ctxt);
- if ( rc == X86EMUL_EXCEPTION )
- {
- /* Squash memory access related exceptions. */
- x86_emul_reset_event(ctxt);
- rc = X86EMUL_OKAY;
- }
-
- op_mask &= ~(1 << i);
- }
-
- if ( rc != X86EMUL_OKAY )
- goto done;
-#endif
-
- state->simd_size = simd_none;
- break;
- }
-
case X86EMUL_OPC(0x0f38, 0xc8): /* sha1nexte xmm/m128,xmm */
case X86EMUL_OPC(0x0f38, 0xc9): /* sha1msg1 xmm/m128,xmm */
case X86EMUL_OPC(0x0f38, 0xca): /* sha1msg2 xmm/m128,xmm */
op_bytes = 16;
goto simd_0f38_common;
- case X86EMUL_OPC_EVEX_66(0x0f38, 0xc8): /* vexp2p{s,d} zmm/mem,zmm{k} */
- case X86EMUL_OPC_EVEX_66(0x0f38, 0xca): /* vrcp28p{s,d} zmm/mem,zmm{k} */
- case X86EMUL_OPC_EVEX_66(0x0f38, 0xcc): /* vrsqrt28p{s,d} zmm/mem,zmm{k} */
- host_and_vcpu_must_have(avx512er);
- generate_exception_if((ea.type != OP_REG || !evex.brs) && evex.lr != 2,
- X86_EXC_UD);
- goto simd_zmm;
-
- case X86EMUL_OPC_EVEX_66(0x0f38, 0xcb): /* vrcp28s{s,d} xmm/mem,xmm,xmm{k} */
- case X86EMUL_OPC_EVEX_66(0x0f38, 0xcd): /* vrsqrt28s{s,d} xmm/mem,xmm,xmm{k} */
- host_and_vcpu_must_have(avx512er);
- goto simd_zmm_scalar_sae;
-
case X86EMUL_OPC_VEX_F2(0x0f38, 0xcb): /* vsha512rnds2 xmm,ymm,ymm */
case X86EMUL_OPC_VEX_F2(0x0f38, 0xcc): /* vsha512msg1 xmm,ymm */
case X86EMUL_OPC_VEX_F2(0x0f38, 0xcd): /* vsha512msg2 ymm,ymm */
XEN_CPUFEATURE(CLFLUSHOPT, 5*32+23) /*A CLFLUSHOPT instruction */
XEN_CPUFEATURE(CLWB, 5*32+24) /*!A CLWB instruction */
XEN_CPUFEATURE(PROC_TRACE, 5*32+25) /* Processor Trace */
-XEN_CPUFEATURE(AVX512PF, 5*32+26) /*A AVX-512 Prefetch Instructions */
-XEN_CPUFEATURE(AVX512ER, 5*32+27) /*A AVX-512 Exponent & Reciprocal Instrs */
+XEN_CPUFEATURE(AVX512PF, 5*32+26) /* Xeon Phi AVX-512 Prefetch Instructions */
+XEN_CPUFEATURE(AVX512ER, 5*32+27) /* Xeon Phi AVX-512 Exponent & Reciprocal Instrs */
XEN_CPUFEATURE(AVX512CD, 5*32+28) /*A AVX-512 Conflict Detection Instrs */
XEN_CPUFEATURE(SHA, 5*32+29) /*A SHA1 & SHA256 instructions */
XEN_CPUFEATURE(AVX512BW, 5*32+30) /*A AVX-512 Byte and Word Instructions */
XEN_CPUFEATURE(IBPB_RET, 8*32+30) /*A IBPB clears RSB/RAS too. */
/* Intel-defined CPU features, CPUID level 0x00000007:0.edx, word 9 */
-XEN_CPUFEATURE(AVX512_4VNNIW, 9*32+ 2) /*A AVX512 Neural Network Instructions */
-XEN_CPUFEATURE(AVX512_4FMAPS, 9*32+ 3) /*A AVX512 Multiply Accumulation Single Precision */
+XEN_CPUFEATURE(AVX512_4VNNIW, 9*32+ 2) /* Xeon Phi AVX512 Neural Network Instructions */
+XEN_CPUFEATURE(AVX512_4FMAPS, 9*32+ 3) /* Xeon Phi AVX512 Multiply Accumulation Single Precision */
XEN_CPUFEATURE(FSRM, 9*32+ 4) /*A Fast Short REP MOVS */
XEN_CPUFEATURE(UINTR, 9*32+ 5) /* User-mode Interrupts */
XEN_CPUFEATURE(AVX512_VP2INTERSECT, 9*32+8) /*a VP2INTERSECT{D,Q} insns */