[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH v4 06/44] x86emul: test for correct EVEX Disp8 scaling
Besides the already existing tests (which are going to be extended once respective ISA extension support is complete), let's also ensure for every individual insn that their Disp8 scaling (and memory access width) are correct. Signed-off-by: Jan Beulich <jbeulich@xxxxxxxx> --- v4: Introduce ESZ_d_WIG. v3: New. --- a/tools/tests/x86_emulator/Makefile +++ b/tools/tests/x86_emulator/Makefile @@ -139,7 +139,7 @@ $(addsuffix .h,$(SIMD) $(FMA) $(SG)): si xop.h: simd-fma.c -$(TARGET): x86-emulate.o test_x86_emulator.o wrappers.o +$(TARGET): x86-emulate.o test_x86_emulator.o evex-disp8.o wrappers.o $(HOSTCC) $(HOSTCFLAGS) -o $@ $^ .PHONY: clean @@ -166,7 +166,7 @@ x86.h := $(addprefix $(XEN_ROOT)/tools/i x86-vendors.h x86-defns.h msr-index.h) x86_emulate.h := x86-emulate.h x86_emulate/x86_emulate.h $(x86.h) -x86-emulate.o test_x86_emulator.o wrappers.o: %.o: %.c $(x86_emulate.h) +x86-emulate.o test_x86_emulator.o evex-disp8.o wrappers.o: %.o: %.c $(x86_emulate.h) $(HOSTCC) $(HOSTCFLAGS) -c -g -o $@ $< x86-emulate.o: x86_emulate/x86_emulate.c --- /dev/null +++ b/tools/tests/x86_emulator/evex-disp8.c @@ -0,0 +1,452 @@ +#include <stdarg.h> +#include <stdio.h> + +#include "x86-emulate.h" + +struct test { + const char *mnemonic; + unsigned int opc:8; + unsigned int spc:2; + unsigned int pfx:2; + unsigned int vsz:3; + unsigned int esz:4; + unsigned int scale:1; + unsigned int ext:3; +}; + +enum spc { + SPC_invalid, + SPC_0f, + SPC_0f38, + SPC_0f3a, +}; + +enum pfx { + PFX_, + PFX_66, + PFX_f3, + PFX_f2 +}; + +enum vl { + VL_128, + VL_256, + VL_512, +}; + +enum scale { + SC_vl, + SC_el, +}; + +enum vsz { + VSZ_vl, + VSZ_vl_2, /* VL / 2 */ + VSZ_vl_4, /* VL / 4 */ + VSZ_vl_8, /* VL / 8 */ + /* "no broadcast" implied from here on. */ + VSZ_el, + VSZ_el_2, /* EL * 2 */ + VSZ_el_4, /* EL * 4 */ + VSZ_el_8, /* EL * 8 */ +}; + +enum esz { + ESZ_d, + ESZ_q, + ESZ_dq, + ESZ_sd, + ESZ_d_nb, + ESZ_q_nb, + /* "no broadcast" implied from here on. */ +#ifdef __i386__ + ESZ_d_WIG, +#endif + ESZ_b, + ESZ_w, + ESZ_bw, +}; + +#ifndef __i386__ +# define ESZ_dq64 ESZ_dq +#else +# define ESZ_dq64 ESZ_d_WIG +#endif + +#define INSNX(m, p, sp, o, e, vs, es, sc) { \ + .mnemonic = #m, .opc = 0x##o, .spc = SPC_##sp, .pfx = PFX_##p, \ + .vsz = VSZ_##vs, .esz = ESZ_##es, .scale = SC_##sc, .ext = 0##e \ +} +#define INSN(m, p, sp, o, vs, es, sc) INSNX(m, p, sp, o, 0, vs, es, sc) +#define INSN_PFP(m, sp, o) \ + INSN(m##pd, 66, sp, o, vl, q, vl), \ + INSN(m##ps, , sp, o, vl, d, vl) +#define INSN_PFP_NB(m, sp, o) \ + INSN(m##pd, 66, sp, o, vl, q_nb, vl), \ + INSN(m##ps, , sp, o, vl, d_nb, vl) +#define INSN_SFP(m, sp, o) \ + INSN(m##sd, f2, sp, o, el, q, el), \ + INSN(m##ss, f3, sp, o, el, d, el) + +#define INSN_FP(m, sp, o) \ + INSN_PFP(m, sp, o), \ + INSN_SFP(m, sp, o) + +static const struct test avx512f_all[] = { + INSN_SFP(mov, 0f, 10), + INSN_SFP(mov, 0f, 11), + INSN_PFP_NB(mova, 0f, 28), + INSN_PFP_NB(mova, 0f, 29), + INSN(movdqa32, 66, 0f, 6f, vl, d_nb, vl), + INSN(movdqa32, 66, 0f, 7f, vl, d_nb, vl), + INSN(movdqa64, 66, 0f, 6f, vl, q_nb, vl), + INSN(movdqa64, 66, 0f, 7f, vl, q_nb, vl), + INSN(movdqu32, f3, 0f, 6f, vl, d_nb, vl), + INSN(movdqu32, f3, 0f, 7f, vl, d_nb, vl), + INSN(movdqu64, f3, 0f, 6f, vl, q_nb, vl), + INSN(movdqu64, f3, 0f, 7f, vl, q_nb, vl), + INSN(movntdq, 66, 0f, e7, vl, d_nb, vl), + INSN(movntdqa, 66, 0f38, 2a, vl, d_nb, vl), + INSN_PFP_NB(movnt, 0f, 2b), + INSN_PFP_NB(movu, 0f, 10), + INSN_PFP_NB(movu, 0f, 11), +}; + +static const struct test avx512f_128[] = { + INSN(mov, 66, 0f, 6e, el, dq64, el), + INSN(mov, 66, 0f, 7e, el, dq64, el), + INSN(movq, f3, 0f, 7e, el, q, el), + INSN(movq, 66, 0f, d6, el, q, el), +}; + +static const struct test avx512bw_all[] = { + INSN(movdqu8, f2, 0f, 6f, vl, b, vl), + INSN(movdqu8, f2, 0f, 7f, vl, b, vl), + INSN(movdqu16, f2, 0f, 6f, vl, w, vl), + INSN(movdqu16, f2, 0f, 7f, vl, w, vl), +}; + +static const unsigned char vl_all[] = { VL_512, VL_128, VL_256 }; +static const unsigned char vl_128[] = { VL_128 }; + +/* + * This table, indicating the presence of an immediate (byte) for an opcode + * space 0f major opcode, is indexed by high major opcode byte nibble, with + * each table element then bit-indexed by low major opcode byte nibble. + */ +static const uint16_t imm0f[16] = { + [0x7] = (1 << 0x0) /* vpshuf* */ | + (1 << 0x1) /* vps{ll,ra,rl}w */ | + (1 << 0x2) /* vps{l,r}ld, vp{rol,ror,sra}{d,q} */ | + (1 << 0x3) /* vps{l,r}l{,d}q */, + [0xc] = (1 << 0x2) /* vcmp{p,s}{d,s} */ | + (1 << 0x4) /* vpinsrw */ | + (1 << 0x5) /* vpextrw */ | + (1 << 0x6) /* vshufp{d,s} */, +}; + +static struct x86_emulate_ops emulops; + +static unsigned int accessed[3 * 64]; + +static bool record_access(enum x86_segment seg, unsigned long offset, + unsigned int bytes) +{ + while ( bytes-- ) + { + if ( offset >= ARRAY_SIZE(accessed) ) + return false; + ++accessed[offset++]; + } + + return true; +} + +static int read(enum x86_segment seg, unsigned long offset, void *p_data, + unsigned int bytes, struct x86_emulate_ctxt *ctxt) +{ + if ( !record_access(seg, offset, bytes) ) + return X86EMUL_UNHANDLEABLE; + memset(p_data, 0, bytes); + return X86EMUL_OKAY; +} + +static int write(enum x86_segment seg, unsigned long offset, void *p_data, + unsigned int bytes, struct x86_emulate_ctxt *ctxt) +{ + if ( !record_access(seg, offset, bytes) ) + return X86EMUL_UNHANDLEABLE; + return X86EMUL_OKAY; +} + +static void test_one(const struct test *test, enum vl vl, + unsigned char *instr, struct x86_emulate_ctxt *ctxt) +{ + unsigned int vsz, esz, i; + int rc; + bool sg = strstr(test->mnemonic, "gather") || + strstr(test->mnemonic, "scatter"); + bool imm = test->spc == SPC_0f3a || + (test->spc == SPC_0f && + (imm0f[test->opc >> 4] & (1 << (test->opc & 0xf)))); + union evex { + uint8_t raw[3]; + struct { + uint8_t opcx:2; + uint8_t mbz:2; + uint8_t R:1; + uint8_t b:1; + uint8_t x:1; + uint8_t r:1; + uint8_t pfx:2; + uint8_t mbs:1; + uint8_t reg:4; + uint8_t w:1; + uint8_t opmsk:3; + uint8_t RX:1; + uint8_t bcst:1; + uint8_t lr:2; + uint8_t z:1; + }; + } evex = { + .opcx = test->spc, .pfx = test->pfx, .lr = vl, + .R = 1, .b = 1, .x = 1, .r = 1, .mbs = 1, + .reg = 0xf, .RX = 1, .opmsk = sg, + }; + + switch ( test->esz ) + { + case ESZ_b: + esz = 1; + break; + + case ESZ_w: + esz = 2; + evex.w = 1; + break; + +#ifdef __i386__ + case ESZ_d_WIG: + evex.w = 1; + /* fall through */ +#endif + case ESZ_d: case ESZ_d_nb: + esz = 4; + break; + + case ESZ_q: case ESZ_q_nb: + esz = 8; + evex.w = 1; + break; + + default: + ASSERT_UNREACHABLE(); + } + + switch ( test->vsz ) + { + case VSZ_vl: + vsz = 16 << vl; + break; + + case VSZ_vl_2: + vsz = 8 << vl; + break; + + case VSZ_vl_4: + vsz = 4 << vl; + break; + + case VSZ_vl_8: + vsz = 2 << vl; + break; + + case VSZ_el: + vsz = esz; + break; + + case VSZ_el_2: + vsz = esz * 2; + break; + + case VSZ_el_4: + vsz = esz * 4; + break; + + case VSZ_el_8: + vsz = esz * 8; + break; + + default: + ASSERT_UNREACHABLE(); + } + + /* + * Note: SIB addressing is used here, such that S/G insns can be handled + * without extra conditionals. + */ + instr[0] = 0x62; + instr[1] = evex.raw[0]; + instr[2] = evex.raw[1]; + instr[3] = evex.raw[2]; + instr[4] = test->opc; + instr[5] = 0x44 | (test->ext << 3); /* ModR/M */ + instr[6] = 0x12; /* SIB: base rDX, index none / xMM4 */ + instr[7] = 1; /* Disp8 */ + instr[8] = 0; /* immediate, if any */ + + asm volatile ( "kxnorw %k1, %k1, %k1" ); + asm volatile ( "vxorps %xmm4, %xmm4, %xmm4" ); + + ctxt->regs->eip = (unsigned long)&instr[0]; + ctxt->regs->edx = 0; + memset(accessed, 0, sizeof(accessed)); + + rc = x86_emulate(ctxt, &emulops); + if ( rc != X86EMUL_OKAY || + (ctxt->regs->eip != (unsigned long)&instr[8 + imm]) ) + goto fail; + + for ( i = 0; i < vsz; ++i ) + if ( accessed[i] ) + goto fail; + for ( ; i < (test->scale == SC_vl ? vsz : esz) + (sg ? esz : vsz); ++i ) + if ( accessed[i] != (sg ? vsz / esz : 1) ) + goto fail; + for ( ; i < ARRAY_SIZE(accessed); ++i ) + if ( accessed[i] ) + goto fail; + + /* Also check the broadcast case, if available. */ + if ( test->vsz >= VSZ_el || test->scale != SC_vl ) + return; + + switch ( test->esz ) + { + case ESZ_d_nb: case ESZ_q_nb: + case ESZ_b: case ESZ_w: case ESZ_bw: + return; + + case ESZ_d: case ESZ_q: + break; + + default: + ASSERT_UNREACHABLE(); + } + + evex.bcst = 1; + instr[3] = evex.raw[2]; + + ctxt->regs->eip = (unsigned long)&instr[0]; + memset(accessed, 0, sizeof(accessed)); + + rc = x86_emulate(ctxt, &emulops); + if ( rc != X86EMUL_OKAY || + (ctxt->regs->eip != (unsigned long)&instr[8 + imm]) ) + goto fail; + + for ( i = 0; i < esz; ++i ) + if ( accessed[i] ) + goto fail; + for ( ; i < esz * 2; ++i ) + if ( accessed[i] != 1 ) + goto fail; + for ( ; i < ARRAY_SIZE(accessed); ++i ) + if ( accessed[i] ) + goto fail; + + return; + + fail: + printf("failed (v%s%s %u-bit)\n", test->mnemonic, + evex.bcst ? "/bcst" : "", 128 << vl); + exit(1); +} + +static void test_pair(const struct test *tmpl, enum vl vl, + enum esz esz1, const char *suffix1, + enum esz esz2, const char *suffix2, + unsigned char *instr, struct x86_emulate_ctxt *ctxt) +{ + struct test test = *tmpl; + char mnemonic[24]; + + test.esz = esz1; + snprintf(mnemonic, ARRAY_SIZE(mnemonic), "%s%s", tmpl->mnemonic, suffix1); + test.mnemonic = mnemonic; + test_one(&test, vl, instr, ctxt); + + test.esz = esz2; + snprintf(mnemonic, ARRAY_SIZE(mnemonic), "%s%s", tmpl->mnemonic, suffix2); + test.mnemonic = mnemonic; + test_one(&test, vl, instr, ctxt); +} + +static void test_group(const struct test tests[], unsigned int nr_test, + const unsigned char vl[], unsigned int nr_vl, + void *instr, struct x86_emulate_ctxt *ctxt) +{ + unsigned int i, j; + + for ( i = 0; i < nr_test; ++i ) + { + for ( j = 0; j < nr_vl; ++j ) + { + if ( vl[0] == VL_512 && vl[j] != VL_512 && !cpu_has_avx512vl ) + continue; + + switch ( tests[i].esz ) + { + default: + test_one(&tests[i], vl[j], instr, ctxt); + break; + + case ESZ_bw: + test_pair(&tests[i], vl[j], ESZ_b, "b", ESZ_w, "w", + instr, ctxt); + break; + + case ESZ_dq: + test_pair(&tests[i], vl[j], ESZ_d, "d", ESZ_q, "q", + instr, ctxt); + break; + +#ifdef __i386__ + case ESZ_d_WIG: + test_pair(&tests[i], vl[j], ESZ_d, "/W0", + ESZ_d_WIG, "/W1", instr, ctxt); + break; +#endif + + case ESZ_sd: + test_pair(&tests[i], vl[j], + ESZ_d, tests[i].vsz < VSZ_el ? "ps" : "ss", + ESZ_q, tests[i].vsz < VSZ_el ? "pd" : "sd", + instr, ctxt); + break; + } + } + } +} + +void evex_disp8_test(void *instr, struct x86_emulate_ctxt *ctxt, + const struct x86_emulate_ops *ops) +{ + emulops = *ops; + emulops.read = read; + emulops.write = write; + +#define RUN(feat, vl) do { \ + if ( cpu_has_##feat ) \ + { \ + printf("%-40s", "Testing " #feat "/" #vl " disp8 handling..."); \ + test_group(feat ## _ ## vl, ARRAY_SIZE(feat ## _ ## vl), \ + vl_ ## vl, ARRAY_SIZE(vl_ ## vl), instr, ctxt); \ + printf("okay\n"); \ + } \ +} while ( false ) + + RUN(avx512f, all); + RUN(avx512f, 128); + RUN(avx512bw, all); +} --- a/tools/tests/x86_emulator/test_x86_emulator.c +++ b/tools/tests/x86_emulator/test_x86_emulator.c @@ -3795,6 +3795,9 @@ int main(int argc, char **argv) else printf("skipped\n"); + if ( stack_exec ) + evex_disp8_test(instr, &ctxt, &emulops); + for ( j = 0; j < ARRAY_SIZE(blobs); j++ ) { if ( blobs[j].check_cpu && !blobs[j].check_cpu() ) --- a/tools/tests/x86_emulator/x86-emulate.h +++ b/tools/tests/x86_emulator/x86-emulate.h @@ -71,6 +71,9 @@ WRAP(puts); #include "x86_emulate/x86_emulate.h" +void evex_disp8_test(void *instr, struct x86_emulate_ctxt *ctxt, + const struct x86_emulate_ops *ops); + static inline uint64_t xgetbv(uint32_t xcr) { uint32_t lo, hi; _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxxxxxxxxx https://lists.xenproject.org/mailman/listinfo/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |