[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH 3/8] x86emul: support MMX/SSE/SSE2 moves
Previously supported insns are being converted to the new model, and several new ones are being added. To keep the stub handling reasonably simple, integrate SET_SSE_PREFIX() into copy_REX_VEX(), at once switching the stubs to use an empty REX prefix instead of a double DS: one (no byte registers are being accessed, so an empty REX prefix has no effect), except (of course) for the 32-bit test harness build. Signed-off-by: Jan Beulich <jbeulich@xxxxxxxx> --- a/tools/tests/x86_emulator/test_x86_emulator.c +++ b/tools/tests/x86_emulator/test_x86_emulator.c @@ -1557,6 +1557,29 @@ int main(int argc, char **argv) else printf("skipped\n"); + printf("%-40s", "Testing movq 32(%ecx),%xmm1..."); + if ( stack_exec && cpu_has_sse2 ) + { + decl_insn(movq_from_mem2); + + asm volatile ( "pcmpeqb %%xmm1, %%xmm1\n" + put_insn(movq_from_mem2, "movq 32(%0), %%xmm1") + :: "c" (NULL) ); + + set_insn(movq_from_mem2); + rc = x86_emulate(&ctxt, &emulops); + if ( rc != X86EMUL_OKAY || !check_eip(movq_from_mem2) ) + goto fail; + asm ( "pcmpgtb %%xmm0, %%xmm0\n\t" + "pcmpeqb %%xmm1, %%xmm0\n\t" + "pmovmskb %%xmm0, %0" : "=r" (rc) ); + if ( rc != 0xffff ) + goto fail; + printf("okay\n"); + } + else + printf("skipped\n"); + printf("%-40s", "Testing vmovq %xmm1,32(%edx)..."); if ( stack_exec && cpu_has_avx ) { @@ -1581,6 +1604,29 @@ int main(int argc, char **argv) else printf("skipped\n"); + printf("%-40s", "Testing vmovq 32(%edx),%xmm0..."); + if ( stack_exec && cpu_has_avx ) + { + decl_insn(vmovq_from_mem); + + asm volatile ( "pcmpeqb %%xmm0, %%xmm0\n" + put_insn(vmovq_from_mem, "vmovq 32(%0), %%xmm0") + :: "d" (NULL) ); + + set_insn(vmovq_from_mem); + rc = x86_emulate(&ctxt, &emulops); + if ( rc != X86EMUL_OKAY || !check_eip(vmovq_from_mem) ) + goto fail; + asm ( "pcmpgtb %%xmm1, %%xmm1\n\t" + "pcmpeqb %%xmm0, %%xmm1\n\t" + "pmovmskb %%xmm1, %0" : "=r" (rc) ); + if ( rc != 0xffff ) + goto fail; + printf("okay\n"); + } + else + printf("skipped\n"); + printf("%-40s", "Testing movdqu %xmm2,(%ecx)..."); if ( stack_exec && cpu_has_sse2 ) { @@ -1812,6 +1858,33 @@ int main(int argc, char **argv) else printf("skipped\n"); + printf("%-40s", "Testing movd 32(%ecx),%mm4..."); + if ( stack_exec && cpu_has_mmx ) + { + decl_insn(movd_from_mem); + + asm volatile ( "pcmpgtb %%mm4, %%mm4\n" + put_insn(movd_from_mem, "movd 32(%0), %%mm4") + :: "c" (NULL) ); + + set_insn(movd_from_mem); + rc = x86_emulate(&ctxt, &emulops); + if ( rc != X86EMUL_OKAY || !check_eip(movd_from_mem) ) + goto fail; + asm ( "pxor %%mm2,%%mm2\n\t" + "pcmpeqb %%mm4, %%mm2\n\t" + "pmovmskb %%mm2, %0" : "=r" (rc) ); + if ( rc != 0xf0 ) + goto fail; + asm ( "pcmpeqb %%mm4, %%mm3\n\t" + "pmovmskb %%mm3, %0" : "=r" (rc) ); + if ( rc != 0x0f ) + goto fail; + printf("okay\n"); + } + else + printf("skipped\n"); + printf("%-40s", "Testing movd %xmm2,32(%edx)..."); if ( stack_exec && cpu_has_sse2 ) { @@ -1836,6 +1909,34 @@ int main(int argc, char **argv) else printf("skipped\n"); + printf("%-40s", "Testing movd 32(%edx),%xmm3..."); + if ( stack_exec && cpu_has_sse2 ) + { + decl_insn(movd_from_mem2); + + asm volatile ( "pcmpeqb %%xmm3, %%xmm3\n" + put_insn(movd_from_mem2, "movd 32(%0), %%xmm3") + :: "d" (NULL) ); + + set_insn(movd_from_mem2); + rc = x86_emulate(&ctxt, &emulops); + if ( rc != X86EMUL_OKAY || !check_eip(movd_from_mem2) ) + goto fail; + asm ( "pxor %%xmm1,%%xmm1\n\t" + "pcmpeqb %%xmm3, %%xmm1\n\t" + "pmovmskb %%xmm1, %0" : "=r" (rc) ); + if ( rc != 0xfff0 ) + goto fail; + asm ( "pcmpeqb %%xmm2, %%xmm2\n\t" + "pcmpeqb %%xmm3, %%xmm2\n\t" + "pmovmskb %%xmm2, %0" : "=r" (rc) ); + if ( rc != 0x000f ) + goto fail; + printf("okay\n"); + } + else + printf("skipped\n"); + printf("%-40s", "Testing vmovd %xmm1,32(%ecx)..."); if ( stack_exec && cpu_has_avx ) { @@ -1860,6 +1961,34 @@ int main(int argc, char **argv) else printf("skipped\n"); + printf("%-40s", "Testing vmovd 32(%ecx),%xmm2..."); + if ( stack_exec && cpu_has_avx ) + { + decl_insn(vmovd_from_mem); + + asm volatile ( "pcmpeqb %%xmm2, %%xmm2\n" + put_insn(vmovd_from_mem, "vmovd 32(%0), %%xmm2") + :: "c" (NULL) ); + + set_insn(vmovd_from_mem); + rc = x86_emulate(&ctxt, &emulops); + if ( rc != X86EMUL_OKAY || !check_eip(vmovd_from_mem) ) + goto fail; + asm ( "pxor %%xmm0,%%xmm0\n\t" + "pcmpeqb %%xmm2, %%xmm0\n\t" + "pmovmskb %%xmm0, %0" : "=r" (rc) ); + if ( rc != 0xfff0 ) + goto fail; + asm ( "pcmpeqb %%xmm1, %%xmm1\n\t" + "pcmpeqb %%xmm2, %%xmm1\n\t" + "pmovmskb %%xmm1, %0" : "=r" (rc) ); + if ( rc != 0x000f ) + goto fail; + printf("okay\n"); + } + else + printf("skipped\n"); + printf("%-40s", "Testing movd %mm3,%ebx..."); if ( stack_exec && cpu_has_mmx ) { @@ -1890,6 +2019,34 @@ int main(int argc, char **argv) else printf("skipped\n"); + printf("%-40s", "Testing movd %ebx,%mm4..."); + if ( stack_exec && cpu_has_mmx ) + { + decl_insn(movd_from_reg); + + /* See comment next to movd above. */ + asm volatile ( "pcmpgtb %%mm4, %%mm4\n" + put_insn(movd_from_reg, "movd %%ebx, %%mm4") + :: ); + + set_insn(movd_from_reg); + rc = x86_emulate(&ctxt, &emulops); + if ( (rc != X86EMUL_OKAY) || !check_eip(movd_from_reg) ) + goto fail; + asm ( "pxor %%mm2,%%mm2\n\t" + "pcmpeqb %%mm4, %%mm2\n\t" + "pmovmskb %%mm2, %0" : "=r" (rc) ); + if ( rc != 0xf0 ) + goto fail; + asm ( "pcmpeqb %%mm4, %%mm3\n\t" + "pmovmskb %%mm3, %0" : "=r" (rc) ); + if ( rc != 0x0f ) + goto fail; + printf("okay\n"); + } + else + printf("skipped\n"); + printf("%-40s", "Testing movd %xmm2,%ebx..."); if ( stack_exec && cpu_has_sse2 ) { @@ -1915,6 +2072,35 @@ int main(int argc, char **argv) else printf("skipped\n"); + printf("%-40s", "Testing movd %ebx,%xmm3..."); + if ( stack_exec && cpu_has_sse2 ) + { + decl_insn(movd_from_reg2); + + /* See comment next to movd above. */ + asm volatile ( "pcmpgtb %%xmm3, %%xmm3\n" + put_insn(movd_from_reg2, "movd %%ebx, %%xmm3") + :: ); + + set_insn(movd_from_reg2); + rc = x86_emulate(&ctxt, &emulops); + if ( (rc != X86EMUL_OKAY) || !check_eip(movd_from_reg2) ) + goto fail; + asm ( "pxor %%xmm1,%%xmm1\n\t" + "pcmpeqb %%xmm3, %%xmm1\n\t" + "pmovmskb %%xmm1, %0" : "=r" (rc) ); + if ( rc != 0xfff0 ) + goto fail; + asm ( "pcmpeqb %%xmm2, %%xmm2\n\t" + "pcmpeqb %%xmm3, %%xmm2\n\t" + "pmovmskb %%xmm2, %0" : "=r" (rc) ); + if ( rc != 0x000f ) + goto fail; + printf("okay\n"); + } + else + printf("skipped\n"); + printf("%-40s", "Testing vmovd %xmm1,%ebx..."); if ( stack_exec && cpu_has_avx ) { @@ -1937,6 +2123,35 @@ int main(int argc, char **argv) goto fail; printf("okay\n"); } + else + printf("skipped\n"); + + printf("%-40s", "Testing vmovd %ebx,%xmm2..."); + if ( stack_exec && cpu_has_avx ) + { + decl_insn(vmovd_from_reg); + + /* See comment next to movd above. */ + asm volatile ( "pcmpgtb %%xmm2, %%xmm2\n" + put_insn(vmovd_from_reg, "vmovd %%ebx, %%xmm2") + :: ); + + set_insn(vmovd_from_reg); + rc = x86_emulate(&ctxt, &emulops); + if ( (rc != X86EMUL_OKAY) || !check_eip(vmovd_from_reg) ) + goto fail; + asm ( "pxor %%xmm0,%%xmm0\n\t" + "pcmpeqb %%xmm2, %%xmm0\n\t" + "pmovmskb %%xmm0, %0" : "=r" (rc) ); + if ( rc != 0xfff0 ) + goto fail; + asm ( "pcmpeqb %%xmm1, %%xmm1\n\t" + "pcmpeqb %%xmm2, %%xmm1\n\t" + "pmovmskb %%xmm1, %0" : "=r" (rc) ); + if ( rc != 0x000f ) + goto fail; + printf("okay\n"); + } else printf("skipped\n"); --- a/xen/arch/x86/x86_emulate/x86_emulate.c +++ b/xen/arch/x86/x86_emulate/x86_emulate.c @@ -236,11 +236,11 @@ static const struct { [0x0f] = { ModRM|SrcImmByte }, [0x10] = { DstImplicit|SrcMem|ModRM|Mov, simd_any_fp }, [0x11] = { DstMem|SrcImplicit|ModRM|Mov, simd_any_fp }, - [0x12] = { ImplicitOps|ModRM }, - [0x13] = { ImplicitOps|ModRM }, + [0x12] = { DstImplicit|SrcMem|ModRM|Mov, simd_other }, + [0x13] = { DstMem|SrcImplicit|ModRM|Mov, simd_other }, [0x14 ... 0x15] = { DstImplicit|SrcMem|ModRM, simd_packed_fp }, - [0x16] = { ImplicitOps|ModRM }, - [0x17] = { ImplicitOps|ModRM }, + [0x16] = { DstImplicit|SrcMem|ModRM|Mov, simd_other }, + [0x17] = { DstMem|SrcImplicit|ModRM|Mov, simd_other }, [0x18] = { ImplicitOps|ModRM }, [0x19] = { ImplicitOps|ModRM }, [0x1a] = { ImplicitOps|ModRM }, @@ -267,7 +267,7 @@ static const struct { [0x38] = { DstReg|SrcMem|ModRM }, [0x3a] = { DstReg|SrcImmByte|ModRM }, [0x40 ... 0x4f] = { DstReg|SrcMem|ModRM|Mov }, - [0x50] = { ModRM }, + [0x50] = { DstReg|SrcImplicit|ModRM|Mov }, [0x51] = { DstImplicit|SrcMem|ModRM|TwoOp, simd_any_fp }, [0x52 ... 0x53] = { DstImplicit|SrcMem|ModRM|TwoOp, simd_single_fp }, [0x54 ... 0x57] = { DstImplicit|SrcMem|ModRM, simd_packed_fp }, @@ -280,7 +280,8 @@ static const struct { [0x68 ... 0x6a] = { DstImplicit|SrcMem|ModRM, simd_other }, [0x6b] = { DstImplicit|SrcMem|ModRM, simd_packed_int }, [0x6c ... 0x6d] = { DstImplicit|SrcMem|ModRM, simd_packed_int }, - [0x6e ... 0x6f] = { ImplicitOps|ModRM }, + [0x6e] = { DstImplicit|SrcMem|ModRM|Mov }, + [0x6f] = { DstImplicit|SrcMem|ModRM|Mov, simd_packed_int }, [0x70] = { SrcImmByte|ModRM|TwoOp, simd_other }, [0x71 ... 0x73] = { SrcImmByte|ModRM }, [0x74 ... 0x76] = { DstImplicit|SrcMem|ModRM, simd_packed_int }, @@ -289,7 +290,8 @@ static const struct { [0x79] = { ModRM }, [0x7c] = { DstImplicit|SrcMem|ModRM, simd_other }, [0x7d] = { DstImplicit|SrcMem|ModRM, simd_other }, - [0x7e ... 0x7f] = { ImplicitOps|ModRM }, + [0x7e] = { DstMem|SrcImplicit|ModRM|Mov }, + [0x7f] = { DstMem|SrcImplicit|ModRM|Mov, simd_packed_int }, [0x80 ... 0x8f] = { DstImplicit|SrcImm }, [0x90 ... 0x9f] = { ByteOp|DstMem|SrcNone|ModRM|Mov }, [0xa0 ... 0xa1] = { ImplicitOps|Mov }, @@ -332,19 +334,19 @@ static const struct { [0xd0] = { DstImplicit|SrcMem|ModRM, simd_other }, [0xd1 ... 0xd3] = { DstImplicit|SrcMem|ModRM, simd_other }, [0xd4 ... 0xd5] = { DstImplicit|SrcMem|ModRM, simd_packed_int }, - [0xd6] = { ImplicitOps|ModRM }, - [0xd7] = { ModRM }, + [0xd6] = { DstMem|SrcImplicit|ModRM|Mov, simd_other }, + [0xd7] = { DstReg|SrcImplicit|ModRM|Mov }, [0xd8 ... 0xdf] = { DstImplicit|SrcMem|ModRM, simd_packed_int }, [0xe0] = { DstImplicit|SrcMem|ModRM, simd_packed_int }, [0xe1 ... 0xe2] = { DstImplicit|SrcMem|ModRM, simd_other }, [0xe3 ... 0xe5] = { DstImplicit|SrcMem|ModRM, simd_packed_int }, [0xe6] = { ModRM }, - [0xe7] = { ImplicitOps|ModRM }, + [0xe7] = { DstMem|SrcImplicit|ModRM|Mov, simd_packed_int }, [0xe8 ... 0xef] = { DstImplicit|SrcMem|ModRM, simd_packed_int }, [0xf0] = { DstImplicit|SrcMem|ModRM|Mov, simd_other }, [0xf1 ... 0xf3] = { DstImplicit|SrcMem|ModRM, simd_other }, [0xf4 ... 0xf6] = { DstImplicit|SrcMem|ModRM, simd_packed_int }, - [0xf7] = { ModRM }, + [0xf7] = { DstMem|SrcMem|ModRM|Mov, simd_packed_int }, [0xf8 ... 0xfe] = { DstImplicit|SrcMem|ModRM, simd_packed_int }, [0xff] = { ModRM } }; @@ -380,11 +382,6 @@ enum vex_pfx { static const uint8_t sse_prefix[] = { 0x66, 0xf3, 0xf2 }; -#define SET_SSE_PREFIX(dst, vex_pfx) do { \ - if ( vex_pfx ) \ - (dst) = sse_prefix[(vex_pfx) - 1]; \ -} while (0) - union vex { uint8_t raw[2]; struct { @@ -399,15 +396,35 @@ union vex { }; }; +#ifdef __x86_64__ +# define PFX2 REX_PREFIX +#else +# define PFX2 0x3e +#endif +#define PFX_BYTES 3 +#define init_prefixes(stub) ({ \ + uint8_t *buf_ = get_stub(stub); \ + buf_[0] = 0x3e; \ + buf_[1] = PFX2; \ + buf_[2] = 0x0f; \ + buf_ + 3; \ +}) + #define copy_REX_VEX(ptr, rex, vex) do { \ if ( (vex).opcx != vex_none ) \ { \ if ( !mode_64bit() ) \ vex.reg |= 8; \ - ptr[0] = 0xc4, ptr[1] = (vex).raw[0], ptr[2] = (vex).raw[1]; \ + (ptr)[0 - PFX_BYTES] = 0xc4; \ + (ptr)[1 - PFX_BYTES] = (vex).raw[0]; \ + (ptr)[2 - PFX_BYTES] = (vex).raw[1]; \ + } \ + else \ + { \ + if ( (vex).pfx ) \ + (ptr)[0 - PFX_BYTES] = sse_prefix[(vex).pfx - 1]; \ + (ptr)[1 - PFX_BYTES] |= rex; \ } \ - else if ( mode_64bit() ) \ - ptr[1] = rex | REX_PREFIX; \ } while (0) union evex { @@ -2180,7 +2197,8 @@ x86_decode_twobyte( case 0x10 ... 0x18: case 0x28 ... 0x2f: case 0x50 ... 0x77: - case 0x79 ... 0x7f: + case 0x79 ... 0x7d: + case 0x7f: case 0xae: case 0xc2 ... 0xc3: case 0xc5 ... 0xc6: @@ -2200,6 +2218,18 @@ x86_decode_twobyte( op_bytes = mode_64bit() ? 8 : 4; break; + case 0x7e: + ctxt->opcode |= MASK_INSR(vex.pfx, X86EMUL_OPC_PFX_MASK); + if ( vex.pfx == vex_f3 ) /* movq xmm/m64,xmm */ + { + case X86EMUL_OPC_VEX_F3(0, 0x7e): /* vmovq xmm/m64,xmm */ + state->desc = DstImplicit | SrcMem | ModRM | Mov; + state->simd_size = simd_other; + /* Avoid the state->desc adjustment below. */ + return X86EMUL_OKAY; + } + break; + case 0xb8: /* jmpe / popcnt */ if ( rep_prefix() ) ctxt->opcode |= MASK_INSR(vex.pfx, X86EMUL_OPC_PFX_MASK); @@ -2797,7 +2827,7 @@ x86_emulate( struct cpu_user_regs _regs = *ctxt->regs; struct x86_emulate_state state; int rc; - uint8_t b, d; + uint8_t b, d, *opc = NULL; bool singlestep = (_regs._eflags & EFLG_TF) && !is_branch_step(ctxt, ops); bool sfence = false; struct operand src = { .reg = PTR_POISON }; @@ -5276,6 +5306,7 @@ x86_emulate( CASE_SIMD_ALL_FP(_VEX, 0x0f, 0x5e): /* vdiv{p,s}{s,d} {x,y}mm/mem,{x,y}mm,{x,y}mm */ CASE_SIMD_ALL_FP(, 0x0f, 0x5f): /* max{p,s}{s,d} xmm/mem,xmm */ CASE_SIMD_ALL_FP(_VEX, 0x0f, 0x5f): /* vmax{p,s}{s,d} {x,y}mm/mem,{x,y}mm,{x,y}mm */ + simd_0f_fp: if ( vex.opcx == vex_none ) { if ( vex.pfx & VEX_PREFIX_DOUBLE_MASK ) @@ -5291,24 +5322,62 @@ x86_emulate( get_fpu(X86EMUL_FPU_ymm, &fic); } simd_0f_common: - { - uint8_t *buf = get_stub(stub); - - buf[0] = 0x3e; - buf[1] = 0x3e; - buf[2] = 0x0f; - buf[3] = b; - buf[4] = modrm; + opc = init_prefixes(stub); + opc[0] = b; + opc[1] = modrm; if ( ea.type == OP_MEM ) { /* convert memory operand to (%rAX) */ rex_prefix &= ~REX_B; vex.b = 1; - buf[4] &= 0x38; + opc[1] &= 0x38; } - fic.insn_bytes = 5; + fic.insn_bytes = PFX_BYTES + 2; break; - } + + case X86EMUL_OPC_66(0x0f, 0x12): /* movlpd m64,xmm */ + case X86EMUL_OPC_VEX_66(0x0f, 0x12): /* vmovlpd m64,xmm,xmm */ + CASE_SIMD_PACKED_FP(, 0x0f, 0x13): /* movlp{s,d} xmm,m64 */ + CASE_SIMD_PACKED_FP(_VEX, 0x0f, 0x13): /* vmovlp{s,d} xmm,m64 */ + case X86EMUL_OPC_66(0x0f, 0x16): /* movhpd m64,xmm */ + case X86EMUL_OPC_VEX_66(0x0f, 0x16): /* vmovhpd m64,xmm,xmm */ + CASE_SIMD_PACKED_FP(, 0x0f, 0x17): /* movhp{s,d} xmm,m64 */ + CASE_SIMD_PACKED_FP(_VEX, 0x0f, 0x17): /* vmovhp{s,d} xmm,m64 */ + generate_exception_if(ea.type != OP_MEM, EXC_UD); + /* fall through */ + case X86EMUL_OPC(0x0f, 0x12): /* movlps m64,xmm */ + /* movhlps xmm,xmm */ + case X86EMUL_OPC_VEX(0x0f, 0x12): /* vmovlps m64,xmm,xmm */ + /* vmovhlps xmm,xmm,xmm */ + case X86EMUL_OPC(0x0f, 0x16): /* movhps m64,xmm */ + /* movlhps xmm,xmm */ + case X86EMUL_OPC_VEX(0x0f, 0x16): /* vmovhps m64,xmm,xmm */ + /* vmovlhps xmm,xmm,xmm */ + generate_exception_if(vex.l, EXC_UD); + d &= ~TwoOp; + op_bytes = 8; + goto simd_0f_fp; + + case X86EMUL_OPC_F3(0x0f, 0x12): /* movsldup xmm/m128,xmm */ + case X86EMUL_OPC_VEX_F3(0x0f, 0x12): /* vmovsldup {x,y}mm/mem,{x,y}mm */ + case X86EMUL_OPC_F2(0x0f, 0x12): /* movddup xmm/m64,xmm */ + case X86EMUL_OPC_VEX_F2(0x0f, 0x12): /* vmovddup {x,y}mm/mem,{x,y}mm */ + case X86EMUL_OPC_F3(0x0f, 0x16): /* movshdup xmm/m128,xmm */ + case X86EMUL_OPC_VEX_F3(0x0f, 0x16): /* vmovshdup {x,y}mm/mem,{x,y}mm */ + d |= TwoOp; + op_bytes = !(vex.pfx & VEX_PREFIX_DOUBLE_MASK) || vex.l + ? 16 << vex.l : 8; + if ( vex.opcx == vex_none ) + { + host_and_vcpu_must_have(sse3); + get_fpu(X86EMUL_FPU_xmm, &fic); + } + else + { + host_and_vcpu_must_have(avx); + get_fpu(X86EMUL_FPU_ymm, &fic); + } + goto simd_0f_common; case X86EMUL_OPC(0x0f, 0x20): /* mov cr,reg */ case X86EMUL_OPC(0x0f, 0x21): /* mov dr,reg */ @@ -5469,6 +5538,57 @@ x86_emulate( break; } + CASE_SIMD_PACKED_FP(, 0x0f, 0x50): /* movmskp{s,d} xmm,reg */ + CASE_SIMD_PACKED_FP(_VEX, 0x0f, 0x50): /* vmovmskp{s,d} {x,y}mm,reg */ + CASE_SIMD_PACKED_INT(0x0f, 0xd7): /* pmovmskb {,x}mm,reg */ + case X86EMUL_OPC_VEX_66(0x0f, 0xd7): /* vpmovmskb {x,y}mm,reg */ + generate_exception_if(ea.type != OP_REG, EXC_UD); + + if ( vex.opcx == vex_none ) + { + if ( vex.pfx & VEX_PREFIX_DOUBLE_MASK ) + vcpu_must_have(sse2); + else + { + if ( b != 0x50 ) + host_and_vcpu_must_have(mmx); + vcpu_must_have(sse); + } + if ( b == 0x50 || (vex.pfx & VEX_PREFIX_DOUBLE_MASK) ) + get_fpu(X86EMUL_FPU_xmm, &fic); + else + get_fpu(X86EMUL_FPU_mmx, &fic); + } + else + { + generate_exception_if(vex.reg != 0xf, EXC_UD); + if ( b == 0x50 || !vex.l ) + host_and_vcpu_must_have(avx); + else + host_and_vcpu_must_have(avx2); + get_fpu(X86EMUL_FPU_ymm, &fic); + } + + opc = init_prefixes(stub); + opc[0] = b; + /* Convert GPR destination to %rAX. */ + rex_prefix &= ~REX_R; + vex.r = 1; + if ( !mode_64bit() ) + vex.w = 0; + opc[1] = modrm & 0xc7; + fic.insn_bytes = PFX_BYTES + 2; + opc[2] = 0xc3; + + copy_REX_VEX(opc, rex_prefix, vex); + invoke_stub("", "", "=a" (dst.val) : [dummy] "i" (0)); + + put_stub(stub); + put_fpu(&fic); + + dst.bytes = 4; + break; + CASE_SIMD_PACKED_INT(0x0f, 0x60): /* punpcklbw {,x}mm/mem,{,x}mm */ case X86EMUL_OPC_VEX_66(0x0f, 0x60): /* vpunpcklbw {x,y}mm/mem,{x,y}mm,{x,y}mm */ CASE_SIMD_PACKED_INT(0x0f, 0x61): /* punpcklwd {,x}mm/mem,{,x}mm */ @@ -5588,134 +5708,72 @@ x86_emulate( } goto simd_0f_common; - case X86EMUL_OPC(0x0f, 0xe7): /* movntq mm,m64 */ - case X86EMUL_OPC_66(0x0f, 0xe7): /* movntdq xmm,m128 */ - case X86EMUL_OPC_VEX_66(0x0f, 0xe7): /* vmovntdq xmm,m128 */ - /* vmovntdq ymm,m256 */ - fail_if(ea.type != OP_MEM); - /* fall through */ - case X86EMUL_OPC(0x0f, 0x6f): /* movq mm/m64,mm */ - case X86EMUL_OPC_66(0x0f, 0x6f): /* movdqa xmm/m128,xmm */ - case X86EMUL_OPC_F3(0x0f, 0x6f): /* movdqu xmm/m128,xmm */ - case X86EMUL_OPC_VEX_66(0x0f, 0x6f): /* vmovdqa xmm/m128,xmm */ - /* vmovdqa ymm/m256,ymm */ - case X86EMUL_OPC_VEX_F3(0x0f, 0x6f): /* vmovdqu xmm/m128,xmm */ - /* vmovdqu ymm/m256,ymm */ - case X86EMUL_OPC(0x0f, 0x7e): /* movd mm,r/m32 */ - /* movq mm,r/m64 */ - case X86EMUL_OPC_66(0x0f, 0x7e): /* movd xmm,r/m32 */ - /* movq xmm,r/m64 */ - case X86EMUL_OPC_VEX_66(0x0f, 0x7e): /* vmovd xmm,r/m32 */ - /* vmovq xmm,r/m64 */ - case X86EMUL_OPC(0x0f, 0x7f): /* movq mm,mm/m64 */ - case X86EMUL_OPC_66(0x0f, 0x7f): /* movdqa xmm,xmm/m128 */ - case X86EMUL_OPC_VEX_66(0x0f, 0x7f): /* vmovdqa xmm,xmm/m128 */ - /* vmovdqa ymm,ymm/m256 */ - case X86EMUL_OPC_F3(0x0f, 0x7f): /* movdqu xmm,xmm/m128 */ - case X86EMUL_OPC_VEX_F3(0x0f, 0x7f): /* vmovdqu xmm,xmm/m128 */ - /* vmovdqu ymm,ymm/m256 */ - case X86EMUL_OPC_66(0x0f, 0xd6): /* movq xmm,xmm/m64 */ - case X86EMUL_OPC_VEX_66(0x0f, 0xd6): /* vmovq xmm,xmm/m64 */ - { - uint8_t *buf = get_stub(stub); - - fic.insn_bytes = 5; - buf[0] = 0x3e; - buf[1] = 0x3e; - buf[2] = 0x0f; - buf[3] = b; - buf[4] = modrm; - buf[5] = 0xc3; - if ( vex.opcx == vex_none ) - { - switch ( vex.pfx ) - { - case vex_66: - case vex_f3: - vcpu_must_have(sse2); - /* Converting movdqu to movdqa here: Our buffer is aligned. */ - buf[0] = 0x66; - get_fpu(X86EMUL_FPU_xmm, &fic); - ea.bytes = 16; - break; - case vex_none: - if ( b != 0xe7 ) - host_and_vcpu_must_have(mmx); - else - vcpu_must_have(sse); - get_fpu(X86EMUL_FPU_mmx, &fic); - ea.bytes = 8; - break; - default: - goto cannot_emulate; - } - } - else + CASE_SIMD_PACKED_INT(0x0f, 0x6e): /* mov{d,q} r/m,{,x}mm */ + case X86EMUL_OPC_VEX_66(0x0f, 0x6e): /* vmov{d,q} r/m,xmm */ + CASE_SIMD_PACKED_INT(0x0f, 0x7e): /* mov{d,q} {,x}mm,r/m */ + case X86EMUL_OPC_VEX_66(0x0f, 0x7e): /* vmov{d,q} xmm,r/m */ + if ( vex.opcx != vex_none ) { - fail_if(vex.reg != 0xf); + generate_exception_if(vex.l || vex.reg != 0xf, EXC_UD); host_and_vcpu_must_have(avx); get_fpu(X86EMUL_FPU_ymm, &fic); - ea.bytes = 16 << vex.l; } - switch ( b ) + else if ( vex.pfx ) { - case 0x7e: - generate_exception_if(vex.l, EXC_UD); - ea.bytes = op_bytes; - break; - case 0xd6: - generate_exception_if(vex.l, EXC_UD); - ea.bytes = 8; - break; + vcpu_must_have(sse2); + get_fpu(X86EMUL_FPU_xmm, &fic); } - if ( ea.type == OP_MEM ) + else { - uint32_t mxcsr = 0; - - if ( ea.bytes < 16 || vex.pfx == vex_f3 ) - mxcsr = MXCSR_MM; - else if ( vcpu_has_misalignsse() ) - asm ( "stmxcsr %0" : "=m" (mxcsr) ); - generate_exception_if(!(mxcsr & MXCSR_MM) && - !is_aligned(ea.mem.seg, ea.mem.off, ea.bytes, - ctxt, ops), - EXC_GP, 0); - if ( b == 0x6f ) - rc = ops->read(ea.mem.seg, ea.mem.off+0, mmvalp, - ea.bytes, ctxt); - else - fail_if(!ops->write); /* Check before running the stub. */ + host_and_vcpu_must_have(mmx); + get_fpu(X86EMUL_FPU_mmx, &fic); } - if ( ea.type == OP_MEM || b == 0x7e ) + + opc = init_prefixes(stub); + opc[0] = b; + /* Convert memory/GPR operand to (%rAX). */ + rex_prefix &= ~REX_B; + vex.b = 1; + if ( !mode_64bit() ) + vex.w = 0; + opc[1] = modrm & 0x38; + fic.insn_bytes = PFX_BYTES + 2; + opc[2] = 0xc3; + + copy_REX_VEX(opc, rex_prefix, vex); + invoke_stub("", "", "+m" (src.val) : "a" (&src.val)); + dst.val = src.val; + + put_stub(stub); + put_fpu(&fic); + break; + + case X86EMUL_OPC_66(0x0f, 0xe7): /* movntdq xmm,m128 */ + case X86EMUL_OPC_VEX_66(0x0f, 0xe7): /* vmovntdq {x,y}mm,mem */ + generate_exception_if(ea.type != OP_MEM, EXC_UD); + sfence = true; + /* fall through */ + case X86EMUL_OPC_66(0x0f, 0x6f): /* movdqa xmm/m128,xmm */ + case X86EMUL_OPC_VEX_66(0x0f, 0x6f): /* vmovdqa {x,y}mm/mem,{x,y}mm */ + case X86EMUL_OPC_F3(0x0f, 0x6f): /* movdqu xmm/m128,xmm */ + case X86EMUL_OPC_VEX_F3(0x0f, 0x6f): /* vmovdqu {x,y}mm/mem,{x,y}mm */ + case X86EMUL_OPC_66(0x0f, 0x7f): /* movdqa xmm,xmm/m128 */ + case X86EMUL_OPC_VEX_66(0x0f, 0x7f): /* vmovdqa {x,y}mm,{x,y}mm/m128 */ + case X86EMUL_OPC_F3(0x0f, 0x7f): /* movdqu xmm,xmm/m128 */ + case X86EMUL_OPC_VEX_F3(0x0f, 0x7f): /* vmovdqu {x,y}mm,{x,y}mm/mem */ + if ( vex.opcx != vex_none ) { - /* Convert memory operand or GPR destination to (%rAX) */ - rex_prefix &= ~REX_B; - vex.b = 1; - buf[4] &= 0x38; - if ( ea.type == OP_MEM ) - ea.reg = (void *)mmvalp; - else /* Ensure zero-extension of a 32-bit result. */ - *ea.reg = 0; - } - if ( !rc ) - { - copy_REX_VEX(buf, rex_prefix, vex); - asm volatile ( "call *%0" : : "r" (stub.func), "a" (ea.reg) - : "memory" ); + host_and_vcpu_must_have(avx); + get_fpu(X86EMUL_FPU_ymm, &fic); } - put_fpu(&fic); - put_stub(stub); - if ( !rc && (b != 0x6f) && (ea.type == OP_MEM) ) + else { - ASSERT(ops->write); /* See the fail_if() above. */ - rc = ops->write(ea.mem.seg, ea.mem.off, mmvalp, - ea.bytes, ctxt); + vcpu_must_have(sse2); + get_fpu(X86EMUL_FPU_xmm, &fic); } - if ( rc ) - goto done; - dst.type = OP_NONE; - break; - } + d |= TwoOp; + op_bytes = 16 << vex.l; + goto simd_0f_common; CASE_SIMD_PACKED_INT(0x0f, 0x70): /* pshuf{w,d} $imm8,{,x}mm/mem,{,x}mm */ case X86EMUL_OPC_VEX_66(0x0f, 0x70): /* vpshufd $imm8,{x,y}mm/mem,{x,y}mm */ @@ -5744,25 +5802,25 @@ x86_emulate( get_fpu(X86EMUL_FPU_mmx, &fic); } simd_0f_imm8: - { - uint8_t *buf = get_stub(stub); - - buf[0] = 0x3e; - buf[1] = 0x3e; - buf[2] = 0x0f; - buf[3] = b; - buf[4] = modrm; + opc = init_prefixes(stub); + opc[0] = b; + opc[1] = modrm; if ( ea.type == OP_MEM ) { /* Convert memory operand to (%rAX). */ rex_prefix &= ~REX_B; vex.b = 1; - buf[4] &= 0x38; + opc[1] &= 0x38; } - buf[5] = imm1; - fic.insn_bytes = 6; + opc[2] = imm1; + fic.insn_bytes = PFX_BYTES + 3; break; - } + + case X86EMUL_OPC_F3(0x0f, 0x7e): /* movq xmm/m64,xmm */ + case X86EMUL_OPC_VEX_F3(0x0f, 0x7e): /* vmovq xmm/m64,xmm */ + generate_exception_if(vex.l, EXC_UD); + op_bytes = 8; + goto simd_0f_int; case X86EMUL_OPC_F2(0x0f, 0xf0): /* lddqu m128,xmm */ case X86EMUL_OPC_VEX_F2(0x0f, 0xf0): /* vlddqu mem,{x,y}mm */ @@ -6335,6 +6393,27 @@ x86_emulate( get_fpu(X86EMUL_FPU_mmx, &fic); goto simd_0f_common; + case X86EMUL_OPC_VEX_66(0x0f, 0xd6): /* vmovq xmm,xmm/m64 */ + generate_exception_if(vex.l, EXC_UD); + /* fall through */ + case X86EMUL_OPC_66(0x0f, 0xd6): /* movq xmm,xmm/m64 */ + case X86EMUL_OPC(0x0f, 0x6f): /* movq mm/m64,mm */ + case X86EMUL_OPC(0x0f, 0x7f): /* movq mm,mm/m64 */ + op_bytes = 8; + d |= TwoOp; + goto simd_0f_int; + + case X86EMUL_OPC_F3(0x0f, 0xd6): /* movq2dq mm,xmm */ + case X86EMUL_OPC_F2(0x0f, 0xd6): /* movdq2q xmm,mm */ + generate_exception_if(ea.type != OP_REG, EXC_UD); + op_bytes = 8; + host_and_vcpu_must_have(mmx); + goto simd_0f_int; + + case X86EMUL_OPC(0x0f, 0xe7): /* movntq mm,m64 */ + generate_exception_if(ea.type != OP_MEM, EXC_UD); + sfence = true; + /* fall through */ case X86EMUL_OPC(0x0f, 0xda): /* pminub mm/m64,mm */ case X86EMUL_OPC(0x0f, 0xde): /* pmaxub mm/m64,mm */ case X86EMUL_OPC(0x0f, 0xea): /* pminsw mm/m64,mm */ @@ -6348,6 +6427,73 @@ x86_emulate( get_fpu(X86EMUL_FPU_mmx, &fic); goto simd_0f_common; + CASE_SIMD_PACKED_INT(0x0f, 0xf7): /* maskmov{q,dqu} {,x}mm,{,x}mm */ + case X86EMUL_OPC_VEX_66(0x0f, 0xf7): /* vmaskmovdqu xmm,xmm */ + generate_exception_if(ea.type != OP_REG, EXC_UD); + if ( vex.opcx != vex_none ) + { + generate_exception_if(vex.l || vex.reg != 0xf, EXC_UD); + host_and_vcpu_must_have(avx); + get_fpu(X86EMUL_FPU_ymm, &fic); + } + else if ( vex.pfx ) + { + vcpu_must_have(sse2); + get_fpu(X86EMUL_FPU_xmm, &fic); + } + else + { + host_and_vcpu_must_have(mmx); + vcpu_must_have(sse); + get_fpu(X86EMUL_FPU_mmx, &fic); + } + + /* + * While we can't reasonably provide fully correct behavior here + * (in particular avoiding the memory read in anticipation of all + * bytes in the range eventually being written), we can (and should) + * still suppress the memory access if all mask bits are clear. Read + * the mask bits via {,v}pmovmskb for that purpose. + */ + opc = init_prefixes(stub); + opc[0] = 0xd7; /* {,v}pmovmskb */ + /* (Ab)use "sfence" for latching the original REX.R / VEX.R. */ + sfence = rex_prefix & REX_R; + /* Convert GPR destination to %rAX. */ + rex_prefix &= ~REX_R; + vex.r = 1; + if ( !mode_64bit() ) + vex.w = 0; + opc[1] = modrm & 0xc7; + fic.insn_bytes = PFX_BYTES + 2; + opc[2] = 0xc3; + + copy_REX_VEX(opc, rex_prefix, vex); + invoke_stub("", "", "=a" (ea.val) : [dummy] "i" (0)); + + put_stub(stub); + if ( !ea.val ) + { + put_fpu(&fic); + goto complete_insn; + } + + opc = init_prefixes(stub); + opc[0] = b; + opc[1] = modrm; + /* Restore high bit of XMM destination. */ + if ( sfence ) + { + rex_prefix |= REX_R; + vex.r = 0; + } + + d |= TwoOp; + ea.type = OP_MEM; + ea.mem.off = truncate_ea(_regs.r(di)); + sfence = true; + break; + case X86EMUL_OPC(0x0f38, 0xf0): /* movbe m,r */ case X86EMUL_OPC(0x0f38, 0xf1): /* movbe r,m */ vcpu_must_have(movbe); @@ -6611,23 +6757,14 @@ x86_emulate( if ( state->simd_size ) { -#ifdef __XEN__ - uint8_t *buf = stub.ptr; -#else - uint8_t *buf = get_stub(stub); -#endif - generate_exception_if(!op_bytes, EXC_UD); generate_exception_if(vex.opcx && (d & TwoOp) && vex.reg != 0xf, EXC_UD); - if ( !buf ) + if ( !opc ) BUG(); - if ( vex.opcx == vex_none ) - SET_SSE_PREFIX(buf[0], vex.pfx); - - buf[fic.insn_bytes] = 0xc3; - copy_REX_VEX(buf, rex_prefix, vex); + opc[fic.insn_bytes - PFX_BYTES] = 0xc3; + copy_REX_VEX(opc, rex_prefix, vex); if ( ea.type == OP_MEM ) { @@ -6635,10 +6772,16 @@ x86_emulate( if ( op_bytes < 16 || (vex.opcx - ? /* vmov{a,nt}p{s,d} are exceptions. */ - ext != ext_0f || ((b | 1) != 0x29 && b != 0x2b) - : /* movup{s,d} and lddqu are exceptions. */ - ext == ext_0f && ((b | 1) == 0x11 || b == 0xf0)) ) + ? /* vmov{{a,nt}p{s,d},dqa,ntdq} are exceptions. */ + ext != ext_0f || + ((b | 1) != 0x29 && b != 0x2b && + ((b | 0x10) != 0x7f || vex.pfx != vex_66) && + b != 0xe7) + : /* movup{s,d}, {,mask}movdqu, and lddqu are exceptions. */ + ext == ext_0f && + ((b | 1) == 0x11 || + ((b | 0x10) == 0x7f && vex.pfx == vex_f3) || + b == 0xf7 || b == 0xf0)) ) mxcsr = MXCSR_MM; else if ( vcpu_has_misalignsse() ) asm ( "stmxcsr %0" : "=m" (mxcsr) ); @@ -6646,14 +6789,25 @@ x86_emulate( !is_aligned(ea.mem.seg, ea.mem.off, op_bytes, ctxt, ops), EXC_GP, 0); - if ( (d & SrcMask) == SrcMem ) + switch ( d & SrcMask ) { + case SrcMem: rc = ops->read(ea.mem.seg, ea.mem.off, mmvalp, op_bytes, ctxt); if ( rc != X86EMUL_OKAY ) goto done; + /* fall through */ + case SrcMem16: dst.type = OP_NONE; + break; + default: + if ( (d & DstMask) != DstMem ) + { + ASSERT_UNREACHABLE(); + return X86EMUL_UNHANDLEABLE; + } + break; } - else if ( (d & DstMask) == DstMem ) + if ( (d & DstMask) == DstMem ) { fail_if(!ops->write); /* Check before running the stub. */ ASSERT(d & Mov); @@ -6661,13 +6815,6 @@ x86_emulate( dst.bytes = op_bytes; dst.mem = ea.mem; } - else if ( (d & SrcMask) == SrcMem16 ) - dst.type = OP_NONE; - else - { - ASSERT_UNREACHABLE(); - return X86EMUL_UNHANDLEABLE; - } } else dst.type = OP_NONE; @@ -6928,6 +7075,8 @@ x86_insn_is_mem_access(const struct x86_ case 0xa4 ... 0xa7: /* MOVS / CMPS */ case 0xaa ... 0xaf: /* STOS / LODS / SCAS */ case 0xd7: /* XLAT */ + CASE_SIMD_PACKED_INT(0x0f, 0xf7): /* MASKMOV{Q,DQU} */ + case X86EMUL_OPC_VEX_66(0x0f, 0xf7): /* VMASKMOVDQU */ return true; case X86EMUL_OPC(0x0f, 0x01): @@ -6945,7 +7094,8 @@ x86_insn_is_mem_write(const struct x86_e switch ( state->desc & DstMask ) { case DstMem: - return state->modrm_mod != 3; + /* The SrcMem check is to cover {,V}MASKMOV{Q,DQU}. */ + return state->modrm_mod != 3 || (state->desc & SrcMask) == SrcMem; case DstBitBase: case DstImplicit: @@ -6965,22 +7115,9 @@ x86_insn_is_mem_write(const struct x86_e case 0x6c: case 0x6d: /* INS */ case 0xa4: case 0xa5: /* MOVS */ case 0xaa: case 0xab: /* STOS */ - case X86EMUL_OPC(0x0f, 0x7e): /* MOVD/MOVQ */ - case X86EMUL_OPC_66(0x0f, 0x7e): /* MOVD/MOVQ */ - case X86EMUL_OPC_VEX_66(0x0f, 0x7e): /* VMOVD/VMOVQ */ - case X86EMUL_OPC(0x0f, 0x7f): /* VMOVQ */ - case X86EMUL_OPC_66(0x0f, 0x7f): /* MOVDQA */ - case X86EMUL_OPC_VEX_66(0x0f, 0x7f): /* VMOVDQA */ - case X86EMUL_OPC_F3(0x0f, 0x7f): /* MOVDQU */ - case X86EMUL_OPC_VEX_F3(0x0f, 0x7f): /* VMOVDQU */ case X86EMUL_OPC(0x0f, 0xab): /* BTS */ case X86EMUL_OPC(0x0f, 0xb3): /* BTR */ case X86EMUL_OPC(0x0f, 0xbb): /* BTC */ - case X86EMUL_OPC_66(0x0f, 0xd6): /* MOVQ */ - case X86EMUL_OPC_VEX_66(0x0f, 0xd6): /* VMOVQ */ - case X86EMUL_OPC(0x0f, 0xe7): /* MOVNTQ */ - case X86EMUL_OPC_66(0x0f, 0xe7): /* MOVNTDQ */ - case X86EMUL_OPC_VEX_66(0x0f, 0xe7): /* VMOVNTDQ */ return true; case 0xd9: Attachment:
x86emul-SSE-AVX-0f-mov.patch _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx https://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |