+2019-07-01 Jan Beulich <jbeulich@suse.com>
+
+ * config/tc-i386.c (optimize_encoding): Make j unsigned. Handle
+ vpand{d,q}, vpandn{d,q}, vpor{d,q}, and vpxor{d,q}. Also check/
+ clear broadcast. Eliminate a loop.
+ * doc/c-i386.texi: Update -O1 documentation.
+ * testsuite/gas/i386/optimize-1.s,
+ testsuite/gas/i386/optimize-2.s,
+ testsuite/gas/i386/optimize-3.s,
+ testsuite/gas/i386/optimize-5.s,
+ testsuite/gas/i386/x86-64-optimize-2.s,
+ testsuite/gas/i386/x86-64-optimize-3.s,
+ testsuite/gas/i386/x86-64-optimize-4.s,
+ testsuite/gas/i386/x86-64-optimize-6.s: Add vpand{d,q},
+ vpandn{d,q}, vpor{d,q}, and vpxor{d,q} cases.
+ testsuite/gas/i386/optimize-1.d,
+ testsuite/gas/i386/optimize-1a.d,
+ testsuite/gas/i386/optimize-2.d,
+ testsuite/gas/i386/optimize-3.d,
+ testsuite/gas/i386/optimize-4.d,
+ testsuite/gas/i386/optimize-5.d,
+ testsuite/gas/i386/x86-64-optimize-2.d,
+ testsuite/gas/i386/x86-64-optimize-2a.d,
+ testsuite/gas/i386/x86-64-optimize-2b.d,
+ testsuite/gas/i386/x86-64-optimize-3.d,
+ testsuite/gas/i386/x86-64-optimize-4.d,
+ testsuite/gas/i386/x86-64-optimize-5.d,
+ testsuite/gas/i386/x86-64-optimize-6.d: Adjust expectations.
+
2019-07-01 Jan Beulich <jbeulich@suse.com>
* testsuite/gas/i386/avx512f_vpclmulqdq.s,
static void
optimize_encoding (void)
{
- int j;
+ unsigned int j;
if (optimize_for_space
&& i.reg_operands == 1
&& !i.types[0].bitfield.zmmword
&& !i.types[1].bitfield.zmmword
&& !i.mask
+ && !i.broadcast
&& is_evex_encoding (&i.tm)
&& ((i.tm.base_opcode & ~Opcode_SIMD_IntD) == 0x666f
|| (i.tm.base_opcode & ~Opcode_SIMD_IntD) == 0xf36f
- || (i.tm.base_opcode & ~Opcode_SIMD_IntD) == 0xf26f)
+ || (i.tm.base_opcode & ~Opcode_SIMD_IntD) == 0xf26f
+ || (i.tm.base_opcode & ~4) == 0x66db
+ || (i.tm.base_opcode & ~4) == 0x66eb)
&& i.tm.extension_opcode == None)
{
/* Optimize: -O1:
-> VEX mvmovdqa|vmovdquem, %xmmN (N < 16)
EVEX VOP mem, %ymmN
-> VEX vmovdqa|vmovdqu mem, %ymmN (N < 16)
+ VOP, one of vpand, vpandn, vpor, vpxor:
+ EVEX VOP{d,q} %xmmL, %xmmM, %xmmN
+ -> VEX VOP %xmmL, %xmmM, %xmmN (L, M, and N < 16)
+ EVEX VOP{d,q} %ymmL, %ymmM, %ymmN
+ -> VEX VOP %ymmL, %ymmM, %ymmN (L, M, and N < 16)
+ EVEX VOP{d,q} mem, %xmmM, %xmmN
+ -> VEX VOP mem, %xmmM, %xmmN (M and N < 16)
+ EVEX VOP{d,q} mem, %ymmM, %ymmN
+ -> VEX VOP mem, %ymmM, %ymmN (M and N < 16)
*/
- for (j = 0; j < 2; j++)
+ for (j = 0; j < i.operands; j++)
if (operand_type_check (i.types[j], disp)
&& i.op[j].disps->X_op == O_constant)
{
i.tm.opcode_modifier.vexw = VEXW0;
i.tm.opcode_modifier.evex = 0;
i.tm.opcode_modifier.masking = 0;
+ i.tm.opcode_modifier.broadcast = 0;
i.tm.opcode_modifier.disp8memshift = 0;
i.memshift = 0;
- for (j = 0; j < 2; j++)
- if (operand_type_check (i.types[j], disp)
- && i.op[j].disps->X_op == O_constant)
- {
- i.types[j].bitfield.disp8
- = fits_in_disp8 (i.op[j].disps->X_add_number);
- break;
- }
+ if (j < i.operands)
+ i.types[j].bitfield.disp8
+ = fits_in_disp8 (i.op[j].disps->X_add_number);
}
}
and @samp{-O1} encode 64-bit register load instructions with 64-bit
immediate as 32-bit register load instructions with 31-bit or 32-bits
immediates, encode 64-bit register clearing instructions with 32-bit
-register clearing instructions and encode 256-bit/512-bit VEX/EVEX
-vector register clearing instructions with 128-bit VEX vector register
-clearing instructions as well as encode 128-bit/256-bit EVEX vector
+register clearing instructions, encode 256-bit/512-bit VEX/EVEX vector
+register clearing instructions with 128-bit VEX vector register
+clearing instructions, encode 128-bit/256-bit EVEX vector
register load/store instructions with VEX vector register load/store
-instructions. @samp{-O2} includes @samp{-O1} optimization plus
-encodes 256-bit/512-bit EVEX vector register clearing instructions with
-128-bit EVEX vector register clearing instructions.
+instructions, and encode 128-bit/256-bit EVEX packed integer logical
+instructions with 128-bit/256-bit VEX packed integer logical.
+
+@samp{-O2} includes @samp{-O1} optimization plus encodes
+256-bit/512-bit EVEX vector register clearing instructions with 128-bit
+EVEX vector register clearing instructions.
+
@samp{-Os} includes @samp{-O2} optimization plus encodes 16-bit, 32-bit
and 64-bit register tests with immediate as 8-bit register test with
immediate. @samp{-O0} turns off this optimization.
+[a-f0-9]+: 62 f1 7e 28 7f 48 04 vmovdqu32 %ymm1,0x80\(%eax\)
+[a-f0-9]+: 62 f1 fe 28 7f 48 04 vmovdqu64 %ymm1,0x80\(%eax\)
+[a-f0-9]+: 62 f1 7d 48 6f 10 vmovdqa32 \(%eax\),%zmm2
+ +[a-f0-9]+: c5 .* vpand %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpand %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpandn %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpandn %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpor %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpor %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpxor %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpxor %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpand %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpand %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpandn %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpandn %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpor %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpor %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpxor %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpxor %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpand 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpand 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpandn 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpandn 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpor 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpor 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpxor 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpxor 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandd 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandq 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandnd 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandnq 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpord 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vporq 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpxord 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpxorq 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpand 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpand 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpandn 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpandn 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpor 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpor 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpxor 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpxor 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandd 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandq 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandnd 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandnq 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpord 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vporq 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpxord 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpxorq 0x80\(%eax\),%ymm2,%ymm3
#pass
vmovdqu64 %ymm1, 128(%eax)
vmovdqa32 (%eax), %zmm2
+
+ vpandd %xmm2, %xmm3, %xmm4
+ vpandq %xmm2, %xmm3, %xmm4
+ vpandnd %xmm2, %xmm3, %xmm4
+ vpandnq %xmm2, %xmm3, %xmm4
+ vpord %xmm2, %xmm3, %xmm4
+ vporq %xmm2, %xmm3, %xmm4
+ vpxord %xmm2, %xmm3, %xmm4
+ vpxorq %xmm2, %xmm3, %xmm4
+
+ vpandd %ymm2, %ymm3, %ymm4
+ vpandq %ymm2, %ymm3, %ymm4
+ vpandnd %ymm2, %ymm3, %ymm4
+ vpandnq %ymm2, %ymm3, %ymm4
+ vpord %ymm2, %ymm3, %ymm4
+ vporq %ymm2, %ymm3, %ymm4
+ vpxord %ymm2, %ymm3, %ymm4
+ vpxorq %ymm2, %ymm3, %ymm4
+
+ vpandd 112(%eax), %xmm2, %xmm3
+ vpandq 112(%eax), %xmm2, %xmm3
+ vpandnd 112(%eax), %xmm2, %xmm3
+ vpandnq 112(%eax), %xmm2, %xmm3
+ vpord 112(%eax), %xmm2, %xmm3
+ vporq 112(%eax), %xmm2, %xmm3
+ vpxord 112(%eax), %xmm2, %xmm3
+ vpxorq 112(%eax), %xmm2, %xmm3
+
+ vpandd 128(%eax), %xmm2, %xmm3
+ vpandq 128(%eax), %xmm2, %xmm3
+ vpandnd 128(%eax), %xmm2, %xmm3
+ vpandnq 128(%eax), %xmm2, %xmm3
+ vpord 128(%eax), %xmm2, %xmm3
+ vporq 128(%eax), %xmm2, %xmm3
+ vpxord 128(%eax), %xmm2, %xmm3
+ vpxorq 128(%eax), %xmm2, %xmm3
+
+ vpandd 96(%eax), %ymm2, %ymm3
+ vpandq 96(%eax), %ymm2, %ymm3
+ vpandnd 96(%eax), %ymm2, %ymm3
+ vpandnq 96(%eax), %ymm2, %ymm3
+ vpord 96(%eax), %ymm2, %ymm3
+ vporq 96(%eax), %ymm2, %ymm3
+ vpxord 96(%eax), %ymm2, %ymm3
+ vpxorq 96(%eax), %ymm2, %ymm3
+
+ vpandd 128(%eax), %ymm2, %ymm3
+ vpandq 128(%eax), %ymm2, %ymm3
+ vpandnd 128(%eax), %ymm2, %ymm3
+ vpandnq 128(%eax), %ymm2, %ymm3
+ vpord 128(%eax), %ymm2, %ymm3
+ vporq 128(%eax), %ymm2, %ymm3
+ vpxord 128(%eax), %ymm2, %ymm3
+ vpxorq 128(%eax), %ymm2, %ymm3
+[a-f0-9]+: 62 f1 7e 28 7f 48 04 vmovdqu32 %ymm1,0x80\(%eax\)
+[a-f0-9]+: 62 f1 fe 28 7f 48 04 vmovdqu64 %ymm1,0x80\(%eax\)
+[a-f0-9]+: 62 f1 7d 48 6f 10 vmovdqa32 \(%eax\),%zmm2
+ +[a-f0-9]+: c5 .* vpand %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpand %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpandn %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpandn %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpor %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpor %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpxor %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpxor %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpand %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpand %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpandn %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpandn %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpor %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpor %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpxor %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpxor %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpand 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpand 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpandn 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpandn 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpor 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpor 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpxor 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpxor 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandd 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandq 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandnd 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandnq 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpord 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vporq 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpxord 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpxorq 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpand 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpand 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpandn 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpandn 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpor 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpor 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpxor 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpxor 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandd 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandq 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandnd 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandnq 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpord 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vporq 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpxord 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpxorq 0x80\(%eax\),%ymm2,%ymm3
#pass
+[a-f0-9]+: 62 f1 ff 89 6f d1 vmovdqu16 %xmm1,%xmm2\{%k1\}\{z\}
+[a-f0-9]+: 62 f1 7e 89 6f d1 vmovdqu32 %xmm1,%xmm2\{%k1\}\{z\}
+[a-f0-9]+: 62 f1 fe 89 6f d1 vmovdqu64 %xmm1,%xmm2\{%k1\}\{z\}
+ +[a-f0-9]+: c5 .* vpand %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpand %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpandn %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpandn %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpor %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpor %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpxor %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpxor %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpand %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpand %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpandn %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpandn %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpor %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpor %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpxor %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpxor %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpand 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpand 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpandn 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpandn 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpor 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpor 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpxor 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpxor 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandd 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandq 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandnd 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandnq 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpord 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vporq 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpxord 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpxorq 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpand 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpand 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpandn 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpandn 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpor 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpor 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpxor 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpxor 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandd 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandq 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandnd 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandnq 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpord 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vporq 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpxord 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpxorq 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandd %xmm2,%xmm3,%xmm4\{%k5\}
+ +[a-f0-9]+: 62 .* vpandq %ymm2,%ymm3,%ymm4\{%k5\}
+ +[a-f0-9]+: 62 .* vpandnd %ymm2,%ymm3,%ymm4\{%k5\}
+ +[a-f0-9]+: 62 .* vpandnq %xmm2,%xmm3,%xmm4\{%k5\}
+ +[a-f0-9]+: 62 .* vpord %xmm2,%xmm3,%xmm4\{%k5\}
+ +[a-f0-9]+: 62 .* vporq %ymm2,%ymm3,%ymm4\{%k5\}
+ +[a-f0-9]+: 62 .* vpxord %ymm2,%ymm3,%ymm4\{%k5\}
+ +[a-f0-9]+: 62 .* vpxorq %xmm2,%xmm3,%xmm4\{%k5\}
+ +[a-f0-9]+: 62 .* vpandd \(%eax\)\{1to8\},%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandq \(%eax\)\{1to2\},%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandnd \(%eax\)\{1to4\},%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandnq \(%eax\)\{1to4\},%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpord \(%eax\)\{1to8\},%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vporq \(%eax\)\{1to2\},%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpxord \(%eax\)\{1to4\},%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpxorq \(%eax\)\{1to4\},%ymm2,%ymm3
#pass
vmovdqu16 %xmm1, %xmm2{%k1}{z}
vmovdqu32 %xmm1, %xmm2{%k1}{z}
vmovdqu64 %xmm1, %xmm2{%k1}{z}
+
+ vpandd %xmm2, %xmm3, %xmm4
+ vpandq %xmm2, %xmm3, %xmm4
+ vpandnd %xmm2, %xmm3, %xmm4
+ vpandnq %xmm2, %xmm3, %xmm4
+ vpord %xmm2, %xmm3, %xmm4
+ vporq %xmm2, %xmm3, %xmm4
+ vpxord %xmm2, %xmm3, %xmm4
+ vpxorq %xmm2, %xmm3, %xmm4
+
+ vpandd %ymm2, %ymm3, %ymm4
+ vpandq %ymm2, %ymm3, %ymm4
+ vpandnd %ymm2, %ymm3, %ymm4
+ vpandnq %ymm2, %ymm3, %ymm4
+ vpord %ymm2, %ymm3, %ymm4
+ vporq %ymm2, %ymm3, %ymm4
+ vpxord %ymm2, %ymm3, %ymm4
+ vpxorq %ymm2, %ymm3, %ymm4
+
+ vpandd 112(%eax), %xmm2, %xmm3
+ vpandq 112(%eax), %xmm2, %xmm3
+ vpandnd 112(%eax), %xmm2, %xmm3
+ vpandnq 112(%eax), %xmm2, %xmm3
+ vpord 112(%eax), %xmm2, %xmm3
+ vporq 112(%eax), %xmm2, %xmm3
+ vpxord 112(%eax), %xmm2, %xmm3
+ vpxorq 112(%eax), %xmm2, %xmm3
+
+ vpandd 128(%eax), %xmm2, %xmm3
+ vpandq 128(%eax), %xmm2, %xmm3
+ vpandnd 128(%eax), %xmm2, %xmm3
+ vpandnq 128(%eax), %xmm2, %xmm3
+ vpord 128(%eax), %xmm2, %xmm3
+ vporq 128(%eax), %xmm2, %xmm3
+ vpxord 128(%eax), %xmm2, %xmm3
+ vpxorq 128(%eax), %xmm2, %xmm3
+
+ vpandd 96(%eax), %ymm2, %ymm3
+ vpandq 96(%eax), %ymm2, %ymm3
+ vpandnd 96(%eax), %ymm2, %ymm3
+ vpandnq 96(%eax), %ymm2, %ymm3
+ vpord 96(%eax), %ymm2, %ymm3
+ vporq 96(%eax), %ymm2, %ymm3
+ vpxord 96(%eax), %ymm2, %ymm3
+ vpxorq 96(%eax), %ymm2, %ymm3
+
+ vpandd 128(%eax), %ymm2, %ymm3
+ vpandq 128(%eax), %ymm2, %ymm3
+ vpandnd 128(%eax), %ymm2, %ymm3
+ vpandnq 128(%eax), %ymm2, %ymm3
+ vpord 128(%eax), %ymm2, %ymm3
+ vporq 128(%eax), %ymm2, %ymm3
+ vpxord 128(%eax), %ymm2, %ymm3
+ vpxorq 128(%eax), %ymm2, %ymm3
+
+ vpandd %xmm2, %xmm3, %xmm4{%k5}
+ vpandq %ymm2, %ymm3, %ymm4{%k5}
+ vpandnd %ymm2, %ymm3, %ymm4{%k5}
+ vpandnq %xmm2, %xmm3, %xmm4{%k5}
+ vpord %xmm2, %xmm3, %xmm4{%k5}
+ vporq %ymm2, %ymm3, %ymm4{%k5}
+ vpxord %ymm2, %ymm3, %ymm4{%k5}
+ vpxorq %xmm2, %xmm3, %xmm4{%k5}
+
+ vpandd (%eax){1to8}, %ymm2, %ymm3
+ vpandq (%eax){1to2}, %xmm2, %xmm3
+ vpandnd (%eax){1to4}, %xmm2, %xmm3
+ vpandnq (%eax){1to4}, %ymm2, %ymm3
+ vpord (%eax){1to8}, %ymm2, %ymm3
+ vporq (%eax){1to2}, %xmm2, %xmm3
+ vpxord (%eax){1to4}, %xmm2, %xmm3
+ vpxorq (%eax){1to4}, %ymm2, %ymm3
+[a-f0-9]+: 62 f1 ff 08 6f d1 vmovdqu16 %xmm1,%xmm2
+[a-f0-9]+: 62 f1 7e 08 6f d1 vmovdqu32 %xmm1,%xmm2
+[a-f0-9]+: 62 f1 fe 08 6f d1 vmovdqu64 %xmm1,%xmm2
+ +[a-f0-9]+: 62 .* vpandd %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: 62 .* vpandq %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: 62 .* vpandnd %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: 62 .* vpandnq %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: 62 .* vpord %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: 62 .* vporq %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: 62 .* vpxord %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: 62 .* vpxorq %xmm2,%xmm3,%xmm4
#pass
{nooptimize} vmovdqu16 %xmm1, %xmm2
{nooptimize} vmovdqu32 %xmm1, %xmm2
{nooptimize} vmovdqu64 %xmm1, %xmm2
+
+ {nooptimize} vpandd %xmm2, %xmm3, %xmm4
+ {nooptimize} vpandq %ymm2, %ymm3, %ymm4
+ {nooptimize} vpandnd %ymm2, %ymm3, %ymm4
+ {nooptimize} vpandnq %xmm2, %xmm3, %xmm4
+ {nooptimize} vpord %xmm2, %xmm3, %xmm4
+ {nooptimize} vporq %ymm2, %ymm3, %ymm4
+ {nooptimize} vpxord %ymm2, %ymm3, %ymm4
+ {nooptimize} vpxorq %xmm2, %xmm3, %xmm4
+[a-f0-9]+: 62 f1 7e 28 7f 48 04 vmovdqu32 %ymm1,0x80\(%eax\)
+[a-f0-9]+: 62 f1 fe 28 7f 48 04 vmovdqu64 %ymm1,0x80\(%eax\)
+[a-f0-9]+: 62 f1 7d 48 6f 10 vmovdqa32 \(%eax\),%zmm2
+ +[a-f0-9]+: c5 .* vpand %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpand %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpandn %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpandn %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpor %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpor %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpxor %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpxor %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpand %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpand %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpandn %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpandn %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpor %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpor %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpxor %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpxor %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpand 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpand 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpandn 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpandn 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpor 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpor 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpxor 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpxor 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandd 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandq 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandnd 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandnq 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpord 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vporq 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpxord 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpxorq 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpand 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpand 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpandn 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpandn 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpor 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpor 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpxor 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpxor 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandd 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandq 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandnd 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandnq 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpord 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vporq 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpxord 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpxorq 0x80\(%eax\),%ymm2,%ymm3
+[a-f0-9]+: 62 f1 f5 08 55 e9 vandnpd %xmm1,%xmm1,%xmm5
+[a-f0-9]+: 62 f1 f5 08 55 e9 vandnpd %xmm1,%xmm1,%xmm5
#pass
+[a-f0-9]+: 62 f1 7e 28 7f 48 04 vmovdqu32 %ymm1,0x80\(%eax\)
+[a-f0-9]+: 62 f1 fe 28 7f 48 04 vmovdqu64 %ymm1,0x80\(%eax\)
+[a-f0-9]+: 62 f1 7d 48 6f 10 vmovdqa32 \(%eax\),%zmm2
+ +[a-f0-9]+: c5 .* vpand %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpand %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpandn %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpandn %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpor %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpor %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpxor %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpxor %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpand %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpand %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpandn %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpandn %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpor %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpor %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpxor %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpxor %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpand 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpand 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpandn 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpandn 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpor 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpor 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpxor 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpxor 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandd 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandq 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandnd 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandnq 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpord 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vporq 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpxord 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpxorq 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpand 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpand 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpandn 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpandn 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpor 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpor 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpxor 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpxor 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandd 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandq 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandnd 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandnq 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpord 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vporq 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpxord 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpxorq 0x80\(%eax\),%ymm2,%ymm3
+[a-f0-9]+: 62 f1 f5 08 55 e9 vandnpd %xmm1,%xmm1,%xmm5
+[a-f0-9]+: 62 f1 f5 08 55 e9 vandnpd %xmm1,%xmm1,%xmm5
+[a-f0-9]+: 62 f1 7d 28 6f d1 vmovdqa32 %ymm1,%ymm2
+[a-f0-9]+: 62 f1 ff 08 6f d1 vmovdqu16 %xmm1,%xmm2
+[a-f0-9]+: 62 f1 7e 08 6f d1 vmovdqu32 %xmm1,%xmm2
+[a-f0-9]+: 62 f1 fe 08 6f d1 vmovdqu64 %xmm1,%xmm2
+ +[a-f0-9]+: 62 .* vpandd %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: 62 .* vpandq %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: 62 .* vpandnd %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: 62 .* vpandnq %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: 62 .* vpord %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: 62 .* vporq %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: 62 .* vpxord %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: 62 .* vpxorq %xmm2,%xmm3,%xmm4
#pass
{evex} vmovdqu16 %xmm1, %xmm2
{evex} vmovdqu32 %xmm1, %xmm2
{evex} vmovdqu64 %xmm1, %xmm2
+
+ {evex} vpandd %xmm2, %xmm3, %xmm4
+ {evex} vpandq %ymm2, %ymm3, %ymm4
+ {evex} vpandnd %ymm2, %ymm3, %ymm4
+ {evex} vpandnq %xmm2, %xmm3, %xmm4
+ {evex} vpord %xmm2, %xmm3, %xmm4
+ {evex} vporq %ymm2, %ymm3, %ymm4
+ {evex} vpxord %ymm2, %ymm3, %ymm4
+ {evex} vpxorq %xmm2, %xmm3, %xmm4
+[a-f0-9]+: 62 f1 7e 28 7f 48 04 vmovdqu32 %ymm1,0x80\(%rax\)
+[a-f0-9]+: 62 f1 fe 28 7f 48 04 vmovdqu64 %ymm1,0x80\(%rax\)
+[a-f0-9]+: 62 f1 7d 48 6f 10 vmovdqa32 \(%rax\),%zmm2
+ +[a-f0-9]+: c5 .* vpand %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c4 .* vpand %xmm12,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpandn %xmm2,%xmm13,%xmm4
+ +[a-f0-9]+: c5 .* vpandn %xmm2,%xmm3,%xmm14
+ +[a-f0-9]+: c5 .* vpor %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c4 .* vpor %xmm12,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpxor %xmm2,%xmm13,%xmm4
+ +[a-f0-9]+: c5 .* vpxor %xmm2,%xmm3,%xmm14
+ +[a-f0-9]+: c5 .* vpand %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c4 .* vpand %ymm12,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpandn %ymm2,%ymm13,%ymm4
+ +[a-f0-9]+: c5 .* vpandn %ymm2,%ymm3,%ymm14
+ +[a-f0-9]+: c5 .* vpor %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c4 .* vpor %ymm12,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpxor %ymm2,%ymm13,%ymm4
+ +[a-f0-9]+: c5 .* vpxor %ymm2,%ymm3,%ymm14
+ +[a-f0-9]+: c5 .* vpand 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpand 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpandn 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpandn 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpor 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpor 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpxor 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpxor 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandd 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandq 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandnd 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandnq 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpord 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vporq 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpxord 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpxorq 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpand 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpand 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpandn 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpandn 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpor 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpor 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpxor 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpxor 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandd 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandq 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandnd 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandnq 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpord 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vporq 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpxord 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpxorq 0x80\(%rax\),%ymm2,%ymm3
#pass
vmovdqu64 %ymm1, 128(%rax)
vmovdqa32 (%rax), %zmm2
+
+ vpandd %xmm2, %xmm3, %xmm4
+ vpandq %xmm12, %xmm3, %xmm4
+ vpandnd %xmm2, %xmm13, %xmm4
+ vpandnq %xmm2, %xmm3, %xmm14
+ vpord %xmm2, %xmm3, %xmm4
+ vporq %xmm12, %xmm3, %xmm4
+ vpxord %xmm2, %xmm13, %xmm4
+ vpxorq %xmm2, %xmm3, %xmm14
+
+ vpandd %ymm2, %ymm3, %ymm4
+ vpandq %ymm12, %ymm3, %ymm4
+ vpandnd %ymm2, %ymm13, %ymm4
+ vpandnq %ymm2, %ymm3, %ymm14
+ vpord %ymm2, %ymm3, %ymm4
+ vporq %ymm12, %ymm3, %ymm4
+ vpxord %ymm2, %ymm13, %ymm4
+ vpxorq %ymm2, %ymm3, %ymm14
+
+ vpandd 112(%rax), %xmm2, %xmm3
+ vpandq 112(%rax), %xmm2, %xmm3
+ vpandnd 112(%rax), %xmm2, %xmm3
+ vpandnq 112(%rax), %xmm2, %xmm3
+ vpord 112(%rax), %xmm2, %xmm3
+ vporq 112(%rax), %xmm2, %xmm3
+ vpxord 112(%rax), %xmm2, %xmm3
+ vpxorq 112(%rax), %xmm2, %xmm3
+
+ vpandd 128(%rax), %xmm2, %xmm3
+ vpandq 128(%rax), %xmm2, %xmm3
+ vpandnd 128(%rax), %xmm2, %xmm3
+ vpandnq 128(%rax), %xmm2, %xmm3
+ vpord 128(%rax), %xmm2, %xmm3
+ vporq 128(%rax), %xmm2, %xmm3
+ vpxord 128(%rax), %xmm2, %xmm3
+ vpxorq 128(%rax), %xmm2, %xmm3
+
+ vpandd 96(%rax), %ymm2, %ymm3
+ vpandq 96(%rax), %ymm2, %ymm3
+ vpandnd 96(%rax), %ymm2, %ymm3
+ vpandnq 96(%rax), %ymm2, %ymm3
+ vpord 96(%rax), %ymm2, %ymm3
+ vporq 96(%rax), %ymm2, %ymm3
+ vpxord 96(%rax), %ymm2, %ymm3
+ vpxorq 96(%rax), %ymm2, %ymm3
+
+ vpandd 128(%rax), %ymm2, %ymm3
+ vpandq 128(%rax), %ymm2, %ymm3
+ vpandnd 128(%rax), %ymm2, %ymm3
+ vpandnq 128(%rax), %ymm2, %ymm3
+ vpord 128(%rax), %ymm2, %ymm3
+ vporq 128(%rax), %ymm2, %ymm3
+ vpxord 128(%rax), %ymm2, %ymm3
+ vpxorq 128(%rax), %ymm2, %ymm3
+[a-f0-9]+: 62 f1 7e 28 7f 48 04 vmovdqu32 %ymm1,0x80\(%rax\)
+[a-f0-9]+: 62 f1 fe 28 7f 48 04 vmovdqu64 %ymm1,0x80\(%rax\)
+[a-f0-9]+: 62 f1 7d 48 6f 10 vmovdqa32 \(%rax\),%zmm2
+ +[a-f0-9]+: c5 .* vpand %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c4 .* vpand %xmm12,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpandn %xmm2,%xmm13,%xmm4
+ +[a-f0-9]+: c5 .* vpandn %xmm2,%xmm3,%xmm14
+ +[a-f0-9]+: c5 .* vpor %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c4 .* vpor %xmm12,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpxor %xmm2,%xmm13,%xmm4
+ +[a-f0-9]+: c5 .* vpxor %xmm2,%xmm3,%xmm14
+ +[a-f0-9]+: c5 .* vpand %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c4 .* vpand %ymm12,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpandn %ymm2,%ymm13,%ymm4
+ +[a-f0-9]+: c5 .* vpandn %ymm2,%ymm3,%ymm14
+ +[a-f0-9]+: c5 .* vpor %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c4 .* vpor %ymm12,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpxor %ymm2,%ymm13,%ymm4
+ +[a-f0-9]+: c5 .* vpxor %ymm2,%ymm3,%ymm14
+ +[a-f0-9]+: c5 .* vpand 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpand 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpandn 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpandn 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpor 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpor 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpxor 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpxor 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandd 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandq 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandnd 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandnq 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpord 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vporq 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpxord 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpxorq 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpand 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpand 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpandn 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpandn 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpor 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpor 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpxor 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpxor 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandd 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandq 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandnd 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandnq 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpord 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vporq 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpxord 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpxorq 0x80\(%rax\),%ymm2,%ymm3
#pass
+[a-f0-9]+: 62 f1 7e 28 7f 48 04 vmovdqu32 %ymm1,0x80\(%rax\)
+[a-f0-9]+: 62 f1 fe 28 7f 48 04 vmovdqu64 %ymm1,0x80\(%rax\)
+[a-f0-9]+: 62 f1 7d 48 6f 10 vmovdqa32 \(%rax\),%zmm2
+ +[a-f0-9]+: c5 .* vpand %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c4 .* vpand %xmm12,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpandn %xmm2,%xmm13,%xmm4
+ +[a-f0-9]+: c5 .* vpandn %xmm2,%xmm3,%xmm14
+ +[a-f0-9]+: c5 .* vpor %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c4 .* vpor %xmm12,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpxor %xmm2,%xmm13,%xmm4
+ +[a-f0-9]+: c5 .* vpxor %xmm2,%xmm3,%xmm14
+ +[a-f0-9]+: c5 .* vpand %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c4 .* vpand %ymm12,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpandn %ymm2,%ymm13,%ymm4
+ +[a-f0-9]+: c5 .* vpandn %ymm2,%ymm3,%ymm14
+ +[a-f0-9]+: c5 .* vpor %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c4 .* vpor %ymm12,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpxor %ymm2,%ymm13,%ymm4
+ +[a-f0-9]+: c5 .* vpxor %ymm2,%ymm3,%ymm14
+ +[a-f0-9]+: c5 .* vpand 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpand 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpandn 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpandn 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpor 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpor 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpxor 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpxor 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandd 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandq 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandnd 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandnq 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpord 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vporq 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpxord 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpxorq 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpand 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpand 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpandn 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpandn 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpor 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpor 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpxor 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpxor 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandd 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandq 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandnd 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandnq 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpord 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vporq 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpxord 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpxorq 0x80\(%rax\),%ymm2,%ymm3
#pass
+[a-f0-9]+: 62 f1 ff 89 6f d1 vmovdqu16 %xmm1,%xmm2\{%k1\}\{z\}
+[a-f0-9]+: 62 f1 7e 89 6f d1 vmovdqu32 %xmm1,%xmm2\{%k1\}\{z\}
+[a-f0-9]+: 62 f1 fe 89 6f d1 vmovdqu64 %xmm1,%xmm2\{%k1\}\{z\}
+ +[a-f0-9]+: c5 .* vpand %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c4 .* vpand %xmm12,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpandn %xmm2,%xmm13,%xmm4
+ +[a-f0-9]+: c5 .* vpandn %xmm2,%xmm3,%xmm14
+ +[a-f0-9]+: c5 .* vpor %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c4 .* vpor %xmm12,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpxor %xmm2,%xmm13,%xmm4
+ +[a-f0-9]+: c5 .* vpxor %xmm2,%xmm3,%xmm14
+ +[a-f0-9]+: c5 .* vpand %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c4 .* vpand %ymm12,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpandn %ymm2,%ymm13,%ymm4
+ +[a-f0-9]+: c5 .* vpandn %ymm2,%ymm3,%ymm14
+ +[a-f0-9]+: c5 .* vpor %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c4 .* vpor %ymm12,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpxor %ymm2,%ymm13,%ymm4
+ +[a-f0-9]+: c5 .* vpxor %ymm2,%ymm3,%ymm14
+ +[a-f0-9]+: c5 .* vpand 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpand 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpandn 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpandn 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpor 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpor 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpxor 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpxor 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandd 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandq 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandnd 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandnq 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpord 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vporq 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpxord 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpxorq 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpand 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpand 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpandn 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpandn 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpor 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpor 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpxor 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpxor 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandd 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandq 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandnd 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandnq 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpord 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vporq 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpxord 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpxorq 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandd %xmm22,%xmm23,%xmm24
+ +[a-f0-9]+: 62 .* vpandq %ymm22,%ymm3,%ymm4
+ +[a-f0-9]+: 62 .* vpandnd %ymm2,%ymm23,%ymm4
+ +[a-f0-9]+: 62 .* vpandnq %xmm2,%xmm3,%xmm24
+ +[a-f0-9]+: 62 .* vpord %xmm22,%xmm23,%xmm24
+ +[a-f0-9]+: 62 .* vporq %ymm22,%ymm3,%ymm4
+ +[a-f0-9]+: 62 .* vpxord %ymm2,%ymm23,%ymm4
+ +[a-f0-9]+: 62 .* vpxorq %xmm2,%xmm3,%xmm24
+ +[a-f0-9]+: 62 .* vpandd %xmm2,%xmm3,%xmm4\{%k5\}
+ +[a-f0-9]+: 62 .* vpandq %ymm12,%ymm3,%ymm4\{%k5\}
+ +[a-f0-9]+: 62 .* vpandnd %ymm2,%ymm13,%ymm4\{%k5\}
+ +[a-f0-9]+: 62 .* vpandnq %xmm2,%xmm3,%xmm14\{%k5\}
+ +[a-f0-9]+: 62 .* vpord %xmm2,%xmm3,%xmm4\{%k5\}
+ +[a-f0-9]+: 62 .* vporq %ymm12,%ymm3,%ymm4\{%k5\}
+ +[a-f0-9]+: 62 .* vpxord %ymm2,%ymm13,%ymm4\{%k5\}
+ +[a-f0-9]+: 62 .* vpxorq %xmm2,%xmm3,%xmm14\{%k5\}
+ +[a-f0-9]+: 62 .* vpandd \(%rax\)\{1to8\},%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandq \(%rax\)\{1to2\},%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandnd \(%rax\)\{1to4\},%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandnq \(%rax\)\{1to4\},%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpord \(%rax\)\{1to8\},%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vporq \(%rax\)\{1to2\},%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpxord \(%rax\)\{1to4\},%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpxorq \(%rax\)\{1to4\},%ymm2,%ymm3
#pass
vmovdqu16 %xmm1, %xmm2{%k1}{z}
vmovdqu32 %xmm1, %xmm2{%k1}{z}
vmovdqu64 %xmm1, %xmm2{%k1}{z}
+
+ vpandd %xmm2, %xmm3, %xmm4
+ vpandq %xmm12, %xmm3, %xmm4
+ vpandnd %xmm2, %xmm13, %xmm4
+ vpandnq %xmm2, %xmm3, %xmm14
+ vpord %xmm2, %xmm3, %xmm4
+ vporq %xmm12, %xmm3, %xmm4
+ vpxord %xmm2, %xmm13, %xmm4
+ vpxorq %xmm2, %xmm3, %xmm14
+
+ vpandd %ymm2, %ymm3, %ymm4
+ vpandq %ymm12, %ymm3, %ymm4
+ vpandnd %ymm2, %ymm13, %ymm4
+ vpandnq %ymm2, %ymm3, %ymm14
+ vpord %ymm2, %ymm3, %ymm4
+ vporq %ymm12, %ymm3, %ymm4
+ vpxord %ymm2, %ymm13, %ymm4
+ vpxorq %ymm2, %ymm3, %ymm14
+
+ vpandd 112(%rax), %xmm2, %xmm3
+ vpandq 112(%rax), %xmm2, %xmm3
+ vpandnd 112(%rax), %xmm2, %xmm3
+ vpandnq 112(%rax), %xmm2, %xmm3
+ vpord 112(%rax), %xmm2, %xmm3
+ vporq 112(%rax), %xmm2, %xmm3
+ vpxord 112(%rax), %xmm2, %xmm3
+ vpxorq 112(%rax), %xmm2, %xmm3
+
+ vpandd 128(%rax), %xmm2, %xmm3
+ vpandq 128(%rax), %xmm2, %xmm3
+ vpandnd 128(%rax), %xmm2, %xmm3
+ vpandnq 128(%rax), %xmm2, %xmm3
+ vpord 128(%rax), %xmm2, %xmm3
+ vporq 128(%rax), %xmm2, %xmm3
+ vpxord 128(%rax), %xmm2, %xmm3
+ vpxorq 128(%rax), %xmm2, %xmm3
+
+ vpandd 96(%rax), %ymm2, %ymm3
+ vpandq 96(%rax), %ymm2, %ymm3
+ vpandnd 96(%rax), %ymm2, %ymm3
+ vpandnq 96(%rax), %ymm2, %ymm3
+ vpord 96(%rax), %ymm2, %ymm3
+ vporq 96(%rax), %ymm2, %ymm3
+ vpxord 96(%rax), %ymm2, %ymm3
+ vpxorq 96(%rax), %ymm2, %ymm3
+
+ vpandd 128(%rax), %ymm2, %ymm3
+ vpandq 128(%rax), %ymm2, %ymm3
+ vpandnd 128(%rax), %ymm2, %ymm3
+ vpandnq 128(%rax), %ymm2, %ymm3
+ vpord 128(%rax), %ymm2, %ymm3
+ vporq 128(%rax), %ymm2, %ymm3
+ vpxord 128(%rax), %ymm2, %ymm3
+ vpxorq 128(%rax), %ymm2, %ymm3
+
+ vpandd %xmm22, %xmm23, %xmm24
+ vpandq %ymm22, %ymm3, %ymm4
+ vpandnd %ymm2, %ymm23, %ymm4
+ vpandnq %xmm2, %xmm3, %xmm24
+ vpord %xmm22, %xmm23, %xmm24
+ vporq %ymm22, %ymm3, %ymm4
+ vpxord %ymm2, %ymm23, %ymm4
+ vpxorq %xmm2, %xmm3, %xmm24
+
+ vpandd %xmm2, %xmm3, %xmm4{%k5}
+ vpandq %ymm12, %ymm3, %ymm4{%k5}
+ vpandnd %ymm2, %ymm13, %ymm4{%k5}
+ vpandnq %xmm2, %xmm3, %xmm14{%k5}
+ vpord %xmm2, %xmm3, %xmm4{%k5}
+ vporq %ymm12, %ymm3, %ymm4{%k5}
+ vpxord %ymm2, %ymm13, %ymm4{%k5}
+ vpxorq %xmm2, %xmm3, %xmm14{%k5}
+
+ vpandd (%rax){1to8}, %ymm2, %ymm3
+ vpandq (%rax){1to2}, %xmm2, %xmm3
+ vpandnd (%rax){1to4}, %xmm2, %xmm3
+ vpandnq (%rax){1to4}, %ymm2, %ymm3
+ vpord (%rax){1to8}, %ymm2, %ymm3
+ vporq (%rax){1to2}, %xmm2, %xmm3
+ vpxord (%rax){1to4}, %xmm2, %xmm3
+ vpxorq (%rax){1to4}, %ymm2, %ymm3
+[a-f0-9]+: 62 f1 ff 08 6f d1 vmovdqu16 %xmm1,%xmm2
+[a-f0-9]+: 62 f1 7e 08 6f d1 vmovdqu32 %xmm1,%xmm2
+[a-f0-9]+: 62 f1 fe 08 6f d1 vmovdqu64 %xmm1,%xmm2
+ +[a-f0-9]+: 62 .* vpandd %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: 62 .* vpandq %ymm12,%ymm3,%ymm4
+ +[a-f0-9]+: 62 .* vpandnd %ymm2,%ymm13,%ymm4
+ +[a-f0-9]+: 62 .* vpandnq %xmm2,%xmm3,%xmm14
+ +[a-f0-9]+: 62 .* vpord %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: 62 .* vporq %ymm12,%ymm3,%ymm4
+ +[a-f0-9]+: 62 .* vpxord %ymm2,%ymm13,%ymm4
+ +[a-f0-9]+: 62 .* vpxorq %xmm2,%xmm3,%xmm14
#pass
{nooptimize} vmovdqu16 %xmm1, %xmm2
{nooptimize} vmovdqu32 %xmm1, %xmm2
{nooptimize} vmovdqu64 %xmm1, %xmm2
+
+ {nooptimize} vpandd %xmm2, %xmm3, %xmm4
+ {nooptimize} vpandq %ymm12, %ymm3, %ymm4
+ {nooptimize} vpandnd %ymm2, %ymm13, %ymm4
+ {nooptimize} vpandnq %xmm2, %xmm3, %xmm14
+ {nooptimize} vpord %xmm2, %xmm3, %xmm4
+ {nooptimize} vporq %ymm12, %ymm3, %ymm4
+ {nooptimize} vpxord %ymm2, %ymm13, %ymm4
+ {nooptimize} vpxorq %xmm2, %xmm3, %xmm14
+[a-f0-9]+: 62 f1 7e 28 7f 48 04 vmovdqu32 %ymm1,0x80\(%rax\)
+[a-f0-9]+: 62 f1 fe 28 7f 48 04 vmovdqu64 %ymm1,0x80\(%rax\)
+[a-f0-9]+: 62 f1 7d 48 6f 10 vmovdqa32 \(%rax\),%zmm2
+ +[a-f0-9]+: c5 .* vpand %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c4 .* vpand %xmm12,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpandn %xmm2,%xmm13,%xmm4
+ +[a-f0-9]+: c5 .* vpandn %xmm2,%xmm3,%xmm14
+ +[a-f0-9]+: c5 .* vpor %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c4 .* vpor %xmm12,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpxor %xmm2,%xmm13,%xmm4
+ +[a-f0-9]+: c5 .* vpxor %xmm2,%xmm3,%xmm14
+ +[a-f0-9]+: c5 .* vpand %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c4 .* vpand %ymm12,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpandn %ymm2,%ymm13,%ymm4
+ +[a-f0-9]+: c5 .* vpandn %ymm2,%ymm3,%ymm14
+ +[a-f0-9]+: c5 .* vpor %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c4 .* vpor %ymm12,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpxor %ymm2,%ymm13,%ymm4
+ +[a-f0-9]+: c5 .* vpxor %ymm2,%ymm3,%ymm14
+ +[a-f0-9]+: c5 .* vpand 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpand 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpandn 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpandn 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpor 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpor 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpxor 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpxor 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandd 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandq 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandnd 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandnq 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpord 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vporq 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpxord 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpxorq 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpand 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpand 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpandn 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpandn 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpor 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpor 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpxor 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpxor 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandd 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandq 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandnd 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandnq 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpord 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vporq 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpxord 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpxorq 0x80\(%rax\),%ymm2,%ymm3
+[a-f0-9]+: 62 f1 f5 08 55 e9 vandnpd %xmm1,%xmm1,%xmm5
+[a-f0-9]+: 62 f1 f5 08 55 e9 vandnpd %xmm1,%xmm1,%xmm5
+[a-f0-9]+: 62 f1 7d 28 6f d1 vmovdqa32 %ymm1,%ymm2
+[a-f0-9]+: 62 f1 7e 28 7f 48 04 vmovdqu32 %ymm1,0x80\(%rax\)
+[a-f0-9]+: 62 f1 fe 28 7f 48 04 vmovdqu64 %ymm1,0x80\(%rax\)
+[a-f0-9]+: 62 f1 7d 48 6f 10 vmovdqa32 \(%rax\),%zmm2
+ +[a-f0-9]+: c5 .* vpand %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c4 .* vpand %xmm12,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpandn %xmm2,%xmm13,%xmm4
+ +[a-f0-9]+: c5 .* vpandn %xmm2,%xmm3,%xmm14
+ +[a-f0-9]+: c5 .* vpor %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c4 .* vpor %xmm12,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpxor %xmm2,%xmm13,%xmm4
+ +[a-f0-9]+: c5 .* vpxor %xmm2,%xmm3,%xmm14
+ +[a-f0-9]+: c5 .* vpand %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c4 .* vpand %ymm12,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpandn %ymm2,%ymm13,%ymm4
+ +[a-f0-9]+: c5 .* vpandn %ymm2,%ymm3,%ymm14
+ +[a-f0-9]+: c5 .* vpor %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c4 .* vpor %ymm12,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpxor %ymm2,%ymm13,%ymm4
+ +[a-f0-9]+: c5 .* vpxor %ymm2,%ymm3,%ymm14
+ +[a-f0-9]+: c5 .* vpand 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpand 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpandn 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpandn 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpor 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpor 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpxor 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpxor 0x70\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandd 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandq 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandnd 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandnq 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpord 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vporq 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpxord 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpxorq 0x80\(%rax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpand 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpand 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpandn 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpandn 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpor 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpor 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpxor 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpxor 0x60\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandd 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandq 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandnd 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandnq 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpord 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vporq 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpxord 0x80\(%rax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpxorq 0x80\(%rax\),%ymm2,%ymm3
+[a-f0-9]+: 62 f1 f5 08 55 e9 vandnpd %xmm1,%xmm1,%xmm5
+[a-f0-9]+: 62 f1 f5 08 55 e9 vandnpd %xmm1,%xmm1,%xmm5
+[a-f0-9]+: 62 f1 7d 28 6f d1 vmovdqa32 %ymm1,%ymm2
+[a-f0-9]+: 62 f1 ff 08 6f d1 vmovdqu16 %xmm1,%xmm2
+[a-f0-9]+: 62 f1 7e 08 6f d1 vmovdqu32 %xmm1,%xmm2
+[a-f0-9]+: 62 f1 fe 08 6f d1 vmovdqu64 %xmm1,%xmm2
+ +[a-f0-9]+: 62 .* vpandd %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: 62 .* vpandq %ymm12,%ymm3,%ymm4
+ +[a-f0-9]+: 62 .* vpandnd %ymm2,%ymm13,%ymm4
+ +[a-f0-9]+: 62 .* vpandnq %xmm2,%xmm3,%xmm14
+ +[a-f0-9]+: 62 .* vpord %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: 62 .* vporq %ymm12,%ymm3,%ymm4
+ +[a-f0-9]+: 62 .* vpxord %ymm2,%ymm13,%ymm4
+ +[a-f0-9]+: 62 .* vpxorq %xmm2,%xmm3,%xmm14
#pass
{evex} vmovdqu16 %xmm1, %xmm2
{evex} vmovdqu32 %xmm1, %xmm2
{evex} vmovdqu64 %xmm1, %xmm2
+
+ {evex} vpandd %xmm2, %xmm3, %xmm4
+ {evex} vpandq %ymm12, %ymm3, %ymm4
+ {evex} vpandnd %ymm2, %ymm13, %ymm4
+ {evex} vpandnq %xmm2, %xmm3, %xmm14
+ {evex} vpord %xmm2, %xmm3, %xmm4
+ {evex} vporq %ymm12, %ymm3, %ymm4
+ {evex} vpxord %ymm2, %ymm13, %ymm4
+ {evex} vpxorq %xmm2, %xmm3, %xmm14
+2019-07-01 Jan Beulich <jbeulich@suse.com>
+
+ * i386-opc.tbl (and, or): Add Optimize to forms allowing two
+ register operands.
+ * i386-tbl.h: Re-generate.
+
2019-07-01 Jan Beulich <jbeulich@suse.com>
* i386-dis-evex-prefix.h: Use PCLMUL for vpclmulqdq.
vrsqrt14pd, 2, 0x664E, None, 1, CpuAVX512F, Modrm|Masking=3|VexOpcode=1|VexW=2|Broadcast|Disp8ShiftVL|CheckRegSize|No_bSuf|No_wSuf|No_lSuf|No_sSuf|No_qSuf|No_ldSuf, { RegXMM|RegYMM|RegZMM|Qword|Unspecified|BaseIndex, RegXMM|RegYMM|RegZMM }
vpaddd, 3, 0x66FE, None, 1, CpuAVX512F, Modrm|Masking=3|VexOpcode=0|VexVVVV=1|VexW=1|Broadcast|Disp8ShiftVL|CheckRegSize|No_bSuf|No_wSuf|No_lSuf|No_sSuf|No_qSuf|No_ldSuf, { RegXMM|RegYMM|RegZMM|Dword|Unspecified|BaseIndex, RegXMM|RegYMM|RegZMM, RegXMM|RegYMM|RegZMM }
-vpandd, 3, 0x66DB, None, 1, CpuAVX512F, Modrm|Masking=3|VexOpcode=0|VexVVVV=1|VexW=1|Broadcast|Disp8ShiftVL|CheckRegSize|No_bSuf|No_wSuf|No_lSuf|No_sSuf|No_qSuf|No_ldSuf, { RegXMM|RegYMM|RegZMM|Dword|Unspecified|BaseIndex, RegXMM|RegYMM|RegZMM, RegXMM|RegYMM|RegZMM }
+vpandd, 3, 0x66DB, None, 1, CpuAVX512F, Modrm|Masking=3|VexOpcode=0|VexVVVV=1|VexW=1|Broadcast|Disp8ShiftVL|CheckRegSize|No_bSuf|No_wSuf|No_lSuf|No_sSuf|No_qSuf|No_ldSuf|Optimize, { RegXMM|RegYMM|RegZMM|Dword|Unspecified|BaseIndex, RegXMM|RegYMM|RegZMM, RegXMM|RegYMM|RegZMM }
vpandnd, 3, 0x66DF, None, 1, CpuAVX512F, Modrm|Masking=3|VexOpcode=0|VexVVVV=1|VexW=1|Broadcast|Disp8ShiftVL|CheckRegSize|No_bSuf|No_wSuf|No_lSuf|No_sSuf|No_qSuf|No_ldSuf|Optimize, { RegXMM|RegYMM|RegZMM|Dword|Unspecified|BaseIndex, RegXMM|RegYMM|RegZMM, RegXMM|RegYMM|RegZMM }
-vpord, 3, 0x66EB, None, 1, CpuAVX512F, Modrm|Masking=3|VexOpcode=0|VexVVVV=1|VexW=1|Broadcast|Disp8ShiftVL|CheckRegSize|No_bSuf|No_wSuf|No_lSuf|No_sSuf|No_qSuf|No_ldSuf, { RegXMM|RegYMM|RegZMM|Dword|Unspecified|BaseIndex, RegXMM|RegYMM|RegZMM, RegXMM|RegYMM|RegZMM }
+vpord, 3, 0x66EB, None, 1, CpuAVX512F, Modrm|Masking=3|VexOpcode=0|VexVVVV=1|VexW=1|Broadcast|Disp8ShiftVL|CheckRegSize|No_bSuf|No_wSuf|No_lSuf|No_sSuf|No_qSuf|No_ldSuf|Optimize, { RegXMM|RegYMM|RegZMM|Dword|Unspecified|BaseIndex, RegXMM|RegYMM|RegZMM, RegXMM|RegYMM|RegZMM }
vpsubd, 3, 0x66FA, None, 1, CpuAVX512F, Modrm|Masking=3|VexOpcode=0|VexVVVV=1|VexW=1|Broadcast|Disp8ShiftVL|CheckRegSize|No_bSuf|No_wSuf|No_lSuf|No_sSuf|No_qSuf|No_ldSuf|Optimize, { RegXMM|RegYMM|RegZMM|Dword|Unspecified|BaseIndex, RegXMM|RegYMM|RegZMM, RegXMM|RegYMM|RegZMM }
vpunpckhdq, 3, 0x666A, None, 1, CpuAVX512F, Modrm|Masking=3|VexOpcode=0|VexVVVV=1|VexW=1|Broadcast|Disp8ShiftVL|CheckRegSize|No_bSuf|No_wSuf|No_lSuf|No_sSuf|No_qSuf|No_ldSuf, { RegXMM|RegYMM|RegZMM|Dword|Unspecified|BaseIndex, RegXMM|RegYMM|RegZMM, RegXMM|RegYMM|RegZMM }
vpunpckldq, 3, 0x6662, None, 1, CpuAVX512F, Modrm|Masking=3|VexOpcode=0|VexVVVV=1|VexW=1|Broadcast|Disp8ShiftVL|CheckRegSize|No_bSuf|No_wSuf|No_lSuf|No_sSuf|No_qSuf|No_ldSuf, { RegXMM|RegYMM|RegZMM|Dword|Unspecified|BaseIndex, RegXMM|RegYMM|RegZMM, RegXMM|RegYMM|RegZMM }
vpaddq, 3, 0x66D4, None, 1, CpuAVX512F, Modrm|Masking=3|VexOpcode=0|VexVVVV=1|VexW=2|Broadcast|Disp8ShiftVL|CheckRegSize|No_bSuf|No_wSuf|No_lSuf|No_sSuf|No_qSuf|No_ldSuf, { RegXMM|RegYMM|RegZMM|Qword|Unspecified|BaseIndex, RegXMM|RegYMM|RegZMM, RegXMM|RegYMM|RegZMM }
vpandnq, 3, 0x66DF, None, 1, CpuAVX512F, Modrm|Masking=3|VexOpcode=0|VexVVVV=1|VexW=2|Broadcast|Disp8ShiftVL|CheckRegSize|No_bSuf|No_wSuf|No_lSuf|No_sSuf|No_qSuf|No_ldSuf|Optimize, { RegXMM|RegYMM|RegZMM|Qword|Unspecified|BaseIndex, RegXMM|RegYMM|RegZMM, RegXMM|RegYMM|RegZMM }
-vpandq, 3, 0x66DB, None, 1, CpuAVX512F, Modrm|Masking=3|VexOpcode=0|VexVVVV=1|VexW=2|Broadcast|Disp8ShiftVL|CheckRegSize|No_bSuf|No_wSuf|No_lSuf|No_sSuf|No_qSuf|No_ldSuf, { RegXMM|RegYMM|RegZMM|Qword|Unspecified|BaseIndex, RegXMM|RegYMM|RegZMM, RegXMM|RegYMM|RegZMM }
+vpandq, 3, 0x66DB, None, 1, CpuAVX512F, Modrm|Masking=3|VexOpcode=0|VexVVVV=1|VexW=2|Broadcast|Disp8ShiftVL|CheckRegSize|No_bSuf|No_wSuf|No_lSuf|No_sSuf|No_qSuf|No_ldSuf|Optimize, { RegXMM|RegYMM|RegZMM|Qword|Unspecified|BaseIndex, RegXMM|RegYMM|RegZMM, RegXMM|RegYMM|RegZMM }
vpmuludq, 3, 0x66F4, None, 1, CpuAVX512F, Modrm|Masking=3|VexOpcode=0|VexVVVV=1|VexW=2|Broadcast|Disp8ShiftVL|CheckRegSize|No_bSuf|No_wSuf|No_lSuf|No_sSuf|No_qSuf|No_ldSuf, { RegXMM|RegYMM|RegZMM|Qword|Unspecified|BaseIndex, RegXMM|RegYMM|RegZMM, RegXMM|RegYMM|RegZMM }
-vporq, 3, 0x66EB, None, 1, CpuAVX512F, Modrm|Masking=3|VexOpcode=0|VexVVVV=1|VexW=2|Broadcast|Disp8ShiftVL|CheckRegSize|No_bSuf|No_wSuf|No_lSuf|No_sSuf|No_qSuf|No_ldSuf, { RegXMM|RegYMM|RegZMM|Qword|Unspecified|BaseIndex, RegXMM|RegYMM|RegZMM, RegXMM|RegYMM|RegZMM }
+vporq, 3, 0x66EB, None, 1, CpuAVX512F, Modrm|Masking=3|VexOpcode=0|VexVVVV=1|VexW=2|Broadcast|Disp8ShiftVL|CheckRegSize|No_bSuf|No_wSuf|No_lSuf|No_sSuf|No_qSuf|No_ldSuf|Optimize, { RegXMM|RegYMM|RegZMM|Qword|Unspecified|BaseIndex, RegXMM|RegYMM|RegZMM, RegXMM|RegYMM|RegZMM }
vpsubq, 3, 0x66FB, None, 1, CpuAVX512F, Modrm|Masking=3|VexOpcode=0|VexVVVV=1|VexW=2|Broadcast|Disp8ShiftVL|CheckRegSize|No_bSuf|No_wSuf|No_lSuf|No_sSuf|No_qSuf|No_ldSuf|Optimize, { RegXMM|RegYMM|RegZMM|Qword|Unspecified|BaseIndex, RegXMM|RegYMM|RegZMM, RegXMM|RegYMM|RegZMM }
vpunpckhqdq, 3, 0x666D, None, 1, CpuAVX512F, Modrm|Masking=3|VexOpcode=0|VexVVVV=1|VexW=2|Broadcast|Disp8ShiftVL|CheckRegSize|No_bSuf|No_wSuf|No_lSuf|No_sSuf|No_qSuf|No_ldSuf, { RegXMM|RegYMM|RegZMM|Qword|Unspecified|BaseIndex, RegXMM|RegYMM|RegZMM, RegXMM|RegYMM|RegZMM }
vpunpcklqdq, 3, 0x666C, None, 1, CpuAVX512F, Modrm|Masking=3|VexOpcode=0|VexVVVV=1|VexW=2|Broadcast|Disp8ShiftVL|CheckRegSize|No_bSuf|No_wSuf|No_lSuf|No_sSuf|No_qSuf|No_ldSuf, { RegXMM|RegYMM|RegZMM|Qword|Unspecified|BaseIndex, RegXMM|RegYMM|RegZMM, RegXMM|RegYMM|RegZMM }
0, 0, 0, 0, 0, 0, 0, 0, 0 } },
{ 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 1, 1, 1, 1, 1,
1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1,
- 1, 0, 0, 0, 0, 0, 0, 3, 3, 0, 0, 7, 0, 0, 0, 0, 0, 0, 0, 0 },
+ 1, 0, 0, 0, 0, 0, 0, 3, 3, 0, 0, 7, 0, 0, 1, 0, 0, 0, 0, 0 },
{ { { 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1,
0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 1, 1, 1, 1, 0, 0,
0, 0 } },
0, 0, 0, 0, 0, 0, 0, 0, 0 } },
{ 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 1, 1, 1, 1, 1,
1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1,
- 1, 0, 0, 0, 0, 0, 0, 3, 3, 0, 0, 7, 0, 0, 0, 0, 0, 0, 0, 0 },
+ 1, 0, 0, 0, 0, 0, 0, 3, 3, 0, 0, 7, 0, 0, 1, 0, 0, 0, 0, 0 },
{ { { 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1,
0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 1, 1, 1, 1, 0, 0,
0, 0 } },
0, 0, 0, 0, 0, 0, 0, 0, 0 } },
{ 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 1, 1, 1, 1, 1,
1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1,
- 2, 0, 0, 0, 0, 0, 0, 3, 4, 0, 0, 7, 0, 0, 0, 0, 0, 0, 0, 0 },
+ 2, 0, 0, 0, 0, 0, 0, 3, 4, 0, 0, 7, 0, 0, 1, 0, 0, 0, 0, 0 },
{ { { 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1,
0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 1, 1, 1, 1, 0, 0,
0, 0 } },
0, 0, 0, 0, 0, 0, 0, 0, 0 } },
{ 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 1, 1, 1, 1, 1,
1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1,
- 2, 0, 0, 0, 0, 0, 0, 3, 4, 0, 0, 7, 0, 0, 0, 0, 0, 0, 0, 0 },
+ 2, 0, 0, 0, 0, 0, 0, 3, 4, 0, 0, 7, 0, 0, 1, 0, 0, 0, 0, 0 },
{ { { 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1,
0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 1, 1, 1, 1, 0, 0,
0, 0 } },