From: Dmitrij Pochepko Date: Fri, 17 Jul 2020 09:24:46 +0000 (+0100) Subject: vector creation from two parts of two vectors produces TBL rather than ins (PR 93720) X-Git-Url: https://git.libre-soc.org/?a=commitdiff_plain;h=c9c87e6f9c795bb36e4570a07501fc182eaad035;p=gcc.git vector creation from two parts of two vectors produces TBL rather than ins (PR 93720) The following patch enables vector permutations optimization by trying to use ins instruction instead of slow and generic tbl. example: vector float f0(vector float a, vector float b) { return __builtin_shuffle (a, a, (vector int){3, 1, 2, 3}); } was compiled into: ... adrp x0, .LC0 ldr q1, [x0, #:lo12:.LC0] tbl v0.16b, {v0.16b}, v1.16b ... and after patch: ... ins v0.s[0], v0.s[3] ... bootstrapped and tested on aarch64-linux-gnu with no regressions gcc/ChangeLog: 2020-07-17 Andrew Pinski PR target/93720 * config/aarch64/aarch64.c (aarch64_evpc_ins): New function. (aarch64_expand_vec_perm_const_1): Call it. * config/aarch64/aarch64-simd.md (aarch64_simd_vec_copy_lane): Make public, and add a "@" prefix. gcc/testsuite/ChangeLog: 2020-07-17 Andrew Pinski PR target/93720 * gcc.target/aarch64/vins-1.c: New test. * gcc.target/aarch64/vins-2.c: New test. * gcc.target/aarch64/vins-3.c: New test. Co-Authored-By: Dmitrij Pochepko --- diff --git a/gcc/config/aarch64/aarch64-simd.md b/gcc/config/aarch64/aarch64-simd.md index 9f0e2bd1e6f..11ebf5b93c4 100644 --- a/gcc/config/aarch64/aarch64-simd.md +++ b/gcc/config/aarch64/aarch64-simd.md @@ -958,7 +958,7 @@ [(set_attr "type" "neon_ins, neon_from_gp, neon_load1_one_lane")] ) -(define_insn "*aarch64_simd_vec_copy_lane" +(define_insn "@aarch64_simd_vec_copy_lane" [(set (match_operand:VALL_F16 0 "register_operand" "=w") (vec_merge:VALL_F16 (vec_duplicate:VALL_F16 diff --git a/gcc/config/aarch64/aarch64.c b/gcc/config/aarch64/aarch64.c index 110f49d0c73..2a0fd714e29 100644 --- a/gcc/config/aarch64/aarch64.c +++ b/gcc/config/aarch64/aarch64.c @@ -20623,6 +20623,81 @@ aarch64_evpc_sel (struct expand_vec_perm_d *d) return true; } +/* Recognize patterns suitable for the INS instructions. */ +static bool +aarch64_evpc_ins (struct expand_vec_perm_d *d) +{ + machine_mode mode = d->vmode; + unsigned HOST_WIDE_INT nelt; + + if (d->vec_flags != VEC_ADVSIMD) + return false; + + /* to_constant is safe since this routine is specific to Advanced SIMD + vectors. */ + nelt = d->perm.length ().to_constant (); + rtx insv = d->op0; + + HOST_WIDE_INT idx = -1; + + for (unsigned HOST_WIDE_INT i = 0; i < nelt; i++) + { + HOST_WIDE_INT elt; + if (!d->perm[i].is_constant (&elt)) + return false; + if (elt == (HOST_WIDE_INT) i) + continue; + if (idx != -1) + { + idx = -1; + break; + } + idx = i; + } + + if (idx == -1) + { + insv = d->op1; + for (unsigned HOST_WIDE_INT i = 0; i < nelt; i++) + { + if (d->perm[i].to_constant () == (HOST_WIDE_INT) (i + nelt)) + continue; + if (idx != -1) + return false; + idx = i; + } + + if (idx == -1) + return false; + } + + if (d->testing_p) + return true; + + gcc_assert (idx != -1); + + unsigned extractindex = d->perm[idx].to_constant (); + rtx extractv = d->op0; + if (extractindex >= nelt) + { + extractv = d->op1; + extractindex -= nelt; + } + gcc_assert (extractindex < nelt); + + emit_move_insn (d->target, insv); + insn_code icode = code_for_aarch64_simd_vec_copy_lane (mode); + expand_operand ops[5]; + create_output_operand (&ops[0], d->target, mode); + create_input_operand (&ops[1], d->target, mode); + create_integer_operand (&ops[2], 1 << idx); + create_input_operand (&ops[3], extractv, mode); + create_integer_operand (&ops[4], extractindex); + expand_insn (icode, 5, ops); + + return true; +} + static bool aarch64_expand_vec_perm_const_1 (struct expand_vec_perm_d *d) { @@ -20657,6 +20732,8 @@ aarch64_expand_vec_perm_const_1 (struct expand_vec_perm_d *d) return true; else if (aarch64_evpc_sel (d)) return true; + else if (aarch64_evpc_ins (d)) + return true; else if (aarch64_evpc_reencode (d)) return true; if (d->vec_flags == VEC_SVE_DATA) diff --git a/gcc/testsuite/gcc.target/aarch64/vins-1.c b/gcc/testsuite/gcc.target/aarch64/vins-1.c new file mode 100644 index 00000000000..eefb93c262a --- /dev/null +++ b/gcc/testsuite/gcc.target/aarch64/vins-1.c @@ -0,0 +1,23 @@ +/* { dg-do compile } */ +/* { dg-options "-O2" } */ + +#define vector __attribute__((vector_size(4*sizeof(float)))) + +vector float f0(vector float a, vector float b) +{ + return __builtin_shuffle (a, a, (vector int){3, 1, 2, 3}); +} +vector float f1(vector float a, vector float b) +{ + return __builtin_shuffle (a, a, (vector int){0, 0, 2, 3}); +} +vector float f2(vector float a, vector float b) +{ + return __builtin_shuffle (a, a, (vector int){0, 1, 0, 3}); +} +vector float f3(vector float a, vector float b) +{ + return __builtin_shuffle (a, a, (vector int){0, 1, 2, 0}); +} + +/* { dg-final { scan-assembler-times {[ \t]*ins[ \t]+v[0-9]+\.s} 4 } } */ diff --git a/gcc/testsuite/gcc.target/aarch64/vins-2.c b/gcc/testsuite/gcc.target/aarch64/vins-2.c new file mode 100644 index 00000000000..1501b600c03 --- /dev/null +++ b/gcc/testsuite/gcc.target/aarch64/vins-2.c @@ -0,0 +1,23 @@ +/* { dg-do compile } */ +/* { dg-options "-O2" } */ + +#define vector __attribute__((vector_size(8*sizeof(short)))) + +vector short f0(vector short a, vector short b) +{ + return __builtin_shuffle (a, a, (vector short){6, 1, 2, 3, 4, 5, 6, 7}); +} +vector short f2(vector short a, vector short b) +{ + return __builtin_shuffle (a, a, (vector short){0, 1, 2, 1, 4, 5, 6, 7}); +} +vector short f4(vector short a, vector short b) +{ + return __builtin_shuffle (a, a, (vector short){0, 1, 2, 3, 0, 5, 6, 7}); +} +vector short f6(vector short a, vector short b) +{ + return __builtin_shuffle (a, a, (vector short){0, 1, 2, 3, 4, 5, 6, 1}); +} + +/* { dg-final { scan-assembler-times {[ \t]*ins[ \t]+v[0-9]+\.h} 4 } } */ diff --git a/gcc/testsuite/gcc.target/aarch64/vins-3.c b/gcc/testsuite/gcc.target/aarch64/vins-3.c new file mode 100644 index 00000000000..3b087019ce1 --- /dev/null +++ b/gcc/testsuite/gcc.target/aarch64/vins-3.c @@ -0,0 +1,23 @@ +/* { dg-do compile } */ +/* { dg-options "-O2" } */ + +#define vector __attribute__((vector_size(4*sizeof(float)))) + +vector float f0(vector float a, vector float b) +{ + return __builtin_shuffle (a, b, (vector int){0, 5, 6, 7}); +} +vector float f1(vector float a, vector float b) +{ + return __builtin_shuffle (a, b, (vector int){4, 0, 6, 7}); +} +vector float f2(vector float a, vector float b) +{ + return __builtin_shuffle (a, b, (vector int){4, 5, 0, 7}); +} +vector float f3(vector float a, vector float b) +{ + return __builtin_shuffle (a, b, (vector int){4, 5, 6, 0}); +} + +/* { dg-final { scan-assembler-times {[ \t]*ins[ \t]+v[0-9]+\.s} 4 } } */