From bf592b2ff776aef71c91924cdb5e0d10488496cf Mon Sep 17 00:00:00 2001 From: xiezhiheng Date: Fri, 17 Jul 2020 10:00:37 +0100 Subject: [PATCH] AArch64: Add flags in built-in functions [PR94442] 2020-07-17 Zhiheng Xie gcc/ChangeLog: * config/aarch64/aarch64-builtins.c (enum aarch64_type_qualifiers): Add new field flags. (VAR1): Add new field FLAG in macro. (VAR2): Likewise. (VAR3): Likewise. (VAR4): Likewise. (VAR5): Likewise. (VAR6): Likewise. (VAR7): Likewise. (VAR8): Likewise. (VAR9): Likewise. (VAR10): Likewise. (VAR11): Likewise. (VAR12): Likewise. (VAR13): Likewise. (VAR14): Likewise. (VAR15): Likewise. (VAR16): Likewise. (aarch64_general_fold_builtin): Likewise. (aarch64_general_gimple_fold_builtin): Likewise. * config/aarch64/aarch64-simd-builtins.def: Add default flag for each built-in function. * config/aarch64/geniterators.sh: Add new field in BUILTIN macro. --- gcc/config/aarch64/aarch64-builtins.c | 131 ++- gcc/config/aarch64/aarch64-simd-builtins.def | 1014 +++++++++--------- gcc/config/aarch64/geniterators.sh | 4 +- 3 files changed, 581 insertions(+), 568 deletions(-) diff --git a/gcc/config/aarch64/aarch64-builtins.c b/gcc/config/aarch64/aarch64-builtins.c index 49dfbafec3a..d5fb29048c4 100644 --- a/gcc/config/aarch64/aarch64-builtins.c +++ b/gcc/config/aarch64/aarch64-builtins.c @@ -117,6 +117,18 @@ enum aarch64_type_qualifiers qualifier_lane_quadtup_index = 0x1000, }; +/* Flags that describe what a function might do. */ +const unsigned int FLAG_NONE = 0U; +const unsigned int FLAG_READ_FPCR = 1U << 0; +const unsigned int FLAG_RAISE_FP_EXCEPTIONS = 1U << 1; +const unsigned int FLAG_READ_MEMORY = 1U << 2; +const unsigned int FLAG_PREFETCH_MEMORY = 1U << 3; +const unsigned int FLAG_WRITE_MEMORY = 1U << 4; + +const unsigned int FLAG_FP = FLAG_READ_FPCR | FLAG_RAISE_FP_EXCEPTIONS; +const unsigned int FLAG_ALL = FLAG_READ_FPCR | FLAG_RAISE_FP_EXCEPTIONS + | FLAG_READ_MEMORY | FLAG_PREFETCH_MEMORY | FLAG_WRITE_MEMORY; + typedef struct { const char *name; @@ -124,6 +136,7 @@ typedef struct const enum insn_code code; unsigned int fcode; enum aarch64_type_qualifiers *qualifiers; + unsigned int flags; } aarch64_simd_builtin_datum; static enum aarch64_type_qualifiers @@ -336,53 +349,53 @@ aarch64_types_storestruct_lane_qualifiers[SIMD_MAX_BUILTIN_ARGS] #define CF4(N, X) CODE_FOR_##N##X##4 #define CF10(N, X) CODE_FOR_##N##X -#define VAR1(T, N, MAP, A) \ - {#N #A, UP (A), CF##MAP (N, A), 0, TYPES_##T}, -#define VAR2(T, N, MAP, A, B) \ - VAR1 (T, N, MAP, A) \ - VAR1 (T, N, MAP, B) -#define VAR3(T, N, MAP, A, B, C) \ - VAR2 (T, N, MAP, A, B) \ - VAR1 (T, N, MAP, C) -#define VAR4(T, N, MAP, A, B, C, D) \ - VAR3 (T, N, MAP, A, B, C) \ - VAR1 (T, N, MAP, D) -#define VAR5(T, N, MAP, A, B, C, D, E) \ - VAR4 (T, N, MAP, A, B, C, D) \ - VAR1 (T, N, MAP, E) -#define VAR6(T, N, MAP, A, B, C, D, E, F) \ - VAR5 (T, N, MAP, A, B, C, D, E) \ - VAR1 (T, N, MAP, F) -#define VAR7(T, N, MAP, A, B, C, D, E, F, G) \ - VAR6 (T, N, MAP, A, B, C, D, E, F) \ - VAR1 (T, N, MAP, G) -#define VAR8(T, N, MAP, A, B, C, D, E, F, G, H) \ - VAR7 (T, N, MAP, A, B, C, D, E, F, G) \ - VAR1 (T, N, MAP, H) -#define VAR9(T, N, MAP, A, B, C, D, E, F, G, H, I) \ - VAR8 (T, N, MAP, A, B, C, D, E, F, G, H) \ - VAR1 (T, N, MAP, I) -#define VAR10(T, N, MAP, A, B, C, D, E, F, G, H, I, J) \ - VAR9 (T, N, MAP, A, B, C, D, E, F, G, H, I) \ - VAR1 (T, N, MAP, J) -#define VAR11(T, N, MAP, A, B, C, D, E, F, G, H, I, J, K) \ - VAR10 (T, N, MAP, A, B, C, D, E, F, G, H, I, J) \ - VAR1 (T, N, MAP, K) -#define VAR12(T, N, MAP, A, B, C, D, E, F, G, H, I, J, K, L) \ - VAR11 (T, N, MAP, A, B, C, D, E, F, G, H, I, J, K) \ - VAR1 (T, N, MAP, L) -#define VAR13(T, N, MAP, A, B, C, D, E, F, G, H, I, J, K, L, M) \ - VAR12 (T, N, MAP, A, B, C, D, E, F, G, H, I, J, K, L) \ - VAR1 (T, N, MAP, M) -#define VAR14(T, X, MAP, A, B, C, D, E, F, G, H, I, J, K, L, M, N) \ - VAR13 (T, X, MAP, A, B, C, D, E, F, G, H, I, J, K, L, M) \ - VAR1 (T, X, MAP, N) -#define VAR15(T, X, MAP, A, B, C, D, E, F, G, H, I, J, K, L, M, N, O) \ - VAR14 (T, X, MAP, A, B, C, D, E, F, G, H, I, J, K, L, M, N) \ - VAR1 (T, X, MAP, O) -#define VAR16(T, X, MAP, A, B, C, D, E, F, G, H, I, J, K, L, M, N, O, P) \ - VAR15 (T, X, MAP, A, B, C, D, E, F, G, H, I, J, K, L, M, N, O) \ - VAR1 (T, X, MAP, P) +#define VAR1(T, N, MAP, FLAG, A) \ + {#N #A, UP (A), CF##MAP (N, A), 0, TYPES_##T, FLAG_##FLAG}, +#define VAR2(T, N, MAP, FLAG, A, B) \ + VAR1 (T, N, MAP, FLAG, A) \ + VAR1 (T, N, MAP, FLAG, B) +#define VAR3(T, N, MAP, FLAG, A, B, C) \ + VAR2 (T, N, MAP, FLAG, A, B) \ + VAR1 (T, N, MAP, FLAG, C) +#define VAR4(T, N, MAP, FLAG, A, B, C, D) \ + VAR3 (T, N, MAP, FLAG, A, B, C) \ + VAR1 (T, N, MAP, FLAG, D) +#define VAR5(T, N, MAP, FLAG, A, B, C, D, E) \ + VAR4 (T, N, MAP, FLAG, A, B, C, D) \ + VAR1 (T, N, MAP, FLAG, E) +#define VAR6(T, N, MAP, FLAG, A, B, C, D, E, F) \ + VAR5 (T, N, MAP, FLAG, A, B, C, D, E) \ + VAR1 (T, N, MAP, FLAG, F) +#define VAR7(T, N, MAP, FLAG, A, B, C, D, E, F, G) \ + VAR6 (T, N, MAP, FLAG, A, B, C, D, E, F) \ + VAR1 (T, N, MAP, FLAG, G) +#define VAR8(T, N, MAP, FLAG, A, B, C, D, E, F, G, H) \ + VAR7 (T, N, MAP, FLAG, A, B, C, D, E, F, G) \ + VAR1 (T, N, MAP, FLAG, H) +#define VAR9(T, N, MAP, FLAG, A, B, C, D, E, F, G, H, I) \ + VAR8 (T, N, MAP, FLAG, A, B, C, D, E, F, G, H) \ + VAR1 (T, N, MAP, FLAG, I) +#define VAR10(T, N, MAP, FLAG, A, B, C, D, E, F, G, H, I, J) \ + VAR9 (T, N, MAP, FLAG, A, B, C, D, E, F, G, H, I) \ + VAR1 (T, N, MAP, FLAG, J) +#define VAR11(T, N, MAP, FLAG, A, B, C, D, E, F, G, H, I, J, K) \ + VAR10 (T, N, MAP, FLAG, A, B, C, D, E, F, G, H, I, J) \ + VAR1 (T, N, MAP, FLAG, K) +#define VAR12(T, N, MAP, FLAG, A, B, C, D, E, F, G, H, I, J, K, L) \ + VAR11 (T, N, MAP, FLAG, A, B, C, D, E, F, G, H, I, J, K) \ + VAR1 (T, N, MAP, FLAG, L) +#define VAR13(T, N, MAP, FLAG, A, B, C, D, E, F, G, H, I, J, K, L, M) \ + VAR12 (T, N, MAP, FLAG, A, B, C, D, E, F, G, H, I, J, K, L) \ + VAR1 (T, N, MAP, FLAG, M) +#define VAR14(T, X, MAP, FLAG, A, B, C, D, E, F, G, H, I, J, K, L, M, N) \ + VAR13 (T, X, MAP, FLAG, A, B, C, D, E, F, G, H, I, J, K, L, M) \ + VAR1 (T, X, MAP, FLAG, N) +#define VAR15(T, X, MAP, FLAG, A, B, C, D, E, F, G, H, I, J, K, L, M, N, O) \ + VAR14 (T, X, MAP, FLAG, A, B, C, D, E, F, G, H, I, J, K, L, M, N) \ + VAR1 (T, X, MAP, FLAG, O) +#define VAR16(T, X, MAP, FLAG, A, B, C, D, E, F, G, H, I, J, K, L, M, N, O, P) \ + VAR15 (T, X, MAP, FLAG, A, B, C, D, E, F, G, H, I, J, K, L, M, N, O) \ + VAR1 (T, X, MAP, FLAG, P) #include "aarch64-builtin-iterators.h" @@ -438,7 +451,7 @@ typedef struct AARCH64_SIMD_BUILTIN_FCMLA_LANEQ##I##_##M, #undef VAR1 -#define VAR1(T, N, MAP, A) \ +#define VAR1(T, N, MAP, FLAG, A) \ AARCH64_SIMD_BUILTIN_##T##_##N##A, enum aarch64_builtins @@ -2196,7 +2209,7 @@ aarch64_general_builtin_rsqrt (unsigned int fn) } #undef VAR1 -#define VAR1(T, N, MAP, A) \ +#define VAR1(T, N, MAP, FLAG, A) \ case AARCH64_SIMD_BUILTIN_##T##_##N##A: /* Try to fold a call to the built-in function with subcode FCODE. The @@ -2209,11 +2222,11 @@ aarch64_general_fold_builtin (unsigned int fcode, tree type, { switch (fcode) { - BUILTIN_VDQF (UNOP, abs, 2) + BUILTIN_VDQF (UNOP, abs, 2, ALL) return fold_build1 (ABS_EXPR, type, args[0]); - VAR1 (UNOP, floatv2si, 2, v2sf) - VAR1 (UNOP, floatv4si, 2, v4sf) - VAR1 (UNOP, floatv2di, 2, v2df) + VAR1 (UNOP, floatv2si, 2, ALL, v2sf) + VAR1 (UNOP, floatv4si, 2, ALL, v4sf) + VAR1 (UNOP, floatv2di, 2, ALL, v2df) return fold_build1 (FLOAT_EXPR, type, args[0]); default: break; @@ -2239,24 +2252,24 @@ aarch64_general_gimple_fold_builtin (unsigned int fcode, gcall *stmt) the arguments to the __builtin. */ switch (fcode) { - BUILTIN_VALL (UNOP, reduc_plus_scal_, 10) + BUILTIN_VALL (UNOP, reduc_plus_scal_, 10, ALL) new_stmt = gimple_build_call_internal (IFN_REDUC_PLUS, 1, args[0]); gimple_call_set_lhs (new_stmt, gimple_call_lhs (stmt)); break; - BUILTIN_VDQIF (UNOP, reduc_smax_scal_, 10) - BUILTIN_VDQ_BHSI (UNOPU, reduc_umax_scal_, 10) + BUILTIN_VDQIF (UNOP, reduc_smax_scal_, 10, ALL) + BUILTIN_VDQ_BHSI (UNOPU, reduc_umax_scal_, 10, ALL) new_stmt = gimple_build_call_internal (IFN_REDUC_MAX, 1, args[0]); gimple_call_set_lhs (new_stmt, gimple_call_lhs (stmt)); break; - BUILTIN_VDQIF (UNOP, reduc_smin_scal_, 10) - BUILTIN_VDQ_BHSI (UNOPU, reduc_umin_scal_, 10) + BUILTIN_VDQIF (UNOP, reduc_smin_scal_, 10, ALL) + BUILTIN_VDQ_BHSI (UNOPU, reduc_umin_scal_, 10, ALL) new_stmt = gimple_build_call_internal (IFN_REDUC_MIN, 1, args[0]); gimple_call_set_lhs (new_stmt, gimple_call_lhs (stmt)); break; - BUILTIN_GPF (BINOP, fmulx, 0) + BUILTIN_GPF (BINOP, fmulx, 0, ALL) { gcc_assert (nargs == 2); bool a0_cst_p = TREE_CODE (args[0]) == REAL_CST; diff --git a/gcc/config/aarch64/aarch64-simd-builtins.def b/gcc/config/aarch64/aarch64-simd-builtins.def index 332a0b6b1ea..e8650121cd6 100644 --- a/gcc/config/aarch64/aarch64-simd-builtins.def +++ b/gcc/config/aarch64/aarch64-simd-builtins.def @@ -39,693 +39,693 @@ 1-9 - CODE_FOR_<1-9> 10 - CODE_FOR_. */ - BUILTIN_VDC (COMBINE, combine, 0) - VAR1 (COMBINEP, combine, 0, di) - BUILTIN_VB (BINOP, pmul, 0) - BUILTIN_VHSDF_HSDF (BINOP, fmulx, 0) - BUILTIN_VHSDF_DF (UNOP, sqrt, 2) - BUILTIN_VD_BHSI (BINOP, addp, 0) - VAR1 (UNOP, addp, 0, di) - BUILTIN_VDQ_BHSI (UNOP, clrsb, 2) - BUILTIN_VDQ_BHSI (UNOP, clz, 2) - BUILTIN_VS (UNOP, ctz, 2) - BUILTIN_VB (UNOP, popcount, 2) + BUILTIN_VDC (COMBINE, combine, 0, ALL) + VAR1 (COMBINEP, combine, 0, ALL, di) + BUILTIN_VB (BINOP, pmul, 0, ALL) + BUILTIN_VHSDF_HSDF (BINOP, fmulx, 0, ALL) + BUILTIN_VHSDF_DF (UNOP, sqrt, 2, ALL) + BUILTIN_VD_BHSI (BINOP, addp, 0, ALL) + VAR1 (UNOP, addp, 0, ALL, di) + BUILTIN_VDQ_BHSI (UNOP, clrsb, 2, ALL) + BUILTIN_VDQ_BHSI (UNOP, clz, 2, ALL) + BUILTIN_VS (UNOP, ctz, 2, ALL) + BUILTIN_VB (UNOP, popcount, 2, ALL) /* Implemented by aarch64_qshl. */ - BUILTIN_VSDQ_I (BINOP, sqshl, 0) - BUILTIN_VSDQ_I (BINOP_UUS, uqshl, 0) - BUILTIN_VSDQ_I (BINOP, sqrshl, 0) - BUILTIN_VSDQ_I (BINOP_UUS, uqrshl, 0) + BUILTIN_VSDQ_I (BINOP, sqshl, 0, ALL) + BUILTIN_VSDQ_I (BINOP_UUS, uqshl, 0, ALL) + BUILTIN_VSDQ_I (BINOP, sqrshl, 0, ALL) + BUILTIN_VSDQ_I (BINOP_UUS, uqrshl, 0, ALL) /* Implemented by aarch64_. */ - BUILTIN_VSDQ_I (BINOP, sqadd, 0) - BUILTIN_VSDQ_I (BINOPU, uqadd, 0) - BUILTIN_VSDQ_I (BINOP, sqsub, 0) - BUILTIN_VSDQ_I (BINOPU, uqsub, 0) + BUILTIN_VSDQ_I (BINOP, sqadd, 0, ALL) + BUILTIN_VSDQ_I (BINOPU, uqadd, 0, ALL) + BUILTIN_VSDQ_I (BINOP, sqsub, 0, ALL) + BUILTIN_VSDQ_I (BINOPU, uqsub, 0, ALL) /* Implemented by aarch64_qadd. */ - BUILTIN_VSDQ_I (BINOP_SSU, suqadd, 0) - BUILTIN_VSDQ_I (BINOP_UUS, usqadd, 0) + BUILTIN_VSDQ_I (BINOP_SSU, suqadd, 0, ALL) + BUILTIN_VSDQ_I (BINOP_UUS, usqadd, 0, ALL) /* Implemented by aarch64_get_dreg. */ - BUILTIN_VDC (GETREG, get_dregoi, 0) - BUILTIN_VDC (GETREG, get_dregci, 0) - BUILTIN_VDC (GETREG, get_dregxi, 0) - VAR1 (GETREGP, get_dregoi, 0, di) - VAR1 (GETREGP, get_dregci, 0, di) - VAR1 (GETREGP, get_dregxi, 0, di) + BUILTIN_VDC (GETREG, get_dregoi, 0, ALL) + BUILTIN_VDC (GETREG, get_dregci, 0, ALL) + BUILTIN_VDC (GETREG, get_dregxi, 0, ALL) + VAR1 (GETREGP, get_dregoi, 0, ALL, di) + VAR1 (GETREGP, get_dregci, 0, ALL, di) + VAR1 (GETREGP, get_dregxi, 0, ALL, di) /* Implemented by aarch64_get_qreg. */ - BUILTIN_VQ (GETREG, get_qregoi, 0) - BUILTIN_VQ (GETREG, get_qregci, 0) - BUILTIN_VQ (GETREG, get_qregxi, 0) - VAR1 (GETREGP, get_qregoi, 0, v2di) - VAR1 (GETREGP, get_qregci, 0, v2di) - VAR1 (GETREGP, get_qregxi, 0, v2di) + BUILTIN_VQ (GETREG, get_qregoi, 0, ALL) + BUILTIN_VQ (GETREG, get_qregci, 0, ALL) + BUILTIN_VQ (GETREG, get_qregxi, 0, ALL) + VAR1 (GETREGP, get_qregoi, 0, ALL, v2di) + VAR1 (GETREGP, get_qregci, 0, ALL, v2di) + VAR1 (GETREGP, get_qregxi, 0, ALL, v2di) /* Implemented by aarch64_set_qreg. */ - BUILTIN_VQ (SETREG, set_qregoi, 0) - BUILTIN_VQ (SETREG, set_qregci, 0) - BUILTIN_VQ (SETREG, set_qregxi, 0) - VAR1 (SETREGP, set_qregoi, 0, v2di) - VAR1 (SETREGP, set_qregci, 0, v2di) - VAR1 (SETREGP, set_qregxi, 0, v2di) + BUILTIN_VQ (SETREG, set_qregoi, 0, ALL) + BUILTIN_VQ (SETREG, set_qregci, 0, ALL) + BUILTIN_VQ (SETREG, set_qregxi, 0, ALL) + VAR1 (SETREGP, set_qregoi, 0, ALL, v2di) + VAR1 (SETREGP, set_qregci, 0, ALL, v2di) + VAR1 (SETREGP, set_qregxi, 0, ALL, v2di) /* Implemented by aarch64_ld1x2. */ - BUILTIN_VQ (LOADSTRUCT, ld1x2, 0) + BUILTIN_VQ (LOADSTRUCT, ld1x2, 0, ALL) /* Implemented by aarch64_ld1x2. */ - BUILTIN_VDC (LOADSTRUCT, ld1x2, 0) + BUILTIN_VDC (LOADSTRUCT, ld1x2, 0, ALL) /* Implemented by aarch64_ld. */ - BUILTIN_VDC (LOADSTRUCT, ld2, 0) - BUILTIN_VDC (LOADSTRUCT, ld3, 0) - BUILTIN_VDC (LOADSTRUCT, ld4, 0) + BUILTIN_VDC (LOADSTRUCT, ld2, 0, ALL) + BUILTIN_VDC (LOADSTRUCT, ld3, 0, ALL) + BUILTIN_VDC (LOADSTRUCT, ld4, 0, ALL) /* Implemented by aarch64_ld. */ - BUILTIN_VQ (LOADSTRUCT, ld2, 0) - BUILTIN_VQ (LOADSTRUCT, ld3, 0) - BUILTIN_VQ (LOADSTRUCT, ld4, 0) + BUILTIN_VQ (LOADSTRUCT, ld2, 0, ALL) + BUILTIN_VQ (LOADSTRUCT, ld3, 0, ALL) + BUILTIN_VQ (LOADSTRUCT, ld4, 0, ALL) /* Implemented by aarch64_ldr. */ - BUILTIN_VALLDIF (LOADSTRUCT, ld2r, 0) - BUILTIN_VALLDIF (LOADSTRUCT, ld3r, 0) - BUILTIN_VALLDIF (LOADSTRUCT, ld4r, 0) + BUILTIN_VALLDIF (LOADSTRUCT, ld2r, 0, ALL) + BUILTIN_VALLDIF (LOADSTRUCT, ld3r, 0, ALL) + BUILTIN_VALLDIF (LOADSTRUCT, ld4r, 0, ALL) /* Implemented by aarch64_ld_lane. */ - BUILTIN_VALLDIF (LOADSTRUCT_LANE, ld2_lane, 0) - BUILTIN_VALLDIF (LOADSTRUCT_LANE, ld3_lane, 0) - BUILTIN_VALLDIF (LOADSTRUCT_LANE, ld4_lane, 0) + BUILTIN_VALLDIF (LOADSTRUCT_LANE, ld2_lane, 0, ALL) + BUILTIN_VALLDIF (LOADSTRUCT_LANE, ld3_lane, 0, ALL) + BUILTIN_VALLDIF (LOADSTRUCT_LANE, ld4_lane, 0, ALL) /* Implemented by aarch64_st. */ - BUILTIN_VDC (STORESTRUCT, st2, 0) - BUILTIN_VDC (STORESTRUCT, st3, 0) - BUILTIN_VDC (STORESTRUCT, st4, 0) + BUILTIN_VDC (STORESTRUCT, st2, 0, ALL) + BUILTIN_VDC (STORESTRUCT, st3, 0, ALL) + BUILTIN_VDC (STORESTRUCT, st4, 0, ALL) /* Implemented by aarch64_st. */ - BUILTIN_VQ (STORESTRUCT, st2, 0) - BUILTIN_VQ (STORESTRUCT, st3, 0) - BUILTIN_VQ (STORESTRUCT, st4, 0) - - BUILTIN_VALLDIF (STORESTRUCT_LANE, st2_lane, 0) - BUILTIN_VALLDIF (STORESTRUCT_LANE, st3_lane, 0) - BUILTIN_VALLDIF (STORESTRUCT_LANE, st4_lane, 0) - - BUILTIN_VQW (BINOP, saddl2, 0) - BUILTIN_VQW (BINOP, uaddl2, 0) - BUILTIN_VQW (BINOP, ssubl2, 0) - BUILTIN_VQW (BINOP, usubl2, 0) - BUILTIN_VQW (BINOP, saddw2, 0) - BUILTIN_VQW (BINOP, uaddw2, 0) - BUILTIN_VQW (BINOP, ssubw2, 0) - BUILTIN_VQW (BINOP, usubw2, 0) + BUILTIN_VQ (STORESTRUCT, st2, 0, ALL) + BUILTIN_VQ (STORESTRUCT, st3, 0, ALL) + BUILTIN_VQ (STORESTRUCT, st4, 0, ALL) + + BUILTIN_VALLDIF (STORESTRUCT_LANE, st2_lane, 0, ALL) + BUILTIN_VALLDIF (STORESTRUCT_LANE, st3_lane, 0, ALL) + BUILTIN_VALLDIF (STORESTRUCT_LANE, st4_lane, 0, ALL) + + BUILTIN_VQW (BINOP, saddl2, 0, ALL) + BUILTIN_VQW (BINOP, uaddl2, 0, ALL) + BUILTIN_VQW (BINOP, ssubl2, 0, ALL) + BUILTIN_VQW (BINOP, usubl2, 0, ALL) + BUILTIN_VQW (BINOP, saddw2, 0, ALL) + BUILTIN_VQW (BINOP, uaddw2, 0, ALL) + BUILTIN_VQW (BINOP, ssubw2, 0, ALL) + BUILTIN_VQW (BINOP, usubw2, 0, ALL) /* Implemented by aarch64_l. */ - BUILTIN_VD_BHSI (BINOP, saddl, 0) - BUILTIN_VD_BHSI (BINOP, uaddl, 0) - BUILTIN_VD_BHSI (BINOP, ssubl, 0) - BUILTIN_VD_BHSI (BINOP, usubl, 0) + BUILTIN_VD_BHSI (BINOP, saddl, 0, ALL) + BUILTIN_VD_BHSI (BINOP, uaddl, 0, ALL) + BUILTIN_VD_BHSI (BINOP, ssubl, 0, ALL) + BUILTIN_VD_BHSI (BINOP, usubl, 0, ALL) /* Implemented by aarch64_w. */ - BUILTIN_VD_BHSI (BINOP, saddw, 0) - BUILTIN_VD_BHSI (BINOP, uaddw, 0) - BUILTIN_VD_BHSI (BINOP, ssubw, 0) - BUILTIN_VD_BHSI (BINOP, usubw, 0) + BUILTIN_VD_BHSI (BINOP, saddw, 0, ALL) + BUILTIN_VD_BHSI (BINOP, uaddw, 0, ALL) + BUILTIN_VD_BHSI (BINOP, ssubw, 0, ALL) + BUILTIN_VD_BHSI (BINOP, usubw, 0, ALL) /* Implemented by aarch64_h. */ - BUILTIN_VDQ_BHSI (BINOP, shadd, 0) - BUILTIN_VDQ_BHSI (BINOP, shsub, 0) - BUILTIN_VDQ_BHSI (BINOP, uhadd, 0) - BUILTIN_VDQ_BHSI (BINOP, uhsub, 0) - BUILTIN_VDQ_BHSI (BINOP, srhadd, 0) - BUILTIN_VDQ_BHSI (BINOP, urhadd, 0) + BUILTIN_VDQ_BHSI (BINOP, shadd, 0, ALL) + BUILTIN_VDQ_BHSI (BINOP, shsub, 0, ALL) + BUILTIN_VDQ_BHSI (BINOP, uhadd, 0, ALL) + BUILTIN_VDQ_BHSI (BINOP, uhsub, 0, ALL) + BUILTIN_VDQ_BHSI (BINOP, srhadd, 0, ALL) + BUILTIN_VDQ_BHSI (BINOP, urhadd, 0, ALL) /* Implemented by aarch64_hn. */ - BUILTIN_VQN (BINOP, addhn, 0) - BUILTIN_VQN (BINOP, subhn, 0) - BUILTIN_VQN (BINOP, raddhn, 0) - BUILTIN_VQN (BINOP, rsubhn, 0) + BUILTIN_VQN (BINOP, addhn, 0, ALL) + BUILTIN_VQN (BINOP, subhn, 0, ALL) + BUILTIN_VQN (BINOP, raddhn, 0, ALL) + BUILTIN_VQN (BINOP, rsubhn, 0, ALL) /* Implemented by aarch64_hn2. */ - BUILTIN_VQN (TERNOP, addhn2, 0) - BUILTIN_VQN (TERNOP, subhn2, 0) - BUILTIN_VQN (TERNOP, raddhn2, 0) - BUILTIN_VQN (TERNOP, rsubhn2, 0) + BUILTIN_VQN (TERNOP, addhn2, 0, ALL) + BUILTIN_VQN (TERNOP, subhn2, 0, ALL) + BUILTIN_VQN (TERNOP, raddhn2, 0, ALL) + BUILTIN_VQN (TERNOP, rsubhn2, 0, ALL) - BUILTIN_VSQN_HSDI (UNOP, sqmovun, 0) + BUILTIN_VSQN_HSDI (UNOP, sqmovun, 0, ALL) /* Implemented by aarch64_qmovn. */ - BUILTIN_VSQN_HSDI (UNOP, sqmovn, 0) - BUILTIN_VSQN_HSDI (UNOP, uqmovn, 0) + BUILTIN_VSQN_HSDI (UNOP, sqmovn, 0, ALL) + BUILTIN_VSQN_HSDI (UNOP, uqmovn, 0, ALL) /* Implemented by aarch64_s. */ - BUILTIN_VSDQ_I (UNOP, sqabs, 0) - BUILTIN_VSDQ_I (UNOP, sqneg, 0) + BUILTIN_VSDQ_I (UNOP, sqabs, 0, ALL) + BUILTIN_VSDQ_I (UNOP, sqneg, 0, ALL) /* Implemented by aarch64_sqdmll. */ - BUILTIN_VSD_HSI (TERNOP, sqdmlal, 0) - BUILTIN_VSD_HSI (TERNOP, sqdmlsl, 0) + BUILTIN_VSD_HSI (TERNOP, sqdmlal, 0, ALL) + BUILTIN_VSD_HSI (TERNOP, sqdmlsl, 0, ALL) /* Implemented by aarch64_sqdmll_lane. */ - BUILTIN_VSD_HSI (QUADOP_LANE, sqdmlal_lane, 0) - BUILTIN_VSD_HSI (QUADOP_LANE, sqdmlsl_lane, 0) + BUILTIN_VSD_HSI (QUADOP_LANE, sqdmlal_lane, 0, ALL) + BUILTIN_VSD_HSI (QUADOP_LANE, sqdmlsl_lane, 0, ALL) /* Implemented by aarch64_sqdmll_laneq. */ - BUILTIN_VSD_HSI (QUADOP_LANE, sqdmlal_laneq, 0) - BUILTIN_VSD_HSI (QUADOP_LANE, sqdmlsl_laneq, 0) + BUILTIN_VSD_HSI (QUADOP_LANE, sqdmlal_laneq, 0, ALL) + BUILTIN_VSD_HSI (QUADOP_LANE, sqdmlsl_laneq, 0, ALL) /* Implemented by aarch64_sqdmll_n. */ - BUILTIN_VD_HSI (TERNOP, sqdmlal_n, 0) - BUILTIN_VD_HSI (TERNOP, sqdmlsl_n, 0) - - BUILTIN_VQ_HSI (TERNOP, sqdmlal2, 0) - BUILTIN_VQ_HSI (TERNOP, sqdmlsl2, 0) - BUILTIN_VQ_HSI (QUADOP_LANE, sqdmlal2_lane, 0) - BUILTIN_VQ_HSI (QUADOP_LANE, sqdmlsl2_lane, 0) - BUILTIN_VQ_HSI (QUADOP_LANE, sqdmlal2_laneq, 0) - BUILTIN_VQ_HSI (QUADOP_LANE, sqdmlsl2_laneq, 0) - BUILTIN_VQ_HSI (TERNOP, sqdmlal2_n, 0) - BUILTIN_VQ_HSI (TERNOP, sqdmlsl2_n, 0) - - BUILTIN_VD_BHSI (BINOP, intrinsic_vec_smult_lo_, 0) - BUILTIN_VD_BHSI (BINOPU, intrinsic_vec_umult_lo_, 0) - - BUILTIN_VQW (BINOP, vec_widen_smult_hi_, 10) - BUILTIN_VQW (BINOPU, vec_widen_umult_hi_, 10) - - BUILTIN_VD_HSI (TERNOP_LANE, vec_smult_lane_, 0) - BUILTIN_VD_HSI (QUADOP_LANE, vec_smlal_lane_, 0) - BUILTIN_VD_HSI (TERNOP_LANE, vec_smult_laneq_, 0) - BUILTIN_VD_HSI (QUADOP_LANE, vec_smlal_laneq_, 0) - BUILTIN_VD_HSI (TERNOPU_LANE, vec_umult_lane_, 0) - BUILTIN_VD_HSI (QUADOPU_LANE, vec_umlal_lane_, 0) - BUILTIN_VD_HSI (TERNOPU_LANE, vec_umult_laneq_, 0) - BUILTIN_VD_HSI (QUADOPU_LANE, vec_umlal_laneq_, 0) - - BUILTIN_VSD_HSI (BINOP, sqdmull, 0) - BUILTIN_VSD_HSI (TERNOP_LANE, sqdmull_lane, 0) - BUILTIN_VSD_HSI (TERNOP_LANE, sqdmull_laneq, 0) - BUILTIN_VD_HSI (BINOP, sqdmull_n, 0) - BUILTIN_VQ_HSI (BINOP, sqdmull2, 0) - BUILTIN_VQ_HSI (TERNOP_LANE, sqdmull2_lane, 0) - BUILTIN_VQ_HSI (TERNOP_LANE, sqdmull2_laneq, 0) - BUILTIN_VQ_HSI (BINOP, sqdmull2_n, 0) + BUILTIN_VD_HSI (TERNOP, sqdmlal_n, 0, ALL) + BUILTIN_VD_HSI (TERNOP, sqdmlsl_n, 0, ALL) + + BUILTIN_VQ_HSI (TERNOP, sqdmlal2, 0, ALL) + BUILTIN_VQ_HSI (TERNOP, sqdmlsl2, 0, ALL) + BUILTIN_VQ_HSI (QUADOP_LANE, sqdmlal2_lane, 0, ALL) + BUILTIN_VQ_HSI (QUADOP_LANE, sqdmlsl2_lane, 0, ALL) + BUILTIN_VQ_HSI (QUADOP_LANE, sqdmlal2_laneq, 0, ALL) + BUILTIN_VQ_HSI (QUADOP_LANE, sqdmlsl2_laneq, 0, ALL) + BUILTIN_VQ_HSI (TERNOP, sqdmlal2_n, 0, ALL) + BUILTIN_VQ_HSI (TERNOP, sqdmlsl2_n, 0, ALL) + + BUILTIN_VD_BHSI (BINOP, intrinsic_vec_smult_lo_, 0, ALL) + BUILTIN_VD_BHSI (BINOPU, intrinsic_vec_umult_lo_, 0, ALL) + + BUILTIN_VQW (BINOP, vec_widen_smult_hi_, 10, ALL) + BUILTIN_VQW (BINOPU, vec_widen_umult_hi_, 10, ALL) + + BUILTIN_VD_HSI (TERNOP_LANE, vec_smult_lane_, 0, ALL) + BUILTIN_VD_HSI (QUADOP_LANE, vec_smlal_lane_, 0, ALL) + BUILTIN_VD_HSI (TERNOP_LANE, vec_smult_laneq_, 0, ALL) + BUILTIN_VD_HSI (QUADOP_LANE, vec_smlal_laneq_, 0, ALL) + BUILTIN_VD_HSI (TERNOPU_LANE, vec_umult_lane_, 0, ALL) + BUILTIN_VD_HSI (QUADOPU_LANE, vec_umlal_lane_, 0, ALL) + BUILTIN_VD_HSI (TERNOPU_LANE, vec_umult_laneq_, 0, ALL) + BUILTIN_VD_HSI (QUADOPU_LANE, vec_umlal_laneq_, 0, ALL) + + BUILTIN_VSD_HSI (BINOP, sqdmull, 0, ALL) + BUILTIN_VSD_HSI (TERNOP_LANE, sqdmull_lane, 0, ALL) + BUILTIN_VSD_HSI (TERNOP_LANE, sqdmull_laneq, 0, ALL) + BUILTIN_VD_HSI (BINOP, sqdmull_n, 0, ALL) + BUILTIN_VQ_HSI (BINOP, sqdmull2, 0, ALL) + BUILTIN_VQ_HSI (TERNOP_LANE, sqdmull2_lane, 0, ALL) + BUILTIN_VQ_HSI (TERNOP_LANE, sqdmull2_laneq, 0, ALL) + BUILTIN_VQ_HSI (BINOP, sqdmull2_n, 0, ALL) /* Implemented by aarch64_sqdmulh. */ - BUILTIN_VSDQ_HSI (BINOP, sqdmulh, 0) - BUILTIN_VSDQ_HSI (BINOP, sqrdmulh, 0) + BUILTIN_VSDQ_HSI (BINOP, sqdmulh, 0, ALL) + BUILTIN_VSDQ_HSI (BINOP, sqrdmulh, 0, ALL) /* Implemented by aarch64_sqdmulh_lane. */ - BUILTIN_VSDQ_HSI (TERNOP_LANE, sqdmulh_lane, 0) - BUILTIN_VSDQ_HSI (TERNOP_LANE, sqdmulh_laneq, 0) - BUILTIN_VSDQ_HSI (TERNOP_LANE, sqrdmulh_lane, 0) - BUILTIN_VSDQ_HSI (TERNOP_LANE, sqrdmulh_laneq, 0) + BUILTIN_VSDQ_HSI (TERNOP_LANE, sqdmulh_lane, 0, ALL) + BUILTIN_VSDQ_HSI (TERNOP_LANE, sqdmulh_laneq, 0, ALL) + BUILTIN_VSDQ_HSI (TERNOP_LANE, sqrdmulh_lane, 0, ALL) + BUILTIN_VSDQ_HSI (TERNOP_LANE, sqrdmulh_laneq, 0, ALL) - BUILTIN_VSDQ_I_DI (BINOP, ashl, 3) + BUILTIN_VSDQ_I_DI (BINOP, ashl, 3, ALL) /* Implemented by aarch64_shl. */ - BUILTIN_VSDQ_I_DI (BINOP, sshl, 0) - BUILTIN_VSDQ_I_DI (BINOP_UUS, ushl, 0) - BUILTIN_VSDQ_I_DI (BINOP, srshl, 0) - BUILTIN_VSDQ_I_DI (BINOP_UUS, urshl, 0) + BUILTIN_VSDQ_I_DI (BINOP, sshl, 0, ALL) + BUILTIN_VSDQ_I_DI (BINOP_UUS, ushl, 0, ALL) + BUILTIN_VSDQ_I_DI (BINOP, srshl, 0, ALL) + BUILTIN_VSDQ_I_DI (BINOP_UUS, urshl, 0, ALL) /* Implemented by aarch64_{_lane}{q}. */ - BUILTIN_VB (TERNOP, sdot, 0) - BUILTIN_VB (TERNOPU, udot, 0) - BUILTIN_VB (TERNOP_SSUS, usdot, 0) - BUILTIN_VB (QUADOP_LANE, sdot_lane, 0) - BUILTIN_VB (QUADOPU_LANE, udot_lane, 0) - BUILTIN_VB (QUADOP_LANE, sdot_laneq, 0) - BUILTIN_VB (QUADOPU_LANE, udot_laneq, 0) - BUILTIN_VB (QUADOPSSUS_LANE_QUADTUP, usdot_lane, 0) - BUILTIN_VB (QUADOPSSUS_LANE_QUADTUP, usdot_laneq, 0) - BUILTIN_VB (QUADOPSSSU_LANE_QUADTUP, sudot_lane, 0) - BUILTIN_VB (QUADOPSSSU_LANE_QUADTUP, sudot_laneq, 0) + BUILTIN_VB (TERNOP, sdot, 0, ALL) + BUILTIN_VB (TERNOPU, udot, 0, ALL) + BUILTIN_VB (TERNOP_SSUS, usdot, 0, ALL) + BUILTIN_VB (QUADOP_LANE, sdot_lane, 0, ALL) + BUILTIN_VB (QUADOPU_LANE, udot_lane, 0, ALL) + BUILTIN_VB (QUADOP_LANE, sdot_laneq, 0, ALL) + BUILTIN_VB (QUADOPU_LANE, udot_laneq, 0, ALL) + BUILTIN_VB (QUADOPSSUS_LANE_QUADTUP, usdot_lane, 0, ALL) + BUILTIN_VB (QUADOPSSUS_LANE_QUADTUP, usdot_laneq, 0, ALL) + BUILTIN_VB (QUADOPSSSU_LANE_QUADTUP, sudot_lane, 0, ALL) + BUILTIN_VB (QUADOPSSSU_LANE_QUADTUP, sudot_laneq, 0, ALL) /* Implemented by aarch64_fcadd. */ - BUILTIN_VHSDF (BINOP, fcadd90, 0) - BUILTIN_VHSDF (BINOP, fcadd270, 0) + BUILTIN_VHSDF (BINOP, fcadd90, 0, ALL) + BUILTIN_VHSDF (BINOP, fcadd270, 0, ALL) /* Implemented by aarch64_fcmla{_lane}{q}. */ - BUILTIN_VHSDF (TERNOP, fcmla0, 0) - BUILTIN_VHSDF (TERNOP, fcmla90, 0) - BUILTIN_VHSDF (TERNOP, fcmla180, 0) - BUILTIN_VHSDF (TERNOP, fcmla270, 0) - BUILTIN_VHSDF (QUADOP_LANE_PAIR, fcmla_lane0, 0) - BUILTIN_VHSDF (QUADOP_LANE_PAIR, fcmla_lane90, 0) - BUILTIN_VHSDF (QUADOP_LANE_PAIR, fcmla_lane180, 0) - BUILTIN_VHSDF (QUADOP_LANE_PAIR, fcmla_lane270, 0) - - BUILTIN_VQ_HSF (QUADOP_LANE_PAIR, fcmlaq_lane0, 0) - BUILTIN_VQ_HSF (QUADOP_LANE_PAIR, fcmlaq_lane90, 0) - BUILTIN_VQ_HSF (QUADOP_LANE_PAIR, fcmlaq_lane180, 0) - BUILTIN_VQ_HSF (QUADOP_LANE_PAIR, fcmlaq_lane270, 0) - - BUILTIN_VDQ_I (SHIFTIMM, ashr, 3) - VAR1 (SHIFTIMM, ashr_simd, 0, di) - BUILTIN_VDQ_I (SHIFTIMM, lshr, 3) - VAR1 (USHIFTIMM, lshr_simd, 0, di) + BUILTIN_VHSDF (TERNOP, fcmla0, 0, ALL) + BUILTIN_VHSDF (TERNOP, fcmla90, 0, ALL) + BUILTIN_VHSDF (TERNOP, fcmla180, 0, ALL) + BUILTIN_VHSDF (TERNOP, fcmla270, 0, ALL) + BUILTIN_VHSDF (QUADOP_LANE_PAIR, fcmla_lane0, 0, ALL) + BUILTIN_VHSDF (QUADOP_LANE_PAIR, fcmla_lane90, 0, ALL) + BUILTIN_VHSDF (QUADOP_LANE_PAIR, fcmla_lane180, 0, ALL) + BUILTIN_VHSDF (QUADOP_LANE_PAIR, fcmla_lane270, 0, ALL) + + BUILTIN_VQ_HSF (QUADOP_LANE_PAIR, fcmlaq_lane0, 0, ALL) + BUILTIN_VQ_HSF (QUADOP_LANE_PAIR, fcmlaq_lane90, 0, ALL) + BUILTIN_VQ_HSF (QUADOP_LANE_PAIR, fcmlaq_lane180, 0, ALL) + BUILTIN_VQ_HSF (QUADOP_LANE_PAIR, fcmlaq_lane270, 0, ALL) + + BUILTIN_VDQ_I (SHIFTIMM, ashr, 3, ALL) + VAR1 (SHIFTIMM, ashr_simd, 0, ALL, di) + BUILTIN_VDQ_I (SHIFTIMM, lshr, 3, ALL) + VAR1 (USHIFTIMM, lshr_simd, 0, ALL, di) /* Implemented by aarch64_shr_n. */ - BUILTIN_VSDQ_I_DI (SHIFTIMM, srshr_n, 0) - BUILTIN_VSDQ_I_DI (USHIFTIMM, urshr_n, 0) + BUILTIN_VSDQ_I_DI (SHIFTIMM, srshr_n, 0, ALL) + BUILTIN_VSDQ_I_DI (USHIFTIMM, urshr_n, 0, ALL) /* Implemented by aarch64_sra_n. */ - BUILTIN_VSDQ_I_DI (SHIFTACC, ssra_n, 0) - BUILTIN_VSDQ_I_DI (USHIFTACC, usra_n, 0) - BUILTIN_VSDQ_I_DI (SHIFTACC, srsra_n, 0) - BUILTIN_VSDQ_I_DI (USHIFTACC, ursra_n, 0) + BUILTIN_VSDQ_I_DI (SHIFTACC, ssra_n, 0, ALL) + BUILTIN_VSDQ_I_DI (USHIFTACC, usra_n, 0, ALL) + BUILTIN_VSDQ_I_DI (SHIFTACC, srsra_n, 0, ALL) + BUILTIN_VSDQ_I_DI (USHIFTACC, ursra_n, 0, ALL) /* Implemented by aarch64_shll_n. */ - BUILTIN_VD_BHSI (SHIFTIMM, sshll_n, 0) - BUILTIN_VD_BHSI (USHIFTIMM, ushll_n, 0) + BUILTIN_VD_BHSI (SHIFTIMM, sshll_n, 0, ALL) + BUILTIN_VD_BHSI (USHIFTIMM, ushll_n, 0, ALL) /* Implemented by aarch64_shll2_n. */ - BUILTIN_VQW (SHIFTIMM, sshll2_n, 0) - BUILTIN_VQW (SHIFTIMM, ushll2_n, 0) + BUILTIN_VQW (SHIFTIMM, sshll2_n, 0, ALL) + BUILTIN_VQW (SHIFTIMM, ushll2_n, 0, ALL) /* Implemented by aarch64_qshrn_n. */ - BUILTIN_VSQN_HSDI (SHIFTIMM, sqshrun_n, 0) - BUILTIN_VSQN_HSDI (SHIFTIMM, sqrshrun_n, 0) - BUILTIN_VSQN_HSDI (SHIFTIMM, sqshrn_n, 0) - BUILTIN_VSQN_HSDI (USHIFTIMM, uqshrn_n, 0) - BUILTIN_VSQN_HSDI (SHIFTIMM, sqrshrn_n, 0) - BUILTIN_VSQN_HSDI (USHIFTIMM, uqrshrn_n, 0) + BUILTIN_VSQN_HSDI (SHIFTIMM, sqshrun_n, 0, ALL) + BUILTIN_VSQN_HSDI (SHIFTIMM, sqrshrun_n, 0, ALL) + BUILTIN_VSQN_HSDI (SHIFTIMM, sqshrn_n, 0, ALL) + BUILTIN_VSQN_HSDI (USHIFTIMM, uqshrn_n, 0, ALL) + BUILTIN_VSQN_HSDI (SHIFTIMM, sqrshrn_n, 0, ALL) + BUILTIN_VSQN_HSDI (USHIFTIMM, uqrshrn_n, 0, ALL) /* Implemented by aarch64_si_n. */ - BUILTIN_VSDQ_I_DI (SHIFTINSERT, ssri_n, 0) - BUILTIN_VSDQ_I_DI (USHIFTACC, usri_n, 0) - BUILTIN_VSDQ_I_DI (SHIFTINSERT, ssli_n, 0) - VAR2 (SHIFTINSERTP, ssli_n, 0, di, v2di) - BUILTIN_VSDQ_I_DI (USHIFTACC, usli_n, 0) + BUILTIN_VSDQ_I_DI (SHIFTINSERT, ssri_n, 0, ALL) + BUILTIN_VSDQ_I_DI (USHIFTACC, usri_n, 0, ALL) + BUILTIN_VSDQ_I_DI (SHIFTINSERT, ssli_n, 0, ALL) + VAR2 (SHIFTINSERTP, ssli_n, 0, ALL, di, v2di) + BUILTIN_VSDQ_I_DI (USHIFTACC, usli_n, 0, ALL) /* Implemented by aarch64_qshl_n. */ - BUILTIN_VSDQ_I (SHIFTIMM_USS, sqshlu_n, 0) - BUILTIN_VSDQ_I (SHIFTIMM, sqshl_n, 0) - BUILTIN_VSDQ_I (USHIFTIMM, uqshl_n, 0) + BUILTIN_VSDQ_I (SHIFTIMM_USS, sqshlu_n, 0, ALL) + BUILTIN_VSDQ_I (SHIFTIMM, sqshl_n, 0, ALL) + BUILTIN_VSDQ_I (USHIFTIMM, uqshl_n, 0, ALL) /* Implemented by aarch64_reduc_plus_. */ - BUILTIN_VALL (UNOP, reduc_plus_scal_, 10) + BUILTIN_VALL (UNOP, reduc_plus_scal_, 10, ALL) /* Implemented by reduc__scal_ (producing scalar). */ - BUILTIN_VDQIF_F16 (UNOP, reduc_smax_scal_, 10) - BUILTIN_VDQIF_F16 (UNOP, reduc_smin_scal_, 10) - BUILTIN_VDQ_BHSI (UNOPU, reduc_umax_scal_, 10) - BUILTIN_VDQ_BHSI (UNOPU, reduc_umin_scal_, 10) - BUILTIN_VHSDF (UNOP, reduc_smax_nan_scal_, 10) - BUILTIN_VHSDF (UNOP, reduc_smin_nan_scal_, 10) + BUILTIN_VDQIF_F16 (UNOP, reduc_smax_scal_, 10, ALL) + BUILTIN_VDQIF_F16 (UNOP, reduc_smin_scal_, 10, ALL) + BUILTIN_VDQ_BHSI (UNOPU, reduc_umax_scal_, 10, ALL) + BUILTIN_VDQ_BHSI (UNOPU, reduc_umin_scal_, 10, ALL) + BUILTIN_VHSDF (UNOP, reduc_smax_nan_scal_, 10, ALL) + BUILTIN_VHSDF (UNOP, reduc_smin_nan_scal_, 10, ALL) /* Implemented by 3. smax variants map to fmaxnm, smax_nan variants map to fmax. */ - BUILTIN_VDQ_BHSI (BINOP, smax, 3) - BUILTIN_VDQ_BHSI (BINOP, smin, 3) - BUILTIN_VDQ_BHSI (BINOP, umax, 3) - BUILTIN_VDQ_BHSI (BINOP, umin, 3) - BUILTIN_VHSDF_DF (BINOP, smax_nan, 3) - BUILTIN_VHSDF_DF (BINOP, smin_nan, 3) + BUILTIN_VDQ_BHSI (BINOP, smax, 3, ALL) + BUILTIN_VDQ_BHSI (BINOP, smin, 3, ALL) + BUILTIN_VDQ_BHSI (BINOP, umax, 3, ALL) + BUILTIN_VDQ_BHSI (BINOP, umin, 3, ALL) + BUILTIN_VHSDF_DF (BINOP, smax_nan, 3, ALL) + BUILTIN_VHSDF_DF (BINOP, smin_nan, 3, ALL) /* Implemented by 3. */ - BUILTIN_VHSDF_HSDF (BINOP, fmax, 3) - BUILTIN_VHSDF_HSDF (BINOP, fmin, 3) + BUILTIN_VHSDF_HSDF (BINOP, fmax, 3, ALL) + BUILTIN_VHSDF_HSDF (BINOP, fmin, 3, ALL) /* Implemented by aarch64_p. */ - BUILTIN_VDQ_BHSI (BINOP, smaxp, 0) - BUILTIN_VDQ_BHSI (BINOP, sminp, 0) - BUILTIN_VDQ_BHSI (BINOP, umaxp, 0) - BUILTIN_VDQ_BHSI (BINOP, uminp, 0) - BUILTIN_VHSDF (BINOP, smaxp, 0) - BUILTIN_VHSDF (BINOP, sminp, 0) - BUILTIN_VHSDF (BINOP, smax_nanp, 0) - BUILTIN_VHSDF (BINOP, smin_nanp, 0) + BUILTIN_VDQ_BHSI (BINOP, smaxp, 0, ALL) + BUILTIN_VDQ_BHSI (BINOP, sminp, 0, ALL) + BUILTIN_VDQ_BHSI (BINOP, umaxp, 0, ALL) + BUILTIN_VDQ_BHSI (BINOP, uminp, 0, ALL) + BUILTIN_VHSDF (BINOP, smaxp, 0, ALL) + BUILTIN_VHSDF (BINOP, sminp, 0, ALL) + BUILTIN_VHSDF (BINOP, smax_nanp, 0, ALL) + BUILTIN_VHSDF (BINOP, smin_nanp, 0, ALL) /* Implemented by 2. */ - BUILTIN_VHSDF (UNOP, btrunc, 2) - BUILTIN_VHSDF (UNOP, ceil, 2) - BUILTIN_VHSDF (UNOP, floor, 2) - BUILTIN_VHSDF (UNOP, nearbyint, 2) - BUILTIN_VHSDF (UNOP, rint, 2) - BUILTIN_VHSDF (UNOP, round, 2) - BUILTIN_VHSDF_DF (UNOP, frintn, 2) - - VAR1 (UNOP, btrunc, 2, hf) - VAR1 (UNOP, ceil, 2, hf) - VAR1 (UNOP, floor, 2, hf) - VAR1 (UNOP, frintn, 2, hf) - VAR1 (UNOP, nearbyint, 2, hf) - VAR1 (UNOP, rint, 2, hf) - VAR1 (UNOP, round, 2, hf) + BUILTIN_VHSDF (UNOP, btrunc, 2, ALL) + BUILTIN_VHSDF (UNOP, ceil, 2, ALL) + BUILTIN_VHSDF (UNOP, floor, 2, ALL) + BUILTIN_VHSDF (UNOP, nearbyint, 2, ALL) + BUILTIN_VHSDF (UNOP, rint, 2, ALL) + BUILTIN_VHSDF (UNOP, round, 2, ALL) + BUILTIN_VHSDF_DF (UNOP, frintn, 2, ALL) + + VAR1 (UNOP, btrunc, 2, ALL, hf) + VAR1 (UNOP, ceil, 2, ALL, hf) + VAR1 (UNOP, floor, 2, ALL, hf) + VAR1 (UNOP, frintn, 2, ALL, hf) + VAR1 (UNOP, nearbyint, 2, ALL, hf) + VAR1 (UNOP, rint, 2, ALL, hf) + VAR1 (UNOP, round, 2, ALL, hf) /* Implemented by l2. */ - VAR1 (UNOP, lbtruncv4hf, 2, v4hi) - VAR1 (UNOP, lbtruncv8hf, 2, v8hi) - VAR1 (UNOP, lbtruncv2sf, 2, v2si) - VAR1 (UNOP, lbtruncv4sf, 2, v4si) - VAR1 (UNOP, lbtruncv2df, 2, v2di) - - VAR1 (UNOPUS, lbtruncuv4hf, 2, v4hi) - VAR1 (UNOPUS, lbtruncuv8hf, 2, v8hi) - VAR1 (UNOPUS, lbtruncuv2sf, 2, v2si) - VAR1 (UNOPUS, lbtruncuv4sf, 2, v4si) - VAR1 (UNOPUS, lbtruncuv2df, 2, v2di) - - VAR1 (UNOP, lroundv4hf, 2, v4hi) - VAR1 (UNOP, lroundv8hf, 2, v8hi) - VAR1 (UNOP, lroundv2sf, 2, v2si) - VAR1 (UNOP, lroundv4sf, 2, v4si) - VAR1 (UNOP, lroundv2df, 2, v2di) + VAR1 (UNOP, lbtruncv4hf, 2, ALL, v4hi) + VAR1 (UNOP, lbtruncv8hf, 2, ALL, v8hi) + VAR1 (UNOP, lbtruncv2sf, 2, ALL, v2si) + VAR1 (UNOP, lbtruncv4sf, 2, ALL, v4si) + VAR1 (UNOP, lbtruncv2df, 2, ALL, v2di) + + VAR1 (UNOPUS, lbtruncuv4hf, 2, ALL, v4hi) + VAR1 (UNOPUS, lbtruncuv8hf, 2, ALL, v8hi) + VAR1 (UNOPUS, lbtruncuv2sf, 2, ALL, v2si) + VAR1 (UNOPUS, lbtruncuv4sf, 2, ALL, v4si) + VAR1 (UNOPUS, lbtruncuv2df, 2, ALL, v2di) + + VAR1 (UNOP, lroundv4hf, 2, ALL, v4hi) + VAR1 (UNOP, lroundv8hf, 2, ALL, v8hi) + VAR1 (UNOP, lroundv2sf, 2, ALL, v2si) + VAR1 (UNOP, lroundv4sf, 2, ALL, v4si) + VAR1 (UNOP, lroundv2df, 2, ALL, v2di) /* Implemented by l2. */ - BUILTIN_GPI_I16 (UNOP, lroundhf, 2) - VAR1 (UNOP, lroundsf, 2, si) - VAR1 (UNOP, lrounddf, 2, di) - - VAR1 (UNOPUS, lrounduv4hf, 2, v4hi) - VAR1 (UNOPUS, lrounduv8hf, 2, v8hi) - VAR1 (UNOPUS, lrounduv2sf, 2, v2si) - VAR1 (UNOPUS, lrounduv4sf, 2, v4si) - VAR1 (UNOPUS, lrounduv2df, 2, v2di) - BUILTIN_GPI_I16 (UNOPUS, lrounduhf, 2) - VAR1 (UNOPUS, lroundusf, 2, si) - VAR1 (UNOPUS, lroundudf, 2, di) - - VAR1 (UNOP, lceilv4hf, 2, v4hi) - VAR1 (UNOP, lceilv8hf, 2, v8hi) - VAR1 (UNOP, lceilv2sf, 2, v2si) - VAR1 (UNOP, lceilv4sf, 2, v4si) - VAR1 (UNOP, lceilv2df, 2, v2di) - BUILTIN_GPI_I16 (UNOP, lceilhf, 2) - - VAR1 (UNOPUS, lceiluv4hf, 2, v4hi) - VAR1 (UNOPUS, lceiluv8hf, 2, v8hi) - VAR1 (UNOPUS, lceiluv2sf, 2, v2si) - VAR1 (UNOPUS, lceiluv4sf, 2, v4si) - VAR1 (UNOPUS, lceiluv2df, 2, v2di) - BUILTIN_GPI_I16 (UNOPUS, lceiluhf, 2) - VAR1 (UNOPUS, lceilusf, 2, si) - VAR1 (UNOPUS, lceiludf, 2, di) - - VAR1 (UNOP, lfloorv4hf, 2, v4hi) - VAR1 (UNOP, lfloorv8hf, 2, v8hi) - VAR1 (UNOP, lfloorv2sf, 2, v2si) - VAR1 (UNOP, lfloorv4sf, 2, v4si) - VAR1 (UNOP, lfloorv2df, 2, v2di) - BUILTIN_GPI_I16 (UNOP, lfloorhf, 2) - - VAR1 (UNOPUS, lflooruv4hf, 2, v4hi) - VAR1 (UNOPUS, lflooruv8hf, 2, v8hi) - VAR1 (UNOPUS, lflooruv2sf, 2, v2si) - VAR1 (UNOPUS, lflooruv4sf, 2, v4si) - VAR1 (UNOPUS, lflooruv2df, 2, v2di) - BUILTIN_GPI_I16 (UNOPUS, lflooruhf, 2) - VAR1 (UNOPUS, lfloorusf, 2, si) - VAR1 (UNOPUS, lfloorudf, 2, di) - - VAR1 (UNOP, lfrintnv4hf, 2, v4hi) - VAR1 (UNOP, lfrintnv8hf, 2, v8hi) - VAR1 (UNOP, lfrintnv2sf, 2, v2si) - VAR1 (UNOP, lfrintnv4sf, 2, v4si) - VAR1 (UNOP, lfrintnv2df, 2, v2di) - BUILTIN_GPI_I16 (UNOP, lfrintnhf, 2) - VAR1 (UNOP, lfrintnsf, 2, si) - VAR1 (UNOP, lfrintndf, 2, di) - - VAR1 (UNOPUS, lfrintnuv4hf, 2, v4hi) - VAR1 (UNOPUS, lfrintnuv8hf, 2, v8hi) - VAR1 (UNOPUS, lfrintnuv2sf, 2, v2si) - VAR1 (UNOPUS, lfrintnuv4sf, 2, v4si) - VAR1 (UNOPUS, lfrintnuv2df, 2, v2di) - BUILTIN_GPI_I16 (UNOPUS, lfrintnuhf, 2) - VAR1 (UNOPUS, lfrintnusf, 2, si) - VAR1 (UNOPUS, lfrintnudf, 2, di) + BUILTIN_GPI_I16 (UNOP, lroundhf, 2, ALL) + VAR1 (UNOP, lroundsf, 2, ALL, si) + VAR1 (UNOP, lrounddf, 2, ALL, di) + + VAR1 (UNOPUS, lrounduv4hf, 2, ALL, v4hi) + VAR1 (UNOPUS, lrounduv8hf, 2, ALL, v8hi) + VAR1 (UNOPUS, lrounduv2sf, 2, ALL, v2si) + VAR1 (UNOPUS, lrounduv4sf, 2, ALL, v4si) + VAR1 (UNOPUS, lrounduv2df, 2, ALL, v2di) + BUILTIN_GPI_I16 (UNOPUS, lrounduhf, 2, ALL) + VAR1 (UNOPUS, lroundusf, 2, ALL, si) + VAR1 (UNOPUS, lroundudf, 2, ALL, di) + + VAR1 (UNOP, lceilv4hf, 2, ALL, v4hi) + VAR1 (UNOP, lceilv8hf, 2, ALL, v8hi) + VAR1 (UNOP, lceilv2sf, 2, ALL, v2si) + VAR1 (UNOP, lceilv4sf, 2, ALL, v4si) + VAR1 (UNOP, lceilv2df, 2, ALL, v2di) + BUILTIN_GPI_I16 (UNOP, lceilhf, 2, ALL) + + VAR1 (UNOPUS, lceiluv4hf, 2, ALL, v4hi) + VAR1 (UNOPUS, lceiluv8hf, 2, ALL, v8hi) + VAR1 (UNOPUS, lceiluv2sf, 2, ALL, v2si) + VAR1 (UNOPUS, lceiluv4sf, 2, ALL, v4si) + VAR1 (UNOPUS, lceiluv2df, 2, ALL, v2di) + BUILTIN_GPI_I16 (UNOPUS, lceiluhf, 2, ALL) + VAR1 (UNOPUS, lceilusf, 2, ALL, si) + VAR1 (UNOPUS, lceiludf, 2, ALL, di) + + VAR1 (UNOP, lfloorv4hf, 2, ALL, v4hi) + VAR1 (UNOP, lfloorv8hf, 2, ALL, v8hi) + VAR1 (UNOP, lfloorv2sf, 2, ALL, v2si) + VAR1 (UNOP, lfloorv4sf, 2, ALL, v4si) + VAR1 (UNOP, lfloorv2df, 2, ALL, v2di) + BUILTIN_GPI_I16 (UNOP, lfloorhf, 2, ALL) + + VAR1 (UNOPUS, lflooruv4hf, 2, ALL, v4hi) + VAR1 (UNOPUS, lflooruv8hf, 2, ALL, v8hi) + VAR1 (UNOPUS, lflooruv2sf, 2, ALL, v2si) + VAR1 (UNOPUS, lflooruv4sf, 2, ALL, v4si) + VAR1 (UNOPUS, lflooruv2df, 2, ALL, v2di) + BUILTIN_GPI_I16 (UNOPUS, lflooruhf, 2, ALL) + VAR1 (UNOPUS, lfloorusf, 2, ALL, si) + VAR1 (UNOPUS, lfloorudf, 2, ALL, di) + + VAR1 (UNOP, lfrintnv4hf, 2, ALL, v4hi) + VAR1 (UNOP, lfrintnv8hf, 2, ALL, v8hi) + VAR1 (UNOP, lfrintnv2sf, 2, ALL, v2si) + VAR1 (UNOP, lfrintnv4sf, 2, ALL, v4si) + VAR1 (UNOP, lfrintnv2df, 2, ALL, v2di) + BUILTIN_GPI_I16 (UNOP, lfrintnhf, 2, ALL) + VAR1 (UNOP, lfrintnsf, 2, ALL, si) + VAR1 (UNOP, lfrintndf, 2, ALL, di) + + VAR1 (UNOPUS, lfrintnuv4hf, 2, ALL, v4hi) + VAR1 (UNOPUS, lfrintnuv8hf, 2, ALL, v8hi) + VAR1 (UNOPUS, lfrintnuv2sf, 2, ALL, v2si) + VAR1 (UNOPUS, lfrintnuv4sf, 2, ALL, v4si) + VAR1 (UNOPUS, lfrintnuv2df, 2, ALL, v2di) + BUILTIN_GPI_I16 (UNOPUS, lfrintnuhf, 2, ALL) + VAR1 (UNOPUS, lfrintnusf, 2, ALL, si) + VAR1 (UNOPUS, lfrintnudf, 2, ALL, di) /* Implemented by 2. */ - VAR1 (UNOP, floatv4hi, 2, v4hf) - VAR1 (UNOP, floatv8hi, 2, v8hf) - VAR1 (UNOP, floatv2si, 2, v2sf) - VAR1 (UNOP, floatv4si, 2, v4sf) - VAR1 (UNOP, floatv2di, 2, v2df) + VAR1 (UNOP, floatv4hi, 2, ALL, v4hf) + VAR1 (UNOP, floatv8hi, 2, ALL, v8hf) + VAR1 (UNOP, floatv2si, 2, ALL, v2sf) + VAR1 (UNOP, floatv4si, 2, ALL, v4sf) + VAR1 (UNOP, floatv2di, 2, ALL, v2df) - VAR1 (UNOP, floatunsv4hi, 2, v4hf) - VAR1 (UNOP, floatunsv8hi, 2, v8hf) - VAR1 (UNOP, floatunsv2si, 2, v2sf) - VAR1 (UNOP, floatunsv4si, 2, v4sf) - VAR1 (UNOP, floatunsv2di, 2, v2df) + VAR1 (UNOP, floatunsv4hi, 2, ALL, v4hf) + VAR1 (UNOP, floatunsv8hi, 2, ALL, v8hf) + VAR1 (UNOP, floatunsv2si, 2, ALL, v2sf) + VAR1 (UNOP, floatunsv4si, 2, ALL, v4sf) + VAR1 (UNOP, floatunsv2di, 2, ALL, v2df) - VAR5 (UNOPU, bswap, 2, v4hi, v8hi, v2si, v4si, v2di) + VAR5 (UNOPU, bswap, 2, ALL, v4hi, v8hi, v2si, v4si, v2di) - BUILTIN_VB (UNOP, rbit, 0) + BUILTIN_VB (UNOP, rbit, 0, ALL) /* Implemented by aarch64_. */ - BUILTIN_VALL (BINOP, zip1, 0) - BUILTIN_VALL (BINOP, zip2, 0) - BUILTIN_VALL (BINOP, uzp1, 0) - BUILTIN_VALL (BINOP, uzp2, 0) - BUILTIN_VALL (BINOP, trn1, 0) - BUILTIN_VALL (BINOP, trn2, 0) + BUILTIN_VALL (BINOP, zip1, 0, ALL) + BUILTIN_VALL (BINOP, zip2, 0, ALL) + BUILTIN_VALL (BINOP, uzp1, 0, ALL) + BUILTIN_VALL (BINOP, uzp2, 0, ALL) + BUILTIN_VALL (BINOP, trn1, 0, ALL) + BUILTIN_VALL (BINOP, trn2, 0, ALL) - BUILTIN_GPF_F16 (UNOP, frecpe, 0) - BUILTIN_GPF_F16 (UNOP, frecpx, 0) + BUILTIN_GPF_F16 (UNOP, frecpe, 0, ALL) + BUILTIN_GPF_F16 (UNOP, frecpx, 0, ALL) - BUILTIN_VDQ_SI (UNOP, urecpe, 0) + BUILTIN_VDQ_SI (UNOP, urecpe, 0, ALL) - BUILTIN_VHSDF (UNOP, frecpe, 0) - BUILTIN_VHSDF_HSDF (BINOP, frecps, 0) + BUILTIN_VHSDF (UNOP, frecpe, 0, ALL) + BUILTIN_VHSDF_HSDF (BINOP, frecps, 0, ALL) /* Implemented by a mixture of abs2 patterns. Note the DImode builtin is only ever used for the int64x1_t intrinsic, there is no scalar version. */ - BUILTIN_VSDQ_I_DI (UNOP, abs, 0) - BUILTIN_VHSDF (UNOP, abs, 2) - VAR1 (UNOP, abs, 2, hf) + BUILTIN_VSDQ_I_DI (UNOP, abs, 0, ALL) + BUILTIN_VHSDF (UNOP, abs, 2, ALL) + VAR1 (UNOP, abs, 2, ALL, hf) - BUILTIN_VQ_HSF (UNOP, vec_unpacks_hi_, 10) - VAR1 (BINOP, float_truncate_hi_, 0, v4sf) - VAR1 (BINOP, float_truncate_hi_, 0, v8hf) + BUILTIN_VQ_HSF (UNOP, vec_unpacks_hi_, 10, ALL) + VAR1 (BINOP, float_truncate_hi_, 0, ALL, v4sf) + VAR1 (BINOP, float_truncate_hi_, 0, ALL, v8hf) - VAR1 (UNOP, float_extend_lo_, 0, v2df) - VAR1 (UNOP, float_extend_lo_, 0, v4sf) - BUILTIN_VDF (UNOP, float_truncate_lo_, 0) + VAR1 (UNOP, float_extend_lo_, 0, ALL, v2df) + VAR1 (UNOP, float_extend_lo_, 0, ALL, v4sf) + BUILTIN_VDF (UNOP, float_truncate_lo_, 0, ALL) /* Implemented by aarch64_ld1. */ - BUILTIN_VALL_F16 (LOAD1, ld1, 0) - VAR1(STORE1P, ld1, 0, v2di) + BUILTIN_VALL_F16 (LOAD1, ld1, 0, ALL) + VAR1(STORE1P, ld1, 0, ALL, v2di) /* Implemented by aarch64_st1. */ - BUILTIN_VALL_F16 (STORE1, st1, 0) - VAR1(STORE1P, st1, 0, v2di) + BUILTIN_VALL_F16 (STORE1, st1, 0, ALL) + VAR1(STORE1P, st1, 0, ALL, v2di) /* Implemented by aarch64_ld1x3. */ - BUILTIN_VALLDIF (LOADSTRUCT, ld1x3, 0) + BUILTIN_VALLDIF (LOADSTRUCT, ld1x3, 0, ALL) /* Implemented by aarch64_ld1x4. */ - BUILTIN_VALLDIF (LOADSTRUCT, ld1x4, 0) + BUILTIN_VALLDIF (LOADSTRUCT, ld1x4, 0, ALL) /* Implemented by aarch64_st1x2. */ - BUILTIN_VALLDIF (STORESTRUCT, st1x2, 0) + BUILTIN_VALLDIF (STORESTRUCT, st1x2, 0, ALL) /* Implemented by aarch64_st1x3. */ - BUILTIN_VALLDIF (STORESTRUCT, st1x3, 0) + BUILTIN_VALLDIF (STORESTRUCT, st1x3, 0, ALL) /* Implemented by aarch64_st1x4. */ - BUILTIN_VALLDIF (STORESTRUCT, st1x4, 0) + BUILTIN_VALLDIF (STORESTRUCT, st1x4, 0, ALL) /* Implemented by fma4. */ - BUILTIN_VHSDF (TERNOP, fma, 4) - VAR1 (TERNOP, fma, 4, hf) + BUILTIN_VHSDF (TERNOP, fma, 4, ALL) + VAR1 (TERNOP, fma, 4, ALL, hf) /* Implemented by fnma4. */ - BUILTIN_VHSDF (TERNOP, fnma, 4) - VAR1 (TERNOP, fnma, 4, hf) + BUILTIN_VHSDF (TERNOP, fnma, 4, ALL) + VAR1 (TERNOP, fnma, 4, ALL, hf) /* Implemented by aarch64_simd_bsl. */ - BUILTIN_VDQQH (BSL_P, simd_bsl, 0) - VAR2 (BSL_P, simd_bsl,0, di, v2di) - BUILTIN_VSDQ_I_DI (BSL_U, simd_bsl, 0) - BUILTIN_VALLDIF (BSL_S, simd_bsl, 0) + BUILTIN_VDQQH (BSL_P, simd_bsl, 0, ALL) + VAR2 (BSL_P, simd_bsl,0, ALL, di, v2di) + BUILTIN_VSDQ_I_DI (BSL_U, simd_bsl, 0, ALL) + BUILTIN_VALLDIF (BSL_S, simd_bsl, 0, ALL) /* Implemented by aarch64_crypto_aes. */ - VAR1 (BINOPU, crypto_aese, 0, v16qi) - VAR1 (BINOPU, crypto_aesd, 0, v16qi) - VAR1 (UNOPU, crypto_aesmc, 0, v16qi) - VAR1 (UNOPU, crypto_aesimc, 0, v16qi) + VAR1 (BINOPU, crypto_aese, 0, ALL, v16qi) + VAR1 (BINOPU, crypto_aesd, 0, ALL, v16qi) + VAR1 (UNOPU, crypto_aesmc, 0, ALL, v16qi) + VAR1 (UNOPU, crypto_aesimc, 0, ALL, v16qi) /* Implemented by aarch64_crypto_sha1. */ - VAR1 (UNOPU, crypto_sha1h, 0, si) - VAR1 (BINOPU, crypto_sha1su1, 0, v4si) - VAR1 (TERNOPU, crypto_sha1c, 0, v4si) - VAR1 (TERNOPU, crypto_sha1m, 0, v4si) - VAR1 (TERNOPU, crypto_sha1p, 0, v4si) - VAR1 (TERNOPU, crypto_sha1su0, 0, v4si) + VAR1 (UNOPU, crypto_sha1h, 0, ALL, si) + VAR1 (BINOPU, crypto_sha1su1, 0, ALL, v4si) + VAR1 (TERNOPU, crypto_sha1c, 0, ALL, v4si) + VAR1 (TERNOPU, crypto_sha1m, 0, ALL, v4si) + VAR1 (TERNOPU, crypto_sha1p, 0, ALL, v4si) + VAR1 (TERNOPU, crypto_sha1su0, 0, ALL, v4si) /* Implemented by aarch64_crypto_sha256. */ - VAR1 (TERNOPU, crypto_sha256h, 0, v4si) - VAR1 (TERNOPU, crypto_sha256h2, 0, v4si) - VAR1 (BINOPU, crypto_sha256su0, 0, v4si) - VAR1 (TERNOPU, crypto_sha256su1, 0, v4si) + VAR1 (TERNOPU, crypto_sha256h, 0, ALL, v4si) + VAR1 (TERNOPU, crypto_sha256h2, 0, ALL, v4si) + VAR1 (BINOPU, crypto_sha256su0, 0, ALL, v4si) + VAR1 (TERNOPU, crypto_sha256su1, 0, ALL, v4si) /* Implemented by aarch64_crypto_pmull. */ - VAR1 (BINOPP, crypto_pmull, 0, di) - VAR1 (BINOPP, crypto_pmull, 0, v2di) + VAR1 (BINOPP, crypto_pmull, 0, ALL, di) + VAR1 (BINOPP, crypto_pmull, 0, ALL, v2di) /* Implemented by aarch64_tbl3. */ - VAR1 (BINOP, tbl3, 0, v8qi) - VAR1 (BINOP, tbl3, 0, v16qi) + VAR1 (BINOP, tbl3, 0, ALL, v8qi) + VAR1 (BINOP, tbl3, 0, ALL, v16qi) /* Implemented by aarch64_qtbl3. */ - VAR1 (BINOP, qtbl3, 0, v8qi) - VAR1 (BINOP, qtbl3, 0, v16qi) + VAR1 (BINOP, qtbl3, 0, ALL, v8qi) + VAR1 (BINOP, qtbl3, 0, ALL, v16qi) /* Implemented by aarch64_qtbl4. */ - VAR1 (BINOP, qtbl4, 0, v8qi) - VAR1 (BINOP, qtbl4, 0, v16qi) + VAR1 (BINOP, qtbl4, 0, ALL, v8qi) + VAR1 (BINOP, qtbl4, 0, ALL, v16qi) /* Implemented by aarch64_tbx4. */ - VAR1 (TERNOP, tbx4, 0, v8qi) - VAR1 (TERNOP, tbx4, 0, v16qi) + VAR1 (TERNOP, tbx4, 0, ALL, v8qi) + VAR1 (TERNOP, tbx4, 0, ALL, v16qi) /* Implemented by aarch64_qtbx3. */ - VAR1 (TERNOP, qtbx3, 0, v8qi) - VAR1 (TERNOP, qtbx3, 0, v16qi) + VAR1 (TERNOP, qtbx3, 0, ALL, v8qi) + VAR1 (TERNOP, qtbx3, 0, ALL, v16qi) /* Implemented by aarch64_qtbx4. */ - VAR1 (TERNOP, qtbx4, 0, v8qi) - VAR1 (TERNOP, qtbx4, 0, v16qi) + VAR1 (TERNOP, qtbx4, 0, ALL, v8qi) + VAR1 (TERNOP, qtbx4, 0, ALL, v16qi) /* Builtins for ARMv8.1-A Adv.SIMD instructions. */ /* Implemented by aarch64_sqrdmlh. */ - BUILTIN_VSDQ_HSI (TERNOP, sqrdmlah, 0) - BUILTIN_VSDQ_HSI (TERNOP, sqrdmlsh, 0) + BUILTIN_VSDQ_HSI (TERNOP, sqrdmlah, 0, ALL) + BUILTIN_VSDQ_HSI (TERNOP, sqrdmlsh, 0, ALL) /* Implemented by aarch64_sqrdmlh_lane. */ - BUILTIN_VSDQ_HSI (QUADOP_LANE, sqrdmlah_lane, 0) - BUILTIN_VSDQ_HSI (QUADOP_LANE, sqrdmlsh_lane, 0) + BUILTIN_VSDQ_HSI (QUADOP_LANE, sqrdmlah_lane, 0, ALL) + BUILTIN_VSDQ_HSI (QUADOP_LANE, sqrdmlsh_lane, 0, ALL) /* Implemented by aarch64_sqrdmlh_laneq. */ - BUILTIN_VSDQ_HSI (QUADOP_LANE, sqrdmlah_laneq, 0) - BUILTIN_VSDQ_HSI (QUADOP_LANE, sqrdmlsh_laneq, 0) + BUILTIN_VSDQ_HSI (QUADOP_LANE, sqrdmlah_laneq, 0, ALL) + BUILTIN_VSDQ_HSI (QUADOP_LANE, sqrdmlsh_laneq, 0, ALL) /* Implemented by <*><*>3. */ - BUILTIN_VSDQ_HSDI (SHIFTIMM, scvtf, 3) - BUILTIN_VSDQ_HSDI (FCVTIMM_SUS, ucvtf, 3) - BUILTIN_VHSDF_HSDF (SHIFTIMM, fcvtzs, 3) - BUILTIN_VHSDF_HSDF (SHIFTIMM_USS, fcvtzu, 3) - VAR1 (SHIFTIMM, scvtfsi, 3, hf) - VAR1 (SHIFTIMM, scvtfdi, 3, hf) - VAR1 (FCVTIMM_SUS, ucvtfsi, 3, hf) - VAR1 (FCVTIMM_SUS, ucvtfdi, 3, hf) - BUILTIN_GPI (SHIFTIMM, fcvtzshf, 3) - BUILTIN_GPI (SHIFTIMM_USS, fcvtzuhf, 3) + BUILTIN_VSDQ_HSDI (SHIFTIMM, scvtf, 3, ALL) + BUILTIN_VSDQ_HSDI (FCVTIMM_SUS, ucvtf, 3, ALL) + BUILTIN_VHSDF_HSDF (SHIFTIMM, fcvtzs, 3, ALL) + BUILTIN_VHSDF_HSDF (SHIFTIMM_USS, fcvtzu, 3, ALL) + VAR1 (SHIFTIMM, scvtfsi, 3, ALL, hf) + VAR1 (SHIFTIMM, scvtfdi, 3, ALL, hf) + VAR1 (FCVTIMM_SUS, ucvtfsi, 3, ALL, hf) + VAR1 (FCVTIMM_SUS, ucvtfdi, 3, ALL, hf) + BUILTIN_GPI (SHIFTIMM, fcvtzshf, 3, ALL) + BUILTIN_GPI (SHIFTIMM_USS, fcvtzuhf, 3, ALL) /* Implemented by aarch64_rsqrte. */ - BUILTIN_VHSDF_HSDF (UNOP, rsqrte, 0) + BUILTIN_VHSDF_HSDF (UNOP, rsqrte, 0, ALL) /* Implemented by aarch64_rsqrts. */ - BUILTIN_VHSDF_HSDF (BINOP, rsqrts, 0) + BUILTIN_VHSDF_HSDF (BINOP, rsqrts, 0, ALL) /* Implemented by fabd3. */ - BUILTIN_VHSDF_HSDF (BINOP, fabd, 3) + BUILTIN_VHSDF_HSDF (BINOP, fabd, 3, ALL) /* Implemented by aarch64_faddp. */ - BUILTIN_VHSDF (BINOP, faddp, 0) + BUILTIN_VHSDF (BINOP, faddp, 0, ALL) /* Implemented by aarch64_cm. */ - BUILTIN_VHSDF_HSDF (BINOP_USS, cmeq, 0) - BUILTIN_VHSDF_HSDF (BINOP_USS, cmge, 0) - BUILTIN_VHSDF_HSDF (BINOP_USS, cmgt, 0) - BUILTIN_VHSDF_HSDF (BINOP_USS, cmle, 0) - BUILTIN_VHSDF_HSDF (BINOP_USS, cmlt, 0) + BUILTIN_VHSDF_HSDF (BINOP_USS, cmeq, 0, ALL) + BUILTIN_VHSDF_HSDF (BINOP_USS, cmge, 0, ALL) + BUILTIN_VHSDF_HSDF (BINOP_USS, cmgt, 0, ALL) + BUILTIN_VHSDF_HSDF (BINOP_USS, cmle, 0, ALL) + BUILTIN_VHSDF_HSDF (BINOP_USS, cmlt, 0, ALL) /* Implemented by neg2. */ - BUILTIN_VHSDF_HSDF (UNOP, neg, 2) + BUILTIN_VHSDF_HSDF (UNOP, neg, 2, ALL) /* Implemented by aarch64_fac. */ - BUILTIN_VHSDF_HSDF (BINOP_USS, faclt, 0) - BUILTIN_VHSDF_HSDF (BINOP_USS, facle, 0) - BUILTIN_VHSDF_HSDF (BINOP_USS, facgt, 0) - BUILTIN_VHSDF_HSDF (BINOP_USS, facge, 0) + BUILTIN_VHSDF_HSDF (BINOP_USS, faclt, 0, ALL) + BUILTIN_VHSDF_HSDF (BINOP_USS, facle, 0, ALL) + BUILTIN_VHSDF_HSDF (BINOP_USS, facgt, 0, ALL) + BUILTIN_VHSDF_HSDF (BINOP_USS, facge, 0, ALL) /* Implemented by sqrt2. */ - VAR1 (UNOP, sqrt, 2, hf) + VAR1 (UNOP, sqrt, 2, ALL, hf) /* Implemented by hf2. */ - VAR1 (UNOP, floatdi, 2, hf) - VAR1 (UNOP, floatsi, 2, hf) - VAR1 (UNOP, floathi, 2, hf) - VAR1 (UNOPUS, floatunsdi, 2, hf) - VAR1 (UNOPUS, floatunssi, 2, hf) - VAR1 (UNOPUS, floatunshi, 2, hf) - BUILTIN_GPI_I16 (UNOP, fix_trunchf, 2) - BUILTIN_GPI (UNOP, fix_truncsf, 2) - BUILTIN_GPI (UNOP, fix_truncdf, 2) - BUILTIN_GPI_I16 (UNOPUS, fixuns_trunchf, 2) - BUILTIN_GPI (UNOPUS, fixuns_truncsf, 2) - BUILTIN_GPI (UNOPUS, fixuns_truncdf, 2) + VAR1 (UNOP, floatdi, 2, ALL, hf) + VAR1 (UNOP, floatsi, 2, ALL, hf) + VAR1 (UNOP, floathi, 2, ALL, hf) + VAR1 (UNOPUS, floatunsdi, 2, ALL, hf) + VAR1 (UNOPUS, floatunssi, 2, ALL, hf) + VAR1 (UNOPUS, floatunshi, 2, ALL, hf) + BUILTIN_GPI_I16 (UNOP, fix_trunchf, 2, ALL) + BUILTIN_GPI (UNOP, fix_truncsf, 2, ALL) + BUILTIN_GPI (UNOP, fix_truncdf, 2, ALL) + BUILTIN_GPI_I16 (UNOPUS, fixuns_trunchf, 2, ALL) + BUILTIN_GPI (UNOPUS, fixuns_truncsf, 2, ALL) + BUILTIN_GPI (UNOPUS, fixuns_truncdf, 2, ALL) /* Implemented by aarch64_sm3ss1qv4si. */ - VAR1 (TERNOPU, sm3ss1q, 0, v4si) + VAR1 (TERNOPU, sm3ss1q, 0, ALL, v4si) /* Implemented by aarch64_sm3ttqv4si. */ - VAR1 (QUADOPUI, sm3tt1aq, 0, v4si) - VAR1 (QUADOPUI, sm3tt1bq, 0, v4si) - VAR1 (QUADOPUI, sm3tt2aq, 0, v4si) - VAR1 (QUADOPUI, sm3tt2bq, 0, v4si) + VAR1 (QUADOPUI, sm3tt1aq, 0, ALL, v4si) + VAR1 (QUADOPUI, sm3tt1bq, 0, ALL, v4si) + VAR1 (QUADOPUI, sm3tt2aq, 0, ALL, v4si) + VAR1 (QUADOPUI, sm3tt2bq, 0, ALL, v4si) /* Implemented by aarch64_sm3partwqv4si. */ - VAR1 (TERNOPU, sm3partw1q, 0, v4si) - VAR1 (TERNOPU, sm3partw2q, 0, v4si) + VAR1 (TERNOPU, sm3partw1q, 0, ALL, v4si) + VAR1 (TERNOPU, sm3partw2q, 0, ALL, v4si) /* Implemented by aarch64_sm4eqv4si. */ - VAR1 (BINOPU, sm4eq, 0, v4si) + VAR1 (BINOPU, sm4eq, 0, ALL, v4si) /* Implemented by aarch64_sm4ekeyqv4si. */ - VAR1 (BINOPU, sm4ekeyq, 0, v4si) + VAR1 (BINOPU, sm4ekeyq, 0, ALL, v4si) /* Implemented by aarch64_crypto_sha512hqv2di. */ - VAR1 (TERNOPU, crypto_sha512hq, 0, v2di) + VAR1 (TERNOPU, crypto_sha512hq, 0, ALL, v2di) /* Implemented by aarch64_sha512h2qv2di. */ - VAR1 (TERNOPU, crypto_sha512h2q, 0, v2di) + VAR1 (TERNOPU, crypto_sha512h2q, 0, ALL, v2di) /* Implemented by aarch64_crypto_sha512su0qv2di. */ - VAR1 (BINOPU, crypto_sha512su0q, 0, v2di) + VAR1 (BINOPU, crypto_sha512su0q, 0, ALL, v2di) /* Implemented by aarch64_crypto_sha512su1qv2di. */ - VAR1 (TERNOPU, crypto_sha512su1q, 0, v2di) + VAR1 (TERNOPU, crypto_sha512su1q, 0, ALL, v2di) /* Implemented by eor3q4. */ - BUILTIN_VQ_I (TERNOPU, eor3q, 4) - BUILTIN_VQ_I (TERNOP, eor3q, 4) + BUILTIN_VQ_I (TERNOPU, eor3q, 4, ALL) + BUILTIN_VQ_I (TERNOP, eor3q, 4, ALL) /* Implemented by aarch64_rax1qv2di. */ - VAR1 (BINOPU, rax1q, 0, v2di) + VAR1 (BINOPU, rax1q, 0, ALL, v2di) /* Implemented by aarch64_xarqv2di. */ - VAR1 (TERNOPUI, xarq, 0, v2di) + VAR1 (TERNOPUI, xarq, 0, ALL, v2di) /* Implemented by bcaxq4. */ - BUILTIN_VQ_I (TERNOPU, bcaxq, 4) - BUILTIN_VQ_I (TERNOP, bcaxq, 4) + BUILTIN_VQ_I (TERNOPU, bcaxq, 4, ALL) + BUILTIN_VQ_I (TERNOP, bcaxq, 4, ALL) /* Implemented by aarch64_fmll_low. */ - VAR1 (TERNOP, fmlal_low, 0, v2sf) - VAR1 (TERNOP, fmlsl_low, 0, v2sf) - VAR1 (TERNOP, fmlalq_low, 0, v4sf) - VAR1 (TERNOP, fmlslq_low, 0, v4sf) + VAR1 (TERNOP, fmlal_low, 0, ALL, v2sf) + VAR1 (TERNOP, fmlsl_low, 0, ALL, v2sf) + VAR1 (TERNOP, fmlalq_low, 0, ALL, v4sf) + VAR1 (TERNOP, fmlslq_low, 0, ALL, v4sf) /* Implemented by aarch64_fmll_high. */ - VAR1 (TERNOP, fmlal_high, 0, v2sf) - VAR1 (TERNOP, fmlsl_high, 0, v2sf) - VAR1 (TERNOP, fmlalq_high, 0, v4sf) - VAR1 (TERNOP, fmlslq_high, 0, v4sf) + VAR1 (TERNOP, fmlal_high, 0, ALL, v2sf) + VAR1 (TERNOP, fmlsl_high, 0, ALL, v2sf) + VAR1 (TERNOP, fmlalq_high, 0, ALL, v4sf) + VAR1 (TERNOP, fmlslq_high, 0, ALL, v4sf) /* Implemented by aarch64_fmll_lane_lowv2sf. */ - VAR1 (QUADOP_LANE, fmlal_lane_low, 0, v2sf) - VAR1 (QUADOP_LANE, fmlsl_lane_low, 0, v2sf) + VAR1 (QUADOP_LANE, fmlal_lane_low, 0, ALL, v2sf) + VAR1 (QUADOP_LANE, fmlsl_lane_low, 0, ALL, v2sf) /* Implemented by aarch64_fmll_laneq_lowv2sf. */ - VAR1 (QUADOP_LANE, fmlal_laneq_low, 0, v2sf) - VAR1 (QUADOP_LANE, fmlsl_laneq_low, 0, v2sf) + VAR1 (QUADOP_LANE, fmlal_laneq_low, 0, ALL, v2sf) + VAR1 (QUADOP_LANE, fmlsl_laneq_low, 0, ALL, v2sf) /* Implemented by aarch64_fmllq_lane_lowv4sf. */ - VAR1 (QUADOP_LANE, fmlalq_lane_low, 0, v4sf) - VAR1 (QUADOP_LANE, fmlslq_lane_low, 0, v4sf) + VAR1 (QUADOP_LANE, fmlalq_lane_low, 0, ALL, v4sf) + VAR1 (QUADOP_LANE, fmlslq_lane_low, 0, ALL, v4sf) /* Implemented by aarch64_fmllq_laneq_lowv4sf. */ - VAR1 (QUADOP_LANE, fmlalq_laneq_low, 0, v4sf) - VAR1 (QUADOP_LANE, fmlslq_laneq_low, 0, v4sf) + VAR1 (QUADOP_LANE, fmlalq_laneq_low, 0, ALL, v4sf) + VAR1 (QUADOP_LANE, fmlslq_laneq_low, 0, ALL, v4sf) /* Implemented by aarch64_fmll_lane_highv2sf. */ - VAR1 (QUADOP_LANE, fmlal_lane_high, 0, v2sf) - VAR1 (QUADOP_LANE, fmlsl_lane_high, 0, v2sf) + VAR1 (QUADOP_LANE, fmlal_lane_high, 0, ALL, v2sf) + VAR1 (QUADOP_LANE, fmlsl_lane_high, 0, ALL, v2sf) /* Implemented by aarch64_fmll_laneq_highv2sf. */ - VAR1 (QUADOP_LANE, fmlal_laneq_high, 0, v2sf) - VAR1 (QUADOP_LANE, fmlsl_laneq_high, 0, v2sf) + VAR1 (QUADOP_LANE, fmlal_laneq_high, 0, ALL, v2sf) + VAR1 (QUADOP_LANE, fmlsl_laneq_high, 0, ALL, v2sf) /* Implemented by aarch64_fmllq_lane_highv4sf. */ - VAR1 (QUADOP_LANE, fmlalq_lane_high, 0, v4sf) - VAR1 (QUADOP_LANE, fmlslq_lane_high, 0, v4sf) + VAR1 (QUADOP_LANE, fmlalq_lane_high, 0, ALL, v4sf) + VAR1 (QUADOP_LANE, fmlslq_lane_high, 0, ALL, v4sf) /* Implemented by aarch64_fmllq_laneq_highv4sf. */ - VAR1 (QUADOP_LANE, fmlalq_laneq_high, 0, v4sf) - VAR1 (QUADOP_LANE, fmlslq_laneq_high, 0, v4sf) + VAR1 (QUADOP_LANE, fmlalq_laneq_high, 0, ALL, v4sf) + VAR1 (QUADOP_LANE, fmlslq_laneq_high, 0, ALL, v4sf) /* Implemented by aarch64_. */ - BUILTIN_VSFDF (UNOP, frint32z, 0) - BUILTIN_VSFDF (UNOP, frint32x, 0) - BUILTIN_VSFDF (UNOP, frint64z, 0) - BUILTIN_VSFDF (UNOP, frint64x, 0) + BUILTIN_VSFDF (UNOP, frint32z, 0, ALL) + BUILTIN_VSFDF (UNOP, frint32x, 0, ALL) + BUILTIN_VSFDF (UNOP, frint64z, 0, ALL) + BUILTIN_VSFDF (UNOP, frint64x, 0, ALL) /* Implemented by aarch64_bfdot{_lane}{q}. */ - VAR2 (TERNOP, bfdot, 0, v2sf, v4sf) - VAR2 (QUADOP_LANE_PAIR, bfdot_lane, 0, v2sf, v4sf) - VAR2 (QUADOP_LANE_PAIR, bfdot_laneq, 0, v2sf, v4sf) + VAR2 (TERNOP, bfdot, 0, ALL, v2sf, v4sf) + VAR2 (QUADOP_LANE_PAIR, bfdot_lane, 0, ALL, v2sf, v4sf) + VAR2 (QUADOP_LANE_PAIR, bfdot_laneq, 0, ALL, v2sf, v4sf) /* Implemented by aarch64_bfmmlaqv4sf */ - VAR1 (TERNOP, bfmmlaq, 0, v4sf) + VAR1 (TERNOP, bfmmlaq, 0, ALL, v4sf) /* Implemented by aarch64_bfmlal{_lane{q}}v4sf */ - VAR1 (TERNOP, bfmlalb, 0, v4sf) - VAR1 (TERNOP, bfmlalt, 0, v4sf) - VAR1 (QUADOP_LANE, bfmlalb_lane, 0, v4sf) - VAR1 (QUADOP_LANE, bfmlalt_lane, 0, v4sf) - VAR1 (QUADOP_LANE, bfmlalb_lane_q, 0, v4sf) - VAR1 (QUADOP_LANE, bfmlalt_lane_q, 0, v4sf) + VAR1 (TERNOP, bfmlalb, 0, ALL, v4sf) + VAR1 (TERNOP, bfmlalt, 0, ALL, v4sf) + VAR1 (QUADOP_LANE, bfmlalb_lane, 0, ALL, v4sf) + VAR1 (QUADOP_LANE, bfmlalt_lane, 0, ALL, v4sf) + VAR1 (QUADOP_LANE, bfmlalb_lane_q, 0, ALL, v4sf) + VAR1 (QUADOP_LANE, bfmlalt_lane_q, 0, ALL, v4sf) /* Implemented by aarch64_simd_mmlav16qi. */ - VAR1 (TERNOP, simd_smmla, 0, v16qi) - VAR1 (TERNOPU, simd_ummla, 0, v16qi) - VAR1 (TERNOP_SSUS, simd_usmmla, 0, v16qi) + VAR1 (TERNOP, simd_smmla, 0, ALL, v16qi) + VAR1 (TERNOPU, simd_ummla, 0, ALL, v16qi) + VAR1 (TERNOP_SSUS, simd_usmmla, 0, ALL, v16qi) /* Implemented by aarch64_bfcvtn{q}{2} */ - VAR1 (UNOP, bfcvtn, 0, v4bf) - VAR1 (UNOP, bfcvtn_q, 0, v8bf) - VAR1 (BINOP, bfcvtn2, 0, v8bf) - VAR1 (UNOP, bfcvt, 0, bf) + VAR1 (UNOP, bfcvtn, 0, ALL, v4bf) + VAR1 (UNOP, bfcvtn_q, 0, ALL, v8bf) + VAR1 (BINOP, bfcvtn2, 0, ALL, v8bf) + VAR1 (UNOP, bfcvt, 0, ALL, bf) diff --git a/gcc/config/aarch64/geniterators.sh b/gcc/config/aarch64/geniterators.sh index a7420964f85..43feb482ce9 100644 --- a/gcc/config/aarch64/geniterators.sh +++ b/gcc/config/aarch64/geniterators.sh @@ -70,8 +70,8 @@ iterdef { sub(/ *\]/, "", s) n = split(s, a) - printf "#define BUILTIN_" a[1] "(T, N, MAP) \\\n" - printf " VAR" (n-1) " (T, N, MAP" + printf "#define BUILTIN_" a[1] "(T, N, MAP, FLAG) \\\n" + printf " VAR" (n-1) " (T, N, MAP, FLAG" for (i = 2; i <= n; i++) printf ", " tolower(a[i]) printf ")\n" -- 2.30.2