+2020-04-16 Richard Sandiford <richard.sandiford@arm.com>
+
+ PR target/94606
+ * config/aarch64/aarch64.c (aarch64_expand_sve_const_pred_eor): Take
+ the VNx16BI lowpart of the recursively-generated constant.
+
2020-04-16 Martin Liska <mliska@suse.cz>
Jakub Jelinek <jakub@redhat.com>
/* EOR the result with an ELT_SIZE PTRUE. */
rtx mask = aarch64_ptrue_all (elt_size);
mask = force_reg (VNx16BImode, mask);
+ inv = gen_lowpart (VNx16BImode, inv);
target = aarch64_target_reg (target, VNx16BImode);
emit_insn (gen_aarch64_pred_z (XOR, VNx16BImode, target, mask, inv, mask));
return target;
+2020-04-16 Richard Sandiford <richard.sandiford@arm.com>
+
+ PR target/94606
+ * gcc.dg/vect/pr94606.c: New test.
+
2020-04-16 Martin Liska <mliska@suse.cz>
Jakub Jelinek <jakub@redhat.com>
--- /dev/null
+/* { dg-do compile } */
+/* { dg-additional-options "-march=armv8.2-a+sve -msve-vector-bits=256" { target aarch64*-*-* } } */
+
+const short mask[] = { 0, 0, 0, 0, 0, 0, 0, 0,
+ 0, 0, 0, 1, 1, 1, 1, 1 };
+
+int
+foo (short *restrict x, short *restrict y)
+{
+ for (int i = 0; i < 16; ++i)
+ if (mask[i])
+ x[i] += y[i];
+}