void mir_rewrite_index_src_single(midgard_instruction *ins, unsigned old, unsigned new)
{
- if (ins->ssa_args.src0 == old)
- ins->ssa_args.src0 = new;
+ for (unsigned i = 0; i < ARRAY_SIZE(ins->src); ++i) {
+ if (ins->src[i] == old)
+ ins->src[i] = new;
+ }
+}
+
+void mir_rewrite_index_dst_single(midgard_instruction *ins, unsigned old, unsigned new)
+{
+ if (ins->dest == old)
+ ins->dest = new;
+}
+
+unsigned
+mir_get_swizzle(midgard_instruction *ins, unsigned idx)
+{
+ if (ins->type == TAG_ALU_4) {
+ if (idx == 2)
+ return ins->csel_swizzle;
+
+ unsigned b = (idx == 0) ? ins->alu.src1 : ins->alu.src2;
+
+ midgard_vector_alu_src s =
+ vector_alu_from_unsigned(b);
+
+ return s.swizzle;
+ } else if (ins->type == TAG_LOAD_STORE_4) {
+ /* Main swizzle of a load is on the destination */
+ if (!OP_IS_STORE(ins->load_store.op))
+ idx++;
+
+ switch (idx) {
+ case 0:
+ return ins->load_store.swizzle;
+ case 1:
+ case 2: {
+ uint8_t raw =
+ (idx == 2) ? ins->load_store.arg_2 : ins->load_store.arg_1;
+
+ return component_to_swizzle(midgard_ldst_select(raw).component);
+ }
+ default:
+ unreachable("Unknown load/store source");
+ }
+ } else if (ins->type == TAG_TEXTURE_4) {
+ switch (idx) {
+ case 0:
+ return ins->texture.in_reg_swizzle;
+ case 1:
+ /* Swizzle on bias doesn't make sense */
+ return 0;
+ default:
+ unreachable("Unknown texture source");
+ }
+ } else {
+ unreachable("Unknown type");
+ }
+}
+
+void
+mir_set_swizzle(midgard_instruction *ins, unsigned idx, unsigned new)
+{
+ if (ins->type == TAG_ALU_4) {
+ unsigned b = (idx == 0) ? ins->alu.src1 : ins->alu.src2;
+
+ midgard_vector_alu_src s =
+ vector_alu_from_unsigned(b);
+
+ s.swizzle = new;
+ unsigned pack = vector_alu_srco_unsigned(s);
+
+ if (idx == 0)
+ ins->alu.src1 = pack;
+ else
+ ins->alu.src2 = pack;
+ } else if (ins->type == TAG_LOAD_STORE_4) {
+ /* Main swizzle of a load is on the destination */
+ if (!OP_IS_STORE(ins->load_store.op))
+ idx++;
+
+ switch (idx) {
+ case 0:
+ ins->load_store.swizzle = new;
+ break;
+ case 1:
+ case 2: {
+ uint8_t raw =
+ (idx == 2) ? ins->load_store.arg_2 : ins->load_store.arg_1;
+
+ midgard_ldst_register_select sel
+ = midgard_ldst_select(raw);
+ sel.component = swizzle_to_component(new);
+ uint8_t packed = midgard_ldst_pack(sel);
- if (ins->ssa_args.src1 == old &&
- !ins->ssa_args.inline_constant)
- ins->ssa_args.src1 = new;
+ if (idx == 2)
+ ins->load_store.arg_2 = packed;
+ else
+ ins->load_store.arg_1 = packed;
+
+ break;
+ }
+ default:
+ assert(new == 0);
+ break;
+ }
+ } else if (ins->type == TAG_TEXTURE_4) {
+ switch (idx) {
+ case 0:
+ ins->texture.in_reg_swizzle = new;
+ break;
+ default:
+ assert(new == 0);
+ break;
+ }
+ } else {
+ unreachable("Unknown type");
+ }
}
+static void
+mir_rewrite_index_src_single_swizzle(midgard_instruction *ins, unsigned old, unsigned new, unsigned swizzle)
+{
+ for (unsigned i = 0; i < ARRAY_SIZE(ins->src); ++i) {
+ if (ins->src[i] != old) continue;
+
+ ins->src[i] = new;
+
+ mir_set_swizzle(ins, i,
+ pan_compose_swizzle(mir_get_swizzle(ins, i), swizzle));
+ }
+}
void
mir_rewrite_index_src(compiler_context *ctx, unsigned old, unsigned new)
}
}
+void
+mir_rewrite_index_src_swizzle(compiler_context *ctx, unsigned old, unsigned new, unsigned swizzle)
+{
+ mir_foreach_instr_global(ctx, ins) {
+ mir_rewrite_index_src_single_swizzle(ins, old, new, swizzle);
+ }
+}
+
void
mir_rewrite_index_src_tag(compiler_context *ctx, unsigned old, unsigned new, unsigned tag)
{
mir_rewrite_index_dst(compiler_context *ctx, unsigned old, unsigned new)
{
mir_foreach_instr_global(ctx, ins) {
- if (ins->ssa_args.dest == old)
- ins->ssa_args.dest = new;
+ mir_rewrite_index_dst_single(ins, old, new);
}
}
if (ins->type != tag)
continue;
- if (ins->ssa_args.dest == old)
- ins->ssa_args.dest = new;
+ if (ins->dest == old)
+ ins->dest = new;
}
}
bool
mir_single_use(compiler_context *ctx, unsigned value)
{
+ /* We can replicate constants in places so who cares */
+ if (value == SSA_FIXED_REGISTER(REGISTER_CONSTANT))
+ return true;
+
return mir_use_count(ctx, value) <= 1;
}
+static bool
+mir_nontrivial_raw_mod(midgard_vector_alu_src src, bool is_int)
+{
+ if (is_int)
+ return src.mod == midgard_int_shift;
+ else
+ return src.mod;
+}
+
bool
mir_nontrivial_mod(midgard_vector_alu_src src, bool is_int, unsigned mask)
{
- /* abs or neg */
- if (!is_int && src.mod) return true;
-
- /* Other int mods don't matter in isolation */
- if (is_int && src.mod == midgard_int_shift) return true;
+ if (mir_nontrivial_raw_mod(src, is_int)) return true;
/* size-conversion */
if (src.half) return true;
return mir_nontrivial_mod(src2, is_int, ins->mask);
}
+bool
+mir_nontrivial_source2_mod_simple(midgard_instruction *ins)
+{
+ bool is_int = midgard_is_integer_op(ins->alu.op);
+
+ midgard_vector_alu_src src2 =
+ vector_alu_from_unsigned(ins->alu.src2);
+
+ return mir_nontrivial_raw_mod(src2, is_int) || src2.half;
+}
+
bool
mir_nontrivial_outmod(midgard_instruction *ins)
{
bool is_int = midgard_is_integer_op(ins->alu.op);
unsigned mod = ins->alu.outmod;
+ /* Pseudo-outmod */
+ if (ins->invert)
+ return true;
+
/* Type conversion is a sort of outmod */
if (ins->alu.dest_override != midgard_dest_override_none)
return true;
mir_foreach_instr_global(ctx, ins) {
bool is_ldst = ins->type == TAG_LOAD_STORE_4;
bool is_tex = ins->type == TAG_TEXTURE_4;
+ bool is_writeout = ins->compact_branch && ins->writeout;
- if (!(is_ldst || is_tex))
+ if (!(is_ldst || is_tex || is_writeout))
continue;
if (mir_has_arg(ins, idx))
bool
mir_is_written_before(compiler_context *ctx, midgard_instruction *ins, unsigned node)
{
- if ((node < 0) || (node >= SSA_FIXED_MINIMUM))
+ if (node >= SSA_FIXED_MINIMUM)
return true;
mir_foreach_instr_global(ctx, q) {
if (q == ins)
break;
- if (q->ssa_args.dest == node)
+ if (q->dest == node)
return true;
}
return false;
}
+
+/* Creates a mask of the components of a node read by an instruction, by
+ * analyzing the swizzle with respect to the instruction's mask. E.g.:
+ *
+ * fadd r0.xz, r1.yyyy, r2.zwyx
+ *
+ * will return a mask of Z/Y for r2
+ */
+
+static unsigned
+mir_mask_of_read_components_single(unsigned swizzle, unsigned outmask)
+{
+ unsigned mask = 0;
+
+ for (unsigned c = 0; c < 4; ++c) {
+ if (!(outmask & (1 << c))) continue;
+
+ unsigned comp = (swizzle >> (2*c)) & 3;
+ mask |= (1 << comp);
+ }
+
+ return mask;
+}
+
+static unsigned
+mir_source_count(midgard_instruction *ins)
+{
+ if (ins->type == TAG_ALU_4) {
+ /* ALU is always binary */
+ return 2;
+ } else if (ins->type == TAG_LOAD_STORE_4) {
+ bool load = !OP_IS_STORE(ins->load_store.op);
+ return (load ? 2 : 3);
+ } else if (ins->type == TAG_TEXTURE_4) {
+ /* Coords, bias.. TODO: Offsets? */
+ return 2;
+ } else {
+ unreachable("Invalid instruction type");
+ }
+}
+
+static unsigned
+mir_component_count_implicit(midgard_instruction *ins, unsigned i)
+{
+ if (ins->type == TAG_LOAD_STORE_4) {
+ switch (ins->load_store.op) {
+ /* Address implicitly 64-bit */
+ case midgard_op_ld_int4:
+ return (i == 0) ? 1 : 0;
+
+ case midgard_op_st_int4:
+ return (i == 1) ? 1 : 0;
+
+ default:
+ return 0;
+ }
+ }
+
+ return 0;
+}
+
+unsigned
+mir_mask_of_read_components(midgard_instruction *ins, unsigned node)
+{
+ unsigned mask = 0;
+
+ for (unsigned i = 0; i < mir_source_count(ins); ++i) {
+ if (ins->src[i] != node) continue;
+
+ unsigned swizzle = mir_get_swizzle(ins, i);
+ unsigned m = mir_mask_of_read_components_single(swizzle, ins->mask);
+
+ /* Sometimes multi-arg ops are passed implicitly */
+ unsigned implicit = mir_component_count_implicit(ins, i);
+ assert(implicit < 2);
+
+ /* Extend the mask */
+ if (implicit == 1) {
+ /* Ensure it's a single bit currently */
+ assert((m >> __builtin_ctz(m)) == 0x1);
+
+ /* Set the next bit to extend one*/
+ m |= (m << 1);
+ }
+
+ mask |= m;
+ }
+
+ return mask;
+}
+
+unsigned
+mir_ubo_shift(midgard_load_store_op op)
+{
+ switch (op) {
+ case midgard_op_ld_ubo_char:
+ return 0;
+ case midgard_op_ld_ubo_char2:
+ return 1;
+ case midgard_op_ld_ubo_char4:
+ return 2;
+ case midgard_op_ld_ubo_short4:
+ return 3;
+ case midgard_op_ld_ubo_int4:
+ return 4;
+ default:
+ unreachable("Invalid op");
+ }
+}
+
+/* Register allocation occurs after instruction scheduling, which is fine until
+ * we start needing to spill registers and therefore insert instructions into
+ * an already-scheduled program. We don't have to be terribly efficient about
+ * this, since spilling is already slow. So just semantically we need to insert
+ * the instruction into a new bundle before/after the bundle of the instruction
+ * in question */
+
+static midgard_bundle
+mir_bundle_for_op(compiler_context *ctx, midgard_instruction ins)
+{
+ midgard_instruction *u = mir_upload_ins(ctx, ins);
+
+ midgard_bundle bundle = {
+ .tag = ins.type,
+ .instruction_count = 1,
+ .instructions = { u },
+ };
+
+ if (bundle.tag == TAG_ALU_4) {
+ assert(OP_IS_MOVE(u->alu.op));
+ u->unit = UNIT_VMUL;
+
+ size_t bytes_emitted = sizeof(uint32_t) + sizeof(midgard_reg_info) + sizeof(midgard_vector_alu);
+ bundle.padding = ~(bytes_emitted - 1) & 0xF;
+ bundle.control = ins.type | u->unit;
+ }
+
+ return bundle;
+}
+
+static unsigned
+mir_bundle_idx_for_ins(midgard_instruction *tag, midgard_block *block)
+{
+ midgard_bundle *bundles =
+ (midgard_bundle *) block->bundles.data;
+
+ size_t count = (block->bundles.size / sizeof(midgard_bundle));
+
+ for (unsigned i = 0; i < count; ++i) {
+ for (unsigned j = 0; j < bundles[i].instruction_count; ++j) {
+ if (bundles[i].instructions[j] == tag)
+ return i;
+ }
+ }
+
+ mir_print_instruction(tag);
+ unreachable("Instruction not scheduled in block");
+}
+
+void
+mir_insert_instruction_before_scheduled(
+ compiler_context *ctx,
+ midgard_block *block,
+ midgard_instruction *tag,
+ midgard_instruction ins)
+{
+ unsigned before = mir_bundle_idx_for_ins(tag, block);
+ size_t count = util_dynarray_num_elements(&block->bundles, midgard_bundle);
+ UNUSED void *unused = util_dynarray_grow(&block->bundles, midgard_bundle, 1);
+
+ midgard_bundle *bundles = (midgard_bundle *) block->bundles.data;
+ memmove(bundles + before + 1, bundles + before, (count - before) * sizeof(midgard_bundle));
+ midgard_bundle *before_bundle = bundles + before + 1;
+
+ midgard_bundle new = mir_bundle_for_op(ctx, ins);
+ memcpy(bundles + before, &new, sizeof(new));
+
+ list_addtail(&new.instructions[0]->link, &before_bundle->instructions[0]->link);
+}
+
+void
+mir_insert_instruction_after_scheduled(
+ compiler_context *ctx,
+ midgard_block *block,
+ midgard_instruction *tag,
+ midgard_instruction ins)
+{
+ unsigned after = mir_bundle_idx_for_ins(tag, block);
+ size_t count = util_dynarray_num_elements(&block->bundles, midgard_bundle);
+ UNUSED void *unused = util_dynarray_grow(&block->bundles, midgard_bundle, 1);
+
+ midgard_bundle *bundles = (midgard_bundle *) block->bundles.data;
+ memmove(bundles + after + 2, bundles + after + 1, (count - after - 1) * sizeof(midgard_bundle));
+ midgard_bundle *after_bundle_1 = bundles + after + 2;
+
+ midgard_bundle new = mir_bundle_for_op(ctx, ins);
+ memcpy(bundles + after + 1, &new, sizeof(new));
+ list_addtail(&new.instructions[0]->link, &after_bundle_1->instructions[0]->link);
+}