Added few more stubs so that control reaches to DestroyDevice().
[mesa.git] / src / compiler / nir / nir_lower_bit_size.c
index 531e6aa5289f7f4a1a30120d56db7de86ecb9f4c..d145cc1ada1cd3ea615a9b78f47a02115043b87f 100644 (file)
  * a bit-size that is supported natively and then converts the result back to
  * the original bit-size.
  */
-static nir_ssa_def *
-convert_to_bit_size(nir_builder *bld,
-                    nir_ssa_def *src,
-                    nir_alu_type type,
-                    unsigned bit_size)
-{
-   nir_alu_type base_type = nir_alu_type_get_base_type(type);
-   nir_alu_type lowered_type = bit_size | base_type;
-
-   nir_op opcode =
-      nir_type_conversion_op(type, lowered_type, nir_rounding_mode_undef);
-
-   return nir_build_alu(bld, opcode, src, NULL, NULL, NULL);
-}
 
 static void
 lower_instr(nir_builder *bld, nir_alu_instr *alu, unsigned bit_size)
 {
    const nir_op op = alu->op;
+   unsigned dst_bit_size = alu->dest.dest.ssa.bit_size;
 
    bld->cursor = nir_before_instr(&alu->instr);
 
    /* Convert each source to the requested bit-size */
-   nir_ssa_def *srcs[4] = { NULL, NULL, NULL, NULL };
+   nir_ssa_def *srcs[NIR_MAX_VEC_COMPONENTS] = { NULL };
    for (unsigned i = 0; i < nir_op_infos[op].num_inputs; i++) {
       nir_ssa_def *src = nir_ssa_for_alu_src(bld, alu, i);
 
       nir_alu_type type = nir_op_infos[op].input_types[i];
       if (nir_alu_type_get_type_size(type) == 0)
-         srcs[i] = convert_to_bit_size(bld, src, type, bit_size);
-      else
-         srcs[i] = src;
+         src = nir_convert_to_bit_size(bld, src, type, bit_size);
+
+      if (i == 1 && (op == nir_op_ishl || op == nir_op_ishr || op == nir_op_ushr)) {
+         assert(util_is_power_of_two_nonzero(dst_bit_size));
+         src = nir_iand(bld, src, nir_imm_int(bld, dst_bit_size - 1));
+      }
+
+      srcs[i] = src;
    }
 
    /* Emit the lowered ALU instruction */
-   nir_ssa_def *lowered_dst =
-      nir_build_alu(bld, op, srcs[0], srcs[1], srcs[2], srcs[3]);
+   nir_ssa_def *lowered_dst = NULL;
+   if (op == nir_op_imul_high || op == nir_op_umul_high) {
+      assert(dst_bit_size * 2 <= bit_size);
+      nir_ssa_def *lowered_dst = nir_imul(bld, srcs[0], srcs[1]);
+      if (nir_op_infos[op].output_type & nir_type_uint)
+         lowered_dst = nir_ushr_imm(bld, lowered_dst, dst_bit_size);
+      else
+         lowered_dst = nir_ishr_imm(bld, lowered_dst, dst_bit_size);
+   } else {
+      lowered_dst = nir_build_alu_src_arr(bld, op, srcs);
+   }
+
 
    /* Convert result back to the original bit-size */
-   unsigned dst_bit_size = alu->dest.dest.ssa.bit_size;
-   nir_alu_type type = nir_op_infos[op].output_type;
-   nir_ssa_def *dst = convert_to_bit_size(bld, lowered_dst, type, dst_bit_size);
-   nir_ssa_def_rewrite_uses(&alu->dest.dest.ssa, nir_src_for_ssa(dst));
+   if (dst_bit_size != bit_size) {
+      nir_alu_type type = nir_op_infos[op].output_type;
+      nir_ssa_def *dst = nir_convert_to_bit_size(bld, lowered_dst, type, dst_bit_size);
+      nir_ssa_def_rewrite_uses(&alu->dest.dest.ssa, nir_src_for_ssa(dst));
+   } else {
+      nir_ssa_def_rewrite_uses(&alu->dest.dest.ssa, nir_src_for_ssa(lowered_dst));
+   }
 }
 
 static bool
@@ -95,8 +100,6 @@ lower_impl(nir_function_impl *impl,
          if (lower_bit_size == 0)
             continue;
 
-         assert(lower_bit_size != alu->dest.dest.ssa.bit_size);
-
          lower_instr(&b, alu, lower_bit_size);
          progress = true;
       }
@@ -105,6 +108,8 @@ lower_impl(nir_function_impl *impl,
    if (progress) {
       nir_metadata_preserve(impl, nir_metadata_block_index |
                                   nir_metadata_dominance);
+   } else {
+      nir_metadata_preserve(impl, nir_metadata_all);
    }
 
    return progress;
@@ -124,3 +129,93 @@ nir_lower_bit_size(nir_shader *shader,
 
    return progress;
 }
+
+static void
+split_phi(nir_builder *b, nir_phi_instr *phi)
+{
+   nir_phi_instr *lowered[2] = {
+      nir_phi_instr_create(b->shader),
+      nir_phi_instr_create(b->shader)
+   };
+   int num_components = phi->dest.ssa.num_components;
+   assert(phi->dest.ssa.bit_size == 64);
+
+   nir_foreach_phi_src(src, phi) {
+      assert(num_components == src->src.ssa->num_components);
+
+      b->cursor = nir_before_src(&src->src, false);
+
+      nir_ssa_def *x = nir_unpack_64_2x32_split_x(b, src->src.ssa);
+      nir_ssa_def *y = nir_unpack_64_2x32_split_y(b, src->src.ssa);
+
+      nir_phi_src *xsrc = rzalloc(lowered[0], nir_phi_src);
+      xsrc->pred = src->pred;
+      xsrc->src = nir_src_for_ssa(x);
+      exec_list_push_tail(&lowered[0]->srcs, &xsrc->node);
+
+      nir_phi_src *ysrc = rzalloc(lowered[1], nir_phi_src);
+      ysrc->pred = src->pred;
+      ysrc->src = nir_src_for_ssa(y);
+      exec_list_push_tail(&lowered[1]->srcs, &ysrc->node);
+   }
+
+   nir_ssa_dest_init(&lowered[0]->instr, &lowered[0]->dest,
+                     num_components, 32, NULL);
+   nir_ssa_dest_init(&lowered[1]->instr, &lowered[1]->dest,
+                     num_components, 32, NULL);
+
+   b->cursor = nir_before_instr(&phi->instr);
+   nir_builder_instr_insert(b, &lowered[0]->instr);
+   nir_builder_instr_insert(b, &lowered[1]->instr);
+
+   b->cursor = nir_after_phis(nir_cursor_current_block(b->cursor));
+   nir_ssa_def *merged = nir_pack_64_2x32_split(b, &lowered[0]->dest.ssa, &lowered[1]->dest.ssa);
+   nir_ssa_def_rewrite_uses(&phi->dest.ssa, nir_src_for_ssa(merged));
+   nir_instr_remove(&phi->instr);
+}
+
+static bool
+lower_64bit_phi_impl(nir_function_impl *impl)
+{
+   nir_builder b;
+   nir_builder_init(&b, impl);
+   bool progress = false;
+
+   nir_foreach_block(block, impl) {
+      nir_foreach_instr_safe(instr, block) {
+         if (instr->type != nir_instr_type_phi)
+            break;
+
+         nir_phi_instr *phi = nir_instr_as_phi(instr);
+         assert(phi->dest.is_ssa);
+
+         if (phi->dest.ssa.bit_size <= 32)
+            continue;
+
+         split_phi(&b, phi);
+         progress = true;
+      }
+   }
+
+   if (progress) {
+      nir_metadata_preserve(impl, nir_metadata_block_index |
+                                  nir_metadata_dominance);
+   } else {
+      nir_metadata_preserve(impl, nir_metadata_all);
+   }
+
+   return progress;
+}
+
+bool
+nir_lower_64bit_phis(nir_shader *shader)
+{
+   bool progress = false;
+
+   nir_foreach_function(function, shader) {
+      if (function->impl)
+         progress |= lower_64bit_phi_impl(function->impl);
+   }
+
+   return progress;
+}