--- /dev/null
+/*
+ * Copyright © 2018 Intel Corporation
+ *
+ * Permission is hereby granted, free of charge, to any person obtaining a
+ * copy of this software and associated documentation files (the "Software"),
+ * to deal in the Software without restriction, including without limitation
+ * the rights to use, copy, modify, merge, publish, distribute, sublicense,
+ * and/or sell copies of the Software, and to permit persons to whom the
+ * Software is furnished to do so, subject to the following conditions:
+ *
+ * The above copyright notice and this permission notice (including the next
+ * paragraph) shall be included in all copies or substantial portions of the
+ * Software.
+ *
+ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
+ * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
+ * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
+ * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
+ * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
+ * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
+ * IN THE SOFTWARE.
+ */
+
+#include "nir.h"
+#include "nir_builder.h"
+
+nir_deref_var *
+nir_deref_instr_to_deref(nir_deref_instr *instr, void *mem_ctx)
+{
+ nir_deref *deref = NULL;
+
+ while (instr->deref_type != nir_deref_type_var) {
+ nir_deref *nderef;
+ switch (instr->deref_type) {
+ case nir_deref_type_array:
+ case nir_deref_type_array_wildcard: {
+ nir_deref_array *deref_arr = nir_deref_array_create(mem_ctx);
+ if (instr->deref_type == nir_deref_type_array) {
+ nir_const_value *const_index =
+ nir_src_as_const_value(instr->arr.index);
+ if (const_index) {
+ deref_arr->deref_array_type = nir_deref_array_type_direct;
+ deref_arr->base_offset = const_index->u32[0];
+ } else {
+ deref_arr->deref_array_type = nir_deref_array_type_indirect;
+ deref_arr->base_offset = 0;
+ nir_src_copy(&deref_arr->indirect, &instr->arr.index, mem_ctx);
+ }
+ } else {
+ deref_arr->deref_array_type = nir_deref_array_type_wildcard;
+ }
+ nderef = &deref_arr->deref;
+ break;
+ }
+
+ case nir_deref_type_struct:
+ nderef = &nir_deref_struct_create(mem_ctx, instr->strct.index)->deref;
+ break;
+
+ default:
+ unreachable("Invalid deref instruction type");
+ }
+
+ nderef->child = deref;
+ ralloc_steal(nderef, deref);
+ nderef->type = instr->type;
+
+ deref = nderef;
+ assert(instr->parent.is_ssa);
+ instr = nir_src_as_deref(instr->parent);
+ }
+
+ assert(instr->deref_type == nir_deref_type_var);
+ nir_deref_var *deref_var = nir_deref_var_create(mem_ctx, instr->var);
+ deref_var->deref.child = deref;
+ ralloc_steal(deref_var, deref);
+
+ return deref_var;
+}
+
+static nir_deref_var *
+nir_deref_src_to_deref(nir_src src, void *mem_ctx)
+{
+ return nir_deref_instr_to_deref(nir_src_as_deref(src), mem_ctx);
+}
+
+static bool
+nir_lower_deref_instrs_tex(nir_tex_instr *tex)
+{
+ bool progress = false;
+
+ /* Remove the instruction before we modify it. This way we won't mess up
+ * use-def chains when we move sources around.
+ */
+ nir_cursor cursor = nir_instr_remove(&tex->instr);
+
+ unsigned new_num_srcs = 0;
+ for (unsigned i = 0; i < tex->num_srcs; i++) {
+ if (tex->src[i].src_type == nir_tex_src_texture_deref) {
+ tex->texture = nir_deref_src_to_deref(tex->src[i].src, tex);
+ progress = true;
+ continue;
+ } else if (tex->src[i].src_type == nir_tex_src_sampler_deref) {
+ tex->sampler = nir_deref_src_to_deref(tex->src[i].src, tex);
+ progress = true;
+ continue;
+ }
+
+ /* Compact the sources down to remove the deref sources */
+ assert(new_num_srcs <= i);
+ tex->src[new_num_srcs++] = tex->src[i];
+ }
+ tex->num_srcs = new_num_srcs;
+
+ nir_instr_insert(cursor, &tex->instr);
+
+ return progress;
+}
+
+static bool
+nir_lower_deref_instrs_intrin(nir_intrinsic_instr *intrin,
+ enum nir_lower_deref_flags flags)
+{
+ nir_intrinsic_op deref_op = intrin->intrinsic;
+ nir_intrinsic_op var_op;
+
+ switch (deref_op) {
+#define CASE(a) \
+ case nir_intrinsic_##a##_deref: \
+ if (!(flags & nir_lower_load_store_derefs)) \
+ return false; \
+ var_op = nir_intrinsic_##a##_var; \
+ break;
+ CASE(load)
+ CASE(store)
+ CASE(copy)
+#undef CASE
+
+#define CASE(a) \
+ case nir_intrinsic_interp_deref_##a: \
+ if (!(flags & nir_lower_interp_derefs)) \
+ return false; \
+ var_op = nir_intrinsic_interp_var_##a; \
+ break;
+ CASE(at_centroid)
+ CASE(at_sample)
+ CASE(at_offset)
+#undef CASE
+
+#define CASE(a) \
+ case nir_intrinsic_atomic_counter_##a##_deref: \
+ if (!(flags & nir_lower_atomic_counter_derefs)) \
+ return false; \
+ var_op = nir_intrinsic_atomic_counter_##a##_var; \
+ break;
+ CASE(inc)
+ CASE(dec)
+ CASE(read)
+ CASE(add)
+ CASE(min)
+ CASE(max)
+ CASE(and)
+ CASE(or)
+ CASE(xor)
+ CASE(exchange)
+ CASE(comp_swap)
+#undef CASE
+
+#define CASE(a) \
+ case nir_intrinsic_deref_atomic_##a: \
+ if (!(flags & nir_lower_atomic_derefs)) \
+ return false; \
+ var_op = nir_intrinsic_var_atomic_##a; \
+ break;
+ CASE(add)
+ CASE(imin)
+ CASE(umin)
+ CASE(imax)
+ CASE(umax)
+ CASE(and)
+ CASE(or)
+ CASE(xor)
+ CASE(exchange)
+ CASE(comp_swap)
+#undef CASE
+
+#define CASE(a) \
+ case nir_intrinsic_image_deref_##a: \
+ if (!(flags & nir_lower_image_derefs)) \
+ return false; \
+ var_op = nir_intrinsic_image_var_##a; \
+ break;
+ CASE(load)
+ CASE(store)
+ CASE(atomic_add)
+ CASE(atomic_min)
+ CASE(atomic_max)
+ CASE(atomic_and)
+ CASE(atomic_or)
+ CASE(atomic_xor)
+ CASE(atomic_exchange)
+ CASE(atomic_comp_swap)
+ CASE(size)
+ CASE(samples)
+#undef CASE
+
+ default:
+ return false;
+ }
+
+ /* Remove the instruction before we modify it. This way we won't mess up
+ * use-def chains when we move sources around.
+ */
+ nir_cursor cursor = nir_instr_remove(&intrin->instr);
+
+ unsigned num_derefs = nir_intrinsic_infos[var_op].num_variables;
+ assert(nir_intrinsic_infos[var_op].num_srcs + num_derefs ==
+ nir_intrinsic_infos[deref_op].num_srcs);
+
+ /* Move deref sources to variables */
+ for (unsigned i = 0; i < num_derefs; i++)
+ intrin->variables[i] = nir_deref_src_to_deref(intrin->src[i], intrin);
+
+ /* Shift all the other sources down */
+ for (unsigned i = 0; i < nir_intrinsic_infos[var_op].num_srcs; i++)
+ nir_src_copy(&intrin->src[i], &intrin->src[i + num_derefs], intrin);
+
+ /* Rewrite the extra sources to NIR_SRC_INIT just in case */
+ for (unsigned i = 0; i < num_derefs; i++)
+ intrin->src[nir_intrinsic_infos[var_op].num_srcs + i] = NIR_SRC_INIT;
+
+ /* It's safe to just stomp the intrinsic to var intrinsic since every
+ * intrinsic has room for some variables and the number of sources only
+ * shrinks.
+ */
+ intrin->intrinsic = var_op;
+
+ nir_instr_insert(cursor, &intrin->instr);
+
+ return true;
+}
+
+static bool
+nir_lower_deref_instrs_impl(nir_function_impl *impl,
+ enum nir_lower_deref_flags flags)
+{
+ bool progress = false;
+
+ /* Walk the instructions in reverse order so that we can safely clean up
+ * the deref instructions after we clean up their uses.
+ */
+ nir_foreach_block_reverse(block, impl) {
+ nir_foreach_instr_reverse_safe(instr, block) {
+ switch (instr->type) {
+ case nir_instr_type_deref:
+ if (list_empty(&nir_instr_as_deref(instr)->dest.ssa.uses)) {
+ nir_instr_remove(instr);
+ progress = true;
+ }
+ break;
+
+ case nir_instr_type_tex:
+ if (flags & nir_lower_texture_derefs)
+ progress |= nir_lower_deref_instrs_tex(nir_instr_as_tex(instr));
+ break;
+
+ case nir_instr_type_intrinsic:
+ progress |=
+ nir_lower_deref_instrs_intrin(nir_instr_as_intrinsic(instr),
+ flags);
+ break;
+
+ default:
+ break; /* Nothing to do */
+ }
+ }
+ }
+
+ if (progress) {
+ nir_metadata_preserve(impl, nir_metadata_block_index |
+ nir_metadata_dominance);
+ }
+
+ return progress;
+}
+
+bool
+nir_lower_deref_instrs(nir_shader *shader,
+ enum nir_lower_deref_flags flags)
+{
+ bool progress = false;
+
+ nir_foreach_function(function, shader) {
+ if (!function->impl)
+ continue;
+
+ progress |= nir_lower_deref_instrs_impl(function->impl, flags);
+ }
+
+ return progress;
+}