etnaviv_format.c \
etnaviv_format.h \
etnaviv_internal.h \
+ etnaviv_nir.c \
+ etnaviv_nir.h \
etnaviv_perfmon.c \
etnaviv_perfmon.h \
etnaviv_query.c \
#include "etnaviv_context.h"
#include "etnaviv_debug.h"
#include "etnaviv_disasm.h"
+#include "etnaviv_nir.h"
#include "etnaviv_uniforms.h"
#include "etnaviv_util.h"
bool error;
};
-/* io related lowering
- * run after lower_int_to_float because it adds i2f/f2i ops
- */
-static void
-etna_lower_io(nir_shader *shader, struct etna_shader_variant *v)
-{
- nir_foreach_function(function, shader) {
- nir_builder b;
- nir_builder_init(&b, function->impl);
-
- nir_foreach_block(block, function->impl) {
- nir_foreach_instr_safe(instr, block) {
- if (instr->type == nir_instr_type_intrinsic) {
- nir_intrinsic_instr *intr = nir_instr_as_intrinsic(instr);
-
- switch (intr->intrinsic) {
- case nir_intrinsic_load_front_face: {
- /* HW front_face is 0.0/1.0, not 0/~0u for bool
- * lower with a comparison with 0
- */
- intr->dest.ssa.bit_size = 32;
-
- b.cursor = nir_after_instr(instr);
-
- nir_ssa_def *ssa = nir_ine(&b, &intr->dest.ssa, nir_imm_int(&b, 0));
- if (v->key.front_ccw)
- nir_instr_as_alu(ssa->parent_instr)->op = nir_op_ieq;
-
- nir_ssa_def_rewrite_uses_after(&intr->dest.ssa,
- nir_src_for_ssa(ssa),
- ssa->parent_instr);
- } break;
- case nir_intrinsic_store_deref: {
- nir_deref_instr *deref = nir_src_as_deref(intr->src[0]);
- if (shader->info.stage != MESA_SHADER_FRAGMENT || !v->key.frag_rb_swap)
- break;
-
- assert(deref->deref_type == nir_deref_type_var);
-
- if (deref->var->data.location != FRAG_RESULT_COLOR &&
- deref->var->data.location != FRAG_RESULT_DATA0)
- break;
-
- b.cursor = nir_before_instr(instr);
-
- nir_ssa_def *ssa = nir_mov(&b, intr->src[1].ssa);
- nir_alu_instr *alu = nir_instr_as_alu(ssa->parent_instr);
- alu->src[0].swizzle[0] = 2;
- alu->src[0].swizzle[2] = 0;
- nir_instr_rewrite_src(instr, &intr->src[1], nir_src_for_ssa(ssa));
- } break;
- case nir_intrinsic_load_uniform: {
- /* convert indirect load_uniform to load_ubo when possible
- * this is required on HALTI5+ because address register is not implemented
- * address register loads also arent done optimally
- */
- if (v->shader->specs->halti < 2 || nir_src_is_const(intr->src[0]))
- break;
-
- nir_intrinsic_instr *load_ubo =
- nir_intrinsic_instr_create(b.shader, nir_intrinsic_load_ubo);
- load_ubo->num_components = intr->num_components;
- nir_ssa_dest_init(&load_ubo->instr, &load_ubo->dest,
- load_ubo->num_components, 32, NULL);
-
- b.cursor = nir_before_instr(instr);
- load_ubo->src[0] = nir_src_for_ssa(nir_imm_int(&b, 0));
- load_ubo->src[1] = nir_src_for_ssa(nir_iadd(&b,
- nir_imul(&b, intr->src[0].ssa, nir_imm_int(&b, 16)),
- nir_imm_int(&b, nir_intrinsic_base(intr) * 16)));
- nir_builder_instr_insert(&b, &load_ubo->instr);
- nir_ssa_def_rewrite_uses(&intr->dest.ssa,
- nir_src_for_ssa(&load_ubo->dest.ssa));
- nir_instr_remove(&intr->instr);
- } break;
- case nir_intrinsic_load_ubo: {
- nir_const_value *idx = nir_src_as_const_value(intr->src[0]);
- assert(idx);
- /* offset index by 1, index 0 is used for converted load_uniform */
- b.cursor = nir_before_instr(instr);
- nir_instr_rewrite_src(instr, &intr->src[0],
- nir_src_for_ssa(nir_imm_int(&b, idx[0].u32 + 1)));
- } break;
- case nir_intrinsic_load_vertex_id:
- case nir_intrinsic_load_instance_id:
- /* detect use of vertex_id/instance_id */
- v->vs_id_in_reg = v->infile.num_reg;
- break;
- default:
- break;
- }
- }
-
- if (instr->type != nir_instr_type_tex)
- continue;
-
- nir_tex_instr *tex = nir_instr_as_tex(instr);
- nir_src *coord = NULL;
- nir_src *lod_bias = NULL;
- unsigned lod_bias_idx;
-
- assert(tex->sampler_index == tex->texture_index);
-
- for (unsigned i = 0; i < tex->num_srcs; i++) {
- switch (tex->src[i].src_type) {
- case nir_tex_src_coord:
- coord = &tex->src[i].src;
- break;
- case nir_tex_src_bias:
- case nir_tex_src_lod:
- assert(!lod_bias);
- lod_bias = &tex->src[i].src;
- lod_bias_idx = i;
- break;
- case nir_tex_src_comparator:
- break;
- default:
- assert(0);
- break;
- }
- }
-
- if (tex->sampler_dim == GLSL_SAMPLER_DIM_RECT) {
- /* use a dummy load_uniform here to represent texcoord scale */
- b.cursor = nir_before_instr(instr);
- nir_intrinsic_instr *load =
- nir_intrinsic_instr_create(b.shader, nir_intrinsic_load_uniform);
- nir_intrinsic_set_base(load, ~tex->sampler_index);
- load->num_components = 2;
- load->src[0] = nir_src_for_ssa(nir_imm_float(&b, 0.0f));
- nir_ssa_dest_init(&load->instr, &load->dest, 2, 32, NULL);
- nir_intrinsic_set_type(load, nir_type_float);
-
- nir_builder_instr_insert(&b, &load->instr);
-
- nir_ssa_def *new_coord = nir_fmul(&b, coord->ssa, &load->dest.ssa);
- nir_instr_rewrite_src(&tex->instr, coord, nir_src_for_ssa(new_coord));
- }
-
- /* pre HALTI5 needs texture sources in a single source */
-
- if (!lod_bias || v->shader->specs->halti >= 5)
- continue;
-
- assert(coord && lod_bias && tex->coord_components < 4);
-
- nir_alu_instr *vec = nir_alu_instr_create(shader, nir_op_vec4);
- for (unsigned i = 0; i < tex->coord_components; i++) {
- vec->src[i].src = nir_src_for_ssa(coord->ssa);
- vec->src[i].swizzle[0] = i;
- }
- for (unsigned i = tex->coord_components; i < 4; i++)
- vec->src[i].src = nir_src_for_ssa(lod_bias->ssa);
-
- vec->dest.write_mask = 0xf;
- nir_ssa_dest_init(&vec->instr, &vec->dest.dest, 4, 32, NULL);
-
- nir_tex_instr_remove_src(tex, lod_bias_idx);
- nir_instr_rewrite_src(&tex->instr, coord, nir_src_for_ssa(&vec->dest.dest.ssa));
- tex->coord_components = 4;
-
- nir_instr_insert_before(&tex->instr, &vec->instr);
- }
- }
- }
-}
+
static bool
etna_alu_to_scalar_filter_cb(const nir_instr *instr, const void *data)
--- /dev/null
+/*
+ * Copyright (c) 2019 Zodiac Inflight Innovations
+ *
+ * Permission is hereby granted, free of charge, to any person obtaining a
+ * copy of this software and associated documentation files (the "Software"),
+ * to deal in the Software without restriction, including without limitation
+ * the rights to use, copy, modify, merge, publish, distribute, sub license,
+ * and/or sell copies of the Software, and to permit persons to whom the
+ * Software is furnished to do so, subject to the following conditions:
+ *
+ * The above copyright notice and this permission notice (including the
+ * next paragraph) shall be included in all copies or substantial portions
+ * of the Software.
+ *
+ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
+ * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
+ * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL
+ * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
+ * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
+ * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
+ * DEALINGS IN THE SOFTWARE.
+ *
+ * Authors:
+ * Jonathan Marek <jonathan@marek.ca>
+ */
+
+#include "etnaviv_nir.h"
+
+/* io related lowering
+ * run after lower_int_to_float because it adds i2f/f2i ops
+ */
+void
+etna_lower_io(nir_shader *shader, struct etna_shader_variant *v)
+{
+ nir_foreach_function(function, shader) {
+ nir_builder b;
+ nir_builder_init(&b, function->impl);
+
+ nir_foreach_block(block, function->impl) {
+ nir_foreach_instr_safe(instr, block) {
+ if (instr->type == nir_instr_type_intrinsic) {
+ nir_intrinsic_instr *intr = nir_instr_as_intrinsic(instr);
+
+ switch (intr->intrinsic) {
+ case nir_intrinsic_load_front_face: {
+ /* HW front_face is 0.0/1.0, not 0/~0u for bool
+ * lower with a comparison with 0
+ */
+ intr->dest.ssa.bit_size = 32;
+
+ b.cursor = nir_after_instr(instr);
+
+ nir_ssa_def *ssa = nir_ine(&b, &intr->dest.ssa, nir_imm_int(&b, 0));
+ if (v->key.front_ccw)
+ nir_instr_as_alu(ssa->parent_instr)->op = nir_op_ieq;
+
+ nir_ssa_def_rewrite_uses_after(&intr->dest.ssa,
+ nir_src_for_ssa(ssa),
+ ssa->parent_instr);
+ } break;
+ case nir_intrinsic_store_deref: {
+ nir_deref_instr *deref = nir_src_as_deref(intr->src[0]);
+ if (shader->info.stage != MESA_SHADER_FRAGMENT || !v->key.frag_rb_swap)
+ break;
+
+ assert(deref->deref_type == nir_deref_type_var);
+
+ if (deref->var->data.location != FRAG_RESULT_COLOR &&
+ deref->var->data.location != FRAG_RESULT_DATA0)
+ break;
+
+ b.cursor = nir_before_instr(instr);
+
+ nir_ssa_def *ssa = nir_mov(&b, intr->src[1].ssa);
+ nir_alu_instr *alu = nir_instr_as_alu(ssa->parent_instr);
+ alu->src[0].swizzle[0] = 2;
+ alu->src[0].swizzle[2] = 0;
+ nir_instr_rewrite_src(instr, &intr->src[1], nir_src_for_ssa(ssa));
+ } break;
+ case nir_intrinsic_load_uniform: {
+ /* convert indirect load_uniform to load_ubo when possible
+ * this is required on HALTI5+ because address register is not implemented
+ * address register loads also arent done optimally
+ */
+ if (v->shader->specs->halti < 2 || nir_src_is_const(intr->src[0]))
+ break;
+
+ nir_intrinsic_instr *load_ubo =
+ nir_intrinsic_instr_create(b.shader, nir_intrinsic_load_ubo);
+ load_ubo->num_components = intr->num_components;
+ nir_ssa_dest_init(&load_ubo->instr, &load_ubo->dest,
+ load_ubo->num_components, 32, NULL);
+
+ b.cursor = nir_before_instr(instr);
+ load_ubo->src[0] = nir_src_for_ssa(nir_imm_int(&b, 0));
+ load_ubo->src[1] = nir_src_for_ssa(nir_iadd(&b,
+ nir_imul(&b, intr->src[0].ssa, nir_imm_int(&b, 16)),
+ nir_imm_int(&b, nir_intrinsic_base(intr) * 16)));
+ nir_builder_instr_insert(&b, &load_ubo->instr);
+ nir_ssa_def_rewrite_uses(&intr->dest.ssa,
+ nir_src_for_ssa(&load_ubo->dest.ssa));
+ nir_instr_remove(&intr->instr);
+ } break;
+ case nir_intrinsic_load_ubo: {
+ nir_const_value *idx = nir_src_as_const_value(intr->src[0]);
+ assert(idx);
+ /* offset index by 1, index 0 is used for converted load_uniform */
+ b.cursor = nir_before_instr(instr);
+ nir_instr_rewrite_src(instr, &intr->src[0],
+ nir_src_for_ssa(nir_imm_int(&b, idx[0].u32 + 1)));
+ } break;
+ case nir_intrinsic_load_vertex_id:
+ case nir_intrinsic_load_instance_id:
+ /* detect use of vertex_id/instance_id */
+ v->vs_id_in_reg = v->infile.num_reg;
+ break;
+ default:
+ break;
+ }
+ }
+
+ if (instr->type != nir_instr_type_tex)
+ continue;
+
+ nir_tex_instr *tex = nir_instr_as_tex(instr);
+ nir_src *coord = NULL;
+ nir_src *lod_bias = NULL;
+ unsigned lod_bias_idx;
+
+ assert(tex->sampler_index == tex->texture_index);
+
+ for (unsigned i = 0; i < tex->num_srcs; i++) {
+ switch (tex->src[i].src_type) {
+ case nir_tex_src_coord:
+ coord = &tex->src[i].src;
+ break;
+ case nir_tex_src_bias:
+ case nir_tex_src_lod:
+ assert(!lod_bias);
+ lod_bias = &tex->src[i].src;
+ lod_bias_idx = i;
+ break;
+ case nir_tex_src_comparator:
+ break;
+ default:
+ assert(0);
+ break;
+ }
+ }
+
+ if (tex->sampler_dim == GLSL_SAMPLER_DIM_RECT) {
+ /* use a dummy load_uniform here to represent texcoord scale */
+ b.cursor = nir_before_instr(instr);
+ nir_intrinsic_instr *load =
+ nir_intrinsic_instr_create(b.shader, nir_intrinsic_load_uniform);
+ nir_intrinsic_set_base(load, ~tex->sampler_index);
+ load->num_components = 2;
+ load->src[0] = nir_src_for_ssa(nir_imm_float(&b, 0.0f));
+ nir_ssa_dest_init(&load->instr, &load->dest, 2, 32, NULL);
+ nir_intrinsic_set_type(load, nir_type_float);
+
+ nir_builder_instr_insert(&b, &load->instr);
+
+ nir_ssa_def *new_coord = nir_fmul(&b, coord->ssa, &load->dest.ssa);
+ nir_instr_rewrite_src(&tex->instr, coord, nir_src_for_ssa(new_coord));
+ }
+
+ /* pre HALTI5 needs texture sources in a single source */
+
+ if (!lod_bias || v->shader->specs->halti >= 5)
+ continue;
+
+ assert(coord && lod_bias && tex->coord_components < 4);
+
+ nir_alu_instr *vec = nir_alu_instr_create(shader, nir_op_vec4);
+ for (unsigned i = 0; i < tex->coord_components; i++) {
+ vec->src[i].src = nir_src_for_ssa(coord->ssa);
+ vec->src[i].swizzle[0] = i;
+ }
+ for (unsigned i = tex->coord_components; i < 4; i++)
+ vec->src[i].src = nir_src_for_ssa(lod_bias->ssa);
+
+ vec->dest.write_mask = 0xf;
+ nir_ssa_dest_init(&vec->instr, &vec->dest.dest, 4, 32, NULL);
+
+ nir_tex_instr_remove_src(tex, lod_bias_idx);
+ nir_instr_rewrite_src(&tex->instr, coord, nir_src_for_ssa(&vec->dest.dest.ssa));
+ tex->coord_components = 4;
+
+ nir_instr_insert_before(&tex->instr, &vec->instr);
+ }
+ }
+ }
+}
--- /dev/null
+/*
+ * Copyright (c) 2012-2015 Etnaviv Project
+ *
+ * Permission is hereby granted, free of charge, to any person obtaining a
+ * copy of this software and associated documentation files (the "Software"),
+ * to deal in the Software without restriction, including without limitation
+ * the rights to use, copy, modify, merge, publish, distribute, sub license,
+ * and/or sell copies of the Software, and to permit persons to whom the
+ * Software is furnished to do so, subject to the following conditions:
+ *
+ * The above copyright notice and this permission notice (including the
+ * next paragraph) shall be included in all copies or substantial portions
+ * of the Software.
+ *
+ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
+ * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
+ * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL
+ * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
+ * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
+ * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
+ * DEALINGS IN THE SOFTWARE.
+ */
+
+#ifndef H_ETNA_NIR
+#define H_ETNA_NIR
+
+#include "compiler/nir/nir.h"
+#include "compiler/nir/nir_builder.h"
+#include "etnaviv_compiler.h"
+
+void
+etna_lower_io(nir_shader *shader, struct etna_shader_variant *v);
+
+#endif
'etnaviv_format.c',
'etnaviv_format.h',
'etnaviv_internal.h',
+ 'etnaviv_nir.c',
+ 'etnaviv_nir.h',
'etnaviv_perfmon.c',
'etnaviv_perfmon.h',
'etnaviv_query.c',