X-Git-Url: https://git.libre-soc.org/?a=blobdiff_plain;f=src%2Fcompiler%2Fnir%2Fnir_validate.c;h=9d2dfbc820675d7f7385f69824e536674e294e84;hb=e4f07f8bdc602bae665fd57bb69e293a69d89bef;hp=0c32d5fe07a5e83a96deaa66f67f813f821770fa;hpb=3c807607df4ab457037f2daa5ea9e5ce18392b7b;p=mesa.git diff --git a/src/compiler/nir/nir_validate.c b/src/compiler/nir/nir_validate.c index 0c32d5fe07a..9d2dfbc8206 100644 --- a/src/compiler/nir/nir_validate.c +++ b/src/compiler/nir/nir_validate.c @@ -26,6 +26,7 @@ */ #include "nir.h" +#include "c11/threads.h" #include /* @@ -35,7 +36,7 @@ /* Since this file is just a pile of asserts, don't bother compiling it if * we're not building a debug build. */ -#ifdef DEBUG +#ifndef NDEBUG /* * Per-register validation state. @@ -51,15 +52,8 @@ typedef struct { } reg_validate_state; typedef struct { - /* - * equivalent to the uses in nir_ssa_def, but built up by the validator. - * At the end, we verify that the sets have the same entries. - */ - struct set *uses, *if_uses; - nir_function_impl *where_defined; -} ssa_def_validate_state; + void *mem_ctx; -typedef struct { /* map of register -> validation state (struct above) */ struct hash_table *regs; @@ -69,6 +63,9 @@ typedef struct { /* the current instruction being validated */ nir_instr *instr; + /* the current variable being validated */ + nir_variable *var; + /* the current basic block being validated */ nir_block *block; @@ -84,8 +81,8 @@ typedef struct { /* the current function implementation being validated */ nir_function_impl *impl; - /* map of SSA value -> function implementation where it is defined */ - struct hash_table *ssa_defs; + /* Set of seen SSA sources */ + struct set *ssa_srcs; /* bitset of ssa definitions we have found; used to check uniqueness */ BITSET_WORD *ssa_defs_found; @@ -93,84 +90,132 @@ typedef struct { /* bitset of registers we have currently found; used to check uniqueness */ BITSET_WORD *regs_found; - /* map of local variable -> function implementation where it is defined */ + /* map of variable -> function implementation where it is defined or NULL + * if it is a global variable + */ struct hash_table *var_defs; + + /* map of instruction/var/etc to failed assert string */ + struct hash_table *errors; } validate_state; -static void validate_src(nir_src *src, validate_state *state); +static void +log_error(validate_state *state, const char *cond, const char *file, int line) +{ + const void *obj; + + if (state->instr) + obj = state->instr; + else if (state->var) + obj = state->var; + else + obj = cond; + + char *msg = ralloc_asprintf(state->errors, "error: %s (%s:%d)", + cond, file, line); + + _mesa_hash_table_insert(state->errors, obj, msg); +} + +#define validate_assert(state, cond) do { \ + if (!(cond)) \ + log_error(state, #cond, __FILE__, __LINE__); \ + } while (0) + +static void validate_src(nir_src *src, validate_state *state, + unsigned bit_sizes, unsigned num_components); static void -validate_reg_src(nir_src *src, validate_state *state) +validate_num_components(validate_state *state, unsigned num_components) { - assert(src->reg.reg != NULL); + validate_assert(state, nir_num_components_valid(num_components)); +} + +static void +validate_reg_src(nir_src *src, validate_state *state, + unsigned bit_sizes, unsigned num_components) +{ + validate_assert(state, src->reg.reg != NULL); struct hash_entry *entry; entry = _mesa_hash_table_search(state->regs, src->reg.reg); - assert(entry); + validate_assert(state, entry); reg_validate_state *reg_state = (reg_validate_state *) entry->data; if (state->instr) { _mesa_set_add(reg_state->uses, src); } else { - assert(state->if_stmt); + validate_assert(state, state->if_stmt); _mesa_set_add(reg_state->if_uses, src); } - if (!src->reg.reg->is_global) { - assert(reg_state->where_defined == state->impl && - "using a register declared in a different function"); - } + validate_assert(state, reg_state->where_defined == state->impl && + "using a register declared in a different function"); + + if (bit_sizes) + validate_assert(state, src->reg.reg->bit_size & bit_sizes); + if (num_components) + validate_assert(state, src->reg.reg->num_components == num_components); - assert((src->reg.reg->num_array_elems == 0 || + validate_assert(state, (src->reg.reg->num_array_elems == 0 || src->reg.base_offset < src->reg.reg->num_array_elems) && "definitely out-of-bounds array access"); if (src->reg.indirect) { - assert(src->reg.reg->num_array_elems != 0); - assert((src->reg.indirect->is_ssa || + validate_assert(state, src->reg.reg->num_array_elems != 0); + validate_assert(state, (src->reg.indirect->is_ssa || src->reg.indirect->reg.indirect == NULL) && "only one level of indirection allowed"); - validate_src(src->reg.indirect, state); + validate_src(src->reg.indirect, state, 32, 1); } } +#define SET_PTR_BIT(ptr, bit) \ + (void *)(((uintptr_t)(ptr)) | (((uintptr_t)1) << bit)) + static void -validate_ssa_src(nir_src *src, validate_state *state) +validate_ssa_src(nir_src *src, validate_state *state, + unsigned bit_sizes, unsigned num_components) { - assert(src->ssa != NULL); - - struct hash_entry *entry = _mesa_hash_table_search(state->ssa_defs, src->ssa); - - assert(entry); - - ssa_def_validate_state *def_state = (ssa_def_validate_state *)entry->data; - - assert(def_state->where_defined == state->impl && - "using an SSA value defined in a different function"); + validate_assert(state, src->ssa != NULL); + /* As we walk SSA defs, we add every use to this set. We need to make sure + * our use is seen in a use list. + */ + struct set_entry *entry; if (state->instr) { - _mesa_set_add(def_state->uses, src); + entry = _mesa_set_search(state->ssa_srcs, src); } else { - assert(state->if_stmt); - _mesa_set_add(def_state->if_uses, src); + entry = _mesa_set_search(state->ssa_srcs, SET_PTR_BIT(src, 0)); } + validate_assert(state, entry); + + /* This will let us prove that we've seen all the sources */ + if (entry) + _mesa_set_remove(state->ssa_srcs, entry); + + if (bit_sizes) + validate_assert(state, src->ssa->bit_size & bit_sizes); + if (num_components) + validate_assert(state, src->ssa->num_components == num_components); /* TODO validate that the use is dominated by the definition */ } static void -validate_src(nir_src *src, validate_state *state) +validate_src(nir_src *src, validate_state *state, + unsigned bit_sizes, unsigned num_components) { if (state->instr) - assert(src->parent_instr == state->instr); + validate_assert(state, src->parent_instr == state->instr); else - assert(src->parent_if == state->if_stmt); + validate_assert(state, src->parent_if == state->if_stmt); if (src->is_ssa) - validate_ssa_src(src, state); + validate_ssa_src(src, state, bit_sizes, num_components); else - validate_reg_src(src, state); + validate_reg_src(src, state, bit_sizes, num_components); } static void @@ -178,258 +223,450 @@ validate_alu_src(nir_alu_instr *instr, unsigned index, validate_state *state) { nir_alu_src *src = &instr->src[index]; - unsigned num_components; - if (src->src.is_ssa) - num_components = src->src.ssa->num_components; - else { - if (src->src.reg.reg->is_packed) - num_components = 4; /* can't check anything */ - else - num_components = src->src.reg.reg->num_components; - } - for (unsigned i = 0; i < 4; i++) { - assert(src->swizzle[i] < 4); + if (instr->op == nir_op_mov) + assert(!src->abs && !src->negate); + + unsigned num_components = nir_src_num_components(src->src); + for (unsigned i = 0; i < NIR_MAX_VEC_COMPONENTS; i++) { + validate_assert(state, src->swizzle[i] < NIR_MAX_VEC_COMPONENTS); if (nir_alu_instr_channel_used(instr, index, i)) - assert(src->swizzle[i] < num_components); + validate_assert(state, src->swizzle[i] < num_components); } - validate_src(&src->src, state); + validate_src(&src->src, state, 0, 0); } static void -validate_reg_dest(nir_reg_dest *dest, validate_state *state) +validate_reg_dest(nir_reg_dest *dest, validate_state *state, + unsigned bit_sizes, unsigned num_components) { - assert(dest->reg != NULL); + validate_assert(state, dest->reg != NULL); - assert(dest->parent_instr == state->instr); + validate_assert(state, dest->parent_instr == state->instr); struct hash_entry *entry2; entry2 = _mesa_hash_table_search(state->regs, dest->reg); - assert(entry2); + validate_assert(state, entry2); reg_validate_state *reg_state = (reg_validate_state *) entry2->data; _mesa_set_add(reg_state->defs, dest); - if (!dest->reg->is_global) { - assert(reg_state->where_defined == state->impl && - "writing to a register declared in a different function"); - } + validate_assert(state, reg_state->where_defined == state->impl && + "writing to a register declared in a different function"); + + if (bit_sizes) + validate_assert(state, dest->reg->bit_size & bit_sizes); + if (num_components) + validate_assert(state, dest->reg->num_components == num_components); - assert((dest->reg->num_array_elems == 0 || + validate_assert(state, (dest->reg->num_array_elems == 0 || dest->base_offset < dest->reg->num_array_elems) && "definitely out-of-bounds array access"); if (dest->indirect) { - assert(dest->reg->num_array_elems != 0); - assert((dest->indirect->is_ssa || dest->indirect->reg.indirect == NULL) && + validate_assert(state, dest->reg->num_array_elems != 0); + validate_assert(state, (dest->indirect->is_ssa || dest->indirect->reg.indirect == NULL) && "only one level of indirection allowed"); - validate_src(dest->indirect, state); + validate_src(dest->indirect, state, 32, 1); } } static void validate_ssa_def(nir_ssa_def *def, validate_state *state) { - assert(def->index < state->impl->ssa_alloc); - assert(!BITSET_TEST(state->ssa_defs_found, def->index)); + validate_assert(state, def->index < state->impl->ssa_alloc); + validate_assert(state, !BITSET_TEST(state->ssa_defs_found, def->index)); BITSET_SET(state->ssa_defs_found, def->index); - assert(def->parent_instr == state->instr); - - assert(def->num_components <= 4); + validate_assert(state, def->parent_instr == state->instr); + validate_num_components(state, def->num_components); list_validate(&def->uses); - list_validate(&def->if_uses); + nir_foreach_use(src, def) { + validate_assert(state, src->is_ssa); + validate_assert(state, src->ssa == def); + bool already_seen = false; + _mesa_set_search_and_add(state->ssa_srcs, src, &already_seen); + /* A nir_src should only appear once and only in one SSA def use list */ + validate_assert(state, !already_seen); + } - ssa_def_validate_state *def_state = ralloc(state->ssa_defs, - ssa_def_validate_state); - def_state->where_defined = state->impl; - def_state->uses = _mesa_set_create(def_state, _mesa_hash_pointer, - _mesa_key_pointer_equal); - def_state->if_uses = _mesa_set_create(def_state, _mesa_hash_pointer, - _mesa_key_pointer_equal); - _mesa_hash_table_insert(state->ssa_defs, def, def_state); + list_validate(&def->if_uses); + nir_foreach_if_use(src, def) { + validate_assert(state, src->is_ssa); + validate_assert(state, src->ssa == def); + bool already_seen = false; + _mesa_set_search_and_add(state->ssa_srcs, SET_PTR_BIT(src, 0), + &already_seen); + /* A nir_src should only appear once and only in one SSA def use list */ + validate_assert(state, !already_seen); + } } static void -validate_dest(nir_dest *dest, validate_state *state) +validate_dest(nir_dest *dest, validate_state *state, + unsigned bit_sizes, unsigned num_components) { - if (dest->is_ssa) + if (dest->is_ssa) { + if (bit_sizes) + validate_assert(state, dest->ssa.bit_size & bit_sizes); + if (num_components) + validate_assert(state, dest->ssa.num_components == num_components); validate_ssa_def(&dest->ssa, state); - else - validate_reg_dest(&dest->reg, state); + } else { + validate_reg_dest(&dest->reg, state, bit_sizes, num_components); + } } static void -validate_alu_dest(nir_alu_dest *dest, validate_state *state) +validate_alu_dest(nir_alu_instr *instr, validate_state *state) { - unsigned dest_size = - dest->dest.is_ssa ? dest->dest.ssa.num_components - : dest->dest.reg.reg->num_components; - bool is_packed = !dest->dest.is_ssa && dest->dest.reg.reg->is_packed; + nir_alu_dest *dest = &instr->dest; + + if (instr->op == nir_op_mov) + assert(!dest->saturate); + + unsigned dest_size = nir_dest_num_components(dest->dest); /* * validate that the instruction doesn't write to components not in the * register/SSA value */ - assert(is_packed || !(dest->write_mask & ~((1 << dest_size) - 1))); + validate_assert(state, !(dest->write_mask & ~((1 << dest_size) - 1))); /* validate that saturate is only ever used on instructions with * destinations of type float */ nir_alu_instr *alu = nir_instr_as_alu(state->instr); - assert(nir_op_infos[alu->op].output_type == nir_type_float || + validate_assert(state, + (nir_alu_type_get_base_type(nir_op_infos[alu->op].output_type) == + nir_type_float) || !dest->saturate); - validate_dest(&dest->dest, state); + validate_dest(&dest->dest, state, 0, 0); } static void validate_alu_instr(nir_alu_instr *instr, validate_state *state) { - assert(instr->op < nir_num_opcodes); + validate_assert(state, instr->op < nir_num_opcodes); + unsigned instr_bit_size = 0; for (unsigned i = 0; i < nir_op_infos[instr->op].num_inputs; i++) { + nir_alu_type src_type = nir_op_infos[instr->op].input_types[i]; + unsigned src_bit_size = nir_src_bit_size(instr->src[i].src); + if (nir_alu_type_get_type_size(src_type)) { + validate_assert(state, src_bit_size == nir_alu_type_get_type_size(src_type)); + } else if (instr_bit_size) { + validate_assert(state, src_bit_size == instr_bit_size); + } else { + instr_bit_size = src_bit_size; + } + + if (nir_alu_type_get_base_type(src_type) == nir_type_float) { + /* 8-bit float isn't a thing */ + validate_assert(state, src_bit_size == 16 || src_bit_size == 32 || + src_bit_size == 64); + } + validate_alu_src(instr, i, state); } - validate_alu_dest(&instr->dest, state); + nir_alu_type dest_type = nir_op_infos[instr->op].output_type; + unsigned dest_bit_size = nir_dest_bit_size(instr->dest.dest); + if (nir_alu_type_get_type_size(dest_type)) { + validate_assert(state, dest_bit_size == nir_alu_type_get_type_size(dest_type)); + } else if (instr_bit_size) { + validate_assert(state, dest_bit_size == instr_bit_size); + } else { + /* The only unsized thing is the destination so it's vacuously valid */ + } + + if (nir_alu_type_get_base_type(dest_type) == nir_type_float) { + /* 8-bit float isn't a thing */ + validate_assert(state, dest_bit_size == 16 || dest_bit_size == 32 || + dest_bit_size == 64); + } + + validate_alu_dest(instr, state); } static void -validate_deref_chain(nir_deref *deref, validate_state *state) +validate_var_use(nir_variable *var, validate_state *state) { - assert(deref->child == NULL || ralloc_parent(deref->child) == deref); + struct hash_entry *entry = _mesa_hash_table_search(state->var_defs, var); + validate_assert(state, entry); + if (entry && var->data.mode == nir_var_function_temp) + validate_assert(state, (nir_function_impl *) entry->data == state->impl); +} - nir_deref *parent = NULL; - while (deref != NULL) { - switch (deref->deref_type) { - case nir_deref_type_array: - assert(deref->type == glsl_get_array_element(parent->type)); - if (nir_deref_as_array(deref)->deref_array_type == - nir_deref_array_type_indirect) - validate_src(&nir_deref_as_array(deref)->indirect, state); - break; +static void +validate_deref_instr(nir_deref_instr *instr, validate_state *state) +{ + if (instr->deref_type == nir_deref_type_var) { + /* Variable dereferences are stupid simple. */ + validate_assert(state, instr->mode == instr->var->data.mode); + validate_assert(state, instr->type == instr->var->type); + validate_var_use(instr->var, state); + } else if (instr->deref_type == nir_deref_type_cast) { + /* For cast, we simply have to trust the instruction. It's up to + * lowering passes and front/back-ends to make them sane. + */ + validate_src(&instr->parent, state, 0, 0); + + /* We just validate that the type and mode are there */ + validate_assert(state, instr->mode); + validate_assert(state, instr->type); + } else { + /* We require the parent to be SSA. This may be lifted in the future */ + validate_assert(state, instr->parent.is_ssa); + + /* The parent pointer value must have the same number of components + * as the destination. + */ + validate_src(&instr->parent, state, nir_dest_bit_size(instr->dest), + nir_dest_num_components(instr->dest)); + nir_instr *parent_instr = instr->parent.ssa->parent_instr; + + /* The parent must come from another deref instruction */ + validate_assert(state, parent_instr->type == nir_instr_type_deref); + + nir_deref_instr *parent = nir_instr_as_deref(parent_instr); + + validate_assert(state, instr->mode == parent->mode); + + switch (instr->deref_type) { case nir_deref_type_struct: - assert(deref->type == - glsl_get_struct_field(parent->type, - nir_deref_as_struct(deref)->index)); + validate_assert(state, glsl_type_is_struct_or_ifc(parent->type)); + validate_assert(state, + instr->strct.index < glsl_get_length(parent->type)); + validate_assert(state, instr->type == + glsl_get_struct_field(parent->type, instr->strct.index)); + break; + + case nir_deref_type_array: + case nir_deref_type_array_wildcard: + if (instr->mode == nir_var_mem_ubo || + instr->mode == nir_var_mem_ssbo || + instr->mode == nir_var_mem_shared || + instr->mode == nir_var_mem_global) { + /* Shared variables and UBO/SSBOs have a bit more relaxed rules + * because we need to be able to handle array derefs on vectors. + * Fortunately, nir_lower_io handles these just fine. + */ + validate_assert(state, glsl_type_is_array(parent->type) || + glsl_type_is_matrix(parent->type) || + glsl_type_is_vector(parent->type)); + } else { + /* Most of NIR cannot handle array derefs on vectors */ + validate_assert(state, glsl_type_is_array(parent->type) || + glsl_type_is_matrix(parent->type)); + } + validate_assert(state, + instr->type == glsl_get_array_element(parent->type)); + + if (instr->deref_type == nir_deref_type_array) { + validate_src(&instr->arr.index, state, + nir_dest_bit_size(instr->dest), 1); + } break; - case nir_deref_type_var: + case nir_deref_type_ptr_as_array: + /* ptr_as_array derefs must have a parent that is either an array, + * ptr_as_array, or cast. If the parent is a cast, we get the stride + * information (if any) from the cast deref. + */ + validate_assert(state, + parent->deref_type == nir_deref_type_array || + parent->deref_type == nir_deref_type_ptr_as_array || + parent->deref_type == nir_deref_type_cast); + validate_src(&instr->arr.index, state, + nir_dest_bit_size(instr->dest), 1); break; default: - assert(!"Invalid deref type"); - break; + unreachable("Invalid deref instruction type"); } - - parent = deref; - deref = deref->child; } -} -static void -validate_var_use(nir_variable *var, validate_state *state) -{ - if (var->data.mode == nir_var_local) { - struct hash_entry *entry = _mesa_hash_table_search(state->var_defs, var); + /* We intentionally don't validate the size of the destination because we + * want to let other compiler components such as SPIR-V decide how big + * pointers should be. + */ + validate_dest(&instr->dest, state, 0, 0); - assert(entry); - assert((nir_function_impl *) entry->data == state->impl); + /* Deref instructions as if conditions don't make sense because if + * conditions expect well-formed Booleans. If you want to compare with + * NULL, an explicit comparison operation should be used. + */ + validate_assert(state, list_is_empty(&instr->dest.ssa.if_uses)); + + /* Only certain modes can be used as sources for phi instructions. */ + nir_foreach_use(use, &instr->dest.ssa) { + if (use->parent_instr->type == nir_instr_type_phi) { + validate_assert(state, instr->mode == nir_var_mem_ubo || + instr->mode == nir_var_mem_ssbo || + instr->mode == nir_var_mem_shared || + instr->mode == nir_var_mem_global || + instr->mode == nir_var_mem_constant); + } } } -static void -validate_deref_var(void *parent_mem_ctx, nir_deref_var *deref, validate_state *state) +static bool +vectorized_intrinsic(nir_intrinsic_instr *intr) { - assert(deref != NULL); - assert(ralloc_parent(deref) == parent_mem_ctx); - assert(deref->deref.type == deref->var->type); + const nir_intrinsic_info *info = &nir_intrinsic_infos[intr->intrinsic]; + + if (info->dest_components == 0) + return true; - validate_var_use(deref->var, state); + for (unsigned i = 0; i < info->num_srcs; i++) + if (info->src_components[i] == 0) + return true; - validate_deref_chain(&deref->deref, state); + return false; } static void validate_intrinsic_instr(nir_intrinsic_instr *instr, validate_state *state) { - unsigned num_srcs = nir_intrinsic_infos[instr->intrinsic].num_srcs; - for (unsigned i = 0; i < num_srcs; i++) { - unsigned components_read = - nir_intrinsic_infos[instr->intrinsic].src_components[i]; - if (components_read == 0) - components_read = instr->num_components; - - assert(components_read > 0); - - if (instr->src[i].is_ssa) { - assert(components_read <= instr->src[i].ssa->num_components); - } else if (!instr->src[i].reg.reg->is_packed) { - assert(components_read <= instr->src[i].reg.reg->num_components); - } - - validate_src(&instr->src[i], state); + unsigned dest_bit_size = 0; + unsigned src_bit_sizes[NIR_INTRINSIC_MAX_INPUTS] = { 0, }; + switch (instr->intrinsic) { + case nir_intrinsic_load_param: { + unsigned param_idx = nir_intrinsic_param_idx(instr); + validate_assert(state, param_idx < state->impl->function->num_params); + nir_parameter *param = &state->impl->function->params[param_idx]; + validate_assert(state, instr->num_components == param->num_components); + dest_bit_size = param->bit_size; + break; } - unsigned num_vars = nir_intrinsic_infos[instr->intrinsic].num_variables; - for (unsigned i = 0; i < num_vars; i++) { - validate_deref_var(instr, instr->variables[i], state); + case nir_intrinsic_load_deref: { + nir_deref_instr *src = nir_src_as_deref(instr->src[0]); + assert(src); + validate_assert(state, glsl_type_is_vector_or_scalar(src->type) || + (src->mode == nir_var_uniform && + glsl_get_base_type(src->type) == GLSL_TYPE_SUBROUTINE)); + validate_assert(state, instr->num_components == + glsl_get_vector_elements(src->type)); + dest_bit_size = glsl_get_bit_size(src->type); + /* Also allow 32-bit boolean load operations */ + if (glsl_type_is_boolean(src->type)) + dest_bit_size |= 32; + break; } - if (nir_intrinsic_infos[instr->intrinsic].has_dest) { - unsigned components_written = - nir_intrinsic_infos[instr->intrinsic].dest_components; - if (components_written == 0) - components_written = instr->num_components; - - assert(components_written > 0); - - if (instr->dest.is_ssa) { - assert(components_written <= instr->dest.ssa.num_components); - } else if (!instr->dest.reg.reg->is_packed) { - assert(components_written <= instr->dest.reg.reg->num_components); - } - - validate_dest(&instr->dest, state); + case nir_intrinsic_store_deref: { + nir_deref_instr *dst = nir_src_as_deref(instr->src[0]); + assert(dst); + validate_assert(state, glsl_type_is_vector_or_scalar(dst->type)); + validate_assert(state, instr->num_components == + glsl_get_vector_elements(dst->type)); + src_bit_sizes[1] = glsl_get_bit_size(dst->type); + /* Also allow 32-bit boolean store operations */ + if (glsl_type_is_boolean(dst->type)) + src_bit_sizes[1] |= 32; + validate_assert(state, (dst->mode & (nir_var_shader_in | + nir_var_uniform)) == 0); + validate_assert(state, (nir_intrinsic_write_mask(instr) & ~((1 << instr->num_components) - 1)) == 0); + break; } - switch (instr->intrinsic) { - case nir_intrinsic_load_var: { - const struct glsl_type *type = - nir_deref_tail(&instr->variables[0]->deref)->type; - assert(glsl_type_is_vector_or_scalar(type) || - (instr->variables[0]->var->data.mode == nir_var_uniform && - glsl_get_base_type(type) == GLSL_TYPE_SUBROUTINE)); - assert(instr->num_components == glsl_get_vector_elements(type)); + case nir_intrinsic_copy_deref: { + nir_deref_instr *dst = nir_src_as_deref(instr->src[0]); + nir_deref_instr *src = nir_src_as_deref(instr->src[1]); + validate_assert(state, glsl_get_bare_type(dst->type) == + glsl_get_bare_type(src->type)); + validate_assert(state, (dst->mode & (nir_var_shader_in | + nir_var_uniform)) == 0); break; } - case nir_intrinsic_store_var: { - const struct glsl_type *type = - nir_deref_tail(&instr->variables[0]->deref)->type; - assert(glsl_type_is_vector_or_scalar(type) || - (instr->variables[0]->var->data.mode == nir_var_uniform && - glsl_get_base_type(type) == GLSL_TYPE_SUBROUTINE)); - assert(instr->num_components == glsl_get_vector_elements(type)); - assert(instr->variables[0]->var->data.mode != nir_var_shader_in && - instr->variables[0]->var->data.mode != nir_var_uniform && - instr->variables[0]->var->data.mode != nir_var_shader_storage); - assert((nir_intrinsic_write_mask(instr) & ~((1 << instr->num_components) - 1)) == 0); + + case nir_intrinsic_load_ubo_vec4: { + int bit_size = nir_dest_bit_size(instr->dest); + validate_assert(state, bit_size >= 8); + validate_assert(state, (nir_intrinsic_component(instr) + + instr->num_components) * (bit_size / 8) <= 16); break; } - case nir_intrinsic_copy_var: - assert(nir_deref_tail(&instr->variables[0]->deref)->type == - nir_deref_tail(&instr->variables[1]->deref)->type); - assert(instr->variables[0]->var->data.mode != nir_var_shader_in && - instr->variables[0]->var->data.mode != nir_var_uniform && - instr->variables[0]->var->data.mode != nir_var_shader_storage); + + case nir_intrinsic_load_ubo: + case nir_intrinsic_load_ssbo: + case nir_intrinsic_load_shared: + case nir_intrinsic_load_global: + case nir_intrinsic_load_scratch: + case nir_intrinsic_load_constant: + /* These memory load operations must have alignments */ + validate_assert(state, + util_is_power_of_two_nonzero(nir_intrinsic_align_mul(instr))); + validate_assert(state, nir_intrinsic_align_offset(instr) < + nir_intrinsic_align_mul(instr)); + /* Fall through */ + + case nir_intrinsic_load_uniform: + case nir_intrinsic_load_input: + case nir_intrinsic_load_per_vertex_input: + case nir_intrinsic_load_interpolated_input: + case nir_intrinsic_load_output: + case nir_intrinsic_load_per_vertex_output: + case nir_intrinsic_load_push_constant: + /* All memory load operations must load at least a byte */ + validate_assert(state, nir_dest_bit_size(instr->dest) >= 8); + break; + + case nir_intrinsic_store_ssbo: + case nir_intrinsic_store_shared: + case nir_intrinsic_store_global: + case nir_intrinsic_store_scratch: + /* These memory store operations must also have alignments */ + validate_assert(state, + util_is_power_of_two_nonzero(nir_intrinsic_align_mul(instr))); + validate_assert(state, nir_intrinsic_align_offset(instr) < + nir_intrinsic_align_mul(instr)); + /* Fall through */ + + case nir_intrinsic_store_output: + case nir_intrinsic_store_per_vertex_output: + /* All memory store operations must store at least a byte */ + validate_assert(state, nir_src_bit_size(instr->src[0]) >= 8); break; + default: break; } + + if (instr->num_components > 0) + validate_num_components(state, instr->num_components); + + const nir_intrinsic_info *info = &nir_intrinsic_infos[instr->intrinsic]; + unsigned num_srcs = info->num_srcs; + for (unsigned i = 0; i < num_srcs; i++) { + unsigned components_read = nir_intrinsic_src_components(instr, i); + + validate_num_components(state, components_read); + + validate_src(&instr->src[i], state, src_bit_sizes[i], components_read); + } + + if (nir_intrinsic_infos[instr->intrinsic].has_dest) { + unsigned components_written = nir_intrinsic_dest_components(instr); + unsigned bit_sizes = nir_intrinsic_infos[instr->intrinsic].dest_bit_sizes; + + validate_num_components(state, components_written); + if (dest_bit_size && bit_sizes) + validate_assert(state, dest_bit_size & bit_sizes); + else + dest_bit_size = dest_bit_size ? dest_bit_size : bit_sizes; + + validate_dest(&instr->dest, state, dest_bit_size, components_written); + } + + if (!vectorized_intrinsic(instr)) + validate_assert(state, instr->num_components == 0); } static void @@ -440,42 +677,82 @@ validate_tex_instr(nir_tex_instr *instr, validate_state *state) src_type_seen[i] = false; for (unsigned i = 0; i < instr->num_srcs; i++) { - assert(!src_type_seen[instr->src[i].src_type]); + validate_assert(state, !src_type_seen[instr->src[i].src_type]); src_type_seen[instr->src[i].src_type] = true; - validate_src(&instr->src[i].src, state); + validate_src(&instr->src[i].src, state, + 0, nir_tex_instr_src_size(instr, i)); + + switch (instr->src[i].src_type) { + case nir_tex_src_texture_deref: + case nir_tex_src_sampler_deref: + validate_assert(state, instr->src[i].src.is_ssa); + validate_assert(state, + instr->src[i].src.ssa->parent_instr->type == nir_instr_type_deref); + break; + default: + break; + } } - if (instr->texture != NULL) - validate_deref_var(instr, instr->texture, state); - - if (instr->sampler != NULL) - validate_deref_var(instr, instr->sampler, state); + if (nir_tex_instr_has_explicit_tg4_offsets(instr)) { + validate_assert(state, instr->op == nir_texop_tg4); + validate_assert(state, !src_type_seen[nir_tex_src_offset]); + } - validate_dest(&instr->dest, state); + validate_dest(&instr->dest, state, 0, nir_tex_instr_dest_size(instr)); } static void validate_call_instr(nir_call_instr *instr, validate_state *state) { - if (instr->return_deref == NULL) { - assert(glsl_type_is_void(instr->callee->return_type)); - } else { - assert(instr->return_deref->deref.type == instr->callee->return_type); - validate_deref_var(instr, instr->return_deref, state); - } - - assert(instr->num_params == instr->callee->num_params); + validate_assert(state, instr->num_params == instr->callee->num_params); for (unsigned i = 0; i < instr->num_params; i++) { - assert(instr->callee->params[i].type == instr->params[i]->deref.type); - validate_deref_var(instr, instr->params[i], state); + validate_src(&instr->params[i], state, + instr->callee->params[i].bit_size, + instr->callee->params[i].num_components); + } +} + +static void +validate_const_value(nir_const_value *val, unsigned bit_size, + validate_state *state) +{ + /* In order for block copies to work properly for things like instruction + * comparisons and [de]serialization, we require the unused bits of the + * nir_const_value to be zero. + */ + nir_const_value cmp_val; + memset(&cmp_val, 0, sizeof(cmp_val)); + switch (bit_size) { + case 1: + cmp_val.b = val->b; + break; + case 8: + cmp_val.u8 = val->u8; + break; + case 16: + cmp_val.u16 = val->u16; + break; + case 32: + cmp_val.u32 = val->u32; + break; + case 64: + cmp_val.u64 = val->u64; + break; + default: + validate_assert(state, !"Invalid load_const bit size"); } + validate_assert(state, memcmp(val, &cmp_val, sizeof(cmp_val)) == 0); } static void validate_load_const_instr(nir_load_const_instr *instr, validate_state *state) { validate_ssa_def(&instr->def, state); + + for (unsigned i = 0; i < instr->def.num_components; i++) + validate_const_value(&instr->value[i], instr->def.bit_size, state); } static void @@ -492,17 +769,78 @@ validate_phi_instr(nir_phi_instr *instr, validate_state *state) * basic blocks, to avoid validating an SSA use before its definition. */ - validate_dest(&instr->dest, state); + validate_dest(&instr->dest, state, 0, 0); exec_list_validate(&instr->srcs); - assert(exec_list_length(&instr->srcs) == + validate_assert(state, exec_list_length(&instr->srcs) == state->block->predecessors->entries); } +static void +validate_jump_instr(nir_jump_instr *instr, validate_state *state) +{ + nir_block *block = state->block; + validate_assert(state, &instr->instr == nir_block_last_instr(block)); + + switch (instr->type) { + case nir_jump_return: + validate_assert(state, block->successors[0] == state->impl->end_block); + validate_assert(state, block->successors[1] == NULL); + validate_assert(state, instr->target == NULL); + validate_assert(state, instr->else_target == NULL); + break; + + case nir_jump_break: + validate_assert(state, state->impl->structured); + validate_assert(state, state->loop != NULL); + if (state->loop) { + nir_block *after = + nir_cf_node_as_block(nir_cf_node_next(&state->loop->cf_node)); + validate_assert(state, block->successors[0] == after); + } + validate_assert(state, block->successors[1] == NULL); + validate_assert(state, instr->target == NULL); + validate_assert(state, instr->else_target == NULL); + break; + + case nir_jump_continue: + validate_assert(state, state->impl->structured); + validate_assert(state, state->loop != NULL); + if (state->loop) { + nir_block *first = nir_loop_first_block(state->loop); + validate_assert(state, block->successors[0] == first); + } + validate_assert(state, block->successors[1] == NULL); + validate_assert(state, instr->target == NULL); + validate_assert(state, instr->else_target == NULL); + break; + + case nir_jump_goto: + validate_assert(state, !state->impl->structured); + validate_assert(state, instr->target == block->successors[0]); + validate_assert(state, instr->target != NULL); + validate_assert(state, instr->else_target == NULL); + break; + + case nir_jump_goto_if: + validate_assert(state, !state->impl->structured); + validate_assert(state, instr->target == block->successors[1]); + validate_assert(state, instr->else_target == block->successors[0]); + validate_src(&instr->condition, state, 0, 1); + validate_assert(state, instr->target != NULL); + validate_assert(state, instr->else_target != NULL); + break; + + default: + validate_assert(state, !"Invalid jump instruction type"); + break; + } +} + static void validate_instr(nir_instr *instr, validate_state *state) { - assert(instr->block == state->block); + validate_assert(state, instr->block == state->block); state->instr = instr; @@ -511,6 +849,10 @@ validate_instr(nir_instr *instr, validate_state *state) validate_alu_instr(nir_instr_as_alu(instr), state); break; + case nir_instr_type_deref: + validate_deref_instr(nir_instr_as_deref(instr), state); + break; + case nir_instr_type_call: validate_call_instr(nir_instr_as_call(instr), state); break; @@ -536,10 +878,11 @@ validate_instr(nir_instr *instr, validate_state *state) break; case nir_instr_type_jump: + validate_jump_instr(nir_instr_as_jump(instr), state); break; default: - assert(!"Invalid ALU instruction type"); + validate_assert(state, !"Invalid ALU instruction type"); break; } @@ -551,28 +894,26 @@ validate_phi_src(nir_phi_instr *instr, nir_block *pred, validate_state *state) { state->instr = &instr->instr; - assert(instr->dest.is_ssa); + validate_assert(state, instr->dest.is_ssa); exec_list_validate(&instr->srcs); - nir_foreach_phi_src(instr, src) { + nir_foreach_phi_src(src, instr) { if (src->pred == pred) { - assert(src->src.is_ssa); - assert(src->src.ssa->num_components == - instr->dest.ssa.num_components); - - validate_src(&src->src, state); + validate_assert(state, src->src.is_ssa); + validate_src(&src->src, state, instr->dest.ssa.bit_size, + instr->dest.ssa.num_components); state->instr = NULL; return; } } - - abort(); + validate_assert(state, !"Phi does not have a source corresponding to one " + "of its predecessor blocks"); } static void validate_phi_srcs(nir_block *block, nir_block *succ, validate_state *state) { - nir_foreach_instr(succ, instr) { + nir_foreach_instr(instr, succ) { if (instr->type != nir_instr_type_phi) break; @@ -585,78 +926,48 @@ static void validate_cf_node(nir_cf_node *node, validate_state *state); static void validate_block(nir_block *block, validate_state *state) { - assert(block->cf_node.parent == state->parent_node); + validate_assert(state, block->cf_node.parent == state->parent_node); state->block = block; exec_list_validate(&block->instr_list); - nir_foreach_instr(block, instr) { + nir_foreach_instr(instr, block) { if (instr->type == nir_instr_type_phi) { - assert(instr == nir_block_first_instr(block) || + validate_assert(state, instr == nir_block_first_instr(block) || nir_instr_prev(instr)->type == nir_instr_type_phi); } - if (instr->type == nir_instr_type_jump) { - assert(instr == nir_block_last_instr(block)); - } - validate_instr(instr, state); } - assert(block->successors[0] != NULL); - assert(block->successors[0] != block->successors[1]); + validate_assert(state, block->successors[0] != NULL); + validate_assert(state, block->successors[0] != block->successors[1]); for (unsigned i = 0; i < 2; i++) { if (block->successors[i] != NULL) { struct set_entry *entry = _mesa_set_search(block->successors[i]->predecessors, block); - assert(entry); + validate_assert(state, entry); validate_phi_srcs(block, block->successors[i], state); } } - struct set_entry *entry; set_foreach(block->predecessors, entry) { const nir_block *pred = entry->key; - assert(pred->successors[0] == block || + validate_assert(state, pred->successors[0] == block || pred->successors[1] == block); } - if (!exec_list_is_empty(&block->instr_list) && - nir_block_last_instr(block)->type == nir_instr_type_jump) { - assert(block->successors[1] == NULL); - nir_jump_instr *jump = nir_instr_as_jump(nir_block_last_instr(block)); - switch (jump->type) { - case nir_jump_break: { - nir_block *after = - nir_cf_node_as_block(nir_cf_node_next(&state->loop->cf_node)); - assert(block->successors[0] == after); - break; - } - - case nir_jump_continue: { - nir_block *first = - nir_cf_node_as_block(nir_loop_first_cf_node(state->loop)); - assert(block->successors[0] == first); - break; - } - - case nir_jump_return: - assert(block->successors[0] == state->impl->end_block); - break; - - default: - unreachable("bad jump type"); - } - } else { + if (!state->impl->structured) { + validate_assert(state, nir_block_ends_in_jump(block)); + } else if (!nir_block_ends_in_jump(block)) { nir_cf_node *next = nir_cf_node_next(&block->cf_node); if (next == NULL) { switch (state->parent_node->type) { case nir_cf_node_loop: { - nir_block *first = - nir_cf_node_as_block(nir_loop_first_cf_node(state->loop)); - assert(block->successors[0] == first); + nir_block *first = nir_loop_first_block(state->loop); + validate_assert(state, block->successors[0] == first); /* due to the hack for infinite loops, block->successors[1] may * point to the block after the loop. */ @@ -666,14 +977,14 @@ validate_block(nir_block *block, validate_state *state) case nir_cf_node_if: { nir_block *after = nir_cf_node_as_block(nir_cf_node_next(state->parent_node)); - assert(block->successors[0] == after); - assert(block->successors[1] == NULL); + validate_assert(state, block->successors[0] == after); + validate_assert(state, block->successors[1] == NULL); break; } case nir_cf_node_function: - assert(block->successors[0] == state->impl->end_block); - assert(block->successors[1] == NULL); + validate_assert(state, block->successors[0] == state->impl->end_block); + validate_assert(state, block->successors[1] == NULL); break; default: @@ -682,16 +993,18 @@ validate_block(nir_block *block, validate_state *state) } else { if (next->type == nir_cf_node_if) { nir_if *if_stmt = nir_cf_node_as_if(next); - assert(&block->successors[0]->cf_node == - nir_if_first_then_node(if_stmt)); - assert(&block->successors[1]->cf_node == - nir_if_first_else_node(if_stmt)); - } else { - assert(next->type == nir_cf_node_loop); + validate_assert(state, block->successors[0] == + nir_if_first_then_block(if_stmt)); + validate_assert(state, block->successors[1] == + nir_if_first_else_block(if_stmt)); + } else if (next->type == nir_cf_node_loop) { nir_loop *loop = nir_cf_node_as_loop(next); - assert(&block->successors[0]->cf_node == - nir_loop_first_cf_node(loop)); - assert(block->successors[1] == NULL); + validate_assert(state, block->successors[0] == + nir_loop_first_block(loop)); + validate_assert(state, block->successors[1] == NULL); + } else { + validate_assert(state, + !"Structured NIR cannot have consecutive blocks"); } } } @@ -700,20 +1013,22 @@ validate_block(nir_block *block, validate_state *state) static void validate_if(nir_if *if_stmt, validate_state *state) { + validate_assert(state, state->impl->structured); + state->if_stmt = if_stmt; - assert(!exec_node_is_head_sentinel(if_stmt->cf_node.node.prev)); + validate_assert(state, !exec_node_is_head_sentinel(if_stmt->cf_node.node.prev)); nir_cf_node *prev_node = nir_cf_node_prev(&if_stmt->cf_node); - assert(prev_node->type == nir_cf_node_block); + validate_assert(state, prev_node->type == nir_cf_node_block); - assert(!exec_node_is_tail_sentinel(if_stmt->cf_node.node.next)); + validate_assert(state, !exec_node_is_tail_sentinel(if_stmt->cf_node.node.next)); nir_cf_node *next_node = nir_cf_node_next(&if_stmt->cf_node); - assert(next_node->type == nir_cf_node_block); + validate_assert(state, next_node->type == nir_cf_node_block); - validate_src(&if_stmt->condition, state); + validate_src(&if_stmt->condition, state, 0, 1); - assert(!exec_list_is_empty(&if_stmt->then_list)); - assert(!exec_list_is_empty(&if_stmt->else_list)); + validate_assert(state, !exec_list_is_empty(&if_stmt->then_list)); + validate_assert(state, !exec_list_is_empty(&if_stmt->else_list)); nir_cf_node *old_parent = state->parent_node; state->parent_node = &if_stmt->cf_node; @@ -735,15 +1050,17 @@ validate_if(nir_if *if_stmt, validate_state *state) static void validate_loop(nir_loop *loop, validate_state *state) { - assert(!exec_node_is_head_sentinel(loop->cf_node.node.prev)); + validate_assert(state, state->impl->structured); + + validate_assert(state, !exec_node_is_head_sentinel(loop->cf_node.node.prev)); nir_cf_node *prev_node = nir_cf_node_prev(&loop->cf_node); - assert(prev_node->type == nir_cf_node_block); + validate_assert(state, prev_node->type == nir_cf_node_block); - assert(!exec_node_is_tail_sentinel(loop->cf_node.node.next)); + validate_assert(state, !exec_node_is_tail_sentinel(loop->cf_node.node.next)); nir_cf_node *next_node = nir_cf_node_next(&loop->cf_node); - assert(next_node->type == nir_cf_node_block); + validate_assert(state, next_node->type == nir_cf_node_block); - assert(!exec_list_is_empty(&loop->body)); + validate_assert(state, !exec_list_is_empty(&loop->body)); nir_cf_node *old_parent = state->parent_node; state->parent_node = &loop->cf_node; @@ -762,7 +1079,7 @@ validate_loop(nir_loop *loop, validate_state *state) static void validate_cf_node(nir_cf_node *node, validate_state *state) { - assert(node->parent == state->parent_node); + validate_assert(state, node->parent == state->parent_node); switch (node->type) { case nir_cf_node_block: @@ -783,15 +1100,11 @@ validate_cf_node(nir_cf_node *node, validate_state *state) } static void -prevalidate_reg_decl(nir_register *reg, bool is_global, validate_state *state) +prevalidate_reg_decl(nir_register *reg, validate_state *state) { - assert(reg->is_global == is_global); - - if (is_global) - assert(reg->index < state->shader->reg_alloc); - else - assert(reg->index < state->impl->reg_alloc); - assert(!BITSET_TEST(state->regs_found, reg->index)); + validate_assert(state, reg->index < state->impl->reg_alloc); + validate_assert(state, !BITSET_TEST(state->regs_found, reg->index)); + validate_num_components(state, reg->num_components); BITSET_SET(state->regs_found, reg->index); list_validate(®->uses); @@ -799,14 +1112,11 @@ prevalidate_reg_decl(nir_register *reg, bool is_global, validate_state *state) list_validate(®->if_uses); reg_validate_state *reg_state = ralloc(state->regs, reg_validate_state); - reg_state->uses = _mesa_set_create(reg_state, _mesa_hash_pointer, - _mesa_key_pointer_equal); - reg_state->if_uses = _mesa_set_create(reg_state, _mesa_hash_pointer, - _mesa_key_pointer_equal); - reg_state->defs = _mesa_set_create(reg_state, _mesa_hash_pointer, - _mesa_key_pointer_equal); + reg_state->uses = _mesa_pointer_set_create(reg_state); + reg_state->if_uses = _mesa_pointer_set_create(reg_state); + reg_state->defs = _mesa_pointer_set_create(reg_state); - reg_state->where_defined = is_global ? NULL : state->impl; + reg_state->where_defined = state->impl; _mesa_hash_table_insert(state->regs, reg, reg_state); } @@ -816,167 +1126,114 @@ postvalidate_reg_decl(nir_register *reg, validate_state *state) { struct hash_entry *entry = _mesa_hash_table_search(state->regs, reg); + assume(entry); reg_validate_state *reg_state = (reg_validate_state *) entry->data; - nir_foreach_use(reg, src) { + nir_foreach_use(src, reg) { struct set_entry *entry = _mesa_set_search(reg_state->uses, src); - assert(entry); + validate_assert(state, entry); _mesa_set_remove(reg_state->uses, entry); } + validate_assert(state, reg_state->uses->entries == 0); - if (reg_state->uses->entries != 0) { - printf("extra entries in register uses:\n"); - struct set_entry *entry; - set_foreach(reg_state->uses, entry) - printf("%p\n", entry->key); - - abort(); - } - - nir_foreach_if_use(reg, src) { + nir_foreach_if_use(src, reg) { struct set_entry *entry = _mesa_set_search(reg_state->if_uses, src); - assert(entry); + validate_assert(state, entry); _mesa_set_remove(reg_state->if_uses, entry); } + validate_assert(state, reg_state->if_uses->entries == 0); - if (reg_state->if_uses->entries != 0) { - printf("extra entries in register if_uses:\n"); - struct set_entry *entry; - set_foreach(reg_state->if_uses, entry) - printf("%p\n", entry->key); - - abort(); - } - - nir_foreach_def(reg, src) { + nir_foreach_def(src, reg) { struct set_entry *entry = _mesa_set_search(reg_state->defs, src); - assert(entry); + validate_assert(state, entry); _mesa_set_remove(reg_state->defs, entry); } - - if (reg_state->defs->entries != 0) { - printf("extra entries in register defs:\n"); - struct set_entry *entry; - set_foreach(reg_state->defs, entry) - printf("%p\n", entry->key); - - abort(); - } + validate_assert(state, reg_state->defs->entries == 0); } static void -validate_var_decl(nir_variable *var, bool is_global, validate_state *state) -{ - assert(is_global == nir_variable_is_global(var)); - - /* - * TODO validate some things ir_validate.cpp does (requires more GLSL type - * support) - */ - - if (!is_global) { - _mesa_hash_table_insert(state->var_defs, var, state->impl); - } -} - -static bool -postvalidate_ssa_def(nir_ssa_def *def, void *void_state) +validate_var_decl(nir_variable *var, nir_variable_mode valid_modes, + validate_state *state) { - validate_state *state = void_state; - - struct hash_entry *entry = _mesa_hash_table_search(state->ssa_defs, def); - ssa_def_validate_state *def_state = (ssa_def_validate_state *)entry->data; + state->var = var; - nir_foreach_use(def, src) { - struct set_entry *entry = _mesa_set_search(def_state->uses, src); - assert(entry); - _mesa_set_remove(def_state->uses, entry); - } + /* Must have exactly one mode set */ + validate_assert(state, util_is_power_of_two_nonzero(var->data.mode)); + validate_assert(state, var->data.mode & valid_modes); - if (def_state->uses->entries != 0) { - printf("extra entries in register uses:\n"); - struct set_entry *entry; - set_foreach(def_state->uses, entry) - printf("%p\n", entry->key); + if (var->data.compact) { + /* The "compact" flag is only valid on arrays of scalars. */ + assert(glsl_type_is_array(var->type)); - abort(); + const struct glsl_type *type = glsl_get_array_element(var->type); + if (nir_is_per_vertex_io(var, state->shader->info.stage)) { + assert(glsl_type_is_array(type)); + assert(glsl_type_is_scalar(glsl_get_array_element(type))); + } else { + assert(glsl_type_is_scalar(type)); + } } - nir_foreach_if_use(def, src) { - struct set_entry *entry = _mesa_set_search(def_state->if_uses, src); - assert(entry); - _mesa_set_remove(def_state->if_uses, entry); + if (var->num_members > 0) { + const struct glsl_type *without_array = glsl_without_array(var->type); + validate_assert(state, glsl_type_is_struct_or_ifc(without_array)); + validate_assert(state, var->num_members == glsl_get_length(without_array)); + validate_assert(state, var->members != NULL); } - if (def_state->if_uses->entries != 0) { - printf("extra entries in register uses:\n"); - struct set_entry *entry; - set_foreach(def_state->if_uses, entry) - printf("%p\n", entry->key); - - abort(); - } + if (var->data.per_view) + validate_assert(state, glsl_type_is_array(var->type)); - return true; -} + /* + * TODO validate some things ir_validate.cpp does (requires more GLSL type + * support) + */ -static bool -postvalidate_ssa_defs_block(nir_block *block, void *state) -{ - nir_foreach_instr(block, instr) - nir_foreach_ssa_def(instr, postvalidate_ssa_def, state); + _mesa_hash_table_insert(state->var_defs, var, + valid_modes == nir_var_function_temp ? + state->impl : NULL); - return true; + state->var = NULL; } static void validate_function_impl(nir_function_impl *impl, validate_state *state) { - assert(impl->function->impl == impl); - assert(impl->cf_node.parent == NULL); - - assert(impl->num_params == impl->function->num_params); - for (unsigned i = 0; i < impl->num_params; i++) { - assert(impl->params[i]->type == impl->function->params[i].type); - assert(impl->params[i]->data.mode == nir_var_param); - assert(impl->params[i]->data.location == i); - validate_var_decl(impl->params[i], false, state); - } + /* Resize the ssa_srcs set. It's likely that the size of this set will + * never actually hit the number of SSA defs because we remove sources from + * the set as we visit them. (It could actually be much larger because + * each SSA def can be used more than once.) However, growing it now costs + * us very little (the extra memory is already dwarfed by the SSA defs + * themselves) and makes collisions much less likely. + */ + _mesa_set_resize(state->ssa_srcs, impl->ssa_alloc); - if (glsl_type_is_void(impl->function->return_type)) { - assert(impl->return_var == NULL); - } else { - assert(impl->return_var->type == impl->function->return_type); - assert(impl->return_var->data.mode == nir_var_param); - assert(impl->return_var->data.location == -1); - validate_var_decl(impl->return_var, false, state); - } + validate_assert(state, impl->function->impl == impl); + validate_assert(state, impl->cf_node.parent == NULL); - assert(exec_list_is_empty(&impl->end_block->instr_list)); - assert(impl->end_block->successors[0] == NULL); - assert(impl->end_block->successors[1] == NULL); + validate_assert(state, exec_list_is_empty(&impl->end_block->instr_list)); + validate_assert(state, impl->end_block->successors[0] == NULL); + validate_assert(state, impl->end_block->successors[1] == NULL); state->impl = impl; state->parent_node = &impl->cf_node; exec_list_validate(&impl->locals); - nir_foreach_variable(var, &impl->locals) { - validate_var_decl(var, false, state); + nir_foreach_function_temp_variable(var, impl) { + validate_var_decl(var, nir_var_function_temp, state); } - state->regs_found = realloc(state->regs_found, - BITSET_WORDS(impl->reg_alloc) * - sizeof(BITSET_WORD)); + state->regs_found = reralloc(state->mem_ctx, state->regs_found, + BITSET_WORD, BITSET_WORDS(impl->reg_alloc)); memset(state->regs_found, 0, BITSET_WORDS(impl->reg_alloc) * sizeof(BITSET_WORD)); exec_list_validate(&impl->registers); foreach_list_typed(nir_register, reg, node, &impl->registers) { - prevalidate_reg_decl(reg, false, state); + prevalidate_reg_decl(reg, state); } - state->ssa_defs_found = realloc(state->ssa_defs_found, - BITSET_WORDS(impl->ssa_alloc) * - sizeof(BITSET_WORD)); + state->ssa_defs_found = reralloc(state->mem_ctx, state->ssa_defs_found, + BITSET_WORD, BITSET_WORDS(impl->ssa_alloc)); memset(state->ssa_defs_found, 0, BITSET_WORDS(impl->ssa_alloc) * sizeof(BITSET_WORD)); exec_list_validate(&impl->body); @@ -988,14 +1245,15 @@ validate_function_impl(nir_function_impl *impl, validate_state *state) postvalidate_reg_decl(reg, state); } - nir_foreach_block(impl, postvalidate_ssa_defs_block, state); + validate_assert(state, state->ssa_srcs->entries == 0); + _mesa_set_clear(state->ssa_srcs, NULL); } static void validate_function(nir_function *func, validate_state *state) { if (func->impl != NULL) { - assert(func->impl->function == func); + validate_assert(state, func->impl->function == func); validate_function_impl(func->impl, state); } } @@ -1003,83 +1261,96 @@ validate_function(nir_function *func, validate_state *state) static void init_validate_state(validate_state *state) { - state->regs = _mesa_hash_table_create(NULL, _mesa_hash_pointer, - _mesa_key_pointer_equal); - state->ssa_defs = _mesa_hash_table_create(NULL, _mesa_hash_pointer, - _mesa_key_pointer_equal); + state->mem_ctx = ralloc_context(NULL); + state->regs = _mesa_pointer_hash_table_create(state->mem_ctx); + state->ssa_srcs = _mesa_pointer_set_create(state->mem_ctx); state->ssa_defs_found = NULL; state->regs_found = NULL; - state->var_defs = _mesa_hash_table_create(NULL, _mesa_hash_pointer, - _mesa_key_pointer_equal); + state->var_defs = _mesa_pointer_hash_table_create(state->mem_ctx); + state->errors = _mesa_pointer_hash_table_create(state->mem_ctx); + state->loop = NULL; + state->instr = NULL; + state->var = NULL; } static void destroy_validate_state(validate_state *state) { - _mesa_hash_table_destroy(state->regs, NULL); - _mesa_hash_table_destroy(state->ssa_defs, NULL); - free(state->ssa_defs_found); - free(state->regs_found); - _mesa_hash_table_destroy(state->var_defs, NULL); + ralloc_free(state->mem_ctx); } -void -nir_validate_shader(nir_shader *shader) +mtx_t fail_dump_mutex = _MTX_INITIALIZER_NP; + +static void +dump_errors(validate_state *state, const char *when) { - validate_state state; - init_validate_state(&state); + struct hash_table *errors = state->errors; - state.shader = shader; + /* Lock around dumping so that we get clean dumps in a multi-threaded + * scenario + */ + mtx_lock(&fail_dump_mutex); - exec_list_validate(&shader->uniforms); - nir_foreach_variable(var, &shader->uniforms) { - validate_var_decl(var, true, &state); + if (when) { + fprintf(stderr, "NIR validation failed %s\n", when); + fprintf(stderr, "%d errors:\n", _mesa_hash_table_num_entries(errors)); + } else { + fprintf(stderr, "NIR validation failed with %d errors:\n", + _mesa_hash_table_num_entries(errors)); } - exec_list_validate(&shader->inputs); - nir_foreach_variable(var, &shader->inputs) { - validate_var_decl(var, true, &state); - } + nir_print_shader_annotated(state->shader, stderr, errors); - exec_list_validate(&shader->outputs); - nir_foreach_variable(var, &shader->outputs) { - validate_var_decl(var, true, &state); + if (_mesa_hash_table_num_entries(errors) > 0) { + fprintf(stderr, "%d additional errors:\n", + _mesa_hash_table_num_entries(errors)); + hash_table_foreach(errors, entry) { + fprintf(stderr, "%s\n", (char *)entry->data); + } } - exec_list_validate(&shader->shared); - nir_foreach_variable(var, &shader->shared) { - validate_var_decl(var, true, &state); - } + mtx_unlock(&fail_dump_mutex); - exec_list_validate(&shader->globals); - nir_foreach_variable(var, &shader->globals) { - validate_var_decl(var, true, &state); - } + abort(); +} - exec_list_validate(&shader->system_values); - nir_foreach_variable(var, &shader->system_values) { - validate_var_decl(var, true, &state); - } +void +nir_validate_shader(nir_shader *shader, const char *when) +{ + static int should_validate = -1; + if (should_validate < 0) + should_validate = env_var_as_boolean("NIR_VALIDATE", true); + if (!should_validate) + return; - state.regs_found = realloc(state.regs_found, - BITSET_WORDS(shader->reg_alloc) * - sizeof(BITSET_WORD)); - memset(state.regs_found, 0, BITSET_WORDS(shader->reg_alloc) * - sizeof(BITSET_WORD)); - exec_list_validate(&shader->registers); - foreach_list_typed(nir_register, reg, node, &shader->registers) { - prevalidate_reg_decl(reg, true, &state); - } + validate_state state; + init_validate_state(&state); + + state.shader = shader; + + nir_variable_mode valid_modes = + nir_var_shader_in | + nir_var_shader_out | + nir_var_shader_temp | + nir_var_uniform | + nir_var_mem_ubo | + nir_var_system_value | + nir_var_mem_ssbo | + nir_var_mem_shared | + nir_var_mem_constant; + + exec_list_validate(&shader->variables); + nir_foreach_variable_in_shader(var, shader) + validate_var_decl(var, valid_modes, &state); exec_list_validate(&shader->functions); foreach_list_typed(nir_function, func, node, &shader->functions) { validate_function(func, &state); } - foreach_list_typed(nir_register, reg, node, &shader->registers) { - postvalidate_reg_decl(reg, &state); - } + if (_mesa_hash_table_num_entries(state.errors) > 0) + dump_errors(&state, when); destroy_validate_state(&state); }