var->data.how_declared = nir_var_declared_normally;
if ((mode == nir_var_shader_in &&
- shader->info.stage != MESA_SHADER_VERTEX) ||
+ shader->info.stage != MESA_SHADER_VERTEX &&
+ shader->info.stage != MESA_SHADER_KERNEL) ||
(mode == nir_var_shader_out &&
shader->info.stage != MESA_SHADER_FRAGMENT))
var->data.interpolation = INTERP_MODE_SMOOTH;
{
switch (cursor.option) {
case nir_cursor_before_block:
- assert(nir_cf_node_prev(&cursor.block->cf_node) == NULL ||
- nir_cf_node_prev(&cursor.block->cf_node)->type != nir_cf_node_block);
if (exec_list_is_empty(&cursor.block->instr_list)) {
/* Empty block. After is as good as before. */
cursor.option = nir_cursor_after_block;
nir_cf_node_get_function(&instr->block->cf_node);
def->index = impl->ssa_alloc++;
+
+ impl->valid_metadata &= ~nir_metadata_live_ssa_defs;
}
return true;
unsigned bit_size, const char *name)
{
def->name = ralloc_strdup(instr, name);
+ def->live_index = UINT_MAX; /* Something clearly OOB */
def->parent_instr = instr;
list_inithead(&def->uses);
list_inithead(&def->if_uses);
nir_cf_node_get_function(&instr->block->cf_node);
def->index = impl->ssa_alloc++;
+
+ impl->valid_metadata &= ~nir_metadata_live_ssa_defs;
} else {
def->index = UINT_MAX;
}
{
unsigned index = 0;
+ impl->valid_metadata &= ~nir_metadata_live_ssa_defs;
+
nir_foreach_block_unstructured(block, impl) {
nir_foreach_instr(instr, block)
nir_foreach_ssa_def(instr, index_ssa_def_cb, &index);
return nir_intrinsic_load_local_group_size;
case SYSTEM_VALUE_GLOBAL_INVOCATION_ID:
return nir_intrinsic_load_global_invocation_id;
+ case SYSTEM_VALUE_BASE_GLOBAL_INVOCATION_ID:
+ return nir_intrinsic_load_base_global_invocation_id;
case SYSTEM_VALUE_GLOBAL_INVOCATION_INDEX:
return nir_intrinsic_load_global_invocation_index;
case SYSTEM_VALUE_WORK_DIM:
return SYSTEM_VALUE_LOCAL_GROUP_SIZE;
case nir_intrinsic_load_global_invocation_id:
return SYSTEM_VALUE_GLOBAL_INVOCATION_ID;
+ case nir_intrinsic_load_base_global_invocation_id:
+ return SYSTEM_VALUE_BASE_GLOBAL_INVOCATION_ID;
+ case nir_intrinsic_load_global_invocation_index:
+ return SYSTEM_VALUE_GLOBAL_INVOCATION_INDEX;
+ case nir_intrinsic_load_work_dim:
+ return SYSTEM_VALUE_WORK_DIM;
case nir_intrinsic_load_user_data_amd:
return SYSTEM_VALUE_USER_DATA_AMD;
default:
bool bindless)
{
enum gl_access_qualifier access = nir_intrinsic_access(intrin);
+ nir_alu_type type = nir_type_invalid;
+ if (nir_intrinsic_infos[intrin->intrinsic].index_map[NIR_INTRINSIC_TYPE])
+ type = nir_intrinsic_type(intrin);
switch (intrin->intrinsic) {
#define CASE(op) \
nir_intrinsic_set_image_array(intrin, glsl_sampler_type_is_array(deref->type));
nir_intrinsic_set_access(intrin, access | var->data.access);
nir_intrinsic_set_format(intrin, var->data.image.format);
+ if (nir_intrinsic_infos[intrin->intrinsic].index_map[NIR_INTRINSIC_TYPE])
+ nir_intrinsic_set_type(intrin, type);
nir_instr_rewrite_src(&intrin->instr, &intrin->src[0],
nir_src_for_ssa(src));