return nir_if_first_else_block(if_stmt);
assert(block == nir_if_last_else_block(if_stmt));
- /* fall through */
}
+ /* fallthrough */
case nir_cf_node_loop:
return nir_cf_node_as_block(nir_cf_node_next(parent));
return nir_if_last_then_block(if_stmt);
assert(block == nir_if_first_then_block(if_stmt));
- /* fall through */
}
+ /* fallthrough */
case nir_cf_node_loop:
return nir_cf_node_as_block(nir_cf_node_prev(parent));
case nir_intrinsic_demote:
case nir_intrinsic_demote_if:
shader->info.fs.uses_demote = true;
- /* fallthrough: quads with helper lanes only might be discarded entirely */
+ /* fallthrough - quads with helper lanes only might be discarded entirely */
case nir_intrinsic_discard:
case nir_intrinsic_discard_if:
/* Freedreno uses the discard_if intrinsic to end GS invocations that
break;
}
}
+ /* fallthrough */
default:
/* We can't scalarize this type of instruction */
default:
break;
}
+
+ return false;
}
default:
instr->pass_flags = GCM_INSTR_SCHEDULE_EARLIER_ONLY;
break;
- case nir_op_mov:
- if (!is_src_scalarizable(&(nir_instr_as_alu(instr)->src[0].src))) {
- instr->pass_flags = GCM_INSTR_PINNED;
- break;
- }
+ case nir_op_mov:
+ if (!is_src_scalarizable(&(nir_instr_as_alu(instr)->src[0].src))) {
+ instr->pass_flags = GCM_INSTR_PINNED;
+ break;
+ }
+ /* fallthrough */
default:
instr->pass_flags = 0;
nir_intrinsic_instr *intrinsic = nir_instr_as_intrinsic(instr);
if (!(nir_intrinsic_infos[intrinsic->intrinsic].flags & NIR_INTRINSIC_CAN_REORDER))
return not_invariant;
- /* fallthrough */
}
+ /* fallthrough */
default:
return nir_foreach_src(instr, src_is_invariant, loop) ? invariant : not_invariant;
}