break;
case nir_op_unpack_double_2x32_split_x:
- bld.MOV(result, subscript(op[0], BRW_REGISTER_TYPE_UD, 0));
- break;
+ case nir_op_unpack_double_2x32_split_y: {
+ /* Optimize the common case where we are unpacking from a double we have
+ * previously packed. In this case we can just bypass the pack operation
+ * and source directly from its arguments.
+ */
+ unsigned index = (instr->op == nir_op_unpack_double_2x32_split_x) ? 0 : 1;
+ if (instr->src[0].src.is_ssa) {
+ nir_instr *parent_instr = instr->src[0].src.ssa->parent_instr;
+ if (parent_instr->type == nir_instr_type_alu) {
+ nir_alu_instr *alu_parent = nir_instr_as_alu(parent_instr);
+ if (alu_parent->op == nir_op_pack_double_2x32_split &&
+ alu_parent->src[index].src.is_ssa) {
+ op[0] = retype(get_nir_src(alu_parent->src[index].src),
+ BRW_REGISTER_TYPE_UD);
+ op[0] =
+ offset(op[0], bld, alu_parent->src[index].swizzle[channel]);
+ bld.MOV(result, op[0]);
+ break;
+ }
+ }
+ }
- case nir_op_unpack_double_2x32_split_y:
- bld.MOV(result, subscript(op[0], BRW_REGISTER_TYPE_UD, 1));
+ if (instr->op == nir_op_unpack_double_2x32_split_x)
+ bld.MOV(result, subscript(op[0], BRW_REGISTER_TYPE_UD, 0));
+ else
+ bld.MOV(result, subscript(op[0], BRW_REGISTER_TYPE_UD, 1));
break;
+ }
case nir_op_fpow:
inst = bld.emit(SHADER_OPCODE_POW, result, op[0], op[1]);