Temp sc = bld.vop3(aco_opcode::v_cubesc_f32, bld.def(v1), src[0], src[1], src[2]);
Temp tc = bld.vop3(aco_opcode::v_cubetc_f32, bld.def(v1), src[0], src[1], src[2]);
sc = bld.vop2(aco_opcode::v_add_f32, bld.def(v1),
- bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), sc, ma), Operand(0x3f000000u/*0.5*/));
+ Operand(0x3f000000u/*0.5*/),
+ bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), sc, ma));
tc = bld.vop2(aco_opcode::v_add_f32, bld.def(v1),
- bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), tc, ma), Operand(0x3f000000u/*0.5*/));
+ Operand(0x3f000000u/*0.5*/),
+ bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), tc, ma));
bld.pseudo(aco_opcode::p_create_vector, Definition(dst), sc, tc);
break;
}
}
case nir_op_ubfe:
case nir_op_ibfe: {
- Temp base = get_alu_src(ctx, instr->src[0]);
- Temp offset = get_alu_src(ctx, instr->src[1]);
- Temp bits = get_alu_src(ctx, instr->src[2]);
-
if (dst.bytes() != 4)
unreachable("Unsupported BFE bit size");
if (dst.type() == RegType::sgpr) {
+ Temp base = get_alu_src(ctx, instr->src[0]);
+
nir_const_value* const_offset = nir_src_as_const_value(instr->src[1].src);
nir_const_value* const_bits = nir_src_as_const_value(instr->src[2].src);
if (const_offset && const_bits) {
uint32_t extract = (const_bits->u32 << 16) | (const_offset->u32 & 0x1f);
aco_opcode opcode = instr->op == nir_op_ubfe ? aco_opcode::s_bfe_u32 : aco_opcode::s_bfe_i32;
bld.sop2(opcode, Definition(dst), bld.def(s1, scc), base, Operand(extract));
- } else if (instr->op == nir_op_ubfe) {
+ break;
+ }
+
+ Temp offset = get_alu_src(ctx, instr->src[1]);
+ Temp bits = get_alu_src(ctx, instr->src[2]);
+ if (instr->op == nir_op_ubfe) {
Temp mask = bld.sop2(aco_opcode::s_bfm_b32, bld.def(s1), bits, offset);
Temp masked = bld.sop2(aco_opcode::s_and_b32, bld.def(s1), bld.def(s1, scc), base, mask);
bld.sop2(aco_opcode::s_lshr_b32, Definition(dst), bld.def(s1, scc), masked, offset);