X-Git-Url: https://git.libre-soc.org/?a=blobdiff_plain;f=src%2Fgallium%2Fauxiliary%2Fgallivm%2Flp_bld_sample_soa.c;h=a852490f041da9fc8dd49bd461ef9dbac6edaf90;hb=6e25a5a3756546b49bfaf53a2b16cc9b6d313928;hp=337b6f73fd12a9db3aaa86187099656da6da5def;hpb=e1590b9690cfc5c3f304b5339621129226a20f36;p=mesa.git diff --git a/src/gallium/auxiliary/gallivm/lp_bld_sample_soa.c b/src/gallium/auxiliary/gallivm/lp_bld_sample_soa.c index 337b6f73fd1..a852490f041 100644 --- a/src/gallium/auxiliary/gallivm/lp_bld_sample_soa.c +++ b/src/gallium/auxiliary/gallivm/lp_bld_sample_soa.c @@ -40,8 +40,9 @@ #include "util/u_dump.h" #include "util/u_memory.h" #include "util/u_math.h" -#include "util/u_format.h" +#include "util/format/u_format.h" #include "util/u_cpu_detect.h" +#include "util/format_rgb9e5.h" #include "lp_bld_debug.h" #include "lp_bld_type.h" #include "lp_bld_const.h" @@ -59,6 +60,8 @@ #include "lp_bld_struct.h" #include "lp_bld_quad.h" #include "lp_bld_pack.h" +#include "lp_bld_intr.h" +#include "lp_bld_misc.h" /** @@ -72,7 +75,6 @@ */ static void lp_build_sample_texel_soa(struct lp_build_sample_context *bld, - unsigned sampler_unit, LLVMValueRef width, LLVMValueRef height, LLVMValueRef depth, @@ -158,9 +160,10 @@ lp_build_sample_texel_soa(struct lp_build_sample_context *bld, lp_build_fetch_rgba_soa(bld->gallivm, bld->format_desc, - bld->texel_type, + bld->texel_type, TRUE, data_ptr, offset, i, j, + bld->cache, texel_out); /* @@ -179,53 +182,87 @@ lp_build_sample_texel_soa(struct lp_build_sample_context *bld, */ if (use_border) { - /* select texel color or border color depending on use_border */ - LLVMValueRef border_color_ptr = - bld->dynamic_state->border_color(bld->dynamic_state, - bld->gallivm, sampler_unit); + /* select texel color or border color depending on use_border. */ + const struct util_format_description *format_desc = bld->format_desc; int chan; + struct lp_type border_type = bld->texel_type; + border_type.length = 4; + /* + * Only replace channels which are actually present. The others should + * get optimized away eventually by sampler_view swizzle anyway but it's + * easier too. + */ for (chan = 0; chan < 4; chan++) { - LLVMValueRef border_chan = - lp_build_array_get(bld->gallivm, border_color_ptr, - lp_build_const_int32(bld->gallivm, chan)); - LLVMValueRef border_chan_vec = - lp_build_broadcast_scalar(&bld->float_vec_bld, border_chan); - - if (!bld->texel_type.floating) { - border_chan_vec = LLVMBuildBitCast(builder, border_chan_vec, - bld->texel_bld.vec_type, ""); + unsigned chan_s; + /* reverse-map channel... */ + if (util_format_has_stencil(format_desc)) { + if (chan == 0) + chan_s = 0; + else + break; + } + else { + for (chan_s = 0; chan_s < 4; chan_s++) { + if (chan_s == format_desc->swizzle[chan]) { + break; + } + } + } + if (chan_s <= 3) { + /* use the already clamped color */ + LLVMValueRef idx = lp_build_const_int32(bld->gallivm, chan); + LLVMValueRef border_chan; + + border_chan = lp_build_extract_broadcast(bld->gallivm, + border_type, + bld->texel_type, + bld->border_color_clamped, + idx); + texel_out[chan] = lp_build_select(&bld->texel_bld, use_border, + border_chan, texel_out[chan]); } - texel_out[chan] = lp_build_select(&bld->texel_bld, use_border, - border_chan_vec, texel_out[chan]); } } } /** - * Helper to compute the mirror function for the PIPE_WRAP_MIRROR modes. + * Helper to compute the mirror function for the PIPE_WRAP_MIRROR_REPEAT mode. + * (Note that with pot sizes could do this much more easily post-scale + * with some bit arithmetic.) */ static LLVMValueRef lp_build_coord_mirror(struct lp_build_sample_context *bld, - LLVMValueRef coord) + LLVMValueRef coord, boolean posOnly) { struct lp_build_context *coord_bld = &bld->coord_bld; - struct lp_build_context *int_coord_bld = &bld->int_coord_bld; - LLVMValueRef fract, flr, isOdd; - - lp_build_ifloor_fract(coord_bld, coord, &flr, &fract); - - /* isOdd = flr & 1 */ - isOdd = LLVMBuildAnd(bld->gallivm->builder, flr, int_coord_bld->one, ""); + LLVMValueRef fract; + LLVMValueRef half = lp_build_const_vec(bld->gallivm, coord_bld->type, 0.5); - /* make coord positive or negative depending on isOdd */ - coord = lp_build_set_sign(coord_bld, fract, isOdd); + /* + * We can just use 2*(x - round(0.5*x)) to do all the mirroring, + * it all works out. (The result is in range [-1, 1.0], negative if + * the coord is in the "odd" section, otherwise positive.) + */ - /* convert isOdd to float */ - isOdd = lp_build_int_to_float(coord_bld, isOdd); + coord = lp_build_mul(coord_bld, coord, half); + fract = lp_build_round(coord_bld, coord); + fract = lp_build_sub(coord_bld, coord, fract); + coord = lp_build_add(coord_bld, fract, fract); - /* add isOdd to coord */ - coord = lp_build_add(coord_bld, coord, isOdd); + if (posOnly) { + /* + * Theoretically it's not quite 100% accurate because the spec says + * that ultimately a scaled coord of -x.0 should map to int coord + * -x + 1 with mirroring, not -x (this does not matter for bilinear + * filtering). + */ + coord = lp_build_abs(coord_bld, coord); + /* kill off NaNs */ + /* XXX: not safe without arch rounding, fract can be anything. */ + coord = lp_build_max_ext(coord_bld, coord, coord_bld->zero, + GALLIVM_NAN_RETURN_OTHER_SECOND_NONNAN); + } return coord; } @@ -258,10 +295,15 @@ lp_build_coord_repeat_npot_linear(struct lp_build_sample_context *bld, * we avoided the 0.5/length division before the repeat wrap, * now need to fix up edge cases with selects */ + /* + * Note we do a float (unordered) compare so we can eliminate NaNs. + * (Otherwise would need fract_safe above). + */ + mask = lp_build_compare(coord_bld->gallivm, coord_bld->type, + PIPE_FUNC_LESS, coord_f, coord_bld->zero); + /* convert to int, compute lerp weight */ lp_build_ifloor_fract(coord_bld, coord_f, coord0_i, weight_f); - mask = lp_build_compare(int_coord_bld->gallivm, int_coord_bld->type, - PIPE_FUNC_LESS, *coord0_i, int_coord_bld->zero); *coord0_i = lp_build_select(int_coord_bld, mask, length_minus_one, *coord0_i); } @@ -274,6 +316,7 @@ lp_build_coord_repeat_npot_linear(struct lp_build_sample_context *bld, */ static void lp_build_sample_wrap_linear(struct lp_build_sample_context *bld, + boolean is_gather, LLVMValueRef coord, LLVMValueRef length, LLVMValueRef length_f, @@ -336,7 +379,13 @@ lp_build_sample_wrap_linear(struct lp_build_sample_context *bld, coord = lp_build_add(coord_bld, coord, offset); } - /* clamp to [0, length] */ + /* + * clamp to [0, length] + * + * Unlike some other wrap modes, this should be correct for gather + * too. GL_CLAMP explicitly does this clamp on the coord prior to + * actual wrapping (which is per sample). + */ coord = lp_build_clamp(coord_bld, coord, coord_bld->zero, length_f); coord = lp_build_sub(coord_bld, coord, half); @@ -361,14 +410,31 @@ lp_build_sample_wrap_linear(struct lp_build_sample_context *bld, } /* clamp to length max */ - coord = lp_build_min(coord_bld, coord, length_f); - /* subtract 0.5 */ - coord = lp_build_sub(coord_bld, coord, half); - /* clamp to [0, length - 0.5] */ - coord = lp_build_max(coord_bld, coord, coord_bld->zero); - /* convert to int, compute lerp weight */ - lp_build_ifloor_fract(&abs_coord_bld, coord, &coord0, &weight); - coord1 = lp_build_add(int_coord_bld, coord0, int_coord_bld->one); + coord = lp_build_min_ext(coord_bld, coord, length_f, + GALLIVM_NAN_RETURN_OTHER_SECOND_NONNAN); + if (!is_gather) { + /* subtract 0.5 */ + coord = lp_build_sub(coord_bld, coord, half); + /* clamp to [0, length - 0.5] */ + coord = lp_build_max(coord_bld, coord, coord_bld->zero); + /* convert to int, compute lerp weight */ + lp_build_ifloor_fract(&abs_coord_bld, coord, &coord0, &weight); + coord1 = lp_build_add(int_coord_bld, coord0, int_coord_bld->one); + } else { + /* + * The non-gather path will end up with coords 0, 1 if coord was + * smaller than 0.5 (with corresponding weight 0.0 so it doesn't + * really matter what the second coord is). But for gather, we + * really need to end up with coords 0, 0. + */ + coord = lp_build_max(coord_bld, coord, coord_bld->zero); + coord0 = lp_build_sub(coord_bld, coord, half); + coord1 = lp_build_add(coord_bld, coord, half); + /* Values range ([-0.5, length_f - 0.5], [0.5, length_f + 0.5] */ + coord0 = lp_build_itrunc(coord_bld, coord0); + coord1 = lp_build_itrunc(coord_bld, coord1); + weight = coord_bld->undef; + } /* coord1 = min(coord1, length-1) */ coord1 = lp_build_min(int_coord_bld, coord1, length_minus_one); break; @@ -383,8 +449,13 @@ lp_build_sample_wrap_linear(struct lp_build_sample_context *bld, offset = lp_build_int_to_float(coord_bld, offset); coord = lp_build_add(coord_bld, coord, offset); } - /* was: clamp to [-0.5, length + 0.5], then sub 0.5 */ - /* can skip clamp (though might not work for very large coord values */ + /* + * We don't need any clamp. Technically, for very large (pos or neg) + * (or infinite) values, clamp against [-length, length] would be + * correct, but we don't need to guarantee any specific + * result for such coords (the ifloor will be undefined, but for modes + * requiring border all resulting coords are safe). + */ coord = lp_build_sub(coord_bld, coord, half); /* convert to int, compute lerp weight */ lp_build_ifloor_fract(coord_bld, coord, &coord0, &weight); @@ -392,25 +463,69 @@ lp_build_sample_wrap_linear(struct lp_build_sample_context *bld, break; case PIPE_TEX_WRAP_MIRROR_REPEAT: - /* compute mirror function */ - coord = lp_build_coord_mirror(bld, coord); - - /* scale coord to length */ - coord = lp_build_mul(coord_bld, coord, length_f); - coord = lp_build_sub(coord_bld, coord, half); if (offset) { offset = lp_build_int_to_float(coord_bld, offset); + offset = lp_build_div(coord_bld, offset, length_f); coord = lp_build_add(coord_bld, coord, offset); } + if (!is_gather) { + /* compute mirror function */ + coord = lp_build_coord_mirror(bld, coord, TRUE); - /* convert to int, compute lerp weight */ - lp_build_ifloor_fract(coord_bld, coord, &coord0, &weight); - coord1 = lp_build_add(int_coord_bld, coord0, int_coord_bld->one); + /* scale coord to length */ + coord = lp_build_mul(coord_bld, coord, length_f); + coord = lp_build_sub(coord_bld, coord, half); + + /* convert to int, compute lerp weight */ + lp_build_ifloor_fract(coord_bld, coord, &coord0, &weight); + coord1 = lp_build_add(int_coord_bld, coord0, int_coord_bld->one); + + /* coord0 = max(coord0, 0) */ + coord0 = lp_build_max(int_coord_bld, coord0, int_coord_bld->zero); + /* coord1 = min(coord1, length-1) */ + coord1 = lp_build_min(int_coord_bld, coord1, length_minus_one); + } else { + /* + * This is pretty reasonable in the end, all what the tests care + * about is nasty edge cases (scaled coords x.5, so the individual + * coords are actually integers, which is REALLY tricky to get right + * due to this working differently both for negative numbers as well + * as for even/odd cases). But with enough magic it's not too complex + * after all. + * Maybe should try a bit arithmetic one though for POT textures... + */ + LLVMValueRef isNeg; + /* + * Wrapping just once still works, even though it means we can + * get "wrong" sign due to performing mirror in the middle of the + * two coords (because this can only happen very near the odd/even + * edges, so both coords will actually end up as 0 or length - 1 + * in the end). + * For GL4 gather with per-sample offsets we'd need to the mirroring + * per coord too. + */ + coord = lp_build_coord_mirror(bld, coord, FALSE); + coord = lp_build_mul(coord_bld, coord, length_f); + + /* + * NaNs should be safe here, we'll do away with them with + * the ones' complement plus min. + */ + coord0 = lp_build_sub(coord_bld, coord, half); + coord0 = lp_build_ifloor(coord_bld, coord0); + coord1 = lp_build_add(int_coord_bld, coord0, int_coord_bld->one); + /* ones complement for neg numbers (mirror(negX) = X - 1) */ + isNeg = lp_build_cmp(int_coord_bld, PIPE_FUNC_LESS, + coord0, int_coord_bld->zero); + coord0 = lp_build_xor(int_coord_bld, coord0, isNeg); + isNeg = lp_build_cmp(int_coord_bld, PIPE_FUNC_LESS, + coord1, int_coord_bld->zero); + coord1 = lp_build_xor(int_coord_bld, coord1, isNeg); + coord0 = lp_build_min(int_coord_bld, coord0, length_minus_one); + coord1 = lp_build_min(int_coord_bld, coord1, length_minus_one); - /* coord0 = max(coord0, 0) */ - coord0 = lp_build_max(int_coord_bld, coord0, int_coord_bld->zero); - /* coord1 = min(coord1, length-1) */ - coord1 = lp_build_min(int_coord_bld, coord1, length_minus_one); + weight = coord_bld->undef; + } break; case PIPE_TEX_WRAP_MIRROR_CLAMP: @@ -422,10 +537,19 @@ lp_build_sample_wrap_linear(struct lp_build_sample_context *bld, offset = lp_build_int_to_float(coord_bld, offset); coord = lp_build_add(coord_bld, coord, offset); } + /* + * XXX: probably not correct for gather, albeit I'm not + * entirely sure as it's poorly specified. The wrapping looks + * correct according to the spec which is against gl 1.2.1, + * however negative values will be swapped - gl re-specified + * wrapping with newer versions (no more pre-clamp except with + * GL_CLAMP). + */ coord = lp_build_abs(coord_bld, coord); /* clamp to [0, length] */ - coord = lp_build_min(coord_bld, coord, length_f); + coord = lp_build_min_ext(coord_bld, coord, length_f, + GALLIVM_NAN_RETURN_OTHER_SECOND_NONNAN); coord = lp_build_sub(coord_bld, coord, half); @@ -447,20 +571,59 @@ lp_build_sample_wrap_linear(struct lp_build_sample_context *bld, offset = lp_build_int_to_float(coord_bld, offset); coord = lp_build_add(coord_bld, coord, offset); } - coord = lp_build_abs(coord_bld, coord); - - /* clamp to length max */ - coord = lp_build_min(coord_bld, coord, length_f); - /* subtract 0.5 */ - coord = lp_build_sub(coord_bld, coord, half); - /* clamp to [0, length - 0.5] */ - coord = lp_build_max(coord_bld, coord, coord_bld->zero); - - /* convert to int, compute lerp weight */ - lp_build_ifloor_fract(&abs_coord_bld, coord, &coord0, &weight); - coord1 = lp_build_add(int_coord_bld, coord0, int_coord_bld->one); - /* coord1 = min(coord1, length-1) */ - coord1 = lp_build_min(int_coord_bld, coord1, length_minus_one); + if (!is_gather) { + coord = lp_build_abs(coord_bld, coord); + + /* clamp to length max */ + coord = lp_build_min_ext(coord_bld, coord, length_f, + GALLIVM_NAN_RETURN_OTHER_SECOND_NONNAN); + /* subtract 0.5 */ + coord = lp_build_sub(coord_bld, coord, half); + /* clamp to [0, length - 0.5] */ + coord = lp_build_max(coord_bld, coord, coord_bld->zero); + + /* convert to int, compute lerp weight */ + lp_build_ifloor_fract(&abs_coord_bld, coord, &coord0, &weight); + coord1 = lp_build_add(int_coord_bld, coord0, int_coord_bld->one); + /* coord1 = min(coord1, length-1) */ + coord1 = lp_build_min(int_coord_bld, coord1, length_minus_one); + } else { + /* + * The non-gather path will swap coord0/1 if coord was negative, + * which is ok for filtering since the filter weight matches + * accordingly. Also, if coord is close to zero, coord0/1 will + * be 0 and 1, instead of 0 and 0 (again ok due to filter + * weight being 0.0). Both issues need to be fixed for gather. + */ + LLVMValueRef isNeg; + + /* + * Actually wanted to cheat here and use: + * coord1 = lp_build_iround(coord_bld, coord); + * but it's not good enough for some tests (even piglit + * textureGather is set up in a way so the coords area always + * .5, that is right at the crossover points). + * So do ordinary sub/floor, then do ones' complement + * for negative numbers. + * (Note can't just do sub|add/abs/itrunc per coord neither - + * because the spec demands that mirror(3.0) = 3 but + * mirror(-3.0) = 2.) + */ + coord = lp_build_sub(coord_bld, coord, half); + coord0 = lp_build_ifloor(coord_bld, coord); + coord1 = lp_build_add(int_coord_bld, coord0, int_coord_bld->one); + isNeg = lp_build_cmp(int_coord_bld, PIPE_FUNC_LESS, coord0, + int_coord_bld->zero); + coord0 = lp_build_xor(int_coord_bld, isNeg, coord0); + coord0 = lp_build_min(int_coord_bld, coord0, length_minus_one); + + isNeg = lp_build_cmp(int_coord_bld, PIPE_FUNC_LESS, coord1, + int_coord_bld->zero); + coord1 = lp_build_xor(int_coord_bld, isNeg, coord1); + coord1 = lp_build_min(int_coord_bld, coord1, length_minus_one); + + weight = coord_bld->undef; + } } break; @@ -474,11 +637,20 @@ lp_build_sample_wrap_linear(struct lp_build_sample_context *bld, offset = lp_build_int_to_float(coord_bld, offset); coord = lp_build_add(coord_bld, coord, offset); } + /* + * XXX: probably not correct for gather due to swapped + * order if coord is negative (same rationale as for + * MIRROR_CLAMP). + */ coord = lp_build_abs(coord_bld, coord); - /* was: clamp to [-0.5, length + 0.5] then sub 0.5 */ - /* skip clamp - always positive, and other side - only potentially matters for very large coords */ + /* + * We don't need any clamp. Technically, for very large + * (or infinite) values, clamp against length would be + * correct, but we don't need to guarantee any specific + * result for such coords (the ifloor will be undefined, but + * for modes requiring border all resulting coords are safe). + */ coord = lp_build_sub(coord_bld, coord, half); /* convert to int, compute lerp weight */ @@ -554,12 +726,13 @@ lp_build_sample_wrap_nearest(struct lp_build_sample_context *bld, coord = lp_build_mul(coord_bld, coord, length_f); } + if (offset) { + offset = lp_build_int_to_float(coord_bld, offset); + coord = lp_build_add(coord_bld, coord, offset); + } /* floor */ /* use itrunc instead since we clamp to 0 anyway */ icoord = lp_build_itrunc(coord_bld, coord); - if (offset) { - icoord = lp_build_add(int_coord_bld, icoord, offset); - } /* clamp to [0, length - 1]. */ icoord = lp_build_clamp(int_coord_bld, icoord, int_coord_bld->zero, @@ -585,7 +758,7 @@ lp_build_sample_wrap_nearest(struct lp_build_sample_context *bld, coord = lp_build_add(coord_bld, coord, offset); } /* compute mirror function */ - coord = lp_build_coord_mirror(bld, coord); + coord = lp_build_coord_mirror(bld, coord, TRUE); /* scale coord to length */ assert(bld->static_sampler_state->normalized_coords); @@ -612,9 +785,15 @@ lp_build_sample_wrap_nearest(struct lp_build_sample_context *bld, /* itrunc == ifloor here */ icoord = lp_build_itrunc(coord_bld, coord); - - /* clamp to [0, length - 1] */ - icoord = lp_build_min(int_coord_bld, icoord, length_minus_one); + /* + * Use unsigned min due to possible undef values (NaNs, overflow) + */ + { + struct lp_build_context abs_coord_bld = *int_coord_bld; + abs_coord_bld.type.sign = FALSE; + /* clamp to [0, length - 1] */ + icoord = lp_build_min(&abs_coord_bld, icoord, length_minus_one); + } break; case PIPE_TEX_WRAP_MIRROR_CLAMP_TO_BORDER: @@ -641,21 +820,53 @@ lp_build_sample_wrap_nearest(struct lp_build_sample_context *bld, } +/** + * Do shadow test/comparison. + * \param p shadow ref value + * \param texel the texel to compare against + */ +static LLVMValueRef +lp_build_sample_comparefunc(struct lp_build_sample_context *bld, + LLVMValueRef p, + LLVMValueRef texel) +{ + struct lp_build_context *texel_bld = &bld->texel_bld; + LLVMValueRef res; + + if (0) { + //lp_build_print_value(bld->gallivm, "shadow cmp coord", p); + lp_build_print_value(bld->gallivm, "shadow cmp texel", texel); + } + + /* result = (p FUNC texel) ? 1 : 0 */ + /* + * honor d3d10 floating point rules here, which state that comparisons + * are ordered except NOT_EQUAL which is unordered. + */ + if (bld->static_sampler_state->compare_func != PIPE_FUNC_NOTEQUAL) { + res = lp_build_cmp_ordered(texel_bld, bld->static_sampler_state->compare_func, + p, texel); + } + else { + res = lp_build_cmp(texel_bld, bld->static_sampler_state->compare_func, + p, texel); + } + return res; +} + + /** * Generate code to sample a mipmap level with nearest filtering. * If sampling a cube texture, r = cube face in [0,5]. */ static void lp_build_sample_image_nearest(struct lp_build_sample_context *bld, - unsigned sampler_unit, LLVMValueRef size, LLVMValueRef row_stride_vec, LLVMValueRef img_stride_vec, LLVMValueRef data_ptr, LLVMValueRef mipoffsets, - LLVMValueRef s, - LLVMValueRef t, - LLVMValueRef r, + const LLVMValueRef *coords, const LLVMValueRef *offsets, LLVMValueRef colors_out[4]) { @@ -686,60 +897,130 @@ lp_build_sample_image_nearest(struct lp_build_sample_context *bld, /* * Compute integer texcoords. */ - x = lp_build_sample_wrap_nearest(bld, s, width_vec, flt_width_vec, offsets[0], + x = lp_build_sample_wrap_nearest(bld, coords[0], width_vec, + flt_width_vec, offsets[0], bld->static_texture_state->pot_width, bld->static_sampler_state->wrap_s); lp_build_name(x, "tex.x.wrapped"); if (dims >= 2) { - y = lp_build_sample_wrap_nearest(bld, t, height_vec, flt_height_vec, offsets[1], + y = lp_build_sample_wrap_nearest(bld, coords[1], height_vec, + flt_height_vec, offsets[1], bld->static_texture_state->pot_height, bld->static_sampler_state->wrap_t); lp_build_name(y, "tex.y.wrapped"); if (dims == 3) { - z = lp_build_sample_wrap_nearest(bld, r, depth_vec, flt_depth_vec, offsets[2], + z = lp_build_sample_wrap_nearest(bld, coords[2], depth_vec, + flt_depth_vec, offsets[2], bld->static_texture_state->pot_depth, bld->static_sampler_state->wrap_r); lp_build_name(z, "tex.z.wrapped"); } } - if (bld->static_texture_state->target == PIPE_TEXTURE_CUBE || - bld->static_texture_state->target == PIPE_TEXTURE_1D_ARRAY || - bld->static_texture_state->target == PIPE_TEXTURE_2D_ARRAY) { - z = r; + if (has_layer_coord(bld->static_texture_state->target)) { + if (bld->static_texture_state->target == PIPE_TEXTURE_CUBE_ARRAY) { + /* add cube layer to face */ + z = lp_build_add(&bld->int_coord_bld, coords[2], coords[3]); + } + else { + z = coords[2]; + } lp_build_name(z, "tex.z.layer"); } /* * Get texture colors. */ - lp_build_sample_texel_soa(bld, sampler_unit, + lp_build_sample_texel_soa(bld, width_vec, height_vec, depth_vec, x, y, z, row_stride_vec, img_stride_vec, data_ptr, mipoffsets, colors_out); + + if (bld->static_sampler_state->compare_mode != PIPE_TEX_COMPARE_NONE) { + LLVMValueRef cmpval; + cmpval = lp_build_sample_comparefunc(bld, coords[4], colors_out[0]); + /* this is really just a AND 1.0, cmpval but llvm is clever enough */ + colors_out[0] = lp_build_select(&bld->texel_bld, cmpval, + bld->texel_bld.one, bld->texel_bld.zero); + colors_out[1] = colors_out[2] = colors_out[3] = colors_out[0]; + } + +} + + +/** + * Like a lerp, but inputs are 0/~0 masks, so can simplify slightly. + */ +static LLVMValueRef +lp_build_masklerp(struct lp_build_context *bld, + LLVMValueRef weight, + LLVMValueRef mask0, + LLVMValueRef mask1) +{ + struct gallivm_state *gallivm = bld->gallivm; + LLVMBuilderRef builder = gallivm->builder; + LLVMValueRef weight2; + + weight2 = lp_build_sub(bld, bld->one, weight); + weight = LLVMBuildBitCast(builder, weight, + lp_build_int_vec_type(gallivm, bld->type), ""); + weight2 = LLVMBuildBitCast(builder, weight2, + lp_build_int_vec_type(gallivm, bld->type), ""); + weight = LLVMBuildAnd(builder, weight, mask1, ""); + weight2 = LLVMBuildAnd(builder, weight2, mask0, ""); + weight = LLVMBuildBitCast(builder, weight, bld->vec_type, ""); + weight2 = LLVMBuildBitCast(builder, weight2, bld->vec_type, ""); + return lp_build_add(bld, weight, weight2); +} + +/** + * Like a 2d lerp, but inputs are 0/~0 masks, so can simplify slightly. + */ +static LLVMValueRef +lp_build_masklerp2d(struct lp_build_context *bld, + LLVMValueRef weight0, + LLVMValueRef weight1, + LLVMValueRef mask00, + LLVMValueRef mask01, + LLVMValueRef mask10, + LLVMValueRef mask11) +{ + LLVMValueRef val0 = lp_build_masklerp(bld, weight0, mask00, mask01); + LLVMValueRef val1 = lp_build_masklerp(bld, weight0, mask10, mask11); + return lp_build_lerp(bld, weight1, val0, val1, 0); } +/* + * this is a bit excessive code for something OpenGL just recommends + * but does not require. + */ +#define ACCURATE_CUBE_CORNERS 1 /** * Generate code to sample a mipmap level with linear filtering. * If sampling a cube texture, r = cube face in [0,5]. + * If linear_mask is present, only pixels having their mask set + * will receive linear filtering, the rest will use nearest. */ static void lp_build_sample_image_linear(struct lp_build_sample_context *bld, - unsigned sampler_unit, + boolean is_gather, LLVMValueRef size, + LLVMValueRef linear_mask, LLVMValueRef row_stride_vec, LLVMValueRef img_stride_vec, LLVMValueRef data_ptr, LLVMValueRef mipoffsets, - LLVMValueRef s, - LLVMValueRef t, - LLVMValueRef r, + const LLVMValueRef *coords, const LLVMValueRef *offsets, LLVMValueRef colors_out[4]) { + LLVMBuilderRef builder = bld->gallivm->builder; + struct lp_build_context *ivec_bld = &bld->int_coord_bld; + struct lp_build_context *coord_bld = &bld->coord_bld; + struct lp_build_context *texel_bld = &bld->texel_bld; const unsigned dims = bld->dims; LLVMValueRef width_vec; LLVMValueRef height_vec; @@ -748,10 +1029,39 @@ lp_build_sample_image_linear(struct lp_build_sample_context *bld, LLVMValueRef flt_width_vec; LLVMValueRef flt_height_vec; LLVMValueRef flt_depth_vec; - LLVMValueRef x0, y0 = NULL, z0 = NULL, x1, y1 = NULL, z1 = NULL; + LLVMValueRef fall_off[4], have_corners; + LLVMValueRef z1 = NULL; + LLVMValueRef z00 = NULL, z01 = NULL, z10 = NULL, z11 = NULL; + LLVMValueRef x00 = NULL, x01 = NULL, x10 = NULL, x11 = NULL; + LLVMValueRef y00 = NULL, y01 = NULL, y10 = NULL, y11 = NULL; LLVMValueRef s_fpart, t_fpart = NULL, r_fpart = NULL; + LLVMValueRef xs[4], ys[4], zs[4]; LLVMValueRef neighbors[2][2][4]; - int chan; + int chan, texel_index; + boolean seamless_cube_filter, accurate_cube_corners; + unsigned chan_swiz = bld->static_texture_state->swizzle_r; + + if (is_gather) { + switch (bld->gather_comp) { + case 0: chan_swiz = bld->static_texture_state->swizzle_r; break; + case 1: chan_swiz = bld->static_texture_state->swizzle_g; break; + case 2: chan_swiz = bld->static_texture_state->swizzle_b; break; + case 3: chan_swiz = bld->static_texture_state->swizzle_a; break; + default: + break; + } + } + + seamless_cube_filter = (bld->static_texture_state->target == PIPE_TEXTURE_CUBE || + bld->static_texture_state->target == PIPE_TEXTURE_CUBE_ARRAY) && + bld->static_sampler_state->seamless_cube_map; + + /* + * Disable accurate cube corners for integer textures, which should only + * get here in the gather path. + */ + accurate_cube_corners = ACCURATE_CUBE_CORNERS && seamless_cube_filter && + !util_format_is_pure_integer(bld->static_texture_state->format); lp_build_extract_image_sizes(bld, &bld->int_size_bld, @@ -770,133 +1080,655 @@ lp_build_sample_image_linear(struct lp_build_sample_context *bld, /* * Compute integer texcoords. */ - lp_build_sample_wrap_linear(bld, s, width_vec, flt_width_vec, offsets[0], - bld->static_texture_state->pot_width, - bld->static_sampler_state->wrap_s, - &x0, &x1, &s_fpart); - lp_build_name(x0, "tex.x0.wrapped"); - lp_build_name(x1, "tex.x1.wrapped"); - if (dims >= 2) { - lp_build_sample_wrap_linear(bld, t, height_vec, flt_height_vec, offsets[1], - bld->static_texture_state->pot_height, - bld->static_sampler_state->wrap_t, - &y0, &y1, &t_fpart); - lp_build_name(y0, "tex.y0.wrapped"); - lp_build_name(y1, "tex.y1.wrapped"); + if (!seamless_cube_filter) { + lp_build_sample_wrap_linear(bld, is_gather, coords[0], width_vec, + flt_width_vec, offsets[0], + bld->static_texture_state->pot_width, + bld->static_sampler_state->wrap_s, + &x00, &x01, &s_fpart); + lp_build_name(x00, "tex.x0.wrapped"); + lp_build_name(x01, "tex.x1.wrapped"); + x10 = x00; + x11 = x01; - if (dims == 3) { - lp_build_sample_wrap_linear(bld, r, depth_vec, flt_depth_vec, offsets[2], - bld->static_texture_state->pot_depth, - bld->static_sampler_state->wrap_r, - &z0, &z1, &r_fpart); - lp_build_name(z0, "tex.z0.wrapped"); - lp_build_name(z1, "tex.z1.wrapped"); + if (dims >= 2) { + lp_build_sample_wrap_linear(bld, is_gather, coords[1], height_vec, + flt_height_vec, offsets[1], + bld->static_texture_state->pot_height, + bld->static_sampler_state->wrap_t, + &y00, &y10, &t_fpart); + lp_build_name(y00, "tex.y0.wrapped"); + lp_build_name(y10, "tex.y1.wrapped"); + y01 = y00; + y11 = y10; + + if (dims == 3) { + lp_build_sample_wrap_linear(bld, is_gather, coords[2], depth_vec, + flt_depth_vec, offsets[2], + bld->static_texture_state->pot_depth, + bld->static_sampler_state->wrap_r, + &z00, &z1, &r_fpart); + z01 = z10 = z11 = z00; + lp_build_name(z00, "tex.z0.wrapped"); + lp_build_name(z1, "tex.z1.wrapped"); + } + } + if (has_layer_coord(bld->static_texture_state->target)) { + if (bld->static_texture_state->target == PIPE_TEXTURE_CUBE_ARRAY) { + /* add cube layer to face */ + z00 = z01 = z10 = z11 = z1 = + lp_build_add(&bld->int_coord_bld, coords[2], coords[3]); + } + else { + z00 = z01 = z10 = z11 = z1 = coords[2]; /* cube face or layer */ + } + lp_build_name(z00, "tex.z0.layer"); + lp_build_name(z1, "tex.z1.layer"); } } - if (bld->static_texture_state->target == PIPE_TEXTURE_CUBE || - bld->static_texture_state->target == PIPE_TEXTURE_1D_ARRAY || - bld->static_texture_state->target == PIPE_TEXTURE_2D_ARRAY) { - z0 = z1 = r; /* cube face or array layer */ - lp_build_name(z0, "tex.z0.layer"); - lp_build_name(z1, "tex.z1.layer"); + else { + struct lp_build_if_state edge_if; + LLVMTypeRef int1t; + LLVMValueRef new_faces[4], new_xcoords[4][2], new_ycoords[4][2]; + LLVMValueRef coord0, coord1, have_edge, have_corner; + LLVMValueRef fall_off_ym_notxm, fall_off_ym_notxp, fall_off_x, fall_off_y; + LLVMValueRef fall_off_yp_notxm, fall_off_yp_notxp; + LLVMValueRef x0, x1, y0, y1, y0_clamped, y1_clamped; + LLVMValueRef face = coords[2]; + LLVMValueRef half = lp_build_const_vec(bld->gallivm, coord_bld->type, 0.5f); + LLVMValueRef length_minus_one = lp_build_sub(ivec_bld, width_vec, ivec_bld->one); + /* XXX drop height calcs. Could (should) do this without seamless filtering too */ + height_vec = width_vec; + flt_height_vec = flt_width_vec; + + /* XXX the overflow logic is actually sort of duplicated with trilinear, + * since an overflow in one mip should also have a corresponding overflow + * in another. + */ + /* should always have normalized coords, and offsets are undefined */ + assert(bld->static_sampler_state->normalized_coords); + /* + * The coords should all be between [0,1] however we can have NaNs, + * which will wreak havoc. In particular the y1_clamped value below + * can be -INT_MAX (on x86) and be propagated right through (probably + * other values might be bogus in the end too). + * So kill off the NaNs here. + */ + coord0 = lp_build_max_ext(coord_bld, coords[0], coord_bld->zero, + GALLIVM_NAN_RETURN_OTHER_SECOND_NONNAN); + coord0 = lp_build_mul(coord_bld, coord0, flt_width_vec); + /* instead of clamp, build mask if overflowed */ + coord0 = lp_build_sub(coord_bld, coord0, half); + /* convert to int, compute lerp weight */ + /* not ideal with AVX (and no AVX2) */ + lp_build_ifloor_fract(coord_bld, coord0, &x0, &s_fpart); + x1 = lp_build_add(ivec_bld, x0, ivec_bld->one); + coord1 = lp_build_max_ext(coord_bld, coords[1], coord_bld->zero, + GALLIVM_NAN_RETURN_OTHER_SECOND_NONNAN); + coord1 = lp_build_mul(coord_bld, coord1, flt_height_vec); + coord1 = lp_build_sub(coord_bld, coord1, half); + lp_build_ifloor_fract(coord_bld, coord1, &y0, &t_fpart); + y1 = lp_build_add(ivec_bld, y0, ivec_bld->one); + + fall_off[0] = lp_build_cmp(ivec_bld, PIPE_FUNC_LESS, x0, ivec_bld->zero); + fall_off[1] = lp_build_cmp(ivec_bld, PIPE_FUNC_GREATER, x1, length_minus_one); + fall_off[2] = lp_build_cmp(ivec_bld, PIPE_FUNC_LESS, y0, ivec_bld->zero); + fall_off[3] = lp_build_cmp(ivec_bld, PIPE_FUNC_GREATER, y1, length_minus_one); + + fall_off_x = lp_build_or(ivec_bld, fall_off[0], fall_off[1]); + fall_off_y = lp_build_or(ivec_bld, fall_off[2], fall_off[3]); + have_edge = lp_build_or(ivec_bld, fall_off_x, fall_off_y); + have_edge = lp_build_any_true_range(ivec_bld, ivec_bld->type.length, have_edge); + + /* needed for accurate corner filtering branch later, rely on 0 init */ + int1t = LLVMInt1TypeInContext(bld->gallivm->context); + have_corners = lp_build_alloca(bld->gallivm, int1t, "have_corner"); + + for (texel_index = 0; texel_index < 4; texel_index++) { + xs[texel_index] = lp_build_alloca(bld->gallivm, ivec_bld->vec_type, "xs"); + ys[texel_index] = lp_build_alloca(bld->gallivm, ivec_bld->vec_type, "ys"); + zs[texel_index] = lp_build_alloca(bld->gallivm, ivec_bld->vec_type, "zs"); + } + + lp_build_if(&edge_if, bld->gallivm, have_edge); + + have_corner = lp_build_and(ivec_bld, fall_off_x, fall_off_y); + have_corner = lp_build_any_true_range(ivec_bld, ivec_bld->type.length, have_corner); + LLVMBuildStore(builder, have_corner, have_corners); + + /* + * Need to feed clamped values here for cheap corner handling, + * but only for y coord (as when falling off both edges we only + * fall off the x one) - this should be sufficient. + */ + y0_clamped = lp_build_max(ivec_bld, y0, ivec_bld->zero); + y1_clamped = lp_build_min(ivec_bld, y1, length_minus_one); + + /* + * Get all possible new coords. + */ + lp_build_cube_new_coords(ivec_bld, face, + x0, x1, y0_clamped, y1_clamped, + length_minus_one, + new_faces, new_xcoords, new_ycoords); + + /* handle fall off x-, x+ direction */ + /* determine new coords, face (not both fall_off vars can be true at same time) */ + x00 = lp_build_select(ivec_bld, fall_off[0], new_xcoords[0][0], x0); + y00 = lp_build_select(ivec_bld, fall_off[0], new_ycoords[0][0], y0_clamped); + x10 = lp_build_select(ivec_bld, fall_off[0], new_xcoords[0][1], x0); + y10 = lp_build_select(ivec_bld, fall_off[0], new_ycoords[0][1], y1_clamped); + x01 = lp_build_select(ivec_bld, fall_off[1], new_xcoords[1][0], x1); + y01 = lp_build_select(ivec_bld, fall_off[1], new_ycoords[1][0], y0_clamped); + x11 = lp_build_select(ivec_bld, fall_off[1], new_xcoords[1][1], x1); + y11 = lp_build_select(ivec_bld, fall_off[1], new_ycoords[1][1], y1_clamped); + + z00 = z10 = lp_build_select(ivec_bld, fall_off[0], new_faces[0], face); + z01 = z11 = lp_build_select(ivec_bld, fall_off[1], new_faces[1], face); + + /* handle fall off y-, y+ direction */ + /* + * Cheap corner logic: just hack up things so a texel doesn't fall + * off both sides (which means filter weights will be wrong but we'll only + * use valid texels in the filter). + * This means however (y) coords must additionally be clamped (see above). + * This corner handling should be fully OpenGL (but not d3d10) compliant. + */ + fall_off_ym_notxm = lp_build_andnot(ivec_bld, fall_off[2], fall_off[0]); + fall_off_ym_notxp = lp_build_andnot(ivec_bld, fall_off[2], fall_off[1]); + fall_off_yp_notxm = lp_build_andnot(ivec_bld, fall_off[3], fall_off[0]); + fall_off_yp_notxp = lp_build_andnot(ivec_bld, fall_off[3], fall_off[1]); + + x00 = lp_build_select(ivec_bld, fall_off_ym_notxm, new_xcoords[2][0], x00); + y00 = lp_build_select(ivec_bld, fall_off_ym_notxm, new_ycoords[2][0], y00); + x01 = lp_build_select(ivec_bld, fall_off_ym_notxp, new_xcoords[2][1], x01); + y01 = lp_build_select(ivec_bld, fall_off_ym_notxp, new_ycoords[2][1], y01); + x10 = lp_build_select(ivec_bld, fall_off_yp_notxm, new_xcoords[3][0], x10); + y10 = lp_build_select(ivec_bld, fall_off_yp_notxm, new_ycoords[3][0], y10); + x11 = lp_build_select(ivec_bld, fall_off_yp_notxp, new_xcoords[3][1], x11); + y11 = lp_build_select(ivec_bld, fall_off_yp_notxp, new_ycoords[3][1], y11); + + z00 = lp_build_select(ivec_bld, fall_off_ym_notxm, new_faces[2], z00); + z01 = lp_build_select(ivec_bld, fall_off_ym_notxp, new_faces[2], z01); + z10 = lp_build_select(ivec_bld, fall_off_yp_notxm, new_faces[3], z10); + z11 = lp_build_select(ivec_bld, fall_off_yp_notxp, new_faces[3], z11); + + if (bld->static_texture_state->target == PIPE_TEXTURE_CUBE_ARRAY) { + /* now can add cube layer to face (per sample) */ + z00 = lp_build_add(ivec_bld, z00, coords[3]); + z01 = lp_build_add(ivec_bld, z01, coords[3]); + z10 = lp_build_add(ivec_bld, z10, coords[3]); + z11 = lp_build_add(ivec_bld, z11, coords[3]); + } + + LLVMBuildStore(builder, x00, xs[0]); + LLVMBuildStore(builder, x01, xs[1]); + LLVMBuildStore(builder, x10, xs[2]); + LLVMBuildStore(builder, x11, xs[3]); + LLVMBuildStore(builder, y00, ys[0]); + LLVMBuildStore(builder, y01, ys[1]); + LLVMBuildStore(builder, y10, ys[2]); + LLVMBuildStore(builder, y11, ys[3]); + LLVMBuildStore(builder, z00, zs[0]); + LLVMBuildStore(builder, z01, zs[1]); + LLVMBuildStore(builder, z10, zs[2]); + LLVMBuildStore(builder, z11, zs[3]); + + lp_build_else(&edge_if); + + LLVMBuildStore(builder, x0, xs[0]); + LLVMBuildStore(builder, x1, xs[1]); + LLVMBuildStore(builder, x0, xs[2]); + LLVMBuildStore(builder, x1, xs[3]); + LLVMBuildStore(builder, y0, ys[0]); + LLVMBuildStore(builder, y0, ys[1]); + LLVMBuildStore(builder, y1, ys[2]); + LLVMBuildStore(builder, y1, ys[3]); + if (bld->static_texture_state->target == PIPE_TEXTURE_CUBE_ARRAY) { + LLVMValueRef cube_layer = lp_build_add(ivec_bld, face, coords[3]); + LLVMBuildStore(builder, cube_layer, zs[0]); + LLVMBuildStore(builder, cube_layer, zs[1]); + LLVMBuildStore(builder, cube_layer, zs[2]); + LLVMBuildStore(builder, cube_layer, zs[3]); + } + else { + LLVMBuildStore(builder, face, zs[0]); + LLVMBuildStore(builder, face, zs[1]); + LLVMBuildStore(builder, face, zs[2]); + LLVMBuildStore(builder, face, zs[3]); + } + + lp_build_endif(&edge_if); + + x00 = LLVMBuildLoad(builder, xs[0], ""); + x01 = LLVMBuildLoad(builder, xs[1], ""); + x10 = LLVMBuildLoad(builder, xs[2], ""); + x11 = LLVMBuildLoad(builder, xs[3], ""); + y00 = LLVMBuildLoad(builder, ys[0], ""); + y01 = LLVMBuildLoad(builder, ys[1], ""); + y10 = LLVMBuildLoad(builder, ys[2], ""); + y11 = LLVMBuildLoad(builder, ys[3], ""); + z00 = LLVMBuildLoad(builder, zs[0], ""); + z01 = LLVMBuildLoad(builder, zs[1], ""); + z10 = LLVMBuildLoad(builder, zs[2], ""); + z11 = LLVMBuildLoad(builder, zs[3], ""); } + if (linear_mask) { + /* + * Whack filter weights into place. Whatever texel had more weight is + * the one which should have been selected by nearest filtering hence + * just use 100% weight for it. + */ + struct lp_build_context *c_bld = &bld->coord_bld; + LLVMValueRef w1_mask, w1_weight; + LLVMValueRef half = lp_build_const_vec(bld->gallivm, c_bld->type, 0.5f); + + w1_mask = lp_build_cmp(c_bld, PIPE_FUNC_GREATER, s_fpart, half); + /* this select is really just a "and" */ + w1_weight = lp_build_select(c_bld, w1_mask, c_bld->one, c_bld->zero); + s_fpart = lp_build_select(c_bld, linear_mask, s_fpart, w1_weight); + if (dims >= 2) { + w1_mask = lp_build_cmp(c_bld, PIPE_FUNC_GREATER, t_fpart, half); + w1_weight = lp_build_select(c_bld, w1_mask, c_bld->one, c_bld->zero); + t_fpart = lp_build_select(c_bld, linear_mask, t_fpart, w1_weight); + if (dims == 3) { + w1_mask = lp_build_cmp(c_bld, PIPE_FUNC_GREATER, r_fpart, half); + w1_weight = lp_build_select(c_bld, w1_mask, c_bld->one, c_bld->zero); + r_fpart = lp_build_select(c_bld, linear_mask, r_fpart, w1_weight); + } + } + } /* * Get texture colors. */ /* get x0/x1 texels */ - lp_build_sample_texel_soa(bld, sampler_unit, + lp_build_sample_texel_soa(bld, width_vec, height_vec, depth_vec, - x0, y0, z0, + x00, y00, z00, row_stride_vec, img_stride_vec, data_ptr, mipoffsets, neighbors[0][0]); - lp_build_sample_texel_soa(bld, sampler_unit, + lp_build_sample_texel_soa(bld, width_vec, height_vec, depth_vec, - x1, y0, z0, + x01, y01, z01, row_stride_vec, img_stride_vec, data_ptr, mipoffsets, neighbors[0][1]); if (dims == 1) { - /* Interpolate two samples from 1D image to produce one color */ - for (chan = 0; chan < 4; chan++) { - colors_out[chan] = lp_build_lerp(&bld->texel_bld, s_fpart, - neighbors[0][0][chan], - neighbors[0][1][chan], - 0); + assert(!is_gather); + if (bld->static_sampler_state->compare_mode == PIPE_TEX_COMPARE_NONE) { + /* Interpolate two samples from 1D image to produce one color */ + for (chan = 0; chan < 4; chan++) { + colors_out[chan] = lp_build_lerp(texel_bld, s_fpart, + neighbors[0][0][chan], + neighbors[0][1][chan], + 0); + } + } + else { + LLVMValueRef cmpval0, cmpval1; + cmpval0 = lp_build_sample_comparefunc(bld, coords[4], neighbors[0][0][0]); + cmpval1 = lp_build_sample_comparefunc(bld, coords[4], neighbors[0][1][0]); + /* simplified lerp, AND mask with weight and add */ + colors_out[0] = lp_build_masklerp(texel_bld, s_fpart, + cmpval0, cmpval1); + colors_out[1] = colors_out[2] = colors_out[3] = colors_out[0]; } } else { /* 2D/3D texture */ - LLVMValueRef colors0[4]; + struct lp_build_if_state corner_if; + LLVMValueRef colors0[4], colorss[4]; /* get x0/x1 texels at y1 */ - lp_build_sample_texel_soa(bld, sampler_unit, + lp_build_sample_texel_soa(bld, width_vec, height_vec, depth_vec, - x0, y1, z0, + x10, y10, z10, row_stride_vec, img_stride_vec, data_ptr, mipoffsets, neighbors[1][0]); - lp_build_sample_texel_soa(bld, sampler_unit, + lp_build_sample_texel_soa(bld, width_vec, height_vec, depth_vec, - x1, y1, z0, + x11, y11, z11, row_stride_vec, img_stride_vec, data_ptr, mipoffsets, neighbors[1][1]); - /* Bilinear interpolate the four samples from the 2D image / 3D slice */ - for (chan = 0; chan < 4; chan++) { - colors0[chan] = lp_build_lerp_2d(&bld->texel_bld, - s_fpart, t_fpart, - neighbors[0][0][chan], - neighbors[0][1][chan], - neighbors[1][0][chan], - neighbors[1][1][chan], - 0); + /* + * To avoid having to duplicate linear_mask / fetch code use + * another branch (with corner condition though edge would work + * as well) here. + */ + if (accurate_cube_corners) { + LLVMValueRef c00, c01, c10, c11, c00f, c01f, c10f, c11f; + LLVMValueRef have_corner, one_third; + + colorss[0] = lp_build_alloca(bld->gallivm, coord_bld->vec_type, "cs0"); + colorss[1] = lp_build_alloca(bld->gallivm, coord_bld->vec_type, "cs1"); + colorss[2] = lp_build_alloca(bld->gallivm, coord_bld->vec_type, "cs2"); + colorss[3] = lp_build_alloca(bld->gallivm, coord_bld->vec_type, "cs3"); + + have_corner = LLVMBuildLoad(builder, have_corners, ""); + + lp_build_if(&corner_if, bld->gallivm, have_corner); + + one_third = lp_build_const_vec(bld->gallivm, coord_bld->type, + 1.0f/3.0f); + + /* find corner */ + c00 = lp_build_and(ivec_bld, fall_off[0], fall_off[2]); + c00f = LLVMBuildBitCast(builder, c00, coord_bld->vec_type, ""); + c01 = lp_build_and(ivec_bld, fall_off[1], fall_off[2]); + c01f = LLVMBuildBitCast(builder, c01, coord_bld->vec_type, ""); + c10 = lp_build_and(ivec_bld, fall_off[0], fall_off[3]); + c10f = LLVMBuildBitCast(builder, c10, coord_bld->vec_type, ""); + c11 = lp_build_and(ivec_bld, fall_off[1], fall_off[3]); + c11f = LLVMBuildBitCast(builder, c11, coord_bld->vec_type, ""); + + if (!is_gather) { + /* + * we can't use standard 2d lerp as we need per-element weight + * in case of corners, so just calculate bilinear result as + * w00*s00 + w01*s01 + w10*s10 + w11*s11. + * (This is actually less work than using 2d lerp, 7 vs. 9 + * instructions, however calculating the weights needs another 6, + * so actually probably not slower than 2d lerp only for 4 channels + * as weights only need to be calculated once - of course fixing + * the weights has additional cost.) + */ + LLVMValueRef w00, w01, w10, w11, wx0, wy0, c_weight, tmp; + wx0 = lp_build_sub(coord_bld, coord_bld->one, s_fpart); + wy0 = lp_build_sub(coord_bld, coord_bld->one, t_fpart); + w00 = lp_build_mul(coord_bld, wx0, wy0); + w01 = lp_build_mul(coord_bld, s_fpart, wy0); + w10 = lp_build_mul(coord_bld, wx0, t_fpart); + w11 = lp_build_mul(coord_bld, s_fpart, t_fpart); + + /* find corner weight */ + c_weight = lp_build_select(coord_bld, c00, w00, coord_bld->zero); + c_weight = lp_build_select(coord_bld, c01, w01, c_weight); + c_weight = lp_build_select(coord_bld, c10, w10, c_weight); + c_weight = lp_build_select(coord_bld, c11, w11, c_weight); + + /* + * add 1/3 of the corner weight to the weight of the 3 other + * samples and null out corner weight. + */ + c_weight = lp_build_mul(coord_bld, c_weight, one_third); + w00 = lp_build_add(coord_bld, w00, c_weight); + w00 = lp_build_andnot(coord_bld, w00, c00f); + w01 = lp_build_add(coord_bld, w01, c_weight); + w01 = lp_build_andnot(coord_bld, w01, c01f); + w10 = lp_build_add(coord_bld, w10, c_weight); + w10 = lp_build_andnot(coord_bld, w10, c10f); + w11 = lp_build_add(coord_bld, w11, c_weight); + w11 = lp_build_andnot(coord_bld, w11, c11f); + + if (bld->static_sampler_state->compare_mode == + PIPE_TEX_COMPARE_NONE) { + for (chan = 0; chan < 4; chan++) { + colors0[chan] = lp_build_mul(coord_bld, w00, + neighbors[0][0][chan]); + tmp = lp_build_mul(coord_bld, w01, neighbors[0][1][chan]); + colors0[chan] = lp_build_add(coord_bld, tmp, colors0[chan]); + tmp = lp_build_mul(coord_bld, w10, neighbors[1][0][chan]); + colors0[chan] = lp_build_add(coord_bld, tmp, colors0[chan]); + tmp = lp_build_mul(coord_bld, w11, neighbors[1][1][chan]); + colors0[chan] = lp_build_add(coord_bld, tmp, colors0[chan]); + } + } + else { + LLVMValueRef cmpval00, cmpval01, cmpval10, cmpval11; + cmpval00 = lp_build_sample_comparefunc(bld, coords[4], + neighbors[0][0][0]); + cmpval01 = lp_build_sample_comparefunc(bld, coords[4], + neighbors[0][1][0]); + cmpval10 = lp_build_sample_comparefunc(bld, coords[4], + neighbors[1][0][0]); + cmpval11 = lp_build_sample_comparefunc(bld, coords[4], + neighbors[1][1][0]); + /* + * inputs to interpolation are just masks so just add + * masked weights together + */ + cmpval00 = LLVMBuildBitCast(builder, cmpval00, + coord_bld->vec_type, ""); + cmpval01 = LLVMBuildBitCast(builder, cmpval01, + coord_bld->vec_type, ""); + cmpval10 = LLVMBuildBitCast(builder, cmpval10, + coord_bld->vec_type, ""); + cmpval11 = LLVMBuildBitCast(builder, cmpval11, + coord_bld->vec_type, ""); + colors0[0] = lp_build_and(coord_bld, w00, cmpval00); + tmp = lp_build_and(coord_bld, w01, cmpval01); + colors0[0] = lp_build_add(coord_bld, tmp, colors0[0]); + tmp = lp_build_and(coord_bld, w10, cmpval10); + colors0[0] = lp_build_add(coord_bld, tmp, colors0[0]); + tmp = lp_build_and(coord_bld, w11, cmpval11); + colors0[0] = lp_build_add(coord_bld, tmp, colors0[0]); + colors0[1] = colors0[2] = colors0[3] = colors0[0]; + } + } + else { + /* + * We don't have any weights to adjust, so instead calculate + * the fourth texel as simply the average of the other 3. + * (This would work for non-gather too, however we'd have + * a boatload more of the select stuff due to there being + * 4 times as many colors as weights.) + */ + LLVMValueRef col00, col01, col10, col11; + LLVMValueRef colc, colc0, colc1; + col10 = lp_build_swizzle_soa_channel(texel_bld, + neighbors[1][0], chan_swiz); + col11 = lp_build_swizzle_soa_channel(texel_bld, + neighbors[1][1], chan_swiz); + col01 = lp_build_swizzle_soa_channel(texel_bld, + neighbors[0][1], chan_swiz); + col00 = lp_build_swizzle_soa_channel(texel_bld, + neighbors[0][0], chan_swiz); + + /* + * The spec says for comparison filtering, the comparison + * must happen before synthesizing the new value. + * This means all gathered values are always 0 or 1, + * except for the non-existing texel, which can be 0,1/3,2/3,1... + * Seems like we'd be allowed to just return 0 or 1 too, so we + * could simplify and pass down the compare mask values to the + * end (using int arithmetic/compare on the mask values to + * construct the fourth texel) and only there convert to floats + * but it's probably not worth it (it might be easier for the cpu + * but not for the code)... + */ + if (bld->static_sampler_state->compare_mode != + PIPE_TEX_COMPARE_NONE) { + LLVMValueRef cmpval00, cmpval01, cmpval10, cmpval11; + cmpval00 = lp_build_sample_comparefunc(bld, coords[4], col00); + cmpval01 = lp_build_sample_comparefunc(bld, coords[4], col01); + cmpval10 = lp_build_sample_comparefunc(bld, coords[4], col10); + cmpval11 = lp_build_sample_comparefunc(bld, coords[4], col11); + col00 = lp_build_select(texel_bld, cmpval00, + texel_bld->one, texel_bld->zero); + col01 = lp_build_select(texel_bld, cmpval01, + texel_bld->one, texel_bld->zero); + col10 = lp_build_select(texel_bld, cmpval10, + texel_bld->one, texel_bld->zero); + col11 = lp_build_select(texel_bld, cmpval11, + texel_bld->one, texel_bld->zero); + } + + /* + * Null out corner color. + */ + col00 = lp_build_andnot(coord_bld, col00, c00f); + col01 = lp_build_andnot(coord_bld, col01, c01f); + col10 = lp_build_andnot(coord_bld, col10, c10f); + col11 = lp_build_andnot(coord_bld, col11, c11f); + + /* + * New corner texel color is all colors added / 3. + */ + colc0 = lp_build_add(coord_bld, col00, col01); + colc1 = lp_build_add(coord_bld, col10, col11); + colc = lp_build_add(coord_bld, colc0, colc1); + colc = lp_build_mul(coord_bld, one_third, colc); + + /* + * Replace the corner texel color with the new value. + */ + col00 = lp_build_select(coord_bld, c00, colc, col00); + col01 = lp_build_select(coord_bld, c01, colc, col01); + col10 = lp_build_select(coord_bld, c10, colc, col10); + col11 = lp_build_select(coord_bld, c11, colc, col11); + + colors0[0] = col10; + colors0[1] = col11; + colors0[2] = col01; + colors0[3] = col00; + } + + LLVMBuildStore(builder, colors0[0], colorss[0]); + LLVMBuildStore(builder, colors0[1], colorss[1]); + LLVMBuildStore(builder, colors0[2], colorss[2]); + LLVMBuildStore(builder, colors0[3], colorss[3]); + + lp_build_else(&corner_if); + } + + if (bld->static_sampler_state->compare_mode == PIPE_TEX_COMPARE_NONE) { + if (is_gather) { + /* + * Just assign the red channel (no component selection yet). + * This is a bit hackish, we usually do the swizzle at the + * end of sampling (much less values to swizzle), but this + * obviously cannot work when using gather. + */ + colors0[0] = lp_build_swizzle_soa_channel(texel_bld, + neighbors[1][0], + chan_swiz); + colors0[1] = lp_build_swizzle_soa_channel(texel_bld, + neighbors[1][1], + chan_swiz); + colors0[2] = lp_build_swizzle_soa_channel(texel_bld, + neighbors[0][1], + chan_swiz); + colors0[3] = lp_build_swizzle_soa_channel(texel_bld, + neighbors[0][0], + chan_swiz); + } + else { + /* Bilinear interpolate the four samples from the 2D image / 3D slice */ + for (chan = 0; chan < 4; chan++) { + colors0[chan] = lp_build_lerp_2d(texel_bld, + s_fpart, t_fpart, + neighbors[0][0][chan], + neighbors[0][1][chan], + neighbors[1][0][chan], + neighbors[1][1][chan], + 0); + } + } + } + else { + LLVMValueRef cmpval00, cmpval01, cmpval10, cmpval11; + cmpval00 = lp_build_sample_comparefunc(bld, coords[4], neighbors[0][0][0]); + cmpval01 = lp_build_sample_comparefunc(bld, coords[4], neighbors[0][1][0]); + cmpval10 = lp_build_sample_comparefunc(bld, coords[4], neighbors[1][0][0]); + cmpval11 = lp_build_sample_comparefunc(bld, coords[4], neighbors[1][1][0]); + + if (is_gather) { + /* more hacks for swizzling, should be X, ONE or ZERO... */ + colors0[0] = lp_build_select(texel_bld, cmpval10, + texel_bld->one, texel_bld->zero); + colors0[1] = lp_build_select(texel_bld, cmpval11, + texel_bld->one, texel_bld->zero); + colors0[2] = lp_build_select(texel_bld, cmpval01, + texel_bld->one, texel_bld->zero); + colors0[3] = lp_build_select(texel_bld, cmpval00, + texel_bld->one, texel_bld->zero); + } + else { + colors0[0] = lp_build_masklerp2d(texel_bld, s_fpart, t_fpart, + cmpval00, cmpval01, cmpval10, cmpval11); + colors0[1] = colors0[2] = colors0[3] = colors0[0]; + } + } + + if (accurate_cube_corners) { + LLVMBuildStore(builder, colors0[0], colorss[0]); + LLVMBuildStore(builder, colors0[1], colorss[1]); + LLVMBuildStore(builder, colors0[2], colorss[2]); + LLVMBuildStore(builder, colors0[3], colorss[3]); + + lp_build_endif(&corner_if); + + colors0[0] = LLVMBuildLoad(builder, colorss[0], ""); + colors0[1] = LLVMBuildLoad(builder, colorss[1], ""); + colors0[2] = LLVMBuildLoad(builder, colorss[2], ""); + colors0[3] = LLVMBuildLoad(builder, colorss[3], ""); } if (dims == 3) { LLVMValueRef neighbors1[2][2][4]; LLVMValueRef colors1[4]; + assert(!is_gather); + /* get x0/x1/y0/y1 texels at z1 */ - lp_build_sample_texel_soa(bld, sampler_unit, + lp_build_sample_texel_soa(bld, width_vec, height_vec, depth_vec, - x0, y0, z1, + x00, y00, z1, row_stride_vec, img_stride_vec, data_ptr, mipoffsets, neighbors1[0][0]); - lp_build_sample_texel_soa(bld, sampler_unit, + lp_build_sample_texel_soa(bld, width_vec, height_vec, depth_vec, - x1, y0, z1, + x01, y01, z1, row_stride_vec, img_stride_vec, data_ptr, mipoffsets, neighbors1[0][1]); - lp_build_sample_texel_soa(bld, sampler_unit, + lp_build_sample_texel_soa(bld, width_vec, height_vec, depth_vec, - x0, y1, z1, + x10, y10, z1, row_stride_vec, img_stride_vec, data_ptr, mipoffsets, neighbors1[1][0]); - lp_build_sample_texel_soa(bld, sampler_unit, + lp_build_sample_texel_soa(bld, width_vec, height_vec, depth_vec, - x1, y1, z1, + x11, y11, z1, row_stride_vec, img_stride_vec, data_ptr, mipoffsets, neighbors1[1][1]); - /* Bilinear interpolate the four samples from the second Z slice */ - for (chan = 0; chan < 4; chan++) { - colors1[chan] = lp_build_lerp_2d(&bld->texel_bld, - s_fpart, t_fpart, - neighbors1[0][0][chan], - neighbors1[0][1][chan], - neighbors1[1][0][chan], - neighbors1[1][1][chan], - 0); + if (bld->static_sampler_state->compare_mode == PIPE_TEX_COMPARE_NONE) { + /* Bilinear interpolate the four samples from the second Z slice */ + for (chan = 0; chan < 4; chan++) { + colors1[chan] = lp_build_lerp_2d(texel_bld, + s_fpart, t_fpart, + neighbors1[0][0][chan], + neighbors1[0][1][chan], + neighbors1[1][0][chan], + neighbors1[1][1][chan], + 0); + } + /* Linearly interpolate the two samples from the two 3D slices */ + for (chan = 0; chan < 4; chan++) { + colors_out[chan] = lp_build_lerp(texel_bld, + r_fpart, + colors0[chan], colors1[chan], + 0); + } } - - /* Linearly interpolate the two samples from the two 3D slices */ - for (chan = 0; chan < 4; chan++) { - colors_out[chan] = lp_build_lerp(&bld->texel_bld, - r_fpart, - colors0[chan], colors1[chan], - 0); + else { + LLVMValueRef cmpval00, cmpval01, cmpval10, cmpval11; + cmpval00 = lp_build_sample_comparefunc(bld, coords[4], neighbors[0][0][0]); + cmpval01 = lp_build_sample_comparefunc(bld, coords[4], neighbors[0][1][0]); + cmpval10 = lp_build_sample_comparefunc(bld, coords[4], neighbors[1][0][0]); + cmpval11 = lp_build_sample_comparefunc(bld, coords[4], neighbors[1][1][0]); + colors1[0] = lp_build_masklerp2d(texel_bld, s_fpart, t_fpart, + cmpval00, cmpval01, cmpval10, cmpval11); + /* Linearly interpolate the two samples from the two 3D slices */ + colors_out[0] = lp_build_lerp(texel_bld, + r_fpart, + colors0[0], colors1[0], + 0); + colors_out[1] = colors_out[2] = colors_out[3] = colors_out[0]; } } else { @@ -906,23 +1738,41 @@ lp_build_sample_image_linear(struct lp_build_sample_context *bld, } } } + if (is_gather) { + /* + * For gather, we can't do our usual channel swizzling done later, + * so do it here. It only really matters for 0/1 swizzles in case + * of comparison filtering, since in this case the results would be + * wrong, without comparison it should all work out alright but it + * can't hurt to do that here, since it will instantly drop all + * calculations above, though it's a rather stupid idea to do + * gather on a channel which will always return 0 or 1 in any case... + */ + if (chan_swiz == PIPE_SWIZZLE_1) { + for (chan = 0; chan < 4; chan++) { + colors_out[chan] = texel_bld->one; + } + } else if (chan_swiz == PIPE_SWIZZLE_0) { + for (chan = 0; chan < 4; chan++) { + colors_out[chan] = texel_bld->zero; + } + } + } } /** * Sample the texture/mipmap using given image filter and mip filter. - * data0_ptr and data1_ptr point to the two mipmap levels to sample - * from. width0/1_vec, height0/1_vec, depth0/1_vec indicate their sizes. + * ilevel0 and ilevel1 indicate the two mipmap levels to sample + * from (vectors or scalars). * If we're using nearest miplevel sampling the '1' values will be null/unused. */ static void lp_build_sample_mipmap(struct lp_build_sample_context *bld, - unsigned sampler_unit, unsigned img_filter, unsigned mip_filter, - LLVMValueRef s, - LLVMValueRef t, - LLVMValueRef r, + boolean is_gather, + const LLVMValueRef *coords, const LLVMValueRef *offsets, LLVMValueRef ilevel0, LLVMValueRef ilevel1, @@ -947,7 +1797,7 @@ lp_build_sample_mipmap(struct lp_build_sample_context *bld, lp_build_mipmap_level_sizes(bld, ilevel0, &size0, &row_stride0_vec, &img_stride0_vec); - if (bld->num_lods == 1) { + if (bld->num_mips == 1) { data_ptr0 = lp_build_get_mipmap_level(bld, ilevel0); } else { @@ -956,18 +1806,16 @@ lp_build_sample_mipmap(struct lp_build_sample_context *bld, mipoff0 = lp_build_get_mip_offsets(bld, ilevel0); } if (img_filter == PIPE_TEX_FILTER_NEAREST) { - lp_build_sample_image_nearest(bld, sampler_unit, - size0, + lp_build_sample_image_nearest(bld, size0, row_stride0_vec, img_stride0_vec, - data_ptr0, mipoff0, s, t, r, offsets, + data_ptr0, mipoff0, coords, offsets, colors0); } else { assert(img_filter == PIPE_TEX_FILTER_LINEAR); - lp_build_sample_image_linear(bld, sampler_unit, - size0, + lp_build_sample_image_linear(bld, is_gather, size0, NULL, row_stride0_vec, img_stride0_vec, - data_ptr0, mipoff0, s, t, r, offsets, + data_ptr0, mipoff0, coords, offsets, colors0); } @@ -983,7 +1831,7 @@ lp_build_sample_mipmap(struct lp_build_sample_context *bld, /* need_lerp = lod_fpart > 0 */ if (bld->num_lods == 1) { need_lerp = LLVMBuildFCmp(builder, LLVMRealUGT, - lod_fpart, bld->levelf_bld.zero, + lod_fpart, bld->lodf_bld.zero, "need_lerp"); } else { @@ -991,28 +1839,29 @@ lp_build_sample_mipmap(struct lp_build_sample_context *bld, * We'll do mip filtering if any of the quads (or individual * pixel in case of per-pixel lod) need it. * It might be better to split the vectors here and only fetch/filter - * quads which need it. - */ - /* - * We unfortunately need to clamp lod_fpart here since we can get - * negative values which would screw up filtering if not all - * lod_fpart values have same sign. + * quads which need it (if there's one lod per quad). */ - lod_fpart = lp_build_max(&bld->levelf_bld, lod_fpart, - bld->levelf_bld.zero); - need_lerp = lp_build_compare(bld->gallivm, bld->levelf_bld.type, + need_lerp = lp_build_compare(bld->gallivm, bld->lodf_bld.type, PIPE_FUNC_GREATER, - lod_fpart, bld->levelf_bld.zero); - need_lerp = lp_build_any_true_range(&bld->leveli_bld, bld->num_lods, need_lerp); + lod_fpart, bld->lodf_bld.zero); + need_lerp = lp_build_any_true_range(&bld->lodi_bld, bld->num_lods, need_lerp); + lp_build_name(need_lerp, "need_lerp"); } lp_build_if(&if_ctx, bld->gallivm, need_lerp); { + /* + * We unfortunately need to clamp lod_fpart here since we can get + * negative values which would screw up filtering if not all + * lod_fpart values have same sign. + */ + lod_fpart = lp_build_max(&bld->lodf_bld, lod_fpart, + bld->lodf_bld.zero); /* sample the second mipmap level */ lp_build_mipmap_level_sizes(bld, ilevel1, &size1, &row_stride1_vec, &img_stride1_vec); - if (bld->num_lods == 1) { + if (bld->num_mips == 1) { data_ptr1 = lp_build_get_mipmap_level(bld, ilevel1); } else { @@ -1020,17 +1869,15 @@ lp_build_sample_mipmap(struct lp_build_sample_context *bld, mipoff1 = lp_build_get_mip_offsets(bld, ilevel1); } if (img_filter == PIPE_TEX_FILTER_NEAREST) { - lp_build_sample_image_nearest(bld, sampler_unit, - size1, + lp_build_sample_image_nearest(bld, size1, row_stride1_vec, img_stride1_vec, - data_ptr1, mipoff1, s, t, r, offsets, + data_ptr1, mipoff1, coords, offsets, colors1); } else { - lp_build_sample_image_linear(bld, sampler_unit, - size1, + lp_build_sample_image_linear(bld, FALSE, size1, NULL, row_stride1_vec, img_stride1_vec, - data_ptr1, mipoff1, s, t, r, offsets, + data_ptr1, mipoff1, coords, offsets, colors1); } @@ -1038,7 +1885,7 @@ lp_build_sample_mipmap(struct lp_build_sample_context *bld, if (bld->num_lods != bld->coord_type.length) lod_fpart = lp_build_unpack_broadcast_aos_scalars(bld->gallivm, - bld->levelf_bld.type, + bld->lodf_bld.type, bld->texel_bld.type, lod_fpart); @@ -1055,24 +1902,144 @@ lp_build_sample_mipmap(struct lp_build_sample_context *bld, /** - * Build (per-coord) layer value. - * Either clamp layer to valid values or fill in optional out_of_bounds - * value and just return value unclamped. + * Sample the texture/mipmap using given mip filter, and using + * both nearest and linear filtering at the same time depending + * on linear_mask. + * lod can be per quad but linear_mask is always per pixel. + * ilevel0 and ilevel1 indicate the two mipmap levels to sample + * from (vectors or scalars). + * If we're using nearest miplevel sampling the '1' values will be null/unused. */ -static LLVMValueRef -lp_build_layer_coord(struct lp_build_sample_context *bld, - unsigned texture_unit, - LLVMValueRef layer, - LLVMValueRef *out_of_bounds) +static void +lp_build_sample_mipmap_both(struct lp_build_sample_context *bld, + LLVMValueRef linear_mask, + unsigned mip_filter, + const LLVMValueRef *coords, + const LLVMValueRef *offsets, + LLVMValueRef ilevel0, + LLVMValueRef ilevel1, + LLVMValueRef lod_fpart, + LLVMValueRef lod_positive, + LLVMValueRef *colors_out) { - LLVMValueRef num_layers; + LLVMBuilderRef builder = bld->gallivm->builder; + LLVMValueRef size0 = NULL; + LLVMValueRef size1 = NULL; + LLVMValueRef row_stride0_vec = NULL; + LLVMValueRef row_stride1_vec = NULL; + LLVMValueRef img_stride0_vec = NULL; + LLVMValueRef img_stride1_vec = NULL; + LLVMValueRef data_ptr0 = NULL; + LLVMValueRef data_ptr1 = NULL; + LLVMValueRef mipoff0 = NULL; + LLVMValueRef mipoff1 = NULL; + LLVMValueRef colors0[4], colors1[4]; + unsigned chan; + + /* sample the first mipmap level */ + lp_build_mipmap_level_sizes(bld, ilevel0, + &size0, + &row_stride0_vec, &img_stride0_vec); + if (bld->num_mips == 1) { + data_ptr0 = lp_build_get_mipmap_level(bld, ilevel0); + } + else { + /* This path should work for num_lods 1 too but slightly less efficient */ + data_ptr0 = bld->base_ptr; + mipoff0 = lp_build_get_mip_offsets(bld, ilevel0); + } + + lp_build_sample_image_linear(bld, FALSE, size0, linear_mask, + row_stride0_vec, img_stride0_vec, + data_ptr0, mipoff0, coords, offsets, + colors0); + + /* Store the first level's colors in the output variables */ + for (chan = 0; chan < 4; chan++) { + LLVMBuildStore(builder, colors0[chan], colors_out[chan]); + } + + if (mip_filter == PIPE_TEX_MIPFILTER_LINEAR) { + struct lp_build_if_state if_ctx; + LLVMValueRef need_lerp; + + /* + * We'll do mip filtering if any of the quads (or individual + * pixel in case of per-pixel lod) need it. + * Note using lod_positive here not lod_fpart since it may be the same + * condition as that used in the outer "if" in the caller hence llvm + * should be able to merge the branches in this case. + */ + need_lerp = lp_build_any_true_range(&bld->lodi_bld, bld->num_lods, lod_positive); + lp_build_name(need_lerp, "need_lerp"); + + lp_build_if(&if_ctx, bld->gallivm, need_lerp); + { + /* + * We unfortunately need to clamp lod_fpart here since we can get + * negative values which would screw up filtering if not all + * lod_fpart values have same sign. + */ + lod_fpart = lp_build_max(&bld->lodf_bld, lod_fpart, + bld->lodf_bld.zero); + /* sample the second mipmap level */ + lp_build_mipmap_level_sizes(bld, ilevel1, + &size1, + &row_stride1_vec, &img_stride1_vec); + if (bld->num_mips == 1) { + data_ptr1 = lp_build_get_mipmap_level(bld, ilevel1); + } + else { + data_ptr1 = bld->base_ptr; + mipoff1 = lp_build_get_mip_offsets(bld, ilevel1); + } + + lp_build_sample_image_linear(bld, FALSE, size1, linear_mask, + row_stride1_vec, img_stride1_vec, + data_ptr1, mipoff1, coords, offsets, + colors1); + + /* interpolate samples from the two mipmap levels */ + + if (bld->num_lods != bld->coord_type.length) + lod_fpart = lp_build_unpack_broadcast_aos_scalars(bld->gallivm, + bld->lodf_bld.type, + bld->texel_bld.type, + lod_fpart); + + for (chan = 0; chan < 4; chan++) { + colors0[chan] = lp_build_lerp(&bld->texel_bld, lod_fpart, + colors0[chan], colors1[chan], + 0); + LLVMBuildStore(builder, colors0[chan], colors_out[chan]); + } + } + lp_build_endif(&if_ctx); + } +} + + +/** + * Build (per-coord) layer value. + * Either clamp layer to valid values or fill in optional out_of_bounds + * value and just return value unclamped. + */ +static LLVMValueRef +lp_build_layer_coord(struct lp_build_sample_context *bld, + unsigned texture_unit, + boolean is_cube_array, + LLVMValueRef layer, + LLVMValueRef *out_of_bounds) +{ + LLVMValueRef num_layers; struct lp_build_context *int_coord_bld = &bld->int_coord_bld; - num_layers = bld->dynamic_state->depth(bld->dynamic_state, - bld->gallivm, texture_unit); + num_layers = bld->dynamic_state->depth(bld->dynamic_state, bld->gallivm, + bld->context_ptr, texture_unit, NULL); if (out_of_bounds) { LLVMValueRef out1, out; + assert(!is_cube_array); num_layers = lp_build_broadcast_scalar(int_coord_bld, num_layers); out = lp_build_cmp(int_coord_bld, PIPE_FUNC_LESS, layer, int_coord_bld->zero); out1 = lp_build_cmp(int_coord_bld, PIPE_FUNC_GEQUAL, layer, num_layers); @@ -1081,7 +2048,9 @@ lp_build_layer_coord(struct lp_build_sample_context *bld, } else { LLVMValueRef maxlayer; - maxlayer = lp_build_sub(&bld->int_bld, num_layers, bld->int_bld.one); + LLVMValueRef s = is_cube_array ? lp_build_const_int32(bld->gallivm, 6) : + bld->int_bld.one; + maxlayer = lp_build_sub(&bld->int_bld, num_layers, s); maxlayer = lp_build_broadcast_scalar(int_coord_bld, maxlayer); return lp_build_clamp(int_coord_bld, layer, int_coord_bld->zero, maxlayer); } @@ -1093,15 +2062,15 @@ lp_build_layer_coord(struct lp_build_sample_context *bld, */ static void lp_build_sample_common(struct lp_build_sample_context *bld, + boolean is_lodq, unsigned texture_index, unsigned sampler_index, - LLVMValueRef *s, - LLVMValueRef *t, - LLVMValueRef *r, + LLVMValueRef *coords, const struct lp_derivatives *derivs, /* optional */ LLVMValueRef lod_bias, /* optional */ LLVMValueRef explicit_lod, /* optional */ - LLVMValueRef *lod_ipart, + LLVMValueRef *lod_pos_or_zero, + LLVMValueRef *lod, LLVMValueRef *lod_fpart, LLVMValueRef *ilevel0, LLVMValueRef *ilevel1) @@ -1111,6 +2080,8 @@ lp_build_sample_common(struct lp_build_sample_context *bld, const unsigned mag_filter = bld->static_sampler_state->mag_img_filter; const unsigned target = bld->static_texture_state->target; LLVMValueRef first_level, cube_rho = NULL; + LLVMValueRef lod_ipart = NULL; + struct lp_derivatives cube_derivs; /* printf("%s mip %d min %d mag %d\n", __FUNCTION__, @@ -1121,44 +2092,104 @@ lp_build_sample_common(struct lp_build_sample_context *bld, * Choose cube face, recompute texcoords for the chosen face and * compute rho here too (as it requires transform of derivatives). */ - if (target == PIPE_TEXTURE_CUBE) { - LLVMValueRef face, face_s, face_t; + if (target == PIPE_TEXTURE_CUBE || target == PIPE_TEXTURE_CUBE_ARRAY) { boolean need_derivs; need_derivs = ((min_filter != mag_filter || mip_filter != PIPE_TEX_MIPFILTER_NONE) && !bld->static_sampler_state->min_max_lod_equal && !explicit_lod); - lp_build_cube_lookup(bld, *s, *t, *r, derivs, &face, &face_s, &face_t, - &cube_rho, need_derivs); - *s = face_s; /* vec */ - *t = face_t; /* vec */ - /* use 'r' to indicate cube face */ - *r = face; /* vec */ + lp_build_cube_lookup(bld, coords, derivs, &cube_rho, &cube_derivs, need_derivs); + derivs = &cube_derivs; + if (target == PIPE_TEXTURE_CUBE_ARRAY) { + /* calculate cube layer coord now */ + LLVMValueRef layer = lp_build_iround(&bld->coord_bld, coords[3]); + LLVMValueRef six = lp_build_const_int_vec(bld->gallivm, bld->int_coord_type, 6); + layer = lp_build_mul(&bld->int_coord_bld, layer, six); + coords[3] = lp_build_layer_coord(bld, texture_index, TRUE, layer, NULL); + /* because of seamless filtering can't add it to face (coords[2]) here. */ + } } - else if (target == PIPE_TEXTURE_1D_ARRAY) { - *r = lp_build_iround(&bld->coord_bld, *t); - *r = lp_build_layer_coord(bld, texture_index, *r, NULL); + else if (target == PIPE_TEXTURE_1D_ARRAY || + target == PIPE_TEXTURE_2D_ARRAY) { + coords[2] = lp_build_iround(&bld->coord_bld, coords[2]); + coords[2] = lp_build_layer_coord(bld, texture_index, FALSE, coords[2], NULL); } - else if (target == PIPE_TEXTURE_2D_ARRAY) { - *r = lp_build_iround(&bld->coord_bld, *r); - *r = lp_build_layer_coord(bld, texture_index, *r, NULL); + + if (bld->static_sampler_state->compare_mode != PIPE_TEX_COMPARE_NONE) { + /* + * Clamp p coords to [0,1] for fixed function depth texture format here. + * Technically this is not entirely correct for unorm depth as the ref value + * should be converted to the depth format (quantization!) and comparison + * then done in texture format. This would actually help performance (since + * only need to do it once and could save the per-sample conversion of texels + * to floats instead), but it would need more messy code (would need to push + * at least some bits down to actual fetch so conversion could be skipped, + * and would have ugly interaction with border color, would need to convert + * border color to that format too or do some other tricks to make it work). + */ + const struct util_format_description *format_desc = bld->format_desc; + unsigned chan_type; + /* not entirely sure we couldn't end up with non-valid swizzle here */ + chan_type = format_desc->swizzle[0] <= PIPE_SWIZZLE_W ? + format_desc->channel[format_desc->swizzle[0]].type : + UTIL_FORMAT_TYPE_FLOAT; + if (chan_type != UTIL_FORMAT_TYPE_FLOAT) { + coords[4] = lp_build_clamp(&bld->coord_bld, coords[4], + bld->coord_bld.zero, bld->coord_bld.one); + } } /* * Compute the level of detail (float). */ if (min_filter != mag_filter || - mip_filter != PIPE_TEX_MIPFILTER_NONE) { + mip_filter != PIPE_TEX_MIPFILTER_NONE || is_lodq) { /* Need to compute lod either to choose mipmap levels or to * distinguish between minification/magnification with one mipmap level. */ - lp_build_lod_selector(bld, texture_index, sampler_index, - *s, *t, *r, cube_rho, + lp_build_lod_selector(bld, is_lodq, texture_index, sampler_index, + coords[0], coords[1], coords[2], cube_rho, derivs, lod_bias, explicit_lod, - mip_filter, - lod_ipart, lod_fpart); + mip_filter, lod, + &lod_ipart, lod_fpart, lod_pos_or_zero); + if (is_lodq) { + LLVMValueRef last_level; + last_level = bld->dynamic_state->last_level(bld->dynamic_state, + bld->gallivm, + bld->context_ptr, + texture_index, NULL); + first_level = bld->dynamic_state->first_level(bld->dynamic_state, + bld->gallivm, + bld->context_ptr, + texture_index, NULL); + last_level = lp_build_sub(&bld->int_bld, last_level, first_level); + last_level = lp_build_int_to_float(&bld->float_bld, last_level); + last_level = lp_build_broadcast_scalar(&bld->lodf_bld, last_level); + + switch (mip_filter) { + case PIPE_TEX_MIPFILTER_NONE: + *lod_fpart = bld->lodf_bld.zero; + break; + case PIPE_TEX_MIPFILTER_NEAREST: + *lod_fpart = lp_build_round(&bld->lodf_bld, *lod_fpart); + /* fallthrough */ + case PIPE_TEX_MIPFILTER_LINEAR: + *lod_fpart = lp_build_clamp(&bld->lodf_bld, *lod_fpart, + bld->lodf_bld.zero, last_level); + break; + } + return; + } + } else { - *lod_ipart = bld->leveli_bld.zero; + lod_ipart = bld->lodi_bld.zero; + *lod_pos_or_zero = bld->lodi_bld.zero; + } + + if (bld->num_lods != bld->num_mips) { + /* only makes sense if there's just a single mip level */ + assert(bld->num_mips == 1); + lod_ipart = lp_build_extract_range(bld->gallivm, lod_ipart, 0, 1); } /* @@ -1170,35 +2201,254 @@ lp_build_sample_common(struct lp_build_sample_context *bld, /* fall-through */ case PIPE_TEX_MIPFILTER_NONE: /* always use mip level 0 */ - if (HAVE_LLVM == 0x0207 && target == PIPE_TEXTURE_CUBE) { - /* XXX this is a work-around for an apparent bug in LLVM 2.7. - * We should be able to set ilevel0 = const(0) but that causes - * bad x86 code to be emitted. - */ - assert(*lod_ipart); - lp_build_nearest_mip_level(bld, texture_index, *lod_ipart, ilevel0, NULL); - } - else { - first_level = bld->dynamic_state->first_level(bld->dynamic_state, - bld->gallivm, texture_index); - first_level = lp_build_broadcast_scalar(&bld->leveli_bld, first_level); - *ilevel0 = first_level; - } + first_level = bld->dynamic_state->first_level(bld->dynamic_state, + bld->gallivm, bld->context_ptr, + texture_index, NULL); + first_level = lp_build_broadcast_scalar(&bld->leveli_bld, first_level); + *ilevel0 = first_level; break; case PIPE_TEX_MIPFILTER_NEAREST: - assert(*lod_ipart); - lp_build_nearest_mip_level(bld, texture_index, *lod_ipart, ilevel0, NULL); + assert(lod_ipart); + lp_build_nearest_mip_level(bld, texture_index, lod_ipart, ilevel0, NULL); break; case PIPE_TEX_MIPFILTER_LINEAR: - assert(*lod_ipart); + assert(lod_ipart); assert(*lod_fpart); lp_build_linear_mip_levels(bld, texture_index, - *lod_ipart, lod_fpart, + lod_ipart, lod_fpart, ilevel0, ilevel1); break; } } +static void +lp_build_clamp_border_color(struct lp_build_sample_context *bld, + unsigned sampler_unit) +{ + struct gallivm_state *gallivm = bld->gallivm; + LLVMBuilderRef builder = gallivm->builder; + LLVMValueRef border_color_ptr = + bld->dynamic_state->border_color(bld->dynamic_state, gallivm, + bld->context_ptr, sampler_unit); + LLVMValueRef border_color; + const struct util_format_description *format_desc = bld->format_desc; + struct lp_type vec4_type = bld->texel_type; + struct lp_build_context vec4_bld; + LLVMValueRef min_clamp = NULL; + LLVMValueRef max_clamp = NULL; + + /* + * For normalized format need to clamp border color (technically + * probably should also quantize the data). Really sucks doing this + * here but can't avoid at least for now since this is part of + * sampler state and texture format is part of sampler_view state. + * GL expects also expects clamping for uint/sint formats too so + * do that as well (d3d10 can't end up here with uint/sint since it + * only supports them with ld). + */ + vec4_type.length = 4; + lp_build_context_init(&vec4_bld, gallivm, vec4_type); + + /* + * Vectorized clamping of border color. Loading is a bit of a hack since + * we just cast the pointer to float array to pointer to vec4 + * (int or float). + */ + border_color_ptr = lp_build_array_get_ptr(gallivm, border_color_ptr, + lp_build_const_int32(gallivm, 0)); + border_color_ptr = LLVMBuildBitCast(builder, border_color_ptr, + LLVMPointerType(vec4_bld.vec_type, 0), ""); + border_color = LLVMBuildLoad(builder, border_color_ptr, ""); + /* we don't have aligned type in the dynamic state unfortunately */ + LLVMSetAlignment(border_color, 4); + + /* + * Instead of having some incredibly complex logic which will try to figure out + * clamping necessary for each channel, simply use the first channel, and treat + * mixed signed/unsigned normalized formats specially. + * (Mixed non-normalized, which wouldn't work at all here, do not exist for a + * good reason.) + */ + if (format_desc->layout == UTIL_FORMAT_LAYOUT_PLAIN) { + int chan; + /* d/s needs special handling because both present means just sampling depth */ + if (util_format_is_depth_and_stencil(format_desc->format)) { + chan = format_desc->swizzle[0]; + } + else { + chan = util_format_get_first_non_void_channel(format_desc->format); + } + if (chan >= 0 && chan <= PIPE_SWIZZLE_W) { + unsigned chan_type = format_desc->channel[chan].type; + unsigned chan_norm = format_desc->channel[chan].normalized; + unsigned chan_pure = format_desc->channel[chan].pure_integer; + if (chan_type == UTIL_FORMAT_TYPE_SIGNED) { + if (chan_norm) { + min_clamp = lp_build_const_vec(gallivm, vec4_type, -1.0F); + max_clamp = vec4_bld.one; + } + else if (chan_pure) { + /* + * Border color was stored as int, hence need min/max clamp + * only if chan has less than 32 bits.. + */ + unsigned chan_size = format_desc->channel[chan].size; + if (chan_size < 32) { + min_clamp = lp_build_const_int_vec(gallivm, vec4_type, + 0 - (1 << (chan_size - 1))); + max_clamp = lp_build_const_int_vec(gallivm, vec4_type, + (1 << (chan_size - 1)) - 1); + } + } + /* TODO: no idea about non-pure, non-normalized! */ + } + else if (chan_type == UTIL_FORMAT_TYPE_UNSIGNED) { + if (chan_norm) { + min_clamp = vec4_bld.zero; + max_clamp = vec4_bld.one; + } + /* + * Need a ugly hack here, because we don't have Z32_FLOAT_X8X24 + * we use Z32_FLOAT_S8X24 to imply sampling depth component + * and ignoring stencil, which will blow up here if we try to + * do a uint clamp in a float texel build... + * And even if we had that format, mesa st also thinks using z24s8 + * means depth sampling ignoring stencil. + */ + else if (chan_pure) { + /* + * Border color was stored as uint, hence never need min + * clamp, and only need max clamp if chan has less than 32 bits. + */ + unsigned chan_size = format_desc->channel[chan].size; + if (chan_size < 32) { + max_clamp = lp_build_const_int_vec(gallivm, vec4_type, + (1 << chan_size) - 1); + } + /* TODO: no idea about non-pure, non-normalized! */ + } + } + else if (chan_type == UTIL_FORMAT_TYPE_FIXED) { + /* TODO: I have no idea what clamp this would need if any! */ + } + } + /* mixed plain formats (or different pure size) */ + switch (format_desc->format) { + case PIPE_FORMAT_B10G10R10A2_UINT: + case PIPE_FORMAT_R10G10B10A2_UINT: + { + unsigned max10 = (1 << 10) - 1; + max_clamp = lp_build_const_aos(gallivm, vec4_type, max10, max10, + max10, (1 << 2) - 1, NULL); + } + break; + case PIPE_FORMAT_R10SG10SB10SA2U_NORM: + min_clamp = lp_build_const_aos(gallivm, vec4_type, -1.0F, -1.0F, + -1.0F, 0.0F, NULL); + max_clamp = vec4_bld.one; + break; + case PIPE_FORMAT_R8SG8SB8UX8U_NORM: + case PIPE_FORMAT_R5SG5SB6U_NORM: + min_clamp = lp_build_const_aos(gallivm, vec4_type, -1.0F, -1.0F, + 0.0F, 0.0F, NULL); + max_clamp = vec4_bld.one; + break; + default: + break; + } + } + else { + /* cannot figure this out from format description */ + if (format_desc->layout == UTIL_FORMAT_LAYOUT_S3TC) { + /* s3tc formats are always unorm */ + min_clamp = vec4_bld.zero; + max_clamp = vec4_bld.one; + } + else if (format_desc->layout == UTIL_FORMAT_LAYOUT_RGTC || + format_desc->layout == UTIL_FORMAT_LAYOUT_ETC || + format_desc->layout == UTIL_FORMAT_LAYOUT_BPTC) { + switch (format_desc->format) { + case PIPE_FORMAT_RGTC1_UNORM: + case PIPE_FORMAT_RGTC2_UNORM: + case PIPE_FORMAT_LATC1_UNORM: + case PIPE_FORMAT_LATC2_UNORM: + case PIPE_FORMAT_ETC1_RGB8: + case PIPE_FORMAT_BPTC_RGBA_UNORM: + case PIPE_FORMAT_BPTC_SRGBA: + min_clamp = vec4_bld.zero; + max_clamp = vec4_bld.one; + break; + case PIPE_FORMAT_RGTC1_SNORM: + case PIPE_FORMAT_RGTC2_SNORM: + case PIPE_FORMAT_LATC1_SNORM: + case PIPE_FORMAT_LATC2_SNORM: + min_clamp = lp_build_const_vec(gallivm, vec4_type, -1.0F); + max_clamp = vec4_bld.one; + break; + case PIPE_FORMAT_BPTC_RGB_FLOAT: + /* not sure if we should clamp to max half float? */ + break; + case PIPE_FORMAT_BPTC_RGB_UFLOAT: + min_clamp = vec4_bld.zero; + break; + default: + assert(0); + break; + } + } + /* + * all others from subsampled/other group, though we don't care + * about yuv (and should not have any from zs here) + */ + else if (format_desc->colorspace != UTIL_FORMAT_COLORSPACE_YUV){ + switch (format_desc->format) { + case PIPE_FORMAT_R8G8_B8G8_UNORM: + case PIPE_FORMAT_G8R8_G8B8_UNORM: + case PIPE_FORMAT_G8R8_B8R8_UNORM: + case PIPE_FORMAT_R8G8_R8B8_UNORM: + case PIPE_FORMAT_R1_UNORM: /* doesn't make sense but ah well */ + min_clamp = vec4_bld.zero; + max_clamp = vec4_bld.one; + break; + case PIPE_FORMAT_R8G8Bx_SNORM: + min_clamp = lp_build_const_vec(gallivm, vec4_type, -1.0F); + max_clamp = vec4_bld.one; + break; + /* + * Note smallfloat formats usually don't need clamping + * (they still have infinite range) however this is not + * true for r11g11b10 and r9g9b9e5, which can't represent + * negative numbers (and additionally r9g9b9e5 can't represent + * very large numbers). d3d10 seems happy without clamping in + * this case, but gl spec is pretty clear: "for floating + * point and integer formats, border values are clamped to + * the representable range of the format" so do that here. + */ + case PIPE_FORMAT_R11G11B10_FLOAT: + min_clamp = vec4_bld.zero; + break; + case PIPE_FORMAT_R9G9B9E5_FLOAT: + min_clamp = vec4_bld.zero; + max_clamp = lp_build_const_vec(gallivm, vec4_type, MAX_RGB9E5); + break; + default: + assert(0); + break; + } + } + } + + if (min_clamp) { + border_color = lp_build_max(&vec4_bld, border_color, min_clamp); + } + if (max_clamp) { + border_color = lp_build_min(&vec4_bld, border_color, max_clamp); + } + + bld->border_color_clamped = border_color; +} + + /** * General texture sampling codegen. * This function handles texture sampling for all texture targets (1D, @@ -1207,24 +2457,39 @@ lp_build_sample_common(struct lp_build_sample_context *bld, static void lp_build_sample_general(struct lp_build_sample_context *bld, unsigned sampler_unit, - LLVMValueRef s, - LLVMValueRef t, - LLVMValueRef r, + boolean is_gather, + const LLVMValueRef *coords, const LLVMValueRef *offsets, - LLVMValueRef lod_ipart, + LLVMValueRef lod_positive, LLVMValueRef lod_fpart, LLVMValueRef ilevel0, LLVMValueRef ilevel1, LLVMValueRef *colors_out) { - struct lp_build_context *int_bld = &bld->int_bld; LLVMBuilderRef builder = bld->gallivm->builder; - const unsigned mip_filter = bld->static_sampler_state->min_mip_filter; - const unsigned min_filter = bld->static_sampler_state->min_img_filter; - const unsigned mag_filter = bld->static_sampler_state->mag_img_filter; + const struct lp_static_sampler_state *sampler_state = bld->static_sampler_state; + const unsigned mip_filter = sampler_state->min_mip_filter; + const unsigned min_filter = sampler_state->min_img_filter; + const unsigned mag_filter = sampler_state->mag_img_filter; LLVMValueRef texels[4]; unsigned chan; + /* if we need border color, (potentially) clamp it now */ + if (lp_sampler_wrap_mode_uses_border_color(sampler_state->wrap_s, + min_filter, + mag_filter) || + (bld->dims > 1 && + lp_sampler_wrap_mode_uses_border_color(sampler_state->wrap_t, + min_filter, + mag_filter)) || + (bld->dims > 2 && + lp_sampler_wrap_mode_uses_border_color(sampler_state->wrap_r, + min_filter, + mag_filter))) { + lp_build_clamp_border_color(bld, sampler_unit); + } + + /* * Get/interpolate texture colors. */ @@ -1236,55 +2501,101 @@ lp_build_sample_general(struct lp_build_sample_context *bld, if (min_filter == mag_filter) { /* no need to distinguish between minification and magnification */ - lp_build_sample_mipmap(bld, sampler_unit, - min_filter, mip_filter, - s, t, r, offsets, + lp_build_sample_mipmap(bld, min_filter, mip_filter, + is_gather, + coords, offsets, ilevel0, ilevel1, lod_fpart, texels); } else { - /* Emit conditional to choose min image filter or mag image filter - * depending on the lod being > 0 or <= 0, respectively. - */ - struct lp_build_if_state if_ctx; - LLVMValueRef minify; - /* - * XXX this should to all lods into account, if some are min - * some max probably could hack up the coords/weights in the linear - * path with selects to work for nearest. - * If that's just two quads sitting next to each other it seems - * quite ok to do the same filtering method on both though, at - * least unless we have explicit lod (and who uses different - * min/mag filter with that?) + * Could also get rid of the if-logic and always use mipmap_both, both + * for the single lod and multi-lod case if nothing really uses this. */ - if (bld->num_lods > 1) - lod_ipart = LLVMBuildExtractElement(builder, lod_ipart, - lp_build_const_int32(bld->gallivm, 0), ""); + if (bld->num_lods == 1) { + /* Emit conditional to choose min image filter or mag image filter + * depending on the lod being > 0 or <= 0, respectively. + */ + struct lp_build_if_state if_ctx; + + lod_positive = LLVMBuildTrunc(builder, lod_positive, + LLVMInt1TypeInContext(bld->gallivm->context), + "lod_pos"); + + lp_build_if(&if_ctx, bld->gallivm, lod_positive); + { + /* Use the minification filter */ + lp_build_sample_mipmap(bld, min_filter, mip_filter, FALSE, + coords, offsets, + ilevel0, ilevel1, lod_fpart, + texels); + } + lp_build_else(&if_ctx); + { + /* Use the magnification filter */ + lp_build_sample_mipmap(bld, mag_filter, PIPE_TEX_MIPFILTER_NONE, + FALSE, + coords, offsets, + ilevel0, NULL, NULL, + texels); + } + lp_build_endif(&if_ctx); + } + else { + LLVMValueRef need_linear, linear_mask; + unsigned mip_filter_for_nearest; + struct lp_build_if_state if_ctx; - /* minify = lod >= 0.0 */ - minify = LLVMBuildICmp(builder, LLVMIntSGE, - lod_ipart, int_bld->zero, ""); + if (min_filter == PIPE_TEX_FILTER_LINEAR) { + linear_mask = lod_positive; + mip_filter_for_nearest = PIPE_TEX_MIPFILTER_NONE; + } + else { + linear_mask = lp_build_not(&bld->lodi_bld, lod_positive); + mip_filter_for_nearest = mip_filter; + } + need_linear = lp_build_any_true_range(&bld->lodi_bld, bld->num_lods, + linear_mask); + lp_build_name(need_linear, "need_linear"); + + if (bld->num_lods != bld->coord_type.length) { + linear_mask = lp_build_unpack_broadcast_aos_scalars(bld->gallivm, + bld->lodi_type, + bld->int_coord_type, + linear_mask); + } - lp_build_if(&if_ctx, bld->gallivm, minify); - { - /* Use the minification filter */ - lp_build_sample_mipmap(bld, sampler_unit, - min_filter, mip_filter, - s, t, r, offsets, - ilevel0, ilevel1, lod_fpart, - texels); - } - lp_build_else(&if_ctx); - { - /* Use the magnification filter */ - lp_build_sample_mipmap(bld, sampler_unit, - mag_filter, PIPE_TEX_MIPFILTER_NONE, - s, t, r, offsets, - ilevel0, NULL, NULL, - texels); + lp_build_if(&if_ctx, bld->gallivm, need_linear); + { + /* + * Do sampling with both filters simultaneously. This means using + * a linear filter and doing some tricks (with weights) for the pixels + * which need nearest filter. + * Note that it's probably rare some pixels need nearest and some + * linear filter but the fixups required for the nearest pixels + * aren't all that complicated so just always run a combined path + * if at least some pixels require linear. + */ + lp_build_sample_mipmap_both(bld, linear_mask, mip_filter, + coords, offsets, + ilevel0, ilevel1, + lod_fpart, lod_positive, + texels); + } + lp_build_else(&if_ctx); + { + /* + * All pixels require just nearest filtering, which is way + * cheaper than linear, hence do a separate path for that. + */ + lp_build_sample_mipmap(bld, PIPE_TEX_FILTER_NEAREST, + mip_filter_for_nearest, FALSE, + coords, offsets, + ilevel0, ilevel1, lod_fpart, + texels); + } + lp_build_endif(&if_ctx); } - lp_build_endif(&if_ctx); } for (chan = 0; chan < 4; ++chan) { @@ -1305,12 +2616,13 @@ lp_build_sample_general(struct lp_build_sample_context *bld, static void lp_build_fetch_texel(struct lp_build_sample_context *bld, unsigned texture_unit, + LLVMValueRef ms_index, const LLVMValueRef *coords, LLVMValueRef explicit_lod, const LLVMValueRef *offsets, LLVMValueRef *colors_out) { - struct lp_build_context *perquadi_bld = &bld->leveli_bld; + struct lp_build_context *perquadi_bld = &bld->lodi_bld; struct lp_build_context *int_coord_bld = &bld->int_coord_bld; unsigned dims = bld->dims, chan; unsigned target = bld->static_texture_state->target; @@ -1324,7 +2636,7 @@ lp_build_fetch_texel(struct lp_build_sample_context *bld, out_of_bounds = int_coord_bld->zero; if (explicit_lod && bld->static_texture_state->target != PIPE_BUFFER) { - if (bld->num_lods != int_coord_bld->type.length) { + if (bld->num_mips != int_coord_bld->type.length) { ilevel = lp_build_pack_aos_scalars(bld->gallivm, int_coord_bld->type, perquadi_bld->type, explicit_lod, 0); } @@ -1335,10 +2647,10 @@ lp_build_fetch_texel(struct lp_build_sample_context *bld, out_of_bound_ret_zero ? &out_of_bounds : NULL); } else { - assert(bld->num_lods == 1); + assert(bld->num_mips == 1); if (bld->static_texture_state->target != PIPE_BUFFER) { - ilevel = bld->dynamic_state->first_level(bld->dynamic_state, - bld->gallivm, texture_unit); + ilevel = bld->dynamic_state->first_level(bld->dynamic_state, bld->gallivm, + bld->context_ptr, texture_unit, NULL); } else { ilevel = lp_build_const_int32(bld->gallivm, 0); @@ -1352,15 +2664,12 @@ lp_build_fetch_texel(struct lp_build_sample_context *bld, if (target == PIPE_TEXTURE_1D_ARRAY || target == PIPE_TEXTURE_2D_ARRAY) { - if (target == PIPE_TEXTURE_1D_ARRAY) { - z = y; - } if (out_of_bound_ret_zero) { - z = lp_build_layer_coord(bld, texture_unit, z, &out1); + z = lp_build_layer_coord(bld, texture_unit, FALSE, z, &out1); out_of_bounds = lp_build_or(int_coord_bld, out_of_bounds, out1); } else { - z = lp_build_layer_coord(bld, texture_unit, z, NULL); + z = lp_build_layer_coord(bld, texture_unit, FALSE, z, NULL); } } @@ -1407,13 +2716,26 @@ lp_build_fetch_texel(struct lp_build_sample_context *bld, lp_build_get_mip_offsets(bld, ilevel)); } + if (bld->fetch_ms) { + LLVMValueRef num_samples; + num_samples = bld->dynamic_state->num_samples(bld->dynamic_state, bld->gallivm, + bld->context_ptr, texture_unit, NULL); + out1 = lp_build_cmp(int_coord_bld, PIPE_FUNC_LESS, ms_index, int_coord_bld->zero); + out_of_bounds = lp_build_or(int_coord_bld, out_of_bounds, out1); + out1 = lp_build_cmp(int_coord_bld, PIPE_FUNC_GEQUAL, ms_index, lp_build_broadcast_scalar(int_coord_bld, num_samples)); + out_of_bounds = lp_build_or(int_coord_bld, out_of_bounds, out1); + offset = lp_build_add(int_coord_bld, offset, + lp_build_mul(int_coord_bld, bld->sample_stride, ms_index)); + } + offset = lp_build_andnot(int_coord_bld, offset, out_of_bounds); lp_build_fetch_rgba_soa(bld->gallivm, bld->format_desc, - bld->texel_type, + bld->texel_type, TRUE, bld->base_ptr, offset, i, j, + bld->cache, colors_out); if (out_of_bound_ret_zero) { @@ -1430,73 +2752,6 @@ lp_build_fetch_texel(struct lp_build_sample_context *bld, } -/** - * Do shadow test/comparison. - * \param coords incoming texcoords - * \param texel the texel to compare against (use the X channel) - * Ideally this should really be done per-sample. - */ -static void -lp_build_sample_compare(struct lp_build_sample_context *bld, - const LLVMValueRef *coords, - LLVMValueRef texel[4]) -{ - struct lp_build_context *texel_bld = &bld->texel_bld; - LLVMBuilderRef builder = bld->gallivm->builder; - LLVMValueRef res, p; - const unsigned chan = 0; - unsigned chan_type; - const struct util_format_description *format_desc; - - if (bld->static_sampler_state->compare_mode == PIPE_TEX_COMPARE_NONE) - return; - - if (bld->static_texture_state->target == PIPE_TEXTURE_2D_ARRAY || - bld->static_texture_state->target == PIPE_TEXTURE_CUBE) { - p = coords[3]; - } - else { - p = coords[2]; - } - - /* debug code */ - if (0) { - LLVMValueRef indx = lp_build_const_int32(bld->gallivm, 0); - LLVMValueRef coord = LLVMBuildExtractElement(builder, p, indx, ""); - LLVMValueRef tex = LLVMBuildExtractElement(builder, texel[chan], indx, ""); - lp_build_printf(bld->gallivm, "shadow compare coord %f to texture %f\n", - coord, tex); - } - - /* Clamp p coords to [0,1] for fixed function depth texture format */ - format_desc = util_format_description(bld->static_texture_state->format); - /* not entirely sure we couldn't end up with non-valid swizzle here */ - chan_type = format_desc->swizzle[0] <= UTIL_FORMAT_SWIZZLE_W ? - format_desc->channel[format_desc->swizzle[0]].type : - UTIL_FORMAT_TYPE_FLOAT; - if (chan_type != UTIL_FORMAT_TYPE_FLOAT) { - p = lp_build_clamp(&bld->coord_bld, p, - bld->coord_bld.zero, bld->coord_bld.one); - } - - /* - * technically this is not entirely correct for unorm depth as the ref value - * should be converted to the depth format (quantization!) and comparison - * then done in texture format. - */ - /* result = (p FUNC texel) ? 1 : 0 */ - res = lp_build_cmp(texel_bld, bld->static_sampler_state->compare_func, - p, texel[chan]); - res = lp_build_select(texel_bld, res, texel_bld->one, texel_bld->zero); - - /* XXX returning result for default GL_DEPTH_TEXTURE_MODE = GL_LUMINANCE */ - texel[0] = - texel[1] = - texel[2] = res; - texel[3] = texel_bld->one; -} - - /** * Just set texels to white instead of actually sampling the texture. * For debugging. @@ -1517,58 +2772,118 @@ lp_build_sample_nop(struct gallivm_state *gallivm, /** - * Build texture sampling code. + * Build the actual texture sampling code. * 'texel' will return a vector of four LLVMValueRefs corresponding to * R, G, B, A. * \param type vector float type to use for coords, etc. - * \param is_fetch if this is a texel fetch instruction. + * \param sample_key * \param derivs partial derivatives of (s,t,r,q) with respect to x and y */ -void -lp_build_sample_soa(struct gallivm_state *gallivm, - const struct lp_static_texture_state *static_texture_state, - const struct lp_static_sampler_state *static_sampler_state, - struct lp_sampler_dynamic_state *dynamic_state, - struct lp_type type, - boolean is_fetch, - unsigned texture_index, - unsigned sampler_index, - const LLVMValueRef *coords, - const LLVMValueRef *offsets, - const struct lp_derivatives *derivs, /* optional */ - LLVMValueRef lod_bias, /* optional */ - LLVMValueRef explicit_lod, /* optional */ - boolean scalar_lod, - LLVMValueRef texel_out[4]) +static void +lp_build_sample_soa_code(struct gallivm_state *gallivm, + const struct lp_static_texture_state *static_texture_state, + const struct lp_static_sampler_state *static_sampler_state, + struct lp_sampler_dynamic_state *dynamic_state, + struct lp_type type, + unsigned sample_key, + unsigned texture_index, + unsigned sampler_index, + LLVMValueRef context_ptr, + LLVMValueRef thread_data_ptr, + const LLVMValueRef *coords, + const LLVMValueRef *offsets, + const struct lp_derivatives *derivs, /* optional */ + LLVMValueRef lod, /* optional */ + LLVMValueRef ms_index, /* optional */ + LLVMValueRef texel_out[4]) { - unsigned dims = texture_dims(static_texture_state->target); + unsigned target = static_texture_state->target; + unsigned dims = texture_dims(target); unsigned num_quads = type.length / 4; - unsigned mip_filter; + unsigned mip_filter, min_img_filter, mag_img_filter, i; struct lp_build_sample_context bld; struct lp_static_sampler_state derived_sampler_state = *static_sampler_state; LLVMTypeRef i32t = LLVMInt32TypeInContext(gallivm->context); LLVMBuilderRef builder = gallivm->builder; - LLVMValueRef tex_width; - LLVMValueRef s; - LLVMValueRef t; - LLVMValueRef r; + LLVMValueRef tex_width, newcoords[5]; + enum lp_sampler_lod_property lod_property; + enum lp_sampler_lod_control lod_control; + enum lp_sampler_op_type op_type; + LLVMValueRef lod_bias = NULL; + LLVMValueRef explicit_lod = NULL; + boolean op_is_tex, op_is_lodq, op_is_gather, fetch_ms; if (0) { enum pipe_format fmt = static_texture_state->format; debug_printf("Sample from %s\n", util_format_name(fmt)); } + lod_property = (sample_key & LP_SAMPLER_LOD_PROPERTY_MASK) >> + LP_SAMPLER_LOD_PROPERTY_SHIFT; + lod_control = (sample_key & LP_SAMPLER_LOD_CONTROL_MASK) >> + LP_SAMPLER_LOD_CONTROL_SHIFT; + op_type = (sample_key & LP_SAMPLER_OP_TYPE_MASK) >> + LP_SAMPLER_OP_TYPE_SHIFT; + fetch_ms = !!(sample_key & LP_SAMPLER_FETCH_MS); + + op_is_tex = op_type == LP_SAMPLER_OP_TEXTURE; + op_is_lodq = op_type == LP_SAMPLER_OP_LODQ; + op_is_gather = op_type == LP_SAMPLER_OP_GATHER; + + if (lod_control == LP_SAMPLER_LOD_BIAS) { + lod_bias = lod; + assert(lod); + assert(derivs == NULL); + } + else if (lod_control == LP_SAMPLER_LOD_EXPLICIT) { + explicit_lod = lod; + assert(lod); + assert(derivs == NULL); + } + else if (lod_control == LP_SAMPLER_LOD_DERIVATIVES) { + assert(derivs); + assert(lod == NULL); + } + else { + assert(derivs == NULL); + assert(lod == NULL); + } + + if (static_texture_state->format == PIPE_FORMAT_NONE) { + /* + * If there's nothing bound, format is NONE, and we must return + * all zero as mandated by d3d10 in this case. + */ + unsigned chan; + LLVMValueRef zero = lp_build_zero(gallivm, type); + for (chan = 0; chan < 4; chan++) { + texel_out[chan] = zero; + } + return; + } + assert(type.floating); /* Setup our build context */ memset(&bld, 0, sizeof bld); bld.gallivm = gallivm; + bld.context_ptr = context_ptr; bld.static_sampler_state = &derived_sampler_state; bld.static_texture_state = static_texture_state; bld.dynamic_state = dynamic_state; bld.format_desc = util_format_description(static_texture_state->format); bld.dims = dims; + if (gallivm_perf & GALLIVM_PERF_NO_QUAD_LOD || op_is_lodq) { + bld.no_quad_lod = TRUE; + } + if (gallivm_perf & GALLIVM_PERF_NO_RHO_APPROX || op_is_lodq) { + bld.no_rho_approx = TRUE; + } + if (gallivm_perf & GALLIVM_PERF_NO_BRILINEAR || op_is_lodq) { + bld.no_brilinear = TRUE; + } + bld.vector_width = lp_type_width(type); bld.float_type = lp_type_float(32); @@ -1595,20 +2910,56 @@ lp_build_sample_soa(struct gallivm_state *gallivm, else if (util_format_has_stencil(bld.format_desc) && !util_format_has_depth(bld.format_desc)) { /* for stencil only formats, sample stencil (uint) */ - bld.texel_type = lp_type_int_vec(type.width, type.width * type.length); + bld.texel_type = lp_type_uint_vec(type.width, type.width * type.length); } - if (!static_texture_state->level_zero_only) { + if (!static_texture_state->level_zero_only || + !static_sampler_state->max_lod_pos || op_is_lodq) { derived_sampler_state.min_mip_filter = static_sampler_state->min_mip_filter; } else { derived_sampler_state.min_mip_filter = PIPE_TEX_MIPFILTER_NONE; } + if (op_is_gather) { + /* + * gather4 is exactly like GL_LINEAR filtering but in the end skipping + * the actual filtering. Using mostly the same paths, so cube face + * selection, coord wrapping etc. all naturally uses the same code. + */ + derived_sampler_state.min_mip_filter = PIPE_TEX_MIPFILTER_NONE; + derived_sampler_state.min_img_filter = PIPE_TEX_FILTER_LINEAR; + derived_sampler_state.mag_img_filter = PIPE_TEX_FILTER_LINEAR; + } mip_filter = derived_sampler_state.min_mip_filter; if (0) { debug_printf(" .min_mip_filter = %u\n", derived_sampler_state.min_mip_filter); } + if (static_texture_state->target == PIPE_TEXTURE_CUBE || + static_texture_state->target == PIPE_TEXTURE_CUBE_ARRAY) + { + /* + * Seamless filtering ignores wrap modes. + * Setting to CLAMP_TO_EDGE is correct for nearest filtering, for + * bilinear it's not correct but way better than using for instance repeat. + * Note we even set this for non-seamless. Technically GL allows any wrap + * mode, which made sense when supporting true borders (can get seamless + * effect with border and CLAMP_TO_BORDER), but gallium doesn't support + * borders and d3d9 requires wrap modes to be ignored and it's a pain to fix + * up the sampler state (as it makes it texture dependent). + */ + derived_sampler_state.wrap_s = PIPE_TEX_WRAP_CLAMP_TO_EDGE; + derived_sampler_state.wrap_t = PIPE_TEX_WRAP_CLAMP_TO_EDGE; + } + /* + * We could force CLAMP to CLAMP_TO_EDGE here if min/mag filter is nearest, + * so AoS path could be used. Not sure it's worth the trouble... + */ + + min_img_filter = derived_sampler_state.min_img_filter; + mag_img_filter = derived_sampler_state.mag_img_filter; + + /* * This is all a bit complicated different paths are chosen for performance * reasons. @@ -1627,31 +2978,70 @@ lp_build_sample_soa(struct gallivm_state *gallivm, * There are other situations where at least the multiple int lods could be * avoided like min and max lod being equal. */ - if (explicit_lod && !scalar_lod && - ((is_fetch && bld.static_texture_state->target != PIPE_BUFFER) || - (!is_fetch && mip_filter != PIPE_TEX_MIPFILTER_NONE))) - bld.num_lods = type.length; + bld.num_mips = bld.num_lods = 1; + + if (bld.no_quad_lod && bld.no_rho_approx && + ((mip_filter != PIPE_TEX_MIPFILTER_NONE && op_is_tex && + (static_texture_state->target == PIPE_TEXTURE_CUBE || + static_texture_state->target == PIPE_TEXTURE_CUBE_ARRAY)) || + op_is_lodq)) { + /* + * special case for using per-pixel lod even for implicit lod, + * which is generally never required (ok by APIs) except to please + * some (somewhat broken imho) tests (because per-pixel face selection + * can cause derivatives to be different for pixels outside the primitive + * due to the major axis division even if pre-project derivatives are + * looking normal). + * For lodq, we do it to simply avoid scalar pack / unpack (albeit for + * cube maps we do indeed get per-pixel lod values). + */ + bld.num_mips = type.length; + bld.num_lods = type.length; + } + else if (lod_property == LP_SAMPLER_LOD_PER_ELEMENT || + (explicit_lod || lod_bias || derivs)) { + if ((!op_is_tex && target != PIPE_BUFFER) || + (op_is_tex && mip_filter != PIPE_TEX_MIPFILTER_NONE)) { + bld.num_mips = type.length; + bld.num_lods = type.length; + } + else if (op_is_tex && min_img_filter != mag_img_filter) { + bld.num_mips = 1; + bld.num_lods = type.length; + } + } /* TODO: for true scalar_lod should only use 1 lod value */ - else if ((is_fetch && explicit_lod && bld.static_texture_state->target != PIPE_BUFFER ) || - (!is_fetch && mip_filter != PIPE_TEX_MIPFILTER_NONE)) { + else if ((!op_is_tex && explicit_lod && target != PIPE_BUFFER) || + (op_is_tex && mip_filter != PIPE_TEX_MIPFILTER_NONE)) { + bld.num_mips = num_quads; bld.num_lods = num_quads; } - else { - bld.num_lods = 1; + else if (op_is_tex && min_img_filter != mag_img_filter) { + bld.num_mips = 1; + bld.num_lods = num_quads; } - bld.levelf_type = type; + bld.fetch_ms = fetch_ms; + if (op_is_gather) + bld.gather_comp = (sample_key & LP_SAMPLER_GATHER_COMP_MASK) >> LP_SAMPLER_GATHER_COMP_SHIFT; + bld.lodf_type = type; /* we want native vector size to be able to use our intrinsics */ if (bld.num_lods != type.length) { - bld.levelf_type.length = type.length > 4 ? ((type.length + 15) / 16) * 4 : 1; + /* TODO: this currently always has to be per-quad or per-element */ + bld.lodf_type.length = type.length > 4 ? ((type.length + 15) / 16) * 4 : 1; + } + bld.lodi_type = lp_int_type(bld.lodf_type); + bld.levelf_type = bld.lodf_type; + if (bld.num_mips == 1) { + bld.levelf_type.length = 1; } bld.leveli_type = lp_int_type(bld.levelf_type); bld.float_size_type = bld.float_size_in_type; /* Note: size vectors may not be native. They contain minified w/h/d/_ values, * with per-element lod that is w0/h0/d0/_/w1/h1/d1_/... so up to 8x4f32 */ - if (bld.num_lods > 1) { - bld.float_size_type.length = bld.num_lods == type.length ? - bld.num_lods * bld.float_size_in_type.length : + if (bld.num_mips > 1) { + bld.float_size_type.length = bld.num_mips == type.length ? + bld.num_mips * bld.float_size_in_type.length : type.length; } bld.int_size_type = lp_int_type(bld.float_size_type); @@ -1668,18 +3058,30 @@ lp_build_sample_soa(struct gallivm_state *gallivm, lp_build_context_init(&bld.texel_bld, gallivm, bld.texel_type); lp_build_context_init(&bld.levelf_bld, gallivm, bld.levelf_type); lp_build_context_init(&bld.leveli_bld, gallivm, bld.leveli_type); + lp_build_context_init(&bld.lodf_bld, gallivm, bld.lodf_type); + lp_build_context_init(&bld.lodi_bld, gallivm, bld.lodi_type); /* Get the dynamic state */ - tex_width = dynamic_state->width(dynamic_state, gallivm, texture_index); - bld.row_stride_array = dynamic_state->row_stride(dynamic_state, gallivm, texture_index); - bld.img_stride_array = dynamic_state->img_stride(dynamic_state, gallivm, texture_index); - bld.base_ptr = dynamic_state->base_ptr(dynamic_state, gallivm, texture_index); - bld.mip_offsets = dynamic_state->mip_offsets(dynamic_state, gallivm, texture_index); + tex_width = dynamic_state->width(dynamic_state, gallivm, + context_ptr, texture_index, NULL); + bld.row_stride_array = dynamic_state->row_stride(dynamic_state, gallivm, + context_ptr, texture_index, NULL); + bld.img_stride_array = dynamic_state->img_stride(dynamic_state, gallivm, + context_ptr, texture_index, NULL); + bld.base_ptr = dynamic_state->base_ptr(dynamic_state, gallivm, + context_ptr, texture_index, NULL); + bld.mip_offsets = dynamic_state->mip_offsets(dynamic_state, gallivm, + context_ptr, texture_index, NULL); + + if (fetch_ms) + bld.sample_stride = lp_build_broadcast_scalar(&bld.int_coord_bld, dynamic_state->sample_stride(dynamic_state, gallivm, + context_ptr, texture_index, NULL)); /* Note that mip_offsets is an array[level] of offsets to texture images */ - s = coords[0]; - t = coords[1]; - r = coords[2]; + if (dynamic_state->cache_ptr && thread_data_ptr) { + bld.cache = dynamic_state->cache_ptr(dynamic_state, gallivm, + thread_data_ptr, texture_index); + } /* width, height, depth as single int vector */ if (dims <= 1) { @@ -1687,91 +3089,163 @@ lp_build_sample_soa(struct gallivm_state *gallivm, } else { bld.int_size = LLVMBuildInsertElement(builder, bld.int_size_in_bld.undef, - tex_width, LLVMConstInt(i32t, 0, 0), ""); + tex_width, + LLVMConstInt(i32t, 0, 0), ""); if (dims >= 2) { LLVMValueRef tex_height = - dynamic_state->height(dynamic_state, gallivm, texture_index); + dynamic_state->height(dynamic_state, gallivm, + context_ptr, texture_index, NULL); bld.int_size = LLVMBuildInsertElement(builder, bld.int_size, - tex_height, LLVMConstInt(i32t, 1, 0), ""); + tex_height, + LLVMConstInt(i32t, 1, 0), ""); if (dims >= 3) { LLVMValueRef tex_depth = - dynamic_state->depth(dynamic_state, gallivm, texture_index); + dynamic_state->depth(dynamic_state, gallivm, context_ptr, + texture_index, NULL); bld.int_size = LLVMBuildInsertElement(builder, bld.int_size, - tex_depth, LLVMConstInt(i32t, 2, 0), ""); + tex_depth, + LLVMConstInt(i32t, 2, 0), ""); } } } + for (i = 0; i < 5; i++) { + newcoords[i] = coords[i]; + } + + if (util_format_is_pure_integer(static_texture_state->format) && + !util_format_has_depth(bld.format_desc) && op_is_tex && + (static_sampler_state->min_mip_filter == PIPE_TEX_MIPFILTER_LINEAR || + static_sampler_state->min_img_filter == PIPE_TEX_FILTER_LINEAR || + static_sampler_state->mag_img_filter == PIPE_TEX_FILTER_LINEAR)) { + /* + * Bail if impossible filtering is specified (the awkard additional + * depth check is because it is legal in gallium to have things like S8Z24 + * here which would say it's pure int despite such formats should sample + * the depth component). + * In GL such filters make the texture incomplete, this makes it robust + * against gallium frontends which set this up regardless (we'd crash in the + * lerp later otherwise). + * At least in some apis it may be legal to use such filters with lod + * queries and/or gather (at least for gather d3d10 says only the wrap + * bits are really used hence filter bits are likely simply ignored). + * For fetch, we don't get valid samplers either way here. + */ + unsigned chan; + LLVMValueRef zero = lp_build_zero(gallivm, type); + for (chan = 0; chan < 4; chan++) { + texel_out[chan] = zero; + } + return; + } + if (0) { /* For debug: no-op texture sampling */ lp_build_sample_nop(gallivm, bld.texel_type, - coords, + newcoords, texel_out); } - else if (is_fetch) { - lp_build_fetch_texel(&bld, texture_index, coords, - explicit_lod, offsets, + else if (op_type == LP_SAMPLER_OP_FETCH) { + lp_build_fetch_texel(&bld, texture_index, ms_index, newcoords, + lod, offsets, texel_out); } else { - LLVMValueRef lod_ipart = NULL, lod_fpart = NULL; - LLVMValueRef ilevel0 = NULL, ilevel1 = NULL; - boolean use_aos = util_format_fits_8unorm(bld.format_desc) && - lp_is_simple_wrap_mode(static_sampler_state->wrap_s) && - lp_is_simple_wrap_mode(static_sampler_state->wrap_t); + LLVMValueRef lod_fpart = NULL, lod_positive = NULL; + LLVMValueRef ilevel0 = NULL, ilevel1 = NULL, lod = NULL; + boolean use_aos; + + use_aos = util_format_fits_8unorm(bld.format_desc) && + op_is_tex && + /* not sure this is strictly needed or simply impossible */ + derived_sampler_state.compare_mode == PIPE_TEX_COMPARE_NONE && + lp_is_simple_wrap_mode(derived_sampler_state.wrap_s); + + use_aos &= bld.num_lods <= num_quads || + derived_sampler_state.min_img_filter == + derived_sampler_state.mag_img_filter; + + if(gallivm_perf & GALLIVM_PERF_NO_AOS_SAMPLING) { + use_aos = 0; + } + + if (dims > 1) { + use_aos &= lp_is_simple_wrap_mode(derived_sampler_state.wrap_t); + if (dims > 2) { + use_aos &= lp_is_simple_wrap_mode(derived_sampler_state.wrap_r); + } + } + if ((static_texture_state->target == PIPE_TEXTURE_CUBE || + static_texture_state->target == PIPE_TEXTURE_CUBE_ARRAY) && + derived_sampler_state.seamless_cube_map && + (derived_sampler_state.min_img_filter == PIPE_TEX_FILTER_LINEAR || + derived_sampler_state.mag_img_filter == PIPE_TEX_FILTER_LINEAR)) { + /* theoretically possible with AoS filtering but not implemented (complex!) */ + use_aos = 0; + } if ((gallivm_debug & GALLIVM_DEBUG_PERF) && !use_aos && util_format_fits_8unorm(bld.format_desc)) { debug_printf("%s: using floating point linear filtering for %s\n", __FUNCTION__, bld.format_desc->short_name); - debug_printf(" min_img %d mag_img %d mip %d wraps %d wrapt %d\n", - static_sampler_state->min_img_filter, - static_sampler_state->mag_img_filter, - static_sampler_state->min_mip_filter, - static_sampler_state->wrap_s, - static_sampler_state->wrap_t); + debug_printf(" min_img %d mag_img %d mip %d target %d seamless %d" + " wraps %d wrapt %d wrapr %d\n", + derived_sampler_state.min_img_filter, + derived_sampler_state.mag_img_filter, + derived_sampler_state.min_mip_filter, + static_texture_state->target, + derived_sampler_state.seamless_cube_map, + derived_sampler_state.wrap_s, + derived_sampler_state.wrap_t, + derived_sampler_state.wrap_r); } - lp_build_sample_common(&bld, texture_index, sampler_index, - &s, &t, &r, + lp_build_sample_common(&bld, op_is_lodq, texture_index, sampler_index, + newcoords, derivs, lod_bias, explicit_lod, - &lod_ipart, &lod_fpart, + &lod_positive, &lod, &lod_fpart, &ilevel0, &ilevel1); + if (op_is_lodq) { + texel_out[0] = lod_fpart; + texel_out[1] = lod; + texel_out[2] = texel_out[3] = bld.coord_bld.zero; + return; + } + + if (use_aos && static_texture_state->target == PIPE_TEXTURE_CUBE_ARRAY) { + /* The aos path doesn't do seamless filtering so simply add cube layer + * to face now. + */ + newcoords[2] = lp_build_add(&bld.int_coord_bld, newcoords[2], newcoords[3]); + } + /* - * we only try 8-wide sampling with soa as it appears to - * be a loss with aos with AVX (but it should work). - * (It should be faster if we'd support avx2) + * we only try 8-wide sampling with soa or if we have AVX2 + * as it appears to be a loss with just AVX) */ - if (num_quads == 1 || !use_aos) { - - if (num_quads > 1) { - if (mip_filter == PIPE_TEX_MIPFILTER_NONE) { - LLVMValueRef index0 = lp_build_const_int32(gallivm, 0); - /* - * These parameters are the same for all quads, - * could probably simplify. - */ - lod_ipart = LLVMBuildExtractElement(builder, lod_ipart, index0, ""); - ilevel0 = LLVMBuildExtractElement(builder, ilevel0, index0, ""); - } - } + if (num_quads == 1 || !use_aos || + (util_cpu_caps.has_avx2 && + (bld.num_lods == 1 || + derived_sampler_state.min_img_filter == derived_sampler_state.mag_img_filter))) { if (use_aos) { /* do sampling/filtering with fixed pt arithmetic */ lp_build_sample_aos(&bld, sampler_index, - s, t, r, offsets, - lod_ipart, lod_fpart, + newcoords[0], newcoords[1], + newcoords[2], + offsets, lod_positive, lod_fpart, ilevel0, ilevel1, texel_out); } else { lp_build_sample_general(&bld, sampler_index, - s, t, r, offsets, - lod_ipart, lod_fpart, + op_type == LP_SAMPLER_OP_GATHER, + newcoords, offsets, + lod_positive, lod_fpart, ilevel0, ilevel1, texel_out); } @@ -1788,7 +3262,11 @@ lp_build_sample_soa(struct gallivm_state *gallivm, /* Setup our build context */ memset(&bld4, 0, sizeof bld4); + bld4.no_quad_lod = bld.no_quad_lod; + bld4.no_rho_approx = bld.no_rho_approx; + bld4.no_brilinear = bld.no_brilinear; bld4.gallivm = bld.gallivm; + bld4.context_ptr = bld.context_ptr; bld4.static_texture_state = bld.static_texture_state; bld4.static_sampler_state = bld.static_sampler_state; bld4.dynamic_state = bld.dynamic_state; @@ -1799,6 +3277,7 @@ lp_build_sample_soa(struct gallivm_state *gallivm, bld4.base_ptr = bld.base_ptr; bld4.mip_offsets = bld.mip_offsets; bld4.int_size = bld.int_size; + bld4.cache = bld.cache; bld4.vector_width = lp_type_width(type4); @@ -1811,28 +3290,43 @@ lp_build_sample_soa(struct gallivm_state *gallivm, bld4.int_size_in_type = lp_int_type(bld4.float_size_in_type); bld4.texel_type = bld.texel_type; bld4.texel_type.length = 4; - bld4.levelf_type = type4; - /* we want native vector size to be able to use our intrinsics */ - bld4.levelf_type.length = 1; - bld4.leveli_type = lp_int_type(bld4.levelf_type); - if (explicit_lod && !scalar_lod && - ((is_fetch && bld.static_texture_state->target != PIPE_BUFFER) || - (!is_fetch && mip_filter != PIPE_TEX_MIPFILTER_NONE))) + bld4.num_mips = bld4.num_lods = 1; + if (bld4.no_quad_lod && bld4.no_rho_approx && + (static_texture_state->target == PIPE_TEXTURE_CUBE || + static_texture_state->target == PIPE_TEXTURE_CUBE_ARRAY) && + (op_is_tex && mip_filter != PIPE_TEX_MIPFILTER_NONE)) { + bld4.num_mips = type4.length; bld4.num_lods = type4.length; - else - bld4.num_lods = 1; + } + if (lod_property == LP_SAMPLER_LOD_PER_ELEMENT && + (explicit_lod || lod_bias || derivs)) { + if ((!op_is_tex && target != PIPE_BUFFER) || + (op_is_tex && mip_filter != PIPE_TEX_MIPFILTER_NONE)) { + bld4.num_mips = type4.length; + bld4.num_lods = type4.length; + } + else if (op_is_tex && min_img_filter != mag_img_filter) { + bld4.num_mips = 1; + bld4.num_lods = type4.length; + } + } - bld4.levelf_type = type4; /* we want native vector size to be able to use our intrinsics */ + bld4.lodf_type = type4; if (bld4.num_lods != type4.length) { + bld4.lodf_type.length = 1; + } + bld4.lodi_type = lp_int_type(bld4.lodf_type); + bld4.levelf_type = type4; + if (bld4.num_mips != type4.length) { bld4.levelf_type.length = 1; } bld4.leveli_type = lp_int_type(bld4.levelf_type); bld4.float_size_type = bld4.float_size_in_type; - if (bld4.num_lods > 1) { - bld4.float_size_type.length = bld4.num_lods == type4.length ? - bld4.num_lods * bld4.float_size_in_type.length : + if (bld4.num_mips > 1) { + bld4.float_size_type.length = bld4.num_mips == type4.length ? + bld4.num_mips * bld4.float_size_in_type.length : type4.length; } bld4.int_size_type = lp_int_type(bld4.float_size_type); @@ -1849,17 +3343,19 @@ lp_build_sample_soa(struct gallivm_state *gallivm, lp_build_context_init(&bld4.texel_bld, gallivm, bld4.texel_type); lp_build_context_init(&bld4.levelf_bld, gallivm, bld4.levelf_type); lp_build_context_init(&bld4.leveli_bld, gallivm, bld4.leveli_type); + lp_build_context_init(&bld4.lodf_bld, gallivm, bld4.lodf_type); + lp_build_context_init(&bld4.lodi_bld, gallivm, bld4.lodi_type); for (i = 0; i < num_quads; i++) { LLVMValueRef s4, t4, r4; - LLVMValueRef lod_ipart4, lod_fpart4 = NULL; + LLVMValueRef lod_positive4, lod_fpart4 = NULL; LLVMValueRef ilevel04, ilevel14 = NULL; LLVMValueRef offsets4[4] = { NULL }; unsigned num_lods = bld4.num_lods; - s4 = lp_build_extract_range(gallivm, s, 4*i, 4); - t4 = lp_build_extract_range(gallivm, t, 4*i, 4); - r4 = lp_build_extract_range(gallivm, r, 4*i, 4); + s4 = lp_build_extract_range(gallivm, newcoords[0], 4*i, 4); + t4 = lp_build_extract_range(gallivm, newcoords[1], 4*i, 4); + r4 = lp_build_extract_range(gallivm, newcoords[2], 4*i, 4); if (offsets[0]) { offsets4[0] = lp_build_extract_range(gallivm, offsets[0], 4*i, 4); @@ -1870,8 +3366,9 @@ lp_build_sample_soa(struct gallivm_state *gallivm, } } } - lod_ipart4 = lp_build_extract_range(gallivm, lod_ipart, num_lods * i, num_lods); - ilevel04 = lp_build_extract_range(gallivm, ilevel0, num_lods * i, num_lods); + lod_positive4 = lp_build_extract_range(gallivm, lod_positive, num_lods * i, num_lods); + ilevel04 = bld.num_mips == 1 ? ilevel0 : + lp_build_extract_range(gallivm, ilevel0, num_lods * i, num_lods); if (mip_filter == PIPE_TEX_MIPFILTER_LINEAR) { ilevel14 = lp_build_extract_range(gallivm, ilevel1, num_lods * i, num_lods); lod_fpart4 = lp_build_extract_range(gallivm, lod_fpart, num_lods * i, num_lods); @@ -1881,15 +3378,24 @@ lp_build_sample_soa(struct gallivm_state *gallivm, /* do sampling/filtering with fixed pt arithmetic */ lp_build_sample_aos(&bld4, sampler_index, s4, t4, r4, offsets4, - lod_ipart4, lod_fpart4, + lod_positive4, lod_fpart4, ilevel04, ilevel14, texelout4); } else { + /* this path is currently unreachable and hence might break easily... */ + LLVMValueRef newcoords4[5]; + newcoords4[0] = s4; + newcoords4[1] = t4; + newcoords4[2] = r4; + newcoords4[3] = lp_build_extract_range(gallivm, newcoords[3], 4*i, 4); + newcoords4[4] = lp_build_extract_range(gallivm, newcoords[4], 4*i, 4); + lp_build_sample_general(&bld4, sampler_index, - s4, t4, r4, offsets4, - lod_ipart4, lod_fpart4, + op_type == LP_SAMPLER_OP_GATHER, + newcoords4, offsets4, + lod_positive4, lod_fpart4, ilevel04, ilevel14, texelout4); } @@ -1902,11 +3408,9 @@ lp_build_sample_soa(struct gallivm_state *gallivm, texel_out[j] = lp_build_concat(gallivm, texelouttmp[j], type4, num_quads); } } - - lp_build_sample_compare(&bld, coords, texel_out); } - if (static_texture_state->target != PIPE_BUFFER) { + if (target != PIPE_BUFFER && op_type != LP_SAMPLER_OP_GATHER) { apply_sampler_swizzle(&bld, texel_out); } @@ -1923,29 +3427,472 @@ lp_build_sample_soa(struct gallivm_state *gallivm, } } + +#define USE_TEX_FUNC_CALL 1 + +#define LP_MAX_TEX_FUNC_ARGS 32 + +static inline void +get_target_info(enum pipe_texture_target target, + unsigned *num_coords, unsigned *num_derivs, + unsigned *num_offsets, unsigned *layer) +{ + unsigned dims = texture_dims(target); + *num_coords = dims; + *num_offsets = dims; + *num_derivs = (target == PIPE_TEXTURE_CUBE || + target == PIPE_TEXTURE_CUBE_ARRAY) ? 3 : dims; + *layer = has_layer_coord(target) ? 2: 0; + if (target == PIPE_TEXTURE_CUBE_ARRAY) { + /* + * dims doesn't include r coord for cubes - this is handled + * by layer instead, but need to fix up for cube arrays... + */ + *layer = 3; + *num_coords = 3; + } +} + + +/** + * Generate the function body for a texture sampling function. + */ +static void +lp_build_sample_gen_func(struct gallivm_state *gallivm, + const struct lp_static_texture_state *static_texture_state, + const struct lp_static_sampler_state *static_sampler_state, + struct lp_sampler_dynamic_state *dynamic_state, + struct lp_type type, + unsigned texture_index, + unsigned sampler_index, + LLVMValueRef function, + unsigned num_args, + unsigned sample_key) +{ + LLVMBuilderRef old_builder; + LLVMBasicBlockRef block; + LLVMValueRef coords[5]; + LLVMValueRef offsets[3] = { NULL }; + LLVMValueRef lod = NULL; + LLVMValueRef ms_index = NULL; + LLVMValueRef context_ptr; + LLVMValueRef thread_data_ptr = NULL; + LLVMValueRef texel_out[4]; + struct lp_derivatives derivs; + struct lp_derivatives *deriv_ptr = NULL; + unsigned num_param = 0; + unsigned i, num_coords, num_derivs, num_offsets, layer; + enum lp_sampler_lod_control lod_control; + boolean need_cache = FALSE; + + lod_control = (sample_key & LP_SAMPLER_LOD_CONTROL_MASK) >> + LP_SAMPLER_LOD_CONTROL_SHIFT; + + get_target_info(static_texture_state->target, + &num_coords, &num_derivs, &num_offsets, &layer); + + if (dynamic_state->cache_ptr) { + const struct util_format_description *format_desc; + format_desc = util_format_description(static_texture_state->format); + if (format_desc && format_desc->layout == UTIL_FORMAT_LAYOUT_S3TC) { + need_cache = TRUE; + } + } + + /* "unpack" arguments */ + context_ptr = LLVMGetParam(function, num_param++); + if (need_cache) { + thread_data_ptr = LLVMGetParam(function, num_param++); + } + for (i = 0; i < num_coords; i++) { + coords[i] = LLVMGetParam(function, num_param++); + } + for (i = num_coords; i < 5; i++) { + /* This is rather unfortunate... */ + coords[i] = lp_build_undef(gallivm, type); + } + if (layer) { + coords[layer] = LLVMGetParam(function, num_param++); + } + if (sample_key & LP_SAMPLER_SHADOW) { + coords[4] = LLVMGetParam(function, num_param++); + } + if (sample_key & LP_SAMPLER_FETCH_MS) { + ms_index = LLVMGetParam(function, num_param++); + } + if (sample_key & LP_SAMPLER_OFFSETS) { + for (i = 0; i < num_offsets; i++) { + offsets[i] = LLVMGetParam(function, num_param++); + } + } + if (lod_control == LP_SAMPLER_LOD_BIAS || + lod_control == LP_SAMPLER_LOD_EXPLICIT) { + lod = LLVMGetParam(function, num_param++); + } + else if (lod_control == LP_SAMPLER_LOD_DERIVATIVES) { + for (i = 0; i < num_derivs; i++) { + derivs.ddx[i] = LLVMGetParam(function, num_param++); + derivs.ddy[i] = LLVMGetParam(function, num_param++); + } + deriv_ptr = &derivs; + } + + assert(num_args == num_param); + + /* + * Function body + */ + + old_builder = gallivm->builder; + block = LLVMAppendBasicBlockInContext(gallivm->context, function, "entry"); + gallivm->builder = LLVMCreateBuilderInContext(gallivm->context); + LLVMPositionBuilderAtEnd(gallivm->builder, block); + + lp_build_sample_soa_code(gallivm, + static_texture_state, + static_sampler_state, + dynamic_state, + type, + sample_key, + texture_index, + sampler_index, + context_ptr, + thread_data_ptr, + coords, + offsets, + deriv_ptr, + lod, + ms_index, + texel_out); + + LLVMBuildAggregateRet(gallivm->builder, texel_out, 4); + + LLVMDisposeBuilder(gallivm->builder); + gallivm->builder = old_builder; + + gallivm_verify_function(gallivm, function); +} + + +/** + * Call the matching function for texture sampling. + * If there's no match, generate a new one. + */ +static void +lp_build_sample_soa_func(struct gallivm_state *gallivm, + const struct lp_static_texture_state *static_texture_state, + const struct lp_static_sampler_state *static_sampler_state, + struct lp_sampler_dynamic_state *dynamic_state, + const struct lp_sampler_params *params, + int texture_index, int sampler_index, + LLVMValueRef *tex_ret) +{ + LLVMBuilderRef builder = gallivm->builder; + LLVMModuleRef module = LLVMGetGlobalParent(LLVMGetBasicBlockParent( + LLVMGetInsertBlock(builder))); + LLVMValueRef function, inst; + LLVMValueRef args[LP_MAX_TEX_FUNC_ARGS]; + LLVMBasicBlockRef bb; + unsigned num_args = 0; + char func_name[64]; + unsigned i, num_coords, num_derivs, num_offsets, layer; + unsigned sample_key = params->sample_key; + const LLVMValueRef *coords = params->coords; + const LLVMValueRef *offsets = params->offsets; + const struct lp_derivatives *derivs = params->derivs; + enum lp_sampler_lod_control lod_control; + boolean need_cache = FALSE; + + lod_control = (sample_key & LP_SAMPLER_LOD_CONTROL_MASK) >> + LP_SAMPLER_LOD_CONTROL_SHIFT; + + get_target_info(static_texture_state->target, + &num_coords, &num_derivs, &num_offsets, &layer); + + if (dynamic_state->cache_ptr) { + const struct util_format_description *format_desc; + format_desc = util_format_description(static_texture_state->format); + if (format_desc && format_desc->layout == UTIL_FORMAT_LAYOUT_S3TC) { + need_cache = TRUE; + } + } + /* + * texture function matches are found by name. + * Thus the name has to include both the texture and sampler unit + * (which covers all static state) plus the actual texture function + * (including things like offsets, shadow coord, lod control). + * Additionally lod_property has to be included too. + */ + + snprintf(func_name, sizeof(func_name), "texfunc_res_%d_sam_%d_%x", + texture_index, sampler_index, sample_key); + + function = LLVMGetNamedFunction(module, func_name); + + if(!function) { + LLVMTypeRef arg_types[LP_MAX_TEX_FUNC_ARGS]; + LLVMTypeRef ret_type; + LLVMTypeRef function_type; + LLVMTypeRef val_type[4]; + unsigned num_param = 0; + + /* + * Generate the function prototype. + */ + + arg_types[num_param++] = LLVMTypeOf(params->context_ptr); + if (need_cache) { + arg_types[num_param++] = LLVMTypeOf(params->thread_data_ptr); + } + for (i = 0; i < num_coords; i++) { + arg_types[num_param++] = LLVMTypeOf(coords[0]); + assert(LLVMTypeOf(coords[0]) == LLVMTypeOf(coords[i])); + } + if (layer) { + arg_types[num_param++] = LLVMTypeOf(coords[layer]); + assert(LLVMTypeOf(coords[0]) == LLVMTypeOf(coords[layer])); + } + if (sample_key & LP_SAMPLER_SHADOW) { + arg_types[num_param++] = LLVMTypeOf(coords[0]); + } + if (sample_key & LP_SAMPLER_FETCH_MS) { + arg_types[num_param++] = LLVMTypeOf(params->ms_index); + } + if (sample_key & LP_SAMPLER_OFFSETS) { + for (i = 0; i < num_offsets; i++) { + arg_types[num_param++] = LLVMTypeOf(offsets[0]); + assert(LLVMTypeOf(offsets[0]) == LLVMTypeOf(offsets[i])); + } + } + if (lod_control == LP_SAMPLER_LOD_BIAS || + lod_control == LP_SAMPLER_LOD_EXPLICIT) { + arg_types[num_param++] = LLVMTypeOf(params->lod); + } + else if (lod_control == LP_SAMPLER_LOD_DERIVATIVES) { + for (i = 0; i < num_derivs; i++) { + arg_types[num_param++] = LLVMTypeOf(derivs->ddx[i]); + arg_types[num_param++] = LLVMTypeOf(derivs->ddy[i]); + assert(LLVMTypeOf(derivs->ddx[0]) == LLVMTypeOf(derivs->ddx[i])); + assert(LLVMTypeOf(derivs->ddy[0]) == LLVMTypeOf(derivs->ddy[i])); + } + } + + val_type[0] = val_type[1] = val_type[2] = val_type[3] = + lp_build_vec_type(gallivm, params->type); + ret_type = LLVMStructTypeInContext(gallivm->context, val_type, 4, 0); + function_type = LLVMFunctionType(ret_type, arg_types, num_param, 0); + function = LLVMAddFunction(module, func_name, function_type); + + for (i = 0; i < num_param; ++i) { + if(LLVMGetTypeKind(arg_types[i]) == LLVMPointerTypeKind) { + + lp_add_function_attr(function, i + 1, LP_FUNC_ATTR_NOALIAS); + } + } + + LLVMSetFunctionCallConv(function, LLVMFastCallConv); + LLVMSetLinkage(function, LLVMInternalLinkage); + + lp_build_sample_gen_func(gallivm, + static_texture_state, + static_sampler_state, + dynamic_state, + params->type, + texture_index, + sampler_index, + function, + num_param, + sample_key); + } + + num_args = 0; + args[num_args++] = params->context_ptr; + if (need_cache) { + args[num_args++] = params->thread_data_ptr; + } + for (i = 0; i < num_coords; i++) { + args[num_args++] = coords[i]; + } + if (layer) { + args[num_args++] = coords[layer]; + } + if (sample_key & LP_SAMPLER_SHADOW) { + args[num_args++] = coords[4]; + } + if (sample_key & LP_SAMPLER_FETCH_MS) { + args[num_args++] = params->ms_index; + } + if (sample_key & LP_SAMPLER_OFFSETS) { + for (i = 0; i < num_offsets; i++) { + args[num_args++] = offsets[i]; + } + } + if (lod_control == LP_SAMPLER_LOD_BIAS || + lod_control == LP_SAMPLER_LOD_EXPLICIT) { + args[num_args++] = params->lod; + } + else if (lod_control == LP_SAMPLER_LOD_DERIVATIVES) { + for (i = 0; i < num_derivs; i++) { + args[num_args++] = derivs->ddx[i]; + args[num_args++] = derivs->ddy[i]; + } + } + + assert(num_args <= LP_MAX_TEX_FUNC_ARGS); + + *tex_ret = LLVMBuildCall(builder, function, args, num_args, ""); + bb = LLVMGetInsertBlock(builder); + inst = LLVMGetLastInstruction(bb); + LLVMSetInstructionCallConv(inst, LLVMFastCallConv); + +} + + +/** + * Build texture sampling code. + * Either via a function call or inline it directly. + */ +void +lp_build_sample_soa(const struct lp_static_texture_state *static_texture_state, + const struct lp_static_sampler_state *static_sampler_state, + struct lp_sampler_dynamic_state *dynamic_state, + struct gallivm_state *gallivm, + const struct lp_sampler_params *params) +{ + boolean use_tex_func = FALSE; + + /* + * Do not use a function call if the sampling is "simple enough". + * We define this by + * a) format + * b) no mips (either one level only or no mip filter) + * No mips will definitely make the code smaller, though + * the format requirement is a bit iffy - there's some (SoA) formats + * which definitely generate less code. This does happen to catch + * some important cases though which are hurt quite a bit by using + * a call (though not really because of the call overhead but because + * they are reusing the same texture unit with some of the same + * parameters). + * Ideally we'd let llvm recognize this stuff by doing IPO passes. + */ + + if (USE_TEX_FUNC_CALL) { + const struct util_format_description *format_desc; + boolean simple_format; + boolean simple_tex; + enum lp_sampler_op_type op_type; + format_desc = util_format_description(static_texture_state->format); + simple_format = !format_desc || + (util_format_is_rgba8_variant(format_desc) && + format_desc->colorspace == UTIL_FORMAT_COLORSPACE_RGB); + + op_type = (params->sample_key & LP_SAMPLER_OP_TYPE_MASK) >> + LP_SAMPLER_OP_TYPE_SHIFT; + simple_tex = + op_type != LP_SAMPLER_OP_TEXTURE || + ((static_sampler_state->min_mip_filter == PIPE_TEX_MIPFILTER_NONE || + static_texture_state->level_zero_only == TRUE) && + static_sampler_state->min_img_filter == static_sampler_state->mag_img_filter); + + use_tex_func = format_desc && !(simple_format && simple_tex); + } + + if (use_tex_func) { + LLVMValueRef tex_ret; + lp_build_sample_soa_func(gallivm, + static_texture_state, + static_sampler_state, + dynamic_state, + params, params->texture_index, params->sampler_index, &tex_ret); + + for (unsigned i = 0; i < 4; i++) { + params->texel[i] = LLVMBuildExtractValue(gallivm->builder, tex_ret, i, ""); + } + } + else { + lp_build_sample_soa_code(gallivm, + static_texture_state, + static_sampler_state, + dynamic_state, + params->type, + params->sample_key, + params->texture_index, + params->sampler_index, + params->context_ptr, + params->thread_data_ptr, + params->coords, + params->offsets, + params->derivs, + params->lod, + params->ms_index, + params->texel); + } +} + + void lp_build_size_query_soa(struct gallivm_state *gallivm, const struct lp_static_texture_state *static_state, struct lp_sampler_dynamic_state *dynamic_state, - struct lp_type int_type, - unsigned texture_unit, - boolean need_nr_mips, - boolean scalar_lod, - LLVMValueRef explicit_lod, - LLVMValueRef *sizes_out) + const struct lp_sampler_size_query_params *params) { - LLVMValueRef lod; - LLVMValueRef size; + LLVMValueRef lod, level = 0, size; LLVMValueRef first_level = NULL; int dims, i; boolean has_array; - struct lp_build_context bld_int_vec; + unsigned num_lods = 1; + struct lp_build_context bld_int_vec4; + LLVMValueRef context_ptr = params->context_ptr; + unsigned texture_unit = params->texture_unit; + unsigned target = params->target; + LLVMValueRef texture_unit_offset = params->texture_unit_offset; + + if (static_state->format == PIPE_FORMAT_NONE) { + /* + * If there's nothing bound, format is NONE, and we must return + * all zero as mandated by d3d10 in this case. + */ + unsigned chan; + LLVMValueRef zero = lp_build_const_vec(gallivm, params->int_type, 0.0F); + for (chan = 0; chan < 4; chan++) { + params->sizes_out[chan] = zero; + } + return; + } + + /* + * Do some sanity verification about bound texture and shader dcl target. + * Not entirely sure what's possible but assume array/non-array + * always compatible (probably not ok for OpenGL but d3d10 has no + * distinction of arrays at the resource level). + * Everything else looks bogus (though not entirely sure about rect/2d). + * Currently disabled because it causes assertion failures if there's + * nothing bound (or rather a dummy texture, not that this case would + * return the right values). + */ + if (0 && static_state->target != target) { + if (static_state->target == PIPE_TEXTURE_1D) + assert(target == PIPE_TEXTURE_1D_ARRAY); + else if (static_state->target == PIPE_TEXTURE_1D_ARRAY) + assert(target == PIPE_TEXTURE_1D); + else if (static_state->target == PIPE_TEXTURE_2D) + assert(target == PIPE_TEXTURE_2D_ARRAY); + else if (static_state->target == PIPE_TEXTURE_2D_ARRAY) + assert(target == PIPE_TEXTURE_2D); + else if (static_state->target == PIPE_TEXTURE_CUBE) + assert(target == PIPE_TEXTURE_CUBE_ARRAY); + else if (static_state->target == PIPE_TEXTURE_CUBE_ARRAY) + assert(target == PIPE_TEXTURE_CUBE); + else + assert(0); + } - dims = texture_dims(static_state->target); + dims = texture_dims(target); - switch (static_state->target) { + switch (target) { case PIPE_TEXTURE_1D_ARRAY: case PIPE_TEXTURE_2D_ARRAY: + case PIPE_TEXTURE_CUBE_ARRAY: has_array = TRUE; break; default: @@ -1953,67 +3900,110 @@ lp_build_size_query_soa(struct gallivm_state *gallivm, break; } - assert(!int_type.floating); + assert(!params->int_type.floating); - lp_build_context_init(&bld_int_vec, gallivm, lp_type_int_vec(32, 128)); + lp_build_context_init(&bld_int_vec4, gallivm, lp_type_int_vec(32, 128)); - if (explicit_lod) { + if (params->samples_only) { + params->sizes_out[0] = lp_build_broadcast(gallivm, lp_build_vec_type(gallivm, params->int_type), + dynamic_state->num_samples(dynamic_state, gallivm, + context_ptr, texture_unit, + texture_unit_offset)); + return; + } + if (params->explicit_lod) { /* FIXME: this needs to honor per-element lod */ - lod = LLVMBuildExtractElement(gallivm->builder, explicit_lod, lp_build_const_int32(gallivm, 0), ""); - first_level = dynamic_state->first_level(dynamic_state, gallivm, texture_unit); - lod = lp_build_broadcast_scalar(&bld_int_vec, - LLVMBuildAdd(gallivm->builder, lod, first_level, "lod")); - + lod = LLVMBuildExtractElement(gallivm->builder, params->explicit_lod, + lp_build_const_int32(gallivm, 0), ""); + first_level = dynamic_state->first_level(dynamic_state, gallivm, + context_ptr, texture_unit, texture_unit_offset); + level = LLVMBuildAdd(gallivm->builder, lod, first_level, "level"); + lod = lp_build_broadcast_scalar(&bld_int_vec4, level); } else { - lod = bld_int_vec.zero; + lod = bld_int_vec4.zero; } - if (need_nr_mips) { - size = bld_int_vec.zero; - } - else { - size = bld_int_vec.undef; - } + size = bld_int_vec4.undef; size = LLVMBuildInsertElement(gallivm->builder, size, - dynamic_state->width(dynamic_state, gallivm, texture_unit), + dynamic_state->width(dynamic_state, gallivm, + context_ptr, texture_unit, texture_unit_offset), lp_build_const_int32(gallivm, 0), ""); if (dims >= 2) { size = LLVMBuildInsertElement(gallivm->builder, size, - dynamic_state->height(dynamic_state, gallivm, texture_unit), + dynamic_state->height(dynamic_state, gallivm, + context_ptr, texture_unit, texture_unit_offset), lp_build_const_int32(gallivm, 1), ""); } if (dims >= 3) { size = LLVMBuildInsertElement(gallivm->builder, size, - dynamic_state->depth(dynamic_state, gallivm, texture_unit), + dynamic_state->depth(dynamic_state, gallivm, + context_ptr, texture_unit, texture_unit_offset), lp_build_const_int32(gallivm, 2), ""); } - size = lp_build_minify(&bld_int_vec, size, lod); + size = lp_build_minify(&bld_int_vec4, size, lod, TRUE); - if (has_array) - size = LLVMBuildInsertElement(gallivm->builder, size, - dynamic_state->depth(dynamic_state, gallivm, texture_unit), + if (has_array) { + LLVMValueRef layers = dynamic_state->depth(dynamic_state, gallivm, + context_ptr, texture_unit, texture_unit_offset); + if (target == PIPE_TEXTURE_CUBE_ARRAY) { + /* + * It looks like GL wants number of cubes, d3d10.1 has it undefined? + * Could avoid this by passing in number of cubes instead of total + * number of layers (might make things easier elsewhere too). + */ + LLVMValueRef six = lp_build_const_int32(gallivm, 6); + layers = LLVMBuildSDiv(gallivm->builder, layers, six, ""); + } + size = LLVMBuildInsertElement(gallivm->builder, size, layers, lp_build_const_int32(gallivm, dims), ""); + } /* - * XXX for out-of-bounds lod, should set size to zero vector here - * (for dx10-style only, i.e. need_nr_mips) + * d3d10 requires zero for x/y/z values (but not w, i.e. mip levels) + * if level is out of bounds (note this can't cover unbound texture + * here, which also requires returning zero). */ - + if (params->explicit_lod && params->is_sviewinfo) { + LLVMValueRef last_level, out, out1; + struct lp_build_context leveli_bld; + + /* everything is scalar for now */ + lp_build_context_init(&leveli_bld, gallivm, lp_type_int_vec(32, 32)); + last_level = dynamic_state->last_level(dynamic_state, gallivm, + context_ptr, texture_unit, texture_unit_offset); + + out = lp_build_cmp(&leveli_bld, PIPE_FUNC_LESS, level, first_level); + out1 = lp_build_cmp(&leveli_bld, PIPE_FUNC_GREATER, level, last_level); + out = lp_build_or(&leveli_bld, out, out1); + if (num_lods == 1) { + out = lp_build_broadcast_scalar(&bld_int_vec4, out); + } + else { + /* TODO */ + assert(0); + } + size = lp_build_andnot(&bld_int_vec4, size, out); + } for (i = 0; i < dims + (has_array ? 1 : 0); i++) { - sizes_out[i] = lp_build_extract_broadcast(gallivm, bld_int_vec.type, int_type, + params->sizes_out[i] = lp_build_extract_broadcast(gallivm, bld_int_vec4.type, params->int_type, size, lp_build_const_int32(gallivm, i)); } + if (params->is_sviewinfo) { + for (; i < 4; i++) { + params->sizes_out[i] = lp_build_const_vec(gallivm, params->int_type, 0.0); + } + } /* * if there's no explicit_lod (buffers, rects) queries requiring nr of * mips would be illegal. */ - if (need_nr_mips && explicit_lod) { + if (params->is_sviewinfo && params->explicit_lod) { struct lp_build_context bld_int_scalar; LLVMValueRef num_levels; lp_build_context_init(&bld_int_scalar, gallivm, lp_type_int(32)); @@ -2024,11 +4014,306 @@ lp_build_size_query_soa(struct gallivm_state *gallivm, else { LLVMValueRef last_level; - last_level = dynamic_state->last_level(dynamic_state, gallivm, texture_unit); + last_level = dynamic_state->last_level(dynamic_state, gallivm, + context_ptr, texture_unit, texture_unit_offset); num_levels = lp_build_sub(&bld_int_scalar, last_level, first_level); num_levels = lp_build_add(&bld_int_scalar, num_levels, bld_int_scalar.one); } - sizes_out[3] = lp_build_broadcast(gallivm, lp_build_vec_type(gallivm, int_type), + params->sizes_out[3] = lp_build_broadcast(gallivm, lp_build_vec_type(gallivm, params->int_type), num_levels); } } + +static void +lp_build_do_atomic_soa(struct gallivm_state *gallivm, + const struct util_format_description *format_desc, + struct lp_type type, + LLVMValueRef exec_mask, + LLVMValueRef base_ptr, + LLVMValueRef offset, + LLVMValueRef out_of_bounds, + unsigned img_op, + LLVMAtomicRMWBinOp op, + const LLVMValueRef rgba_in[4], + const LLVMValueRef rgba2_in[4], + LLVMValueRef atomic_result[4]) +{ + enum pipe_format format = format_desc->format; + + if (format != PIPE_FORMAT_R32_UINT && format != PIPE_FORMAT_R32_SINT && format != PIPE_FORMAT_R32_FLOAT) { + atomic_result[0] = lp_build_zero(gallivm, type); + return; + } + + LLVMValueRef atom_res = lp_build_alloca(gallivm, + LLVMVectorType(LLVMInt32TypeInContext(gallivm->context), type.length), ""); + + offset = LLVMBuildGEP(gallivm->builder, base_ptr, &offset, 1, ""); + struct lp_build_loop_state loop_state; + lp_build_loop_begin(&loop_state, gallivm, lp_build_const_int32(gallivm, 0)); + struct lp_build_if_state ifthen; + LLVMValueRef cond; + LLVMValueRef packed = rgba_in[0], packed2 = rgba2_in[0]; + + LLVMValueRef should_store_mask = LLVMBuildAnd(gallivm->builder, exec_mask, LLVMBuildNot(gallivm->builder, out_of_bounds, ""), "store_mask"); + assert(exec_mask); + + cond = LLVMBuildICmp(gallivm->builder, LLVMIntNE, should_store_mask, lp_build_const_int_vec(gallivm, type, 0), ""); + cond = LLVMBuildExtractElement(gallivm->builder, cond, loop_state.counter, ""); + lp_build_if(&ifthen, gallivm, cond); + + LLVMValueRef data = LLVMBuildExtractElement(gallivm->builder, packed, loop_state.counter, ""); + LLVMValueRef cast_base_ptr = LLVMBuildExtractElement(gallivm->builder, offset, loop_state.counter, ""); + cast_base_ptr = LLVMBuildBitCast(gallivm->builder, cast_base_ptr, LLVMPointerType(LLVMInt32TypeInContext(gallivm->context), 0), ""); + data = LLVMBuildBitCast(gallivm->builder, data, LLVMInt32TypeInContext(gallivm->context), ""); + + if (img_op == LP_IMG_ATOMIC_CAS) { + LLVMValueRef cas_src_ptr = LLVMBuildExtractElement(gallivm->builder, packed2, loop_state.counter, ""); + LLVMValueRef cas_src = LLVMBuildBitCast(gallivm->builder, cas_src_ptr, LLVMInt32TypeInContext(gallivm->context), ""); + data = LLVMBuildAtomicCmpXchg(gallivm->builder, cast_base_ptr, data, + cas_src, + LLVMAtomicOrderingSequentiallyConsistent, + LLVMAtomicOrderingSequentiallyConsistent, + false); + data = LLVMBuildExtractValue(gallivm->builder, data, 0, ""); + } else { + data = LLVMBuildAtomicRMW(gallivm->builder, op, + cast_base_ptr, data, + LLVMAtomicOrderingSequentiallyConsistent, + false); + } + + LLVMValueRef temp_res = LLVMBuildLoad(gallivm->builder, atom_res, ""); + temp_res = LLVMBuildInsertElement(gallivm->builder, temp_res, data, loop_state.counter, ""); + LLVMBuildStore(gallivm->builder, temp_res, atom_res); + + lp_build_endif(&ifthen); + lp_build_loop_end_cond(&loop_state, lp_build_const_int32(gallivm, type.length), + NULL, LLVMIntUGE); + atomic_result[0] = LLVMBuildLoad(gallivm->builder, atom_res, ""); +} + +static void +lp_build_img_op_no_format(struct gallivm_state *gallivm, + const struct lp_img_params *params, + LLVMValueRef outdata[4]) +{ + /* + * If there's nothing bound, format is NONE, and we must return + * all zero as mandated by d3d10 in this case. + */ + if (params->img_op != LP_IMG_STORE) { + LLVMValueRef zero = lp_build_zero(gallivm, params->type); + for (unsigned chan = 0; chan < (params->img_op == LP_IMG_LOAD ? 4 : 1); chan++) { + outdata[chan] = zero; + } + } +} + +void +lp_build_img_op_soa(const struct lp_static_texture_state *static_texture_state, + struct lp_sampler_dynamic_state *dynamic_state, + struct gallivm_state *gallivm, + const struct lp_img_params *params, + LLVMValueRef outdata[4]) +{ + unsigned target = params->target; + unsigned dims = texture_dims(target); + /** regular scalar int type */ + struct lp_type int_type, int_coord_type; + struct lp_build_context int_bld, int_coord_bld; + const struct util_format_description *format_desc = util_format_description(static_texture_state->format); + LLVMValueRef x = params->coords[0], y = params->coords[1], z = params->coords[2]; + LLVMValueRef ms_index = params->ms_index; + LLVMValueRef row_stride_vec = NULL, img_stride_vec = NULL; + int_type = lp_type_int(32); + int_coord_type = lp_int_type(params->type); + lp_build_context_init(&int_bld, gallivm, int_type); + lp_build_context_init(&int_coord_bld, gallivm, int_coord_type); + + if (static_texture_state->format == PIPE_FORMAT_NONE) { + lp_build_img_op_no_format(gallivm, params, outdata); + return; + } + LLVMValueRef offset, i, j; + + LLVMValueRef row_stride = dynamic_state->row_stride(dynamic_state, gallivm, + params->context_ptr, params->image_index, NULL); + LLVMValueRef img_stride = dynamic_state->img_stride(dynamic_state, gallivm, + params->context_ptr, params->image_index, NULL); + LLVMValueRef base_ptr = dynamic_state->base_ptr(dynamic_state, gallivm, + params->context_ptr, params->image_index, NULL); + LLVMValueRef width = dynamic_state->width(dynamic_state, gallivm, + params->context_ptr, params->image_index, NULL); + LLVMValueRef height = dynamic_state->height(dynamic_state, gallivm, + params->context_ptr, params->image_index, NULL); + LLVMValueRef depth = dynamic_state->depth(dynamic_state, gallivm, + params->context_ptr, params->image_index, NULL); + LLVMValueRef num_samples = NULL, sample_stride = NULL; + if (ms_index) { + num_samples = dynamic_state->num_samples(dynamic_state, gallivm, + params->context_ptr, params->image_index, NULL); + sample_stride = dynamic_state->sample_stride(dynamic_state, gallivm, + params->context_ptr, params->image_index, NULL); + } + + boolean layer_coord = has_layer_coord(target); + + width = lp_build_broadcast_scalar(&int_coord_bld, width); + if (dims >= 2) { + height = lp_build_broadcast_scalar(&int_coord_bld, height); + row_stride_vec = lp_build_broadcast_scalar(&int_coord_bld, row_stride); + } + if (dims >= 3 || layer_coord) { + depth = lp_build_broadcast_scalar(&int_coord_bld, depth); + img_stride_vec = lp_build_broadcast_scalar(&int_coord_bld, img_stride); + } + + LLVMValueRef out_of_bounds = int_coord_bld.zero; + LLVMValueRef out1; + out1 = lp_build_cmp(&int_coord_bld, PIPE_FUNC_GEQUAL, x, width); + out_of_bounds = lp_build_or(&int_coord_bld, out_of_bounds, out1); + + if (dims >= 2) { + out1 = lp_build_cmp(&int_coord_bld, PIPE_FUNC_GEQUAL, y, height); + out_of_bounds = lp_build_or(&int_coord_bld, out_of_bounds, out1); + } + if (dims >= 3) { + out1 = lp_build_cmp(&int_coord_bld, PIPE_FUNC_GEQUAL, z, depth); + out_of_bounds = lp_build_or(&int_coord_bld, out_of_bounds, out1); + } + lp_build_sample_offset(&int_coord_bld, + format_desc, + x, y, z, row_stride_vec, img_stride_vec, + &offset, &i, &j); + + if (ms_index) { + out1 = lp_build_cmp(&int_coord_bld, PIPE_FUNC_GEQUAL, ms_index, lp_build_broadcast_scalar(&int_coord_bld, num_samples)); + out_of_bounds = lp_build_or(&int_coord_bld, out_of_bounds, out1); + + offset = lp_build_add(&int_coord_bld, offset, + lp_build_mul(&int_coord_bld, lp_build_broadcast_scalar(&int_coord_bld, sample_stride), + ms_index)); + } + if (params->img_op == LP_IMG_LOAD) { + struct lp_type texel_type = params->type; + if (format_desc->colorspace == UTIL_FORMAT_COLORSPACE_RGB && + format_desc->channel[0].pure_integer) { + if (format_desc->channel[0].type == UTIL_FORMAT_TYPE_SIGNED) { + texel_type = lp_type_int_vec(params->type.width, params->type.width * params->type.length); + } else if (format_desc->channel[0].type == UTIL_FORMAT_TYPE_UNSIGNED) { + texel_type = lp_type_uint_vec(params->type.width, params->type.width * params->type.length); + } + } + + offset = lp_build_andnot(&int_coord_bld, offset, out_of_bounds); + struct lp_build_context texel_bld; + lp_build_context_init(&texel_bld, gallivm, texel_type); + lp_build_fetch_rgba_soa(gallivm, + format_desc, + texel_type, TRUE, + base_ptr, offset, + i, j, + NULL, + outdata); + + for (unsigned chan = 0; chan < 4; chan++) { + outdata[chan] = lp_build_select(&texel_bld, out_of_bounds, + texel_bld.zero, outdata[chan]); + } + } else if (params->img_op == LP_IMG_STORE) { + lp_build_store_rgba_soa(gallivm, format_desc, params->type, params->exec_mask, base_ptr, offset, out_of_bounds, + params->indata); + } else { + lp_build_do_atomic_soa(gallivm, format_desc, params->type, params->exec_mask, base_ptr, offset, out_of_bounds, + params->img_op, params->op, params->indata, params->indata2, outdata); + } +} + +/* + * These functions are for indirect texture access suppoort. + * + * Indirect textures are implemented using a switch statement, that + * takes the texture index and jumps to the sampler functions for + * that texture unit. + */ + +/* + * Initialise an indexed sampler switch block. + * + * This sets up the switch_info state and adds the LLVM flow control pieces. + */ +void +lp_build_sample_array_init_soa(struct lp_build_sample_array_switch *switch_info, + struct gallivm_state *gallivm, + const struct lp_sampler_params *params, + LLVMValueRef idx, + unsigned base, unsigned range) +{ + switch_info->gallivm = gallivm; + switch_info->params = *params; + switch_info->base = base; + switch_info->range = range; + + /* for generating the switch functions we don't want the texture index offset */ + switch_info->params.texture_index_offset = 0; + + LLVMBasicBlockRef initial_block = LLVMGetInsertBlock(gallivm->builder); + switch_info->merge_ref = lp_build_insert_new_block(gallivm, "texmerge"); + + switch_info->switch_ref = LLVMBuildSwitch(gallivm->builder, idx, + switch_info->merge_ref, range - base); + + LLVMTypeRef val_type[4]; + val_type[0] = val_type[1] = val_type[2] = val_type[3] = + lp_build_vec_type(gallivm, params->type); + LLVMTypeRef ret_type = LLVMStructTypeInContext(gallivm->context, val_type, 4, 0); + + LLVMValueRef undef_val = LLVMGetUndef(ret_type); + + LLVMPositionBuilderAtEnd(gallivm->builder, switch_info->merge_ref); + + switch_info->phi = LLVMBuildPhi(gallivm->builder, ret_type, ""); + LLVMAddIncoming(switch_info->phi, &undef_val, &initial_block, 1); +} + +/* + * Add an individual entry to the indirect texture switch. + * + * This builds the sample function and links a case for it into the switch statement. + */ +void +lp_build_sample_array_case_soa(struct lp_build_sample_array_switch *switch_info, + int idx, + const struct lp_static_texture_state *static_texture_state, + const struct lp_static_sampler_state *static_sampler_state, + struct lp_sampler_dynamic_state *dynamic_texture_state) +{ + struct gallivm_state *gallivm = switch_info->gallivm; + LLVMBasicBlockRef this_block = lp_build_insert_new_block(gallivm, "texblock"); + LLVMValueRef tex_ret; + + LLVMAddCase(switch_info->switch_ref, LLVMConstInt(LLVMInt32TypeInContext(gallivm->context), idx, 0), this_block); + LLVMPositionBuilderAtEnd(gallivm->builder, this_block); + + lp_build_sample_soa_func(gallivm, static_texture_state, + static_sampler_state, dynamic_texture_state, &switch_info->params, idx, idx, + &tex_ret); + + LLVMAddIncoming(switch_info->phi, &tex_ret, &this_block, 1); + LLVMBuildBr(gallivm->builder, switch_info->merge_ref); +} + +/* + * Finish a switch statement. + * + * This handles extract the results from the switch. + */ +void lp_build_sample_array_fini_soa(struct lp_build_sample_array_switch *switch_info) +{ + struct gallivm_state *gallivm = switch_info->gallivm; + + LLVMPositionBuilderAtEnd(gallivm->builder, switch_info->merge_ref); + for (unsigned i = 0; i < 4; i++) + switch_info->params.texel[i] = LLVMBuildExtractValue(gallivm->builder, switch_info->phi, i, ""); +}