1 /**************************************************************************
3 * Copyright 2009 VMware, Inc.
4 * Copyright 2007-2008 Tungsten Graphics, Inc., Cedar Park, Texas.
7 * Permission is hereby granted, free of charge, to any person obtaining a
8 * copy of this software and associated documentation files (the
9 * "Software"), to deal in the Software without restriction, including
10 * without limitation the rights to use, copy, modify, merge, publish,
11 * distribute, sub license, and/or sell copies of the Software, and to
12 * permit persons to whom the Software is furnished to do so, subject to
13 * the following conditions:
15 * The above copyright notice and this permission notice (including the
16 * next paragraph) shall be included in all copies or substantial portions
19 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
20 * OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
21 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT.
22 * IN NO EVENT SHALL TUNGSTEN GRAPHICS AND/OR ITS SUPPLIERS BE LIABLE FOR
23 * ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,
24 * TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
25 * SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
27 **************************************************************************/
31 * TGSI to LLVM IR translation -- SoA.
33 * @author Jose Fonseca <jfonseca@vmware.com>
35 * Based on tgsi_sse2.c code written by Michal Krol, Keith Whitwell,
36 * Brian Paul, and others.
39 #include "pipe/p_config.h"
40 #include "pipe/p_shader_tokens.h"
41 #include "util/u_debug.h"
42 #include "util/u_math.h"
43 #include "util/u_memory.h"
44 #include "tgsi/tgsi_dump.h"
45 #include "tgsi/tgsi_info.h"
46 #include "tgsi/tgsi_parse.h"
47 #include "tgsi/tgsi_util.h"
48 #include "tgsi/tgsi_exec.h"
49 #include "tgsi/tgsi_scan.h"
50 #include "lp_bld_type.h"
51 #include "lp_bld_const.h"
52 #include "lp_bld_arit.h"
53 #include "lp_bld_logic.h"
54 #include "lp_bld_swizzle.h"
55 #include "lp_bld_flow.h"
56 #include "lp_bld_tgsi.h"
57 #include "lp_bld_limits.h"
58 #include "lp_bld_debug.h"
61 #define FOR_EACH_CHANNEL( CHAN )\
62 for (CHAN = 0; CHAN < NUM_CHANNELS; CHAN++)
64 #define IS_DST0_CHANNEL_ENABLED( INST, CHAN )\
65 ((INST)->Dst[0].Register.WriteMask & (1 << (CHAN)))
67 #define IF_IS_DST0_CHANNEL_ENABLED( INST, CHAN )\
68 if (IS_DST0_CHANNEL_ENABLED( INST, CHAN ))
70 #define FOR_EACH_DST0_ENABLED_CHANNEL( INST, CHAN )\
71 FOR_EACH_CHANNEL( CHAN )\
72 IF_IS_DST0_CHANNEL_ENABLED( INST, CHAN )
79 #define QUAD_TOP_LEFT 0
80 #define QUAD_TOP_RIGHT 1
81 #define QUAD_BOTTOM_LEFT 2
82 #define QUAD_BOTTOM_RIGHT 3
86 struct lp_build_context
*bld
;
90 LLVMTypeRef int_vec_type
;
92 LLVMValueRef cond_stack
[LP_MAX_TGSI_NESTING
];
94 LLVMValueRef cond_mask
;
96 LLVMValueRef break_stack
[LP_MAX_TGSI_NESTING
];
98 LLVMValueRef break_mask
;
100 LLVMValueRef cont_stack
[LP_MAX_TGSI_NESTING
];
102 LLVMValueRef cont_mask
;
104 LLVMBasicBlockRef loop_stack
[LP_MAX_TGSI_NESTING
];
106 LLVMBasicBlockRef loop_block
;
109 LLVMValueRef exec_mask
;
112 struct lp_build_tgsi_soa_context
114 struct lp_build_context base
;
116 LLVMValueRef consts_ptr
;
117 const LLVMValueRef
*pos
;
118 const LLVMValueRef (*inputs
)[NUM_CHANNELS
];
119 LLVMValueRef (*outputs
)[NUM_CHANNELS
];
121 struct lp_build_sampler_soa
*sampler
;
123 LLVMValueRef immediates
[LP_MAX_TGSI_IMMEDIATES
][NUM_CHANNELS
];
124 LLVMValueRef temps
[LP_MAX_TGSI_TEMPS
][NUM_CHANNELS
];
125 LLVMValueRef addr
[LP_MAX_TGSI_ADDRS
][NUM_CHANNELS
];
127 /* we allocate an array of temps if we have indirect
128 * addressing and then the temps above is unused */
129 LLVMValueRef temps_array
;
130 boolean has_indirect_addressing
;
132 struct lp_build_mask_context
*mask
;
133 struct lp_exec_mask exec_mask
;
136 static const unsigned char
138 QUAD_TOP_LEFT
, QUAD_TOP_LEFT
,
139 QUAD_BOTTOM_LEFT
, QUAD_BOTTOM_LEFT
142 static const unsigned char
144 QUAD_TOP_RIGHT
, QUAD_TOP_RIGHT
,
145 QUAD_BOTTOM_RIGHT
, QUAD_BOTTOM_RIGHT
148 static const unsigned char
150 QUAD_TOP_LEFT
, QUAD_TOP_RIGHT
,
151 QUAD_TOP_LEFT
, QUAD_TOP_RIGHT
154 static const unsigned char
155 swizzle_bottom
[4] = {
156 QUAD_BOTTOM_LEFT
, QUAD_BOTTOM_RIGHT
,
157 QUAD_BOTTOM_LEFT
, QUAD_BOTTOM_RIGHT
160 static void lp_exec_mask_init(struct lp_exec_mask
*mask
, struct lp_build_context
*bld
)
163 mask
->has_mask
= FALSE
;
164 mask
->cond_stack_size
= 0;
165 mask
->loop_stack_size
= 0;
166 mask
->break_stack_size
= 0;
167 mask
->cont_stack_size
= 0;
169 mask
->int_vec_type
= lp_build_int_vec_type(mask
->bld
->type
);
172 static void lp_exec_mask_update(struct lp_exec_mask
*mask
)
174 if (mask
->loop_stack_size
) {
175 /*for loops we need to update the entire mask at runtime */
177 assert(mask
->break_mask
);
178 tmp
= LLVMBuildAnd(mask
->bld
->builder
,
182 mask
->exec_mask
= LLVMBuildAnd(mask
->bld
->builder
,
187 mask
->exec_mask
= mask
->cond_mask
;
190 mask
->has_mask
= (mask
->cond_stack_size
> 0 ||
191 mask
->loop_stack_size
> 0);
194 static void lp_exec_mask_cond_push(struct lp_exec_mask
*mask
,
197 assert(mask
->cond_stack_size
< LP_MAX_TGSI_NESTING
);
198 mask
->cond_stack
[mask
->cond_stack_size
++] = mask
->cond_mask
;
199 mask
->cond_mask
= LLVMBuildBitCast(mask
->bld
->builder
, val
,
200 mask
->int_vec_type
, "");
202 lp_exec_mask_update(mask
);
205 static void lp_exec_mask_cond_invert(struct lp_exec_mask
*mask
)
207 LLVMValueRef prev_mask
= mask
->cond_stack
[mask
->cond_stack_size
- 1];
208 LLVMValueRef inv_mask
= LLVMBuildNot(mask
->bld
->builder
,
209 mask
->cond_mask
, "");
211 /* means that we didn't have any mask before and that
212 * we were fully enabled */
213 if (mask
->cond_stack_size
<= 1) {
214 prev_mask
= LLVMConstAllOnes(mask
->int_vec_type
);
217 mask
->cond_mask
= LLVMBuildAnd(mask
->bld
->builder
,
220 lp_exec_mask_update(mask
);
223 static void lp_exec_mask_cond_pop(struct lp_exec_mask
*mask
)
225 mask
->cond_mask
= mask
->cond_stack
[--mask
->cond_stack_size
];
226 lp_exec_mask_update(mask
);
229 static void lp_exec_bgnloop(struct lp_exec_mask
*mask
)
232 if (mask
->cont_stack_size
== 0)
233 mask
->cont_mask
= LLVMConstAllOnes(mask
->int_vec_type
);
234 if (mask
->break_stack_size
== 0)
235 mask
->break_mask
= LLVMConstAllOnes(mask
->int_vec_type
);
236 if (mask
->cond_stack_size
== 0)
237 mask
->cond_mask
= LLVMConstAllOnes(mask
->int_vec_type
);
239 assert(mask
->break_stack_size
< LP_MAX_TGSI_NESTING
);
240 assert(mask
->cont_stack_size
< LP_MAX_TGSI_NESTING
);
241 assert(mask
->break_stack_size
< LP_MAX_TGSI_NESTING
);
243 mask
->break_stack
[mask
->break_stack_size
++] = mask
->break_mask
;
244 mask
->cont_stack
[mask
->cont_stack_size
++] = mask
->cont_mask
;
245 mask
->loop_stack
[mask
->loop_stack_size
++] = mask
->loop_block
;
246 mask
->loop_block
= lp_build_insert_new_block(mask
->bld
->builder
, "bgnloop");
247 LLVMBuildBr(mask
->bld
->builder
, mask
->loop_block
);
248 LLVMPositionBuilderAtEnd(mask
->bld
->builder
, mask
->loop_block
);
250 lp_exec_mask_update(mask
);
253 static void lp_exec_break(struct lp_exec_mask
*mask
)
255 LLVMValueRef exec_mask
= LLVMBuildNot(mask
->bld
->builder
,
259 mask
->break_mask
= LLVMBuildAnd(mask
->bld
->builder
,
261 exec_mask
, "break_full");
263 lp_exec_mask_update(mask
);
266 static void lp_exec_continue(struct lp_exec_mask
*mask
)
268 LLVMValueRef exec_mask
= LLVMBuildNot(mask
->bld
->builder
,
272 mask
->cont_mask
= LLVMBuildAnd(mask
->bld
->builder
,
276 lp_exec_mask_update(mask
);
280 static void lp_exec_endloop(struct lp_exec_mask
*mask
)
282 LLVMBasicBlockRef endloop
;
283 LLVMTypeRef reg_type
= LLVMIntType(mask
->bld
->type
.width
*
284 mask
->bld
->type
.length
);
287 assert(mask
->break_mask
);
289 /* i1cond = (mask == 0) */
290 i1cond
= LLVMBuildICmp(
293 LLVMBuildBitCast(mask
->bld
->builder
, mask
->break_mask
, reg_type
, ""),
294 LLVMConstNull(reg_type
), "");
296 endloop
= lp_build_insert_new_block(mask
->bld
->builder
, "endloop");
298 LLVMBuildCondBr(mask
->bld
->builder
,
299 i1cond
, mask
->loop_block
, endloop
);
301 LLVMPositionBuilderAtEnd(mask
->bld
->builder
, endloop
);
303 mask
->loop_block
= mask
->loop_stack
[--mask
->loop_stack_size
];
304 /* pop the cont mask */
305 if (mask
->cont_stack_size
) {
306 mask
->cont_mask
= mask
->cont_stack
[--mask
->cont_stack_size
];
308 /* pop the break mask */
309 if (mask
->break_stack_size
) {
310 mask
->break_mask
= mask
->break_stack
[--mask
->break_stack_size
];
313 lp_exec_mask_update(mask
);
316 /* stores val into an address pointed to by dst.
317 * mask->exec_mask is used to figure out which bits of val
318 * should be stored into the address
319 * (0 means don't store this bit, 1 means do store).
321 static void lp_exec_mask_store(struct lp_exec_mask
*mask
,
325 if (mask
->has_mask
) {
326 LLVMValueRef real_val
, dst_val
;
328 dst_val
= LLVMBuildLoad(mask
->bld
->builder
, dst
, "");
329 real_val
= lp_build_select(mask
->bld
,
333 LLVMBuildStore(mask
->bld
->builder
, real_val
, dst
);
335 LLVMBuildStore(mask
->bld
->builder
, val
, dst
);
340 emit_ddx(struct lp_build_tgsi_soa_context
*bld
,
343 LLVMValueRef src_left
= lp_build_swizzle1_aos(&bld
->base
, src
, swizzle_left
);
344 LLVMValueRef src_right
= lp_build_swizzle1_aos(&bld
->base
, src
, swizzle_right
);
345 return lp_build_sub(&bld
->base
, src_right
, src_left
);
350 emit_ddy(struct lp_build_tgsi_soa_context
*bld
,
353 LLVMValueRef src_top
= lp_build_swizzle1_aos(&bld
->base
, src
, swizzle_top
);
354 LLVMValueRef src_bottom
= lp_build_swizzle1_aos(&bld
->base
, src
, swizzle_bottom
);
355 return lp_build_sub(&bld
->base
, src_top
, src_bottom
);
359 get_temp_ptr(struct lp_build_tgsi_soa_context
*bld
,
365 if (!bld
->has_indirect_addressing
) {
366 return bld
->temps
[index
][swizzle
];
368 LLVMValueRef lindex
=
369 LLVMConstInt(LLVMInt32Type(), index
*4 + swizzle
, 0);
371 lindex
= lp_build_add(&bld
->base
, lindex
, addr
);
372 return LLVMBuildGEP(bld
->base
.builder
, bld
->temps_array
, &lindex
, 1, "");
381 struct lp_build_tgsi_soa_context
*bld
,
382 const struct tgsi_full_instruction
*inst
,
384 const unsigned chan_index
)
386 const struct tgsi_full_src_register
*reg
= &inst
->Src
[index
];
387 unsigned swizzle
= tgsi_util_get_full_src_register_swizzle( reg
, chan_index
);
397 if (reg
->Register
.Indirect
) {
398 LLVMTypeRef int_vec_type
= lp_build_int_vec_type(bld
->base
.type
);
399 unsigned swizzle
= tgsi_util_get_src_register_swizzle( ®
->Indirect
, chan_index
);
400 addr
= LLVMBuildLoad(bld
->base
.builder
,
401 bld
->addr
[reg
->Indirect
.Index
][swizzle
],
403 /* for indexing we want integers */
404 addr
= LLVMBuildFPToSI(bld
->base
.builder
, addr
,
406 addr
= LLVMBuildExtractElement(bld
->base
.builder
,
407 addr
, LLVMConstInt(LLVMInt32Type(), 0, 0),
409 addr
= lp_build_mul(&bld
->base
, addr
, LLVMConstInt(LLVMInt32Type(), 4, 0));
412 switch (reg
->Register
.File
) {
413 case TGSI_FILE_CONSTANT
: {
414 LLVMValueRef index
= LLVMConstInt(LLVMInt32Type(), reg
->Register
.Index
*4 + swizzle
, 0);
415 LLVMValueRef scalar
, scalar_ptr
;
417 if (reg
->Register
.Indirect
) {
418 /*lp_build_printf(bld->base.builder,
419 "\taddr = %d\n", addr);*/
420 index
= lp_build_add(&bld
->base
, index
, addr
);
422 scalar_ptr
= LLVMBuildGEP(bld
->base
.builder
, bld
->consts_ptr
, &index
, 1, "");
423 scalar
= LLVMBuildLoad(bld
->base
.builder
, scalar_ptr
, "");
425 res
= lp_build_broadcast_scalar(&bld
->base
, scalar
);
429 case TGSI_FILE_IMMEDIATE
:
430 res
= bld
->immediates
[reg
->Register
.Index
][swizzle
];
434 case TGSI_FILE_INPUT
:
435 res
= bld
->inputs
[reg
->Register
.Index
][swizzle
];
439 case TGSI_FILE_TEMPORARY
: {
440 LLVMValueRef temp_ptr
= get_temp_ptr(bld
, reg
->Register
.Index
,
442 reg
->Register
.Indirect
,
444 res
= LLVMBuildLoad(bld
->base
.builder
, temp_ptr
, "");
446 return bld
->base
.undef
;
452 return bld
->base
.undef
;
458 return bld
->base
.undef
;
461 switch( tgsi_util_get_full_src_register_sign_mode( reg
, chan_index
) ) {
462 case TGSI_UTIL_SIGN_CLEAR
:
463 res
= lp_build_abs( &bld
->base
, res
);
466 case TGSI_UTIL_SIGN_SET
:
467 /* TODO: Use bitwese OR for floating point */
468 res
= lp_build_abs( &bld
->base
, res
);
469 res
= LLVMBuildNeg( bld
->base
.builder
, res
, "" );
472 case TGSI_UTIL_SIGN_TOGGLE
:
473 res
= LLVMBuildNeg( bld
->base
.builder
, res
, "" );
476 case TGSI_UTIL_SIGN_KEEP
:
485 * Register fetch with derivatives.
489 struct lp_build_tgsi_soa_context
*bld
,
490 const struct tgsi_full_instruction
*inst
,
492 const unsigned chan_index
,
499 src
= emit_fetch(bld
, inst
, index
, chan_index
);
504 /* TODO: use interpolation coeffs for inputs */
507 *ddx
= emit_ddx(bld
, src
);
510 *ddy
= emit_ddy(bld
, src
);
519 struct lp_build_tgsi_soa_context
*bld
,
520 const struct tgsi_full_instruction
*inst
,
525 const struct tgsi_full_dst_register
*reg
= &inst
->Dst
[index
];
528 switch( inst
->Instruction
.Saturate
) {
532 case TGSI_SAT_ZERO_ONE
:
533 value
= lp_build_max(&bld
->base
, value
, bld
->base
.zero
);
534 value
= lp_build_min(&bld
->base
, value
, bld
->base
.one
);
537 case TGSI_SAT_MINUS_PLUS_ONE
:
538 value
= lp_build_max(&bld
->base
, value
, lp_build_const_vec(bld
->base
.type
, -1.0));
539 value
= lp_build_min(&bld
->base
, value
, bld
->base
.one
);
546 if (reg
->Register
.Indirect
) {
547 LLVMTypeRef int_vec_type
= lp_build_int_vec_type(bld
->base
.type
);
548 unsigned swizzle
= tgsi_util_get_src_register_swizzle( ®
->Indirect
, chan_index
);
549 addr
= LLVMBuildLoad(bld
->base
.builder
,
550 bld
->addr
[reg
->Indirect
.Index
][swizzle
],
552 /* for indexing we want integers */
553 addr
= LLVMBuildFPToSI(bld
->base
.builder
, addr
,
555 addr
= LLVMBuildExtractElement(bld
->base
.builder
,
556 addr
, LLVMConstInt(LLVMInt32Type(), 0, 0),
558 addr
= lp_build_mul(&bld
->base
, addr
, LLVMConstInt(LLVMInt32Type(), 4, 0));
561 switch( reg
->Register
.File
) {
562 case TGSI_FILE_OUTPUT
:
563 lp_exec_mask_store(&bld
->exec_mask
, value
,
564 bld
->outputs
[reg
->Register
.Index
][chan_index
]);
567 case TGSI_FILE_TEMPORARY
: {
568 LLVMValueRef temp_ptr
= get_temp_ptr(bld
, reg
->Register
.Index
,
570 reg
->Register
.Indirect
,
572 lp_exec_mask_store(&bld
->exec_mask
, value
, temp_ptr
);
576 case TGSI_FILE_ADDRESS
:
577 lp_exec_mask_store(&bld
->exec_mask
, value
,
578 bld
->addr
[reg
->Indirect
.Index
][chan_index
]);
581 case TGSI_FILE_PREDICATE
:
592 * High-level instruction translators.
596 TEX_MODIFIER_NONE
= 0,
597 TEX_MODIFIER_PROJECTED
,
598 TEX_MODIFIER_LOD_BIAS
,
599 TEX_MODIFIER_EXPLICIT_LOD
,
600 TEX_MODIFIER_EXPLICIT_DERIV
604 emit_tex( struct lp_build_tgsi_soa_context
*bld
,
605 const struct tgsi_full_instruction
*inst
,
606 enum tex_modifier modifier
,
610 LLVMValueRef lod_bias
, explicit_lod
;
611 LLVMValueRef oow
= NULL
;
612 LLVMValueRef coords
[3];
619 _debug_printf("warning: found texture instruction but no sampler generator supplied\n");
620 for (i
= 0; i
< 4; i
++) {
621 texel
[i
] = bld
->base
.undef
;
626 switch (inst
->Texture
.Texture
) {
627 case TGSI_TEXTURE_1D
:
630 case TGSI_TEXTURE_2D
:
631 case TGSI_TEXTURE_RECT
:
634 case TGSI_TEXTURE_SHADOW1D
:
635 case TGSI_TEXTURE_SHADOW2D
:
636 case TGSI_TEXTURE_SHADOWRECT
:
637 case TGSI_TEXTURE_3D
:
638 case TGSI_TEXTURE_CUBE
:
646 if (modifier
== TEX_MODIFIER_LOD_BIAS
) {
647 lod_bias
= emit_fetch( bld
, inst
, 0, 3 );
650 else if (modifier
== TEX_MODIFIER_EXPLICIT_LOD
) {
652 explicit_lod
= emit_fetch( bld
, inst
, 0, 3 );
659 if (modifier
== TEX_MODIFIER_PROJECTED
) {
660 oow
= emit_fetch( bld
, inst
, 0, 3 );
661 oow
= lp_build_rcp(&bld
->base
, oow
);
664 for (i
= 0; i
< num_coords
; i
++) {
665 coords
[i
] = emit_fetch( bld
, inst
, 0, i
);
666 if (modifier
== TEX_MODIFIER_PROJECTED
)
667 coords
[i
] = lp_build_mul(&bld
->base
, coords
[i
], oow
);
669 for (i
= num_coords
; i
< 3; i
++) {
670 coords
[i
] = bld
->base
.undef
;
673 if (modifier
== TEX_MODIFIER_EXPLICIT_DERIV
) {
674 for (i
= 0; i
< num_coords
; i
++) {
675 ddx
[i
] = emit_fetch( bld
, inst
, 1, i
);
676 ddy
[i
] = emit_fetch( bld
, inst
, 2, i
);
678 unit
= inst
->Src
[3].Register
.Index
;
680 for (i
= 0; i
< num_coords
; i
++) {
681 ddx
[i
] = emit_ddx( bld
, coords
[i
] );
682 ddy
[i
] = emit_ddy( bld
, coords
[i
] );
684 unit
= inst
->Src
[1].Register
.Index
;
686 for (i
= num_coords
; i
< 3; i
++) {
687 ddx
[i
] = bld
->base
.undef
;
688 ddy
[i
] = bld
->base
.undef
;
691 bld
->sampler
->emit_fetch_texel(bld
->sampler
,
694 unit
, num_coords
, coords
,
696 lod_bias
, explicit_lod
,
702 * Kill fragment if any of the src register values are negative.
706 struct lp_build_tgsi_soa_context
*bld
,
707 const struct tgsi_full_instruction
*inst
)
709 const struct tgsi_full_src_register
*reg
= &inst
->Src
[0];
710 LLVMValueRef terms
[NUM_CHANNELS
];
714 memset(&terms
, 0, sizeof terms
);
716 FOR_EACH_CHANNEL( chan_index
) {
719 /* Unswizzle channel */
720 swizzle
= tgsi_util_get_full_src_register_swizzle( reg
, chan_index
);
722 /* Check if the component has not been already tested. */
723 assert(swizzle
< NUM_CHANNELS
);
724 if( !terms
[swizzle
] )
725 /* TODO: change the comparison operator instead of setting the sign */
726 terms
[swizzle
] = emit_fetch(bld
, inst
, 0, chan_index
);
730 FOR_EACH_CHANNEL( chan_index
) {
731 if(terms
[chan_index
]) {
732 LLVMValueRef chan_mask
;
735 * If term < 0 then mask = 0 else mask = ~0.
737 chan_mask
= lp_build_cmp(&bld
->base
, PIPE_FUNC_GEQUAL
, terms
[chan_index
], bld
->base
.zero
);
740 mask
= LLVMBuildAnd(bld
->base
.builder
, mask
, chan_mask
, "");
747 lp_build_mask_update(bld
->mask
, mask
);
752 * Predicated fragment kill.
753 * XXX Actually, we do an unconditional kill (as in tgsi_exec.c).
754 * The only predication is the execution mask which will apply if
755 * we're inside a loop or conditional.
758 emit_kilp(struct lp_build_tgsi_soa_context
*bld
,
759 const struct tgsi_full_instruction
*inst
)
763 /* For those channels which are "alive", disable fragment shader
766 if (bld
->exec_mask
.has_mask
) {
767 mask
= LLVMBuildNot(bld
->base
.builder
, bld
->exec_mask
.exec_mask
, "kilp");
770 mask
= bld
->base
.zero
;
773 lp_build_mask_update(bld
->mask
, mask
);
778 struct lp_build_tgsi_soa_context
*bld
,
779 const struct tgsi_full_declaration
*decl
)
781 LLVMTypeRef vec_type
= lp_build_vec_type(bld
->base
.type
);
783 unsigned first
= decl
->Range
.First
;
784 unsigned last
= decl
->Range
.Last
;
787 for (idx
= first
; idx
<= last
; ++idx
) {
788 switch (decl
->Declaration
.File
) {
789 case TGSI_FILE_TEMPORARY
:
790 assert(idx
< LP_MAX_TGSI_TEMPS
);
791 if (bld
->has_indirect_addressing
) {
792 LLVMValueRef val
= LLVMConstInt(LLVMInt32Type(),
794 bld
->temps_array
= lp_build_array_alloca(bld
->base
.builder
,
797 for (i
= 0; i
< NUM_CHANNELS
; i
++)
798 bld
->temps
[idx
][i
] = lp_build_alloca(bld
->base
.builder
,
803 case TGSI_FILE_OUTPUT
:
804 for (i
= 0; i
< NUM_CHANNELS
; i
++)
805 bld
->outputs
[idx
][i
] = lp_build_alloca(bld
->base
.builder
,
809 case TGSI_FILE_ADDRESS
:
810 assert(idx
< LP_MAX_TGSI_ADDRS
);
811 for (i
= 0; i
< NUM_CHANNELS
; i
++)
812 bld
->addr
[idx
][i
] = lp_build_alloca(bld
->base
.builder
,
816 case TGSI_FILE_PREDICATE
:
817 _debug_printf("warning: predicate registers not yet implemented\n");
821 /* don't need to declare other vars */
829 * Emit LLVM for one TGSI instruction.
830 * \param return TRUE for success, FALSE otherwise
834 struct lp_build_tgsi_soa_context
*bld
,
835 const struct tgsi_full_instruction
*inst
,
836 const struct tgsi_opcode_info
*info
)
839 LLVMValueRef src0
, src1
, src2
;
840 LLVMValueRef tmp0
, tmp1
, tmp2
;
841 LLVMValueRef tmp3
= NULL
;
842 LLVMValueRef tmp4
= NULL
;
843 LLVMValueRef tmp5
= NULL
;
844 LLVMValueRef tmp6
= NULL
;
845 LLVMValueRef tmp7
= NULL
;
847 LLVMValueRef dst0
[NUM_CHANNELS
];
850 * Stores and write masks are handled in a general fashion after the long
851 * instruction opcode switch statement.
853 * Although not stricitly necessary, we avoid generating instructions for
854 * channels which won't be stored, in cases where's that easy. For some
855 * complex instructions, like texture sampling, it is more convenient to
856 * assume a full writemask and then let LLVM optimization passes eliminate
860 assert(info
->num_dst
<= 1);
862 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
863 dst0
[chan_index
] = bld
->base
.undef
;
867 switch (inst
->Instruction
.Opcode
) {
868 case TGSI_OPCODE_ARL
:
869 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
870 tmp0
= emit_fetch( bld
, inst
, 0, chan_index
);
871 tmp0
= lp_build_floor(&bld
->base
, tmp0
);
872 dst0
[chan_index
] = tmp0
;
876 case TGSI_OPCODE_MOV
:
877 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
878 dst0
[chan_index
] = emit_fetch( bld
, inst
, 0, chan_index
);
882 case TGSI_OPCODE_LIT
:
883 if( IS_DST0_CHANNEL_ENABLED( inst
, CHAN_X
) ) {
884 dst0
[CHAN_X
] = bld
->base
.one
;
886 if( IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Y
) ) {
887 src0
= emit_fetch( bld
, inst
, 0, CHAN_X
);
888 dst0
[CHAN_Y
] = lp_build_max( &bld
->base
, src0
, bld
->base
.zero
);
890 if( IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Z
) ) {
891 /* XMM[1] = SrcReg[0].yyyy */
892 tmp1
= emit_fetch( bld
, inst
, 0, CHAN_Y
);
893 /* XMM[1] = max(XMM[1], 0) */
894 tmp1
= lp_build_max( &bld
->base
, tmp1
, bld
->base
.zero
);
895 /* XMM[2] = SrcReg[0].wwww */
896 tmp2
= emit_fetch( bld
, inst
, 0, CHAN_W
);
897 tmp1
= lp_build_pow( &bld
->base
, tmp1
, tmp2
);
898 tmp0
= emit_fetch( bld
, inst
, 0, CHAN_X
);
899 tmp2
= lp_build_cmp(&bld
->base
, PIPE_FUNC_GREATER
, tmp0
, bld
->base
.zero
);
900 dst0
[CHAN_Z
] = lp_build_select(&bld
->base
, tmp2
, tmp1
, bld
->base
.zero
);
902 if( IS_DST0_CHANNEL_ENABLED( inst
, CHAN_W
) ) {
903 dst0
[CHAN_W
] = bld
->base
.one
;
907 case TGSI_OPCODE_RCP
:
908 /* TGSI_OPCODE_RECIP */
909 src0
= emit_fetch( bld
, inst
, 0, CHAN_X
);
910 res
= lp_build_rcp(&bld
->base
, src0
);
911 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
912 dst0
[chan_index
] = res
;
916 case TGSI_OPCODE_RSQ
:
917 /* TGSI_OPCODE_RECIPSQRT */
918 src0
= emit_fetch( bld
, inst
, 0, CHAN_X
);
919 src0
= lp_build_abs(&bld
->base
, src0
);
920 res
= lp_build_rsqrt(&bld
->base
, src0
);
921 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
922 dst0
[chan_index
] = res
;
926 case TGSI_OPCODE_EXP
:
927 if (IS_DST0_CHANNEL_ENABLED( inst
, CHAN_X
) ||
928 IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Y
) ||
929 IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Z
)) {
930 LLVMValueRef
*p_exp2_int_part
= NULL
;
931 LLVMValueRef
*p_frac_part
= NULL
;
932 LLVMValueRef
*p_exp2
= NULL
;
934 src0
= emit_fetch( bld
, inst
, 0, CHAN_X
);
936 if (IS_DST0_CHANNEL_ENABLED( inst
, CHAN_X
))
937 p_exp2_int_part
= &tmp0
;
938 if (IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Y
))
940 if (IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Z
))
943 lp_build_exp2_approx(&bld
->base
, src0
, p_exp2_int_part
, p_frac_part
, p_exp2
);
945 if (IS_DST0_CHANNEL_ENABLED( inst
, CHAN_X
))
947 if (IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Y
))
949 if (IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Z
))
953 if (IS_DST0_CHANNEL_ENABLED( inst
, CHAN_W
)) {
954 dst0
[CHAN_W
] = bld
->base
.one
;
958 case TGSI_OPCODE_LOG
:
959 if (IS_DST0_CHANNEL_ENABLED( inst
, CHAN_X
) ||
960 IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Y
) ||
961 IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Z
)) {
962 LLVMValueRef
*p_floor_log2
= NULL
;
963 LLVMValueRef
*p_exp
= NULL
;
964 LLVMValueRef
*p_log2
= NULL
;
966 src0
= emit_fetch( bld
, inst
, 0, CHAN_X
);
967 src0
= lp_build_abs( &bld
->base
, src0
);
969 if (IS_DST0_CHANNEL_ENABLED( inst
, CHAN_X
))
970 p_floor_log2
= &tmp0
;
971 if (IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Y
))
973 if (IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Z
))
976 lp_build_log2_approx(&bld
->base
, src0
, p_exp
, p_floor_log2
, p_log2
);
978 /* dst.x = floor(lg2(abs(src.x))) */
979 if (IS_DST0_CHANNEL_ENABLED( inst
, CHAN_X
))
981 /* dst.y = abs(src)/ex2(floor(lg2(abs(src.x)))) */
982 if (IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Y
)) {
983 dst0
[CHAN_Y
] = lp_build_div( &bld
->base
, src0
, tmp1
);
985 /* dst.z = lg2(abs(src.x)) */
986 if (IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Z
))
990 if (IS_DST0_CHANNEL_ENABLED( inst
, CHAN_W
)) {
991 dst0
[CHAN_W
] = bld
->base
.one
;
995 case TGSI_OPCODE_MUL
:
996 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
997 src0
= emit_fetch( bld
, inst
, 0, chan_index
);
998 src1
= emit_fetch( bld
, inst
, 1, chan_index
);
999 dst0
[chan_index
] = lp_build_mul(&bld
->base
, src0
, src1
);
1003 case TGSI_OPCODE_ADD
:
1004 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1005 src0
= emit_fetch( bld
, inst
, 0, chan_index
);
1006 src1
= emit_fetch( bld
, inst
, 1, chan_index
);
1007 dst0
[chan_index
] = lp_build_add(&bld
->base
, src0
, src1
);
1011 case TGSI_OPCODE_DP3
:
1012 /* TGSI_OPCODE_DOT3 */
1013 tmp0
= emit_fetch( bld
, inst
, 0, CHAN_X
);
1014 tmp1
= emit_fetch( bld
, inst
, 1, CHAN_X
);
1015 tmp0
= lp_build_mul( &bld
->base
, tmp0
, tmp1
);
1016 tmp1
= emit_fetch( bld
, inst
, 0, CHAN_Y
);
1017 tmp2
= emit_fetch( bld
, inst
, 1, CHAN_Y
);
1018 tmp1
= lp_build_mul( &bld
->base
, tmp1
, tmp2
);
1019 tmp0
= lp_build_add( &bld
->base
, tmp0
, tmp1
);
1020 tmp1
= emit_fetch( bld
, inst
, 0, CHAN_Z
);
1021 tmp2
= emit_fetch( bld
, inst
, 1, CHAN_Z
);
1022 tmp1
= lp_build_mul( &bld
->base
, tmp1
, tmp2
);
1023 tmp0
= lp_build_add( &bld
->base
, tmp0
, tmp1
);
1024 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1025 dst0
[chan_index
] = tmp0
;
1029 case TGSI_OPCODE_DP4
:
1030 /* TGSI_OPCODE_DOT4 */
1031 tmp0
= emit_fetch( bld
, inst
, 0, CHAN_X
);
1032 tmp1
= emit_fetch( bld
, inst
, 1, CHAN_X
);
1033 tmp0
= lp_build_mul( &bld
->base
, tmp0
, tmp1
);
1034 tmp1
= emit_fetch( bld
, inst
, 0, CHAN_Y
);
1035 tmp2
= emit_fetch( bld
, inst
, 1, CHAN_Y
);
1036 tmp1
= lp_build_mul( &bld
->base
, tmp1
, tmp2
);
1037 tmp0
= lp_build_add( &bld
->base
, tmp0
, tmp1
);
1038 tmp1
= emit_fetch( bld
, inst
, 0, CHAN_Z
);
1039 tmp2
= emit_fetch( bld
, inst
, 1, CHAN_Z
);
1040 tmp1
= lp_build_mul( &bld
->base
, tmp1
, tmp2
);
1041 tmp0
= lp_build_add( &bld
->base
, tmp0
, tmp1
);
1042 tmp1
= emit_fetch( bld
, inst
, 0, CHAN_W
);
1043 tmp2
= emit_fetch( bld
, inst
, 1, CHAN_W
);
1044 tmp1
= lp_build_mul( &bld
->base
, tmp1
, tmp2
);
1045 tmp0
= lp_build_add( &bld
->base
, tmp0
, tmp1
);
1046 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1047 dst0
[chan_index
] = tmp0
;
1051 case TGSI_OPCODE_DST
:
1052 IF_IS_DST0_CHANNEL_ENABLED( inst
, CHAN_X
) {
1053 dst0
[CHAN_X
] = bld
->base
.one
;
1055 IF_IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Y
) {
1056 tmp0
= emit_fetch( bld
, inst
, 0, CHAN_Y
);
1057 tmp1
= emit_fetch( bld
, inst
, 1, CHAN_Y
);
1058 dst0
[CHAN_Y
] = lp_build_mul( &bld
->base
, tmp0
, tmp1
);
1060 IF_IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Z
) {
1061 dst0
[CHAN_Z
] = emit_fetch( bld
, inst
, 0, CHAN_Z
);
1063 IF_IS_DST0_CHANNEL_ENABLED( inst
, CHAN_W
) {
1064 dst0
[CHAN_W
] = emit_fetch( bld
, inst
, 1, CHAN_W
);
1068 case TGSI_OPCODE_MIN
:
1069 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1070 src0
= emit_fetch( bld
, inst
, 0, chan_index
);
1071 src1
= emit_fetch( bld
, inst
, 1, chan_index
);
1072 dst0
[chan_index
] = lp_build_min( &bld
->base
, src0
, src1
);
1076 case TGSI_OPCODE_MAX
:
1077 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1078 src0
= emit_fetch( bld
, inst
, 0, chan_index
);
1079 src1
= emit_fetch( bld
, inst
, 1, chan_index
);
1080 dst0
[chan_index
] = lp_build_max( &bld
->base
, src0
, src1
);
1084 case TGSI_OPCODE_SLT
:
1085 /* TGSI_OPCODE_SETLT */
1086 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1087 src0
= emit_fetch( bld
, inst
, 0, chan_index
);
1088 src1
= emit_fetch( bld
, inst
, 1, chan_index
);
1089 tmp0
= lp_build_cmp( &bld
->base
, PIPE_FUNC_LESS
, src0
, src1
);
1090 dst0
[chan_index
] = lp_build_select( &bld
->base
, tmp0
, bld
->base
.one
, bld
->base
.zero
);
1094 case TGSI_OPCODE_SGE
:
1095 /* TGSI_OPCODE_SETGE */
1096 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1097 src0
= emit_fetch( bld
, inst
, 0, chan_index
);
1098 src1
= emit_fetch( bld
, inst
, 1, chan_index
);
1099 tmp0
= lp_build_cmp( &bld
->base
, PIPE_FUNC_GEQUAL
, src0
, src1
);
1100 dst0
[chan_index
] = lp_build_select( &bld
->base
, tmp0
, bld
->base
.one
, bld
->base
.zero
);
1104 case TGSI_OPCODE_MAD
:
1105 /* TGSI_OPCODE_MADD */
1106 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1107 tmp0
= emit_fetch( bld
, inst
, 0, chan_index
);
1108 tmp1
= emit_fetch( bld
, inst
, 1, chan_index
);
1109 tmp2
= emit_fetch( bld
, inst
, 2, chan_index
);
1110 tmp0
= lp_build_mul( &bld
->base
, tmp0
, tmp1
);
1111 tmp0
= lp_build_add( &bld
->base
, tmp0
, tmp2
);
1112 dst0
[chan_index
] = tmp0
;
1116 case TGSI_OPCODE_SUB
:
1117 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1118 tmp0
= emit_fetch( bld
, inst
, 0, chan_index
);
1119 tmp1
= emit_fetch( bld
, inst
, 1, chan_index
);
1120 dst0
[chan_index
] = lp_build_sub( &bld
->base
, tmp0
, tmp1
);
1124 case TGSI_OPCODE_LRP
:
1125 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1126 src0
= emit_fetch( bld
, inst
, 0, chan_index
);
1127 src1
= emit_fetch( bld
, inst
, 1, chan_index
);
1128 src2
= emit_fetch( bld
, inst
, 2, chan_index
);
1129 tmp0
= lp_build_sub( &bld
->base
, src1
, src2
);
1130 tmp0
= lp_build_mul( &bld
->base
, src0
, tmp0
);
1131 dst0
[chan_index
] = lp_build_add( &bld
->base
, tmp0
, src2
);
1135 case TGSI_OPCODE_CND
:
1136 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1137 src0
= emit_fetch( bld
, inst
, 0, chan_index
);
1138 src1
= emit_fetch( bld
, inst
, 1, chan_index
);
1139 src2
= emit_fetch( bld
, inst
, 2, chan_index
);
1140 tmp1
= lp_build_const_vec(bld
->base
.type
, 0.5);
1141 tmp0
= lp_build_cmp( &bld
->base
, PIPE_FUNC_GREATER
, src2
, tmp1
);
1142 dst0
[chan_index
] = lp_build_select( &bld
->base
, tmp0
, src0
, src1
);
1146 case TGSI_OPCODE_DP2A
:
1147 tmp0
= emit_fetch( bld
, inst
, 0, CHAN_X
); /* xmm0 = src[0].x */
1148 tmp1
= emit_fetch( bld
, inst
, 1, CHAN_X
); /* xmm1 = src[1].x */
1149 tmp0
= lp_build_mul( &bld
->base
, tmp0
, tmp1
); /* xmm0 = xmm0 * xmm1 */
1150 tmp1
= emit_fetch( bld
, inst
, 0, CHAN_Y
); /* xmm1 = src[0].y */
1151 tmp2
= emit_fetch( bld
, inst
, 1, CHAN_Y
); /* xmm2 = src[1].y */
1152 tmp1
= lp_build_mul( &bld
->base
, tmp1
, tmp2
); /* xmm1 = xmm1 * xmm2 */
1153 tmp0
= lp_build_add( &bld
->base
, tmp0
, tmp1
); /* xmm0 = xmm0 + xmm1 */
1154 tmp1
= emit_fetch( bld
, inst
, 2, CHAN_X
); /* xmm1 = src[2].x */
1155 tmp0
= lp_build_add( &bld
->base
, tmp0
, tmp1
); /* xmm0 = xmm0 + xmm1 */
1156 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1157 dst0
[chan_index
] = tmp0
; /* dest[ch] = xmm0 */
1161 case TGSI_OPCODE_FRC
:
1162 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1163 src0
= emit_fetch( bld
, inst
, 0, chan_index
);
1164 tmp0
= lp_build_floor(&bld
->base
, src0
);
1165 tmp0
= lp_build_sub(&bld
->base
, src0
, tmp0
);
1166 dst0
[chan_index
] = tmp0
;
1170 case TGSI_OPCODE_CLAMP
:
1171 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1172 tmp0
= emit_fetch( bld
, inst
, 0, chan_index
);
1173 src1
= emit_fetch( bld
, inst
, 1, chan_index
);
1174 src2
= emit_fetch( bld
, inst
, 2, chan_index
);
1175 tmp0
= lp_build_max(&bld
->base
, tmp0
, src1
);
1176 tmp0
= lp_build_min(&bld
->base
, tmp0
, src2
);
1177 dst0
[chan_index
] = tmp0
;
1181 case TGSI_OPCODE_FLR
:
1182 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1183 tmp0
= emit_fetch( bld
, inst
, 0, chan_index
);
1184 dst0
[chan_index
] = lp_build_floor(&bld
->base
, tmp0
);
1188 case TGSI_OPCODE_ROUND
:
1189 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1190 tmp0
= emit_fetch( bld
, inst
, 0, chan_index
);
1191 dst0
[chan_index
] = lp_build_round(&bld
->base
, tmp0
);
1195 case TGSI_OPCODE_EX2
: {
1196 tmp0
= emit_fetch( bld
, inst
, 0, CHAN_X
);
1197 tmp0
= lp_build_exp2( &bld
->base
, tmp0
);
1198 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1199 dst0
[chan_index
] = tmp0
;
1204 case TGSI_OPCODE_LG2
:
1205 tmp0
= emit_fetch( bld
, inst
, 0, CHAN_X
);
1206 tmp0
= lp_build_log2( &bld
->base
, tmp0
);
1207 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1208 dst0
[chan_index
] = tmp0
;
1212 case TGSI_OPCODE_POW
:
1213 src0
= emit_fetch( bld
, inst
, 0, CHAN_X
);
1214 src1
= emit_fetch( bld
, inst
, 1, CHAN_X
);
1215 res
= lp_build_pow( &bld
->base
, src0
, src1
);
1216 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1217 dst0
[chan_index
] = res
;
1221 case TGSI_OPCODE_XPD
:
1222 if( IS_DST0_CHANNEL_ENABLED( inst
, CHAN_X
) ||
1223 IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Y
) ) {
1224 tmp1
= emit_fetch( bld
, inst
, 1, CHAN_Z
);
1225 tmp3
= emit_fetch( bld
, inst
, 0, CHAN_Z
);
1227 if( IS_DST0_CHANNEL_ENABLED( inst
, CHAN_X
) ||
1228 IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Z
) ) {
1229 tmp0
= emit_fetch( bld
, inst
, 0, CHAN_Y
);
1230 tmp4
= emit_fetch( bld
, inst
, 1, CHAN_Y
);
1232 IF_IS_DST0_CHANNEL_ENABLED( inst
, CHAN_X
) {
1234 tmp2
= lp_build_mul( &bld
->base
, tmp2
, tmp1
);
1236 tmp5
= lp_build_mul( &bld
->base
, tmp5
, tmp4
);
1237 tmp2
= lp_build_sub( &bld
->base
, tmp2
, tmp5
);
1238 dst0
[CHAN_X
] = tmp2
;
1240 if( IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Y
) ||
1241 IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Z
) ) {
1242 tmp2
= emit_fetch( bld
, inst
, 1, CHAN_X
);
1243 tmp5
= emit_fetch( bld
, inst
, 0, CHAN_X
);
1245 IF_IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Y
) {
1246 tmp3
= lp_build_mul( &bld
->base
, tmp3
, tmp2
);
1247 tmp1
= lp_build_mul( &bld
->base
, tmp1
, tmp5
);
1248 tmp3
= lp_build_sub( &bld
->base
, tmp3
, tmp1
);
1249 dst0
[CHAN_Y
] = tmp3
;
1251 IF_IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Z
) {
1252 tmp5
= lp_build_mul( &bld
->base
, tmp5
, tmp4
);
1253 tmp0
= lp_build_mul( &bld
->base
, tmp0
, tmp2
);
1254 tmp5
= lp_build_sub( &bld
->base
, tmp5
, tmp0
);
1255 dst0
[CHAN_Z
] = tmp5
;
1257 IF_IS_DST0_CHANNEL_ENABLED( inst
, CHAN_W
) {
1258 dst0
[CHAN_W
] = bld
->base
.one
;
1262 case TGSI_OPCODE_ABS
:
1263 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1264 tmp0
= emit_fetch( bld
, inst
, 0, chan_index
);
1265 dst0
[chan_index
] = lp_build_abs( &bld
->base
, tmp0
);
1269 case TGSI_OPCODE_RCC
:
1274 case TGSI_OPCODE_DPH
:
1275 tmp0
= emit_fetch( bld
, inst
, 0, CHAN_X
);
1276 tmp1
= emit_fetch( bld
, inst
, 1, CHAN_X
);
1277 tmp0
= lp_build_mul( &bld
->base
, tmp0
, tmp1
);
1278 tmp1
= emit_fetch( bld
, inst
, 0, CHAN_Y
);
1279 tmp2
= emit_fetch( bld
, inst
, 1, CHAN_Y
);
1280 tmp1
= lp_build_mul( &bld
->base
, tmp1
, tmp2
);
1281 tmp0
= lp_build_add( &bld
->base
, tmp0
, tmp1
);
1282 tmp1
= emit_fetch( bld
, inst
, 0, CHAN_Z
);
1283 tmp2
= emit_fetch( bld
, inst
, 1, CHAN_Z
);
1284 tmp1
= lp_build_mul( &bld
->base
, tmp1
, tmp2
);
1285 tmp0
= lp_build_add( &bld
->base
, tmp0
, tmp1
);
1286 tmp1
= emit_fetch( bld
, inst
, 1, CHAN_W
);
1287 tmp0
= lp_build_add( &bld
->base
, tmp0
, tmp1
);
1288 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1289 dst0
[chan_index
] = tmp0
;
1293 case TGSI_OPCODE_COS
:
1294 tmp0
= emit_fetch( bld
, inst
, 0, CHAN_X
);
1295 tmp0
= lp_build_cos( &bld
->base
, tmp0
);
1296 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1297 dst0
[chan_index
] = tmp0
;
1301 case TGSI_OPCODE_DDX
:
1302 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1303 emit_fetch_deriv( bld
, inst
, 0, chan_index
, NULL
, &dst0
[chan_index
], NULL
);
1307 case TGSI_OPCODE_DDY
:
1308 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1309 emit_fetch_deriv( bld
, inst
, 0, chan_index
, NULL
, NULL
, &dst0
[chan_index
]);
1313 case TGSI_OPCODE_KILP
:
1314 /* predicated kill */
1315 emit_kilp( bld
, inst
);
1318 case TGSI_OPCODE_KIL
:
1319 /* conditional kill */
1320 emit_kil( bld
, inst
);
1323 case TGSI_OPCODE_PK2H
:
1327 case TGSI_OPCODE_PK2US
:
1331 case TGSI_OPCODE_PK4B
:
1335 case TGSI_OPCODE_PK4UB
:
1339 case TGSI_OPCODE_RFL
:
1343 case TGSI_OPCODE_SEQ
:
1344 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1345 src0
= emit_fetch( bld
, inst
, 0, chan_index
);
1346 src1
= emit_fetch( bld
, inst
, 1, chan_index
);
1347 tmp0
= lp_build_cmp( &bld
->base
, PIPE_FUNC_EQUAL
, src0
, src1
);
1348 dst0
[chan_index
] = lp_build_select( &bld
->base
, tmp0
, bld
->base
.one
, bld
->base
.zero
);
1352 case TGSI_OPCODE_SFL
:
1353 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1354 dst0
[chan_index
] = bld
->base
.zero
;
1358 case TGSI_OPCODE_SGT
:
1359 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1360 src0
= emit_fetch( bld
, inst
, 0, chan_index
);
1361 src1
= emit_fetch( bld
, inst
, 1, chan_index
);
1362 tmp0
= lp_build_cmp( &bld
->base
, PIPE_FUNC_GREATER
, src0
, src1
);
1363 dst0
[chan_index
] = lp_build_select( &bld
->base
, tmp0
, bld
->base
.one
, bld
->base
.zero
);
1367 case TGSI_OPCODE_SIN
:
1368 tmp0
= emit_fetch( bld
, inst
, 0, CHAN_X
);
1369 tmp0
= lp_build_sin( &bld
->base
, tmp0
);
1370 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1371 dst0
[chan_index
] = tmp0
;
1375 case TGSI_OPCODE_SLE
:
1376 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1377 src0
= emit_fetch( bld
, inst
, 0, chan_index
);
1378 src1
= emit_fetch( bld
, inst
, 1, chan_index
);
1379 tmp0
= lp_build_cmp( &bld
->base
, PIPE_FUNC_LEQUAL
, src0
, src1
);
1380 dst0
[chan_index
] = lp_build_select( &bld
->base
, tmp0
, bld
->base
.one
, bld
->base
.zero
);
1384 case TGSI_OPCODE_SNE
:
1385 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1386 src0
= emit_fetch( bld
, inst
, 0, chan_index
);
1387 src1
= emit_fetch( bld
, inst
, 1, chan_index
);
1388 tmp0
= lp_build_cmp( &bld
->base
, PIPE_FUNC_NOTEQUAL
, src0
, src1
);
1389 dst0
[chan_index
] = lp_build_select( &bld
->base
, tmp0
, bld
->base
.one
, bld
->base
.zero
);
1393 case TGSI_OPCODE_STR
:
1394 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1395 dst0
[chan_index
] = bld
->base
.one
;
1399 case TGSI_OPCODE_TEX
:
1400 emit_tex( bld
, inst
, TEX_MODIFIER_NONE
, dst0
);
1403 case TGSI_OPCODE_TXD
:
1404 emit_tex( bld
, inst
, TEX_MODIFIER_EXPLICIT_DERIV
, dst0
);
1407 case TGSI_OPCODE_UP2H
:
1413 case TGSI_OPCODE_UP2US
:
1419 case TGSI_OPCODE_UP4B
:
1425 case TGSI_OPCODE_UP4UB
:
1431 case TGSI_OPCODE_X2D
:
1437 case TGSI_OPCODE_ARA
:
1443 case TGSI_OPCODE_ARR
:
1444 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1445 tmp0
= emit_fetch( bld
, inst
, 0, chan_index
);
1446 tmp0
= lp_build_round(&bld
->base
, tmp0
);
1447 dst0
[chan_index
] = tmp0
;
1451 case TGSI_OPCODE_BRA
:
1457 case TGSI_OPCODE_CAL
:
1462 case TGSI_OPCODE_RET
:
1467 case TGSI_OPCODE_END
:
1470 case TGSI_OPCODE_SSG
:
1471 /* TGSI_OPCODE_SGN */
1472 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1473 tmp0
= emit_fetch( bld
, inst
, 0, chan_index
);
1474 dst0
[chan_index
] = lp_build_sgn( &bld
->base
, tmp0
);
1478 case TGSI_OPCODE_CMP
:
1479 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1480 src0
= emit_fetch( bld
, inst
, 0, chan_index
);
1481 src1
= emit_fetch( bld
, inst
, 1, chan_index
);
1482 src2
= emit_fetch( bld
, inst
, 2, chan_index
);
1483 tmp0
= lp_build_cmp( &bld
->base
, PIPE_FUNC_LESS
, src0
, bld
->base
.zero
);
1484 dst0
[chan_index
] = lp_build_select( &bld
->base
, tmp0
, src1
, src2
);
1488 case TGSI_OPCODE_SCS
:
1489 IF_IS_DST0_CHANNEL_ENABLED( inst
, CHAN_X
) {
1490 tmp0
= emit_fetch( bld
, inst
, 0, CHAN_X
);
1491 dst0
[CHAN_X
] = lp_build_cos( &bld
->base
, tmp0
);
1493 IF_IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Y
) {
1494 tmp0
= emit_fetch( bld
, inst
, 0, CHAN_X
);
1495 dst0
[CHAN_Y
] = lp_build_sin( &bld
->base
, tmp0
);
1497 IF_IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Z
) {
1498 dst0
[CHAN_Z
] = bld
->base
.zero
;
1500 IF_IS_DST0_CHANNEL_ENABLED( inst
, CHAN_W
) {
1501 dst0
[CHAN_W
] = bld
->base
.one
;
1505 case TGSI_OPCODE_TXB
:
1506 emit_tex( bld
, inst
, TEX_MODIFIER_LOD_BIAS
, dst0
);
1509 case TGSI_OPCODE_NRM
:
1511 case TGSI_OPCODE_NRM4
:
1512 /* 3 or 4-component normalization */
1514 uint dims
= (inst
->Instruction
.Opcode
== TGSI_OPCODE_NRM
) ? 3 : 4;
1516 if (IS_DST0_CHANNEL_ENABLED(inst
, CHAN_X
) ||
1517 IS_DST0_CHANNEL_ENABLED(inst
, CHAN_Y
) ||
1518 IS_DST0_CHANNEL_ENABLED(inst
, CHAN_Z
) ||
1519 (IS_DST0_CHANNEL_ENABLED(inst
, CHAN_W
) && dims
== 4)) {
1521 /* NOTE: Cannot use xmm regs 2/3 here (see emit_rsqrt() above). */
1524 /* xmm0 = src.x * src.x */
1525 tmp0
= emit_fetch(bld
, inst
, 0, CHAN_X
);
1526 if (IS_DST0_CHANNEL_ENABLED(inst
, CHAN_X
)) {
1529 tmp0
= lp_build_mul( &bld
->base
, tmp0
, tmp0
);
1532 /* xmm0 = xmm0 + src.y * src.y */
1533 tmp1
= emit_fetch(bld
, inst
, 0, CHAN_Y
);
1534 if (IS_DST0_CHANNEL_ENABLED(inst
, CHAN_Y
)) {
1537 tmp1
= lp_build_mul( &bld
->base
, tmp1
, tmp1
);
1538 tmp0
= lp_build_add( &bld
->base
, tmp0
, tmp1
);
1541 /* xmm0 = xmm0 + src.z * src.z */
1542 tmp1
= emit_fetch(bld
, inst
, 0, CHAN_Z
);
1543 if (IS_DST0_CHANNEL_ENABLED(inst
, CHAN_Z
)) {
1546 tmp1
= lp_build_mul( &bld
->base
, tmp1
, tmp1
);
1547 tmp0
= lp_build_add( &bld
->base
, tmp0
, tmp1
);
1551 /* xmm0 = xmm0 + src.w * src.w */
1552 tmp1
= emit_fetch(bld
, inst
, 0, CHAN_W
);
1553 if (IS_DST0_CHANNEL_ENABLED(inst
, CHAN_W
)) {
1556 tmp1
= lp_build_mul( &bld
->base
, tmp1
, tmp1
);
1557 tmp0
= lp_build_add( &bld
->base
, tmp0
, tmp1
);
1560 /* xmm1 = 1 / sqrt(xmm0) */
1561 tmp1
= lp_build_rsqrt( &bld
->base
, tmp0
);
1563 /* dst.x = xmm1 * src.x */
1564 if (IS_DST0_CHANNEL_ENABLED(inst
, CHAN_X
)) {
1565 dst0
[CHAN_X
] = lp_build_mul( &bld
->base
, tmp4
, tmp1
);
1568 /* dst.y = xmm1 * src.y */
1569 if (IS_DST0_CHANNEL_ENABLED(inst
, CHAN_Y
)) {
1570 dst0
[CHAN_Y
] = lp_build_mul( &bld
->base
, tmp5
, tmp1
);
1573 /* dst.z = xmm1 * src.z */
1574 if (IS_DST0_CHANNEL_ENABLED(inst
, CHAN_Z
)) {
1575 dst0
[CHAN_Z
] = lp_build_mul( &bld
->base
, tmp6
, tmp1
);
1578 /* dst.w = xmm1 * src.w */
1579 if (IS_DST0_CHANNEL_ENABLED(inst
, CHAN_X
) && dims
== 4) {
1580 dst0
[CHAN_W
] = lp_build_mul( &bld
->base
, tmp7
, tmp1
);
1585 if (IS_DST0_CHANNEL_ENABLED(inst
, CHAN_W
) && dims
== 3) {
1586 dst0
[CHAN_W
] = bld
->base
.one
;
1591 case TGSI_OPCODE_DIV
:
1597 case TGSI_OPCODE_DP2
:
1598 tmp0
= emit_fetch( bld
, inst
, 0, CHAN_X
); /* xmm0 = src[0].x */
1599 tmp1
= emit_fetch( bld
, inst
, 1, CHAN_X
); /* xmm1 = src[1].x */
1600 tmp0
= lp_build_mul( &bld
->base
, tmp0
, tmp1
); /* xmm0 = xmm0 * xmm1 */
1601 tmp1
= emit_fetch( bld
, inst
, 0, CHAN_Y
); /* xmm1 = src[0].y */
1602 tmp2
= emit_fetch( bld
, inst
, 1, CHAN_Y
); /* xmm2 = src[1].y */
1603 tmp1
= lp_build_mul( &bld
->base
, tmp1
, tmp2
); /* xmm1 = xmm1 * xmm2 */
1604 tmp0
= lp_build_add( &bld
->base
, tmp0
, tmp1
); /* xmm0 = xmm0 + xmm1 */
1605 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1606 dst0
[chan_index
] = tmp0
; /* dest[ch] = xmm0 */
1610 case TGSI_OPCODE_TXL
:
1611 emit_tex( bld
, inst
, TEX_MODIFIER_EXPLICIT_LOD
, dst0
);
1614 case TGSI_OPCODE_TXP
:
1615 emit_tex( bld
, inst
, TEX_MODIFIER_PROJECTED
, dst0
);
1618 case TGSI_OPCODE_BRK
:
1619 lp_exec_break(&bld
->exec_mask
);
1622 case TGSI_OPCODE_IF
:
1623 tmp0
= emit_fetch(bld
, inst
, 0, CHAN_X
);
1624 tmp0
= lp_build_cmp(&bld
->base
, PIPE_FUNC_NOTEQUAL
,
1625 tmp0
, bld
->base
.zero
);
1626 lp_exec_mask_cond_push(&bld
->exec_mask
, tmp0
);
1629 case TGSI_OPCODE_BGNLOOP
:
1630 lp_exec_bgnloop(&bld
->exec_mask
);
1633 case TGSI_OPCODE_ELSE
:
1634 lp_exec_mask_cond_invert(&bld
->exec_mask
);
1637 case TGSI_OPCODE_ENDIF
:
1638 lp_exec_mask_cond_pop(&bld
->exec_mask
);
1641 case TGSI_OPCODE_ENDLOOP
:
1642 lp_exec_endloop(&bld
->exec_mask
);
1645 case TGSI_OPCODE_PUSHA
:
1651 case TGSI_OPCODE_POPA
:
1657 case TGSI_OPCODE_CEIL
:
1658 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1659 tmp0
= emit_fetch( bld
, inst
, 0, chan_index
);
1660 dst0
[chan_index
] = lp_build_ceil(&bld
->base
, tmp0
);
1664 case TGSI_OPCODE_I2F
:
1670 case TGSI_OPCODE_NOT
:
1676 case TGSI_OPCODE_TRUNC
:
1677 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1678 tmp0
= emit_fetch( bld
, inst
, 0, chan_index
);
1679 dst0
[chan_index
] = lp_build_trunc(&bld
->base
, tmp0
);
1683 case TGSI_OPCODE_SHL
:
1689 case TGSI_OPCODE_ISHR
:
1695 case TGSI_OPCODE_AND
:
1701 case TGSI_OPCODE_OR
:
1707 case TGSI_OPCODE_MOD
:
1713 case TGSI_OPCODE_XOR
:
1719 case TGSI_OPCODE_SAD
:
1725 case TGSI_OPCODE_TXF
:
1731 case TGSI_OPCODE_TXQ
:
1737 case TGSI_OPCODE_CONT
:
1738 lp_exec_continue(&bld
->exec_mask
);
1741 case TGSI_OPCODE_EMIT
:
1745 case TGSI_OPCODE_ENDPRIM
:
1749 case TGSI_OPCODE_NOP
:
1757 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1758 emit_store( bld
, inst
, 0, chan_index
, dst0
[chan_index
]);
1767 lp_build_tgsi_soa(LLVMBuilderRef builder
,
1768 const struct tgsi_token
*tokens
,
1769 struct lp_type type
,
1770 struct lp_build_mask_context
*mask
,
1771 LLVMValueRef consts_ptr
,
1772 const LLVMValueRef
*pos
,
1773 const LLVMValueRef (*inputs
)[NUM_CHANNELS
],
1774 LLVMValueRef (*outputs
)[NUM_CHANNELS
],
1775 struct lp_build_sampler_soa
*sampler
,
1776 struct tgsi_shader_info
*info
)
1778 struct lp_build_tgsi_soa_context bld
;
1779 struct tgsi_parse_context parse
;
1780 uint num_immediates
= 0;
1783 /* Setup build context */
1784 memset(&bld
, 0, sizeof bld
);
1785 lp_build_context_init(&bld
.base
, builder
, type
);
1788 bld
.inputs
= inputs
;
1789 bld
.outputs
= outputs
;
1790 bld
.consts_ptr
= consts_ptr
;
1791 bld
.sampler
= sampler
;
1792 bld
.has_indirect_addressing
= info
->opcode_count
[TGSI_OPCODE_ARR
] > 0 ||
1793 info
->opcode_count
[TGSI_OPCODE_ARL
] > 0;
1795 lp_exec_mask_init(&bld
.exec_mask
, &bld
.base
);
1797 tgsi_parse_init( &parse
, tokens
);
1799 while( !tgsi_parse_end_of_tokens( &parse
) ) {
1800 tgsi_parse_token( &parse
);
1802 switch( parse
.FullToken
.Token
.Type
) {
1803 case TGSI_TOKEN_TYPE_DECLARATION
:
1804 /* Inputs already interpolated */
1805 emit_declaration( &bld
, &parse
.FullToken
.FullDeclaration
);
1808 case TGSI_TOKEN_TYPE_INSTRUCTION
:
1810 unsigned opcode
= parse
.FullToken
.FullInstruction
.Instruction
.Opcode
;
1811 const struct tgsi_opcode_info
*opcode_info
= tgsi_get_opcode_info(opcode
);
1812 if (!emit_instruction( &bld
, &parse
.FullToken
.FullInstruction
, opcode_info
))
1813 _debug_printf("warning: failed to translate tgsi opcode %s to LLVM\n",
1814 opcode_info
->mnemonic
);
1819 case TGSI_TOKEN_TYPE_IMMEDIATE
:
1820 /* simply copy the immediate values into the next immediates[] slot */
1822 const uint size
= parse
.FullToken
.FullImmediate
.Immediate
.NrTokens
- 1;
1824 assert(num_immediates
< LP_MAX_TGSI_IMMEDIATES
);
1825 for( i
= 0; i
< size
; ++i
)
1826 bld
.immediates
[num_immediates
][i
] =
1827 lp_build_const_vec(type
, parse
.FullToken
.FullImmediate
.u
[i
].Float
);
1828 for( i
= size
; i
< 4; ++i
)
1829 bld
.immediates
[num_immediates
][i
] = bld
.base
.undef
;
1834 case TGSI_TOKEN_TYPE_PROPERTY
:
1842 LLVMBasicBlockRef block
= LLVMGetInsertBlock(builder
);
1843 LLVMValueRef function
= LLVMGetBasicBlockParent(block
);
1844 debug_printf("11111111111111111111111111111 \n");
1845 tgsi_dump(tokens
, 0);
1846 LLVMDumpValue(function
);
1847 debug_printf("2222222222222222222222222222 \n");
1849 tgsi_parse_free( &parse
);