1 /**************************************************************************
3 * Copyright 2009 VMware, Inc.
4 * Copyright 2007-2008 Tungsten Graphics, Inc., Cedar Park, Texas.
7 * Permission is hereby granted, free of charge, to any person obtaining a
8 * copy of this software and associated documentation files (the
9 * "Software"), to deal in the Software without restriction, including
10 * without limitation the rights to use, copy, modify, merge, publish,
11 * distribute, sub license, and/or sell copies of the Software, and to
12 * permit persons to whom the Software is furnished to do so, subject to
13 * the following conditions:
15 * The above copyright notice and this permission notice (including the
16 * next paragraph) shall be included in all copies or substantial portions
19 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
20 * OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
21 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT.
22 * IN NO EVENT SHALL TUNGSTEN GRAPHICS AND/OR ITS SUPPLIERS BE LIABLE FOR
23 * ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,
24 * TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
25 * SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
27 **************************************************************************/
31 * TGSI to LLVM IR translation -- SoA.
33 * @author Jose Fonseca <jfonseca@vmware.com>
35 * Based on tgsi_sse2.c code written by Michal Krol, Keith Whitwell,
36 * Brian Paul, and others.
39 #include "pipe/p_config.h"
40 #include "pipe/p_shader_tokens.h"
41 #include "cso_cache/cso_hash.h"
42 #include "util/u_debug.h"
43 #include "util/u_math.h"
44 #include "util/u_memory.h"
45 #include "util/u_string.h"
46 #include "tgsi/tgsi_dump.h"
47 #include "tgsi/tgsi_info.h"
48 #include "tgsi/tgsi_parse.h"
49 #include "tgsi/tgsi_util.h"
50 #include "tgsi/tgsi_exec.h"
51 #include "tgsi/tgsi_scan.h"
52 #include "lp_bld_type.h"
53 #include "lp_bld_const.h"
54 #include "lp_bld_arit.h"
55 #include "lp_bld_logic.h"
56 #include "lp_bld_swizzle.h"
57 #include "lp_bld_flow.h"
58 #include "lp_bld_tgsi.h"
59 #include "lp_bld_limits.h"
60 #include "lp_bld_debug.h"
63 #define FOR_EACH_CHANNEL( CHAN )\
64 for (CHAN = 0; CHAN < NUM_CHANNELS; CHAN++)
66 #define IS_DST0_CHANNEL_ENABLED( INST, CHAN )\
67 ((INST)->Dst[0].Register.WriteMask & (1 << (CHAN)))
69 #define IF_IS_DST0_CHANNEL_ENABLED( INST, CHAN )\
70 if (IS_DST0_CHANNEL_ENABLED( INST, CHAN ))
72 #define FOR_EACH_DST0_ENABLED_CHANNEL( INST, CHAN )\
73 FOR_EACH_CHANNEL( CHAN )\
74 IF_IS_DST0_CHANNEL_ENABLED( INST, CHAN )
81 #define QUAD_TOP_LEFT 0
82 #define QUAD_TOP_RIGHT 1
83 #define QUAD_BOTTOM_LEFT 2
84 #define QUAD_BOTTOM_RIGHT 3
88 struct lp_build_context
*bld
;
92 LLVMTypeRef int_vec_type
;
94 LLVMValueRef cond_stack
[LP_MAX_TGSI_NESTING
];
96 LLVMValueRef cond_mask
;
98 LLVMBasicBlockRef loop_block
;
99 LLVMValueRef cont_mask
;
100 LLVMValueRef break_mask
;
101 LLVMValueRef break_var
;
102 LLVMValueRef ret_mask
;
104 LLVMBasicBlockRef loop_block
;
105 LLVMValueRef cont_mask
;
106 LLVMValueRef break_mask
;
107 LLVMValueRef break_var
;
108 } loop_stack
[LP_MAX_TGSI_NESTING
];
111 LLVMValueRef exec_mask
;
114 struct lp_build_tgsi_soa_context
116 struct lp_build_context base
;
118 /* Builder for integer masks and indices */
119 struct lp_build_context int_bld
;
121 LLVMValueRef consts_ptr
;
122 const LLVMValueRef
*pos
;
124 const struct lp_build_sampler_soa
*sampler
;
126 const LLVMValueRef (*inputs
)[NUM_CHANNELS
];
127 LLVMValueRef (*outputs
)[NUM_CHANNELS
];
128 LLVMValueRef immediates
[LP_MAX_TGSI_IMMEDIATES
][NUM_CHANNELS
];
129 LLVMValueRef temps
[LP_MAX_TGSI_TEMPS
][NUM_CHANNELS
];
130 LLVMValueRef addr
[LP_MAX_TGSI_ADDRS
][NUM_CHANNELS
];
131 LLVMValueRef preds
[LP_MAX_TGSI_PREDS
][NUM_CHANNELS
];
133 /* we allocate an array of allocas if we have indirect
134 * addressing and then the temps above is unused */
135 LLVMValueRef temps_array
;
137 LLVMValueRef inputs_array
;
138 LLVMValueRef outputs_array
;
139 LLVMValueRef immediates_array
;
140 LLVMValueRef addrs_array
;
141 LLVMValueRef preds_array
;
143 boolean has_indirect_addressing
;
144 boolean has_function_calls
;
146 struct lp_build_mask_context
*mask
;
147 struct lp_exec_mask exec_mask
;
149 struct cso_hash
*func_hash
;
151 LLVMBasicBlockRef main_block
;
154 struct tgsi_declaration_range inputs
;
155 struct tgsi_declaration_range outputs
;
156 struct tgsi_declaration_range temps
;
157 struct tgsi_declaration_range addrs
;
158 struct tgsi_declaration_range preds
;
162 static const unsigned char
164 QUAD_TOP_LEFT
, QUAD_TOP_LEFT
,
165 QUAD_BOTTOM_LEFT
, QUAD_BOTTOM_LEFT
168 static const unsigned char
170 QUAD_TOP_RIGHT
, QUAD_TOP_RIGHT
,
171 QUAD_BOTTOM_RIGHT
, QUAD_BOTTOM_RIGHT
174 static const unsigned char
176 QUAD_TOP_LEFT
, QUAD_TOP_RIGHT
,
177 QUAD_TOP_LEFT
, QUAD_TOP_RIGHT
180 static const unsigned char
181 swizzle_bottom
[4] = {
182 QUAD_BOTTOM_LEFT
, QUAD_BOTTOM_RIGHT
,
183 QUAD_BOTTOM_LEFT
, QUAD_BOTTOM_RIGHT
188 get_temp_ptr(struct lp_build_tgsi_soa_context
*bld
,
195 if (!bld
->has_indirect_addressing
&&
196 !bld
->has_function_calls
) {
197 return bld
->temps
[index
][chan
];
199 LLVMValueRef lindex
=
200 LLVMConstInt(LLVMInt32Type(), index
* 4 + chan
, 0);
202 lindex
= lp_build_add(&bld
->base
, lindex
, addr
);
203 return LLVMBuildGEP(bld
->base
.builder
, bld
->temps_array
, &lindex
, 1, "temp_ptr");
208 get_input_ptr(struct lp_build_tgsi_soa_context
*bld
,
214 LLVMValueRef lindex
=
215 LLVMConstInt(LLVMInt32Type(), index
*4 + swizzle
, 0);
216 assert(bld
->has_function_calls
);
218 lindex
= lp_build_add(&bld
->base
, lindex
, addr
);
219 return LLVMBuildGEP(bld
->base
.builder
, bld
->inputs_array
, &lindex
, 1, "input_ptr");
223 get_output_ptr(struct lp_build_tgsi_soa_context
*bld
,
229 if (!bld
->has_function_calls
) {
230 return bld
->outputs
[index
][swizzle
];
232 LLVMValueRef lindex
=
233 LLVMConstInt(LLVMInt32Type(), index
*4 + swizzle
, 0);
235 lindex
= lp_build_add(&bld
->base
, lindex
, addr
);
236 return LLVMBuildGEP(bld
->base
.builder
, bld
->outputs_array
, &lindex
, 1, "output_ptr");
241 get_immediates_ptr(struct lp_build_tgsi_soa_context
*bld
,
247 LLVMValueRef lindex
=
248 LLVMConstInt(LLVMInt32Type(), index
*4 + swizzle
, 0);
249 assert(bld
->has_function_calls
);
251 lindex
= lp_build_add(&bld
->base
, lindex
, addr
);
252 return LLVMBuildGEP(bld
->base
.builder
, bld
->immediates_array
, &lindex
, 1, "immediates_ptr");
256 get_addr_ptr(struct lp_build_tgsi_soa_context
*bld
,
262 if (!bld
->has_function_calls
) {
263 return bld
->addr
[index
][swizzle
];
265 LLVMValueRef lindex
=
266 LLVMConstInt(LLVMInt32Type(), index
*4 + swizzle
, 0);
268 lindex
= lp_build_add(&bld
->base
, lindex
, addr
);
269 return LLVMBuildGEP(bld
->base
.builder
, bld
->addrs_array
, &lindex
, 1, "addrs_ptr");
274 get_preds_ptr(struct lp_build_tgsi_soa_context
*bld
,
280 if (!bld
->has_function_calls
) {
281 return bld
->preds
[index
][swizzle
];
283 LLVMValueRef lindex
=
284 LLVMConstInt(LLVMInt32Type(), index
*4 + swizzle
, 0);
286 lindex
= lp_build_add(&bld
->base
, lindex
, addr
);
287 return LLVMBuildGEP(bld
->base
.builder
, bld
->preds_array
, &lindex
, 1, "preds_ptr");
291 static LLVMValueRef
lp_get_function(struct lp_build_tgsi_soa_context
*bld
,
294 struct cso_hash
*hash
= bld
->func_hash
;
295 struct cso_hash_iter iter
= cso_hash_find(hash
, label
);
297 LLVMModuleRef module
= LLVMGetGlobalParent(
298 LLVMGetBasicBlockParent(LLVMGetInsertBlock(bld
->base
.builder
)));
300 if (cso_hash_iter_is_null(iter
)) {
301 LLVMTypeRef func_type
;
302 LLVMTypeRef arg_types
[7];
303 LLVMTypeRef vec_type
= lp_build_vec_type(bld
->base
.type
);
307 util_snprintf(func_name
, 31, "func%d", label
);
309 arg_types
[0] = LLVMPointerType(vec_type
, 0); /* inputs */
310 arg_types
[1] = LLVMPointerType(vec_type
, 0); /* outpus */
311 arg_types
[2] = LLVMTypeOf(bld
->consts_ptr
); /* consts */
312 arg_types
[3] = LLVMPointerType(vec_type
, 0); /* temps */
313 arg_types
[4] = LLVMPointerType(vec_type
, 0); /* addrs */
314 arg_types
[5] = LLVMPointerType(vec_type
, 0); /* preds */
315 arg_types
[6] = LLVMPointerType(vec_type
, 0); /* immediates */
317 func_type
= LLVMFunctionType(LLVMVoidType(), arg_types
, Elements(arg_types
), 0);
319 func
= LLVMAddFunction(module
, func_name
, func_type
);
320 LLVMSetFunctionCallConv(func
, LLVMCCallConv
);
321 for(i
= 0; i
< Elements(arg_types
); ++i
)
322 if(LLVMGetTypeKind(arg_types
[i
]) == LLVMPointerTypeKind
)
323 LLVMAddAttribute(LLVMGetParam(func
, i
), LLVMNoAliasAttribute
);
325 cso_hash_insert(hash
, label
, func
);
327 func
= (LLVMValueRef
)cso_hash_iter_data(iter
);
333 static void lp_exec_mask_init(struct lp_exec_mask
*mask
, struct lp_build_context
*bld
)
336 mask
->has_mask
= FALSE
;
337 mask
->cond_stack_size
= 0;
338 mask
->loop_stack_size
= 0;
341 mask
->int_vec_type
= lp_build_int_vec_type(mask
->bld
->type
);
342 mask
->break_mask
= mask
->cont_mask
= mask
->cond_mask
=
343 LLVMConstAllOnes(mask
->int_vec_type
);
346 static void lp_exec_mask_update(struct lp_exec_mask
*mask
)
348 if (mask
->loop_stack_size
) {
349 /*for loops we need to update the entire mask at runtime */
351 assert(mask
->break_mask
);
352 tmp
= LLVMBuildAnd(mask
->bld
->builder
,
356 mask
->exec_mask
= LLVMBuildAnd(mask
->bld
->builder
,
361 mask
->exec_mask
= mask
->cond_mask
;
363 if (mask
->ret_mask
) {
364 mask
->exec_mask
= LLVMBuildAnd(mask
->bld
->builder
,
371 mask
->has_mask
= (mask
->cond_stack_size
> 0 ||
372 mask
->loop_stack_size
> 0 ||
376 static void lp_exec_mask_cond_push(struct lp_exec_mask
*mask
,
379 assert(mask
->cond_stack_size
< LP_MAX_TGSI_NESTING
);
380 if (mask
->cond_stack_size
== 0) {
381 assert(mask
->cond_mask
== LLVMConstAllOnes(mask
->int_vec_type
));
383 mask
->cond_stack
[mask
->cond_stack_size
++] = mask
->cond_mask
;
384 assert(LLVMTypeOf(val
) == mask
->int_vec_type
);
385 mask
->cond_mask
= val
;
387 lp_exec_mask_update(mask
);
390 static void lp_exec_mask_cond_invert(struct lp_exec_mask
*mask
)
392 LLVMValueRef prev_mask
;
393 LLVMValueRef inv_mask
;
395 assert(mask
->cond_stack_size
);
396 prev_mask
= mask
->cond_stack
[mask
->cond_stack_size
- 1];
397 if (mask
->cond_stack_size
== 1) {
398 assert(prev_mask
== LLVMConstAllOnes(mask
->int_vec_type
));
401 inv_mask
= LLVMBuildNot(mask
->bld
->builder
, mask
->cond_mask
, "");
403 mask
->cond_mask
= LLVMBuildAnd(mask
->bld
->builder
,
406 lp_exec_mask_update(mask
);
409 static void lp_exec_mask_cond_pop(struct lp_exec_mask
*mask
)
411 assert(mask
->cond_stack_size
);
412 mask
->cond_mask
= mask
->cond_stack
[--mask
->cond_stack_size
];
413 lp_exec_mask_update(mask
);
417 static void lp_exec_bgnsub(struct lp_exec_mask
*mask
)
419 mask
->exec_mask
= LLVMConstAllOnes(mask
->int_vec_type
);
420 mask
->ret_mask
= LLVMConstAllOnes(mask
->int_vec_type
);
423 static void lp_exec_bgnloop(struct lp_exec_mask
*mask
)
425 if (mask
->loop_stack_size
== 0) {
426 assert(mask
->loop_block
== NULL
);
427 assert(mask
->cont_mask
== LLVMConstAllOnes(mask
->int_vec_type
));
428 assert(mask
->break_mask
== LLVMConstAllOnes(mask
->int_vec_type
));
429 assert(mask
->break_var
== NULL
);
432 assert(mask
->loop_stack_size
< LP_MAX_TGSI_NESTING
);
434 mask
->loop_stack
[mask
->loop_stack_size
].loop_block
= mask
->loop_block
;
435 mask
->loop_stack
[mask
->loop_stack_size
].cont_mask
= mask
->cont_mask
;
436 mask
->loop_stack
[mask
->loop_stack_size
].break_mask
= mask
->break_mask
;
437 mask
->loop_stack
[mask
->loop_stack_size
].break_var
= mask
->break_var
;
438 ++mask
->loop_stack_size
;
440 mask
->break_var
= lp_build_alloca(mask
->bld
->builder
, mask
->int_vec_type
, "");
441 LLVMBuildStore(mask
->bld
->builder
, mask
->break_mask
, mask
->break_var
);
443 mask
->loop_block
= lp_build_insert_new_block(mask
->bld
->builder
, "bgnloop");
444 LLVMBuildBr(mask
->bld
->builder
, mask
->loop_block
);
445 LLVMPositionBuilderAtEnd(mask
->bld
->builder
, mask
->loop_block
);
447 mask
->break_mask
= LLVMBuildLoad(mask
->bld
->builder
, mask
->break_var
, "");
449 lp_exec_mask_update(mask
);
452 static void lp_exec_break(struct lp_exec_mask
*mask
)
454 LLVMValueRef exec_mask
= LLVMBuildNot(mask
->bld
->builder
,
458 mask
->break_mask
= LLVMBuildAnd(mask
->bld
->builder
,
460 exec_mask
, "break_full");
462 lp_exec_mask_update(mask
);
466 static void lp_exec_ret(struct lp_exec_mask
*mask
)
468 LLVMValueRef exec_mask
= LLVMBuildNot(mask
->bld
->builder
,
472 mask
->ret_mask
= LLVMBuildAnd(mask
->bld
->builder
,
474 exec_mask
, "ret_full");
476 lp_exec_mask_update(mask
);
480 static void lp_exec_continue(struct lp_exec_mask
*mask
)
482 LLVMValueRef exec_mask
= LLVMBuildNot(mask
->bld
->builder
,
486 mask
->cont_mask
= LLVMBuildAnd(mask
->bld
->builder
,
490 lp_exec_mask_update(mask
);
494 static void lp_exec_endloop(struct lp_exec_mask
*mask
)
496 LLVMBasicBlockRef endloop
;
497 LLVMTypeRef reg_type
= LLVMIntType(mask
->bld
->type
.width
*
498 mask
->bld
->type
.length
);
501 assert(mask
->break_mask
);
504 * Restore the cont_mask, but don't pop
506 assert(mask
->loop_stack_size
);
507 mask
->cont_mask
= mask
->loop_stack
[mask
->loop_stack_size
- 1].cont_mask
;
508 lp_exec_mask_update(mask
);
511 * Unlike the continue mask, the break_mask must be preserved across loop
514 LLVMBuildStore(mask
->bld
->builder
, mask
->break_mask
, mask
->break_var
);
516 /* i1cond = (mask == 0) */
517 i1cond
= LLVMBuildICmp(
520 LLVMBuildBitCast(mask
->bld
->builder
, mask
->exec_mask
, reg_type
, ""),
521 LLVMConstNull(reg_type
), "");
523 endloop
= lp_build_insert_new_block(mask
->bld
->builder
, "endloop");
525 LLVMBuildCondBr(mask
->bld
->builder
,
526 i1cond
, mask
->loop_block
, endloop
);
528 LLVMPositionBuilderAtEnd(mask
->bld
->builder
, endloop
);
530 assert(mask
->loop_stack_size
);
531 --mask
->loop_stack_size
;
532 mask
->loop_block
= mask
->loop_stack
[mask
->loop_stack_size
].loop_block
;
533 mask
->cont_mask
= mask
->loop_stack
[mask
->loop_stack_size
].cont_mask
;
534 mask
->break_mask
= mask
->loop_stack
[mask
->loop_stack_size
].break_mask
;
535 mask
->break_var
= mask
->loop_stack
[mask
->loop_stack_size
].break_var
;
537 lp_exec_mask_update(mask
);
540 /* stores val into an address pointed to by dst.
541 * mask->exec_mask is used to figure out which bits of val
542 * should be stored into the address
543 * (0 means don't store this bit, 1 means do store).
545 static void lp_exec_mask_store(struct lp_exec_mask
*mask
,
550 /* Mix the predicate and execution mask */
551 if (mask
->has_mask
) {
553 pred
= LLVMBuildAnd(mask
->bld
->builder
, pred
, mask
->exec_mask
, "");
555 pred
= mask
->exec_mask
;
560 LLVMValueRef real_val
, dst_val
;
562 dst_val
= LLVMBuildLoad(mask
->bld
->builder
, dst
, "");
563 real_val
= lp_build_select(mask
->bld
,
567 LLVMBuildStore(mask
->bld
->builder
, real_val
, dst
);
569 LLVMBuildStore(mask
->bld
->builder
, val
, dst
);
573 emit_vec_alloca_array(struct lp_build_tgsi_soa_context
*bld
,
574 LLVMTypeRef vec_type
,
577 LLVMValueRef val
= LLVMConstInt(LLVMInt32Type(),
579 return lp_build_array_alloca(bld
->base
.builder
,
584 emit_preamble(struct lp_build_tgsi_soa_context
*bld
, uint num_immediates
)
586 LLVMTypeRef vec_type
= lp_build_vec_type(bld
->base
.type
);
589 if (bld
->has_indirect_addressing
||
590 bld
->has_function_calls
) {
591 int size
= bld
->full_range
.temps
.Last
+ 1;
592 bld
->temps_array
= emit_vec_alloca_array(bld
, vec_type
, size
);
594 if (bld
->has_function_calls
) {
596 int size
= bld
->full_range
.outputs
.Last
+ 1;
597 bld
->outputs_array
= emit_vec_alloca_array(bld
, vec_type
, size
);
599 /* we need to insert the created immediates into our array */
600 size
= num_immediates
;
602 bld
->immediates_array
= emit_vec_alloca_array(bld
, vec_type
, size
);
603 for (i
= 0; i
< size
; ++i
) {
605 for (j
= 0; j
< NUM_CHANNELS
; ++j
) {
606 LLVMValueRef ptr
= get_immediates_ptr(bld
,
609 LLVMBuildStore(bld
->base
.builder
,
610 bld
->immediates
[i
][j
],
615 size
= bld
->full_range
.addrs
.Last
+ 1;
616 bld
->addrs_array
= emit_vec_alloca_array(bld
, vec_type
, size
);
618 size
= bld
->full_range
.preds
.Last
+ 1;
619 bld
->preds_array
= emit_vec_alloca_array(bld
, vec_type
, size
);
621 /*inputs also need to be copied*/
622 size
= bld
->full_range
.inputs
.Last
+ 1;
623 bld
->inputs_array
= emit_vec_alloca_array(bld
, vec_type
, size
);
624 for (i
= bld
->full_range
.inputs
.First
; i
< size
; ++i
) {
626 for (j
= 0; j
< NUM_CHANNELS
; ++j
) {
627 LLVMValueRef ptr
= get_input_ptr(bld
,
630 LLVMBuildStore(bld
->base
.builder
,
639 emit_end(struct lp_build_tgsi_soa_context
*bld
)
643 bld
->main_block
= LLVMGetInsertBlock(bld
->base
.builder
);
645 /* if we had function calls we want to propagate the
646 * outputs from the array to the values */
647 if (bld
->has_function_calls
) {
648 int size
= bld
->full_range
.outputs
.Last
+ 1;
649 for (i
= bld
->full_range
.outputs
.First
; i
< size
; ++i
) {
650 for (j
= 0; j
< NUM_CHANNELS
; ++j
) {
651 LLVMValueRef ptr
= get_output_ptr(bld
, i
, j
,
653 bld
->outputs
[i
][j
] = ptr
;
660 emit_bgnsub(struct lp_build_tgsi_soa_context
*bld
)
662 LLVMValueRef func
= lp_get_function(bld
, bld
->instno
);
663 LLVMBasicBlockRef block
;
664 LLVMValueRef inputs_ptr
, outputs_ptr
,
665 consts_ptr
, temps_ptr
, addrs_ptr
, preds_ptr
, imms_ptr
;
667 inputs_ptr
= LLVMGetParam(func
, 0);
668 outputs_ptr
= LLVMGetParam(func
, 1);
669 consts_ptr
= LLVMGetParam(func
, 2);
670 temps_ptr
= LLVMGetParam(func
, 3);
671 addrs_ptr
= LLVMGetParam(func
, 4);
672 preds_ptr
= LLVMGetParam(func
, 5);
673 imms_ptr
= LLVMGetParam(func
, 6);
675 lp_build_name(inputs_ptr
, "inputs");
676 lp_build_name(outputs_ptr
, "outputs");
677 lp_build_name(consts_ptr
, "consts");
678 lp_build_name(temps_ptr
, "temps");
679 lp_build_name(addrs_ptr
, "addrs");
680 lp_build_name(preds_ptr
, "preds");
681 lp_build_name(imms_ptr
, "immediates");
683 bld
->inputs_array
= inputs_ptr
;
684 bld
->outputs_array
= outputs_ptr
;
685 bld
->consts_ptr
= consts_ptr
;
686 bld
->temps_array
= temps_ptr
;
687 bld
->addrs_array
= addrs_ptr
;
688 bld
->preds_array
= preds_ptr
;
689 bld
->immediates_array
= imms_ptr
;
691 block
= LLVMAppendBasicBlock(func
, "entry");
692 LLVMPositionBuilderAtEnd(bld
->base
.builder
, block
);
694 lp_exec_bgnsub(&bld
->exec_mask
);
698 emit_endsub(struct lp_build_tgsi_soa_context
*bld
)
700 LLVMBuildRetVoid(bld
->base
.builder
);
704 emit_ddx(struct lp_build_tgsi_soa_context
*bld
,
707 LLVMValueRef src_left
= lp_build_swizzle1_aos(&bld
->base
, src
, swizzle_left
);
708 LLVMValueRef src_right
= lp_build_swizzle1_aos(&bld
->base
, src
, swizzle_right
);
709 return lp_build_sub(&bld
->base
, src_right
, src_left
);
714 emit_ddy(struct lp_build_tgsi_soa_context
*bld
,
717 LLVMValueRef src_top
= lp_build_swizzle1_aos(&bld
->base
, src
, swizzle_top
);
718 LLVMValueRef src_bottom
= lp_build_swizzle1_aos(&bld
->base
, src
, swizzle_bottom
);
719 return lp_build_sub(&bld
->base
, src_top
, src_bottom
);
727 struct lp_build_tgsi_soa_context
*bld
,
728 const struct tgsi_full_instruction
*inst
,
730 const unsigned chan_index
)
732 const struct tgsi_full_src_register
*reg
= &inst
->Src
[index
];
733 const unsigned swizzle
=
734 tgsi_util_get_full_src_register_swizzle(reg
, chan_index
);
736 LLVMValueRef addr
= NULL
;
739 assert(0 && "invalid swizzle in emit_fetch()");
740 return bld
->base
.undef
;
743 if (reg
->Register
.Indirect
) {
744 LLVMTypeRef int_vec_type
= lp_build_int_vec_type(bld
->base
.type
);
745 unsigned swizzle
= tgsi_util_get_src_register_swizzle( ®
->Indirect
, chan_index
);
746 addr
= LLVMBuildLoad(bld
->base
.builder
,
747 get_addr_ptr(bld
, reg
->Indirect
.Index
, swizzle
, FALSE
, 0),
749 /* for indexing we want integers */
750 addr
= LLVMBuildFPToSI(bld
->base
.builder
, addr
,
752 addr
= LLVMBuildExtractElement(bld
->base
.builder
,
753 addr
, LLVMConstInt(LLVMInt32Type(), 0, 0),
755 addr
= lp_build_mul(&bld
->base
, addr
, LLVMConstInt(LLVMInt32Type(), 4, 0));
758 switch (reg
->Register
.File
) {
759 case TGSI_FILE_CONSTANT
:
761 LLVMValueRef index
= LLVMConstInt(LLVMInt32Type(),
762 reg
->Register
.Index
*4 + swizzle
, 0);
763 LLVMValueRef scalar
, scalar_ptr
;
765 if (reg
->Register
.Indirect
) {
766 /*lp_build_printf(bld->base.builder,
767 "\taddr = %d\n", addr);*/
768 index
= lp_build_add(&bld
->base
, index
, addr
);
770 scalar_ptr
= LLVMBuildGEP(bld
->base
.builder
, bld
->consts_ptr
,
772 scalar
= LLVMBuildLoad(bld
->base
.builder
, scalar_ptr
, "");
774 res
= lp_build_broadcast_scalar(&bld
->base
, scalar
);
778 case TGSI_FILE_IMMEDIATE
:
779 if (bld
->has_function_calls
) {
780 LLVMValueRef ptr
= get_immediates_ptr(bld
,
784 res
= LLVMBuildLoad(bld
->base
.builder
, ptr
, "");
786 res
= bld
->immediates
[reg
->Register
.Index
][swizzle
];
790 case TGSI_FILE_INPUT
:
791 if (bld
->has_function_calls
) {
792 LLVMValueRef ptr
= get_input_ptr(bld
,
796 res
= LLVMBuildLoad(bld
->base
.builder
, ptr
, "");
798 res
= bld
->inputs
[reg
->Register
.Index
][swizzle
];
802 case TGSI_FILE_TEMPORARY
:
804 LLVMValueRef temp_ptr
= get_temp_ptr(bld
, reg
->Register
.Index
,
806 reg
->Register
.Indirect
,
808 res
= LLVMBuildLoad(bld
->base
.builder
, temp_ptr
, "");
810 return bld
->base
.undef
;
815 assert(0 && "invalid src register in emit_fetch()");
816 return bld
->base
.undef
;
819 switch( tgsi_util_get_full_src_register_sign_mode( reg
, chan_index
) ) {
820 case TGSI_UTIL_SIGN_CLEAR
:
821 res
= lp_build_abs( &bld
->base
, res
);
824 case TGSI_UTIL_SIGN_SET
:
825 /* TODO: Use bitwese OR for floating point */
826 res
= lp_build_abs( &bld
->base
, res
);
827 res
= LLVMBuildNeg( bld
->base
.builder
, res
, "" );
830 case TGSI_UTIL_SIGN_TOGGLE
:
831 res
= LLVMBuildNeg( bld
->base
.builder
, res
, "" );
834 case TGSI_UTIL_SIGN_KEEP
:
843 * Register fetch with derivatives.
847 struct lp_build_tgsi_soa_context
*bld
,
848 const struct tgsi_full_instruction
*inst
,
850 const unsigned chan_index
,
857 src
= emit_fetch(bld
, inst
, index
, chan_index
);
862 /* TODO: use interpolation coeffs for inputs */
865 *ddx
= emit_ddx(bld
, src
);
868 *ddy
= emit_ddy(bld
, src
);
876 emit_fetch_predicate(
877 struct lp_build_tgsi_soa_context
*bld
,
878 const struct tgsi_full_instruction
*inst
,
882 unsigned char swizzles
[4];
883 LLVMValueRef unswizzled
[4] = {NULL
, NULL
, NULL
, NULL
};
887 if (!inst
->Instruction
.Predicate
) {
888 FOR_EACH_CHANNEL( chan
) {
894 swizzles
[0] = inst
->Predicate
.SwizzleX
;
895 swizzles
[1] = inst
->Predicate
.SwizzleY
;
896 swizzles
[2] = inst
->Predicate
.SwizzleZ
;
897 swizzles
[3] = inst
->Predicate
.SwizzleW
;
899 index
= inst
->Predicate
.Index
;
900 assert(index
< LP_MAX_TGSI_PREDS
);
902 FOR_EACH_CHANNEL( chan
) {
903 unsigned swizzle
= swizzles
[chan
];
906 * Only fetch the predicate register channels that are actually listed
909 if (!unswizzled
[swizzle
]) {
910 LLVMValueRef pred_ptr
= get_preds_ptr(bld
, index
, swizzle
,
912 value
= LLVMBuildLoad(bld
->base
.builder
,
916 * Convert the value to an integer mask.
918 * TODO: Short-circuit this comparison -- a D3D setp_xx instructions
919 * is needlessly causing two comparisons due to storing the intermediate
920 * result as float vector instead of an integer mask vector.
922 value
= lp_build_compare(bld
->base
.builder
,
927 if (inst
->Predicate
.Negate
) {
928 value
= LLVMBuildNot(bld
->base
.builder
, value
, "");
931 unswizzled
[swizzle
] = value
;
933 value
= unswizzled
[swizzle
];
946 struct lp_build_tgsi_soa_context
*bld
,
947 const struct tgsi_full_instruction
*inst
,
953 const struct tgsi_full_dst_register
*reg
= &inst
->Dst
[index
];
954 LLVMValueRef addr
= NULL
;
956 switch( inst
->Instruction
.Saturate
) {
960 case TGSI_SAT_ZERO_ONE
:
961 value
= lp_build_max(&bld
->base
, value
, bld
->base
.zero
);
962 value
= lp_build_min(&bld
->base
, value
, bld
->base
.one
);
965 case TGSI_SAT_MINUS_PLUS_ONE
:
966 value
= lp_build_max(&bld
->base
, value
, lp_build_const_vec(bld
->base
.type
, -1.0));
967 value
= lp_build_min(&bld
->base
, value
, bld
->base
.one
);
974 if (reg
->Register
.Indirect
) {
975 LLVMTypeRef int_vec_type
= lp_build_int_vec_type(bld
->base
.type
);
976 unsigned swizzle
= tgsi_util_get_src_register_swizzle( ®
->Indirect
, chan_index
);
977 addr
= LLVMBuildLoad(bld
->base
.builder
,
978 get_addr_ptr(bld
, reg
->Indirect
.Index
, swizzle
, FALSE
, 0),
980 /* for indexing we want integers */
981 addr
= LLVMBuildFPToSI(bld
->base
.builder
, addr
,
983 addr
= LLVMBuildExtractElement(bld
->base
.builder
,
984 addr
, LLVMConstInt(LLVMInt32Type(), 0, 0),
986 addr
= lp_build_mul(&bld
->base
, addr
, LLVMConstInt(LLVMInt32Type(), 4, 0));
989 switch( reg
->Register
.File
) {
990 case TGSI_FILE_OUTPUT
:
991 lp_exec_mask_store(&bld
->exec_mask
, pred
, value
,
992 get_output_ptr(bld
, reg
->Register
.Index
, chan_index
,
996 case TGSI_FILE_TEMPORARY
: {
997 LLVMValueRef temp_ptr
= get_temp_ptr(bld
, reg
->Register
.Index
,
999 reg
->Register
.Indirect
,
1001 lp_exec_mask_store(&bld
->exec_mask
, pred
, value
, temp_ptr
);
1005 case TGSI_FILE_ADDRESS
:
1006 lp_exec_mask_store(&bld
->exec_mask
, pred
, value
,
1007 get_addr_ptr(bld
, reg
->Indirect
.Index
, chan_index
,
1011 case TGSI_FILE_PREDICATE
:
1012 lp_exec_mask_store(&bld
->exec_mask
, pred
, value
,
1013 get_preds_ptr(bld
, index
, chan_index
,
1024 * High-level instruction translators.
1028 TEX_MODIFIER_NONE
= 0,
1029 TEX_MODIFIER_PROJECTED
,
1030 TEX_MODIFIER_LOD_BIAS
,
1031 TEX_MODIFIER_EXPLICIT_LOD
,
1032 TEX_MODIFIER_EXPLICIT_DERIV
1036 emit_tex( struct lp_build_tgsi_soa_context
*bld
,
1037 const struct tgsi_full_instruction
*inst
,
1038 enum tex_modifier modifier
,
1039 LLVMValueRef
*texel
)
1042 LLVMValueRef lod_bias
, explicit_lod
;
1043 LLVMValueRef oow
= NULL
;
1044 LLVMValueRef coords
[3];
1045 LLVMValueRef ddx
[3];
1046 LLVMValueRef ddy
[3];
1047 unsigned num_coords
;
1050 if (!bld
->sampler
) {
1051 _debug_printf("warning: found texture instruction but no sampler generator supplied\n");
1052 for (i
= 0; i
< 4; i
++) {
1053 texel
[i
] = bld
->base
.undef
;
1058 switch (inst
->Texture
.Texture
) {
1059 case TGSI_TEXTURE_1D
:
1062 case TGSI_TEXTURE_2D
:
1063 case TGSI_TEXTURE_RECT
:
1066 case TGSI_TEXTURE_SHADOW1D
:
1067 case TGSI_TEXTURE_SHADOW2D
:
1068 case TGSI_TEXTURE_SHADOWRECT
:
1069 case TGSI_TEXTURE_3D
:
1070 case TGSI_TEXTURE_CUBE
:
1078 if (modifier
== TEX_MODIFIER_LOD_BIAS
) {
1079 lod_bias
= emit_fetch( bld
, inst
, 0, 3 );
1080 explicit_lod
= NULL
;
1082 else if (modifier
== TEX_MODIFIER_EXPLICIT_LOD
) {
1084 explicit_lod
= emit_fetch( bld
, inst
, 0, 3 );
1088 explicit_lod
= NULL
;
1091 if (modifier
== TEX_MODIFIER_PROJECTED
) {
1092 oow
= emit_fetch( bld
, inst
, 0, 3 );
1093 oow
= lp_build_rcp(&bld
->base
, oow
);
1096 for (i
= 0; i
< num_coords
; i
++) {
1097 coords
[i
] = emit_fetch( bld
, inst
, 0, i
);
1098 if (modifier
== TEX_MODIFIER_PROJECTED
)
1099 coords
[i
] = lp_build_mul(&bld
->base
, coords
[i
], oow
);
1101 for (i
= num_coords
; i
< 3; i
++) {
1102 coords
[i
] = bld
->base
.undef
;
1105 if (modifier
== TEX_MODIFIER_EXPLICIT_DERIV
) {
1106 for (i
= 0; i
< num_coords
; i
++) {
1107 ddx
[i
] = emit_fetch( bld
, inst
, 1, i
);
1108 ddy
[i
] = emit_fetch( bld
, inst
, 2, i
);
1110 unit
= inst
->Src
[3].Register
.Index
;
1112 for (i
= 0; i
< num_coords
; i
++) {
1113 ddx
[i
] = emit_ddx( bld
, coords
[i
] );
1114 ddy
[i
] = emit_ddy( bld
, coords
[i
] );
1116 unit
= inst
->Src
[1].Register
.Index
;
1118 for (i
= num_coords
; i
< 3; i
++) {
1119 ddx
[i
] = bld
->base
.undef
;
1120 ddy
[i
] = bld
->base
.undef
;
1123 bld
->sampler
->emit_fetch_texel(bld
->sampler
,
1126 unit
, num_coords
, coords
,
1128 lod_bias
, explicit_lod
,
1134 * Kill fragment if any of the src register values are negative.
1138 struct lp_build_tgsi_soa_context
*bld
,
1139 const struct tgsi_full_instruction
*inst
)
1141 const struct tgsi_full_src_register
*reg
= &inst
->Src
[0];
1142 LLVMValueRef terms
[NUM_CHANNELS
];
1144 unsigned chan_index
;
1146 memset(&terms
, 0, sizeof terms
);
1148 FOR_EACH_CHANNEL( chan_index
) {
1151 /* Unswizzle channel */
1152 swizzle
= tgsi_util_get_full_src_register_swizzle( reg
, chan_index
);
1154 /* Check if the component has not been already tested. */
1155 assert(swizzle
< NUM_CHANNELS
);
1156 if( !terms
[swizzle
] )
1157 /* TODO: change the comparison operator instead of setting the sign */
1158 terms
[swizzle
] = emit_fetch(bld
, inst
, 0, chan_index
);
1162 FOR_EACH_CHANNEL( chan_index
) {
1163 if(terms
[chan_index
]) {
1164 LLVMValueRef chan_mask
;
1167 * If term < 0 then mask = 0 else mask = ~0.
1169 chan_mask
= lp_build_cmp(&bld
->base
, PIPE_FUNC_GEQUAL
, terms
[chan_index
], bld
->base
.zero
);
1172 mask
= LLVMBuildAnd(bld
->base
.builder
, mask
, chan_mask
, "");
1179 lp_build_mask_update(bld
->mask
, mask
);
1184 * Predicated fragment kill.
1185 * XXX Actually, we do an unconditional kill (as in tgsi_exec.c).
1186 * The only predication is the execution mask which will apply if
1187 * we're inside a loop or conditional.
1190 emit_kilp(struct lp_build_tgsi_soa_context
*bld
,
1191 const struct tgsi_full_instruction
*inst
)
1195 /* For those channels which are "alive", disable fragment shader
1198 if (bld
->exec_mask
.has_mask
) {
1199 mask
= LLVMBuildNot(bld
->base
.builder
, bld
->exec_mask
.exec_mask
, "kilp");
1202 mask
= bld
->base
.zero
;
1205 lp_build_mask_update(bld
->mask
, mask
);
1209 range_check(struct tgsi_declaration_range
*range
,
1210 unsigned new_first
, unsigned new_last
)
1212 range
->First
= MIN2(range
->First
, new_first
);
1213 range
->Last
= MAX2(range
->Last
, new_last
);
1218 struct lp_build_tgsi_soa_context
*bld
,
1219 const struct tgsi_full_declaration
*decl
)
1221 LLVMTypeRef vec_type
= lp_build_vec_type(bld
->base
.type
);
1223 unsigned first
= decl
->Range
.First
;
1224 unsigned last
= decl
->Range
.Last
;
1227 for (idx
= first
; idx
<= last
; ++idx
) {
1228 switch (decl
->Declaration
.File
) {
1229 case TGSI_FILE_TEMPORARY
:
1230 assert(idx
< LP_MAX_TGSI_TEMPS
);
1231 range_check(&bld
->full_range
.temps
,
1233 if (!bld
->has_indirect_addressing
&&
1234 !bld
->has_function_calls
) {
1235 for (i
= 0; i
< NUM_CHANNELS
; i
++)
1236 bld
->temps
[idx
][i
] = lp_build_alloca(bld
->base
.builder
,
1241 case TGSI_FILE_OUTPUT
:
1242 range_check(&bld
->full_range
.outputs
,
1244 if (!bld
->has_function_calls
) {
1245 for (i
= 0; i
< NUM_CHANNELS
; i
++)
1246 bld
->outputs
[idx
][i
] = lp_build_alloca(bld
->base
.builder
,
1251 case TGSI_FILE_ADDRESS
:
1252 assert(idx
< LP_MAX_TGSI_ADDRS
);
1253 range_check(&bld
->full_range
.addrs
,
1255 if (!bld
->has_function_calls
) {
1256 for (i
= 0; i
< NUM_CHANNELS
; i
++)
1257 bld
->addr
[idx
][i
] = lp_build_alloca(bld
->base
.builder
,
1262 case TGSI_FILE_PREDICATE
:
1263 assert(idx
< LP_MAX_TGSI_PREDS
);
1264 range_check(&bld
->full_range
.preds
,
1266 if (!bld
->has_function_calls
) {
1267 for (i
= 0; i
< NUM_CHANNELS
; i
++)
1268 bld
->preds
[idx
][i
] = lp_build_alloca(bld
->base
.builder
,
1273 case TGSI_FILE_INPUT
:
1274 range_check(&bld
->full_range
.inputs
,
1279 /* don't need to declare other vars */
1287 * Emit LLVM for one TGSI instruction.
1288 * \param return TRUE for success, FALSE otherwise
1292 struct lp_build_tgsi_soa_context
*bld
,
1293 const struct tgsi_full_instruction
*inst
,
1294 const struct tgsi_opcode_info
*info
)
1296 unsigned chan_index
;
1297 LLVMValueRef src0
, src1
, src2
;
1298 LLVMValueRef tmp0
, tmp1
, tmp2
;
1299 LLVMValueRef tmp3
= NULL
;
1300 LLVMValueRef tmp4
= NULL
;
1301 LLVMValueRef tmp5
= NULL
;
1302 LLVMValueRef tmp6
= NULL
;
1303 LLVMValueRef tmp7
= NULL
;
1305 LLVMValueRef dst0
[NUM_CHANNELS
];
1308 * Stores and write masks are handled in a general fashion after the long
1309 * instruction opcode switch statement.
1311 * Although not stricitly necessary, we avoid generating instructions for
1312 * channels which won't be stored, in cases where's that easy. For some
1313 * complex instructions, like texture sampling, it is more convenient to
1314 * assume a full writemask and then let LLVM optimization passes eliminate
1318 assert(info
->num_dst
<= 1);
1319 if (info
->num_dst
) {
1320 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1321 dst0
[chan_index
] = bld
->base
.undef
;
1325 switch (inst
->Instruction
.Opcode
) {
1326 case TGSI_OPCODE_ARL
:
1327 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1328 tmp0
= emit_fetch( bld
, inst
, 0, chan_index
);
1329 tmp0
= lp_build_floor(&bld
->base
, tmp0
);
1330 dst0
[chan_index
] = tmp0
;
1334 case TGSI_OPCODE_MOV
:
1335 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1336 dst0
[chan_index
] = emit_fetch( bld
, inst
, 0, chan_index
);
1340 case TGSI_OPCODE_LIT
:
1341 if( IS_DST0_CHANNEL_ENABLED( inst
, CHAN_X
) ) {
1342 dst0
[CHAN_X
] = bld
->base
.one
;
1344 if( IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Y
) ) {
1345 src0
= emit_fetch( bld
, inst
, 0, CHAN_X
);
1346 dst0
[CHAN_Y
] = lp_build_max( &bld
->base
, src0
, bld
->base
.zero
);
1348 if( IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Z
) ) {
1349 /* XMM[1] = SrcReg[0].yyyy */
1350 tmp1
= emit_fetch( bld
, inst
, 0, CHAN_Y
);
1351 /* XMM[1] = max(XMM[1], 0) */
1352 tmp1
= lp_build_max( &bld
->base
, tmp1
, bld
->base
.zero
);
1353 /* XMM[2] = SrcReg[0].wwww */
1354 tmp2
= emit_fetch( bld
, inst
, 0, CHAN_W
);
1355 tmp1
= lp_build_pow( &bld
->base
, tmp1
, tmp2
);
1356 tmp0
= emit_fetch( bld
, inst
, 0, CHAN_X
);
1357 tmp2
= lp_build_cmp(&bld
->base
, PIPE_FUNC_GREATER
, tmp0
, bld
->base
.zero
);
1358 dst0
[CHAN_Z
] = lp_build_select(&bld
->base
, tmp2
, tmp1
, bld
->base
.zero
);
1360 if( IS_DST0_CHANNEL_ENABLED( inst
, CHAN_W
) ) {
1361 dst0
[CHAN_W
] = bld
->base
.one
;
1365 case TGSI_OPCODE_RCP
:
1366 /* TGSI_OPCODE_RECIP */
1367 src0
= emit_fetch( bld
, inst
, 0, CHAN_X
);
1368 res
= lp_build_rcp(&bld
->base
, src0
);
1369 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1370 dst0
[chan_index
] = res
;
1374 case TGSI_OPCODE_RSQ
:
1375 /* TGSI_OPCODE_RECIPSQRT */
1376 src0
= emit_fetch( bld
, inst
, 0, CHAN_X
);
1377 src0
= lp_build_abs(&bld
->base
, src0
);
1378 res
= lp_build_rsqrt(&bld
->base
, src0
);
1379 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1380 dst0
[chan_index
] = res
;
1384 case TGSI_OPCODE_EXP
:
1385 if (IS_DST0_CHANNEL_ENABLED( inst
, CHAN_X
) ||
1386 IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Y
) ||
1387 IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Z
)) {
1388 LLVMValueRef
*p_exp2_int_part
= NULL
;
1389 LLVMValueRef
*p_frac_part
= NULL
;
1390 LLVMValueRef
*p_exp2
= NULL
;
1392 src0
= emit_fetch( bld
, inst
, 0, CHAN_X
);
1394 if (IS_DST0_CHANNEL_ENABLED( inst
, CHAN_X
))
1395 p_exp2_int_part
= &tmp0
;
1396 if (IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Y
))
1397 p_frac_part
= &tmp1
;
1398 if (IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Z
))
1401 lp_build_exp2_approx(&bld
->base
, src0
, p_exp2_int_part
, p_frac_part
, p_exp2
);
1403 if (IS_DST0_CHANNEL_ENABLED( inst
, CHAN_X
))
1404 dst0
[CHAN_X
] = tmp0
;
1405 if (IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Y
))
1406 dst0
[CHAN_Y
] = tmp1
;
1407 if (IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Z
))
1408 dst0
[CHAN_Z
] = tmp2
;
1411 if (IS_DST0_CHANNEL_ENABLED( inst
, CHAN_W
)) {
1412 dst0
[CHAN_W
] = bld
->base
.one
;
1416 case TGSI_OPCODE_LOG
:
1417 if (IS_DST0_CHANNEL_ENABLED( inst
, CHAN_X
) ||
1418 IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Y
) ||
1419 IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Z
)) {
1420 LLVMValueRef
*p_floor_log2
= NULL
;
1421 LLVMValueRef
*p_exp
= NULL
;
1422 LLVMValueRef
*p_log2
= NULL
;
1424 src0
= emit_fetch( bld
, inst
, 0, CHAN_X
);
1425 src0
= lp_build_abs( &bld
->base
, src0
);
1427 if (IS_DST0_CHANNEL_ENABLED( inst
, CHAN_X
))
1428 p_floor_log2
= &tmp0
;
1429 if (IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Y
))
1431 if (IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Z
))
1434 lp_build_log2_approx(&bld
->base
, src0
, p_exp
, p_floor_log2
, p_log2
);
1436 /* dst.x = floor(lg2(abs(src.x))) */
1437 if (IS_DST0_CHANNEL_ENABLED( inst
, CHAN_X
))
1438 dst0
[CHAN_X
] = tmp0
;
1439 /* dst.y = abs(src)/ex2(floor(lg2(abs(src.x)))) */
1440 if (IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Y
)) {
1441 dst0
[CHAN_Y
] = lp_build_div( &bld
->base
, src0
, tmp1
);
1443 /* dst.z = lg2(abs(src.x)) */
1444 if (IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Z
))
1445 dst0
[CHAN_Z
] = tmp2
;
1448 if (IS_DST0_CHANNEL_ENABLED( inst
, CHAN_W
)) {
1449 dst0
[CHAN_W
] = bld
->base
.one
;
1453 case TGSI_OPCODE_MUL
:
1454 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1455 src0
= emit_fetch( bld
, inst
, 0, chan_index
);
1456 src1
= emit_fetch( bld
, inst
, 1, chan_index
);
1457 dst0
[chan_index
] = lp_build_mul(&bld
->base
, src0
, src1
);
1461 case TGSI_OPCODE_ADD
:
1462 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1463 src0
= emit_fetch( bld
, inst
, 0, chan_index
);
1464 src1
= emit_fetch( bld
, inst
, 1, chan_index
);
1465 dst0
[chan_index
] = lp_build_add(&bld
->base
, src0
, src1
);
1469 case TGSI_OPCODE_DP3
:
1470 /* TGSI_OPCODE_DOT3 */
1471 tmp0
= emit_fetch( bld
, inst
, 0, CHAN_X
);
1472 tmp1
= emit_fetch( bld
, inst
, 1, CHAN_X
);
1473 tmp0
= lp_build_mul( &bld
->base
, tmp0
, tmp1
);
1474 tmp1
= emit_fetch( bld
, inst
, 0, CHAN_Y
);
1475 tmp2
= emit_fetch( bld
, inst
, 1, CHAN_Y
);
1476 tmp1
= lp_build_mul( &bld
->base
, tmp1
, tmp2
);
1477 tmp0
= lp_build_add( &bld
->base
, tmp0
, tmp1
);
1478 tmp1
= emit_fetch( bld
, inst
, 0, CHAN_Z
);
1479 tmp2
= emit_fetch( bld
, inst
, 1, CHAN_Z
);
1480 tmp1
= lp_build_mul( &bld
->base
, tmp1
, tmp2
);
1481 tmp0
= lp_build_add( &bld
->base
, tmp0
, tmp1
);
1482 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1483 dst0
[chan_index
] = tmp0
;
1487 case TGSI_OPCODE_DP4
:
1488 /* TGSI_OPCODE_DOT4 */
1489 tmp0
= emit_fetch( bld
, inst
, 0, CHAN_X
);
1490 tmp1
= emit_fetch( bld
, inst
, 1, CHAN_X
);
1491 tmp0
= lp_build_mul( &bld
->base
, tmp0
, tmp1
);
1492 tmp1
= emit_fetch( bld
, inst
, 0, CHAN_Y
);
1493 tmp2
= emit_fetch( bld
, inst
, 1, CHAN_Y
);
1494 tmp1
= lp_build_mul( &bld
->base
, tmp1
, tmp2
);
1495 tmp0
= lp_build_add( &bld
->base
, tmp0
, tmp1
);
1496 tmp1
= emit_fetch( bld
, inst
, 0, CHAN_Z
);
1497 tmp2
= emit_fetch( bld
, inst
, 1, CHAN_Z
);
1498 tmp1
= lp_build_mul( &bld
->base
, tmp1
, tmp2
);
1499 tmp0
= lp_build_add( &bld
->base
, tmp0
, tmp1
);
1500 tmp1
= emit_fetch( bld
, inst
, 0, CHAN_W
);
1501 tmp2
= emit_fetch( bld
, inst
, 1, CHAN_W
);
1502 tmp1
= lp_build_mul( &bld
->base
, tmp1
, tmp2
);
1503 tmp0
= lp_build_add( &bld
->base
, tmp0
, tmp1
);
1504 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1505 dst0
[chan_index
] = tmp0
;
1509 case TGSI_OPCODE_DST
:
1510 IF_IS_DST0_CHANNEL_ENABLED( inst
, CHAN_X
) {
1511 dst0
[CHAN_X
] = bld
->base
.one
;
1513 IF_IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Y
) {
1514 tmp0
= emit_fetch( bld
, inst
, 0, CHAN_Y
);
1515 tmp1
= emit_fetch( bld
, inst
, 1, CHAN_Y
);
1516 dst0
[CHAN_Y
] = lp_build_mul( &bld
->base
, tmp0
, tmp1
);
1518 IF_IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Z
) {
1519 dst0
[CHAN_Z
] = emit_fetch( bld
, inst
, 0, CHAN_Z
);
1521 IF_IS_DST0_CHANNEL_ENABLED( inst
, CHAN_W
) {
1522 dst0
[CHAN_W
] = emit_fetch( bld
, inst
, 1, CHAN_W
);
1526 case TGSI_OPCODE_MIN
:
1527 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1528 src0
= emit_fetch( bld
, inst
, 0, chan_index
);
1529 src1
= emit_fetch( bld
, inst
, 1, chan_index
);
1530 dst0
[chan_index
] = lp_build_min( &bld
->base
, src0
, src1
);
1534 case TGSI_OPCODE_MAX
:
1535 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1536 src0
= emit_fetch( bld
, inst
, 0, chan_index
);
1537 src1
= emit_fetch( bld
, inst
, 1, chan_index
);
1538 dst0
[chan_index
] = lp_build_max( &bld
->base
, src0
, src1
);
1542 case TGSI_OPCODE_SLT
:
1543 /* TGSI_OPCODE_SETLT */
1544 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1545 src0
= emit_fetch( bld
, inst
, 0, chan_index
);
1546 src1
= emit_fetch( bld
, inst
, 1, chan_index
);
1547 tmp0
= lp_build_cmp( &bld
->base
, PIPE_FUNC_LESS
, src0
, src1
);
1548 dst0
[chan_index
] = lp_build_select( &bld
->base
, tmp0
, bld
->base
.one
, bld
->base
.zero
);
1552 case TGSI_OPCODE_SGE
:
1553 /* TGSI_OPCODE_SETGE */
1554 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1555 src0
= emit_fetch( bld
, inst
, 0, chan_index
);
1556 src1
= emit_fetch( bld
, inst
, 1, chan_index
);
1557 tmp0
= lp_build_cmp( &bld
->base
, PIPE_FUNC_GEQUAL
, src0
, src1
);
1558 dst0
[chan_index
] = lp_build_select( &bld
->base
, tmp0
, bld
->base
.one
, bld
->base
.zero
);
1562 case TGSI_OPCODE_MAD
:
1563 /* TGSI_OPCODE_MADD */
1564 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1565 tmp0
= emit_fetch( bld
, inst
, 0, chan_index
);
1566 tmp1
= emit_fetch( bld
, inst
, 1, chan_index
);
1567 tmp2
= emit_fetch( bld
, inst
, 2, chan_index
);
1568 tmp0
= lp_build_mul( &bld
->base
, tmp0
, tmp1
);
1569 tmp0
= lp_build_add( &bld
->base
, tmp0
, tmp2
);
1570 dst0
[chan_index
] = tmp0
;
1574 case TGSI_OPCODE_SUB
:
1575 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1576 tmp0
= emit_fetch( bld
, inst
, 0, chan_index
);
1577 tmp1
= emit_fetch( bld
, inst
, 1, chan_index
);
1578 dst0
[chan_index
] = lp_build_sub( &bld
->base
, tmp0
, tmp1
);
1582 case TGSI_OPCODE_LRP
:
1583 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1584 src0
= emit_fetch( bld
, inst
, 0, chan_index
);
1585 src1
= emit_fetch( bld
, inst
, 1, chan_index
);
1586 src2
= emit_fetch( bld
, inst
, 2, chan_index
);
1587 tmp0
= lp_build_sub( &bld
->base
, src1
, src2
);
1588 tmp0
= lp_build_mul( &bld
->base
, src0
, tmp0
);
1589 dst0
[chan_index
] = lp_build_add( &bld
->base
, tmp0
, src2
);
1593 case TGSI_OPCODE_CND
:
1594 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1595 src0
= emit_fetch( bld
, inst
, 0, chan_index
);
1596 src1
= emit_fetch( bld
, inst
, 1, chan_index
);
1597 src2
= emit_fetch( bld
, inst
, 2, chan_index
);
1598 tmp1
= lp_build_const_vec(bld
->base
.type
, 0.5);
1599 tmp0
= lp_build_cmp( &bld
->base
, PIPE_FUNC_GREATER
, src2
, tmp1
);
1600 dst0
[chan_index
] = lp_build_select( &bld
->base
, tmp0
, src0
, src1
);
1604 case TGSI_OPCODE_DP2A
:
1605 tmp0
= emit_fetch( bld
, inst
, 0, CHAN_X
); /* xmm0 = src[0].x */
1606 tmp1
= emit_fetch( bld
, inst
, 1, CHAN_X
); /* xmm1 = src[1].x */
1607 tmp0
= lp_build_mul( &bld
->base
, tmp0
, tmp1
); /* xmm0 = xmm0 * xmm1 */
1608 tmp1
= emit_fetch( bld
, inst
, 0, CHAN_Y
); /* xmm1 = src[0].y */
1609 tmp2
= emit_fetch( bld
, inst
, 1, CHAN_Y
); /* xmm2 = src[1].y */
1610 tmp1
= lp_build_mul( &bld
->base
, tmp1
, tmp2
); /* xmm1 = xmm1 * xmm2 */
1611 tmp0
= lp_build_add( &bld
->base
, tmp0
, tmp1
); /* xmm0 = xmm0 + xmm1 */
1612 tmp1
= emit_fetch( bld
, inst
, 2, CHAN_X
); /* xmm1 = src[2].x */
1613 tmp0
= lp_build_add( &bld
->base
, tmp0
, tmp1
); /* xmm0 = xmm0 + xmm1 */
1614 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1615 dst0
[chan_index
] = tmp0
; /* dest[ch] = xmm0 */
1619 case TGSI_OPCODE_FRC
:
1620 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1621 src0
= emit_fetch( bld
, inst
, 0, chan_index
);
1622 tmp0
= lp_build_floor(&bld
->base
, src0
);
1623 tmp0
= lp_build_sub(&bld
->base
, src0
, tmp0
);
1624 dst0
[chan_index
] = tmp0
;
1628 case TGSI_OPCODE_CLAMP
:
1629 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1630 tmp0
= emit_fetch( bld
, inst
, 0, chan_index
);
1631 src1
= emit_fetch( bld
, inst
, 1, chan_index
);
1632 src2
= emit_fetch( bld
, inst
, 2, chan_index
);
1633 tmp0
= lp_build_max(&bld
->base
, tmp0
, src1
);
1634 tmp0
= lp_build_min(&bld
->base
, tmp0
, src2
);
1635 dst0
[chan_index
] = tmp0
;
1639 case TGSI_OPCODE_FLR
:
1640 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1641 tmp0
= emit_fetch( bld
, inst
, 0, chan_index
);
1642 dst0
[chan_index
] = lp_build_floor(&bld
->base
, tmp0
);
1646 case TGSI_OPCODE_ROUND
:
1647 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1648 tmp0
= emit_fetch( bld
, inst
, 0, chan_index
);
1649 dst0
[chan_index
] = lp_build_round(&bld
->base
, tmp0
);
1653 case TGSI_OPCODE_EX2
: {
1654 tmp0
= emit_fetch( bld
, inst
, 0, CHAN_X
);
1655 tmp0
= lp_build_exp2( &bld
->base
, tmp0
);
1656 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1657 dst0
[chan_index
] = tmp0
;
1662 case TGSI_OPCODE_LG2
:
1663 tmp0
= emit_fetch( bld
, inst
, 0, CHAN_X
);
1664 tmp0
= lp_build_log2( &bld
->base
, tmp0
);
1665 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1666 dst0
[chan_index
] = tmp0
;
1670 case TGSI_OPCODE_POW
:
1671 src0
= emit_fetch( bld
, inst
, 0, CHAN_X
);
1672 src1
= emit_fetch( bld
, inst
, 1, CHAN_X
);
1673 res
= lp_build_pow( &bld
->base
, src0
, src1
);
1674 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1675 dst0
[chan_index
] = res
;
1679 case TGSI_OPCODE_XPD
:
1680 if( IS_DST0_CHANNEL_ENABLED( inst
, CHAN_X
) ||
1681 IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Y
) ) {
1682 tmp1
= emit_fetch( bld
, inst
, 1, CHAN_Z
);
1683 tmp3
= emit_fetch( bld
, inst
, 0, CHAN_Z
);
1685 if( IS_DST0_CHANNEL_ENABLED( inst
, CHAN_X
) ||
1686 IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Z
) ) {
1687 tmp0
= emit_fetch( bld
, inst
, 0, CHAN_Y
);
1688 tmp4
= emit_fetch( bld
, inst
, 1, CHAN_Y
);
1690 IF_IS_DST0_CHANNEL_ENABLED( inst
, CHAN_X
) {
1692 tmp2
= lp_build_mul( &bld
->base
, tmp2
, tmp1
);
1694 tmp5
= lp_build_mul( &bld
->base
, tmp5
, tmp4
);
1695 tmp2
= lp_build_sub( &bld
->base
, tmp2
, tmp5
);
1696 dst0
[CHAN_X
] = tmp2
;
1698 if( IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Y
) ||
1699 IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Z
) ) {
1700 tmp2
= emit_fetch( bld
, inst
, 1, CHAN_X
);
1701 tmp5
= emit_fetch( bld
, inst
, 0, CHAN_X
);
1703 IF_IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Y
) {
1704 tmp3
= lp_build_mul( &bld
->base
, tmp3
, tmp2
);
1705 tmp1
= lp_build_mul( &bld
->base
, tmp1
, tmp5
);
1706 tmp3
= lp_build_sub( &bld
->base
, tmp3
, tmp1
);
1707 dst0
[CHAN_Y
] = tmp3
;
1709 IF_IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Z
) {
1710 tmp5
= lp_build_mul( &bld
->base
, tmp5
, tmp4
);
1711 tmp0
= lp_build_mul( &bld
->base
, tmp0
, tmp2
);
1712 tmp5
= lp_build_sub( &bld
->base
, tmp5
, tmp0
);
1713 dst0
[CHAN_Z
] = tmp5
;
1715 IF_IS_DST0_CHANNEL_ENABLED( inst
, CHAN_W
) {
1716 dst0
[CHAN_W
] = bld
->base
.one
;
1720 case TGSI_OPCODE_ABS
:
1721 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1722 tmp0
= emit_fetch( bld
, inst
, 0, chan_index
);
1723 dst0
[chan_index
] = lp_build_abs( &bld
->base
, tmp0
);
1727 case TGSI_OPCODE_RCC
:
1732 case TGSI_OPCODE_DPH
:
1733 tmp0
= emit_fetch( bld
, inst
, 0, CHAN_X
);
1734 tmp1
= emit_fetch( bld
, inst
, 1, CHAN_X
);
1735 tmp0
= lp_build_mul( &bld
->base
, tmp0
, tmp1
);
1736 tmp1
= emit_fetch( bld
, inst
, 0, CHAN_Y
);
1737 tmp2
= emit_fetch( bld
, inst
, 1, CHAN_Y
);
1738 tmp1
= lp_build_mul( &bld
->base
, tmp1
, tmp2
);
1739 tmp0
= lp_build_add( &bld
->base
, tmp0
, tmp1
);
1740 tmp1
= emit_fetch( bld
, inst
, 0, CHAN_Z
);
1741 tmp2
= emit_fetch( bld
, inst
, 1, CHAN_Z
);
1742 tmp1
= lp_build_mul( &bld
->base
, tmp1
, tmp2
);
1743 tmp0
= lp_build_add( &bld
->base
, tmp0
, tmp1
);
1744 tmp1
= emit_fetch( bld
, inst
, 1, CHAN_W
);
1745 tmp0
= lp_build_add( &bld
->base
, tmp0
, tmp1
);
1746 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1747 dst0
[chan_index
] = tmp0
;
1751 case TGSI_OPCODE_COS
:
1752 tmp0
= emit_fetch( bld
, inst
, 0, CHAN_X
);
1753 tmp0
= lp_build_cos( &bld
->base
, tmp0
);
1754 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1755 dst0
[chan_index
] = tmp0
;
1759 case TGSI_OPCODE_DDX
:
1760 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1761 emit_fetch_deriv( bld
, inst
, 0, chan_index
, NULL
, &dst0
[chan_index
], NULL
);
1765 case TGSI_OPCODE_DDY
:
1766 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1767 emit_fetch_deriv( bld
, inst
, 0, chan_index
, NULL
, NULL
, &dst0
[chan_index
]);
1771 case TGSI_OPCODE_KILP
:
1772 /* predicated kill */
1773 emit_kilp( bld
, inst
);
1776 case TGSI_OPCODE_KIL
:
1777 /* conditional kill */
1778 emit_kil( bld
, inst
);
1781 case TGSI_OPCODE_PK2H
:
1785 case TGSI_OPCODE_PK2US
:
1789 case TGSI_OPCODE_PK4B
:
1793 case TGSI_OPCODE_PK4UB
:
1797 case TGSI_OPCODE_RFL
:
1801 case TGSI_OPCODE_SEQ
:
1802 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1803 src0
= emit_fetch( bld
, inst
, 0, chan_index
);
1804 src1
= emit_fetch( bld
, inst
, 1, chan_index
);
1805 tmp0
= lp_build_cmp( &bld
->base
, PIPE_FUNC_EQUAL
, src0
, src1
);
1806 dst0
[chan_index
] = lp_build_select( &bld
->base
, tmp0
, bld
->base
.one
, bld
->base
.zero
);
1810 case TGSI_OPCODE_SFL
:
1811 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1812 dst0
[chan_index
] = bld
->base
.zero
;
1816 case TGSI_OPCODE_SGT
:
1817 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1818 src0
= emit_fetch( bld
, inst
, 0, chan_index
);
1819 src1
= emit_fetch( bld
, inst
, 1, chan_index
);
1820 tmp0
= lp_build_cmp( &bld
->base
, PIPE_FUNC_GREATER
, src0
, src1
);
1821 dst0
[chan_index
] = lp_build_select( &bld
->base
, tmp0
, bld
->base
.one
, bld
->base
.zero
);
1825 case TGSI_OPCODE_SIN
:
1826 tmp0
= emit_fetch( bld
, inst
, 0, CHAN_X
);
1827 tmp0
= lp_build_sin( &bld
->base
, tmp0
);
1828 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1829 dst0
[chan_index
] = tmp0
;
1833 case TGSI_OPCODE_SLE
:
1834 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1835 src0
= emit_fetch( bld
, inst
, 0, chan_index
);
1836 src1
= emit_fetch( bld
, inst
, 1, chan_index
);
1837 tmp0
= lp_build_cmp( &bld
->base
, PIPE_FUNC_LEQUAL
, src0
, src1
);
1838 dst0
[chan_index
] = lp_build_select( &bld
->base
, tmp0
, bld
->base
.one
, bld
->base
.zero
);
1842 case TGSI_OPCODE_SNE
:
1843 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1844 src0
= emit_fetch( bld
, inst
, 0, chan_index
);
1845 src1
= emit_fetch( bld
, inst
, 1, chan_index
);
1846 tmp0
= lp_build_cmp( &bld
->base
, PIPE_FUNC_NOTEQUAL
, src0
, src1
);
1847 dst0
[chan_index
] = lp_build_select( &bld
->base
, tmp0
, bld
->base
.one
, bld
->base
.zero
);
1851 case TGSI_OPCODE_STR
:
1852 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1853 dst0
[chan_index
] = bld
->base
.one
;
1857 case TGSI_OPCODE_TEX
:
1858 emit_tex( bld
, inst
, TEX_MODIFIER_NONE
, dst0
);
1861 case TGSI_OPCODE_TXD
:
1862 emit_tex( bld
, inst
, TEX_MODIFIER_EXPLICIT_DERIV
, dst0
);
1865 case TGSI_OPCODE_UP2H
:
1871 case TGSI_OPCODE_UP2US
:
1877 case TGSI_OPCODE_UP4B
:
1883 case TGSI_OPCODE_UP4UB
:
1889 case TGSI_OPCODE_X2D
:
1895 case TGSI_OPCODE_ARA
:
1901 case TGSI_OPCODE_ARR
:
1902 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1903 tmp0
= emit_fetch( bld
, inst
, 0, chan_index
);
1904 tmp0
= lp_build_round(&bld
->base
, tmp0
);
1905 dst0
[chan_index
] = tmp0
;
1909 case TGSI_OPCODE_BRA
:
1915 case TGSI_OPCODE_CAL
: {
1916 LLVMValueRef args
[7];
1917 LLVMValueRef func
= lp_get_function(bld
, inst
->Label
.Label
);
1918 args
[0] = bld
->inputs_array
;
1919 args
[1] = bld
->outputs_array
;
1920 args
[2] = bld
->consts_ptr
;
1921 args
[3] = bld
->temps_array
;
1922 args
[4] = bld
->addrs_array
;
1923 args
[5] = bld
->preds_array
;
1924 args
[6] = bld
->immediates_array
;
1925 LLVMBuildCall(bld
->base
.builder
, func
, args
, Elements(args
), "");
1929 case TGSI_OPCODE_RET
:
1930 lp_exec_ret(&bld
->exec_mask
);
1933 case TGSI_OPCODE_END
:
1937 case TGSI_OPCODE_SSG
:
1938 /* TGSI_OPCODE_SGN */
1939 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1940 tmp0
= emit_fetch( bld
, inst
, 0, chan_index
);
1941 dst0
[chan_index
] = lp_build_sgn( &bld
->base
, tmp0
);
1945 case TGSI_OPCODE_CMP
:
1946 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
1947 src0
= emit_fetch( bld
, inst
, 0, chan_index
);
1948 src1
= emit_fetch( bld
, inst
, 1, chan_index
);
1949 src2
= emit_fetch( bld
, inst
, 2, chan_index
);
1950 tmp0
= lp_build_cmp( &bld
->base
, PIPE_FUNC_LESS
, src0
, bld
->base
.zero
);
1951 dst0
[chan_index
] = lp_build_select( &bld
->base
, tmp0
, src1
, src2
);
1955 case TGSI_OPCODE_SCS
:
1956 IF_IS_DST0_CHANNEL_ENABLED( inst
, CHAN_X
) {
1957 tmp0
= emit_fetch( bld
, inst
, 0, CHAN_X
);
1958 dst0
[CHAN_X
] = lp_build_cos( &bld
->base
, tmp0
);
1960 IF_IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Y
) {
1961 tmp0
= emit_fetch( bld
, inst
, 0, CHAN_X
);
1962 dst0
[CHAN_Y
] = lp_build_sin( &bld
->base
, tmp0
);
1964 IF_IS_DST0_CHANNEL_ENABLED( inst
, CHAN_Z
) {
1965 dst0
[CHAN_Z
] = bld
->base
.zero
;
1967 IF_IS_DST0_CHANNEL_ENABLED( inst
, CHAN_W
) {
1968 dst0
[CHAN_W
] = bld
->base
.one
;
1972 case TGSI_OPCODE_TXB
:
1973 emit_tex( bld
, inst
, TEX_MODIFIER_LOD_BIAS
, dst0
);
1976 case TGSI_OPCODE_NRM
:
1978 case TGSI_OPCODE_NRM4
:
1979 /* 3 or 4-component normalization */
1981 uint dims
= (inst
->Instruction
.Opcode
== TGSI_OPCODE_NRM
) ? 3 : 4;
1983 if (IS_DST0_CHANNEL_ENABLED(inst
, CHAN_X
) ||
1984 IS_DST0_CHANNEL_ENABLED(inst
, CHAN_Y
) ||
1985 IS_DST0_CHANNEL_ENABLED(inst
, CHAN_Z
) ||
1986 (IS_DST0_CHANNEL_ENABLED(inst
, CHAN_W
) && dims
== 4)) {
1988 /* NOTE: Cannot use xmm regs 2/3 here (see emit_rsqrt() above). */
1991 /* xmm0 = src.x * src.x */
1992 tmp0
= emit_fetch(bld
, inst
, 0, CHAN_X
);
1993 if (IS_DST0_CHANNEL_ENABLED(inst
, CHAN_X
)) {
1996 tmp0
= lp_build_mul( &bld
->base
, tmp0
, tmp0
);
1999 /* xmm0 = xmm0 + src.y * src.y */
2000 tmp1
= emit_fetch(bld
, inst
, 0, CHAN_Y
);
2001 if (IS_DST0_CHANNEL_ENABLED(inst
, CHAN_Y
)) {
2004 tmp1
= lp_build_mul( &bld
->base
, tmp1
, tmp1
);
2005 tmp0
= lp_build_add( &bld
->base
, tmp0
, tmp1
);
2008 /* xmm0 = xmm0 + src.z * src.z */
2009 tmp1
= emit_fetch(bld
, inst
, 0, CHAN_Z
);
2010 if (IS_DST0_CHANNEL_ENABLED(inst
, CHAN_Z
)) {
2013 tmp1
= lp_build_mul( &bld
->base
, tmp1
, tmp1
);
2014 tmp0
= lp_build_add( &bld
->base
, tmp0
, tmp1
);
2018 /* xmm0 = xmm0 + src.w * src.w */
2019 tmp1
= emit_fetch(bld
, inst
, 0, CHAN_W
);
2020 if (IS_DST0_CHANNEL_ENABLED(inst
, CHAN_W
)) {
2023 tmp1
= lp_build_mul( &bld
->base
, tmp1
, tmp1
);
2024 tmp0
= lp_build_add( &bld
->base
, tmp0
, tmp1
);
2027 /* xmm1 = 1 / sqrt(xmm0) */
2028 tmp1
= lp_build_rsqrt( &bld
->base
, tmp0
);
2030 /* dst.x = xmm1 * src.x */
2031 if (IS_DST0_CHANNEL_ENABLED(inst
, CHAN_X
)) {
2032 dst0
[CHAN_X
] = lp_build_mul( &bld
->base
, tmp4
, tmp1
);
2035 /* dst.y = xmm1 * src.y */
2036 if (IS_DST0_CHANNEL_ENABLED(inst
, CHAN_Y
)) {
2037 dst0
[CHAN_Y
] = lp_build_mul( &bld
->base
, tmp5
, tmp1
);
2040 /* dst.z = xmm1 * src.z */
2041 if (IS_DST0_CHANNEL_ENABLED(inst
, CHAN_Z
)) {
2042 dst0
[CHAN_Z
] = lp_build_mul( &bld
->base
, tmp6
, tmp1
);
2045 /* dst.w = xmm1 * src.w */
2046 if (IS_DST0_CHANNEL_ENABLED(inst
, CHAN_X
) && dims
== 4) {
2047 dst0
[CHAN_W
] = lp_build_mul( &bld
->base
, tmp7
, tmp1
);
2052 if (IS_DST0_CHANNEL_ENABLED(inst
, CHAN_W
) && dims
== 3) {
2053 dst0
[CHAN_W
] = bld
->base
.one
;
2058 case TGSI_OPCODE_DIV
:
2064 case TGSI_OPCODE_DP2
:
2065 tmp0
= emit_fetch( bld
, inst
, 0, CHAN_X
); /* xmm0 = src[0].x */
2066 tmp1
= emit_fetch( bld
, inst
, 1, CHAN_X
); /* xmm1 = src[1].x */
2067 tmp0
= lp_build_mul( &bld
->base
, tmp0
, tmp1
); /* xmm0 = xmm0 * xmm1 */
2068 tmp1
= emit_fetch( bld
, inst
, 0, CHAN_Y
); /* xmm1 = src[0].y */
2069 tmp2
= emit_fetch( bld
, inst
, 1, CHAN_Y
); /* xmm2 = src[1].y */
2070 tmp1
= lp_build_mul( &bld
->base
, tmp1
, tmp2
); /* xmm1 = xmm1 * xmm2 */
2071 tmp0
= lp_build_add( &bld
->base
, tmp0
, tmp1
); /* xmm0 = xmm0 + xmm1 */
2072 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
2073 dst0
[chan_index
] = tmp0
; /* dest[ch] = xmm0 */
2077 case TGSI_OPCODE_TXL
:
2078 emit_tex( bld
, inst
, TEX_MODIFIER_EXPLICIT_LOD
, dst0
);
2081 case TGSI_OPCODE_TXP
:
2082 emit_tex( bld
, inst
, TEX_MODIFIER_PROJECTED
, dst0
);
2085 case TGSI_OPCODE_BRK
:
2086 lp_exec_break(&bld
->exec_mask
);
2089 case TGSI_OPCODE_IF
:
2090 tmp0
= emit_fetch(bld
, inst
, 0, CHAN_X
);
2091 tmp0
= lp_build_cmp(&bld
->base
, PIPE_FUNC_NOTEQUAL
,
2092 tmp0
, bld
->base
.zero
);
2093 lp_exec_mask_cond_push(&bld
->exec_mask
, tmp0
);
2096 case TGSI_OPCODE_BGNLOOP
:
2097 lp_exec_bgnloop(&bld
->exec_mask
);
2100 case TGSI_OPCODE_BGNSUB
:
2104 case TGSI_OPCODE_ELSE
:
2105 lp_exec_mask_cond_invert(&bld
->exec_mask
);
2108 case TGSI_OPCODE_ENDIF
:
2109 lp_exec_mask_cond_pop(&bld
->exec_mask
);
2112 case TGSI_OPCODE_ENDLOOP
:
2113 lp_exec_endloop(&bld
->exec_mask
);
2116 case TGSI_OPCODE_ENDSUB
:
2120 case TGSI_OPCODE_PUSHA
:
2126 case TGSI_OPCODE_POPA
:
2132 case TGSI_OPCODE_CEIL
:
2133 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
2134 tmp0
= emit_fetch( bld
, inst
, 0, chan_index
);
2135 dst0
[chan_index
] = lp_build_ceil(&bld
->base
, tmp0
);
2139 case TGSI_OPCODE_I2F
:
2145 case TGSI_OPCODE_NOT
:
2151 case TGSI_OPCODE_TRUNC
:
2152 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
2153 tmp0
= emit_fetch( bld
, inst
, 0, chan_index
);
2154 dst0
[chan_index
] = lp_build_trunc(&bld
->base
, tmp0
);
2158 case TGSI_OPCODE_SHL
:
2164 case TGSI_OPCODE_ISHR
:
2170 case TGSI_OPCODE_AND
:
2176 case TGSI_OPCODE_OR
:
2182 case TGSI_OPCODE_MOD
:
2188 case TGSI_OPCODE_XOR
:
2194 case TGSI_OPCODE_SAD
:
2200 case TGSI_OPCODE_TXF
:
2206 case TGSI_OPCODE_TXQ
:
2212 case TGSI_OPCODE_CONT
:
2213 lp_exec_continue(&bld
->exec_mask
);
2216 case TGSI_OPCODE_EMIT
:
2220 case TGSI_OPCODE_ENDPRIM
:
2224 case TGSI_OPCODE_NOP
:
2232 LLVMValueRef pred
[NUM_CHANNELS
];
2234 emit_fetch_predicate( bld
, inst
, pred
);
2236 FOR_EACH_DST0_ENABLED_CHANNEL( inst
, chan_index
) {
2237 emit_store( bld
, inst
, 0, chan_index
, pred
[chan_index
], dst0
[chan_index
]);
2246 lp_build_tgsi_soa(LLVMBuilderRef builder
,
2247 const struct tgsi_token
*tokens
,
2248 struct lp_type type
,
2249 struct lp_build_mask_context
*mask
,
2250 LLVMValueRef consts_ptr
,
2251 const LLVMValueRef
*pos
,
2252 const LLVMValueRef (*inputs
)[NUM_CHANNELS
],
2253 LLVMValueRef (*outputs
)[NUM_CHANNELS
],
2254 struct lp_build_sampler_soa
*sampler
,
2255 const struct tgsi_shader_info
*info
)
2257 struct lp_build_tgsi_soa_context bld
;
2258 struct tgsi_parse_context parse
;
2259 uint num_immediates
= 0;
2262 /* Setup build context */
2263 memset(&bld
, 0, sizeof bld
);
2264 lp_build_context_init(&bld
.base
, builder
, type
);
2265 lp_build_context_init(&bld
.int_bld
, builder
, lp_int_type(type
));
2268 bld
.inputs
= inputs
;
2269 bld
.outputs
= outputs
;
2270 bld
.consts_ptr
= consts_ptr
;
2271 bld
.sampler
= sampler
;
2272 bld
.has_indirect_addressing
= info
->opcode_count
[TGSI_OPCODE_ARR
] > 0 ||
2273 info
->opcode_count
[TGSI_OPCODE_ARL
] > 0;
2274 bld
.has_function_calls
= info
->opcode_count
[TGSI_OPCODE_CAL
] > 0;
2275 bld
.func_hash
= cso_hash_create();
2277 lp_exec_mask_init(&bld
.exec_mask
, &bld
.base
);
2279 tgsi_parse_init( &parse
, tokens
);
2281 while( !tgsi_parse_end_of_tokens( &parse
) ) {
2282 tgsi_parse_token( &parse
);
2284 switch( parse
.FullToken
.Token
.Type
) {
2285 case TGSI_TOKEN_TYPE_DECLARATION
:
2286 /* Inputs already interpolated */
2287 emit_declaration( &bld
, &parse
.FullToken
.FullDeclaration
);
2290 case TGSI_TOKEN_TYPE_INSTRUCTION
:
2292 unsigned opcode
= parse
.FullToken
.FullInstruction
.Instruction
.Opcode
;
2293 const struct tgsi_opcode_info
*opcode_info
= tgsi_get_opcode_info(opcode
);
2294 /* we finished processing declarations, emit anything that needs
2295 * to go before the first instruction */
2296 if (bld
.instno
== 0) {
2297 emit_preamble(&bld
, num_immediates
);
2299 if (!emit_instruction( &bld
, &parse
.FullToken
.FullInstruction
, opcode_info
))
2300 _debug_printf("warning: failed to translate tgsi opcode %s to LLVM\n",
2301 opcode_info
->mnemonic
);
2307 case TGSI_TOKEN_TYPE_IMMEDIATE
:
2308 /* simply copy the immediate values into the next immediates[] slot */
2310 const uint size
= parse
.FullToken
.FullImmediate
.Immediate
.NrTokens
- 1;
2312 assert(num_immediates
< LP_MAX_TGSI_IMMEDIATES
);
2313 for( i
= 0; i
< size
; ++i
)
2314 bld
.immediates
[num_immediates
][i
] =
2315 lp_build_const_vec(type
, parse
.FullToken
.FullImmediate
.u
[i
].Float
);
2316 for( i
= size
; i
< 4; ++i
)
2317 bld
.immediates
[num_immediates
][i
] = bld
.base
.undef
;
2322 case TGSI_TOKEN_TYPE_PROPERTY
:
2329 /* we have to make sure we're at the end of the main block
2330 * (which won't be the case if we had more than one TGSI function
2331 * in the given shader) to let the calling function append
2332 * whatever it needs at the end of the main function */
2333 LLVMPositionBuilderAtEnd(bld
.base
.builder
, bld
.main_block
);
2336 LLVMBasicBlockRef block
= LLVMGetInsertBlock(builder
);
2337 LLVMValueRef function
= LLVMGetBasicBlockParent(block
);
2338 tgsi_dump(tokens
, 0);
2339 lp_debug_dump_value(function
);
2341 tgsi_parse_free( &parse
);
2343 cso_hash_delete(bld
.func_hash
);