2 Copyright (C) Intel Corp. 2006. All Rights Reserved.
3 Intel funded Tungsten Graphics (http://www.tungstengraphics.com) to
4 develop this 3D driver.
6 Permission is hereby granted, free of charge, to any person obtaining
7 a copy of this software and associated documentation files (the
8 "Software"), to deal in the Software without restriction, including
9 without limitation the rights to use, copy, modify, merge, publish,
10 distribute, sublicense, and/or sell copies of the Software, and to
11 permit persons to whom the Software is furnished to do so, subject to
12 the following conditions:
14 The above copyright notice and this permission notice (including the
15 next paragraph) shall be included in all copies or substantial
16 portions of the Software.
18 THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
19 EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
20 MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.
21 IN NO EVENT SHALL THE COPYRIGHT OWNER(S) AND/OR ITS SUPPLIERS BE
22 LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
23 OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
24 WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
26 **********************************************************************/
29 * Keith Whitwell <keith@tungstengraphics.com>
33 #include "main/macros.h"
34 #include "brw_context.h"
37 static GLboolean
can_do_pln(struct intel_context
*intel
,
38 const struct brw_reg
*deltas
)
40 struct brw_context
*brw
= brw_context(&intel
->ctx
);
45 if (deltas
[1].nr
!= deltas
[0].nr
+ 1)
48 if (intel
->gen
< 6 && ((deltas
[0].nr
& 1) != 0))
54 /* Not quite sure how correct this is - need to understand horiz
55 * vs. vertical strides a little better.
57 static INLINE
struct brw_reg
sechalf( struct brw_reg reg
)
64 /* Return the SrcReg index of the channels that can be immediate float operands
65 * instead of usage of PROGRAM_CONSTANT values through push/pull.
68 brw_wm_arg_can_be_immediate(enum prog_opcode opcode
, int arg
)
70 int opcode_array
[] = {
90 /* These opcodes get broken down in a way that allow two
91 * args to be immediates.
93 if (opcode
== OPCODE_MAD
|| opcode
== OPCODE_LRP
) {
94 if (arg
== 1 || arg
== 2)
98 if (opcode
> ARRAY_SIZE(opcode_array
))
101 return arg
== opcode_array
[opcode
] - 1;
105 * Computes the screen-space x,y position of the pixels.
107 * This will be used by emit_delta_xy() or emit_wpos_xy() for
108 * interpolation of attributes..
112 * R0.0 -- pixel mask, one bit for each of 4 pixels in 4 tiles,
113 * corresponding to each of the 16 execution channels.
115 * R1.0 -- triangle vertex 0.X
116 * R1.1 -- triangle vertex 0.Y
117 * R1.2 -- tile 0 x,y coords (2 packed uwords)
118 * R1.3 -- tile 1 x,y coords (2 packed uwords)
119 * R1.4 -- tile 2 x,y coords (2 packed uwords)
120 * R1.5 -- tile 3 x,y coords (2 packed uwords)
125 void emit_pixel_xy(struct brw_wm_compile
*c
,
126 const struct brw_reg
*dst
,
129 struct brw_compile
*p
= &c
->func
;
130 struct brw_reg r1
= brw_vec1_grf(1, 0);
131 struct brw_reg r1_uw
= retype(r1
, BRW_REGISTER_TYPE_UW
);
132 struct brw_reg dst0_uw
, dst1_uw
;
134 brw_push_insn_state(p
);
135 brw_set_compression_control(p
, BRW_COMPRESSION_NONE
);
137 if (c
->dispatch_width
== 16) {
138 dst0_uw
= vec16(retype(dst
[0], BRW_REGISTER_TYPE_UW
));
139 dst1_uw
= vec16(retype(dst
[1], BRW_REGISTER_TYPE_UW
));
141 dst0_uw
= vec8(retype(dst
[0], BRW_REGISTER_TYPE_UW
));
142 dst1_uw
= vec8(retype(dst
[1], BRW_REGISTER_TYPE_UW
));
145 /* Calculate pixel centers by adding 1 or 0 to each of the
146 * micro-tile coordinates passed in r1.
148 if (mask
& WRITEMASK_X
) {
151 stride(suboffset(r1_uw
, 4), 2, 4, 0),
152 brw_imm_v(0x10101010));
155 if (mask
& WRITEMASK_Y
) {
158 stride(suboffset(r1_uw
,5), 2, 4, 0),
159 brw_imm_v(0x11001100));
161 brw_pop_insn_state(p
);
165 * Computes the screen-space x,y distance of the pixels from the start
168 * This will be used in linterp or pinterp with the start vertex value
169 * and the Cx, Cy, and C0 coefficients passed in from the setup engine
170 * to produce interpolated attribute values.
172 void emit_delta_xy(struct brw_compile
*p
,
173 const struct brw_reg
*dst
,
175 const struct brw_reg
*arg0
)
177 struct intel_context
*intel
= &p
->brw
->intel
;
178 struct brw_reg r1
= brw_vec1_grf(1, 0);
183 assert(mask
== WRITEMASK_XY
);
185 if (intel
->gen
>= 6) {
186 /* XXX Gen6 WM doesn't have Xstart/Ystart in payload r1.0/r1.1.
187 Just add them with 0.0 for dst reg.. */
188 r1
= brw_imm_v(0x00000000);
191 retype(arg0
[0], BRW_REGISTER_TYPE_UW
),
195 retype(arg0
[1], BRW_REGISTER_TYPE_UW
),
200 /* Calc delta X,Y by subtracting origin in r1 from the pixel
201 * centers produced by emit_pixel_xy().
205 retype(arg0
[0], BRW_REGISTER_TYPE_UW
),
209 retype(arg0
[1], BRW_REGISTER_TYPE_UW
),
210 negate(suboffset(r1
,1)));
214 * Computes the pixel offset from the window origin for gl_FragCoord().
216 void emit_wpos_xy(struct brw_wm_compile
*c
,
217 const struct brw_reg
*dst
,
219 const struct brw_reg
*arg0
)
221 struct brw_compile
*p
= &c
->func
;
223 if (mask
& WRITEMASK_X
) {
224 if (c
->fp
->program
.PixelCenterInteger
) {
228 retype(arg0
[0], BRW_REGISTER_TYPE_W
));
233 retype(arg0
[0], BRW_REGISTER_TYPE_W
),
238 if (mask
& WRITEMASK_Y
) {
239 if (c
->fp
->program
.OriginUpperLeft
) {
240 if (c
->fp
->program
.PixelCenterInteger
) {
244 retype(arg0
[1], BRW_REGISTER_TYPE_W
));
249 retype(arg0
[1], BRW_REGISTER_TYPE_W
),
253 float center_offset
= c
->fp
->program
.PixelCenterInteger
? 0.0 : 0.5;
255 /* Y' = (height - 1) - Y + center */
258 negate(retype(arg0
[1], BRW_REGISTER_TYPE_W
)),
259 brw_imm_f(c
->key
.drawable_height
- 1 + center_offset
));
265 void emit_pixel_w(struct brw_wm_compile
*c
,
266 const struct brw_reg
*dst
,
268 const struct brw_reg
*arg0
,
269 const struct brw_reg
*deltas
)
271 struct brw_compile
*p
= &c
->func
;
272 struct intel_context
*intel
= &p
->brw
->intel
;
274 struct brw_reg temp_dst
;
279 temp_dst
= brw_message_reg(2);
281 assert(intel
->gen
< 6);
283 /* Don't need this if all you are doing is interpolating color, for
286 if (mask
& WRITEMASK_W
) {
287 struct brw_reg interp3
= brw_vec1_grf(arg0
[0].nr
+1, 4);
289 /* Calc 1/w - just linterp wpos[3] optimized by putting the
290 * result straight into a message reg.
292 if (can_do_pln(intel
, deltas
)) {
293 brw_PLN(p
, temp_dst
, interp3
, deltas
[0]);
295 brw_LINE(p
, brw_null_reg(), interp3
, deltas
[0]);
296 brw_MAC(p
, temp_dst
, suboffset(interp3
, 1), deltas
[1]);
303 src
= brw_null_reg();
305 if (c
->dispatch_width
== 16) {
306 brw_math_16(p
, dst
[3],
307 BRW_MATH_FUNCTION_INV
,
308 BRW_MATH_SATURATE_NONE
,
310 BRW_MATH_PRECISION_FULL
);
313 BRW_MATH_FUNCTION_INV
,
314 BRW_MATH_SATURATE_NONE
,
316 BRW_MATH_DATA_VECTOR
,
317 BRW_MATH_PRECISION_FULL
);
322 void emit_linterp(struct brw_compile
*p
,
323 const struct brw_reg
*dst
,
325 const struct brw_reg
*arg0
,
326 const struct brw_reg
*deltas
)
328 struct intel_context
*intel
= &p
->brw
->intel
;
329 struct brw_reg interp
[4];
330 GLuint nr
= arg0
[0].nr
;
333 interp
[0] = brw_vec1_grf(nr
, 0);
334 interp
[1] = brw_vec1_grf(nr
, 4);
335 interp
[2] = brw_vec1_grf(nr
+1, 0);
336 interp
[3] = brw_vec1_grf(nr
+1, 4);
338 for (i
= 0; i
< 4; i
++) {
340 if (intel
->gen
>= 6) {
341 brw_PLN(p
, dst
[i
], interp
[i
], brw_vec8_grf(2, 0));
342 } else if (can_do_pln(intel
, deltas
)) {
343 brw_PLN(p
, dst
[i
], interp
[i
], deltas
[0]);
345 brw_LINE(p
, brw_null_reg(), interp
[i
], deltas
[0]);
346 brw_MAC(p
, dst
[i
], suboffset(interp
[i
],1), deltas
[1]);
353 void emit_pinterp(struct brw_compile
*p
,
354 const struct brw_reg
*dst
,
356 const struct brw_reg
*arg0
,
357 const struct brw_reg
*deltas
,
358 const struct brw_reg
*w
)
360 struct intel_context
*intel
= &p
->brw
->intel
;
361 struct brw_reg interp
[4];
362 GLuint nr
= arg0
[0].nr
;
365 if (intel
->gen
>= 6) {
366 emit_linterp(p
, dst
, mask
, arg0
, interp
);
370 interp
[0] = brw_vec1_grf(nr
, 0);
371 interp
[1] = brw_vec1_grf(nr
, 4);
372 interp
[2] = brw_vec1_grf(nr
+1, 0);
373 interp
[3] = brw_vec1_grf(nr
+1, 4);
375 for (i
= 0; i
< 4; i
++) {
377 if (can_do_pln(intel
, deltas
)) {
378 brw_PLN(p
, dst
[i
], interp
[i
], deltas
[0]);
380 brw_LINE(p
, brw_null_reg(), interp
[i
], deltas
[0]);
381 brw_MAC(p
, dst
[i
], suboffset(interp
[i
],1), deltas
[1]);
385 for (i
= 0; i
< 4; i
++) {
387 brw_MUL(p
, dst
[i
], dst
[i
], w
[3]);
393 void emit_cinterp(struct brw_compile
*p
,
394 const struct brw_reg
*dst
,
396 const struct brw_reg
*arg0
)
398 struct brw_reg interp
[4];
399 GLuint nr
= arg0
[0].nr
;
402 interp
[0] = brw_vec1_grf(nr
, 0);
403 interp
[1] = brw_vec1_grf(nr
, 4);
404 interp
[2] = brw_vec1_grf(nr
+1, 0);
405 interp
[3] = brw_vec1_grf(nr
+1, 4);
407 for (i
= 0; i
< 4; i
++) {
409 brw_MOV(p
, dst
[i
], suboffset(interp
[i
],3)); /* TODO: optimize away like other moves */
414 /* Sets the destination channels to 1.0 or 0.0 according to glFrontFacing. */
415 void emit_frontfacing(struct brw_compile
*p
,
416 const struct brw_reg
*dst
,
419 struct brw_reg r1_6ud
= retype(brw_vec1_grf(1, 6), BRW_REGISTER_TYPE_UD
);
422 if (!(mask
& WRITEMASK_XYZW
))
425 for (i
= 0; i
< 4; i
++) {
427 brw_MOV(p
, dst
[i
], brw_imm_f(0.0));
431 /* bit 31 is "primitive is back face", so checking < (1 << 31) gives
434 brw_CMP(p
, brw_null_reg(), BRW_CONDITIONAL_L
, r1_6ud
, brw_imm_ud(1 << 31));
435 for (i
= 0; i
< 4; i
++) {
437 brw_MOV(p
, dst
[i
], brw_imm_f(1.0));
440 brw_set_predicate_control_flag_value(p
, 0xff);
443 /* For OPCODE_DDX and OPCODE_DDY, per channel of output we've got input
446 * arg0: ss0.tl ss0.tr ss0.bl ss0.br ss1.tl ss1.tr ss1.bl ss1.br
448 * and we're trying to produce:
451 * dst: (ss0.tr - ss0.tl) (ss0.tl - ss0.bl)
452 * (ss0.tr - ss0.tl) (ss0.tr - ss0.br)
453 * (ss0.br - ss0.bl) (ss0.tl - ss0.bl)
454 * (ss0.br - ss0.bl) (ss0.tr - ss0.br)
455 * (ss1.tr - ss1.tl) (ss1.tl - ss1.bl)
456 * (ss1.tr - ss1.tl) (ss1.tr - ss1.br)
457 * (ss1.br - ss1.bl) (ss1.tl - ss1.bl)
458 * (ss1.br - ss1.bl) (ss1.tr - ss1.br)
460 * and add another set of two more subspans if in 16-pixel dispatch mode.
462 * For DDX, it ends up being easy: width = 2, horiz=0 gets us the same result
463 * for each pair, and vertstride = 2 jumps us 2 elements after processing a
464 * pair. But for DDY, it's harder, as we want to produce the pairs swizzled
465 * between each other. We could probably do it like ddx and swizzle the right
466 * order later, but bail for now and just produce
467 * ((ss0.tl - ss0.bl)x4 (ss1.tl - ss1.bl)x4)
469 void emit_ddxy(struct brw_compile
*p
,
470 const struct brw_reg
*dst
,
473 const struct brw_reg
*arg0
)
476 struct brw_reg src0
, src1
;
479 brw_set_saturate(p
, 1);
480 for (i
= 0; i
< 4; i
++ ) {
483 src0
= brw_reg(arg0
[i
].file
, arg0
[i
].nr
, 1,
485 BRW_VERTICAL_STRIDE_2
,
487 BRW_HORIZONTAL_STRIDE_0
,
488 BRW_SWIZZLE_XYZW
, WRITEMASK_XYZW
);
489 src1
= brw_reg(arg0
[i
].file
, arg0
[i
].nr
, 0,
491 BRW_VERTICAL_STRIDE_2
,
493 BRW_HORIZONTAL_STRIDE_0
,
494 BRW_SWIZZLE_XYZW
, WRITEMASK_XYZW
);
496 src0
= brw_reg(arg0
[i
].file
, arg0
[i
].nr
, 0,
498 BRW_VERTICAL_STRIDE_4
,
500 BRW_HORIZONTAL_STRIDE_0
,
501 BRW_SWIZZLE_XYZW
, WRITEMASK_XYZW
);
502 src1
= brw_reg(arg0
[i
].file
, arg0
[i
].nr
, 2,
504 BRW_VERTICAL_STRIDE_4
,
506 BRW_HORIZONTAL_STRIDE_0
,
507 BRW_SWIZZLE_XYZW
, WRITEMASK_XYZW
);
509 brw_ADD(p
, dst
[i
], src0
, negate(src1
));
513 brw_set_saturate(p
, 0);
516 void emit_alu1(struct brw_compile
*p
,
517 struct brw_instruction
*(*func
)(struct brw_compile
*,
520 const struct brw_reg
*dst
,
522 const struct brw_reg
*arg0
)
527 brw_set_saturate(p
, 1);
529 for (i
= 0; i
< 4; i
++) {
531 func(p
, dst
[i
], arg0
[i
]);
536 brw_set_saturate(p
, 0);
540 void emit_alu2(struct brw_compile
*p
,
541 struct brw_instruction
*(*func
)(struct brw_compile
*,
545 const struct brw_reg
*dst
,
547 const struct brw_reg
*arg0
,
548 const struct brw_reg
*arg1
)
553 brw_set_saturate(p
, 1);
555 for (i
= 0; i
< 4; i
++) {
557 func(p
, dst
[i
], arg0
[i
], arg1
[i
]);
562 brw_set_saturate(p
, 0);
566 void emit_mad(struct brw_compile
*p
,
567 const struct brw_reg
*dst
,
569 const struct brw_reg
*arg0
,
570 const struct brw_reg
*arg1
,
571 const struct brw_reg
*arg2
)
575 for (i
= 0; i
< 4; i
++) {
577 brw_MUL(p
, dst
[i
], arg0
[i
], arg1
[i
]);
579 brw_set_saturate(p
, (mask
& SATURATE
) ? 1 : 0);
580 brw_ADD(p
, dst
[i
], dst
[i
], arg2
[i
]);
581 brw_set_saturate(p
, 0);
586 void emit_lrp(struct brw_compile
*p
,
587 const struct brw_reg
*dst
,
589 const struct brw_reg
*arg0
,
590 const struct brw_reg
*arg1
,
591 const struct brw_reg
*arg2
)
595 /* Uses dst as a temporary:
597 for (i
= 0; i
< 4; i
++) {
599 /* Can I use the LINE instruction for this?
601 brw_ADD(p
, dst
[i
], negate(arg0
[i
]), brw_imm_f(1.0));
602 brw_MUL(p
, brw_null_reg(), dst
[i
], arg2
[i
]);
604 brw_set_saturate(p
, (mask
& SATURATE
) ? 1 : 0);
605 brw_MAC(p
, dst
[i
], arg0
[i
], arg1
[i
]);
606 brw_set_saturate(p
, 0);
611 void emit_sop(struct brw_compile
*p
,
612 const struct brw_reg
*dst
,
615 const struct brw_reg
*arg0
,
616 const struct brw_reg
*arg1
)
620 for (i
= 0; i
< 4; i
++) {
622 brw_push_insn_state(p
);
623 brw_CMP(p
, brw_null_reg(), cond
, arg0
[i
], arg1
[i
]);
624 brw_set_predicate_control(p
, BRW_PREDICATE_NONE
);
625 brw_MOV(p
, dst
[i
], brw_imm_f(0));
626 brw_set_predicate_control(p
, BRW_PREDICATE_NORMAL
);
627 brw_MOV(p
, dst
[i
], brw_imm_f(1.0));
628 brw_pop_insn_state(p
);
633 static void emit_slt( struct brw_compile
*p
,
634 const struct brw_reg
*dst
,
636 const struct brw_reg
*arg0
,
637 const struct brw_reg
*arg1
)
639 emit_sop(p
, dst
, mask
, BRW_CONDITIONAL_L
, arg0
, arg1
);
642 static void emit_sle( struct brw_compile
*p
,
643 const struct brw_reg
*dst
,
645 const struct brw_reg
*arg0
,
646 const struct brw_reg
*arg1
)
648 emit_sop(p
, dst
, mask
, BRW_CONDITIONAL_LE
, arg0
, arg1
);
651 static void emit_sgt( struct brw_compile
*p
,
652 const struct brw_reg
*dst
,
654 const struct brw_reg
*arg0
,
655 const struct brw_reg
*arg1
)
657 emit_sop(p
, dst
, mask
, BRW_CONDITIONAL_G
, arg0
, arg1
);
660 static void emit_sge( struct brw_compile
*p
,
661 const struct brw_reg
*dst
,
663 const struct brw_reg
*arg0
,
664 const struct brw_reg
*arg1
)
666 emit_sop(p
, dst
, mask
, BRW_CONDITIONAL_GE
, arg0
, arg1
);
669 static void emit_seq( struct brw_compile
*p
,
670 const struct brw_reg
*dst
,
672 const struct brw_reg
*arg0
,
673 const struct brw_reg
*arg1
)
675 emit_sop(p
, dst
, mask
, BRW_CONDITIONAL_EQ
, arg0
, arg1
);
678 static void emit_sne( struct brw_compile
*p
,
679 const struct brw_reg
*dst
,
681 const struct brw_reg
*arg0
,
682 const struct brw_reg
*arg1
)
684 emit_sop(p
, dst
, mask
, BRW_CONDITIONAL_NEQ
, arg0
, arg1
);
687 void emit_cmp(struct brw_compile
*p
,
688 const struct brw_reg
*dst
,
690 const struct brw_reg
*arg0
,
691 const struct brw_reg
*arg1
,
692 const struct brw_reg
*arg2
)
696 for (i
= 0; i
< 4; i
++) {
698 brw_CMP(p
, brw_null_reg(), BRW_CONDITIONAL_L
, arg0
[i
], brw_imm_f(0));
700 brw_set_saturate(p
, (mask
& SATURATE
) ? 1 : 0);
701 brw_SEL(p
, dst
[i
], arg1
[i
], arg2
[i
]);
702 brw_set_saturate(p
, 0);
703 brw_set_predicate_control_flag_value(p
, 0xff);
708 void emit_sign(struct brw_compile
*p
,
709 const struct brw_reg
*dst
,
711 const struct brw_reg
*arg0
)
715 for (i
= 0; i
< 4; i
++) {
717 brw_MOV(p
, dst
[i
], brw_imm_f(0.0));
719 brw_CMP(p
, brw_null_reg(), BRW_CONDITIONAL_L
, arg0
[i
], brw_imm_f(0));
720 brw_MOV(p
, dst
[i
], brw_imm_f(-1.0));
721 brw_set_predicate_control(p
, BRW_PREDICATE_NONE
);
723 brw_CMP(p
, brw_null_reg(), BRW_CONDITIONAL_G
, arg0
[i
], brw_imm_f(0));
724 brw_MOV(p
, dst
[i
], brw_imm_f(1.0));
725 brw_set_predicate_control(p
, BRW_PREDICATE_NONE
);
730 void emit_max(struct brw_compile
*p
,
731 const struct brw_reg
*dst
,
733 const struct brw_reg
*arg0
,
734 const struct brw_reg
*arg1
)
738 for (i
= 0; i
< 4; i
++) {
740 brw_CMP(p
, brw_null_reg(), BRW_CONDITIONAL_GE
, arg0
[i
], arg1
[i
]);
742 brw_set_saturate(p
, (mask
& SATURATE
) ? 1 : 0);
743 brw_SEL(p
, dst
[i
], arg0
[i
], arg1
[i
]);
744 brw_set_saturate(p
, 0);
745 brw_set_predicate_control_flag_value(p
, 0xff);
750 void emit_min(struct brw_compile
*p
,
751 const struct brw_reg
*dst
,
753 const struct brw_reg
*arg0
,
754 const struct brw_reg
*arg1
)
758 for (i
= 0; i
< 4; i
++) {
760 brw_CMP(p
, brw_null_reg(), BRW_CONDITIONAL_L
, arg0
[i
], arg1
[i
]);
762 brw_set_saturate(p
, (mask
& SATURATE
) ? 1 : 0);
763 brw_SEL(p
, dst
[i
], arg0
[i
], arg1
[i
]);
764 brw_set_saturate(p
, 0);
765 brw_set_predicate_control_flag_value(p
, 0xff);
771 void emit_dp2(struct brw_compile
*p
,
772 const struct brw_reg
*dst
,
774 const struct brw_reg
*arg0
,
775 const struct brw_reg
*arg1
)
777 int dst_chan
= _mesa_ffs(mask
& WRITEMASK_XYZW
) - 1;
779 if (!(mask
& WRITEMASK_XYZW
))
780 return; /* Do not emit dead code */
782 assert(is_power_of_two(mask
& WRITEMASK_XYZW
));
784 brw_MUL(p
, brw_null_reg(), arg0
[0], arg1
[0]);
786 brw_set_saturate(p
, (mask
& SATURATE
) ? 1 : 0);
787 brw_MAC(p
, dst
[dst_chan
], arg0
[1], arg1
[1]);
788 brw_set_saturate(p
, 0);
792 void emit_dp3(struct brw_compile
*p
,
793 const struct brw_reg
*dst
,
795 const struct brw_reg
*arg0
,
796 const struct brw_reg
*arg1
)
798 int dst_chan
= _mesa_ffs(mask
& WRITEMASK_XYZW
) - 1;
800 if (!(mask
& WRITEMASK_XYZW
))
801 return; /* Do not emit dead code */
803 assert(is_power_of_two(mask
& WRITEMASK_XYZW
));
805 brw_MUL(p
, brw_null_reg(), arg0
[0], arg1
[0]);
806 brw_MAC(p
, brw_null_reg(), arg0
[1], arg1
[1]);
808 brw_set_saturate(p
, (mask
& SATURATE
) ? 1 : 0);
809 brw_MAC(p
, dst
[dst_chan
], arg0
[2], arg1
[2]);
810 brw_set_saturate(p
, 0);
814 void emit_dp4(struct brw_compile
*p
,
815 const struct brw_reg
*dst
,
817 const struct brw_reg
*arg0
,
818 const struct brw_reg
*arg1
)
820 int dst_chan
= _mesa_ffs(mask
& WRITEMASK_XYZW
) - 1;
822 if (!(mask
& WRITEMASK_XYZW
))
823 return; /* Do not emit dead code */
825 assert(is_power_of_two(mask
& WRITEMASK_XYZW
));
827 brw_MUL(p
, brw_null_reg(), arg0
[0], arg1
[0]);
828 brw_MAC(p
, brw_null_reg(), arg0
[1], arg1
[1]);
829 brw_MAC(p
, brw_null_reg(), arg0
[2], arg1
[2]);
831 brw_set_saturate(p
, (mask
& SATURATE
) ? 1 : 0);
832 brw_MAC(p
, dst
[dst_chan
], arg0
[3], arg1
[3]);
833 brw_set_saturate(p
, 0);
837 void emit_dph(struct brw_compile
*p
,
838 const struct brw_reg
*dst
,
840 const struct brw_reg
*arg0
,
841 const struct brw_reg
*arg1
)
843 const int dst_chan
= _mesa_ffs(mask
& WRITEMASK_XYZW
) - 1;
845 if (!(mask
& WRITEMASK_XYZW
))
846 return; /* Do not emit dead code */
848 assert(is_power_of_two(mask
& WRITEMASK_XYZW
));
850 brw_MUL(p
, brw_null_reg(), arg0
[0], arg1
[0]);
851 brw_MAC(p
, brw_null_reg(), arg0
[1], arg1
[1]);
852 brw_MAC(p
, dst
[dst_chan
], arg0
[2], arg1
[2]);
854 brw_set_saturate(p
, (mask
& SATURATE
) ? 1 : 0);
855 brw_ADD(p
, dst
[dst_chan
], dst
[dst_chan
], arg1
[3]);
856 brw_set_saturate(p
, 0);
860 void emit_xpd(struct brw_compile
*p
,
861 const struct brw_reg
*dst
,
863 const struct brw_reg
*arg0
,
864 const struct brw_reg
*arg1
)
868 assert((mask
& WRITEMASK_W
) != WRITEMASK_W
);
870 for (i
= 0 ; i
< 3; i
++) {
875 brw_MUL(p
, brw_null_reg(), negate(arg0
[i2
]), arg1
[i1
]);
877 brw_set_saturate(p
, (mask
& SATURATE
) ? 1 : 0);
878 brw_MAC(p
, dst
[i
], arg0
[i1
], arg1
[i2
]);
879 brw_set_saturate(p
, 0);
885 void emit_math1(struct brw_wm_compile
*c
,
887 const struct brw_reg
*dst
,
889 const struct brw_reg
*arg0
)
891 struct brw_compile
*p
= &c
->func
;
892 struct intel_context
*intel
= &p
->brw
->intel
;
893 int dst_chan
= _mesa_ffs(mask
& WRITEMASK_XYZW
) - 1;
894 GLuint saturate
= ((mask
& SATURATE
) ?
895 BRW_MATH_SATURATE_SATURATE
:
896 BRW_MATH_SATURATE_NONE
);
899 if (intel
->gen
>= 6 && ((arg0
[0].hstride
== BRW_HORIZONTAL_STRIDE_0
||
900 arg0
[0].file
!= BRW_GENERAL_REGISTER_FILE
) ||
901 arg0
[0].negate
|| arg0
[0].abs
)) {
902 /* Gen6 math requires that source and dst horizontal stride be 1,
903 * and that the argument be in the GRF.
905 * The hardware ignores source modifiers (negate and abs) on math
906 * instructions, so we also move to a temp to set those up.
909 brw_MOV(p
, src
, arg0
[0]);
914 if (!(mask
& WRITEMASK_XYZW
))
915 return; /* Do not emit dead code */
917 assert(is_power_of_two(mask
& WRITEMASK_XYZW
));
919 /* Send two messages to perform all 16 operations:
921 brw_push_insn_state(p
);
922 brw_set_compression_control(p
, BRW_COMPRESSION_NONE
);
929 BRW_MATH_DATA_VECTOR
,
930 BRW_MATH_PRECISION_FULL
);
932 if (c
->dispatch_width
== 16) {
933 brw_set_compression_control(p
, BRW_COMPRESSION_2NDHALF
);
935 offset(dst
[dst_chan
],1),
940 BRW_MATH_DATA_VECTOR
,
941 BRW_MATH_PRECISION_FULL
);
943 brw_pop_insn_state(p
);
947 void emit_math2(struct brw_wm_compile
*c
,
949 const struct brw_reg
*dst
,
951 const struct brw_reg
*arg0
,
952 const struct brw_reg
*arg1
)
954 struct brw_compile
*p
= &c
->func
;
955 struct intel_context
*intel
= &p
->brw
->intel
;
956 int dst_chan
= _mesa_ffs(mask
& WRITEMASK_XYZW
) - 1;
958 if (!(mask
& WRITEMASK_XYZW
))
959 return; /* Do not emit dead code */
961 assert(is_power_of_two(mask
& WRITEMASK_XYZW
));
963 brw_push_insn_state(p
);
965 /* math can only operate on up to a vec8 at a time, so in
966 * dispatch_width==16 we have to do the second half manually.
968 if (intel
->gen
>= 6) {
969 struct brw_reg src0
= arg0
[0];
970 struct brw_reg src1
= arg1
[0];
971 struct brw_reg temp_dst
= dst
[dst_chan
];
973 if (arg0
[0].hstride
== BRW_HORIZONTAL_STRIDE_0
) {
974 if (arg1
[0].hstride
== BRW_HORIZONTAL_STRIDE_0
) {
975 /* Both scalar arguments. Do scalar calc. */
976 src0
.hstride
= BRW_HORIZONTAL_STRIDE_1
;
977 src1
.hstride
= BRW_HORIZONTAL_STRIDE_1
;
978 temp_dst
.hstride
= BRW_HORIZONTAL_STRIDE_1
;
979 temp_dst
.width
= BRW_WIDTH_1
;
981 if (arg0
[0].subnr
!= 0) {
982 brw_MOV(p
, temp_dst
, src0
);
985 /* Ouch. We've used the temp as a dst, and we still
986 * need a temp to store arg1 in, because src and dst
987 * offsets have to be equal. Leaving this up to
988 * glsl2-965 to handle correctly.
990 assert(arg1
[0].subnr
== 0);
991 } else if (arg1
[0].subnr
!= 0) {
992 brw_MOV(p
, temp_dst
, src1
);
996 brw_MOV(p
, temp_dst
, src0
);
999 } else if (arg1
[0].hstride
== BRW_HORIZONTAL_STRIDE_0
) {
1000 brw_MOV(p
, temp_dst
, src1
);
1004 brw_set_saturate(p
, (mask
& SATURATE
) ? 1 : 0);
1005 brw_set_compression_control(p
, BRW_COMPRESSION_NONE
);
1011 if (c
->dispatch_width
== 16) {
1012 brw_set_compression_control(p
, BRW_COMPRESSION_2NDHALF
);
1020 /* Splat a scalar result into all the channels. */
1021 if (arg0
[0].hstride
== BRW_HORIZONTAL_STRIDE_0
&&
1022 arg1
[0].hstride
== BRW_HORIZONTAL_STRIDE_0
) {
1023 temp_dst
.hstride
= BRW_HORIZONTAL_STRIDE_0
;
1024 temp_dst
.vstride
= BRW_VERTICAL_STRIDE_0
;
1025 brw_MOV(p
, dst
[dst_chan
], temp_dst
);
1028 GLuint saturate
= ((mask
& SATURATE
) ?
1029 BRW_MATH_SATURATE_SATURATE
:
1030 BRW_MATH_SATURATE_NONE
);
1032 brw_set_compression_control(p
, BRW_COMPRESSION_NONE
);
1033 brw_MOV(p
, brw_message_reg(3), arg1
[0]);
1034 if (c
->dispatch_width
== 16) {
1035 brw_set_compression_control(p
, BRW_COMPRESSION_2NDHALF
);
1036 brw_MOV(p
, brw_message_reg(5), sechalf(arg1
[0]));
1039 brw_set_compression_control(p
, BRW_COMPRESSION_NONE
);
1046 BRW_MATH_DATA_VECTOR
,
1047 BRW_MATH_PRECISION_FULL
);
1049 /* Send two messages to perform all 16 operations:
1051 if (c
->dispatch_width
== 16) {
1052 brw_set_compression_control(p
, BRW_COMPRESSION_2NDHALF
);
1054 offset(dst
[dst_chan
],1),
1059 BRW_MATH_DATA_VECTOR
,
1060 BRW_MATH_PRECISION_FULL
);
1063 brw_pop_insn_state(p
);
1067 void emit_tex(struct brw_wm_compile
*c
,
1068 struct brw_reg
*dst
,
1070 struct brw_reg
*arg
,
1071 struct brw_reg depth_payload
,
1076 struct brw_compile
*p
= &c
->func
;
1077 struct intel_context
*intel
= &p
->brw
->intel
;
1078 struct brw_reg dst_retyped
;
1079 GLuint cur_mrf
= 2, response_length
;
1080 GLuint i
, nr_texcoords
;
1083 GLuint mrf_per_channel
;
1086 if (c
->dispatch_width
== 16) {
1087 mrf_per_channel
= 2;
1088 response_length
= 8;
1089 dst_retyped
= retype(vec16(dst
[0]), BRW_REGISTER_TYPE_UW
);
1090 simd_mode
= BRW_SAMPLER_SIMD_MODE_SIMD16
;
1092 mrf_per_channel
= 1;
1093 response_length
= 4;
1094 dst_retyped
= retype(vec8(dst
[0]), BRW_REGISTER_TYPE_UW
);
1095 simd_mode
= BRW_SAMPLER_SIMD_MODE_SIMD8
;
1098 /* How many input regs are there?
1101 case TEXTURE_1D_INDEX
:
1105 case TEXTURE_2D_INDEX
:
1106 case TEXTURE_RECT_INDEX
:
1107 emit
= WRITEMASK_XY
;
1110 case TEXTURE_3D_INDEX
:
1111 case TEXTURE_CUBE_INDEX
:
1112 emit
= WRITEMASK_XYZ
;
1116 /* unexpected target */
1120 /* Pre-Ironlake, the 8-wide sampler always took u,v,r. */
1121 if (intel
->gen
< 5 && c
->dispatch_width
== 8)
1124 /* For shadow comparisons, we have to supply u,v,r. */
1128 /* Emit the texcoords. */
1129 for (i
= 0; i
< nr_texcoords
; i
++) {
1131 brw_MOV(p
, brw_message_reg(cur_mrf
), arg
[i
]);
1133 brw_MOV(p
, brw_message_reg(cur_mrf
), brw_imm_f(0));
1134 cur_mrf
+= mrf_per_channel
;
1137 /* Fill in the shadow comparison reference value. */
1139 if (intel
->gen
>= 5) {
1140 /* Fill in the cube map array index value. */
1141 brw_MOV(p
, brw_message_reg(cur_mrf
), brw_imm_f(0));
1142 cur_mrf
+= mrf_per_channel
;
1143 } else if (c
->dispatch_width
== 8) {
1144 /* Fill in the LOD bias value. */
1145 brw_MOV(p
, brw_message_reg(cur_mrf
), brw_imm_f(0));
1146 cur_mrf
+= mrf_per_channel
;
1148 brw_MOV(p
, brw_message_reg(cur_mrf
), arg
[2]);
1149 cur_mrf
+= mrf_per_channel
;
1152 if (intel
->gen
>= 5) {
1154 msg_type
= BRW_SAMPLER_MESSAGE_SAMPLE_COMPARE_GEN5
;
1156 msg_type
= BRW_SAMPLER_MESSAGE_SAMPLE_GEN5
;
1158 /* Note that G45 and older determines shadow compare and dispatch width
1159 * from message length for most messages.
1161 if (c
->dispatch_width
== 16 && shadow
)
1162 msg_type
= BRW_SAMPLER_MESSAGE_SIMD16_SAMPLE_COMPARE
;
1164 msg_type
= BRW_SAMPLER_MESSAGE_SIMD16_SAMPLE
;
1170 retype(depth_payload
, BRW_REGISTER_TYPE_UW
),
1171 SURF_INDEX_TEXTURE(sampler
),
1173 dst_flags
& WRITEMASK_XYZW
,
1183 void emit_txb(struct brw_wm_compile
*c
,
1184 struct brw_reg
*dst
,
1186 struct brw_reg
*arg
,
1187 struct brw_reg depth_payload
,
1191 struct brw_compile
*p
= &c
->func
;
1192 struct intel_context
*intel
= &p
->brw
->intel
;
1195 GLuint mrf_per_channel
;
1196 GLuint response_length
;
1197 struct brw_reg dst_retyped
;
1199 /* The G45 and older chipsets don't support 8-wide dispatch for LOD biased
1200 * samples, so we'll use the 16-wide instruction, leave the second halves
1201 * undefined, and trust the execution mask to keep the undefined pixels
1204 if (c
->dispatch_width
== 16 || intel
->gen
< 5) {
1205 if (intel
->gen
>= 5)
1206 msg_type
= BRW_SAMPLER_MESSAGE_SAMPLE_BIAS_GEN5
;
1208 msg_type
= BRW_SAMPLER_MESSAGE_SIMD16_SAMPLE_BIAS
;
1209 mrf_per_channel
= 2;
1210 dst_retyped
= retype(vec16(dst
[0]), BRW_REGISTER_TYPE_UW
);
1211 response_length
= 8;
1213 msg_type
= BRW_SAMPLER_MESSAGE_SAMPLE_BIAS_GEN5
;
1214 mrf_per_channel
= 1;
1215 dst_retyped
= retype(vec8(dst
[0]), BRW_REGISTER_TYPE_UW
);
1216 response_length
= 4;
1219 /* Shadow ignored for txb. */
1221 case TEXTURE_1D_INDEX
:
1222 brw_MOV(p
, brw_message_reg(2 + 0 * mrf_per_channel
), arg
[0]);
1223 brw_MOV(p
, brw_message_reg(2 + 1 * mrf_per_channel
), brw_imm_f(0));
1224 brw_MOV(p
, brw_message_reg(2 + 2 * mrf_per_channel
), brw_imm_f(0));
1226 case TEXTURE_2D_INDEX
:
1227 case TEXTURE_RECT_INDEX
:
1228 brw_MOV(p
, brw_message_reg(2 + 0 * mrf_per_channel
), arg
[0]);
1229 brw_MOV(p
, brw_message_reg(2 + 1 * mrf_per_channel
), arg
[1]);
1230 brw_MOV(p
, brw_message_reg(2 + 2 * mrf_per_channel
), brw_imm_f(0));
1232 case TEXTURE_3D_INDEX
:
1233 case TEXTURE_CUBE_INDEX
:
1234 brw_MOV(p
, brw_message_reg(2 + 0 * mrf_per_channel
), arg
[0]);
1235 brw_MOV(p
, brw_message_reg(2 + 1 * mrf_per_channel
), arg
[1]);
1236 brw_MOV(p
, brw_message_reg(2 + 2 * mrf_per_channel
), arg
[2]);
1239 /* unexpected target */
1243 brw_MOV(p
, brw_message_reg(2 + 3 * mrf_per_channel
), arg
[3]);
1244 msgLength
= 2 + 4 * mrf_per_channel
- 1;
1249 retype(depth_payload
, BRW_REGISTER_TYPE_UW
),
1250 SURF_INDEX_TEXTURE(sampler
),
1252 dst_flags
& WRITEMASK_XYZW
,
1258 BRW_SAMPLER_SIMD_MODE_SIMD16
);
1262 static void emit_lit(struct brw_wm_compile
*c
,
1263 const struct brw_reg
*dst
,
1265 const struct brw_reg
*arg0
)
1267 struct brw_compile
*p
= &c
->func
;
1269 assert((mask
& WRITEMASK_XW
) == 0);
1271 if (mask
& WRITEMASK_Y
) {
1272 brw_set_saturate(p
, (mask
& SATURATE
) ? 1 : 0);
1273 brw_MOV(p
, dst
[1], arg0
[0]);
1274 brw_set_saturate(p
, 0);
1277 if (mask
& WRITEMASK_Z
) {
1278 emit_math2(c
, BRW_MATH_FUNCTION_POW
,
1280 WRITEMASK_X
| (mask
& SATURATE
),
1285 /* Ordinarily you'd use an iff statement to skip or shortcircuit
1286 * some of the POW calculations above, but 16-wide iff statements
1287 * seem to lock c1 hardware, so this is a nasty workaround:
1289 brw_CMP(p
, brw_null_reg(), BRW_CONDITIONAL_LE
, arg0
[0], brw_imm_f(0));
1291 if (mask
& WRITEMASK_Y
)
1292 brw_MOV(p
, dst
[1], brw_imm_f(0));
1294 if (mask
& WRITEMASK_Z
)
1295 brw_MOV(p
, dst
[2], brw_imm_f(0));
1297 brw_set_predicate_control(p
, BRW_PREDICATE_NONE
);
1301 /* Kill pixel - set execution mask to zero for those pixels which
1304 static void emit_kil( struct brw_wm_compile
*c
,
1305 struct brw_reg
*arg0
)
1307 struct brw_compile
*p
= &c
->func
;
1308 struct intel_context
*intel
= &p
->brw
->intel
;
1309 struct brw_reg pixelmask
;
1312 if (intel
->gen
>= 6)
1313 pixelmask
= retype(brw_vec1_grf(1, 7), BRW_REGISTER_TYPE_UW
);
1315 pixelmask
= retype(brw_vec1_grf(0, 0), BRW_REGISTER_TYPE_UW
);
1317 for (i
= 0; i
< 4; i
++) {
1318 /* Check if we've already done the comparison for this reg
1319 * -- common when someone does KIL TEMP.wwww.
1321 for (j
= 0; j
< i
; j
++) {
1322 if (memcmp(&arg0
[j
], &arg0
[i
], sizeof(arg0
[0])) == 0)
1328 brw_push_insn_state(p
);
1329 brw_CMP(p
, brw_null_reg(), BRW_CONDITIONAL_GE
, arg0
[i
], brw_imm_f(0));
1330 brw_set_predicate_control_flag_value(p
, 0xff);
1331 brw_set_compression_control(p
, BRW_COMPRESSION_NONE
);
1332 brw_AND(p
, pixelmask
, brw_flag_reg(), pixelmask
);
1333 brw_pop_insn_state(p
);
1337 static void fire_fb_write( struct brw_wm_compile
*c
,
1343 struct brw_compile
*p
= &c
->func
;
1344 struct intel_context
*intel
= &p
->brw
->intel
;
1347 if (c
->dispatch_width
== 16)
1348 dst
= retype(vec16(brw_null_reg()), BRW_REGISTER_TYPE_UW
);
1350 dst
= retype(vec8(brw_null_reg()), BRW_REGISTER_TYPE_UW
);
1352 /* Pass through control information:
1354 /* mov (8) m1.0<1>:ud r1.0<8;8,1>:ud { Align1 NoMask } */
1355 if (intel
->gen
< 6) /* gen6, use headerless for fb write */
1357 brw_push_insn_state(p
);
1358 brw_set_mask_control(p
, BRW_MASK_DISABLE
); /* ? */
1359 brw_set_compression_control(p
, BRW_COMPRESSION_NONE
);
1361 brw_message_reg(base_reg
+ 1),
1362 brw_vec8_grf(1, 0));
1363 brw_pop_insn_state(p
);
1366 /* Send framebuffer write message: */
1367 /* send (16) null.0<1>:uw m0 r0.0<8;8,1>:uw 0x85a04000:ud { Align1 EOT } */
1372 retype(brw_vec8_grf(0, 0), BRW_REGISTER_TYPE_UW
),
1380 static void emit_aa( struct brw_wm_compile
*c
,
1381 struct brw_reg
*arg1
,
1384 struct brw_compile
*p
= &c
->func
;
1385 GLuint comp
= c
->aa_dest_stencil_reg
/ 2;
1386 GLuint off
= c
->aa_dest_stencil_reg
% 2;
1387 struct brw_reg aa
= offset(arg1
[comp
], off
);
1389 brw_push_insn_state(p
);
1390 brw_set_compression_control(p
, BRW_COMPRESSION_NONE
); /* ?? */
1391 brw_MOV(p
, brw_message_reg(reg
), aa
);
1392 brw_pop_insn_state(p
);
1396 /* Post-fragment-program processing. Send the results to the
1398 * \param arg0 the fragment color
1399 * \param arg1 the pass-through depth value
1400 * \param arg2 the shader-computed depth value
1402 void emit_fb_write(struct brw_wm_compile
*c
,
1403 struct brw_reg
*arg0
,
1404 struct brw_reg
*arg1
,
1405 struct brw_reg
*arg2
,
1409 struct brw_compile
*p
= &c
->func
;
1410 struct brw_context
*brw
= p
->brw
;
1411 struct intel_context
*intel
= &brw
->intel
;
1415 /* Reserve a space for AA - may not be needed:
1417 if (c
->aa_dest_stencil_reg
)
1420 /* I don't really understand how this achieves the color interleave
1421 * (ie RGBARGBA) in the result: [Do the saturation here]
1423 brw_push_insn_state(p
);
1425 for (channel
= 0; channel
< 4; channel
++) {
1426 if (intel
->gen
>= 6) {
1427 /* gen6 SIMD16 single source DP write looks like:
1437 if (c
->dispatch_width
== 16) {
1438 brw_MOV(p
, brw_message_reg(nr
+ channel
* 2), arg0
[channel
]);
1440 brw_MOV(p
, brw_message_reg(nr
+ channel
), arg0
[channel
]);
1442 } else if (c
->dispatch_width
== 16 && brw
->has_compr4
) {
1443 /* pre-gen6 SIMD16 single source DP write looks like:
1453 * By setting the high bit of the MRF register number, we indicate
1454 * that we want COMPR4 mode - instead of doing the usual destination
1455 * + 1 for the second half we get destination + 4.
1458 brw_message_reg(nr
+ channel
+ BRW_MRF_COMPR4
),
1461 /* mov (8) m2.0<1>:ud r28.0<8;8,1>:ud { Align1 } */
1462 /* mov (8) m6.0<1>:ud r29.0<8;8,1>:ud { Align1 SecHalf } */
1463 brw_set_compression_control(p
, BRW_COMPRESSION_NONE
);
1465 brw_message_reg(nr
+ channel
),
1468 if (c
->dispatch_width
== 16) {
1469 brw_set_compression_control(p
, BRW_COMPRESSION_2NDHALF
);
1471 brw_message_reg(nr
+ channel
+ 4),
1472 sechalf(arg0
[channel
]));
1476 /* skip over the regs populated above:
1478 if (c
->dispatch_width
== 16)
1483 brw_pop_insn_state(p
);
1485 if (c
->source_depth_to_render_target
)
1487 if (c
->computes_depth
)
1488 brw_MOV(p
, brw_message_reg(nr
), arg2
[2]);
1490 brw_MOV(p
, brw_message_reg(nr
), arg1
[1]); /* ? */
1495 if (c
->dest_depth_reg
)
1497 GLuint comp
= c
->dest_depth_reg
/ 2;
1498 GLuint off
= c
->dest_depth_reg
% 2;
1501 brw_push_insn_state(p
);
1502 brw_set_compression_control(p
, BRW_COMPRESSION_NONE
);
1504 brw_MOV(p
, brw_message_reg(nr
), offset(arg1
[comp
],1));
1506 brw_MOV(p
, brw_message_reg(nr
+1), arg1
[comp
+1]);
1507 brw_pop_insn_state(p
);
1510 brw_MOV(p
, brw_message_reg(nr
), arg1
[comp
]);
1515 if (intel
->gen
>= 6) {
1516 /* Load the message header. There's no implied move from src0
1517 * to the base mrf on gen6.
1519 brw_push_insn_state(p
);
1520 brw_set_mask_control(p
, BRW_MASK_DISABLE
);
1521 brw_MOV(p
, brw_message_reg(0), brw_vec8_grf(0, 0));
1522 brw_pop_insn_state(p
);
1525 brw_MOV(p
, retype(brw_vec1_reg(BRW_MESSAGE_REGISTER_FILE
,
1527 2), BRW_REGISTER_TYPE_UD
),
1528 brw_imm_ud(target
));
1532 if (!c
->runtime_check_aads_emit
) {
1533 if (c
->aa_dest_stencil_reg
)
1534 emit_aa(c
, arg1
, 2);
1536 fire_fb_write(c
, 0, nr
, target
, eot
);
1539 struct brw_reg v1_null_ud
= vec1(retype(brw_null_reg(), BRW_REGISTER_TYPE_UD
));
1540 struct brw_reg ip
= brw_ip_reg();
1541 struct brw_instruction
*jmp
;
1543 brw_set_compression_control(p
, BRW_COMPRESSION_NONE
);
1544 brw_set_conditionalmod(p
, BRW_CONDITIONAL_Z
);
1547 get_element_ud(brw_vec8_grf(1,0), 6),
1550 jmp
= brw_JMPI(p
, ip
, ip
, brw_imm_w(0));
1552 emit_aa(c
, arg1
, 2);
1553 fire_fb_write(c
, 0, nr
, target
, eot
);
1554 /* note - thread killed in subroutine */
1556 brw_land_fwd_jump(p
, jmp
);
1558 /* ELSE: Shuffle up one register to fill in the hole left for AA:
1560 fire_fb_write(c
, 1, nr
-1, target
, eot
);
1565 * Move a GPR to scratch memory.
1567 static void emit_spill( struct brw_wm_compile
*c
,
1571 struct brw_compile
*p
= &c
->func
;
1574 mov (16) m2.0<1>:ud r2.0<8;8,1>:ud { Align1 Compr }
1576 brw_MOV(p
, brw_message_reg(2), reg
);
1579 mov (1) r0.2<1>:d 0x00000080:d { Align1 NoMask }
1580 send (16) null.0<1>:uw m1 r0.0<8;8,1>:uw 0x053003ff:ud { Align1 }
1582 brw_oword_block_write_scratch(p
, brw_message_reg(1), 2, slot
);
1587 * Load a GPR from scratch memory.
1589 static void emit_unspill( struct brw_wm_compile
*c
,
1593 struct brw_compile
*p
= &c
->func
;
1595 /* Slot 0 is the undef value.
1598 brw_MOV(p
, reg
, brw_imm_f(0));
1603 mov (1) r0.2<1>:d 0x000000c0:d { Align1 NoMask }
1604 send (16) r110.0<1>:uw m1 r0.0<8;8,1>:uw 0x041243ff:ud { Align1 }
1607 brw_oword_block_read(p
, vec16(reg
), brw_message_reg(1), 2, slot
);
1612 * Retrieve up to 4 GEN4 register pairs for the given wm reg:
1613 * Args with unspill_reg != 0 will be loaded from scratch memory.
1615 static void get_argument_regs( struct brw_wm_compile
*c
,
1616 struct brw_wm_ref
*arg
[],
1617 struct brw_reg
*regs
)
1621 for (i
= 0; i
< 4; i
++) {
1623 if (arg
[i
]->unspill_reg
)
1625 brw_vec8_grf(arg
[i
]->unspill_reg
, 0),
1626 arg
[i
]->value
->spill_slot
);
1628 regs
[i
] = arg
[i
]->hw_reg
;
1631 regs
[i
] = brw_null_reg();
1638 * For values that have a spill_slot!=0, write those regs to scratch memory.
1640 static void spill_values( struct brw_wm_compile
*c
,
1641 struct brw_wm_value
*values
,
1646 for (i
= 0; i
< nr
; i
++)
1647 if (values
[i
].spill_slot
)
1648 emit_spill(c
, values
[i
].hw_reg
, values
[i
].spill_slot
);
1652 /* Emit the fragment program instructions here.
1654 void brw_wm_emit( struct brw_wm_compile
*c
)
1656 struct brw_compile
*p
= &c
->func
;
1657 struct intel_context
*intel
= &p
->brw
->intel
;
1660 brw_set_compression_control(p
, BRW_COMPRESSION_COMPRESSED
);
1661 if (intel
->gen
>= 6)
1662 brw_set_acc_write_control(p
, 1);
1664 /* Check if any of the payload regs need to be spilled:
1666 spill_values(c
, c
->payload
.depth
, 4);
1667 spill_values(c
, c
->creg
, c
->nr_creg
);
1668 spill_values(c
, c
->payload
.input_interp
, FRAG_ATTRIB_MAX
);
1671 for (insn
= 0; insn
< c
->nr_insns
; insn
++) {
1673 struct brw_wm_instruction
*inst
= &c
->instruction
[insn
];
1674 struct brw_reg args
[3][4], dst
[4];
1675 GLuint i
, dst_flags
;
1677 /* Get argument regs:
1679 for (i
= 0; i
< 3; i
++)
1680 get_argument_regs(c
, inst
->src
[i
], args
[i
]);
1684 for (i
= 0; i
< 4; i
++)
1686 dst
[i
] = inst
->dst
[i
]->hw_reg
;
1688 dst
[i
] = brw_null_reg();
1692 dst_flags
= inst
->writemask
;
1694 dst_flags
|= SATURATE
;
1696 switch (inst
->opcode
) {
1697 /* Generated instructions for calculating triangle interpolants:
1700 emit_pixel_xy(c
, dst
, dst_flags
);
1704 emit_delta_xy(p
, dst
, dst_flags
, args
[0]);
1708 emit_wpos_xy(c
, dst
, dst_flags
, args
[0]);
1712 emit_pixel_w(c
, dst
, dst_flags
, args
[0], args
[1]);
1716 emit_linterp(p
, dst
, dst_flags
, args
[0], args
[1]);
1720 emit_pinterp(p
, dst
, dst_flags
, args
[0], args
[1], args
[2]);
1724 emit_cinterp(p
, dst
, dst_flags
, args
[0]);
1728 emit_fb_write(c
, args
[0], args
[1], args
[2], inst
->target
, inst
->eot
);
1731 case WM_FRONTFACING
:
1732 emit_frontfacing(p
, dst
, dst_flags
);
1735 /* Straightforward arithmetic:
1738 emit_alu2(p
, brw_ADD
, dst
, dst_flags
, args
[0], args
[1]);
1742 emit_alu1(p
, brw_FRC
, dst
, dst_flags
, args
[0]);
1746 emit_alu1(p
, brw_RNDD
, dst
, dst_flags
, args
[0]);
1750 emit_ddxy(p
, dst
, dst_flags
, GL_TRUE
, args
[0]);
1754 emit_ddxy(p
, dst
, dst_flags
, GL_FALSE
, args
[0]);
1758 emit_dp2(p
, dst
, dst_flags
, args
[0], args
[1]);
1762 emit_dp3(p
, dst
, dst_flags
, args
[0], args
[1]);
1766 emit_dp4(p
, dst
, dst_flags
, args
[0], args
[1]);
1770 emit_dph(p
, dst
, dst_flags
, args
[0], args
[1]);
1774 for (i
= 0; i
< 4; i
++) {
1775 if (dst_flags
& (1<<i
)) {
1776 brw_RNDZ(p
, dst
[i
], args
[0][i
]);
1782 emit_lrp(p
, dst
, dst_flags
, args
[0], args
[1], args
[2]);
1786 emit_mad(p
, dst
, dst_flags
, args
[0], args
[1], args
[2]);
1791 emit_alu1(p
, brw_MOV
, dst
, dst_flags
, args
[0]);
1795 emit_alu2(p
, brw_MUL
, dst
, dst_flags
, args
[0], args
[1]);
1799 emit_xpd(p
, dst
, dst_flags
, args
[0], args
[1]);
1802 /* Higher math functions:
1805 emit_math1(c
, BRW_MATH_FUNCTION_INV
, dst
, dst_flags
, args
[0]);
1809 emit_math1(c
, BRW_MATH_FUNCTION_RSQ
, dst
, dst_flags
, args
[0]);
1813 emit_math1(c
, BRW_MATH_FUNCTION_SIN
, dst
, dst_flags
, args
[0]);
1817 emit_math1(c
, BRW_MATH_FUNCTION_COS
, dst
, dst_flags
, args
[0]);
1821 emit_math1(c
, BRW_MATH_FUNCTION_EXP
, dst
, dst_flags
, args
[0]);
1825 emit_math1(c
, BRW_MATH_FUNCTION_LOG
, dst
, dst_flags
, args
[0]);
1829 /* There is an scs math function, but it would need some
1830 * fixup for 16-element execution.
1832 if (dst_flags
& WRITEMASK_X
)
1833 emit_math1(c
, BRW_MATH_FUNCTION_COS
, dst
, (dst_flags
&SATURATE
)|WRITEMASK_X
, args
[0]);
1834 if (dst_flags
& WRITEMASK_Y
)
1835 emit_math1(c
, BRW_MATH_FUNCTION_SIN
, dst
+1, (dst_flags
&SATURATE
)|WRITEMASK_X
, args
[0]);
1839 emit_math2(c
, BRW_MATH_FUNCTION_POW
, dst
, dst_flags
, args
[0], args
[1]);
1845 emit_cmp(p
, dst
, dst_flags
, args
[0], args
[1], args
[2]);
1849 emit_max(p
, dst
, dst_flags
, args
[0], args
[1]);
1853 emit_min(p
, dst
, dst_flags
, args
[0], args
[1]);
1857 emit_slt(p
, dst
, dst_flags
, args
[0], args
[1]);
1861 emit_sle(p
, dst
, dst_flags
, args
[0], args
[1]);
1864 emit_sgt(p
, dst
, dst_flags
, args
[0], args
[1]);
1867 emit_sge(p
, dst
, dst_flags
, args
[0], args
[1]);
1870 emit_seq(p
, dst
, dst_flags
, args
[0], args
[1]);
1873 emit_sne(p
, dst
, dst_flags
, args
[0], args
[1]);
1877 emit_sign(p
, dst
, dst_flags
, args
[0]);
1881 emit_lit(c
, dst
, dst_flags
, args
[0]);
1884 /* Texturing operations:
1887 emit_tex(c
, dst
, dst_flags
, args
[0], c
->payload
.depth
[0].hw_reg
,
1888 inst
->tex_idx
, inst
->tex_unit
,
1893 emit_txb(c
, dst
, dst_flags
, args
[0], c
->payload
.depth
[0].hw_reg
,
1894 inst
->tex_idx
, inst
->tex_unit
);
1898 emit_kil(c
, args
[0]);
1902 printf("Unsupported opcode %i (%s) in fragment shader\n",
1903 inst
->opcode
, inst
->opcode
< MAX_OPCODE
?
1904 _mesa_opcode_string(inst
->opcode
) :
1908 for (i
= 0; i
< 4; i
++)
1909 if (inst
->dst
[i
] && inst
->dst
[i
]->spill_slot
)
1911 inst
->dst
[i
]->hw_reg
,
1912 inst
->dst
[i
]->spill_slot
);
1915 /* Only properly tested on ILK */
1916 if (p
->brw
->intel
.gen
== 5) {
1917 brw_remove_duplicate_mrf_moves(p
);
1918 if (c
->dispatch_width
== 16)
1919 brw_remove_grf_to_mrf_moves(p
);
1922 if (unlikely(INTEL_DEBUG
& DEBUG_WM
)) {
1925 printf("wm-native:\n");
1926 for (i
= 0; i
< p
->nr_insn
; i
++)
1927 brw_disasm(stdout
, &p
->store
[i
], p
->brw
->intel
.gen
);