llvmpipe/draw: handle UBOs that are < 16 bytes.
[mesa.git] / src / gallium / drivers / llvmpipe / lp_setup.c
1 /**************************************************************************
2 *
3 * Copyright 2007 VMware, Inc.
4 * All Rights Reserved.
5 *
6 * Permission is hereby granted, free of charge, to any person obtaining a
7 * copy of this software and associated documentation files (the
8 * "Software"), to deal in the Software without restriction, including
9 * without limitation the rights to use, copy, modify, merge, publish,
10 * distribute, sub license, and/or sell copies of the Software, and to
11 * permit persons to whom the Software is furnished to do so, subject to
12 * the following conditions:
13 *
14 * The above copyright notice and this permission notice (including the
15 * next paragraph) shall be included in all copies or substantial portions
16 * of the Software.
17 *
18 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
19 * OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
20 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT.
21 * IN NO EVENT SHALL VMWARE AND/OR ITS SUPPLIERS BE LIABLE FOR
22 * ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,
23 * TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
24 * SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
25 *
26 **************************************************************************/
27
28 /**
29 * Tiling engine.
30 *
31 * Builds per-tile display lists and executes them on calls to
32 * lp_setup_flush().
33 */
34
35 #include <limits.h>
36
37 #include "pipe/p_defines.h"
38 #include "util/u_framebuffer.h"
39 #include "util/u_inlines.h"
40 #include "util/u_memory.h"
41 #include "util/u_pack_color.h"
42 #include "util/u_viewport.h"
43 #include "draw/draw_pipe.h"
44 #include "util/os_time.h"
45 #include "lp_context.h"
46 #include "lp_memory.h"
47 #include "lp_scene.h"
48 #include "lp_texture.h"
49 #include "lp_debug.h"
50 #include "lp_fence.h"
51 #include "lp_query.h"
52 #include "lp_rast.h"
53 #include "lp_setup_context.h"
54 #include "lp_screen.h"
55 #include "lp_state.h"
56 #include "state_tracker/sw_winsys.h"
57
58 #include "draw/draw_context.h"
59 #include "draw/draw_vbuf.h"
60
61
62 static boolean set_scene_state( struct lp_setup_context *, enum setup_state,
63 const char *reason);
64 static boolean try_update_scene_state( struct lp_setup_context *setup );
65
66
67 static void
68 lp_setup_get_empty_scene(struct lp_setup_context *setup)
69 {
70 assert(setup->scene == NULL);
71
72 setup->scene_idx++;
73 setup->scene_idx %= ARRAY_SIZE(setup->scenes);
74
75 setup->scene = setup->scenes[setup->scene_idx];
76
77 if (setup->scene->fence) {
78 if (LP_DEBUG & DEBUG_SETUP)
79 debug_printf("%s: wait for scene %d\n",
80 __FUNCTION__, setup->scene->fence->id);
81
82 lp_fence_wait(setup->scene->fence);
83 }
84
85 lp_scene_begin_binning(setup->scene, &setup->fb);
86
87 }
88
89
90 static void
91 first_triangle( struct lp_setup_context *setup,
92 const float (*v0)[4],
93 const float (*v1)[4],
94 const float (*v2)[4])
95 {
96 assert(setup->state == SETUP_ACTIVE);
97 lp_setup_choose_triangle( setup );
98 setup->triangle( setup, v0, v1, v2 );
99 }
100
101 static void
102 first_line( struct lp_setup_context *setup,
103 const float (*v0)[4],
104 const float (*v1)[4])
105 {
106 assert(setup->state == SETUP_ACTIVE);
107 lp_setup_choose_line( setup );
108 setup->line( setup, v0, v1 );
109 }
110
111 static void
112 first_point( struct lp_setup_context *setup,
113 const float (*v0)[4])
114 {
115 assert(setup->state == SETUP_ACTIVE);
116 lp_setup_choose_point( setup );
117 setup->point( setup, v0 );
118 }
119
120 void lp_setup_reset( struct lp_setup_context *setup )
121 {
122 unsigned i;
123
124 LP_DBG(DEBUG_SETUP, "%s\n", __FUNCTION__);
125
126 /* Reset derived state */
127 for (i = 0; i < ARRAY_SIZE(setup->constants); ++i) {
128 setup->constants[i].stored_size = 0;
129 setup->constants[i].stored_data = NULL;
130 }
131 setup->fs.stored = NULL;
132 setup->dirty = ~0;
133
134 /* no current bin */
135 setup->scene = NULL;
136
137 /* Reset some state:
138 */
139 memset(&setup->clear, 0, sizeof setup->clear);
140
141 /* Have an explicit "start-binning" call and get rid of this
142 * pointer twiddling?
143 */
144 setup->line = first_line;
145 setup->point = first_point;
146 setup->triangle = first_triangle;
147 }
148
149
150 /** Rasterize all scene's bins */
151 static void
152 lp_setup_rasterize_scene( struct lp_setup_context *setup )
153 {
154 struct lp_scene *scene = setup->scene;
155 struct llvmpipe_screen *screen = llvmpipe_screen(scene->pipe->screen);
156
157 scene->num_active_queries = setup->active_binned_queries;
158 memcpy(scene->active_queries, setup->active_queries,
159 scene->num_active_queries * sizeof(scene->active_queries[0]));
160
161 lp_scene_end_binning(scene);
162
163 lp_fence_reference(&setup->last_fence, scene->fence);
164
165 if (setup->last_fence)
166 setup->last_fence->issued = TRUE;
167
168 mtx_lock(&screen->rast_mutex);
169
170 /* FIXME: We enqueue the scene then wait on the rasterizer to finish.
171 * This means we never actually run any vertex stuff in parallel to
172 * rasterization (not in the same context at least) which is what the
173 * multiple scenes per setup is about - when we get a new empty scene
174 * any old one is already empty again because we waited here for
175 * raster tasks to be finished. Ideally, we shouldn't need to wait here
176 * and rely on fences elsewhere when waiting is necessary.
177 * Certainly, lp_scene_end_rasterization() would need to be deferred too
178 * and there's probably other bits why this doesn't actually work.
179 */
180 lp_rast_queue_scene(screen->rast, scene);
181 lp_rast_finish(screen->rast);
182 mtx_unlock(&screen->rast_mutex);
183
184 lp_scene_end_rasterization(setup->scene);
185 lp_setup_reset( setup );
186
187 LP_DBG(DEBUG_SETUP, "%s done \n", __FUNCTION__);
188 }
189
190
191
192 static boolean
193 begin_binning( struct lp_setup_context *setup )
194 {
195 struct lp_scene *scene = setup->scene;
196 boolean need_zsload = FALSE;
197 boolean ok;
198
199 assert(scene);
200 assert(scene->fence == NULL);
201
202 /* Always create a fence:
203 */
204 scene->fence = lp_fence_create(MAX2(1, setup->num_threads));
205 if (!scene->fence)
206 return FALSE;
207
208 ok = try_update_scene_state(setup);
209 if (!ok)
210 return FALSE;
211
212 if (setup->fb.zsbuf &&
213 ((setup->clear.flags & PIPE_CLEAR_DEPTHSTENCIL) != PIPE_CLEAR_DEPTHSTENCIL) &&
214 util_format_is_depth_and_stencil(setup->fb.zsbuf->format))
215 need_zsload = TRUE;
216
217 LP_DBG(DEBUG_SETUP, "%s color clear bufs: %x depth: %s\n", __FUNCTION__,
218 setup->clear.flags >> 2,
219 need_zsload ? "clear": "load");
220
221 if (setup->clear.flags & PIPE_CLEAR_COLOR) {
222 unsigned cbuf;
223 for (cbuf = 0; cbuf < setup->fb.nr_cbufs; cbuf++) {
224 assert(PIPE_CLEAR_COLOR0 == 1 << 2);
225 if (setup->clear.flags & (1 << (2 + cbuf))) {
226 union lp_rast_cmd_arg clearrb_arg;
227 struct lp_rast_clear_rb *cc_scene =
228 (struct lp_rast_clear_rb *)
229 lp_scene_alloc(scene, sizeof(struct lp_rast_clear_rb));
230
231 if (!cc_scene) {
232 return FALSE;
233 }
234
235 cc_scene->cbuf = cbuf;
236 cc_scene->color_val = setup->clear.color_val[cbuf];
237 clearrb_arg.clear_rb = cc_scene;
238
239 if (!lp_scene_bin_everywhere(scene,
240 LP_RAST_OP_CLEAR_COLOR,
241 clearrb_arg))
242 return FALSE;
243 }
244 }
245 }
246
247 if (setup->fb.zsbuf) {
248 if (setup->clear.flags & PIPE_CLEAR_DEPTHSTENCIL) {
249 ok = lp_scene_bin_everywhere( scene,
250 LP_RAST_OP_CLEAR_ZSTENCIL,
251 lp_rast_arg_clearzs(
252 setup->clear.zsvalue,
253 setup->clear.zsmask));
254 if (!ok)
255 return FALSE;
256 }
257 }
258
259 setup->clear.flags = 0;
260 setup->clear.zsmask = 0;
261 setup->clear.zsvalue = 0;
262
263 scene->had_queries = !!setup->active_binned_queries;
264
265 LP_DBG(DEBUG_SETUP, "%s done\n", __FUNCTION__);
266 return TRUE;
267 }
268
269
270 /* This basically bins and then flushes any outstanding full-screen
271 * clears.
272 *
273 * TODO: fast path for fullscreen clears and no triangles.
274 */
275 static boolean
276 execute_clears( struct lp_setup_context *setup )
277 {
278 LP_DBG(DEBUG_SETUP, "%s\n", __FUNCTION__);
279
280 return begin_binning( setup );
281 }
282
283 const char *states[] = {
284 "FLUSHED",
285 "CLEARED",
286 "ACTIVE "
287 };
288
289
290 static boolean
291 set_scene_state( struct lp_setup_context *setup,
292 enum setup_state new_state,
293 const char *reason)
294 {
295 unsigned old_state = setup->state;
296
297 if (old_state == new_state)
298 return TRUE;
299
300 if (LP_DEBUG & DEBUG_SCENE) {
301 debug_printf("%s old %s new %s%s%s\n",
302 __FUNCTION__,
303 states[old_state],
304 states[new_state],
305 (new_state == SETUP_FLUSHED) ? ": " : "",
306 (new_state == SETUP_FLUSHED) ? reason : "");
307
308 if (new_state == SETUP_FLUSHED && setup->scene)
309 lp_debug_draw_bins_by_cmd_length(setup->scene);
310 }
311
312 /* wait for a free/empty scene
313 */
314 if (old_state == SETUP_FLUSHED)
315 lp_setup_get_empty_scene(setup);
316
317 switch (new_state) {
318 case SETUP_CLEARED:
319 break;
320
321 case SETUP_ACTIVE:
322 if (!begin_binning( setup ))
323 goto fail;
324 break;
325
326 case SETUP_FLUSHED:
327 if (old_state == SETUP_CLEARED)
328 if (!execute_clears( setup ))
329 goto fail;
330
331 lp_setup_rasterize_scene( setup );
332 assert(setup->scene == NULL);
333 break;
334
335 default:
336 assert(0 && "invalid setup state mode");
337 goto fail;
338 }
339
340 setup->state = new_state;
341 return TRUE;
342
343 fail:
344 if (setup->scene) {
345 lp_scene_end_rasterization(setup->scene);
346 setup->scene = NULL;
347 }
348
349 setup->state = SETUP_FLUSHED;
350 lp_setup_reset( setup );
351 return FALSE;
352 }
353
354
355 void
356 lp_setup_flush( struct lp_setup_context *setup,
357 struct pipe_fence_handle **fence,
358 const char *reason)
359 {
360 set_scene_state( setup, SETUP_FLUSHED, reason );
361
362 if (fence) {
363 lp_fence_reference((struct lp_fence **)fence, setup->last_fence);
364 if (!*fence)
365 *fence = (struct pipe_fence_handle *)lp_fence_create(0);
366 }
367 }
368
369
370 void
371 lp_setup_bind_framebuffer( struct lp_setup_context *setup,
372 const struct pipe_framebuffer_state *fb )
373 {
374 LP_DBG(DEBUG_SETUP, "%s\n", __FUNCTION__);
375
376 /* Flush any old scene.
377 */
378 set_scene_state( setup, SETUP_FLUSHED, __FUNCTION__ );
379
380 /*
381 * Ensure the old scene is not reused.
382 */
383 assert(!setup->scene);
384
385 /* Set new state. This will be picked up later when we next need a
386 * scene.
387 */
388 util_copy_framebuffer_state(&setup->fb, fb);
389 setup->framebuffer.x0 = 0;
390 setup->framebuffer.y0 = 0;
391 setup->framebuffer.x1 = fb->width-1;
392 setup->framebuffer.y1 = fb->height-1;
393 setup->dirty |= LP_SETUP_NEW_SCISSOR;
394 }
395
396
397 /*
398 * Try to clear one color buffer of the attached fb, either by binning a clear
399 * command or queuing up the clear for later (when binning is started).
400 */
401 static boolean
402 lp_setup_try_clear_color_buffer(struct lp_setup_context *setup,
403 const union pipe_color_union *color,
404 unsigned cbuf)
405 {
406 union lp_rast_cmd_arg clearrb_arg;
407 union util_color uc;
408 enum pipe_format format = setup->fb.cbufs[cbuf]->format;
409
410 LP_DBG(DEBUG_SETUP, "%s state %d\n", __FUNCTION__, setup->state);
411
412 if (util_format_is_pure_integer(format)) {
413 /*
414 * We expect int/uint clear values here, though some APIs
415 * might disagree (but in any case util_pack_color()
416 * couldn't handle it)...
417 */
418 if (util_format_is_pure_sint(format)) {
419 util_format_write_4i(format, color->i, 0, &uc, 0, 0, 0, 1, 1);
420 }
421 else {
422 assert(util_format_is_pure_uint(format));
423 util_format_write_4ui(format, color->ui, 0, &uc, 0, 0, 0, 1, 1);
424 }
425 }
426 else {
427 util_pack_color(color->f, format, &uc);
428 }
429
430 if (setup->state == SETUP_ACTIVE) {
431 struct lp_scene *scene = setup->scene;
432
433 /* Add the clear to existing scene. In the unusual case where
434 * both color and depth-stencil are being cleared when there's
435 * already been some rendering, we could discard the currently
436 * binned scene and start again, but I don't see that as being
437 * a common usage.
438 */
439 struct lp_rast_clear_rb *cc_scene =
440 (struct lp_rast_clear_rb *)
441 lp_scene_alloc_aligned(scene, sizeof(struct lp_rast_clear_rb), 8);
442
443 if (!cc_scene) {
444 return FALSE;
445 }
446
447 cc_scene->cbuf = cbuf;
448 cc_scene->color_val = uc;
449 clearrb_arg.clear_rb = cc_scene;
450
451 if (!lp_scene_bin_everywhere(scene,
452 LP_RAST_OP_CLEAR_COLOR,
453 clearrb_arg))
454 return FALSE;
455 }
456 else {
457 /* Put ourselves into the 'pre-clear' state, specifically to try
458 * and accumulate multiple clears to color and depth_stencil
459 * buffers which the app or state-tracker might issue
460 * separately.
461 */
462 set_scene_state( setup, SETUP_CLEARED, __FUNCTION__ );
463
464 assert(PIPE_CLEAR_COLOR0 == (1 << 2));
465 setup->clear.flags |= 1 << (cbuf + 2);
466 setup->clear.color_val[cbuf] = uc;
467 }
468
469 return TRUE;
470 }
471
472 static boolean
473 lp_setup_try_clear_zs(struct lp_setup_context *setup,
474 double depth,
475 unsigned stencil,
476 unsigned flags)
477 {
478 uint64_t zsmask = 0;
479 uint64_t zsvalue = 0;
480 uint32_t zmask32;
481 uint8_t smask8;
482 enum pipe_format format = setup->fb.zsbuf->format;
483
484 LP_DBG(DEBUG_SETUP, "%s state %d\n", __FUNCTION__, setup->state);
485
486 zmask32 = (flags & PIPE_CLEAR_DEPTH) ? ~0 : 0;
487 smask8 = (flags & PIPE_CLEAR_STENCIL) ? ~0 : 0;
488
489 zsvalue = util_pack64_z_stencil(format, depth, stencil);
490
491 zsmask = util_pack64_mask_z_stencil(format, zmask32, smask8);
492
493 zsvalue &= zsmask;
494
495 if (format == PIPE_FORMAT_Z24X8_UNORM ||
496 format == PIPE_FORMAT_X8Z24_UNORM) {
497 /*
498 * Make full mask if there's "X" bits so we can do full
499 * clear (without rmw).
500 */
501 uint32_t zsmask_full = 0;
502 zsmask_full = util_pack_mask_z_stencil(format, ~0, ~0);
503 zsmask |= ~zsmask_full;
504 }
505
506 if (setup->state == SETUP_ACTIVE) {
507 struct lp_scene *scene = setup->scene;
508
509 /* Add the clear to existing scene. In the unusual case where
510 * both color and depth-stencil are being cleared when there's
511 * already been some rendering, we could discard the currently
512 * binned scene and start again, but I don't see that as being
513 * a common usage.
514 */
515 if (!lp_scene_bin_everywhere(scene,
516 LP_RAST_OP_CLEAR_ZSTENCIL,
517 lp_rast_arg_clearzs(zsvalue, zsmask)))
518 return FALSE;
519 }
520 else {
521 /* Put ourselves into the 'pre-clear' state, specifically to try
522 * and accumulate multiple clears to color and depth_stencil
523 * buffers which the app or state-tracker might issue
524 * separately.
525 */
526 set_scene_state( setup, SETUP_CLEARED, __FUNCTION__ );
527
528 setup->clear.flags |= flags;
529
530 setup->clear.zsmask |= zsmask;
531 setup->clear.zsvalue =
532 (setup->clear.zsvalue & ~zsmask) | (zsvalue & zsmask);
533 }
534
535 return TRUE;
536 }
537
538 void
539 lp_setup_clear( struct lp_setup_context *setup,
540 const union pipe_color_union *color,
541 double depth,
542 unsigned stencil,
543 unsigned flags )
544 {
545 unsigned i;
546
547 /*
548 * Note any of these (max 9) clears could fail (but at most there should
549 * be just one failure!). This avoids doing the previous succeeded
550 * clears again (we still clear tiles twice if a clear command succeeded
551 * partially for one buffer).
552 */
553 if (flags & PIPE_CLEAR_DEPTHSTENCIL) {
554 unsigned flagszs = flags & PIPE_CLEAR_DEPTHSTENCIL;
555 if (!lp_setup_try_clear_zs(setup, depth, stencil, flagszs)) {
556 lp_setup_flush(setup, NULL, __FUNCTION__);
557
558 if (!lp_setup_try_clear_zs(setup, depth, stencil, flagszs))
559 assert(0);
560 }
561 }
562
563 if (flags & PIPE_CLEAR_COLOR) {
564 assert(PIPE_CLEAR_COLOR0 == (1 << 2));
565 for (i = 0; i < setup->fb.nr_cbufs; i++) {
566 if ((flags & (1 << (2 + i))) && setup->fb.cbufs[i]) {
567 if (!lp_setup_try_clear_color_buffer(setup, color, i)) {
568 lp_setup_flush(setup, NULL, __FUNCTION__);
569
570 if (!lp_setup_try_clear_color_buffer(setup, color, i))
571 assert(0);
572 }
573 }
574 }
575 }
576 }
577
578
579
580 void
581 lp_setup_set_triangle_state( struct lp_setup_context *setup,
582 unsigned cull_mode,
583 boolean ccw_is_frontface,
584 boolean scissor,
585 boolean half_pixel_center,
586 boolean bottom_edge_rule)
587 {
588 LP_DBG(DEBUG_SETUP, "%s\n", __FUNCTION__);
589
590 setup->ccw_is_frontface = ccw_is_frontface;
591 setup->cullmode = cull_mode;
592 setup->triangle = first_triangle;
593 setup->pixel_offset = half_pixel_center ? 0.5f : 0.0f;
594 setup->bottom_edge_rule = bottom_edge_rule;
595
596 if (setup->scissor_test != scissor) {
597 setup->dirty |= LP_SETUP_NEW_SCISSOR;
598 setup->scissor_test = scissor;
599 }
600 }
601
602 void
603 lp_setup_set_line_state( struct lp_setup_context *setup,
604 float line_width)
605 {
606 LP_DBG(DEBUG_SETUP, "%s\n", __FUNCTION__);
607
608 setup->line_width = line_width;
609 }
610
611 void
612 lp_setup_set_point_state( struct lp_setup_context *setup,
613 float point_size,
614 boolean point_size_per_vertex,
615 uint sprite_coord_enable,
616 uint sprite_coord_origin)
617 {
618 LP_DBG(DEBUG_SETUP, "%s\n", __FUNCTION__);
619
620 setup->point_size = point_size;
621 setup->sprite_coord_enable = sprite_coord_enable;
622 setup->sprite_coord_origin = sprite_coord_origin;
623 setup->point_size_per_vertex = point_size_per_vertex;
624 }
625
626 void
627 lp_setup_set_setup_variant( struct lp_setup_context *setup,
628 const struct lp_setup_variant *variant)
629 {
630 LP_DBG(DEBUG_SETUP, "%s\n", __FUNCTION__);
631
632 setup->setup.variant = variant;
633 }
634
635 void
636 lp_setup_set_fs_variant( struct lp_setup_context *setup,
637 struct lp_fragment_shader_variant *variant)
638 {
639 LP_DBG(DEBUG_SETUP, "%s %p\n", __FUNCTION__,
640 variant);
641 /* FIXME: reference count */
642
643 setup->fs.current.variant = variant;
644 setup->dirty |= LP_SETUP_NEW_FS;
645 }
646
647 void
648 lp_setup_set_fs_constants(struct lp_setup_context *setup,
649 unsigned num,
650 struct pipe_constant_buffer *buffers)
651 {
652 unsigned i;
653
654 LP_DBG(DEBUG_SETUP, "%s %p\n", __FUNCTION__, (void *) buffers);
655
656 assert(num <= ARRAY_SIZE(setup->constants));
657
658 for (i = 0; i < num; ++i) {
659 util_copy_constant_buffer(&setup->constants[i].current, &buffers[i]);
660 }
661 for (; i < ARRAY_SIZE(setup->constants); i++) {
662 util_copy_constant_buffer(&setup->constants[i].current, NULL);
663 }
664 setup->dirty |= LP_SETUP_NEW_CONSTANTS;
665 }
666
667 void
668 lp_setup_set_fs_ssbos(struct lp_setup_context *setup,
669 unsigned num,
670 struct pipe_shader_buffer *buffers)
671 {
672 unsigned i;
673
674 LP_DBG(DEBUG_SETUP, "%s %p\n", __FUNCTION__, (void *) buffers);
675
676 assert(num <= ARRAY_SIZE(setup->ssbos));
677
678 for (i = 0; i < num; ++i) {
679 util_copy_shader_buffer(&setup->ssbos[i].current, &buffers[i]);
680 }
681 for (; i < ARRAY_SIZE(setup->ssbos); i++) {
682 util_copy_shader_buffer(&setup->ssbos[i].current, NULL);
683 }
684 setup->dirty |= LP_SETUP_NEW_SSBOS;
685 }
686
687 void
688 lp_setup_set_fs_images(struct lp_setup_context *setup,
689 unsigned num,
690 struct pipe_image_view *images)
691 {
692 unsigned i;
693
694 LP_DBG(DEBUG_SETUP, "%s %p\n", __FUNCTION__, (void *) images);
695
696 assert(num <= ARRAY_SIZE(setup->images));
697
698 for (i = 0; i < num; ++i) {
699 struct pipe_image_view *image = &images[i];
700 util_copy_image_view(&setup->images[i].current, &images[i]);
701
702 struct pipe_resource *res = image->resource;
703 struct llvmpipe_resource *lp_res = llvmpipe_resource(res);
704 struct lp_jit_image *jit_image;
705
706 jit_image = &setup->fs.current.jit_context.images[i];
707 if (!lp_res)
708 continue;
709 if (!lp_res->dt) {
710 /* regular texture - setup array of mipmap level offsets */
711 if (llvmpipe_resource_is_texture(res)) {
712 jit_image->base = lp_res->tex_data;
713 } else
714 jit_image->base = lp_res->data;
715
716 jit_image->width = res->width0;
717 jit_image->height = res->height0;
718 jit_image->depth = res->depth0;
719
720 if (llvmpipe_resource_is_texture(res)) {
721 uint32_t mip_offset = lp_res->mip_offsets[image->u.tex.level];
722
723 jit_image->width = u_minify(jit_image->width, image->u.tex.level);
724 jit_image->height = u_minify(jit_image->height, image->u.tex.level);
725
726 if (res->target == PIPE_TEXTURE_1D_ARRAY ||
727 res->target == PIPE_TEXTURE_2D_ARRAY ||
728 res->target == PIPE_TEXTURE_3D ||
729 res->target == PIPE_TEXTURE_CUBE ||
730 res->target == PIPE_TEXTURE_CUBE_ARRAY) {
731 /*
732 * For array textures, we don't have first_layer, instead
733 * adjust last_layer (stored as depth) plus the mip level offsets
734 * (as we have mip-first layout can't just adjust base ptr).
735 * XXX For mip levels, could do something similar.
736 */
737 jit_image->depth = image->u.tex.last_layer - image->u.tex.first_layer + 1;
738 mip_offset += image->u.tex.first_layer * lp_res->img_stride[image->u.tex.level];
739 } else
740 jit_image->depth = u_minify(jit_image->depth, image->u.tex.level);
741
742 jit_image->row_stride = lp_res->row_stride[image->u.tex.level];
743 jit_image->img_stride = lp_res->img_stride[image->u.tex.level];
744 jit_image->base = (uint8_t *)jit_image->base + mip_offset;
745 }
746 else {
747 unsigned view_blocksize = util_format_get_blocksize(image->format);
748 jit_image->width = image->u.buf.size / view_blocksize;
749 jit_image->base = (uint8_t *)jit_image->base + image->u.buf.offset;
750 }
751 }
752 }
753 for (; i < ARRAY_SIZE(setup->images); i++) {
754 util_copy_image_view(&setup->images[i].current, NULL);
755 }
756 setup->dirty |= LP_SETUP_NEW_IMAGES;
757 }
758
759 void
760 lp_setup_set_alpha_ref_value( struct lp_setup_context *setup,
761 float alpha_ref_value )
762 {
763 LP_DBG(DEBUG_SETUP, "%s %f\n", __FUNCTION__, alpha_ref_value);
764
765 if(setup->fs.current.jit_context.alpha_ref_value != alpha_ref_value) {
766 setup->fs.current.jit_context.alpha_ref_value = alpha_ref_value;
767 setup->dirty |= LP_SETUP_NEW_FS;
768 }
769 }
770
771 void
772 lp_setup_set_stencil_ref_values( struct lp_setup_context *setup,
773 const ubyte refs[2] )
774 {
775 LP_DBG(DEBUG_SETUP, "%s %d %d\n", __FUNCTION__, refs[0], refs[1]);
776
777 if (setup->fs.current.jit_context.stencil_ref_front != refs[0] ||
778 setup->fs.current.jit_context.stencil_ref_back != refs[1]) {
779 setup->fs.current.jit_context.stencil_ref_front = refs[0];
780 setup->fs.current.jit_context.stencil_ref_back = refs[1];
781 setup->dirty |= LP_SETUP_NEW_FS;
782 }
783 }
784
785 void
786 lp_setup_set_blend_color( struct lp_setup_context *setup,
787 const struct pipe_blend_color *blend_color )
788 {
789 LP_DBG(DEBUG_SETUP, "%s\n", __FUNCTION__);
790
791 assert(blend_color);
792
793 if(memcmp(&setup->blend_color.current, blend_color, sizeof *blend_color) != 0) {
794 memcpy(&setup->blend_color.current, blend_color, sizeof *blend_color);
795 setup->dirty |= LP_SETUP_NEW_BLEND_COLOR;
796 }
797 }
798
799
800 void
801 lp_setup_set_scissors( struct lp_setup_context *setup,
802 const struct pipe_scissor_state *scissors )
803 {
804 unsigned i;
805 LP_DBG(DEBUG_SETUP, "%s\n", __FUNCTION__);
806
807 assert(scissors);
808
809 for (i = 0; i < PIPE_MAX_VIEWPORTS; ++i) {
810 setup->scissors[i].x0 = scissors[i].minx;
811 setup->scissors[i].x1 = scissors[i].maxx-1;
812 setup->scissors[i].y0 = scissors[i].miny;
813 setup->scissors[i].y1 = scissors[i].maxy-1;
814 }
815 setup->dirty |= LP_SETUP_NEW_SCISSOR;
816 }
817
818
819 void
820 lp_setup_set_flatshade_first(struct lp_setup_context *setup,
821 boolean flatshade_first)
822 {
823 setup->flatshade_first = flatshade_first;
824 }
825
826 void
827 lp_setup_set_rasterizer_discard(struct lp_setup_context *setup,
828 boolean rasterizer_discard)
829 {
830 if (setup->rasterizer_discard != rasterizer_discard) {
831 setup->rasterizer_discard = rasterizer_discard;
832 setup->line = first_line;
833 setup->point = first_point;
834 setup->triangle = first_triangle;
835 }
836 }
837
838 void
839 lp_setup_set_vertex_info(struct lp_setup_context *setup,
840 struct vertex_info *vertex_info)
841 {
842 /* XXX: just silently holding onto the pointer:
843 */
844 setup->vertex_info = vertex_info;
845 }
846
847
848 /**
849 * Called during state validation when LP_NEW_VIEWPORT is set.
850 */
851 void
852 lp_setup_set_viewports(struct lp_setup_context *setup,
853 unsigned num_viewports,
854 const struct pipe_viewport_state *viewports)
855 {
856 struct llvmpipe_context *lp = llvmpipe_context(setup->pipe);
857 unsigned i;
858
859 LP_DBG(DEBUG_SETUP, "%s\n", __FUNCTION__);
860
861 assert(num_viewports <= PIPE_MAX_VIEWPORTS);
862 assert(viewports);
863
864 /*
865 * For use in lp_state_fs.c, propagate the viewport values for all viewports.
866 */
867 for (i = 0; i < num_viewports; i++) {
868 float min_depth;
869 float max_depth;
870 util_viewport_zmin_zmax(&viewports[i], lp->rasterizer->clip_halfz,
871 &min_depth, &max_depth);
872
873 if (setup->viewports[i].min_depth != min_depth ||
874 setup->viewports[i].max_depth != max_depth) {
875 setup->viewports[i].min_depth = min_depth;
876 setup->viewports[i].max_depth = max_depth;
877 setup->dirty |= LP_SETUP_NEW_VIEWPORTS;
878 }
879 }
880 }
881
882
883 /**
884 * Called during state validation when LP_NEW_SAMPLER_VIEW is set.
885 */
886 void
887 lp_setup_set_fragment_sampler_views(struct lp_setup_context *setup,
888 unsigned num,
889 struct pipe_sampler_view **views)
890 {
891 unsigned i, max_tex_num;
892
893 LP_DBG(DEBUG_SETUP, "%s\n", __FUNCTION__);
894
895 assert(num <= PIPE_MAX_SHADER_SAMPLER_VIEWS);
896
897 max_tex_num = MAX2(num, setup->fs.current_tex_num);
898
899 for (i = 0; i < max_tex_num; i++) {
900 struct pipe_sampler_view *view = i < num ? views[i] : NULL;
901
902 if (view) {
903 struct pipe_resource *res = view->texture;
904 struct llvmpipe_resource *lp_tex = llvmpipe_resource(res);
905 struct lp_jit_texture *jit_tex;
906 jit_tex = &setup->fs.current.jit_context.textures[i];
907
908 /* We're referencing the texture's internal data, so save a
909 * reference to it.
910 */
911 pipe_resource_reference(&setup->fs.current_tex[i], res);
912
913 if (!lp_tex->dt) {
914 /* regular texture - setup array of mipmap level offsets */
915 int j;
916 unsigned first_level = 0;
917 unsigned last_level = 0;
918
919 if (llvmpipe_resource_is_texture(res)) {
920 first_level = view->u.tex.first_level;
921 last_level = view->u.tex.last_level;
922 assert(first_level <= last_level);
923 assert(last_level <= res->last_level);
924 jit_tex->base = lp_tex->tex_data;
925 }
926 else {
927 jit_tex->base = lp_tex->data;
928 }
929
930 if (LP_PERF & PERF_TEX_MEM) {
931 /* use dummy tile memory */
932 jit_tex->base = lp_dummy_tile;
933 jit_tex->width = TILE_SIZE/8;
934 jit_tex->height = TILE_SIZE/8;
935 jit_tex->depth = 1;
936 jit_tex->first_level = 0;
937 jit_tex->last_level = 0;
938 jit_tex->mip_offsets[0] = 0;
939 jit_tex->row_stride[0] = 0;
940 jit_tex->img_stride[0] = 0;
941 }
942 else {
943 jit_tex->width = res->width0;
944 jit_tex->height = res->height0;
945 jit_tex->depth = res->depth0;
946 jit_tex->first_level = first_level;
947 jit_tex->last_level = last_level;
948
949 if (llvmpipe_resource_is_texture(res)) {
950 for (j = first_level; j <= last_level; j++) {
951 jit_tex->mip_offsets[j] = lp_tex->mip_offsets[j];
952 jit_tex->row_stride[j] = lp_tex->row_stride[j];
953 jit_tex->img_stride[j] = lp_tex->img_stride[j];
954 }
955
956 if (res->target == PIPE_TEXTURE_1D_ARRAY ||
957 res->target == PIPE_TEXTURE_2D_ARRAY ||
958 res->target == PIPE_TEXTURE_CUBE ||
959 res->target == PIPE_TEXTURE_CUBE_ARRAY) {
960 /*
961 * For array textures, we don't have first_layer, instead
962 * adjust last_layer (stored as depth) plus the mip level offsets
963 * (as we have mip-first layout can't just adjust base ptr).
964 * XXX For mip levels, could do something similar.
965 */
966 jit_tex->depth = view->u.tex.last_layer - view->u.tex.first_layer + 1;
967 for (j = first_level; j <= last_level; j++) {
968 jit_tex->mip_offsets[j] += view->u.tex.first_layer *
969 lp_tex->img_stride[j];
970 }
971 if (view->target == PIPE_TEXTURE_CUBE ||
972 view->target == PIPE_TEXTURE_CUBE_ARRAY) {
973 assert(jit_tex->depth % 6 == 0);
974 }
975 assert(view->u.tex.first_layer <= view->u.tex.last_layer);
976 assert(view->u.tex.last_layer < res->array_size);
977 }
978 }
979 else {
980 /*
981 * For buffers, we don't have "offset", instead adjust
982 * the size (stored as width) plus the base pointer.
983 */
984 unsigned view_blocksize = util_format_get_blocksize(view->format);
985 /* probably don't really need to fill that out */
986 jit_tex->mip_offsets[0] = 0;
987 jit_tex->row_stride[0] = 0;
988 jit_tex->img_stride[0] = 0;
989
990 /* everything specified in number of elements here. */
991 jit_tex->width = view->u.buf.size / view_blocksize;
992 jit_tex->base = (uint8_t *)jit_tex->base + view->u.buf.offset;
993 /* XXX Unsure if we need to sanitize parameters? */
994 assert(view->u.buf.offset + view->u.buf.size <= res->width0);
995 }
996 }
997 }
998 else {
999 /* display target texture/surface */
1000 /*
1001 * XXX: Where should this be unmapped?
1002 */
1003 struct llvmpipe_screen *screen = llvmpipe_screen(res->screen);
1004 struct sw_winsys *winsys = screen->winsys;
1005 jit_tex->base = winsys->displaytarget_map(winsys, lp_tex->dt,
1006 PIPE_TRANSFER_READ);
1007 jit_tex->row_stride[0] = lp_tex->row_stride[0];
1008 jit_tex->img_stride[0] = lp_tex->img_stride[0];
1009 jit_tex->mip_offsets[0] = 0;
1010 jit_tex->width = res->width0;
1011 jit_tex->height = res->height0;
1012 jit_tex->depth = res->depth0;
1013 jit_tex->first_level = jit_tex->last_level = 0;
1014 assert(jit_tex->base);
1015 }
1016 }
1017 else {
1018 pipe_resource_reference(&setup->fs.current_tex[i], NULL);
1019 }
1020 }
1021 setup->fs.current_tex_num = num;
1022
1023 setup->dirty |= LP_SETUP_NEW_FS;
1024 }
1025
1026
1027 /**
1028 * Called during state validation when LP_NEW_SAMPLER is set.
1029 */
1030 void
1031 lp_setup_set_fragment_sampler_state(struct lp_setup_context *setup,
1032 unsigned num,
1033 struct pipe_sampler_state **samplers)
1034 {
1035 unsigned i;
1036
1037 LP_DBG(DEBUG_SETUP, "%s\n", __FUNCTION__);
1038
1039 assert(num <= PIPE_MAX_SAMPLERS);
1040
1041 for (i = 0; i < PIPE_MAX_SAMPLERS; i++) {
1042 const struct pipe_sampler_state *sampler = i < num ? samplers[i] : NULL;
1043
1044 if (sampler) {
1045 struct lp_jit_sampler *jit_sam;
1046 jit_sam = &setup->fs.current.jit_context.samplers[i];
1047
1048 jit_sam->min_lod = sampler->min_lod;
1049 jit_sam->max_lod = sampler->max_lod;
1050 jit_sam->lod_bias = sampler->lod_bias;
1051 COPY_4V(jit_sam->border_color, sampler->border_color.f);
1052 }
1053 }
1054
1055 setup->dirty |= LP_SETUP_NEW_FS;
1056 }
1057
1058
1059 /**
1060 * Is the given texture referenced by any scene?
1061 * Note: we have to check all scenes including any scenes currently
1062 * being rendered and the current scene being built.
1063 */
1064 unsigned
1065 lp_setup_is_resource_referenced( const struct lp_setup_context *setup,
1066 const struct pipe_resource *texture )
1067 {
1068 unsigned i;
1069
1070 /* check the render targets */
1071 for (i = 0; i < setup->fb.nr_cbufs; i++) {
1072 if (setup->fb.cbufs[i] && setup->fb.cbufs[i]->texture == texture)
1073 return LP_REFERENCED_FOR_READ | LP_REFERENCED_FOR_WRITE;
1074 }
1075 if (setup->fb.zsbuf && setup->fb.zsbuf->texture == texture) {
1076 return LP_REFERENCED_FOR_READ | LP_REFERENCED_FOR_WRITE;
1077 }
1078
1079 /* check textures referenced by the scene */
1080 for (i = 0; i < ARRAY_SIZE(setup->scenes); i++) {
1081 if (lp_scene_is_resource_referenced(setup->scenes[i], texture)) {
1082 return LP_REFERENCED_FOR_READ;
1083 }
1084 }
1085
1086 for (i = 0; i < ARRAY_SIZE(setup->ssbos); i++) {
1087 if (setup->ssbos[i].current.buffer == texture)
1088 return LP_REFERENCED_FOR_READ | LP_REFERENCED_FOR_WRITE;
1089 }
1090
1091 for (i = 0; i < ARRAY_SIZE(setup->images); i++) {
1092 if (setup->images[i].current.resource == texture)
1093 return LP_REFERENCED_FOR_READ | LP_REFERENCED_FOR_WRITE;
1094 }
1095
1096 return LP_UNREFERENCED;
1097 }
1098
1099
1100 /**
1101 * Called by vbuf code when we're about to draw something.
1102 *
1103 * This function stores all dirty state in the current scene's display list
1104 * memory, via lp_scene_alloc(). We can not pass pointers of mutable state to
1105 * the JIT functions, as the JIT functions will be called later on, most likely
1106 * on a different thread.
1107 *
1108 * When processing dirty state it is imperative that we don't refer to any
1109 * pointers previously allocated with lp_scene_alloc() in this function (or any
1110 * function) as they may belong to a scene freed since then.
1111 */
1112 static boolean
1113 try_update_scene_state( struct lp_setup_context *setup )
1114 {
1115 static const float fake_const_buf[4];
1116 boolean new_scene = (setup->fs.stored == NULL);
1117 struct lp_scene *scene = setup->scene;
1118 unsigned i;
1119
1120 assert(scene);
1121
1122 if (setup->dirty & LP_SETUP_NEW_VIEWPORTS) {
1123 /*
1124 * Record new depth range state for changes due to viewport updates.
1125 *
1126 * TODO: Collapse the existing viewport and depth range information
1127 * into one structure, for access by JIT.
1128 */
1129 struct lp_jit_viewport *stored;
1130
1131 stored = (struct lp_jit_viewport *)
1132 lp_scene_alloc(scene, sizeof setup->viewports);
1133
1134 if (!stored) {
1135 assert(!new_scene);
1136 return FALSE;
1137 }
1138
1139 memcpy(stored, setup->viewports, sizeof setup->viewports);
1140
1141 setup->fs.current.jit_context.viewports = stored;
1142 setup->dirty |= LP_SETUP_NEW_FS;
1143 }
1144
1145 if(setup->dirty & LP_SETUP_NEW_BLEND_COLOR) {
1146 uint8_t *stored;
1147 float* fstored;
1148 unsigned i, j;
1149 unsigned size;
1150
1151 /* Alloc u8_blend_color (16 x i8) and f_blend_color (4 or 8 x f32) */
1152 size = 4 * 16 * sizeof(uint8_t);
1153 size += (LP_MAX_VECTOR_LENGTH / 4) * sizeof(float);
1154 stored = lp_scene_alloc_aligned(scene, size, LP_MIN_VECTOR_ALIGN);
1155
1156 if (!stored) {
1157 assert(!new_scene);
1158 return FALSE;
1159 }
1160
1161 /* Store floating point colour */
1162 fstored = (float*)(stored + 4*16);
1163 for (i = 0; i < (LP_MAX_VECTOR_LENGTH / 4); ++i) {
1164 fstored[i] = setup->blend_color.current.color[i % 4];
1165 }
1166
1167 /* smear each blend color component across 16 ubyte elements */
1168 for (i = 0; i < 4; ++i) {
1169 uint8_t c = float_to_ubyte(setup->blend_color.current.color[i]);
1170 for (j = 0; j < 16; ++j)
1171 stored[i*16 + j] = c;
1172 }
1173
1174 setup->blend_color.stored = stored;
1175 setup->fs.current.jit_context.u8_blend_color = stored;
1176 setup->fs.current.jit_context.f_blend_color = fstored;
1177 setup->dirty |= LP_SETUP_NEW_FS;
1178 }
1179
1180 if (setup->dirty & LP_SETUP_NEW_CONSTANTS) {
1181 for (i = 0; i < ARRAY_SIZE(setup->constants); ++i) {
1182 struct pipe_resource *buffer = setup->constants[i].current.buffer;
1183 const unsigned current_size = MIN2(setup->constants[i].current.buffer_size,
1184 LP_MAX_TGSI_CONST_BUFFER_SIZE);
1185 const ubyte *current_data = NULL;
1186 int num_constants;
1187
1188 STATIC_ASSERT(DATA_BLOCK_SIZE >= LP_MAX_TGSI_CONST_BUFFER_SIZE);
1189
1190 if (buffer) {
1191 /* resource buffer */
1192 current_data = (ubyte *) llvmpipe_resource_data(buffer);
1193 }
1194 else if (setup->constants[i].current.user_buffer) {
1195 /* user-space buffer */
1196 current_data = (ubyte *) setup->constants[i].current.user_buffer;
1197 }
1198
1199 if (current_data) {
1200 current_data += setup->constants[i].current.buffer_offset;
1201
1202 /* TODO: copy only the actually used constants? */
1203
1204 if (setup->constants[i].stored_size != current_size ||
1205 !setup->constants[i].stored_data ||
1206 memcmp(setup->constants[i].stored_data,
1207 current_data,
1208 current_size) != 0) {
1209 void *stored;
1210
1211 stored = lp_scene_alloc(scene, current_size);
1212 if (!stored) {
1213 assert(!new_scene);
1214 return FALSE;
1215 }
1216
1217 memcpy(stored,
1218 current_data,
1219 current_size);
1220 setup->constants[i].stored_size = current_size;
1221 setup->constants[i].stored_data = stored;
1222 }
1223 setup->fs.current.jit_context.constants[i] =
1224 setup->constants[i].stored_data;
1225 }
1226 else {
1227 setup->constants[i].stored_size = 0;
1228 setup->constants[i].stored_data = NULL;
1229 setup->fs.current.jit_context.constants[i] = fake_const_buf;
1230 }
1231
1232 num_constants =
1233 DIV_ROUND_UP(setup->constants[i].stored_size, (sizeof(float) * 4));
1234 setup->fs.current.jit_context.num_constants[i] = num_constants;
1235 setup->dirty |= LP_SETUP_NEW_FS;
1236 }
1237 }
1238
1239 if (setup->dirty & LP_SETUP_NEW_SSBOS) {
1240 for (i = 0; i < ARRAY_SIZE(setup->ssbos); ++i) {
1241 struct pipe_resource *buffer = setup->ssbos[i].current.buffer;
1242 const ubyte *current_data = NULL;
1243
1244 if (!buffer)
1245 continue;
1246 /* resource buffer */
1247 current_data = (ubyte *) llvmpipe_resource_data(buffer);
1248 if (current_data) {
1249 current_data += setup->ssbos[i].current.buffer_offset;
1250
1251 setup->fs.current.jit_context.ssbos[i] = (const uint32_t *)current_data;
1252 setup->fs.current.jit_context.num_ssbos[i] = setup->ssbos[i].current.buffer_size;
1253 } else {
1254 setup->fs.current.jit_context.ssbos[i] = NULL;
1255 setup->fs.current.jit_context.num_ssbos[i] = 0;
1256 }
1257 setup->dirty |= LP_SETUP_NEW_FS;
1258 }
1259 }
1260 if (setup->dirty & LP_SETUP_NEW_FS) {
1261 if (!setup->fs.stored ||
1262 memcmp(setup->fs.stored,
1263 &setup->fs.current,
1264 sizeof setup->fs.current) != 0)
1265 {
1266 struct lp_rast_state *stored;
1267
1268 /* The fs state that's been stored in the scene is different from
1269 * the new, current state. So allocate a new lp_rast_state object
1270 * and append it to the bin's setup data buffer.
1271 */
1272 stored = (struct lp_rast_state *) lp_scene_alloc(scene, sizeof *stored);
1273 if (!stored) {
1274 assert(!new_scene);
1275 return FALSE;
1276 }
1277
1278 memcpy(stored,
1279 &setup->fs.current,
1280 sizeof setup->fs.current);
1281 setup->fs.stored = stored;
1282
1283 /* The scene now references the textures in the rasterization
1284 * state record. Note that now.
1285 */
1286 for (i = 0; i < ARRAY_SIZE(setup->fs.current_tex); i++) {
1287 if (setup->fs.current_tex[i]) {
1288 if (!lp_scene_add_resource_reference(scene,
1289 setup->fs.current_tex[i],
1290 new_scene)) {
1291 assert(!new_scene);
1292 return FALSE;
1293 }
1294 }
1295 }
1296 }
1297 }
1298
1299 if (setup->dirty & LP_SETUP_NEW_SCISSOR) {
1300 unsigned i;
1301 for (i = 0; i < PIPE_MAX_VIEWPORTS; ++i) {
1302 setup->draw_regions[i] = setup->framebuffer;
1303 if (setup->scissor_test) {
1304 u_rect_possible_intersection(&setup->scissors[i],
1305 &setup->draw_regions[i]);
1306 }
1307 }
1308 }
1309
1310 setup->dirty = 0;
1311
1312 assert(setup->fs.stored);
1313 return TRUE;
1314 }
1315
1316 boolean
1317 lp_setup_update_state( struct lp_setup_context *setup,
1318 boolean update_scene )
1319 {
1320 /* Some of the 'draw' pipeline stages may have changed some driver state.
1321 * Make sure we've processed those state changes before anything else.
1322 *
1323 * XXX this is the only place where llvmpipe_context is used in the
1324 * setup code. This may get refactored/changed...
1325 */
1326 {
1327 struct llvmpipe_context *lp = llvmpipe_context(setup->pipe);
1328 if (lp->dirty) {
1329 llvmpipe_update_derived(lp);
1330 }
1331
1332 if (lp->setup->dirty) {
1333 llvmpipe_update_setup(lp);
1334 }
1335
1336 assert(setup->setup.variant);
1337
1338 /* Will probably need to move this somewhere else, just need
1339 * to know about vertex shader point size attribute.
1340 */
1341 setup->psize_slot = lp->psize_slot;
1342 setup->viewport_index_slot = lp->viewport_index_slot;
1343 setup->layer_slot = lp->layer_slot;
1344 setup->face_slot = lp->face_slot;
1345
1346 assert(lp->dirty == 0);
1347
1348 assert(lp->setup_variant.key.size ==
1349 setup->setup.variant->key.size);
1350
1351 assert(memcmp(&lp->setup_variant.key,
1352 &setup->setup.variant->key,
1353 setup->setup.variant->key.size) == 0);
1354 }
1355
1356 if (update_scene && setup->state != SETUP_ACTIVE) {
1357 if (!set_scene_state( setup, SETUP_ACTIVE, __FUNCTION__ ))
1358 return FALSE;
1359 }
1360
1361 /* Only call into update_scene_state() if we already have a
1362 * scene:
1363 */
1364 if (update_scene && setup->scene) {
1365 assert(setup->state == SETUP_ACTIVE);
1366
1367 if (try_update_scene_state(setup))
1368 return TRUE;
1369
1370 /* Update failed, try to restart the scene.
1371 *
1372 * Cannot call lp_setup_flush_and_restart() directly here
1373 * because of potential recursion.
1374 */
1375 if (!set_scene_state(setup, SETUP_FLUSHED, __FUNCTION__))
1376 return FALSE;
1377
1378 if (!set_scene_state(setup, SETUP_ACTIVE, __FUNCTION__))
1379 return FALSE;
1380
1381 if (!setup->scene)
1382 return FALSE;
1383
1384 return try_update_scene_state(setup);
1385 }
1386
1387 return TRUE;
1388 }
1389
1390
1391
1392 /* Only caller is lp_setup_vbuf_destroy()
1393 */
1394 void
1395 lp_setup_destroy( struct lp_setup_context *setup )
1396 {
1397 uint i;
1398
1399 lp_setup_reset( setup );
1400
1401 util_unreference_framebuffer_state(&setup->fb);
1402
1403 for (i = 0; i < ARRAY_SIZE(setup->fs.current_tex); i++) {
1404 pipe_resource_reference(&setup->fs.current_tex[i], NULL);
1405 }
1406
1407 for (i = 0; i < ARRAY_SIZE(setup->constants); i++) {
1408 pipe_resource_reference(&setup->constants[i].current.buffer, NULL);
1409 }
1410
1411 for (i = 0; i < ARRAY_SIZE(setup->ssbos); i++) {
1412 pipe_resource_reference(&setup->ssbos[i].current.buffer, NULL);
1413 }
1414
1415 /* free the scenes in the 'empty' queue */
1416 for (i = 0; i < ARRAY_SIZE(setup->scenes); i++) {
1417 struct lp_scene *scene = setup->scenes[i];
1418
1419 if (scene->fence)
1420 lp_fence_wait(scene->fence);
1421
1422 lp_scene_destroy(scene);
1423 }
1424
1425 lp_fence_reference(&setup->last_fence, NULL);
1426
1427 FREE( setup );
1428 }
1429
1430
1431 /**
1432 * Create a new primitive tiling engine. Plug it into the backend of
1433 * the draw module. Currently also creates a rasterizer to use with
1434 * it.
1435 */
1436 struct lp_setup_context *
1437 lp_setup_create( struct pipe_context *pipe,
1438 struct draw_context *draw )
1439 {
1440 struct llvmpipe_screen *screen = llvmpipe_screen(pipe->screen);
1441 struct lp_setup_context *setup;
1442 unsigned i;
1443
1444 setup = CALLOC_STRUCT(lp_setup_context);
1445 if (!setup) {
1446 goto no_setup;
1447 }
1448
1449 lp_setup_init_vbuf(setup);
1450
1451 /* Used only in update_state():
1452 */
1453 setup->pipe = pipe;
1454
1455
1456 setup->num_threads = screen->num_threads;
1457 setup->vbuf = draw_vbuf_stage(draw, &setup->base);
1458 if (!setup->vbuf) {
1459 goto no_vbuf;
1460 }
1461
1462 draw_set_rasterize_stage(draw, setup->vbuf);
1463 draw_set_render(draw, &setup->base);
1464
1465 /* create some empty scenes */
1466 for (i = 0; i < MAX_SCENES; i++) {
1467 setup->scenes[i] = lp_scene_create( pipe );
1468 if (!setup->scenes[i]) {
1469 goto no_scenes;
1470 }
1471 }
1472
1473 setup->triangle = first_triangle;
1474 setup->line = first_line;
1475 setup->point = first_point;
1476
1477 setup->dirty = ~0;
1478
1479 /* Initialize empty default fb correctly, so the rect is empty */
1480 setup->framebuffer.x1 = -1;
1481 setup->framebuffer.y1 = -1;
1482
1483 return setup;
1484
1485 no_scenes:
1486 for (i = 0; i < MAX_SCENES; i++) {
1487 if (setup->scenes[i]) {
1488 lp_scene_destroy(setup->scenes[i]);
1489 }
1490 }
1491
1492 setup->vbuf->destroy(setup->vbuf);
1493 no_vbuf:
1494 FREE(setup);
1495 no_setup:
1496 return NULL;
1497 }
1498
1499
1500 /**
1501 * Put a BeginQuery command into all bins.
1502 */
1503 void
1504 lp_setup_begin_query(struct lp_setup_context *setup,
1505 struct llvmpipe_query *pq)
1506 {
1507
1508 set_scene_state(setup, SETUP_ACTIVE, "begin_query");
1509
1510 if (!(pq->type == PIPE_QUERY_OCCLUSION_COUNTER ||
1511 pq->type == PIPE_QUERY_OCCLUSION_PREDICATE ||
1512 pq->type == PIPE_QUERY_OCCLUSION_PREDICATE_CONSERVATIVE ||
1513 pq->type == PIPE_QUERY_PIPELINE_STATISTICS))
1514 return;
1515
1516 /* init the query to its beginning state */
1517 assert(setup->active_binned_queries < LP_MAX_ACTIVE_BINNED_QUERIES);
1518 /* exceeding list size so just ignore the query */
1519 if (setup->active_binned_queries >= LP_MAX_ACTIVE_BINNED_QUERIES) {
1520 return;
1521 }
1522 assert(setup->active_queries[setup->active_binned_queries] == NULL);
1523 setup->active_queries[setup->active_binned_queries] = pq;
1524 setup->active_binned_queries++;
1525
1526 assert(setup->scene);
1527 if (setup->scene) {
1528 if (!lp_scene_bin_everywhere(setup->scene,
1529 LP_RAST_OP_BEGIN_QUERY,
1530 lp_rast_arg_query(pq))) {
1531
1532 if (!lp_setup_flush_and_restart(setup))
1533 return;
1534
1535 if (!lp_scene_bin_everywhere(setup->scene,
1536 LP_RAST_OP_BEGIN_QUERY,
1537 lp_rast_arg_query(pq))) {
1538 return;
1539 }
1540 }
1541 setup->scene->had_queries |= TRUE;
1542 }
1543 }
1544
1545
1546 /**
1547 * Put an EndQuery command into all bins.
1548 */
1549 void
1550 lp_setup_end_query(struct lp_setup_context *setup, struct llvmpipe_query *pq)
1551 {
1552 set_scene_state(setup, SETUP_ACTIVE, "end_query");
1553
1554 assert(setup->scene);
1555 if (setup->scene) {
1556 /* pq->fence should be the fence of the *last* scene which
1557 * contributed to the query result.
1558 */
1559 lp_fence_reference(&pq->fence, setup->scene->fence);
1560
1561 if (pq->type == PIPE_QUERY_OCCLUSION_COUNTER ||
1562 pq->type == PIPE_QUERY_OCCLUSION_PREDICATE ||
1563 pq->type == PIPE_QUERY_OCCLUSION_PREDICATE_CONSERVATIVE ||
1564 pq->type == PIPE_QUERY_PIPELINE_STATISTICS ||
1565 pq->type == PIPE_QUERY_TIMESTAMP) {
1566 if (pq->type == PIPE_QUERY_TIMESTAMP &&
1567 !(setup->scene->tiles_x | setup->scene->tiles_y)) {
1568 /*
1569 * If there's a zero width/height framebuffer, there's no bins and
1570 * hence no rast task is ever run. So fill in something here instead.
1571 */
1572 pq->end[0] = os_time_get_nano();
1573 }
1574
1575 if (!lp_scene_bin_everywhere(setup->scene,
1576 LP_RAST_OP_END_QUERY,
1577 lp_rast_arg_query(pq))) {
1578 if (!lp_setup_flush_and_restart(setup))
1579 goto fail;
1580
1581 if (!lp_scene_bin_everywhere(setup->scene,
1582 LP_RAST_OP_END_QUERY,
1583 lp_rast_arg_query(pq))) {
1584 goto fail;
1585 }
1586 }
1587 setup->scene->had_queries |= TRUE;
1588 }
1589 }
1590 else {
1591 lp_fence_reference(&pq->fence, setup->last_fence);
1592 }
1593
1594 fail:
1595 /* Need to do this now not earlier since it still needs to be marked as
1596 * active when binning it would cause a flush.
1597 */
1598 if (pq->type == PIPE_QUERY_OCCLUSION_COUNTER ||
1599 pq->type == PIPE_QUERY_OCCLUSION_PREDICATE ||
1600 pq->type == PIPE_QUERY_OCCLUSION_PREDICATE_CONSERVATIVE ||
1601 pq->type == PIPE_QUERY_PIPELINE_STATISTICS) {
1602 unsigned i;
1603
1604 /* remove from active binned query list */
1605 for (i = 0; i < setup->active_binned_queries; i++) {
1606 if (setup->active_queries[i] == pq)
1607 break;
1608 }
1609 assert(i < setup->active_binned_queries);
1610 if (i == setup->active_binned_queries)
1611 return;
1612 setup->active_binned_queries--;
1613 setup->active_queries[i] = setup->active_queries[setup->active_binned_queries];
1614 setup->active_queries[setup->active_binned_queries] = NULL;
1615 }
1616 }
1617
1618
1619 boolean
1620 lp_setup_flush_and_restart(struct lp_setup_context *setup)
1621 {
1622 if (0) debug_printf("%s\n", __FUNCTION__);
1623
1624 assert(setup->state == SETUP_ACTIVE);
1625
1626 if (!set_scene_state(setup, SETUP_FLUSHED, __FUNCTION__))
1627 return FALSE;
1628
1629 if (!lp_setup_update_state(setup, TRUE))
1630 return FALSE;
1631
1632 return TRUE;
1633 }
1634
1635