llvmpipe: enable support for primitives generated outside streamout
[mesa.git] / src / gallium / drivers / llvmpipe / lp_query.c
1 /**************************************************************************
2 *
3 * Copyright 2007 VMware, Inc.
4 * Copyright 2010 VMware, Inc.
5 * All Rights Reserved.
6 *
7 * Permission is hereby granted, free of charge, to any person obtaining a
8 * copy of this software and associated documentation files (the
9 * "Software"), to deal in the Software without restriction, including
10 * without limitation the rights to use, copy, modify, merge, publish,
11 * distribute, sub license, and/or sell copies of the Software, and to
12 * permit persons to whom the Software is furnished to do so, subject to
13 * the following conditions:
14 *
15 * The above copyright notice and this permission notice (including the
16 * next paragraph) shall be included in all copies or substantial portions
17 * of the Software.
18 *
19 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
20 * OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
21 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT.
22 * IN NO EVENT SHALL THE AUTHORS AND/OR ITS SUPPLIERS BE LIABLE FOR
23 * ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,
24 * TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
25 * SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
26 *
27 **************************************************************************/
28
29 /* Authors:
30 * Keith Whitwell, Qicheng Christopher Li, Brian Paul
31 */
32
33 #include "draw/draw_context.h"
34 #include "pipe/p_defines.h"
35 #include "util/u_memory.h"
36 #include "util/os_time.h"
37 #include "lp_context.h"
38 #include "lp_flush.h"
39 #include "lp_fence.h"
40 #include "lp_query.h"
41 #include "lp_screen.h"
42 #include "lp_state.h"
43 #include "lp_rast.h"
44
45
46 static struct llvmpipe_query *llvmpipe_query( struct pipe_query *p )
47 {
48 return (struct llvmpipe_query *)p;
49 }
50
51 static struct pipe_query *
52 llvmpipe_create_query(struct pipe_context *pipe,
53 unsigned type,
54 unsigned index)
55 {
56 struct llvmpipe_query *pq;
57
58 assert(type < PIPE_QUERY_TYPES);
59
60 pq = CALLOC_STRUCT( llvmpipe_query );
61
62 if (pq) {
63 pq->type = type;
64 }
65
66 return (struct pipe_query *) pq;
67 }
68
69
70 static void
71 llvmpipe_destroy_query(struct pipe_context *pipe, struct pipe_query *q)
72 {
73 struct llvmpipe_query *pq = llvmpipe_query(q);
74
75 /* Ideally we would refcount queries & not get destroyed until the
76 * last scene had finished with us.
77 */
78 if (pq->fence) {
79 if (!lp_fence_issued(pq->fence))
80 llvmpipe_flush(pipe, NULL, __FUNCTION__);
81
82 if (!lp_fence_signalled(pq->fence))
83 lp_fence_wait(pq->fence);
84
85 lp_fence_reference(&pq->fence, NULL);
86 }
87
88 FREE(pq);
89 }
90
91
92 static bool
93 llvmpipe_get_query_result(struct pipe_context *pipe,
94 struct pipe_query *q,
95 bool wait,
96 union pipe_query_result *vresult)
97 {
98 struct llvmpipe_screen *screen = llvmpipe_screen(pipe->screen);
99 unsigned num_threads = MAX2(1, screen->num_threads);
100 struct llvmpipe_query *pq = llvmpipe_query(q);
101 uint64_t *result = (uint64_t *)vresult;
102 int i;
103
104 if (pq->fence) {
105 /* only have a fence if there was a scene */
106 if (!lp_fence_signalled(pq->fence)) {
107 if (!lp_fence_issued(pq->fence))
108 llvmpipe_flush(pipe, NULL, __FUNCTION__);
109
110 if (!wait)
111 return false;
112
113 lp_fence_wait(pq->fence);
114 }
115 }
116
117 /* Sum the results from each of the threads:
118 */
119 *result = 0;
120
121 switch (pq->type) {
122 case PIPE_QUERY_OCCLUSION_COUNTER:
123 for (i = 0; i < num_threads; i++) {
124 *result += pq->end[i];
125 }
126 break;
127 case PIPE_QUERY_OCCLUSION_PREDICATE:
128 case PIPE_QUERY_OCCLUSION_PREDICATE_CONSERVATIVE:
129 for (i = 0; i < num_threads; i++) {
130 /* safer (still not guaranteed) when there's an overflow */
131 vresult->b = vresult->b || pq->end[i];
132 }
133 break;
134 case PIPE_QUERY_TIMESTAMP:
135 for (i = 0; i < num_threads; i++) {
136 if (pq->end[i] > *result) {
137 *result = pq->end[i];
138 }
139 }
140 break;
141 case PIPE_QUERY_TIMESTAMP_DISJOINT: {
142 struct pipe_query_data_timestamp_disjoint *td =
143 (struct pipe_query_data_timestamp_disjoint *)vresult;
144 /* os_get_time_nano return nanoseconds */
145 td->frequency = UINT64_C(1000000000);
146 td->disjoint = false;
147 }
148 break;
149 case PIPE_QUERY_GPU_FINISHED:
150 vresult->b = true;
151 break;
152 case PIPE_QUERY_PRIMITIVES_GENERATED:
153 *result = pq->num_primitives_generated;
154 break;
155 case PIPE_QUERY_PRIMITIVES_EMITTED:
156 *result = pq->num_primitives_written;
157 break;
158 case PIPE_QUERY_SO_OVERFLOW_PREDICATE:
159 case PIPE_QUERY_SO_OVERFLOW_ANY_PREDICATE:
160 vresult->b = pq->num_primitives_generated > pq->num_primitives_written;
161 break;
162 case PIPE_QUERY_SO_STATISTICS: {
163 struct pipe_query_data_so_statistics *stats =
164 (struct pipe_query_data_so_statistics *)vresult;
165 stats->num_primitives_written = pq->num_primitives_written;
166 stats->primitives_storage_needed = pq->num_primitives_generated;
167 }
168 break;
169 case PIPE_QUERY_PIPELINE_STATISTICS: {
170 struct pipe_query_data_pipeline_statistics *stats =
171 (struct pipe_query_data_pipeline_statistics *)vresult;
172 /* only ps_invocations come from binned query */
173 for (i = 0; i < num_threads; i++) {
174 pq->stats.ps_invocations += pq->end[i];
175 }
176 pq->stats.ps_invocations *= LP_RASTER_BLOCK_SIZE * LP_RASTER_BLOCK_SIZE;
177 *stats = pq->stats;
178 }
179 break;
180 default:
181 assert(0);
182 break;
183 }
184
185 return true;
186 }
187
188 static void
189 llvmpipe_get_query_result_resource(struct pipe_context *pipe,
190 struct pipe_query *q,
191 bool wait,
192 enum pipe_query_value_type result_type,
193 int index,
194 struct pipe_resource *resource,
195 unsigned offset)
196 {
197 struct llvmpipe_screen *screen = llvmpipe_screen(pipe->screen);
198 unsigned num_threads = MAX2(1, screen->num_threads);
199 struct llvmpipe_query *pq = llvmpipe_query(q);
200 struct llvmpipe_resource *lpr = llvmpipe_resource(resource);
201 bool unflushed = false;
202 bool unsignalled = false;
203 if (pq->fence) {
204 /* only have a fence if there was a scene */
205 if (!lp_fence_signalled(pq->fence)) {
206 unsignalled = true;
207 if (!lp_fence_issued(pq->fence))
208 unflushed = true;
209 }
210 }
211
212
213 uint64_t value = 0;
214 if (index == -1)
215 if (unsignalled)
216 value = 0;
217 else
218 value = 1;
219 else {
220 unsigned i;
221
222 if (unflushed) {
223 llvmpipe_flush(pipe, NULL, __FUNCTION__);
224
225 if (!wait)
226 return;
227
228 lp_fence_wait(pq->fence);
229 }
230
231 switch (pq->type) {
232 case PIPE_QUERY_OCCLUSION_COUNTER:
233 for (i = 0; i < num_threads; i++) {
234 value += pq->end[i];
235 }
236 break;
237 case PIPE_QUERY_OCCLUSION_PREDICATE:
238 case PIPE_QUERY_OCCLUSION_PREDICATE_CONSERVATIVE:
239 for (i = 0; i < num_threads; i++) {
240 /* safer (still not guaranteed) when there's an overflow */
241 value = value || pq->end[i];
242 }
243 break;
244 case PIPE_QUERY_PRIMITIVES_GENERATED:
245 value = pq->num_primitives_generated;
246 break;
247 case PIPE_QUERY_PRIMITIVES_EMITTED:
248 value = pq->num_primitives_written;
249 break;
250 case PIPE_QUERY_TIMESTAMP:
251 for (i = 0; i < num_threads; i++) {
252 if (pq->end[i] > value) {
253 value = pq->end[i];
254 }
255 }
256 break;
257 case PIPE_QUERY_SO_OVERFLOW_PREDICATE:
258 case PIPE_QUERY_SO_OVERFLOW_ANY_PREDICATE:
259 value = !!(pq->num_primitives_generated > pq->num_primitives_written);
260 break;
261 case PIPE_QUERY_PIPELINE_STATISTICS:
262 switch ((enum pipe_statistics_query_index)index) {
263 case PIPE_STAT_QUERY_IA_VERTICES:
264 value = pq->stats.ia_vertices;
265 break;
266 case PIPE_STAT_QUERY_IA_PRIMITIVES:
267 value = pq->stats.ia_primitives;
268 break;
269 case PIPE_STAT_QUERY_VS_INVOCATIONS:
270 value = pq->stats.vs_invocations;
271 break;
272 case PIPE_STAT_QUERY_GS_INVOCATIONS:
273 value = pq->stats.gs_invocations;
274 break;
275 case PIPE_STAT_QUERY_GS_PRIMITIVES:
276 value = pq->stats.gs_primitives;
277 break;
278 case PIPE_STAT_QUERY_C_INVOCATIONS:
279 value = pq->stats.c_invocations;
280 break;
281 case PIPE_STAT_QUERY_C_PRIMITIVES:
282 value = pq->stats.c_primitives;
283 break;
284 case PIPE_STAT_QUERY_PS_INVOCATIONS:
285 value = 0;
286 for (i = 0; i < num_threads; i++) {
287 value += pq->end[i];
288 }
289 value *= LP_RASTER_BLOCK_SIZE * LP_RASTER_BLOCK_SIZE;
290 break;
291 case PIPE_STAT_QUERY_HS_INVOCATIONS:
292 value = pq->stats.hs_invocations;
293 break;
294 case PIPE_STAT_QUERY_DS_INVOCATIONS:
295 value = pq->stats.ds_invocations;
296 break;
297 case PIPE_STAT_QUERY_CS_INVOCATIONS:
298 value = pq->stats.cs_invocations;
299 break;
300 }
301 break;
302 default:
303 fprintf(stderr, "Unknown query type %d\n", pq->type);
304 break;
305 }
306 }
307
308 void *dst = (uint8_t *)lpr->data + offset;
309 switch (result_type) {
310 case PIPE_QUERY_TYPE_I32: {
311 int32_t *iptr = (int32_t *)dst;
312 if (value > 0x7fffffff)
313 *iptr = 0x7fffffff;
314 else
315 *iptr = (int32_t)value;
316 break;
317 }
318 case PIPE_QUERY_TYPE_U32: {
319 uint32_t *uptr = (uint32_t *)dst;
320 if (value > 0xffffffff)
321 *uptr = 0xffffffff;
322 else
323 *uptr = (uint32_t)value;
324 break;
325 }
326 case PIPE_QUERY_TYPE_I64: {
327 int64_t *iptr = (int64_t *)dst;
328 *iptr = (int64_t)value;
329 break;
330 }
331 case PIPE_QUERY_TYPE_U64: {
332 uint64_t *uptr = (uint64_t *)dst;
333 *uptr = (uint64_t)value;
334 break;
335 }
336 }
337 }
338
339 static bool
340 llvmpipe_begin_query(struct pipe_context *pipe, struct pipe_query *q)
341 {
342 struct llvmpipe_context *llvmpipe = llvmpipe_context( pipe );
343 struct llvmpipe_query *pq = llvmpipe_query(q);
344
345 /* Check if the query is already in the scene. If so, we need to
346 * flush the scene now. Real apps shouldn't re-use a query in a
347 * frame of rendering.
348 */
349 if (pq->fence && !lp_fence_issued(pq->fence)) {
350 llvmpipe_finish(pipe, __FUNCTION__);
351 }
352
353
354 memset(pq->start, 0, sizeof(pq->start));
355 memset(pq->end, 0, sizeof(pq->end));
356 lp_setup_begin_query(llvmpipe->setup, pq);
357
358 switch (pq->type) {
359 case PIPE_QUERY_PRIMITIVES_EMITTED:
360 pq->num_primitives_written = llvmpipe->so_stats.num_primitives_written;
361 break;
362 case PIPE_QUERY_PRIMITIVES_GENERATED:
363 pq->num_primitives_generated = llvmpipe->so_stats.primitives_storage_needed;
364 llvmpipe->active_primgen_queries++;
365 break;
366 case PIPE_QUERY_SO_STATISTICS:
367 pq->num_primitives_written = llvmpipe->so_stats.num_primitives_written;
368 pq->num_primitives_generated = llvmpipe->so_stats.primitives_storage_needed;
369 break;
370 case PIPE_QUERY_SO_OVERFLOW_PREDICATE:
371 case PIPE_QUERY_SO_OVERFLOW_ANY_PREDICATE:
372 pq->num_primitives_written = llvmpipe->so_stats.num_primitives_written;
373 pq->num_primitives_generated = llvmpipe->so_stats.primitives_storage_needed;
374 break;
375 case PIPE_QUERY_PIPELINE_STATISTICS:
376 /* reset our cache */
377 if (llvmpipe->active_statistics_queries == 0) {
378 memset(&llvmpipe->pipeline_statistics, 0,
379 sizeof(llvmpipe->pipeline_statistics));
380 }
381 memcpy(&pq->stats, &llvmpipe->pipeline_statistics, sizeof(pq->stats));
382 llvmpipe->active_statistics_queries++;
383 break;
384 case PIPE_QUERY_OCCLUSION_COUNTER:
385 case PIPE_QUERY_OCCLUSION_PREDICATE:
386 case PIPE_QUERY_OCCLUSION_PREDICATE_CONSERVATIVE:
387 llvmpipe->active_occlusion_queries++;
388 llvmpipe->dirty |= LP_NEW_OCCLUSION_QUERY;
389 break;
390 default:
391 break;
392 }
393 return true;
394 }
395
396
397 static bool
398 llvmpipe_end_query(struct pipe_context *pipe, struct pipe_query *q)
399 {
400 struct llvmpipe_context *llvmpipe = llvmpipe_context( pipe );
401 struct llvmpipe_query *pq = llvmpipe_query(q);
402
403 lp_setup_end_query(llvmpipe->setup, pq);
404
405 switch (pq->type) {
406
407 case PIPE_QUERY_PRIMITIVES_EMITTED:
408 pq->num_primitives_written =
409 llvmpipe->so_stats.num_primitives_written - pq->num_primitives_written;
410 break;
411 case PIPE_QUERY_PRIMITIVES_GENERATED:
412 assert(llvmpipe->active_primgen_queries);
413 llvmpipe->active_primgen_queries--;
414 pq->num_primitives_generated =
415 llvmpipe->so_stats.primitives_storage_needed - pq->num_primitives_generated;
416 break;
417 case PIPE_QUERY_SO_STATISTICS:
418 pq->num_primitives_written =
419 llvmpipe->so_stats.num_primitives_written - pq->num_primitives_written;
420 pq->num_primitives_generated =
421 llvmpipe->so_stats.primitives_storage_needed - pq->num_primitives_generated;
422 break;
423 case PIPE_QUERY_SO_OVERFLOW_PREDICATE:
424 case PIPE_QUERY_SO_OVERFLOW_ANY_PREDICATE:
425 pq->num_primitives_written =
426 llvmpipe->so_stats.num_primitives_written - pq->num_primitives_written;
427 pq->num_primitives_generated =
428 llvmpipe->so_stats.primitives_storage_needed - pq->num_primitives_generated;
429 break;
430 case PIPE_QUERY_PIPELINE_STATISTICS:
431 pq->stats.ia_vertices =
432 llvmpipe->pipeline_statistics.ia_vertices - pq->stats.ia_vertices;
433 pq->stats.ia_primitives =
434 llvmpipe->pipeline_statistics.ia_primitives - pq->stats.ia_primitives;
435 pq->stats.vs_invocations =
436 llvmpipe->pipeline_statistics.vs_invocations - pq->stats.vs_invocations;
437 pq->stats.gs_invocations =
438 llvmpipe->pipeline_statistics.gs_invocations - pq->stats.gs_invocations;
439 pq->stats.gs_primitives =
440 llvmpipe->pipeline_statistics.gs_primitives - pq->stats.gs_primitives;
441 pq->stats.c_invocations =
442 llvmpipe->pipeline_statistics.c_invocations - pq->stats.c_invocations;
443 pq->stats.c_primitives =
444 llvmpipe->pipeline_statistics.c_primitives - pq->stats.c_primitives;
445 pq->stats.ps_invocations =
446 llvmpipe->pipeline_statistics.ps_invocations - pq->stats.ps_invocations;
447 pq->stats.cs_invocations =
448 llvmpipe->pipeline_statistics.cs_invocations - pq->stats.cs_invocations;
449 llvmpipe->active_statistics_queries--;
450 break;
451 case PIPE_QUERY_OCCLUSION_COUNTER:
452 case PIPE_QUERY_OCCLUSION_PREDICATE:
453 case PIPE_QUERY_OCCLUSION_PREDICATE_CONSERVATIVE:
454 assert(llvmpipe->active_occlusion_queries);
455 llvmpipe->active_occlusion_queries--;
456 llvmpipe->dirty |= LP_NEW_OCCLUSION_QUERY;
457 break;
458 default:
459 break;
460 }
461
462 return true;
463 }
464
465 boolean
466 llvmpipe_check_render_cond(struct llvmpipe_context *lp)
467 {
468 struct pipe_context *pipe = &lp->pipe;
469 boolean b, wait;
470 uint64_t result;
471
472 if (!lp->render_cond_query)
473 return TRUE; /* no query predicate, draw normally */
474
475 wait = (lp->render_cond_mode == PIPE_RENDER_COND_WAIT ||
476 lp->render_cond_mode == PIPE_RENDER_COND_BY_REGION_WAIT);
477
478 b = pipe->get_query_result(pipe, lp->render_cond_query, wait, (void*)&result);
479 if (b)
480 return ((!result) == lp->render_cond_cond);
481 else
482 return TRUE;
483 }
484
485 static void
486 llvmpipe_set_active_query_state(struct pipe_context *pipe, bool enable)
487 {
488 struct llvmpipe_context *llvmpipe = llvmpipe_context(pipe);
489
490 llvmpipe->queries_disabled = !enable;
491 /* for OQs we need to regenerate the fragment shader */
492 llvmpipe->dirty |= LP_NEW_OCCLUSION_QUERY;
493 }
494
495 void llvmpipe_init_query_funcs(struct llvmpipe_context *llvmpipe )
496 {
497 llvmpipe->pipe.create_query = llvmpipe_create_query;
498 llvmpipe->pipe.destroy_query = llvmpipe_destroy_query;
499 llvmpipe->pipe.begin_query = llvmpipe_begin_query;
500 llvmpipe->pipe.end_query = llvmpipe_end_query;
501 llvmpipe->pipe.get_query_result = llvmpipe_get_query_result;
502 llvmpipe->pipe.get_query_result_resource = llvmpipe_get_query_result_resource;
503 llvmpipe->pipe.set_active_query_state = llvmpipe_set_active_query_state;
504 }
505
506