/**************************************************************************
*
- * Copyright 2007 Tungsten Graphics, Inc., Cedar Park, Texas.
+ * Copyright 2007 VMware, Inc.
* Copyright 2010 VMware, Inc.
* All Rights Reserved.
*
#include "draw/draw_context.h"
#include "pipe/p_defines.h"
#include "util/u_memory.h"
+#include "os/os_time.h"
#include "lp_context.h"
#include "lp_flush.h"
+#include "lp_fence.h"
#include "lp_query.h"
-#include "lp_rast.h"
-#include "lp_rast_priv.h"
+#include "lp_screen.h"
#include "lp_state.h"
+#include "lp_rast.h"
static struct llvmpipe_query *llvmpipe_query( struct pipe_query *p )
static struct pipe_query *
llvmpipe_create_query(struct pipe_context *pipe,
- unsigned type)
+ unsigned type,
+ unsigned index)
{
struct llvmpipe_query *pq;
- assert(type == PIPE_QUERY_OCCLUSION_COUNTER);
+ assert(type < PIPE_QUERY_TYPES);
pq = CALLOC_STRUCT( llvmpipe_query );
+
if (pq) {
- pipe_mutex_init(pq->mutex);
+ pq->type = type;
}
return (struct pipe_query *) pq;
llvmpipe_destroy_query(struct pipe_context *pipe, struct pipe_query *q)
{
struct llvmpipe_query *pq = llvmpipe_query(q);
- /* query might still be in process if we never waited for the result */
- if (!pq->done) {
- struct pipe_fence_handle *fence = NULL;
- llvmpipe_flush(pipe, 0, &fence);
- if (fence) {
- pipe->screen->fence_finish(pipe->screen, fence, 0);
- pipe->screen->fence_reference(pipe->screen, &fence, NULL);
- }
+
+ /* Ideally we would refcount queries & not get destroyed until the
+ * last scene had finished with us.
+ */
+ if (pq->fence) {
+ if (!lp_fence_issued(pq->fence))
+ llvmpipe_flush(pipe, NULL, __FUNCTION__);
+
+ if (!lp_fence_signalled(pq->fence))
+ lp_fence_wait(pq->fence);
+
+ lp_fence_reference(&pq->fence, NULL);
}
- pipe_mutex_destroy(pq->mutex);
FREE(pq);
}
llvmpipe_get_query_result(struct pipe_context *pipe,
struct pipe_query *q,
boolean wait,
- void *vresult)
+ union pipe_query_result *vresult)
{
+ struct llvmpipe_screen *screen = llvmpipe_screen(pipe->screen);
+ unsigned num_threads = MAX2(1, screen->num_threads);
struct llvmpipe_query *pq = llvmpipe_query(q);
uint64_t *result = (uint64_t *)vresult;
+ int i;
+
+ if (pq->fence) {
+ /* only have a fence if there was a scene */
+ if (!lp_fence_signalled(pq->fence)) {
+ if (!lp_fence_issued(pq->fence))
+ llvmpipe_flush(pipe, NULL, __FUNCTION__);
+
+ if (!wait)
+ return FALSE;
- if (!pq->done) {
- if (wait) {
- struct pipe_fence_handle *fence = NULL;
- llvmpipe_flush(pipe, 0, &fence);
- if (fence) {
- pipe->screen->fence_finish(pipe->screen, fence, 0);
- pipe->screen->fence_reference(pipe->screen, &fence, NULL);
+ lp_fence_wait(pq->fence);
+ }
+ }
+
+ /* Sum the results from each of the threads:
+ */
+ *result = 0;
+
+ switch (pq->type) {
+ case PIPE_QUERY_OCCLUSION_COUNTER:
+ for (i = 0; i < num_threads; i++) {
+ *result += pq->end[i];
+ }
+ break;
+ case PIPE_QUERY_OCCLUSION_PREDICATE:
+ for (i = 0; i < num_threads; i++) {
+ /* safer (still not guaranteed) when there's an overflow */
+ vresult->b = vresult->b || pq->end[i];
+ }
+ break;
+ case PIPE_QUERY_TIMESTAMP:
+ for (i = 0; i < num_threads; i++) {
+ if (pq->end[i] > *result) {
+ *result = pq->end[i];
}
}
- /* this is a bit inconsequent but should be ok */
- else {
- llvmpipe_flush(pipe, 0, NULL);
+ break;
+ case PIPE_QUERY_TIMESTAMP_DISJOINT: {
+ struct pipe_query_data_timestamp_disjoint *td =
+ (struct pipe_query_data_timestamp_disjoint *)vresult;
+ /* os_get_time_nano return nanoseconds */
+ td->frequency = UINT64_C(1000000000);
+ td->disjoint = FALSE;
+ }
+ break;
+ case PIPE_QUERY_GPU_FINISHED:
+ vresult->b = TRUE;
+ break;
+ case PIPE_QUERY_PRIMITIVES_GENERATED:
+ *result = pq->num_primitives_generated;
+ break;
+ case PIPE_QUERY_PRIMITIVES_EMITTED:
+ *result = pq->num_primitives_written;
+ break;
+ case PIPE_QUERY_SO_OVERFLOW_PREDICATE:
+ vresult->b = pq->num_primitives_generated > pq->num_primitives_written;
+ break;
+ case PIPE_QUERY_SO_STATISTICS: {
+ struct pipe_query_data_so_statistics *stats =
+ (struct pipe_query_data_so_statistics *)vresult;
+ stats->num_primitives_written = pq->num_primitives_written;
+ stats->primitives_storage_needed = pq->num_primitives_generated;
+ }
+ break;
+ case PIPE_QUERY_PIPELINE_STATISTICS: {
+ struct pipe_query_data_pipeline_statistics *stats =
+ (struct pipe_query_data_pipeline_statistics *)vresult;
+ /* only ps_invocations come from binned query */
+ for (i = 0; i < num_threads; i++) {
+ pq->stats.ps_invocations += pq->end[i];
}
+ pq->stats.ps_invocations *= LP_RASTER_BLOCK_SIZE * LP_RASTER_BLOCK_SIZE;
+ *stats = pq->stats;
}
-
- if (pq->done) {
- *result = pq->result;
+ break;
+ default:
+ assert(0);
+ break;
}
- return pq->done;
+ return TRUE;
}
-static void
+static boolean
llvmpipe_begin_query(struct pipe_context *pipe, struct pipe_query *q)
{
struct llvmpipe_context *llvmpipe = llvmpipe_context( pipe );
* flush the scene now. Real apps shouldn't re-use a query in a
* frame of rendering.
*/
- if (pq->binned) {
- struct pipe_fence_handle *fence;
- llvmpipe_flush(pipe, 0, &fence);
- if (fence) {
- pipe->screen->fence_finish(pipe->screen, fence, 0);
- pipe->screen->fence_reference(pipe->screen, &fence, NULL);
- }
+ if (pq->fence && !lp_fence_issued(pq->fence)) {
+ llvmpipe_finish(pipe, __FUNCTION__);
}
+
+ memset(pq->start, 0, sizeof(pq->start));
+ memset(pq->end, 0, sizeof(pq->end));
lp_setup_begin_query(llvmpipe->setup, pq);
- llvmpipe->active_query_count++;
- llvmpipe->dirty |= LP_NEW_QUERY;
+ switch (pq->type) {
+ case PIPE_QUERY_PRIMITIVES_EMITTED:
+ pq->num_primitives_written = llvmpipe->so_stats.num_primitives_written;
+ break;
+ case PIPE_QUERY_PRIMITIVES_GENERATED:
+ pq->num_primitives_generated = llvmpipe->so_stats.primitives_storage_needed;
+ break;
+ case PIPE_QUERY_SO_STATISTICS:
+ pq->num_primitives_written = llvmpipe->so_stats.num_primitives_written;
+ pq->num_primitives_generated = llvmpipe->so_stats.primitives_storage_needed;
+ break;
+ case PIPE_QUERY_SO_OVERFLOW_PREDICATE:
+ pq->num_primitives_written = llvmpipe->so_stats.num_primitives_written;
+ pq->num_primitives_generated = llvmpipe->so_stats.primitives_storage_needed;
+ break;
+ case PIPE_QUERY_PIPELINE_STATISTICS:
+ /* reset our cache */
+ if (llvmpipe->active_statistics_queries == 0) {
+ memset(&llvmpipe->pipeline_statistics, 0,
+ sizeof(llvmpipe->pipeline_statistics));
+ }
+ memcpy(&pq->stats, &llvmpipe->pipeline_statistics, sizeof(pq->stats));
+ llvmpipe->active_statistics_queries++;
+ break;
+ case PIPE_QUERY_OCCLUSION_COUNTER:
+ case PIPE_QUERY_OCCLUSION_PREDICATE:
+ llvmpipe->active_occlusion_queries++;
+ llvmpipe->dirty |= LP_NEW_OCCLUSION_QUERY;
+ break;
+ default:
+ break;
+ }
+ return true;
}
-static void
+static bool
llvmpipe_end_query(struct pipe_context *pipe, struct pipe_query *q)
{
struct llvmpipe_context *llvmpipe = llvmpipe_context( pipe );
lp_setup_end_query(llvmpipe->setup, pq);
- assert(llvmpipe->active_query_count);
- llvmpipe->active_query_count--;
- llvmpipe->dirty |= LP_NEW_QUERY;
+ switch (pq->type) {
+
+ case PIPE_QUERY_PRIMITIVES_EMITTED:
+ pq->num_primitives_written =
+ llvmpipe->so_stats.num_primitives_written - pq->num_primitives_written;
+ break;
+ case PIPE_QUERY_PRIMITIVES_GENERATED:
+ pq->num_primitives_generated =
+ llvmpipe->so_stats.primitives_storage_needed - pq->num_primitives_generated;
+ break;
+ case PIPE_QUERY_SO_STATISTICS:
+ pq->num_primitives_written =
+ llvmpipe->so_stats.num_primitives_written - pq->num_primitives_written;
+ pq->num_primitives_generated =
+ llvmpipe->so_stats.primitives_storage_needed - pq->num_primitives_generated;
+ break;
+ case PIPE_QUERY_SO_OVERFLOW_PREDICATE:
+ pq->num_primitives_written =
+ llvmpipe->so_stats.num_primitives_written - pq->num_primitives_written;
+ pq->num_primitives_generated =
+ llvmpipe->so_stats.primitives_storage_needed - pq->num_primitives_generated;
+ break;
+ case PIPE_QUERY_PIPELINE_STATISTICS:
+ pq->stats.ia_vertices =
+ llvmpipe->pipeline_statistics.ia_vertices - pq->stats.ia_vertices;
+ pq->stats.ia_primitives =
+ llvmpipe->pipeline_statistics.ia_primitives - pq->stats.ia_primitives;
+ pq->stats.vs_invocations =
+ llvmpipe->pipeline_statistics.vs_invocations - pq->stats.vs_invocations;
+ pq->stats.gs_invocations =
+ llvmpipe->pipeline_statistics.gs_invocations - pq->stats.gs_invocations;
+ pq->stats.gs_primitives =
+ llvmpipe->pipeline_statistics.gs_primitives - pq->stats.gs_primitives;
+ pq->stats.c_invocations =
+ llvmpipe->pipeline_statistics.c_invocations - pq->stats.c_invocations;
+ pq->stats.c_primitives =
+ llvmpipe->pipeline_statistics.c_primitives - pq->stats.c_primitives;
+ pq->stats.ps_invocations =
+ llvmpipe->pipeline_statistics.ps_invocations - pq->stats.ps_invocations;
+
+ llvmpipe->active_statistics_queries--;
+ break;
+ case PIPE_QUERY_OCCLUSION_COUNTER:
+ case PIPE_QUERY_OCCLUSION_PREDICATE:
+ assert(llvmpipe->active_occlusion_queries);
+ llvmpipe->active_occlusion_queries--;
+ llvmpipe->dirty |= LP_NEW_OCCLUSION_QUERY;
+ break;
+ default:
+ break;
+ }
+
+ return true;
+}
+
+boolean
+llvmpipe_check_render_cond(struct llvmpipe_context *lp)
+{
+ struct pipe_context *pipe = &lp->pipe;
+ boolean b, wait;
+ uint64_t result;
+
+ if (!lp->render_cond_query)
+ return TRUE; /* no query predicate, draw normally */
+
+ wait = (lp->render_cond_mode == PIPE_RENDER_COND_WAIT ||
+ lp->render_cond_mode == PIPE_RENDER_COND_BY_REGION_WAIT);
+
+ b = pipe->get_query_result(pipe, lp->render_cond_query, wait, (void*)&result);
+ if (b)
+ return ((!result) == lp->render_cond_cond);
+ else
+ return TRUE;
}
+static void
+llvmpipe_set_active_query_state(struct pipe_context *pipe, boolean enable)
+{
+}
void llvmpipe_init_query_funcs(struct llvmpipe_context *llvmpipe )
{
llvmpipe->pipe.begin_query = llvmpipe_begin_query;
llvmpipe->pipe.end_query = llvmpipe_end_query;
llvmpipe->pipe.get_query_result = llvmpipe_get_query_result;
+ llvmpipe->pipe.set_active_query_state = llvmpipe_set_active_query_state;
}