gen6_blorp_emit_batch_head(struct brw_context *brw,
const brw_blorp_params *params)
{
- struct gl_context *ctx = &brw->ctx;
-
/* To ensure that the batch contains only the resolve, flush the batch
* before beginning and after finishing emitting the resolve packets.
*/
- intel_flush(ctx);
+ intel_batchbuffer_flush(brw);
}
} else {
perf_debug("Stalling on the GPU for mapping a busy buffer "
"object\n");
- intel_flush(ctx);
+ intel_batchbuffer_flush(brw);
}
} else if (drm_intel_bo_busy(intel_obj->buffer) &&
(access & GL_MAP_INVALIDATE_BUFFER_BIT)) {
brw->NewGLState |= new_state;
}
-void
-_intel_flush(struct gl_context *ctx, const char *file, int line)
-{
- struct brw_context *brw = brw_context(ctx);
-
- if (brw->batch.used)
- _intel_batchbuffer_flush(brw, file, line);
-}
-
static void
intel_glFlush(struct gl_context *ctx)
{
struct brw_context *brw = brw_context(ctx);
- intel_flush(ctx);
+ intel_batchbuffer_flush(brw);
intel_flush_front(ctx);
if (brw->is_front_buffer_rendering)
brw->need_throttle = true;
{
struct brw_context *brw = brw_context(ctx);
- intel_flush(ctx);
+ intel_batchbuffer_flush(brw);
intel_flush_front(ctx);
if (brw->batch.last_bo)
* query, we need to make sure all the pending drawing has landed in the
* real front buffer.
*/
- intel_flush(&brw->ctx);
+ intel_batchbuffer_flush(brw);
intel_flush_front(&brw->ctx);
attachments[i++] = __DRI_BUFFER_FRONT_LEFT;
* So before doing the query, make sure all the pending drawing has
* landed in the real front buffer.
*/
- intel_flush(&brw->ctx);
+ intel_batchbuffer_flush(brw);
intel_flush_front(&brw->ctx);
}
unsigned *dri_ctx_error);
extern void intelFinish(struct gl_context * ctx);
-extern void _intel_flush(struct gl_context * ctx, const char *file, int line);
-
-#define intel_flush(ctx) _intel_flush(ctx, __FILE__, __LINE__)
extern void intelInitDriverFunctions(struct dd_function_table *functions);
void *
intel_miptree_map_raw(struct brw_context *brw, struct intel_mipmap_tree *mt)
{
- struct gl_context *ctx = &brw->ctx;
/* CPU accesses to color buffers don't understand fast color clears, so
* resolve any pending fast color clears before we map.
*/
}
}
- intel_flush(ctx);
+ intel_batchbuffer_flush(brw);
if (mt->region->tiling != I915_TILING_NONE)
drm_intel_gem_bo_map_gtt(bo);
#include "intel_pixel.h"
#include "intel_fbo.h"
#include "intel_blit.h"
+#include "intel_batchbuffer.h"
#define FILE_DEBUG_FLAG DEBUG_PIXEL
intel_prepare_render(brw);
- intel_flush(&brw->ctx);
+ intel_batchbuffer_flush(brw);
/* Clip to destination buffer. */
orig_dstx = dstx;
sync->bo = brw->batch.bo;
drm_intel_bo_reference(sync->bo);
- intel_flush(ctx);
+ intel_batchbuffer_flush(brw);
}
static void intel_client_wait_sync(struct gl_context *ctx, struct gl_sync_object *s,