uint32_t offset,
int size)
{
- struct intel_batchbuffer *batch = &brw->intel.batch;
+ struct intel_batchbuffer *batch = &brw->batch;
if (!brw->state_batch_list) {
/* Our structs are always aligned to at least 32 bytes, so
* is annotated according to the type of each data structure.
*/
void
-brw_annotate_aub(struct intel_context *intel)
+brw_annotate_aub(struct brw_context *brw)
{
- struct brw_context *brw = brw_context(&intel->ctx);
-
unsigned annotation_count = 2 * brw->state_batch_count + 1;
drm_intel_aub_annotation annotations[annotation_count];
int a = 0;
make_annotation(&annotations[a++], AUB_TRACE_TYPE_BATCH, 0,
- 4*intel->batch.used);
+ 4*brw->batch.used);
for (int i = brw->state_batch_count; i-- > 0; ) {
uint32_t type = brw->state_batch_list[i].type;
uint32_t start_offset = brw->state_batch_list[i].offset;
AUB_TRACE_SUBTYPE(type), end_offset);
}
assert(a == annotation_count);
- drm_intel_bufmgr_gem_set_aub_annotations(intel->batch.bo, annotations,
+ drm_intel_bufmgr_gem_set_aub_annotations(brw->batch.bo, annotations,
annotation_count);
}
int alignment,
uint32_t *out_offset)
{
- struct intel_batchbuffer *batch = &brw->intel.batch;
+ struct intel_batchbuffer *batch = &brw->batch;
uint32_t offset;
assert(size < batch->bo->size);
*/
if (batch->state_batch_offset < size ||
offset < 4*batch->used + batch->reserved_space) {
- intel_batchbuffer_flush(&brw->intel);
+ intel_batchbuffer_flush(brw);
offset = ROUND_DOWN_TO(batch->state_batch_offset - size, alignment);
}