projects
/
mesa.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
i965: Use {} to initialize GENX_* structs.
[mesa.git]
/
src
/
mesa
/
drivers
/
dri
/
i965
/
intel_batchbuffer.c
diff --git
a/src/mesa/drivers/dri/i965/intel_batchbuffer.c
b/src/mesa/drivers/dri/i965/intel_batchbuffer.c
index de93aeb0b2911259bfbb9b93a46044e42176dce6..e2f208a3d1fdc2f8491dbab996e8d5c93a99407e 100644
(file)
--- a/
src/mesa/drivers/dri/i965/intel_batchbuffer.c
+++ b/
src/mesa/drivers/dri/i965/intel_batchbuffer.c
@@
-78,8
+78,8
@@
intel_batchbuffer_init(struct intel_batchbuffer *batch,
batch->exec_array_size = 100;
batch->exec_bos =
malloc(batch->exec_array_size * sizeof(batch->exec_bos[0]));
batch->exec_array_size = 100;
batch->exec_bos =
malloc(batch->exec_array_size * sizeof(batch->exec_bos[0]));
- batch->
exec_objects
=
- malloc(batch->exec_array_size * sizeof(batch->
exec_objects
[0]));
+ batch->
validation_list
=
+ malloc(batch->exec_array_size * sizeof(batch->
validation_list
[0]));
if (INTEL_DEBUG & DEBUG_BATCH) {
batch->state_batch_sizes =
if (INTEL_DEBUG & DEBUG_BATCH) {
batch->state_batch_sizes =
@@
-100,7
+100,7
@@
intel_batchbuffer_reset(struct intel_batchbuffer *batch,
batch->bo = brw_bo_alloc(bufmgr, "batchbuffer", BATCH_SZ, 4096);
if (has_llc) {
batch->bo = brw_bo_alloc(bufmgr, "batchbuffer", BATCH_SZ, 4096);
if (has_llc) {
- batch->map = brw_bo_map
_cpu
(NULL, batch->bo, MAP_READ | MAP_WRITE);
+ batch->map = brw_bo_map(NULL, batch->bo, MAP_READ | MAP_WRITE);
}
batch->map_next = batch->map;
}
batch->map_next = batch->map;
@@
-162,7
+162,7
@@
intel_batchbuffer_free(struct intel_batchbuffer *batch)
}
free(batch->relocs);
free(batch->exec_bos);
}
free(batch->relocs);
free(batch->exec_bos);
- free(batch->
exec_objects
);
+ free(batch->
validation_list
);
brw_bo_unreference(batch->last_bo);
brw_bo_unreference(batch->bo);
brw_bo_unreference(batch->last_bo);
brw_bo_unreference(batch->bo);
@@
-239,7
+239,7
@@
do_batch_dump(struct brw_context *brw)
if (batch->ring != RENDER_RING)
return;
if (batch->ring != RENDER_RING)
return;
- void *map = brw_bo_map
_cpu
(brw, batch->bo, MAP_READ);
+ void *map = brw_bo_map(brw, batch->bo, MAP_READ);
if (map == NULL) {
fprintf(stderr,
"WARNING: failed to map batchbuffer, "
if (map == NULL) {
fprintf(stderr,
"WARNING: failed to map batchbuffer, "
@@
-465,12
+465,6
@@
brw_finish_batch(struct brw_context *brw)
PIPE_CONTROL_CS_STALL);
}
}
PIPE_CONTROL_CS_STALL);
}
}
-
- /* Mark that the current program cache BO has been used by the GPU.
- * It will be reallocated if we need to put new programs in for the
- * next batch.
- */
- brw->cache.bo_used_by_gpu = true;
}
static void
}
static void
@@
-497,7
+491,7
@@
throttle(struct brw_context *brw)
/* Pass NULL rather than brw so we avoid perf_debug warnings;
* stalling is common and expected here...
*/
/* Pass NULL rather than brw so we avoid perf_debug warnings;
* stalling is common and expected here...
*/
- brw_bo_wait_rendering(
NULL,
brw->throttle_batch[1]);
+ brw_bo_wait_rendering(brw->throttle_batch[1]);
}
brw_bo_unreference(brw->throttle_batch[1]);
}
}
brw_bo_unreference(brw->throttle_batch[1]);
}
@@
-532,13
+526,13
@@
add_exec_bo(struct intel_batchbuffer *batch, struct brw_bo *bo)
batch->exec_bos =
realloc(batch->exec_bos,
batch->exec_array_size * sizeof(batch->exec_bos[0]));
batch->exec_bos =
realloc(batch->exec_bos,
batch->exec_array_size * sizeof(batch->exec_bos[0]));
- batch->
exec_objects
=
- realloc(batch->
exec_objects
,
- batch->exec_array_size * sizeof(batch->
exec_objects
[0]));
+ batch->
validation_list
=
+ realloc(batch->
validation_list
,
+ batch->exec_array_size * sizeof(batch->
validation_list
[0]));
}
struct drm_i915_gem_exec_object2 *validation_entry =
}
struct drm_i915_gem_exec_object2 *validation_entry =
- &batch->
exec_objects
[batch->exec_count];
+ &batch->
validation_list
[batch->exec_count];
validation_entry->handle = bo->gem_handle;
if (bo == batch->bo) {
validation_entry->relocation_count = batch->reloc_count;
validation_entry->handle = bo->gem_handle;
if (bo == batch->bo) {
validation_entry->relocation_count = batch->reloc_count;
@@
-568,7
+562,7
@@
execbuffer(int fd,
int flags)
{
struct drm_i915_gem_execbuffer2 execbuf = {
int flags)
{
struct drm_i915_gem_execbuffer2 execbuf = {
- .buffers_ptr = (uintptr_t) batch->
exec_objects
,
+ .buffers_ptr = (uintptr_t) batch->
validation_list
,
.buffer_count = batch->exec_count,
.batch_start_offset = 0,
.batch_len = used,
.buffer_count = batch->exec_count,
.batch_start_offset = 0,
.batch_len = used,
@@
-599,10
+593,10
@@
execbuffer(int fd,
bo->idle = false;
/* Update brw_bo::offset64 */
bo->idle = false;
/* Update brw_bo::offset64 */
- if (batch->
exec_objects
[i].offset != bo->offset64) {
+ if (batch->
validation_list
[i].offset != bo->offset64) {
DBG("BO %d migrated: 0x%" PRIx64 " -> 0x%llx\n",
DBG("BO %d migrated: 0x%" PRIx64 " -> 0x%llx\n",
- bo->gem_handle, bo->offset64, batch->
exec_objects
[i].offset);
- bo->offset64 = batch->
exec_objects
[i].offset;
+ bo->gem_handle, bo->offset64, batch->
validation_list
[i].offset);
+ bo->offset64 = batch->
validation_list
[i].offset;
}
}
}
}
@@
-723,7
+717,7
@@
_intel_batchbuffer_flush_fence(struct brw_context *brw,
if (unlikely(INTEL_DEBUG & DEBUG_SYNC)) {
fprintf(stderr, "waiting for idle\n");
if (unlikely(INTEL_DEBUG & DEBUG_SYNC)) {
fprintf(stderr, "waiting for idle\n");
- brw_bo_wait_rendering(brw
, brw
->batch.bo);
+ brw_bo_wait_rendering(brw->batch.bo);
}
/* Start a new batch buffer. */
}
/* Start a new batch buffer. */