struct blorp_address address, uint32_t delta)
{
struct anv_cmd_buffer *cmd_buffer = batch->driver_batch;
- anv_reloc_list_add(&cmd_buffer->surface_relocs, &cmd_buffer->pool->alloc,
- ss_offset, address.buffer, address.offset + delta);
+ VkResult result =
+ anv_reloc_list_add(&cmd_buffer->surface_relocs, &cmd_buffer->pool->alloc,
+ ss_offset, address.buffer, address.offset + delta);
+ if (result != VK_SUCCESS)
+ anv_batch_set_error(&cmd_buffer->batch, result);
}
static void *
struct anv_cmd_buffer *cmd_buffer = batch->driver_batch;
uint32_t state_offset;
- struct anv_state bt_state =
- anv_cmd_buffer_alloc_binding_table(cmd_buffer, num_entries,
- &state_offset);
- if (bt_state.map == NULL) {
- /* We ran out of space. Grab a new binding table block. */
- VkResult result = anv_cmd_buffer_new_binding_table_block(cmd_buffer);
- assert(result == VK_SUCCESS);
-
- /* Re-emit state base addresses so we get the new surface state base
- * address before we start emitting binding tables etc.
- */
- anv_cmd_buffer_emit_state_base_address(cmd_buffer);
-
- bt_state = anv_cmd_buffer_alloc_binding_table(cmd_buffer, num_entries,
- &state_offset);
- assert(bt_state.map != NULL);
- }
+ struct anv_state bt_state;
+
+ VkResult result =
+ anv_cmd_buffer_alloc_blorp_binding_table(cmd_buffer, num_entries,
+ &state_offset, &bt_state);
+ if (result != VK_SUCCESS)
+ return;
uint32_t *bt_map = bt_state.map;
*bt_offset = bt_state.offset;
surface_offsets[i] = surface_state.offset;
surface_maps[i] = surface_state.map;
}
+
+ anv_state_flush(cmd_buffer->device, bt_state);
}
static void *
return vb_state.map;
}
+static void
+blorp_flush_range(struct blorp_batch *batch, void *start, size_t size)
+{
+ struct anv_device *device = batch->blorp->driver_ctx;
+ if (!device->info.has_llc)
+ anv_flush_range(start, size);
+}
+
static void
blorp_emit_urb_config(struct blorp_batch *batch, unsigned vs_entry_size)
{
struct anv_device *device = batch->blorp->driver_ctx;
struct anv_cmd_buffer *cmd_buffer = batch->driver_batch;
+ const unsigned entry_size[4] = { vs_entry_size, 1, 1, 1 };
+
genX(emit_urb_setup)(device, &cmd_buffer->batch,
+ cmd_buffer->state.current_l3_config,
VK_SHADER_STAGE_VERTEX_BIT |
VK_SHADER_STAGE_FRAGMENT_BIT,
- vs_entry_size, 0,
- cmd_buffer->state.current_l3_config);
+ entry_size);
}
-void genX(blorp_exec)(struct blorp_batch *batch,
- const struct blorp_params *params);
-
void
genX(blorp_exec)(struct blorp_batch *batch,
const struct blorp_params *params)
genX(flush_pipeline_select_3d)(cmd_buffer);
+ genX(cmd_buffer_emit_gen7_depth_flush)(cmd_buffer);
+
+ /* BLORP doesn't do anything fancy with depth such as discards, so we want
+ * the PMA fix off. Also, off is always the safe option.
+ */
+ genX(cmd_buffer_enable_pma_fix)(cmd_buffer, false);
+
blorp_exec(batch, params);
cmd_buffer->state.vb_dirty = ~0;