iris_binder_reserve(struct iris_binder *binder, unsigned size,
uint32_t *out_offset)
{
+ /* XXX: if we ever make this allocate a new BO, then make binder_reserve
+ * return the BO, so at least verify use_pinned_bo gets the right one
+ */
/* XXX: Implement a real ringbuffer, for now just croak if run out */
assert(size > 0);
assert(binder->insert_point + size <= BINDER_SIZE);
uint32_t *bt_map = iris_binder_reserve(&ice->state.binder,
num_entries * sizeof(uint32_t),
bt_offset);
+ iris_use_pinned_bo(batch, ice->state.binder.bo, false);
for (unsigned i = 0; i < num_entries; i++) {
surface_maps[i] = stream_state(batch, ice->state.surface_uploader,