static bool is_stateobj(struct fd_ringbuffer *ring);
-static void emit_const(struct fd_ringbuffer *ring,
- const struct ir3_shader_variant *v, uint32_t dst_offset,
- uint32_t offset, uint32_t size,
- const void *user_buffer, struct pipe_resource *buffer);
+static void emit_const_user(struct fd_ringbuffer *ring,
+ const struct ir3_shader_variant *v, uint32_t regid,
+ uint32_t size, const uint32_t *user_buffer);
static void emit_const_bo(struct fd_ringbuffer *ring,
+ const struct ir3_shader_variant *v, uint32_t regid,
+ uint32_t offset, uint32_t size,
+ struct fd_bo *bo);
+
+static void emit_const_prsc(struct fd_ringbuffer *ring,
+ const struct ir3_shader_variant *v, uint32_t regid,
+ uint32_t offset, uint32_t size,
+ struct pipe_resource *buffer)
+{
+ struct fd_resource *rsc = fd_resource(buffer);
+ emit_const_bo(ring, v, regid, offset, size, rsc->bo);
+}
+
+static void emit_const_ptrs(struct fd_ringbuffer *ring,
const struct ir3_shader_variant *v, uint32_t dst_offset,
uint32_t num, struct pipe_resource **prscs, uint32_t *offsets);
+static void
+emit_const_asserts(struct fd_ringbuffer *ring,
+ const struct ir3_shader_variant *v,
+ uint32_t regid, uint32_t sizedwords)
+{
+ assert((regid % 4) == 0);
+ assert((sizedwords % 4) == 0);
+ assert(regid + sizedwords <= v->constlen * 4);
+}
static void
ring_wfi(struct fd_batch *batch, struct fd_ringbuffer *ring)
const struct ir3_ubo_analysis_state *state = &const_state->ubo_state;
for (unsigned i = 0; i < state->num_enabled; i++) {
- assert(!state->range[i].bindless);
- unsigned ubo = state->range[i].block;
+ assert(!state->range[i].ubo.bindless);
+ unsigned ubo = state->range[i].ubo.block;
if (!(constbuf->enabled_mask & (1 << ubo)))
continue;
struct pipe_constant_buffer *cb = &constbuf->cb[ubo];
debug_assert((size % 16) == 0);
debug_assert((offset % 16) == 0);
- emit_const(ring, v, state->range[i].offset / 4,
- offset, size / 4, cb->user_buffer, cb->buffer);
+ if (cb->user_buffer) {
+ emit_const_user(ring, v, state->range[i].offset / 4,
+ size / 4, cb->user_buffer + state->range[i].start);
+ } else {
+ emit_const_prsc(ring, v, state->range[i].offset / 4,
+ offset, size / 4, cb->buffer);
+ }
}
}
assert(offset * 4 + params <= v->constlen * 4);
- emit_const_bo(ring, v, offset * 4, params, prscs, offsets);
+ emit_const_ptrs(ring, v, offset * 4, params, prscs, offsets);
}
}
sizes[off] = sb->sb[index].buffer_size;
}
- emit_const(ring, v, offset * 4, 0, ARRAY_SIZE(sizes), sizes, NULL);
+ emit_const_user(ring, v, offset * 4, ARRAY_SIZE(sizes), sizes);
}
}
* be the same, so use original dimensions for y and z
* stride:
*/
- dims[off + 1] = slice->pitch;
+ dims[off + 1] = fd_resource_pitch(rsc, img->u.tex.level);
/* see corresponding logic in fd_resource_offset(): */
if (rsc->layout.layer_first) {
dims[off + 2] = rsc->layout.layer_size;
}
uint32_t size = MIN2(ARRAY_SIZE(dims), v->constlen * 4 - offset * 4);
- emit_const(ring, v, offset * 4, 0, size, dims, NULL);
+ emit_const_user(ring, v, offset * 4, size, dims);
}
}
{
const struct ir3_const_state *const_state = ir3_const_state(v);
uint32_t base = const_state->offsets.immediate;
- int size = const_state->immediates_count;
+ int size = DIV_ROUND_UP(const_state->immediates_count, 4);
/* truncate size to avoid writing constants that shader
* does not use:
size *= 4;
if (size > 0)
- emit_const(ring, v, base, 0, size, const_state->immediates[0].val, NULL);
+ emit_const_user(ring, v, base, size, const_state->immediates);
}
static inline void
size *= 4;
if (size > 0)
- emit_const(ring, v, base, 0, size, patch_locs, NULL);
+ emit_const_user(ring, v, base, size, patch_locs);
}
/* emit stream-out buffers: */
assert(offset * 4 + params <= v->constlen * 4);
- emit_const_bo(ring, v, offset * 4, params, prscs, offsets);
+ emit_const_ptrs(ring, v, offset * 4, params, prscs, offsets);
}
}
const struct ir3_const_state *const_state = ir3_const_state(v);
uint32_t offset = const_state->offsets.driver_param;
uint32_t vertex_params[IR3_DP_VS_COUNT] = {
- [IR3_DP_VTXID_BASE] = info->index_size ?
+ [IR3_DP_DRAWID] = 0, /* filled by hw (CP_DRAW_INDIRECT_MULTI) */
+ [IR3_DP_VTXID_BASE] = info->index_size ?
info->index_bias : info->start,
- [IR3_DP_VTXCNT_MAX] = max_tf_vtx(ctx, v),
+ [IR3_DP_INSTID_BASE] = info->start_instance,
+ [IR3_DP_VTXCNT_MAX] = max_tf_vtx(ctx, v),
};
- /* if no user-clip-planes, we don't need to emit the
- * entire thing:
- */
- uint32_t vertex_params_size = 4;
-
if (v->key.ucp_enables) {
struct pipe_clip_state *ucp = &ctx->ucp;
unsigned pos = IR3_DP_UCP0_X;
pos++;
}
}
- vertex_params_size = ARRAY_SIZE(vertex_params);
}
- vertex_params_size = MAX2(vertex_params_size, const_state->num_driver_params);
+ /* Only emit as many params as needed, i.e. up to the highest enabled UCP
+ * plane. However a binning pass may drop even some of these, so limit to
+ * program max.
+ */
+ const uint32_t vertex_params_size = MIN2(
+ const_state->num_driver_params,
+ (v->constlen - offset) * 4);
+ assert(vertex_params_size <= IR3_DP_VS_COUNT);
bool needs_vtxid_base =
ir3_find_sysval_regid(v, SYSTEM_VALUE_VERTEX_ID_ZERO_BASE) != regid(63, 0);
ctx->screen->mem_to_mem(ring, vertex_params_rsc, 0,
indirect->buffer, src_off, 1);
- emit_const(ring, v, offset * 4, 0,
- vertex_params_size, NULL, vertex_params_rsc);
+ emit_const_prsc(ring, v, offset * 4, 0,
+ vertex_params_size, vertex_params_rsc);
pipe_resource_reference(&vertex_params_rsc, NULL);
} else {
- emit_const(ring, v, offset * 4, 0,
- vertex_params_size, vertex_params, NULL);
+ emit_const_user(ring, v, offset * 4,
+ vertex_params_size, vertex_params);
}
/* if needed, emit stream-out buffer addresses: */
indirect_offset = info->indirect_offset;
}
- emit_const(ring, v, offset * 4, indirect_offset, 4, NULL, indirect);
+ emit_const_prsc(ring, v, offset * 4, indirect_offset, 4, indirect);
pipe_resource_reference(&indirect, NULL);
} else {
uint32_t size = MIN2(const_state->num_driver_params,
v->constlen * 4 - offset * 4);
- emit_const(ring, v, offset * 4, 0, size, compute_params, NULL);
+ emit_const_user(ring, v, offset * 4, size, compute_params);
}
}
}