#include <stdint.h>
#include <stdbool.h>
+#include <string.h>
+
+#include "util/u_dynarray.h"
+
+#include "drm-uapi/i915_drm.h"
+#include "common/gen_decoder.h"
+
+#include "iris_fence.h"
/* The kernel assumes batchbuffers are smaller than 256kB. */
#define MAX_BATCH_SIZE (256 * 1024)
-/* 3DSTATE_BINDING_TABLE_POINTERS has a U16 offset from Surface State Base
- * Address, which means that we can't put binding tables beyond 64kB. This
- * effectively limits the maximum statebuffer size to 64kB.
- */
-#define MAX_STATE_SIZE (64 * 1024)
+/* Our target batch size - flush approximately at this point. */
+#define BATCH_SZ (20 * 1024)
-struct iris_address {
- struct iris_bo *bo;
- unsigned reloc_flags;
- uint64_t offset;
+enum iris_batch_name {
+ IRIS_BATCH_RENDER,
+ IRIS_BATCH_COMPUTE,
};
-struct iris_batch_buffer {
- struct iris_bo *bo;
- void *map;
- void *map_next;
+#define IRIS_BATCH_COUNT 2
- struct iris_bo *partial_bo;
- unsigned partial_bytes;
+struct iris_address {
+ struct iris_bo *bo;
+ uint64_t offset;
+ bool write;
};
struct iris_batch {
struct iris_screen *screen;
+ struct iris_vtable *vtbl;
struct pipe_debug_callback *dbg;
+ struct pipe_device_reset_callback *reset;
+
+ /** What batch is this? (e.g. IRIS_BATCH_RENDER/COMPUTE) */
+ enum iris_batch_name name;
/** Current batchbuffer being queued up. */
- struct iris_batch_buffer cmdbuf;
+ struct iris_bo *bo;
+ void *map;
+ void *map_next;
+ /** Size of the primary batch if we've moved on to a secondary. */
+ unsigned primary_batch_size;
- /** Last BO submitted to the hardware. Used for glFinish(). */
- struct iris_bo *last_cmd_bo;
+ /** Last Surface State Base Address set in this hardware context. */
+ uint64_t last_surface_base_address;
uint32_t hw_ctx_id;
- /** Which ring this batch targets - a I915_EXEC_RING_MASK value */
- uint8_t ring;
-
- bool no_wrap;
+ /** Which engine this batch targets - a I915_EXEC_RING_MASK value */
+ uint8_t engine;
/** The validation list */
struct drm_i915_gem_exec_object2 *validation_list;
int exec_count;
int exec_array_size;
+ /**
+ * A list of iris_syncpts associated with this batch.
+ *
+ * The first list entry will always be a signalling sync-point, indicating
+ * that this batch has completed. The others are likely to be sync-points
+ * to wait on before executing the batch.
+ */
+ struct util_dynarray syncpts;
+
+ /** A list of drm_i915_exec_fences to have execbuf signal or wait on */
+ struct util_dynarray exec_fences;
+
/** The amount of aperture space (in bytes) used by all exec_bos */
int aperture_space;
- /** Map from batch offset to iris_alloc_state data (with DEBUG_BATCH) */
- struct hash_table *state_sizes;
-
- void (*emit_state_base_address)(struct iris_batch *batch);
+ /** A sync-point for the last batch that was submitted. */
+ struct iris_syncpt *last_syncpt;
+
+ /** List of other batches which we might need to flush to use a BO */
+ struct iris_batch *other_batches[IRIS_BATCH_COUNT - 1];
+
+ struct {
+ /**
+ * Set of struct brw_bo * that have been rendered to within this
+ * batchbuffer and would need flushing before being used from another
+ * cache domain that isn't coherent with it (i.e. the sampler).
+ */
+ struct hash_table *render;
+
+ /**
+ * Set of struct brw_bo * that have been used as a depth buffer within
+ * this batchbuffer and would need flushing before being used from
+ * another cache domain that isn't coherent with it (i.e. the sampler).
+ */
+ struct set *depth;
+ } cache;
+
+ struct gen_batch_decode_ctx decoder;
+ struct hash_table_u64 *state_sizes;
+
+ /** Have we emitted any draw calls to this batch? */
+ bool contains_draw;
};
void iris_init_batch(struct iris_batch *batch,
struct iris_screen *screen,
+ struct iris_vtable *vtbl,
struct pipe_debug_callback *dbg,
- uint8_t ring);
+ struct pipe_device_reset_callback *reset,
+ struct hash_table_u64 *state_sizes,
+ struct iris_batch *all_batches,
+ enum iris_batch_name name,
+ uint8_t ring,
+ int priority);
+void iris_chain_to_new_batch(struct iris_batch *batch);
void iris_batch_free(struct iris_batch *batch);
-void iris_require_command_space(struct iris_batch *batch, unsigned size);
-void iris_batch_emit(struct iris_batch *batch, const void *data, unsigned size);
+void iris_batch_maybe_flush(struct iris_batch *batch, unsigned estimate);
-int _iris_batch_flush_fence(struct iris_batch *batch,
- int in_fence_fd, int *out_fence_fd,
- const char *file, int line);
+void _iris_batch_flush(struct iris_batch *batch, const char *file, int line);
+#define iris_batch_flush(batch) _iris_batch_flush((batch), __FILE__, __LINE__)
+bool iris_batch_references(struct iris_batch *batch, struct iris_bo *bo);
-#define iris_batch_flush_fence(batch, in_fence_fd, out_fence_fd) \
- _iris_batch_flush_fence((batch), (in_fence_fd), (out_fence_fd), \
- __FILE__, __LINE__)
+#define RELOC_WRITE EXEC_OBJECT_WRITE
-#define iris_batch_flush(batch) iris_batch_flush_fence((batch), -1, NULL)
+void iris_use_pinned_bo(struct iris_batch *batch, struct iris_bo *bo,
+ bool writable);
-bool iris_batch_references(struct iris_batch *batch, struct iris_bo *bo);
+enum pipe_reset_status iris_batch_check_for_reset(struct iris_batch *batch);
-#define RELOC_WRITE EXEC_OBJECT_WRITE
+static inline unsigned
+iris_batch_bytes_used(struct iris_batch *batch)
+{
+ return batch->map_next - batch->map;
+}
-void iris_use_pinned_bo(struct iris_batch *batch, struct iris_bo *bo);
+/**
+ * Ensure the current command buffer has \param size bytes of space
+ * remaining. If not, this creates a secondary batch buffer and emits
+ * a jump from the primary batch to the start of the secondary.
+ *
+ * Most callers want iris_get_command_space() instead.
+ */
+static inline void
+iris_require_command_space(struct iris_batch *batch, unsigned size)
+{
+ const unsigned required_bytes = iris_batch_bytes_used(batch) + size;
+
+ if (required_bytes >= BATCH_SZ) {
+ iris_chain_to_new_batch(batch);
+ }
+}
+
+/**
+ * Allocate space in the current command buffer, and return a pointer
+ * to the mapped area so the caller can write commands there.
+ *
+ * This should be called whenever emitting commands.
+ */
+static inline void *
+iris_get_command_space(struct iris_batch *batch, unsigned bytes)
+{
+ iris_require_command_space(batch, bytes);
+ void *map = batch->map_next;
+ batch->map_next += bytes;
+ return map;
+}
+
+/**
+ * Helper to emit GPU commands - allocates space, copies them there.
+ */
+static inline void
+iris_batch_emit(struct iris_batch *batch, const void *data, unsigned size)
+{
+ void *map = iris_get_command_space(batch, size);
+ memcpy(map, data, size);
+}
+
+/**
+ * Get a pointer to the batch's signalling syncpt. Does not refcount.
+ */
+static inline struct iris_syncpt *
+iris_batch_get_signal_syncpt(struct iris_batch *batch)
+{
+ /* The signalling syncpt is the first one in the list. */
+ struct iris_syncpt *syncpt =
+ ((struct iris_syncpt **) util_dynarray_begin(&batch->syncpts))[0];
+ return syncpt;
+}
+
+
+/**
+ * Take a reference to the batch's signalling syncpt.
+ *
+ * Callers can use this to wait for the the current batch under construction
+ * to complete (after flushing it).
+ */
+static inline void
+iris_batch_reference_signal_syncpt(struct iris_batch *batch,
+ struct iris_syncpt **out_syncpt)
+{
+ struct iris_syncpt *syncpt = iris_batch_get_signal_syncpt(batch);
+ iris_syncpt_reference(batch->screen, out_syncpt, syncpt);
+}
+
+/**
+ * Record the size of a piece of state for use in INTEL_DEBUG=bat printing.
+ */
+static inline void
+iris_record_state_size(struct hash_table_u64 *ht,
+ uint32_t offset_from_base,
+ uint32_t size)
+{
+ if (ht) {
+ _mesa_hash_table_u64_insert(ht, offset_from_base,
+ (void *)(uintptr_t) size);
+ }
+}
#endif