2 * Copyright © 2017 Intel Corporation
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
11 * The above copyright notice and this permission notice (including the next
12 * paragraph) shall be included in all copies or substantial portions of the
15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
18 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
19 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
20 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
24 #ifndef IRIS_BATCH_DOT_H
25 #define IRIS_BATCH_DOT_H
31 #include "util/u_dynarray.h"
34 #include "common/gen_decoder.h"
36 /* The kernel assumes batchbuffers are smaller than 256kB. */
37 #define MAX_BATCH_SIZE (256 * 1024)
39 /* Our target batch size - flush approximately at this point. */
40 #define BATCH_SZ (20 * 1024)
42 #define IRIS_BATCH_COUNT 2
51 struct iris_screen
*screen
;
52 struct iris_vtable
*vtbl
;
53 struct pipe_debug_callback
*dbg
;
55 /** The name of this batch for debug info (e.g. "render") */
58 /** Current batchbuffer being queued up. */
62 /** Size of the primary batch if we've moved on to a secondary. */
63 unsigned primary_batch_size
;
65 /** Last BO submitted to the hardware. Used for glFinish(). */
66 struct iris_bo
*last_bo
;
68 /** Last Surface State Base Address set in this hardware context. */
69 uint64_t last_surface_base_address
;
73 /** Which engine this batch targets - a I915_EXEC_RING_MASK value */
76 /** The validation list */
77 struct drm_i915_gem_exec_object2
*validation_list
;
78 struct iris_bo
**exec_bos
;
82 /** A list of drm_i915_exec_fences to have execbuf signal or wait on */
83 struct util_dynarray exec_fences
;
85 /** The amount of aperture space (in bytes) used by all exec_bos */
88 /** List of other batches which we might need to flush to use a BO */
89 struct iris_batch
*other_batches
[IRIS_BATCH_COUNT
- 1];
93 * Set of struct brw_bo * that have been rendered to within this
94 * batchbuffer and would need flushing before being used from another
95 * cache domain that isn't coherent with it (i.e. the sampler).
97 struct hash_table
*render
;
100 * Set of struct brw_bo * that have been used as a depth buffer within
101 * this batchbuffer and would need flushing before being used from
102 * another cache domain that isn't coherent with it (i.e. the sampler).
107 /** Map from batch offset to iris_alloc_state data (with DEBUG_BATCH) */
109 struct hash_table
*state_sizes
;
110 struct gen_batch_decode_ctx decoder
;
112 /** Have we emitted any draw calls to this batch? */
116 void iris_init_batch(struct iris_batch
*batch
,
117 struct iris_screen
*screen
,
118 struct iris_vtable
*vtbl
,
119 struct pipe_debug_callback
*dbg
,
120 struct iris_batch
**other_batches
,
123 void iris_chain_to_new_batch(struct iris_batch
*batch
);
124 void iris_batch_free(struct iris_batch
*batch
);
125 void iris_batch_maybe_flush(struct iris_batch
*batch
, unsigned estimate
);
127 void _iris_batch_flush(struct iris_batch
*batch
, const char *file
, int line
);
128 #define iris_batch_flush(batch) _iris_batch_flush((batch), __FILE__, __LINE__)
130 bool iris_batch_references(struct iris_batch
*batch
, struct iris_bo
*bo
);
132 #define RELOC_WRITE EXEC_OBJECT_WRITE
134 void iris_use_pinned_bo(struct iris_batch
*batch
, struct iris_bo
*bo
,
137 static inline unsigned
138 iris_batch_bytes_used(struct iris_batch
*batch
)
140 return batch
->map_next
- batch
->map
;
144 * Ensure the current command buffer has \param size bytes of space
145 * remaining. If not, this creates a secondary batch buffer and emits
146 * a jump from the primary batch to the start of the secondary.
148 * Most callers want iris_get_command_space() instead.
151 iris_require_command_space(struct iris_batch
*batch
, unsigned size
)
153 const unsigned required_bytes
= iris_batch_bytes_used(batch
) + size
;
155 if (required_bytes
>= BATCH_SZ
) {
156 iris_chain_to_new_batch(batch
);
161 * Allocate space in the current command buffer, and return a pointer
162 * to the mapped area so the caller can write commands there.
164 * This should be called whenever emitting commands.
167 iris_get_command_space(struct iris_batch
*batch
, unsigned bytes
)
169 iris_require_command_space(batch
, bytes
);
170 void *map
= batch
->map_next
;
171 batch
->map_next
+= bytes
;
176 * Helper to emit GPU commands - allocates space, copies them there.
179 iris_batch_emit(struct iris_batch
*batch
, const void *data
, unsigned size
)
181 void *map
= iris_get_command_space(batch
, size
);
182 memcpy(map
, data
, size
);