2 * Copyright © 2011 Intel Corporation
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
11 * The above copyright notice and this permission notice (including the next
12 * paragraph) shall be included in all copies or substantial portions of the
15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
18 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
19 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
20 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
25 * @file gen7_sol_state.c
27 * Controls the stream output logic (SOL) stage of the gen7 hardware, which is
28 * used to implement GL_EXT_transform_feedback.
31 #include "brw_context.h"
32 #include "brw_state.h"
33 #include "brw_defines.h"
34 #include "intel_batchbuffer.h"
35 #include "intel_buffer_objects.h"
38 upload_3dstate_so_buffers(struct brw_context
*brw
)
40 struct intel_context
*intel
= &brw
->intel
;
41 struct gl_context
*ctx
= &intel
->ctx
;
42 /* BRW_NEW_VERTEX_PROGRAM */
43 const struct gl_shader_program
*vs_prog
=
44 ctx
->Shader
.CurrentVertexProgram
;
45 const struct gl_transform_feedback_info
*linked_xfb_info
=
46 &vs_prog
->LinkedTransformFeedback
;
47 /* _NEW_TRANSFORM_FEEDBACK */
48 struct gl_transform_feedback_object
*xfb_obj
=
49 ctx
->TransformFeedback
.CurrentObject
;
52 /* Set up the up to 4 output buffers. These are the ranges defined in the
53 * gl_transform_feedback_object.
55 for (i
= 0; i
< 4; i
++) {
56 struct intel_buffer_object
*bufferobj
=
57 intel_buffer_object(xfb_obj
->Buffers
[i
]);
62 if (!xfb_obj
->Buffers
[i
]) {
63 /* The pitch of 0 in this command indicates that the buffer is
64 * unbound and won't be written to.
67 OUT_BATCH(_3DSTATE_SO_BUFFER
<< 16 | (4 - 2));
68 OUT_BATCH((i
<< SO_BUFFER_INDEX_SHIFT
));
76 bo
= intel_bufferobj_buffer(intel
, bufferobj
, INTEL_WRITE_PART
);
77 stride
= linked_xfb_info
->BufferStride
[i
] * 4;
79 start
= xfb_obj
->Offset
[i
];
80 assert(start
% 4 == 0);
81 end
= ALIGN(start
+ xfb_obj
->Size
[i
], 4);
82 assert(end
<= bo
->size
);
84 /* Offset the starting offset by the current vertex index into the
85 * feedback buffer, offset register is always set to 0 at the start of the
88 start
+= brw
->sol
.offset_0_batch_start
* stride
;
92 OUT_BATCH(_3DSTATE_SO_BUFFER
<< 16 | (4 - 2));
93 OUT_BATCH((i
<< SO_BUFFER_INDEX_SHIFT
) | stride
);
94 OUT_RELOC(bo
, I915_GEM_DOMAIN_RENDER
, I915_GEM_DOMAIN_RENDER
, start
);
95 OUT_RELOC(bo
, I915_GEM_DOMAIN_RENDER
, I915_GEM_DOMAIN_RENDER
, end
);
101 * Outputs the 3DSTATE_SO_DECL_LIST command.
103 * The data output is a series of 64-bit entries containing a SO_DECL per
104 * stream. We only have one stream of rendering coming out of the GS unit, so
105 * we only emit stream 0 (low 16 bits) SO_DECLs.
108 upload_3dstate_so_decl_list(struct brw_context
*brw
,
109 struct brw_vue_map
*vue_map
)
111 struct intel_context
*intel
= &brw
->intel
;
112 struct gl_context
*ctx
= &intel
->ctx
;
113 /* BRW_NEW_VERTEX_PROGRAM */
114 const struct gl_shader_program
*vs_prog
=
115 ctx
->Shader
.CurrentVertexProgram
;
116 /* _NEW_TRANSFORM_FEEDBACK */
117 const struct gl_transform_feedback_info
*linked_xfb_info
=
118 &vs_prog
->LinkedTransformFeedback
;
120 uint16_t so_decl
[128];
122 int next_offset
[4] = {0, 0, 0, 0};
124 STATIC_ASSERT(ARRAY_SIZE(so_decl
) >= MAX_PROGRAM_OUTPUTS
);
126 /* Construct the list of SO_DECLs to be emitted. The formatting of the
127 * command is feels strange -- each dword pair contains a SO_DECL per stream.
129 for (i
= 0; i
< linked_xfb_info
->NumOutputs
; i
++) {
130 int buffer
= linked_xfb_info
->Outputs
[i
].OutputBuffer
;
132 int vert_result
= linked_xfb_info
->Outputs
[i
].OutputRegister
;
133 unsigned component_mask
=
134 (1 << linked_xfb_info
->Outputs
[i
].NumComponents
) - 1;
136 /* gl_PointSize is stored in VERT_RESULT_PSIZ.w. */
137 if (vert_result
== VERT_RESULT_PSIZ
) {
138 assert(linked_xfb_info
->Outputs
[i
].NumComponents
== 1);
139 component_mask
<<= 3;
141 component_mask
<<= linked_xfb_info
->Outputs
[i
].ComponentOffset
;
144 buffer_mask
|= 1 << buffer
;
146 decl
|= buffer
<< SO_DECL_OUTPUT_BUFFER_SLOT_SHIFT
;
147 decl
|= vue_map
->vert_result_to_slot
[vert_result
] <<
148 SO_DECL_REGISTER_INDEX_SHIFT
;
149 decl
|= component_mask
<< SO_DECL_COMPONENT_MASK_SHIFT
;
151 /* This assert should be true until GL_ARB_transform_feedback_instanced
152 * is added and we start using the hole flag.
154 assert(linked_xfb_info
->Outputs
[i
].DstOffset
== next_offset
[buffer
]);
156 next_offset
[buffer
] += linked_xfb_info
->Outputs
[i
].NumComponents
;
161 BEGIN_BATCH(linked_xfb_info
->NumOutputs
* 2 + 3);
162 OUT_BATCH(_3DSTATE_SO_DECL_LIST
<< 16 |
163 (linked_xfb_info
->NumOutputs
* 2 + 1));
165 OUT_BATCH((buffer_mask
<< SO_STREAM_TO_BUFFER_SELECTS_0_SHIFT
) |
166 (0 << SO_STREAM_TO_BUFFER_SELECTS_1_SHIFT
) |
167 (0 << SO_STREAM_TO_BUFFER_SELECTS_2_SHIFT
) |
168 (0 << SO_STREAM_TO_BUFFER_SELECTS_3_SHIFT
));
170 OUT_BATCH((linked_xfb_info
->NumOutputs
<< SO_NUM_ENTRIES_0_SHIFT
) |
171 (0 << SO_NUM_ENTRIES_1_SHIFT
) |
172 (0 << SO_NUM_ENTRIES_2_SHIFT
) |
173 (0 << SO_NUM_ENTRIES_3_SHIFT
));
175 for (i
= 0; i
< linked_xfb_info
->NumOutputs
; i
++) {
176 OUT_BATCH(so_decl
[i
]);
184 upload_3dstate_streamout(struct brw_context
*brw
, bool active
,
185 struct brw_vue_map
*vue_map
)
187 struct intel_context
*intel
= &brw
->intel
;
188 struct gl_context
*ctx
= &intel
->ctx
;
189 /* _NEW_TRANSFORM_FEEDBACK */
190 struct gl_transform_feedback_object
*xfb_obj
=
191 ctx
->TransformFeedback
.CurrentObject
;
192 uint32_t dw1
= 0, dw2
= 0;
195 /* _NEW_RASTERIZER_DISCARD */
196 if (ctx
->RasterDiscard
)
197 dw1
|= SO_RENDERING_DISABLE
;
200 int urb_entry_read_offset
= 0;
201 int urb_entry_read_length
= (vue_map
->num_slots
+ 1) / 2 -
202 urb_entry_read_offset
;
204 dw1
|= SO_FUNCTION_ENABLE
;
205 dw1
|= SO_STATISTICS_ENABLE
;
208 if (ctx
->Light
.ProvokingVertex
!= GL_FIRST_VERTEX_CONVENTION
)
209 dw1
|= SO_REORDER_TRAILING
;
211 for (i
= 0; i
< 4; i
++) {
212 if (xfb_obj
->Buffers
[i
]) {
213 dw1
|= SO_BUFFER_ENABLE(i
);
217 /* We always read the whole vertex. This could be reduced at some
218 * point by reading less and offsetting the register index in the
221 dw2
|= urb_entry_read_offset
<< SO_STREAM_0_VERTEX_READ_OFFSET_SHIFT
;
222 dw2
|= (urb_entry_read_length
- 1) <<
223 SO_STREAM_0_VERTEX_READ_LENGTH_SHIFT
;
227 OUT_BATCH(_3DSTATE_STREAMOUT
<< 16 | (3 - 2));
234 upload_sol_state(struct brw_context
*brw
)
236 struct intel_context
*intel
= &brw
->intel
;
237 struct gl_context
*ctx
= &intel
->ctx
;
238 /* _NEW_TRANSFORM_FEEDBACK */
239 struct gl_transform_feedback_object
*xfb_obj
=
240 ctx
->TransformFeedback
.CurrentObject
;
241 bool active
= xfb_obj
->Active
&& !xfb_obj
->Paused
;
242 struct brw_vue_map vue_map
;
244 /* CACHE_NEW_VS_PROG */
245 brw_compute_vue_map(&vue_map
, intel
, brw
->vs
.prog_data
);
248 upload_3dstate_so_buffers(brw
);
249 upload_3dstate_so_decl_list(brw
, &vue_map
);
251 intel
->batch
.needs_sol_reset
= true;
254 /* Finally, set up the SOL stage. This command must always follow updates to
255 * the nonpipelined SOL state (3DSTATE_SO_BUFFER, 3DSTATE_SO_DECL_LIST) or
256 * MMIO register updates (current performed by the kernel at each batch
259 upload_3dstate_streamout(brw
, active
, &vue_map
);
262 const struct brw_tracked_state gen7_sol_state
= {
264 .mesa
= (_NEW_RASTERIZER_DISCARD
|
266 _NEW_TRANSFORM_FEEDBACK
),
267 .brw
= (BRW_NEW_BATCH
|
268 BRW_NEW_VERTEX_PROGRAM
),
269 .cache
= CACHE_NEW_VS_PROG
,
271 .emit
= upload_sol_state
,
275 gen7_end_transform_feedback(struct gl_context
*ctx
,
276 struct gl_transform_feedback_object
*obj
)
278 /* Because we have to rely on the kernel to reset our SO write offsets, and
279 * we only get to do it once per batchbuffer, flush the batch after feedback
280 * so another transform feedback can get the write offset reset it needs.
282 * This also covers any cache flushing required.
284 struct brw_context
*brw
= brw_context(ctx
);
285 struct intel_context
*intel
= &brw
->intel
;
287 intel_batchbuffer_flush(intel
);