2 Copyright (C) The Weather Channel, Inc. 2002. All Rights Reserved.
4 The Weather Channel (TM) funded Tungsten Graphics to develop the
5 initial release of the Radeon 8500 driver under the XFree86 license.
6 This notice must be preserved.
8 Permission is hereby granted, free of charge, to any person obtaining
9 a copy of this software and associated documentation files (the
10 "Software"), to deal in the Software without restriction, including
11 without limitation the rights to use, copy, modify, merge, publish,
12 distribute, sublicense, and/or sell copies of the Software, and to
13 permit persons to whom the Software is furnished to do so, subject to
14 the following conditions:
16 The above copyright notice and this permission notice (including the
17 next paragraph) shall be included in all copies or substantial
18 portions of the Software.
20 THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
21 EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
22 MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.
23 IN NO EVENT SHALL THE COPYRIGHT OWNER(S) AND/OR ITS SUPPLIERS BE
24 LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
25 OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
26 WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
28 **************************************************************************/
32 * Nicolai Haehnle <prefect_@gmx.net>
40 #include "swrast/swrast.h"
41 #include "simple_list.h"
44 #include "radeon_drm.h"
46 #include "radeon_ioctl.h"
47 #include "r300_context.h"
48 #include "r300_ioctl.h"
49 #include "radeon_reg.h"
51 #include "r300_cmdbuf.h"
52 #include "r300_emit.h"
55 // Set this to 1 for extremely verbose debugging of command buffers
56 #define DEBUG_CMDBUF 0
60 * Send the current command buffer via ioctl to the hardware.
62 int r300FlushCmdBufLocked(r300ContextPtr r300
, const char* caller
)
66 drm_radeon_cmd_buffer_t cmd
;
69 if (r300
->radeon
.lost_context
)
72 start
= r300
->cmdbuf
.count_reemit
;
74 if (RADEON_DEBUG
& DEBUG_IOCTL
) {
75 fprintf(stderr
, "%s from %s - %i cliprects\n",
76 __FUNCTION__
, caller
, r300
->radeon
.numClipRects
);
78 if (DEBUG_CMDBUF
&& RADEON_DEBUG
& DEBUG_VERBOSE
)
79 for (i
= start
; i
< r300
->cmdbuf
.count_used
; ++i
)
80 fprintf(stderr
, "%d: %08x\n", i
,
81 r300
->cmdbuf
.cmd_buf
[i
]);
84 cmd
.buf
= (char*)(r300
->cmdbuf
.cmd_buf
+ start
);
85 cmd
.bufsz
= (r300
->cmdbuf
.count_used
- start
) * 4;
87 if (r300
->radeon
.state
.scissor
.enabled
) {
88 cmd
.nbox
= r300
->radeon
.state
.scissor
.numClipRects
;
89 cmd
.boxes
= (drm_clip_rect_t
*)r300
->radeon
.state
.scissor
.pClipRects
;
91 cmd
.nbox
= r300
->radeon
.numClipRects
;
92 cmd
.boxes
= (drm_clip_rect_t
*)r300
->radeon
.pClipRects
;
96 ret
= drmCommandWrite(r300
->radeon
.dri
.fd
,
97 DRM_RADEON_CMDBUF
, &cmd
, sizeof(cmd
));
99 if (RADEON_DEBUG
& DEBUG_SYNC
) {
100 fprintf(stderr
, "Syncing in %s\n\n", __FUNCTION__
);
101 radeonWaitForIdleLocked(&r300
->radeon
);
105 if (RADEON_DEBUG
& DEBUG_IOCTL
)
106 fprintf(stderr
, "%s: No cliprects\n", __FUNCTION__
);
109 r300
->cmdbuf
.count_used
= 0;
110 r300
->cmdbuf
.count_reemit
= 0;
116 int r300FlushCmdBuf(r300ContextPtr r300
, const char* caller
)
120 drm_radeon_cmd_buffer_t cmd
;
123 LOCK_HARDWARE(&r300
->radeon
);
125 ret
=r300FlushCmdBufLocked(r300
, caller
);
127 UNLOCK_HARDWARE(&r300
->radeon
);
130 fprintf(stderr
, "drmRadeonCmdBuffer: %d (exiting)\n", ret
);
138 static void print_state_atom(struct r300_state_atom
*state
, int dwords
)
142 fprintf(stderr
, " emit %s/%d/%d\n", state
->name
, dwords
, state
->cmd_size
);
144 if (RADEON_DEBUG
& DEBUG_VERBOSE
)
145 for (i
= 0; i
< dwords
; i
++)
146 fprintf(stderr
, " %s[%d]: %08X\n", state
->name
, i
,
151 * Emit all atoms with a dirty field equal to dirty.
153 * The caller must have ensured that there is enough space in the command
156 static __inline__
void r300DoEmitState(r300ContextPtr r300
, GLboolean dirty
)
158 struct r300_state_atom
* atom
;
161 dest
= r300
->cmdbuf
.cmd_buf
+ r300
->cmdbuf
.count_used
;
163 if (DEBUG_CMDBUF
&& RADEON_DEBUG
& DEBUG_STATE
) {
164 foreach(atom
, &r300
->hw
.atomlist
) {
165 if ((atom
->dirty
|| r300
->hw
.all_dirty
) == dirty
) {
166 int dwords
= (*atom
->check
)(r300
, atom
);
169 print_state_atom(atom
, dwords
);
171 fprintf(stderr
, " skip state %s\n",
177 foreach(atom
, &r300
->hw
.atomlist
) {
178 if ((atom
->dirty
|| r300
->hw
.all_dirty
) == dirty
) {
179 int dwords
= (*atom
->check
)(r300
, atom
);
182 memcpy(dest
, atom
->cmd
, dwords
*4);
184 r300
->cmdbuf
.count_used
+= dwords
;
185 atom
->dirty
= GL_FALSE
;
193 * Copy dirty hardware state atoms into the command buffer.
195 * We also copy out clean state if we're at the start of a buffer. That makes
196 * it easy to recover from lost contexts.
198 void r300EmitState(r300ContextPtr r300
)
200 if (RADEON_DEBUG
& (DEBUG_STATE
| DEBUG_PRIMS
))
201 fprintf(stderr
, "%s\n", __FUNCTION__
);
203 if (r300
->cmdbuf
.count_used
&& !r300
->hw
.is_dirty
&& !r300
->hw
.all_dirty
)
206 /* To avoid going across the entire set of states multiple times, just check
207 * for enough space for the case of emitting all state, and inline the
208 * r300AllocCmdBuf code here without all the checks.
210 r300EnsureCmdBufSpace(r300
, r300
->hw
.max_state_size
, __FUNCTION__
);
212 if (!r300
->cmdbuf
.count_used
) {
213 if (RADEON_DEBUG
& DEBUG_STATE
)
214 fprintf(stderr
, "Begin reemit state\n");
216 r300DoEmitState(r300
, GL_FALSE
);
217 r300
->cmdbuf
.count_reemit
= r300
->cmdbuf
.count_used
;
220 if (RADEON_DEBUG
& DEBUG_STATE
)
221 fprintf(stderr
, "Begin dirty state\n");
223 r300DoEmitState(r300
, GL_TRUE
);
225 assert(r300
->cmdbuf
.count_used
< r300
->cmdbuf
.size
);
227 r300
->hw
.is_dirty
= GL_FALSE
;
228 r300
->hw
.all_dirty
= GL_FALSE
;
233 static __inline__
uint32_t cmducs(int reg
, int count
)
235 drm_r300_cmd_header_t cmd
;
237 cmd
.unchecked_state
.cmd_type
= R300_CMD_UNCHECKED_STATE
;
238 cmd
.unchecked_state
.count
= count
;
239 cmd
.unchecked_state
.reghi
= ((unsigned int)reg
& 0xFF00) >> 8;
240 cmd
.unchecked_state
.reglo
= ((unsigned int)reg
& 0x00FF);
245 static __inline__
uint32_t cmdvpu(int addr
, int count
)
247 drm_r300_cmd_header_t cmd
;
249 cmd
.vpu
.cmd_type
= R300_CMD_VPU
;
250 cmd
.vpu
.count
= count
;
251 cmd
.vpu
.adrhi
= ((unsigned int)addr
& 0xFF00) >> 8;
252 cmd
.vpu
.adrlo
= ((unsigned int)addr
& 0x00FF);
258 #define CHECK( NM, COUNT ) \
259 static int check_##NM( r300ContextPtr r300, \
260 struct r300_state_atom* atom ) \
262 (void) atom; (void) r300; \
266 #define ucscount(ptr) (((drm_r300_cmd_header_t*)(ptr))->unchecked_state.count)
267 #define vpucount(ptr) (((drm_r300_cmd_header_t*)(ptr))->vpu.count)
269 CHECK( always
, atom
->cmd_size
)
271 CHECK( variable
, ucscount(atom
->cmd
) ? (1 + ucscount(atom
->cmd
)) : 0 )
272 CHECK( vpu
, vpucount(atom
->cmd
) ? (1 + vpucount(atom
->cmd
)*4) : 0 )
276 #define ALLOC_STATE( ATOM, CHK, SZ, NM, IDX ) \
278 r300->hw.ATOM.cmd_size = SZ; \
279 r300->hw.ATOM.cmd = (uint32_t*)CALLOC(SZ * sizeof(uint32_t)); \
280 r300->hw.ATOM.name = NM; \
281 r300->hw.ATOM.idx = IDX; \
282 r300->hw.ATOM.check = check_##CHK; \
283 r300->hw.ATOM.dirty = GL_FALSE; \
284 r300->hw.max_state_size += SZ; \
289 * Allocate memory for the command buffer and initialize the state atom
290 * list. Note that the initial hardware state is set by r300InitState().
292 void r300InitCmdBuf(r300ContextPtr r300
)
296 r300
->hw
.max_state_size
= 0;
298 mtu
= r300
->radeon
.glCtx
->Const
.MaxTextureUnits
;
299 fprintf(stderr
, "Using %d maximum texture units..\n", mtu
);
301 /* Initialize state atoms */
302 ALLOC_STATE( vpt
, always
, R300_VPT_CMDSIZE
, "vpt", 0 );
303 r300
->hw
.vpt
.cmd
[R300_VPT_CMD_0
] = cmducs(R300_SE_VPORT_XSCALE
, 6);
304 ALLOC_STATE( unk2080
, always
, 2, "unk2080", 0 );
305 r300
->hw
.unk2080
.cmd
[0] = cmducs(0x2080, 1);
306 ALLOC_STATE( vte
, always
, 3, "vte", 0 );
307 r300
->hw
.vte
.cmd
[0] = cmducs(R300_SE_VTE_CNTL
, 2);
308 ALLOC_STATE( unk2134
, always
, 3, "unk2134", 0 );
309 r300
->hw
.unk2134
.cmd
[0] = cmducs(0x2134, 2);
310 ALLOC_STATE( unk2140
, always
, 2, "unk2140", 0 );
311 r300
->hw
.unk2140
.cmd
[0] = cmducs(0x2140, 1);
312 ALLOC_STATE( vir
[0], variable
, R300_VIR_CMDSIZE
, "vir/0", 0 );
313 r300
->hw
.vir
[0].cmd
[R300_VIR_CMD_0
] = cmducs(R300_VAP_INPUT_ROUTE_0_0
, 1);
314 ALLOC_STATE( vir
[1], variable
, R300_VIR_CMDSIZE
, "vir/1", 1 );
315 r300
->hw
.vir
[1].cmd
[R300_VIR_CMD_0
] = cmducs(R300_VAP_INPUT_ROUTE_1_0
, 1);
316 ALLOC_STATE( vic
, always
, R300_VIC_CMDSIZE
, "vic", 0 );
317 r300
->hw
.vic
.cmd
[R300_VIC_CMD_0
] = cmducs(R300_VAP_INPUT_CNTL_0
, 2);
318 ALLOC_STATE( unk21DC
, always
, 2, "unk21DC", 0 );
319 r300
->hw
.unk21DC
.cmd
[0] = cmducs(0x21DC, 1);
320 ALLOC_STATE( unk221C
, always
, 2, "unk221C", 0 );
321 r300
->hw
.unk221C
.cmd
[0] = cmducs(0x221C, 1);
322 ALLOC_STATE( unk2220
, always
, 5, "unk2220", 0 );
323 r300
->hw
.unk2220
.cmd
[0] = cmducs(0x2220, 4);
324 ALLOC_STATE( unk2288
, always
, 2, "unk2288", 0 );
325 r300
->hw
.unk2288
.cmd
[0] = cmducs(0x2288, 1);
326 ALLOC_STATE( vof
, always
, R300_VOF_CMDSIZE
, "vof", 0 );
327 r300
->hw
.vof
.cmd
[R300_VOF_CMD_0
] = cmducs(R300_VAP_OUTPUT_VTX_FMT_0
, 2);
328 ALLOC_STATE( pvs
, always
, R300_PVS_CMDSIZE
, "pvs", 0 );
329 r300
->hw
.pvs
.cmd
[R300_PVS_CMD_0
] = cmducs(R300_VAP_PVS_CNTL_1
, 3);
330 ALLOC_STATE( gb_enable
, always
, 2, "gb_enable", 0 );
331 r300
->hw
.gb_enable
.cmd
[0] = cmducs(R300_GB_ENABLE
, 1);
332 ALLOC_STATE( gb_misc
, always
, R300_GB_MISC_CMDSIZE
, "gb_misc", 0 );
333 r300
->hw
.gb_misc
.cmd
[0] = cmducs(R300_GB_MSPOS0
, 5);
334 ALLOC_STATE( txe
, always
, R300_TXE_CMDSIZE
, "txe", 0 );
335 r300
->hw
.txe
.cmd
[R300_TXE_CMD_0
] = cmducs(R300_TX_ENABLE
, 1);
336 ALLOC_STATE( unk4200
, always
, 5, "unk4200", 0 );
337 r300
->hw
.unk4200
.cmd
[0] = cmducs(0x4200, 4);
338 ALLOC_STATE( unk4214
, always
, 2, "unk4214", 0 );
339 r300
->hw
.unk4214
.cmd
[0] = cmducs(0x4214, 1);
340 ALLOC_STATE( ps
, always
, R300_PS_CMDSIZE
, "ps", 0 );
341 r300
->hw
.ps
.cmd
[0] = cmducs(R300_RE_POINTSIZE
, 1);
342 ALLOC_STATE( unk4230
, always
, 4, "unk4230", 0 );
343 r300
->hw
.unk4230
.cmd
[0] = cmducs(0x4230, 3);
344 ALLOC_STATE( lcntl
, always
, 2, "lcntl", 0 );
345 r300
->hw
.lcntl
.cmd
[0] = cmducs(R300_RE_LINE_CNT
, 1);
347 ALLOC_STATE( lsf
, always
, 2, "lsf", 0 );
348 r300
->hw
.lsf
.cmd
[0] = cmducs(R300_RE_LINE_STIPPLE_FACTOR
, 1);
350 ALLOC_STATE( unk4260
, always
, 4, "unk4260", 0 );
351 r300
->hw
.unk4260
.cmd
[0] = cmducs(0x4260, 3);
352 ALLOC_STATE( unk4274
, always
, 5, "unk4274", 0 );
353 r300
->hw
.unk4274
.cmd
[0] = cmducs(0x4274, 4);
354 ALLOC_STATE( unk4288
, always
, 6, "unk4288", 0 );
355 r300
->hw
.unk4288
.cmd
[0] = cmducs(0x4288, 5);
356 ALLOC_STATE( unk42A0
, always
, 2, "unk42A0", 0 );
357 r300
->hw
.unk42A0
.cmd
[0] = cmducs(0x42A0, 1);
358 ALLOC_STATE( zbs
, always
, R300_ZBS_CMDSIZE
, "zbs", 0 );
359 r300
->hw
.zbs
.cmd
[R300_ZBS_CMD_0
] = cmducs(R300_RE_ZBIAS_T_FACTOR
, 4);
360 ALLOC_STATE( unk42B4
, always
, 2, "unk42B4", 0 );
361 r300
->hw
.unk42B4
.cmd
[0] = cmducs(0x42B4, 1);
362 ALLOC_STATE( cul
, always
, R300_CUL_CMDSIZE
, "cul", 0 );
363 r300
->hw
.cul
.cmd
[R300_CUL_CMD_0
] = cmducs(R300_RE_CULL_CNTL
, 1);
364 ALLOC_STATE( unk42C0
, always
, 3, "unk42C0", 0 );
365 r300
->hw
.unk42C0
.cmd
[0] = cmducs(0x42C0, 2);
366 ALLOC_STATE( rc
, always
, R300_RC_CMDSIZE
, "rc", 0 );
367 r300
->hw
.rc
.cmd
[R300_RC_CMD_0
] = cmducs(R300_RS_CNTL_0
, 2);
368 ALLOC_STATE( ri
, always
, R300_RI_CMDSIZE
, "ri", 0 );
369 r300
->hw
.ri
.cmd
[R300_RI_CMD_0
] = cmducs(R300_RS_INTERP_0
, 8);
370 ALLOC_STATE( rr
, variable
, R300_RR_CMDSIZE
, "rr", 0 );
371 r300
->hw
.rr
.cmd
[R300_RR_CMD_0
] = cmducs(R300_RS_ROUTE_0
, 1);
372 ALLOC_STATE( unk43A4
, always
, 3, "unk43A4", 0 );
373 r300
->hw
.unk43A4
.cmd
[0] = cmducs(0x43A4, 2);
374 ALLOC_STATE( unk43E8
, always
, 2, "unk43E8", 0 );
375 r300
->hw
.unk43E8
.cmd
[0] = cmducs(0x43E8, 1);
376 ALLOC_STATE( fp
, always
, R300_FP_CMDSIZE
, "fp", 0 );
377 r300
->hw
.fp
.cmd
[R300_FP_CMD_0
] = cmducs(R300_PFS_CNTL_0
, 3);
378 r300
->hw
.fp
.cmd
[R300_FP_CMD_1
] = cmducs(R300_PFS_NODE_0
, 4);
379 ALLOC_STATE( fpt
, variable
, R300_FPT_CMDSIZE
, "fpt", 0 );
380 r300
->hw
.fpt
.cmd
[R300_FPT_CMD_0
] = cmducs(R300_PFS_TEXI_0
, 0);
381 ALLOC_STATE( unk46A4
, always
, 6, "unk46A4", 0 );
382 r300
->hw
.unk46A4
.cmd
[0] = cmducs(0x46A4, 5);
383 ALLOC_STATE( fpi
[0], variable
, R300_FPI_CMDSIZE
, "fpi/0", 0 );
384 r300
->hw
.fpi
[0].cmd
[R300_FPI_CMD_0
] = cmducs(R300_PFS_INSTR0_0
, 1);
385 ALLOC_STATE( fpi
[1], variable
, R300_FPI_CMDSIZE
, "fpi/1", 1 );
386 r300
->hw
.fpi
[1].cmd
[R300_FPI_CMD_0
] = cmducs(R300_PFS_INSTR1_0
, 1);
387 ALLOC_STATE( fpi
[2], variable
, R300_FPI_CMDSIZE
, "fpi/2", 2 );
388 r300
->hw
.fpi
[2].cmd
[R300_FPI_CMD_0
] = cmducs(R300_PFS_INSTR2_0
, 1);
389 ALLOC_STATE( fpi
[3], variable
, R300_FPI_CMDSIZE
, "fpi/3", 3 );
390 r300
->hw
.fpi
[3].cmd
[R300_FPI_CMD_0
] = cmducs(R300_PFS_INSTR3_0
, 1);
391 ALLOC_STATE( unk4BC0
, always
, 2, "unk4BC0", 0 );
392 r300
->hw
.unk4BC0
.cmd
[0] = cmducs(0x4BC0, 1);
393 ALLOC_STATE( unk4BC8
, always
, 4, "unk4BC8", 0 );
394 r300
->hw
.unk4BC8
.cmd
[0] = cmducs(0x4BC8, 3);
395 ALLOC_STATE( at
, always
, R300_AT_CMDSIZE
, "at", 0 );
396 r300
->hw
.at
.cmd
[R300_AT_CMD_0
] = cmducs(R300_PP_ALPHA_TEST
, 2);
397 ALLOC_STATE( unk4BD8
, always
, 2, "unk4BD8", 0 );
398 r300
->hw
.unk4BD8
.cmd
[0] = cmducs(0x4BD8, 1);
399 ALLOC_STATE( fpp
, variable
, R300_FPP_CMDSIZE
, "fpp", 0 );
400 r300
->hw
.fpp
.cmd
[R300_FPP_CMD_0
] = cmducs(R300_PFS_PARAM_0_X
, 0);
401 ALLOC_STATE( unk4E00
, always
, 2, "unk4E00", 0 );
402 r300
->hw
.unk4E00
.cmd
[0] = cmducs(0x4E00, 1);
403 ALLOC_STATE( bld
, always
, R300_BLD_CMDSIZE
, "bld", 0 );
404 r300
->hw
.bld
.cmd
[R300_BLD_CMD_0
] = cmducs(R300_RB3D_CBLEND
, 2);
405 ALLOC_STATE( cmk
, always
, R300_CMK_CMDSIZE
, "cmk", 0 );
406 r300
->hw
.cmk
.cmd
[R300_CMK_CMD_0
] = cmducs(R300_RB3D_COLORMASK
, 1);
407 ALLOC_STATE( unk4E10
, always
, 4, "unk4E10", 0 );
408 r300
->hw
.unk4E10
.cmd
[0] = cmducs(0x4E10, 3);
409 ALLOC_STATE( cb
, always
, R300_CB_CMDSIZE
, "cb", 0 );
410 r300
->hw
.cb
.cmd
[R300_CB_CMD_0
] = cmducs(R300_RB3D_COLOROFFSET0
, 1);
411 r300
->hw
.cb
.cmd
[R300_CB_CMD_1
] = cmducs(R300_RB3D_COLORPITCH0
, 1);
412 ALLOC_STATE( unk4E50
, always
, 10, "unk4E50", 0 );
413 r300
->hw
.unk4E50
.cmd
[0] = cmducs(0x4E50, 9);
414 ALLOC_STATE( unk4E88
, always
, 2, "unk4E88", 0 );
415 r300
->hw
.unk4E88
.cmd
[0] = cmducs(0x4E88, 1);
416 ALLOC_STATE( unk4EA0
, always
, 3, "unk4EA0 R350 only", 0 );
417 r300
->hw
.unk4EA0
.cmd
[0] = cmducs(0x4EA0, 2);
418 ALLOC_STATE( zs
, always
, R300_ZS_CMDSIZE
, "zstencil", 0 );
419 r300
->hw
.zs
.cmd
[R300_ZS_CMD_0
] = cmducs(R300_RB3D_ZSTENCIL_CNTL_0
, 3);
420 ALLOC_STATE( unk4F10
, always
, 5, "unk4F10", 0 );
421 r300
->hw
.unk4F10
.cmd
[0] = cmducs(0x4F10, 4);
422 ALLOC_STATE( zb
, always
, R300_ZB_CMDSIZE
, "zb", 0 );
423 r300
->hw
.zb
.cmd
[R300_ZB_CMD_0
] = cmducs(R300_RB3D_DEPTHOFFSET
, 2);
424 ALLOC_STATE( unk4F28
, always
, 2, "unk4F28", 0 );
425 r300
->hw
.unk4F28
.cmd
[0] = cmducs(0x4F28, 1);
426 ALLOC_STATE( unk4F30
, always
, 3, "unk4F30", 0 );
427 r300
->hw
.unk4F30
.cmd
[0] = cmducs(0x4F30, 2);
428 ALLOC_STATE( unk4F44
, always
, 2, "unk4F44", 0 );
429 r300
->hw
.unk4F44
.cmd
[0] = cmducs(0x4F44, 1);
430 ALLOC_STATE( unk4F54
, always
, 2, "unk4F54", 0 );
431 r300
->hw
.unk4F54
.cmd
[0] = cmducs(0x4F54, 1);
433 ALLOC_STATE( vpi
, vpu
, R300_VPI_CMDSIZE
, "vpi", 0 );
434 r300
->hw
.vpi
.cmd
[R300_VPI_CMD_0
] = cmdvpu(R300_PVS_UPLOAD_PROGRAM
, 0);
435 ALLOC_STATE( vpp
, vpu
, R300_VPP_CMDSIZE
, "vpp", 0 );
436 r300
->hw
.vpp
.cmd
[R300_VPP_CMD_0
] = cmdvpu(R300_PVS_UPLOAD_PARAMETERS
, 0);
437 ALLOC_STATE( vps
, vpu
, R300_VPS_CMDSIZE
, "vps", 0 );
438 r300
->hw
.vps
.cmd
[R300_VPS_CMD_0
] = cmdvpu(R300_PVS_UPLOAD_POINTSIZE
, 1);
441 ALLOC_STATE( tex
.filter
, variable
, mtu
+1, "tex_filter", 0 );
442 r300
->hw
.tex
.filter
.cmd
[R300_TEX_CMD_0
] = cmducs(R300_TX_FILTER_0
, 0);
444 ALLOC_STATE( tex
.unknown1
, variable
, mtu
+1, "tex_unknown1", 0 );
445 r300
->hw
.tex
.unknown1
.cmd
[R300_TEX_CMD_0
] = cmducs(R300_TX_UNK1_0
, 0);
447 ALLOC_STATE( tex
.size
, variable
, mtu
+1, "tex_size", 0 );
448 r300
->hw
.tex
.size
.cmd
[R300_TEX_CMD_0
] = cmducs(R300_TX_SIZE_0
, 0);
450 ALLOC_STATE( tex
.format
, variable
, mtu
+1, "tex_format", 0 );
451 r300
->hw
.tex
.format
.cmd
[R300_TEX_CMD_0
] = cmducs(R300_TX_FORMAT_0
, 0);
453 ALLOC_STATE( tex
.offset
, variable
, mtu
+1, "tex_offset", 0 );
454 r300
->hw
.tex
.offset
.cmd
[R300_TEX_CMD_0
] = cmducs(R300_TX_OFFSET_0
, 0);
456 ALLOC_STATE( tex
.unknown4
, variable
, mtu
+1, "tex_unknown4", 0 );
457 r300
->hw
.tex
.unknown4
.cmd
[R300_TEX_CMD_0
] = cmducs(R300_TX_UNK4_0
, 0);
459 ALLOC_STATE( tex
.border_color
, variable
, mtu
+1, "tex_border_color", 0 );
460 r300
->hw
.tex
.border_color
.cmd
[R300_TEX_CMD_0
] = cmducs(R300_TX_BORDER_COLOR_0
, 0);
463 /* Setup the atom linked list */
464 make_empty_list(&r300
->hw
.atomlist
);
465 r300
->hw
.atomlist
.name
= "atom-list";
467 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.vpt
);
468 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.unk2080
);
469 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.vte
);
470 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.unk2134
);
471 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.unk2140
);
472 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.vir
[0]);
473 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.vir
[1]);
474 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.vic
);
475 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.unk21DC
);
476 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.unk221C
);
477 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.unk2220
);
478 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.unk2288
);
479 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.vof
);
480 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.pvs
);
481 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.gb_enable
);
482 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.gb_misc
);
483 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.txe
);
484 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.unk4200
);
485 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.unk4214
);
486 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.ps
);
487 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.unk4230
);
488 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.lcntl
);
490 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.lsf
);
492 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.unk4260
);
493 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.unk4274
);
494 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.unk4288
);
495 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.unk42A0
);
496 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.zbs
);
497 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.unk42B4
);
498 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.cul
);
499 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.unk42C0
);
500 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.rc
);
501 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.ri
);
502 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.rr
);
503 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.unk43A4
);
504 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.unk43E8
);
505 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.fp
);
506 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.fpt
);
507 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.unk46A4
);
508 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.fpi
[0]);
509 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.fpi
[1]);
510 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.fpi
[2]);
511 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.fpi
[3]);
512 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.unk4BC0
);
513 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.unk4BC8
);
514 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.at
);
515 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.unk4BD8
);
516 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.fpp
);
517 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.unk4E00
);
518 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.bld
);
519 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.cmk
);
520 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.unk4E10
);
521 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.cb
);
522 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.unk4E50
);
523 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.unk4E88
);
524 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.unk4EA0
);
525 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.zs
);
526 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.unk4F10
);
527 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.zb
);
528 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.unk4F28
);
529 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.unk4F30
);
530 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.unk4F44
);
531 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.unk4F54
);
533 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.vpi
);
534 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.vpp
);
535 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.vps
);
537 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.tex
.filter
);
538 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.tex
.unknown1
);
539 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.tex
.size
);
540 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.tex
.format
);
541 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.tex
.offset
);
542 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.tex
.unknown4
);
543 insert_at_tail(&r300
->hw
.atomlist
, &r300
->hw
.tex
.border_color
);
545 r300
->hw
.is_dirty
= GL_TRUE
;
546 r300
->hw
.all_dirty
= GL_TRUE
;
548 /* Initialize command buffer */
549 size
= 256 * driQueryOptioni(&r300
->radeon
.optionCache
, "command_buffer_size");
550 if (size
< 2*r300
->hw
.max_state_size
){
551 size
= 2*r300
->hw
.max_state_size
+65535;
554 if (1 || RADEON_DEBUG
& DEBUG_IOCTL
){
555 fprintf(stderr
, "sizeof(drm_r300_cmd_header_t)=%d\n",
556 sizeof(drm_r300_cmd_header_t
));
557 fprintf(stderr
, "sizeof(drm_radeon_cmd_buffer_t)=%d\n",
558 sizeof(drm_radeon_cmd_buffer_t
));
560 "Allocating %d bytes command buffer (max state is %d bytes)\n",
561 size
*4, r300
->hw
.max_state_size
*4);
564 r300
->cmdbuf
.size
= size
;
565 r300
->cmdbuf
.cmd_buf
= (uint32_t*)CALLOC(size
*4);
566 r300
->cmdbuf
.count_used
= 0;
567 r300
->cmdbuf
.count_reemit
= 0;
572 * Destroy the command buffer and state atoms.
574 void r300DestroyCmdBuf(r300ContextPtr r300
)
576 struct r300_state_atom
* atom
;
578 FREE(r300
->cmdbuf
.cmd_buf
);
580 foreach(atom
, &r300
->hw
.atomlist
) {
585 void r300EmitBlit(r300ContextPtr rmesa
,
591 GLint srcx
, GLint srcy
,
592 GLint dstx
, GLint dsty
, GLuint w
, GLuint h
)
594 drm_radeon_cmd_header_t
*cmd
;
596 if (RADEON_DEBUG
& DEBUG_IOCTL
)
598 "%s src %x/%x %d,%d dst: %x/%x %d,%d sz: %dx%d\n",
599 __FUNCTION__
, src_pitch
, src_offset
, srcx
, srcy
,
600 dst_pitch
, dst_offset
, dstx
, dsty
, w
, h
);
602 assert((src_pitch
& 63) == 0);
603 assert((dst_pitch
& 63) == 0);
604 assert((src_offset
& 1023) == 0);
605 assert((dst_offset
& 1023) == 0);
606 assert(w
< (1 << 16));
607 assert(h
< (1 << 16));
610 (drm_radeon_cmd_header_t
*) r300AllocCmdBuf(rmesa
, 8 * sizeof(int),
613 cmd
[0].header
.cmd_type
= RADEON_CMD_PACKET3
;
614 cmd
[1].i
= R200_CP_CMD_BITBLT_MULTI
| (5 << 16);
615 cmd
[2].i
= (RADEON_GMC_SRC_PITCH_OFFSET_CNTL
|
616 RADEON_GMC_DST_PITCH_OFFSET_CNTL
|
617 RADEON_GMC_BRUSH_NONE
|
619 RADEON_GMC_SRC_DATATYPE_COLOR
|
621 RADEON_DP_SRC_SOURCE_MEMORY
|
622 RADEON_GMC_CLR_CMP_CNTL_DIS
| RADEON_GMC_WR_MSK_DIS
);
624 cmd
[3].i
= ((src_pitch
/ 64) << 22) | (src_offset
>> 10);
625 cmd
[4].i
= ((dst_pitch
/ 64) << 22) | (dst_offset
>> 10);
626 cmd
[5].i
= (srcx
<< 16) | srcy
;
627 cmd
[6].i
= (dstx
<< 16) | dsty
; /* dst */
628 cmd
[7].i
= (w
<< 16) | h
;
631 void r300EmitWait(r300ContextPtr rmesa
, GLuint flags
)
633 if (rmesa
->radeon
.dri
.drmMinor
>= 6) {
634 drm_radeon_cmd_header_t
*cmd
;
636 assert(!(flags
& ~(RADEON_WAIT_2D
| RADEON_WAIT_3D
)));
639 (drm_radeon_cmd_header_t
*) r300AllocCmdBuf(rmesa
,
643 cmd
[0].wait
.cmd_type
= RADEON_CMD_WAIT
;
644 cmd
[0].wait
.flags
= flags
;
648 void r300EmitAOS(r300ContextPtr rmesa
, GLuint nr
, GLuint offset
)
650 if (RADEON_DEBUG
& DEBUG_VERTS
)
651 fprintf(stderr
, "%s: nr=%d, ofs=0x%08x\n", __func__
, nr
, offset
);
652 int sz
= 1 + (nr
>> 1) * 3 + (nr
& 1) * 2;
656 start_packet3(RADEON_CP_PACKET3_3D_LOAD_VBPNTR
, sz
-1);
658 for(i
=0;i
+1<nr
;i
+=2){
659 e32( (rmesa
->state
.aos
[i
].aos_size
<< 0)
660 |(rmesa
->state
.aos
[i
].aos_stride
<< 8)
661 |(rmesa
->state
.aos
[i
+1].aos_size
<< 16)
662 |(rmesa
->state
.aos
[i
+1].aos_stride
<< 24)
664 e32(rmesa
->state
.aos
[i
].aos_offset
+offset
*4*rmesa
->state
.aos
[i
].aos_stride
);
665 e32(rmesa
->state
.aos
[i
+1].aos_offset
+offset
*4*rmesa
->state
.aos
[i
+1].aos_stride
);
668 e32( (rmesa
->state
.aos
[nr
-1].aos_size
<< 0)
669 |(rmesa
->state
.aos
[nr
-1].aos_stride
<< 8)
671 e32(rmesa
->state
.aos
[nr
-1].aos_offset
+offset
*4*rmesa
->state
.aos
[nr
-1].aos_stride
);