2 #include "util/u_math.h"
4 #include "nvc0_context.h"
7 nvc0_validate_zcull(struct nvc0_context
*nvc0
)
9 struct nouveau_channel
*chan
= nvc0
->screen
->base
.channel
;
10 struct pipe_framebuffer_state
*fb
= &nvc0
->framebuffer
;
11 struct nvc0_surface
*sf
= nvc0_surface(fb
->zsbuf
);
12 struct nvc0_miptree
*mt
= nvc0_miptree(sf
->base
.texture
);
13 struct nouveau_bo
*bo
= mt
->base
.bo
;
15 uint32_t offset
= align(mt
->total_size
, 1 << 17);
16 unsigned width
, height
;
18 assert(mt
->base
.base
.depth0
== 1 && mt
->base
.base
.array_size
< 2);
20 size
= mt
->total_size
* 2;
22 height
= align(fb
->height
, 32);
23 width
= fb
->width
% 224;
25 width
= fb
->width
+ (224 - width
);
29 MARK_RING (chan
, 23, 4);
30 BEGIN_RING(chan
, RING_3D_(0x1590), 1); /* ZCULL_REGION_INDEX (bits 0x3f) */
32 BEGIN_RING(chan
, RING_3D_(0x07e8), 2); /* ZCULL_ADDRESS_A_HIGH */
33 OUT_RELOCh(chan
, bo
, offset
, NOUVEAU_BO_VRAM
| NOUVEAU_BO_RDWR
);
34 OUT_RELOCl(chan
, bo
, offset
, NOUVEAU_BO_VRAM
| NOUVEAU_BO_RDWR
);
36 BEGIN_RING(chan
, RING_3D_(0x07f0), 2); /* ZCULL_ADDRESS_B_HIGH */
37 OUT_RELOCh(chan
, bo
, offset
, NOUVEAU_BO_VRAM
| NOUVEAU_BO_RDWR
);
38 OUT_RELOCl(chan
, bo
, offset
, NOUVEAU_BO_VRAM
| NOUVEAU_BO_RDWR
);
39 BEGIN_RING(chan
, RING_3D_(0x07e0), 2);
40 OUT_RING (chan
, size
);
41 OUT_RING (chan
, size
>> 16);
42 BEGIN_RING(chan
, RING_3D_(0x15c8), 1); /* bits 0x3 */
44 BEGIN_RING(chan
, RING_3D_(0x07c0), 4); /* ZCULL dimensions */
45 OUT_RING (chan
, width
);
46 OUT_RING (chan
, height
);
49 BEGIN_RING(chan
, RING_3D_(0x15fc), 2);
50 OUT_RING (chan
, 0); /* bits 0xffff */
51 OUT_RING (chan
, 0); /* bits 0xffff */
52 BEGIN_RING(chan
, RING_3D_(0x1958), 1);
53 OUT_RING (chan
, 0); /* bits ~0 */
57 nvc0_validate_fb(struct nvc0_context
*nvc0
)
59 struct nouveau_channel
*chan
= nvc0
->screen
->base
.channel
;
60 struct pipe_framebuffer_state
*fb
= &nvc0
->framebuffer
;
62 boolean serialize
= FALSE
;
64 nvc0_bufctx_reset(nvc0
, NVC0_BUFCTX_FRAME
);
66 BEGIN_RING(chan
, RING_3D(RT_CONTROL
), 1);
67 OUT_RING (chan
, (076543210 << 4) | fb
->nr_cbufs
);
68 BEGIN_RING(chan
, RING_3D(SCREEN_SCISSOR_HORIZ
), 2);
69 OUT_RING (chan
, fb
->width
<< 16);
70 OUT_RING (chan
, fb
->height
<< 16);
72 MARK_RING(chan
, 9 * fb
->nr_cbufs
, 2 * fb
->nr_cbufs
);
74 for (i
= 0; i
< fb
->nr_cbufs
; ++i
) {
75 struct nvc0_miptree
*mt
= nvc0_miptree(fb
->cbufs
[i
]->texture
);
76 struct nvc0_surface
*sf
= nvc0_surface(fb
->cbufs
[i
]);
77 struct nouveau_bo
*bo
= mt
->base
.bo
;
78 uint32_t offset
= sf
->offset
;
80 BEGIN_RING(chan
, RING_3D(RT_ADDRESS_HIGH(i
)), 9);
81 OUT_RELOCh(chan
, bo
, offset
, NOUVEAU_BO_VRAM
| NOUVEAU_BO_RDWR
);
82 OUT_RELOCl(chan
, bo
, offset
, NOUVEAU_BO_VRAM
| NOUVEAU_BO_RDWR
);
83 OUT_RING (chan
, sf
->width
);
84 OUT_RING (chan
, sf
->height
);
85 OUT_RING (chan
, nvc0_format_table
[sf
->base
.format
].rt
);
86 OUT_RING (chan
, (mt
->layout_3d
<< 16) |
87 mt
->level
[sf
->base
.u
.tex
.level
].tile_mode
);
88 OUT_RING (chan
, sf
->base
.u
.tex
.first_layer
+ sf
->depth
);
89 OUT_RING (chan
, mt
->layer_stride
>> 2);
90 OUT_RING (chan
, sf
->base
.u
.tex
.first_layer
);
92 if (mt
->base
.status
& NOUVEAU_BUFFER_STATUS_GPU_READING
)
94 mt
->base
.status
|= NOUVEAU_BUFFER_STATUS_GPU_WRITING
;
95 mt
->base
.status
&= ~NOUVEAU_BUFFER_STATUS_GPU_READING
;
97 /* only register for writing, otherwise we'd always serialize here */
98 nvc0_bufctx_add_resident(nvc0
, NVC0_BUFCTX_FRAME
, &mt
->base
,
99 NOUVEAU_BO_VRAM
| NOUVEAU_BO_WR
);
103 struct nvc0_miptree
*mt
= nvc0_miptree(fb
->zsbuf
->texture
);
104 struct nvc0_surface
*sf
= nvc0_surface(fb
->zsbuf
);
105 struct nouveau_bo
*bo
= mt
->base
.bo
;
106 int unk
= mt
->base
.base
.target
== PIPE_TEXTURE_2D
;
107 uint32_t offset
= sf
->offset
;
109 MARK_RING (chan
, 12, 2);
110 BEGIN_RING(chan
, RING_3D(ZETA_ADDRESS_HIGH
), 5);
111 OUT_RELOCh(chan
, bo
, offset
, NOUVEAU_BO_VRAM
| NOUVEAU_BO_RDWR
);
112 OUT_RELOCl(chan
, bo
, offset
, NOUVEAU_BO_VRAM
| NOUVEAU_BO_RDWR
);
113 OUT_RING (chan
, nvc0_format_table
[fb
->zsbuf
->format
].rt
);
114 OUT_RING (chan
, mt
->level
[sf
->base
.u
.tex
.level
].tile_mode
);
115 OUT_RING (chan
, mt
->layer_stride
>> 2);
116 BEGIN_RING(chan
, RING_3D(ZETA_ENABLE
), 1);
118 BEGIN_RING(chan
, RING_3D(ZETA_HORIZ
), 3);
119 OUT_RING (chan
, sf
->width
);
120 OUT_RING (chan
, sf
->height
);
121 OUT_RING (chan
, (unk
<< 16) |
122 (sf
->base
.u
.tex
.first_layer
+ sf
->depth
));
123 BEGIN_RING(chan
, RING_3D(ZETA_BASE_LAYER
), 1);
124 OUT_RING (chan
, sf
->base
.u
.tex
.first_layer
);
126 if (mt
->base
.status
& NOUVEAU_BUFFER_STATUS_GPU_READING
)
128 mt
->base
.status
|= NOUVEAU_BUFFER_STATUS_GPU_WRITING
;
129 mt
->base
.status
&= ~NOUVEAU_BUFFER_STATUS_GPU_READING
;
131 nvc0_bufctx_add_resident(nvc0
, NVC0_BUFCTX_FRAME
, &mt
->base
,
132 NOUVEAU_BO_VRAM
| NOUVEAU_BO_WR
);
134 BEGIN_RING(chan
, RING_3D(ZETA_ENABLE
), 1);
139 BEGIN_RING(chan
, RING_3D(SERIALIZE
), 1);
145 nvc0_validate_blend_colour(struct nvc0_context
*nvc0
)
147 struct nouveau_channel
*chan
= nvc0
->screen
->base
.channel
;
149 BEGIN_RING(chan
, RING_3D(BLEND_COLOR(0)), 4);
150 OUT_RINGf (chan
, nvc0
->blend_colour
.color
[0]);
151 OUT_RINGf (chan
, nvc0
->blend_colour
.color
[1]);
152 OUT_RINGf (chan
, nvc0
->blend_colour
.color
[2]);
153 OUT_RINGf (chan
, nvc0
->blend_colour
.color
[3]);
157 nvc0_validate_stencil_ref(struct nvc0_context
*nvc0
)
159 struct nouveau_channel
*chan
= nvc0
->screen
->base
.channel
;
160 const ubyte
*ref
= &nvc0
->stencil_ref
.ref_value
[0];
162 IMMED_RING(chan
, RING_3D(STENCIL_FRONT_FUNC_REF
), ref
[0]);
163 IMMED_RING(chan
, RING_3D(STENCIL_BACK_FUNC_REF
), ref
[1]);
167 nvc0_validate_stipple(struct nvc0_context
*nvc0
)
169 struct nouveau_channel
*chan
= nvc0
->screen
->base
.channel
;
172 BEGIN_RING(chan
, RING_3D(POLYGON_STIPPLE_PATTERN(0)), 32);
173 for (i
= 0; i
< 32; ++i
)
174 OUT_RING(chan
, util_bswap32(nvc0
->stipple
.stipple
[i
]));
178 nvc0_validate_scissor(struct nvc0_context
*nvc0
)
180 struct nouveau_channel
*chan
= nvc0
->screen
->base
.channel
;
181 struct pipe_scissor_state
*s
= &nvc0
->scissor
;
183 if (!(nvc0
->dirty
& NVC0_NEW_SCISSOR
) &&
184 nvc0
->rast
->pipe
.scissor
== nvc0
->state
.scissor
)
186 nvc0
->state
.scissor
= nvc0
->rast
->pipe
.scissor
;
188 BEGIN_RING(chan
, RING_3D(SCISSOR_HORIZ(0)), 2);
189 if (nvc0
->rast
->pipe
.scissor
) {
190 OUT_RING(chan
, (s
->maxx
<< 16) | s
->minx
);
191 OUT_RING(chan
, (s
->maxy
<< 16) | s
->miny
);
193 OUT_RING(chan
, (0xffff << 16) | 0);
194 OUT_RING(chan
, (0xffff << 16) | 0);
199 nvc0_validate_viewport(struct nvc0_context
*nvc0
)
201 struct nouveau_channel
*chan
= nvc0
->screen
->base
.channel
;
202 struct pipe_viewport_state
*vp
= &nvc0
->viewport
;
206 BEGIN_RING(chan
, RING_3D(VIEWPORT_TRANSLATE_X(0)), 3);
207 OUT_RINGf (chan
, vp
->translate
[0]);
208 OUT_RINGf (chan
, vp
->translate
[1]);
209 OUT_RINGf (chan
, vp
->translate
[2]);
210 BEGIN_RING(chan
, RING_3D(VIEWPORT_SCALE_X(0)), 3);
211 OUT_RINGf (chan
, vp
->scale
[0]);
212 OUT_RINGf (chan
, vp
->scale
[1]);
213 OUT_RINGf (chan
, vp
->scale
[2]);
215 /* now set the viewport rectangle to viewport dimensions for clipping */
217 x
= util_iround(MAX2(0.0f
, vp
->translate
[0] - fabsf(vp
->scale
[0])));
218 y
= util_iround(MAX2(0.0f
, vp
->translate
[1] - fabsf(vp
->scale
[1])));
219 w
= util_iround(vp
->translate
[0] + fabsf(vp
->scale
[0])) - x
;
220 h
= util_iround(vp
->translate
[1] + fabsf(vp
->scale
[1])) - y
;
222 zmin
= vp
->translate
[2] - fabsf(vp
->scale
[2]);
223 zmax
= vp
->translate
[2] + fabsf(vp
->scale
[2]);
225 BEGIN_RING(chan
, RING_3D(VIEWPORT_HORIZ(0)), 2);
226 OUT_RING (chan
, (w
<< 16) | x
);
227 OUT_RING (chan
, (h
<< 16) | y
);
228 BEGIN_RING(chan
, RING_3D(DEPTH_RANGE_NEAR(0)), 2);
229 OUT_RINGf (chan
, zmin
);
230 OUT_RINGf (chan
, zmax
);
234 nvc0_validate_clip(struct nvc0_context
*nvc0
)
236 struct nouveau_channel
*chan
= nvc0
->screen
->base
.channel
;
239 if (nvc0
->clip
.depth_clamp
) {
241 NVC0_3D_VIEW_VOLUME_CLIP_CTRL_UNK1_UNK1
|
242 NVC0_3D_VIEW_VOLUME_CLIP_CTRL_DEPTH_CLAMP_NEAR
|
243 NVC0_3D_VIEW_VOLUME_CLIP_CTRL_DEPTH_CLAMP_FAR
|
244 NVC0_3D_VIEW_VOLUME_CLIP_CTRL_UNK12_UNK2
;
246 clip
= NVC0_3D_VIEW_VOLUME_CLIP_CTRL_UNK1_UNK1
;
249 BEGIN_RING(chan
, RING_3D(VIEW_VOLUME_CLIP_CTRL
), 1);
250 OUT_RING (chan
, clip
);
253 struct nouveau_bo
*bo
= nvc0
->screen
->uniforms
;
255 MARK_RING (chan
, 6 + nvc0
->clip
.nr
* 4, 2);
256 BEGIN_RING(chan
, RING_3D(CB_SIZE
), 3);
257 OUT_RING (chan
, 256);
258 OUT_RELOCh(chan
, bo
, 5 << 16, NOUVEAU_BO_VRAM
| NOUVEAU_BO_RD
);
259 OUT_RELOCl(chan
, bo
, 5 << 16, NOUVEAU_BO_VRAM
| NOUVEAU_BO_RD
);
260 BEGIN_RING_1I(chan
, RING_3D(CB_POS
), nvc0
->clip
.nr
* 4 + 1);
262 OUT_RINGp (chan
, &nvc0
->clip
.ucp
[0][0], nvc0
->clip
.nr
* 4);
264 BEGIN_RING(chan
, RING_3D(VP_CLIP_DISTANCE_ENABLE
), 1);
265 OUT_RING (chan
, (1 << nvc0
->clip
.nr
) - 1);
267 IMMED_RING(chan
, RING_3D(VP_CLIP_DISTANCE_ENABLE
), 0);
272 nvc0_validate_blend(struct nvc0_context
*nvc0
)
274 struct nouveau_channel
*chan
= nvc0
->screen
->base
.channel
;
276 WAIT_RING(chan
, nvc0
->blend
->size
);
277 OUT_RINGp(chan
, nvc0
->blend
->state
, nvc0
->blend
->size
);
281 nvc0_validate_zsa(struct nvc0_context
*nvc0
)
283 struct nouveau_channel
*chan
= nvc0
->screen
->base
.channel
;
285 WAIT_RING(chan
, nvc0
->zsa
->size
);
286 OUT_RINGp(chan
, nvc0
->zsa
->state
, nvc0
->zsa
->size
);
290 nvc0_validate_rasterizer(struct nvc0_context
*nvc0
)
292 struct nouveau_channel
*chan
= nvc0
->screen
->base
.channel
;
294 WAIT_RING(chan
, nvc0
->rast
->size
);
295 OUT_RINGp(chan
, nvc0
->rast
->state
, nvc0
->rast
->size
);
299 nvc0_constbufs_validate(struct nvc0_context
*nvc0
)
301 struct nouveau_channel
*chan
= nvc0
->screen
->base
.channel
;
302 struct nouveau_bo
*bo
;
305 for (s
= 0; s
< 5; ++s
) {
306 struct nv04_resource
*res
;
309 while (nvc0
->constbuf_dirty
[s
]) {
311 unsigned offset
= 0, words
= 0;
312 boolean rebind
= TRUE
;
314 i
= ffs(nvc0
->constbuf_dirty
[s
]) - 1;
315 nvc0
->constbuf_dirty
[s
] &= ~(1 << i
);
317 res
= nv04_resource(nvc0
->constbuf
[s
][i
]);
319 BEGIN_RING(chan
, RING_3D(CB_BIND(s
)), 1);
320 OUT_RING (chan
, (i
<< 4) | 0);
322 nvc0
->state
.uniform_buffer_bound
[s
] = 0;
326 if (!nouveau_resource_mapped_by_gpu(&res
->base
)) {
329 bo
= nvc0
->screen
->uniforms
;
331 if (nvc0
->state
.uniform_buffer_bound
[s
] >= res
->base
.width0
)
334 nvc0
->state
.uniform_buffer_bound
[s
] =
335 align(res
->base
.width0
, 0x100);
340 nvc0_m2mf_push_linear(nvc0
, bo
, NOUVEAU_BO_VRAM
,
341 base
, res
->base
.width0
, res
->data
);
342 BEGIN_RING(chan
, RING_3D_(0x021c), 1);
343 OUT_RING (chan
, 0x1111);
345 words
= res
->base
.width0
/ 4;
350 nvc0
->state
.uniform_buffer_bound
[s
] = 0;
353 if (bo
!= nvc0
->screen
->uniforms
)
354 nvc0_bufctx_add_resident(nvc0
, NVC0_BUFCTX_CONSTANT
, res
,
355 NOUVEAU_BO_VRAM
| NOUVEAU_BO_RD
);
358 MARK_RING (chan
, 4, 2);
359 BEGIN_RING(chan
, RING_3D(CB_SIZE
), 3);
360 OUT_RING (chan
, align(res
->base
.width0
, 0x100));
361 OUT_RELOCh(chan
, bo
, base
, NOUVEAU_BO_VRAM
| NOUVEAU_BO_RD
);
362 OUT_RELOCl(chan
, bo
, base
, NOUVEAU_BO_VRAM
| NOUVEAU_BO_RD
);
363 BEGIN_RING(chan
, RING_3D(CB_BIND(s
)), 1);
364 OUT_RING (chan
, (i
<< 4) | 1);
368 unsigned nr
= AVAIL_RING(chan
);
374 nr
= MIN2(MIN2(nr
- 6, words
), NV04_PFIFO_MAX_PACKET_LEN
- 1);
376 MARK_RING (chan
, nr
+ 5, 2);
377 BEGIN_RING(chan
, RING_3D(CB_SIZE
), 3);
378 OUT_RING (chan
, align(res
->base
.width0
, 0x100));
379 OUT_RELOCh(chan
, bo
, base
, NOUVEAU_BO_VRAM
| NOUVEAU_BO_RD
);
380 OUT_RELOCl(chan
, bo
, base
, NOUVEAU_BO_VRAM
| NOUVEAU_BO_RD
);
381 BEGIN_RING_1I(chan
, RING_3D(CB_POS
), nr
+ 1);
382 OUT_RING (chan
, offset
);
383 OUT_RINGp (chan
, &res
->data
[offset
], nr
);
393 nvc0_validate_derived_1(struct nvc0_context
*nvc0
)
395 struct nouveau_channel
*chan
= nvc0
->screen
->base
.channel
;
398 early_z
= nvc0
->fragprog
->fp
.early_z
&& !nvc0
->zsa
->pipe
.alpha
.enabled
;
400 if (early_z
!= nvc0
->state
.early_z
) {
401 nvc0
->state
.early_z
= early_z
;
402 IMMED_RING(chan
, RING_3D(EARLY_FRAGMENT_TESTS
), early_z
);
407 nvc0_switch_pipe_context(struct nvc0_context
*ctx_to
)
409 struct nvc0_context
*ctx_from
= ctx_to
->screen
->cur_ctx
;
412 ctx_to
->state
= ctx_from
->state
;
417 ctx_to
->dirty
&= ~(NVC0_NEW_VERTEX
| NVC0_NEW_ARRAYS
);
419 if (!ctx_to
->vertprog
)
420 ctx_to
->dirty
&= ~NVC0_NEW_VERTPROG
;
421 if (!ctx_to
->fragprog
)
422 ctx_to
->dirty
&= ~NVC0_NEW_FRAGPROG
;
425 ctx_to
->dirty
&= ~NVC0_NEW_BLEND
;
427 ctx_to
->dirty
&= ~NVC0_NEW_RASTERIZER
;
429 ctx_to
->dirty
&= ~NVC0_NEW_ZSA
;
431 ctx_to
->screen
->base
.channel
->user_private
= ctx_to
->screen
->cur_ctx
=
435 static struct state_validate
{
436 void (*func
)(struct nvc0_context
*);
438 } validate_list
[] = {
439 { nvc0_validate_fb
, NVC0_NEW_FRAMEBUFFER
},
440 { nvc0_validate_blend
, NVC0_NEW_BLEND
},
441 { nvc0_validate_zsa
, NVC0_NEW_ZSA
},
442 { nvc0_validate_rasterizer
, NVC0_NEW_RASTERIZER
},
443 { nvc0_validate_blend_colour
, NVC0_NEW_BLEND_COLOUR
},
444 { nvc0_validate_stencil_ref
, NVC0_NEW_STENCIL_REF
},
445 { nvc0_validate_stipple
, NVC0_NEW_STIPPLE
},
446 { nvc0_validate_scissor
, NVC0_NEW_SCISSOR
| NVC0_NEW_RASTERIZER
},
447 { nvc0_validate_viewport
, NVC0_NEW_VIEWPORT
},
448 { nvc0_validate_clip
, NVC0_NEW_CLIP
},
449 { nvc0_vertprog_validate
, NVC0_NEW_VERTPROG
},
450 { nvc0_tctlprog_validate
, NVC0_NEW_TCTLPROG
},
451 { nvc0_tevlprog_validate
, NVC0_NEW_TEVLPROG
},
452 { nvc0_gmtyprog_validate
, NVC0_NEW_GMTYPROG
},
453 { nvc0_fragprog_validate
, NVC0_NEW_FRAGPROG
},
454 { nvc0_validate_derived_1
, NVC0_NEW_FRAGPROG
| NVC0_NEW_ZSA
},
455 { nvc0_constbufs_validate
, NVC0_NEW_CONSTBUF
},
456 { nvc0_validate_textures
, NVC0_NEW_TEXTURES
},
457 { nvc0_validate_samplers
, NVC0_NEW_SAMPLERS
},
458 { nvc0_vertex_arrays_validate
, NVC0_NEW_VERTEX
| NVC0_NEW_ARRAYS
},
459 { nvc0_tfb_validate
, NVC0_NEW_TFB
| NVC0_NEW_TFB_BUFFERS
}
461 #define validate_list_len (sizeof(validate_list) / sizeof(validate_list[0]))
464 nvc0_state_validate(struct nvc0_context
*nvc0
)
468 if (nvc0
->screen
->cur_ctx
!= nvc0
)
469 nvc0_switch_pipe_context(nvc0
);
472 for (i
= 0; i
< validate_list_len
; ++i
) {
473 struct state_validate
*validate
= &validate_list
[i
];
475 if (nvc0
->dirty
& validate
->states
)
476 validate
->func(nvc0
);
481 nvc0_bufctx_emit_relocs(nvc0
);