radeon: Move gfx/dma cs cleanup to r600_common_context_cleanup
[mesa.git] / src / gallium / drivers / r600 / r600_pipe.c
1 /*
2 * Copyright 2010 Jerome Glisse <glisse@freedesktop.org>
3 *
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * on the rights to use, copy, modify, merge, publish, distribute, sub
8 * license, and/or sell copies of the Software, and to permit persons to whom
9 * the Software is furnished to do so, subject to the following conditions:
10 *
11 * The above copyright notice and this permission notice (including the next
12 * paragraph) shall be included in all copies or substantial portions of the
13 * Software.
14 *
15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17 * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL
18 * THE AUTHOR(S) AND/OR THEIR SUPPLIERS BE LIABLE FOR ANY CLAIM,
19 * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
20 * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
21 * USE OR OTHER DEALINGS IN THE SOFTWARE.
22 */
23 #include "r600_pipe.h"
24 #include "r600_public.h"
25 #include "r600_isa.h"
26 #include "evergreen_compute.h"
27 #include "r600d.h"
28
29 #include "sb/sb_public.h"
30
31 #include <errno.h>
32 #include "pipe/p_shader_tokens.h"
33 #include "util/u_blitter.h"
34 #include "util/u_debug.h"
35 #include "util/u_memory.h"
36 #include "util/u_simple_shaders.h"
37 #include "util/u_upload_mgr.h"
38 #include "util/u_math.h"
39 #include "vl/vl_decoder.h"
40 #include "vl/vl_video_buffer.h"
41 #include "radeon/radeon_uvd.h"
42 #include "os/os_time.h"
43
44 static const struct debug_named_value r600_debug_options[] = {
45 /* features */
46 #if defined(R600_USE_LLVM)
47 { "nollvm", DBG_NO_LLVM, "Disable the LLVM shader compiler" },
48 #endif
49 { "nocpdma", DBG_NO_CP_DMA, "Disable CP DMA" },
50 { "nodma", DBG_NO_ASYNC_DMA, "Disable asynchronous DMA" },
51
52 /* shader backend */
53 { "nosb", DBG_NO_SB, "Disable sb backend for graphics shaders" },
54 { "sbcl", DBG_SB_CS, "Enable sb backend for compute shaders" },
55 { "sbdry", DBG_SB_DRY_RUN, "Don't use optimized bytecode (just print the dumps)" },
56 { "sbstat", DBG_SB_STAT, "Print optimization statistics for shaders" },
57 { "sbdump", DBG_SB_DUMP, "Print IR dumps after some optimization passes" },
58 { "sbnofallback", DBG_SB_NO_FALLBACK, "Abort on errors instead of fallback" },
59 { "sbdisasm", DBG_SB_DISASM, "Use sb disassembler for shader dumps" },
60 { "sbsafemath", DBG_SB_SAFEMATH, "Disable unsafe math optimizations" },
61
62 DEBUG_NAMED_VALUE_END /* must be last */
63 };
64
65 /*
66 * pipe_context
67 */
68
69 static void r600_flush(struct pipe_context *ctx, unsigned flags)
70 {
71 struct r600_context *rctx = (struct r600_context *)ctx;
72 struct pipe_query *render_cond = NULL;
73 unsigned render_cond_mode = 0;
74 boolean render_cond_cond = FALSE;
75
76 if (rctx->b.rings.gfx.cs->cdw == rctx->initial_gfx_cs_size)
77 return;
78
79 rctx->b.rings.gfx.flushing = true;
80 /* Disable render condition. */
81 if (rctx->current_render_cond) {
82 render_cond = rctx->current_render_cond;
83 render_cond_cond = rctx->current_render_cond_cond;
84 render_cond_mode = rctx->current_render_cond_mode;
85 ctx->render_condition(ctx, NULL, FALSE, 0);
86 }
87
88 r600_context_flush(rctx, flags);
89 rctx->b.rings.gfx.flushing = false;
90 r600_begin_new_cs(rctx);
91
92 /* Re-enable render condition. */
93 if (render_cond) {
94 ctx->render_condition(ctx, render_cond, render_cond_cond, render_cond_mode);
95 }
96
97 rctx->initial_gfx_cs_size = rctx->b.rings.gfx.cs->cdw;
98 }
99
100 static void r600_flush_from_st(struct pipe_context *ctx,
101 struct pipe_fence_handle **fence,
102 unsigned flags)
103 {
104 struct r600_context *rctx = (struct r600_context *)ctx;
105 unsigned fflags;
106
107 fflags = flags & PIPE_FLUSH_END_OF_FRAME ? RADEON_FLUSH_END_OF_FRAME : 0;
108 if (fence) {
109 *fence = rctx->b.ws->cs_create_fence(rctx->b.rings.gfx.cs);
110 }
111 /* flush gfx & dma ring, order does not matter as only one can be live */
112 if (rctx->b.rings.dma.cs) {
113 rctx->b.rings.dma.flush(rctx, fflags);
114 }
115 rctx->b.rings.gfx.flush(rctx, fflags);
116 }
117
118 static void r600_flush_gfx_ring(void *ctx, unsigned flags)
119 {
120 r600_flush((struct pipe_context*)ctx, flags);
121 }
122
123 static void r600_flush_dma_ring(void *ctx, unsigned flags)
124 {
125 struct r600_context *rctx = (struct r600_context *)ctx;
126 struct radeon_winsys_cs *cs = rctx->b.rings.dma.cs;
127
128 if (!cs->cdw) {
129 return;
130 }
131
132 rctx->b.rings.dma.flushing = true;
133 rctx->b.ws->cs_flush(cs, flags, 0);
134 rctx->b.rings.dma.flushing = false;
135 }
136
137 static void r600_flush_from_winsys(void *ctx, unsigned flags)
138 {
139 struct r600_context *rctx = (struct r600_context *)ctx;
140
141 rctx->b.rings.gfx.flush(rctx, flags);
142 }
143
144 static void r600_flush_dma_from_winsys(void *ctx, unsigned flags)
145 {
146 struct r600_context *rctx = (struct r600_context *)ctx;
147
148 rctx->b.rings.dma.flush(rctx, flags);
149 }
150
151 static void r600_destroy_context(struct pipe_context *context)
152 {
153 struct r600_context *rctx = (struct r600_context *)context;
154
155 r600_isa_destroy(rctx->isa);
156
157 r600_sb_context_destroy(rctx->sb_context);
158
159 pipe_resource_reference((struct pipe_resource**)&rctx->dummy_cmask, NULL);
160 pipe_resource_reference((struct pipe_resource**)&rctx->dummy_fmask, NULL);
161
162 if (rctx->dummy_pixel_shader) {
163 rctx->b.b.delete_fs_state(&rctx->b.b, rctx->dummy_pixel_shader);
164 }
165 if (rctx->custom_dsa_flush) {
166 rctx->b.b.delete_depth_stencil_alpha_state(&rctx->b.b, rctx->custom_dsa_flush);
167 }
168 if (rctx->custom_blend_resolve) {
169 rctx->b.b.delete_blend_state(&rctx->b.b, rctx->custom_blend_resolve);
170 }
171 if (rctx->custom_blend_decompress) {
172 rctx->b.b.delete_blend_state(&rctx->b.b, rctx->custom_blend_decompress);
173 }
174 if (rctx->custom_blend_fastclear) {
175 rctx->b.b.delete_blend_state(&rctx->b.b, rctx->custom_blend_fastclear);
176 }
177 util_unreference_framebuffer_state(&rctx->framebuffer.state);
178
179 if (rctx->blitter) {
180 util_blitter_destroy(rctx->blitter);
181 }
182 if (rctx->allocator_fetch_shader) {
183 u_suballocator_destroy(rctx->allocator_fetch_shader);
184 }
185
186 r600_release_command_buffer(&rctx->start_cs_cmd);
187
188 FREE(rctx->start_compute_cs_cmd.buf);
189
190 r600_common_context_cleanup(&rctx->b);
191 FREE(rctx);
192 }
193
194 static struct pipe_context *r600_create_context(struct pipe_screen *screen, void *priv)
195 {
196 struct r600_context *rctx = CALLOC_STRUCT(r600_context);
197 struct r600_screen* rscreen = (struct r600_screen *)screen;
198
199 if (rctx == NULL)
200 return NULL;
201
202 rctx->b.b.screen = screen;
203 rctx->b.b.priv = priv;
204 rctx->b.b.destroy = r600_destroy_context;
205 rctx->b.b.flush = r600_flush_from_st;
206
207 if (!r600_common_context_init(&rctx->b, &rscreen->b))
208 goto fail;
209
210 rctx->screen = rscreen;
211 rctx->keep_tiling_flags = rscreen->b.info.drm_minor >= 12;
212
213 LIST_INITHEAD(&rctx->active_nontimer_queries);
214
215 r600_init_blit_functions(rctx);
216 r600_init_query_functions(rctx);
217 r600_init_context_resource_functions(rctx);
218
219 if (rscreen->b.info.has_uvd) {
220 rctx->b.b.create_video_codec = r600_uvd_create_decoder;
221 rctx->b.b.create_video_buffer = r600_video_buffer_create;
222 } else {
223 rctx->b.b.create_video_codec = vl_create_decoder;
224 rctx->b.b.create_video_buffer = vl_video_buffer_create;
225 }
226
227 r600_init_common_state_functions(rctx);
228
229 switch (rctx->b.chip_class) {
230 case R600:
231 case R700:
232 r600_init_state_functions(rctx);
233 r600_init_atom_start_cs(rctx);
234 rctx->max_db = 4;
235 rctx->custom_dsa_flush = r600_create_db_flush_dsa(rctx);
236 rctx->custom_blend_resolve = rctx->b.chip_class == R700 ? r700_create_resolve_blend(rctx)
237 : r600_create_resolve_blend(rctx);
238 rctx->custom_blend_decompress = r600_create_decompress_blend(rctx);
239 rctx->has_vertex_cache = !(rctx->b.family == CHIP_RV610 ||
240 rctx->b.family == CHIP_RV620 ||
241 rctx->b.family == CHIP_RS780 ||
242 rctx->b.family == CHIP_RS880 ||
243 rctx->b.family == CHIP_RV710);
244 break;
245 case EVERGREEN:
246 case CAYMAN:
247 evergreen_init_state_functions(rctx);
248 evergreen_init_atom_start_cs(rctx);
249 evergreen_init_atom_start_compute_cs(rctx);
250 rctx->max_db = 8;
251 rctx->custom_dsa_flush = evergreen_create_db_flush_dsa(rctx);
252 rctx->custom_blend_resolve = evergreen_create_resolve_blend(rctx);
253 rctx->custom_blend_decompress = evergreen_create_decompress_blend(rctx);
254 rctx->custom_blend_fastclear = evergreen_create_fastclear_blend(rctx);
255 rctx->has_vertex_cache = !(rctx->b.family == CHIP_CEDAR ||
256 rctx->b.family == CHIP_PALM ||
257 rctx->b.family == CHIP_SUMO ||
258 rctx->b.family == CHIP_SUMO2 ||
259 rctx->b.family == CHIP_CAICOS ||
260 rctx->b.family == CHIP_CAYMAN ||
261 rctx->b.family == CHIP_ARUBA);
262 break;
263 default:
264 R600_ERR("Unsupported chip class %d.\n", rctx->b.chip_class);
265 goto fail;
266 }
267
268 if (rscreen->trace_bo) {
269 rctx->b.rings.gfx.cs = rctx->b.ws->cs_create(rctx->b.ws, RING_GFX, rscreen->trace_bo->cs_buf);
270 } else {
271 rctx->b.rings.gfx.cs = rctx->b.ws->cs_create(rctx->b.ws, RING_GFX, NULL);
272 }
273 rctx->b.rings.gfx.flush = r600_flush_gfx_ring;
274 rctx->b.ws->cs_set_flush_callback(rctx->b.rings.gfx.cs, r600_flush_from_winsys, rctx);
275 rctx->b.rings.gfx.flushing = false;
276
277 rctx->b.rings.dma.cs = NULL;
278 if (rscreen->b.info.r600_has_dma && !(rscreen->b.debug_flags & DBG_NO_ASYNC_DMA)) {
279 rctx->b.rings.dma.cs = rctx->b.ws->cs_create(rctx->b.ws, RING_DMA, NULL);
280 rctx->b.rings.dma.flush = r600_flush_dma_ring;
281 rctx->b.ws->cs_set_flush_callback(rctx->b.rings.dma.cs, r600_flush_dma_from_winsys, rctx);
282 rctx->b.rings.dma.flushing = false;
283 }
284
285 rctx->allocator_fetch_shader = u_suballocator_create(&rctx->b.b, 64 * 1024, 256,
286 0, PIPE_USAGE_STATIC, FALSE);
287 if (!rctx->allocator_fetch_shader)
288 goto fail;
289
290 rctx->isa = calloc(1, sizeof(struct r600_isa));
291 if (!rctx->isa || r600_isa_init(rctx, rctx->isa))
292 goto fail;
293
294 rctx->blitter = util_blitter_create(&rctx->b.b);
295 if (rctx->blitter == NULL)
296 goto fail;
297 util_blitter_set_texture_multisample(rctx->blitter, rscreen->has_msaa);
298 rctx->blitter->draw_rectangle = r600_draw_rectangle;
299
300 r600_begin_new_cs(rctx);
301 r600_get_backend_mask(rctx); /* this emits commands and must be last */
302
303 rctx->dummy_pixel_shader =
304 util_make_fragment_cloneinput_shader(&rctx->b.b, 0,
305 TGSI_SEMANTIC_GENERIC,
306 TGSI_INTERPOLATE_CONSTANT);
307 rctx->b.b.bind_fs_state(&rctx->b.b, rctx->dummy_pixel_shader);
308
309 return &rctx->b.b;
310
311 fail:
312 r600_destroy_context(&rctx->b.b);
313 return NULL;
314 }
315
316 /*
317 * pipe_screen
318 */
319 static const char* r600_get_vendor(struct pipe_screen* pscreen)
320 {
321 return "X.Org";
322 }
323
324 static const char *r600_get_family_name(enum radeon_family family)
325 {
326 switch(family) {
327 case CHIP_R600: return "AMD R600";
328 case CHIP_RV610: return "AMD RV610";
329 case CHIP_RV630: return "AMD RV630";
330 case CHIP_RV670: return "AMD RV670";
331 case CHIP_RV620: return "AMD RV620";
332 case CHIP_RV635: return "AMD RV635";
333 case CHIP_RS780: return "AMD RS780";
334 case CHIP_RS880: return "AMD RS880";
335 case CHIP_RV770: return "AMD RV770";
336 case CHIP_RV730: return "AMD RV730";
337 case CHIP_RV710: return "AMD RV710";
338 case CHIP_RV740: return "AMD RV740";
339 case CHIP_CEDAR: return "AMD CEDAR";
340 case CHIP_REDWOOD: return "AMD REDWOOD";
341 case CHIP_JUNIPER: return "AMD JUNIPER";
342 case CHIP_CYPRESS: return "AMD CYPRESS";
343 case CHIP_HEMLOCK: return "AMD HEMLOCK";
344 case CHIP_PALM: return "AMD PALM";
345 case CHIP_SUMO: return "AMD SUMO";
346 case CHIP_SUMO2: return "AMD SUMO2";
347 case CHIP_BARTS: return "AMD BARTS";
348 case CHIP_TURKS: return "AMD TURKS";
349 case CHIP_CAICOS: return "AMD CAICOS";
350 case CHIP_CAYMAN: return "AMD CAYMAN";
351 case CHIP_ARUBA: return "AMD ARUBA";
352 default: return "AMD unknown";
353 }
354 }
355
356 static const char* r600_get_name(struct pipe_screen* pscreen)
357 {
358 struct r600_screen *rscreen = (struct r600_screen *)pscreen;
359
360 return r600_get_family_name(rscreen->b.family);
361 }
362
363 static int r600_get_param(struct pipe_screen* pscreen, enum pipe_cap param)
364 {
365 struct r600_screen *rscreen = (struct r600_screen *)pscreen;
366 enum radeon_family family = rscreen->b.family;
367
368 switch (param) {
369 /* Supported features (boolean caps). */
370 case PIPE_CAP_NPOT_TEXTURES:
371 case PIPE_CAP_MIXED_FRAMEBUFFER_SIZES:
372 case PIPE_CAP_TWO_SIDED_STENCIL:
373 case PIPE_CAP_ANISOTROPIC_FILTER:
374 case PIPE_CAP_POINT_SPRITE:
375 case PIPE_CAP_OCCLUSION_QUERY:
376 case PIPE_CAP_TEXTURE_SHADOW_MAP:
377 case PIPE_CAP_TEXTURE_MIRROR_CLAMP:
378 case PIPE_CAP_BLEND_EQUATION_SEPARATE:
379 case PIPE_CAP_TEXTURE_SWIZZLE:
380 case PIPE_CAP_DEPTH_CLIP_DISABLE:
381 case PIPE_CAP_SHADER_STENCIL_EXPORT:
382 case PIPE_CAP_VERTEX_ELEMENT_INSTANCE_DIVISOR:
383 case PIPE_CAP_MIXED_COLORBUFFER_FORMATS:
384 case PIPE_CAP_TGSI_FS_COORD_ORIGIN_UPPER_LEFT:
385 case PIPE_CAP_TGSI_FS_COORD_PIXEL_CENTER_HALF_INTEGER:
386 case PIPE_CAP_SM3:
387 case PIPE_CAP_SEAMLESS_CUBE_MAP:
388 case PIPE_CAP_PRIMITIVE_RESTART:
389 case PIPE_CAP_CONDITIONAL_RENDER:
390 case PIPE_CAP_TEXTURE_BARRIER:
391 case PIPE_CAP_VERTEX_COLOR_UNCLAMPED:
392 case PIPE_CAP_QUADS_FOLLOW_PROVOKING_VERTEX_CONVENTION:
393 case PIPE_CAP_TGSI_INSTANCEID:
394 case PIPE_CAP_VERTEX_BUFFER_OFFSET_4BYTE_ALIGNED_ONLY:
395 case PIPE_CAP_VERTEX_BUFFER_STRIDE_4BYTE_ALIGNED_ONLY:
396 case PIPE_CAP_VERTEX_ELEMENT_SRC_OFFSET_4BYTE_ALIGNED_ONLY:
397 case PIPE_CAP_USER_INDEX_BUFFERS:
398 case PIPE_CAP_USER_CONSTANT_BUFFERS:
399 case PIPE_CAP_COMPUTE:
400 case PIPE_CAP_START_INSTANCE:
401 case PIPE_CAP_MAX_DUAL_SOURCE_RENDER_TARGETS:
402 case PIPE_CAP_TEXTURE_BUFFER_OBJECTS:
403 case PIPE_CAP_PREFER_BLIT_BASED_TEXTURE_TRANSFER:
404 case PIPE_CAP_QUERY_PIPELINE_STATISTICS:
405 case PIPE_CAP_TEXTURE_MULTISAMPLE:
406 return 1;
407
408 case PIPE_CAP_TGSI_TEXCOORD:
409 return 0;
410
411 case PIPE_CAP_MAX_TEXTURE_BUFFER_SIZE:
412 return MIN2(rscreen->b.info.vram_size, 0xFFFFFFFF);
413
414 case PIPE_CAP_MIN_MAP_BUFFER_ALIGNMENT:
415 return R600_MAP_BUFFER_ALIGNMENT;
416
417 case PIPE_CAP_CONSTANT_BUFFER_OFFSET_ALIGNMENT:
418 return 256;
419
420 case PIPE_CAP_TEXTURE_BUFFER_OFFSET_ALIGNMENT:
421 return 1;
422
423 case PIPE_CAP_GLSL_FEATURE_LEVEL:
424 return 140;
425
426 /* Supported except the original R600. */
427 case PIPE_CAP_INDEP_BLEND_ENABLE:
428 case PIPE_CAP_INDEP_BLEND_FUNC:
429 /* R600 doesn't support per-MRT blends */
430 return family == CHIP_R600 ? 0 : 1;
431
432 /* Supported on Evergreen. */
433 case PIPE_CAP_SEAMLESS_CUBE_MAP_PER_TEXTURE:
434 case PIPE_CAP_CUBE_MAP_ARRAY:
435 return family >= CHIP_CEDAR ? 1 : 0;
436
437 /* Unsupported features. */
438 case PIPE_CAP_TGSI_FS_COORD_ORIGIN_LOWER_LEFT:
439 case PIPE_CAP_TGSI_FS_COORD_PIXEL_CENTER_INTEGER:
440 case PIPE_CAP_SCALED_RESOLVE:
441 case PIPE_CAP_TGSI_CAN_COMPACT_CONSTANTS:
442 case PIPE_CAP_FRAGMENT_COLOR_CLAMPED:
443 case PIPE_CAP_VERTEX_COLOR_CLAMPED:
444 case PIPE_CAP_USER_VERTEX_BUFFERS:
445 case PIPE_CAP_TGSI_VS_LAYER:
446 return 0;
447
448 /* Stream output. */
449 case PIPE_CAP_MAX_STREAM_OUTPUT_BUFFERS:
450 return rscreen->b.has_streamout ? 4 : 0;
451 case PIPE_CAP_STREAM_OUTPUT_PAUSE_RESUME:
452 return rscreen->b.has_streamout ? 1 : 0;
453 case PIPE_CAP_MAX_STREAM_OUTPUT_SEPARATE_COMPONENTS:
454 case PIPE_CAP_MAX_STREAM_OUTPUT_INTERLEAVED_COMPONENTS:
455 return 32*4;
456
457 /* Texturing. */
458 case PIPE_CAP_MAX_TEXTURE_2D_LEVELS:
459 case PIPE_CAP_MAX_TEXTURE_3D_LEVELS:
460 case PIPE_CAP_MAX_TEXTURE_CUBE_LEVELS:
461 if (family >= CHIP_CEDAR)
462 return 15;
463 else
464 return 14;
465 case PIPE_CAP_MAX_TEXTURE_ARRAY_LAYERS:
466 return rscreen->b.info.drm_minor >= 9 ?
467 (family >= CHIP_CEDAR ? 16384 : 8192) : 0;
468 case PIPE_CAP_MAX_COMBINED_SAMPLERS:
469 return 32;
470
471 /* Render targets. */
472 case PIPE_CAP_MAX_RENDER_TARGETS:
473 /* XXX some r6xx are buggy and can only do 4 */
474 return 8;
475
476 case PIPE_CAP_MAX_VIEWPORTS:
477 return 1;
478
479 /* Timer queries, present when the clock frequency is non zero. */
480 case PIPE_CAP_QUERY_TIME_ELAPSED:
481 return rscreen->b.info.r600_clock_crystal_freq != 0;
482 case PIPE_CAP_QUERY_TIMESTAMP:
483 return rscreen->b.info.drm_minor >= 20 &&
484 rscreen->b.info.r600_clock_crystal_freq != 0;
485
486 case PIPE_CAP_MIN_TEXEL_OFFSET:
487 return -8;
488
489 case PIPE_CAP_MAX_TEXEL_OFFSET:
490 return 7;
491
492 case PIPE_CAP_TEXTURE_BORDER_COLOR_QUIRK:
493 return PIPE_QUIRK_TEXTURE_BORDER_COLOR_SWIZZLE_R600;
494 case PIPE_CAP_ENDIANNESS:
495 return PIPE_ENDIAN_LITTLE;
496 }
497 return 0;
498 }
499
500 static float r600_get_paramf(struct pipe_screen* pscreen,
501 enum pipe_capf param)
502 {
503 struct r600_screen *rscreen = (struct r600_screen *)pscreen;
504 enum radeon_family family = rscreen->b.family;
505
506 switch (param) {
507 case PIPE_CAPF_MAX_LINE_WIDTH:
508 case PIPE_CAPF_MAX_LINE_WIDTH_AA:
509 case PIPE_CAPF_MAX_POINT_WIDTH:
510 case PIPE_CAPF_MAX_POINT_WIDTH_AA:
511 if (family >= CHIP_CEDAR)
512 return 16384.0f;
513 else
514 return 8192.0f;
515 case PIPE_CAPF_MAX_TEXTURE_ANISOTROPY:
516 return 16.0f;
517 case PIPE_CAPF_MAX_TEXTURE_LOD_BIAS:
518 return 16.0f;
519 case PIPE_CAPF_GUARD_BAND_LEFT:
520 case PIPE_CAPF_GUARD_BAND_TOP:
521 case PIPE_CAPF_GUARD_BAND_RIGHT:
522 case PIPE_CAPF_GUARD_BAND_BOTTOM:
523 return 0.0f;
524 }
525 return 0.0f;
526 }
527
528 static int r600_get_shader_param(struct pipe_screen* pscreen, unsigned shader, enum pipe_shader_cap param)
529 {
530 switch(shader)
531 {
532 case PIPE_SHADER_FRAGMENT:
533 case PIPE_SHADER_VERTEX:
534 case PIPE_SHADER_COMPUTE:
535 break;
536 case PIPE_SHADER_GEOMETRY:
537 /* XXX: support and enable geometry programs */
538 return 0;
539 default:
540 /* XXX: support tessellation on Evergreen */
541 return 0;
542 }
543
544 switch (param) {
545 case PIPE_SHADER_CAP_MAX_INSTRUCTIONS:
546 case PIPE_SHADER_CAP_MAX_ALU_INSTRUCTIONS:
547 case PIPE_SHADER_CAP_MAX_TEX_INSTRUCTIONS:
548 case PIPE_SHADER_CAP_MAX_TEX_INDIRECTIONS:
549 return 16384;
550 case PIPE_SHADER_CAP_MAX_CONTROL_FLOW_DEPTH:
551 return 32;
552 case PIPE_SHADER_CAP_MAX_INPUTS:
553 return 32;
554 case PIPE_SHADER_CAP_MAX_TEMPS:
555 return 256; /* Max native temporaries. */
556 case PIPE_SHADER_CAP_MAX_ADDRS:
557 /* XXX Isn't this equal to TEMPS? */
558 return 1; /* Max native address registers */
559 case PIPE_SHADER_CAP_MAX_CONSTS:
560 return R600_MAX_CONST_BUFFER_SIZE;
561 case PIPE_SHADER_CAP_MAX_CONST_BUFFERS:
562 return R600_MAX_USER_CONST_BUFFERS;
563 case PIPE_SHADER_CAP_MAX_PREDS:
564 return 0; /* nothing uses this */
565 case PIPE_SHADER_CAP_TGSI_CONT_SUPPORTED:
566 return 1;
567 case PIPE_SHADER_CAP_TGSI_SQRT_SUPPORTED:
568 return 0;
569 case PIPE_SHADER_CAP_INDIRECT_INPUT_ADDR:
570 case PIPE_SHADER_CAP_INDIRECT_OUTPUT_ADDR:
571 case PIPE_SHADER_CAP_INDIRECT_TEMP_ADDR:
572 case PIPE_SHADER_CAP_INDIRECT_CONST_ADDR:
573 return 1;
574 case PIPE_SHADER_CAP_SUBROUTINES:
575 return 0;
576 case PIPE_SHADER_CAP_INTEGERS:
577 return 1;
578 case PIPE_SHADER_CAP_MAX_TEXTURE_SAMPLERS:
579 case PIPE_SHADER_CAP_MAX_SAMPLER_VIEWS:
580 return 16;
581 case PIPE_SHADER_CAP_PREFERRED_IR:
582 if (shader == PIPE_SHADER_COMPUTE) {
583 return PIPE_SHADER_IR_LLVM;
584 } else {
585 return PIPE_SHADER_IR_TGSI;
586 }
587 }
588 return 0;
589 }
590
591 static int r600_get_video_param(struct pipe_screen *screen,
592 enum pipe_video_profile profile,
593 enum pipe_video_entrypoint entrypoint,
594 enum pipe_video_cap param)
595 {
596 switch (param) {
597 case PIPE_VIDEO_CAP_SUPPORTED:
598 return vl_profile_supported(screen, profile, entrypoint);
599 case PIPE_VIDEO_CAP_NPOT_TEXTURES:
600 return 1;
601 case PIPE_VIDEO_CAP_MAX_WIDTH:
602 case PIPE_VIDEO_CAP_MAX_HEIGHT:
603 return vl_video_buffer_max_size(screen);
604 case PIPE_VIDEO_CAP_PREFERED_FORMAT:
605 return PIPE_FORMAT_NV12;
606 case PIPE_VIDEO_CAP_PREFERS_INTERLACED:
607 return false;
608 case PIPE_VIDEO_CAP_SUPPORTS_INTERLACED:
609 return false;
610 case PIPE_VIDEO_CAP_SUPPORTS_PROGRESSIVE:
611 return true;
612 case PIPE_VIDEO_CAP_MAX_LEVEL:
613 return vl_level_supported(screen, profile);
614 default:
615 return 0;
616 }
617 }
618
619 const char * r600_llvm_gpu_string(enum radeon_family family)
620 {
621 const char * gpu_family;
622
623 switch (family) {
624 case CHIP_R600:
625 case CHIP_RV630:
626 case CHIP_RV635:
627 case CHIP_RV670:
628 gpu_family = "r600";
629 break;
630 case CHIP_RV610:
631 case CHIP_RV620:
632 case CHIP_RS780:
633 case CHIP_RS880:
634 gpu_family = "rs880";
635 break;
636 case CHIP_RV710:
637 gpu_family = "rv710";
638 break;
639 case CHIP_RV730:
640 gpu_family = "rv730";
641 break;
642 case CHIP_RV740:
643 case CHIP_RV770:
644 gpu_family = "rv770";
645 break;
646 case CHIP_PALM:
647 case CHIP_CEDAR:
648 gpu_family = "cedar";
649 break;
650 case CHIP_SUMO:
651 case CHIP_SUMO2:
652 gpu_family = "sumo";
653 break;
654 case CHIP_REDWOOD:
655 gpu_family = "redwood";
656 break;
657 case CHIP_JUNIPER:
658 gpu_family = "juniper";
659 break;
660 case CHIP_HEMLOCK:
661 case CHIP_CYPRESS:
662 gpu_family = "cypress";
663 break;
664 case CHIP_BARTS:
665 gpu_family = "barts";
666 break;
667 case CHIP_TURKS:
668 gpu_family = "turks";
669 break;
670 case CHIP_CAICOS:
671 gpu_family = "caicos";
672 break;
673 case CHIP_CAYMAN:
674 case CHIP_ARUBA:
675 gpu_family = "cayman";
676 break;
677 default:
678 gpu_family = "";
679 fprintf(stderr, "Chip not supported by r600 llvm "
680 "backend, please file a bug at " PACKAGE_BUGREPORT "\n");
681 break;
682 }
683 return gpu_family;
684 }
685
686
687 static int r600_get_compute_param(struct pipe_screen *screen,
688 enum pipe_compute_cap param,
689 void *ret)
690 {
691 struct r600_screen *rscreen = (struct r600_screen *)screen;
692 //TODO: select these params by asic
693 switch (param) {
694 case PIPE_COMPUTE_CAP_IR_TARGET: {
695 const char *gpu = r600_llvm_gpu_string(rscreen->b.family);
696 if (ret) {
697 sprintf(ret, "%s-r600--", gpu);
698 }
699 return (8 + strlen(gpu)) * sizeof(char);
700 }
701 case PIPE_COMPUTE_CAP_GRID_DIMENSION:
702 if (ret) {
703 uint64_t * grid_dimension = ret;
704 grid_dimension[0] = 3;
705 }
706 return 1 * sizeof(uint64_t);
707
708 case PIPE_COMPUTE_CAP_MAX_GRID_SIZE:
709 if (ret) {
710 uint64_t * grid_size = ret;
711 grid_size[0] = 65535;
712 grid_size[1] = 65535;
713 grid_size[2] = 1;
714 }
715 return 3 * sizeof(uint64_t) ;
716
717 case PIPE_COMPUTE_CAP_MAX_BLOCK_SIZE:
718 if (ret) {
719 uint64_t * block_size = ret;
720 block_size[0] = 256;
721 block_size[1] = 256;
722 block_size[2] = 256;
723 }
724 return 3 * sizeof(uint64_t);
725
726 case PIPE_COMPUTE_CAP_MAX_THREADS_PER_BLOCK:
727 if (ret) {
728 uint64_t * max_threads_per_block = ret;
729 *max_threads_per_block = 256;
730 }
731 return sizeof(uint64_t);
732
733 case PIPE_COMPUTE_CAP_MAX_GLOBAL_SIZE:
734 if (ret) {
735 uint64_t * max_global_size = ret;
736 /* XXX: This is what the proprietary driver reports, we
737 * may want to use a different value. */
738 *max_global_size = 201326592;
739 }
740 return sizeof(uint64_t);
741
742 case PIPE_COMPUTE_CAP_MAX_INPUT_SIZE:
743 if (ret) {
744 uint64_t * max_input_size = ret;
745 *max_input_size = 1024;
746 }
747 return sizeof(uint64_t);
748
749 case PIPE_COMPUTE_CAP_MAX_LOCAL_SIZE:
750 if (ret) {
751 uint64_t * max_local_size = ret;
752 /* XXX: This is what the proprietary driver reports, we
753 * may want to use a different value. */
754 *max_local_size = 32768;
755 }
756 return sizeof(uint64_t);
757
758 case PIPE_COMPUTE_CAP_MAX_MEM_ALLOC_SIZE:
759 if (ret) {
760 uint64_t max_global_size;
761 uint64_t * max_mem_alloc_size = ret;
762 r600_get_compute_param(screen,
763 PIPE_COMPUTE_CAP_MAX_GLOBAL_SIZE,
764 &max_global_size);
765 /* OpenCL requres this value be at least
766 * max(MAX_GLOBAL_SIZE / 4, 128 * 1024 *1024)
767 * I'm really not sure what value to report here, but
768 * MAX_GLOBAL_SIZE / 4 seems resonable.
769 */
770 *max_mem_alloc_size = max_global_size / 4;
771 }
772 return sizeof(uint64_t);
773
774 default:
775 fprintf(stderr, "unknown PIPE_COMPUTE_CAP %d\n", param);
776 return 0;
777 }
778 }
779
780 static void r600_destroy_screen(struct pipe_screen* pscreen)
781 {
782 struct r600_screen *rscreen = (struct r600_screen *)pscreen;
783
784 if (rscreen == NULL)
785 return;
786
787 if (!radeon_winsys_unref(rscreen->b.ws))
788 return;
789
790 r600_common_screen_cleanup(&rscreen->b);
791
792 if (rscreen->global_pool) {
793 compute_memory_pool_delete(rscreen->global_pool);
794 }
795
796 if (rscreen->trace_bo) {
797 rscreen->b.ws->buffer_unmap(rscreen->trace_bo->cs_buf);
798 pipe_resource_reference((struct pipe_resource**)&rscreen->trace_bo, NULL);
799 }
800
801 rscreen->b.ws->destroy(rscreen->b.ws);
802 FREE(rscreen);
803 }
804
805 static uint64_t r600_get_timestamp(struct pipe_screen *screen)
806 {
807 struct r600_screen *rscreen = (struct r600_screen*)screen;
808
809 return 1000000 * rscreen->b.ws->query_value(rscreen->b.ws, RADEON_TIMESTAMP) /
810 rscreen->b.info.r600_clock_crystal_freq;
811 }
812
813 static int r600_get_driver_query_info(struct pipe_screen *screen,
814 unsigned index,
815 struct pipe_driver_query_info *info)
816 {
817 struct r600_screen *rscreen = (struct r600_screen*)screen;
818 struct pipe_driver_query_info list[] = {
819 {"draw-calls", R600_QUERY_DRAW_CALLS, 0},
820 {"requested-VRAM", R600_QUERY_REQUESTED_VRAM, rscreen->b.info.vram_size, TRUE},
821 {"requested-GTT", R600_QUERY_REQUESTED_GTT, rscreen->b.info.gart_size, TRUE},
822 {"buffer-wait-time", R600_QUERY_BUFFER_WAIT_TIME, 0, FALSE}
823 };
824
825 if (!info)
826 return Elements(list);
827
828 if (index >= Elements(list))
829 return 0;
830
831 *info = list[index];
832 return 1;
833 }
834
835 struct pipe_screen *r600_screen_create(struct radeon_winsys *ws)
836 {
837 struct r600_screen *rscreen = CALLOC_STRUCT(r600_screen);
838
839 if (rscreen == NULL) {
840 return NULL;
841 }
842
843 ws->query_info(ws, &rscreen->b.info);
844
845 /* Set functions first. */
846 rscreen->b.b.context_create = r600_create_context;
847 rscreen->b.b.destroy = r600_destroy_screen;
848 rscreen->b.b.get_name = r600_get_name;
849 rscreen->b.b.get_vendor = r600_get_vendor;
850 rscreen->b.b.get_param = r600_get_param;
851 rscreen->b.b.get_shader_param = r600_get_shader_param;
852 rscreen->b.b.get_paramf = r600_get_paramf;
853 rscreen->b.b.get_compute_param = r600_get_compute_param;
854 rscreen->b.b.get_timestamp = r600_get_timestamp;
855 if (rscreen->b.info.chip_class >= EVERGREEN) {
856 rscreen->b.b.is_format_supported = evergreen_is_format_supported;
857 } else {
858 rscreen->b.b.is_format_supported = r600_is_format_supported;
859 }
860 rscreen->b.b.get_driver_query_info = r600_get_driver_query_info;
861 if (rscreen->b.info.has_uvd) {
862 rscreen->b.b.get_video_param = ruvd_get_video_param;
863 rscreen->b.b.is_video_format_supported = ruvd_is_format_supported;
864 } else {
865 rscreen->b.b.get_video_param = r600_get_video_param;
866 rscreen->b.b.is_video_format_supported = vl_video_buffer_is_format_supported;
867 }
868 r600_init_screen_resource_functions(&rscreen->b.b);
869
870 if (!r600_common_screen_init(&rscreen->b, ws)) {
871 FREE(rscreen);
872 return NULL;
873 }
874
875 rscreen->b.debug_flags |= debug_get_flags_option("R600_DEBUG", r600_debug_options, 0);
876 if (debug_get_bool_option("R600_DEBUG_COMPUTE", FALSE))
877 rscreen->b.debug_flags |= DBG_COMPUTE;
878 if (debug_get_bool_option("R600_DUMP_SHADERS", FALSE))
879 rscreen->b.debug_flags |= DBG_FS | DBG_VS | DBG_GS | DBG_PS | DBG_CS;
880 if (!debug_get_bool_option("R600_HYPERZ", TRUE))
881 rscreen->b.debug_flags |= DBG_NO_HYPERZ;
882 if (!debug_get_bool_option("R600_LLVM", TRUE))
883 rscreen->b.debug_flags |= DBG_NO_LLVM;
884
885 if (rscreen->b.family == CHIP_UNKNOWN) {
886 fprintf(stderr, "r600: Unknown chipset 0x%04X\n", rscreen->b.info.pci_id);
887 FREE(rscreen);
888 return NULL;
889 }
890
891 /* Figure out streamout kernel support. */
892 switch (rscreen->b.chip_class) {
893 case R600:
894 if (rscreen->b.family < CHIP_RS780) {
895 rscreen->b.has_streamout = rscreen->b.info.drm_minor >= 14;
896 } else {
897 rscreen->b.has_streamout = rscreen->b.info.drm_minor >= 23;
898 }
899 break;
900 case R700:
901 rscreen->b.has_streamout = rscreen->b.info.drm_minor >= 17;
902 break;
903 case EVERGREEN:
904 case CAYMAN:
905 rscreen->b.has_streamout = rscreen->b.info.drm_minor >= 14;
906 break;
907 default:
908 rscreen->b.has_streamout = FALSE;
909 break;
910 }
911
912 /* MSAA support. */
913 switch (rscreen->b.chip_class) {
914 case R600:
915 case R700:
916 rscreen->has_msaa = rscreen->b.info.drm_minor >= 22;
917 rscreen->has_compressed_msaa_texturing = false;
918 break;
919 case EVERGREEN:
920 rscreen->has_msaa = rscreen->b.info.drm_minor >= 19;
921 rscreen->has_compressed_msaa_texturing = rscreen->b.info.drm_minor >= 24;
922 break;
923 case CAYMAN:
924 rscreen->has_msaa = rscreen->b.info.drm_minor >= 19;
925 rscreen->has_compressed_msaa_texturing = true;
926 break;
927 default:
928 rscreen->has_msaa = FALSE;
929 rscreen->has_compressed_msaa_texturing = false;
930 }
931
932 rscreen->b.has_cp_dma = rscreen->b.info.drm_minor >= 27 &&
933 !(rscreen->b.debug_flags & DBG_NO_CP_DMA);
934
935 rscreen->global_pool = compute_memory_pool_new(rscreen);
936
937 rscreen->cs_count = 0;
938 if (rscreen->b.info.drm_minor >= 28 && (rscreen->b.debug_flags & DBG_TRACE_CS)) {
939 rscreen->trace_bo = (struct r600_resource*)pipe_buffer_create(&rscreen->b.b,
940 PIPE_BIND_CUSTOM,
941 PIPE_USAGE_STAGING,
942 4096);
943 if (rscreen->trace_bo) {
944 rscreen->trace_ptr = rscreen->b.ws->buffer_map(rscreen->trace_bo->cs_buf, NULL,
945 PIPE_TRANSFER_UNSYNCHRONIZED);
946 }
947 }
948
949 /* Create the auxiliary context. This must be done last. */
950 rscreen->b.aux_context = rscreen->b.b.context_create(&rscreen->b.b, NULL);
951
952 #if 0 /* This is for testing whether aux_context and buffer clearing work correctly. */
953 struct pipe_resource templ = {};
954
955 templ.width0 = 4;
956 templ.height0 = 2048;
957 templ.depth0 = 1;
958 templ.array_size = 1;
959 templ.target = PIPE_TEXTURE_2D;
960 templ.format = PIPE_FORMAT_R8G8B8A8_UNORM;
961 templ.usage = PIPE_USAGE_STATIC;
962
963 struct r600_resource *res = r600_resource(rscreen->screen.resource_create(&rscreen->screen, &templ));
964 unsigned char *map = ws->buffer_map(res->cs_buf, NULL, PIPE_TRANSFER_WRITE);
965
966 memset(map, 0, 256);
967
968 r600_screen_clear_buffer(rscreen, &res->b.b, 4, 4, 0xCC);
969 r600_screen_clear_buffer(rscreen, &res->b.b, 8, 4, 0xDD);
970 r600_screen_clear_buffer(rscreen, &res->b.b, 12, 4, 0xEE);
971 r600_screen_clear_buffer(rscreen, &res->b.b, 20, 4, 0xFF);
972 r600_screen_clear_buffer(rscreen, &res->b.b, 32, 20, 0x87);
973
974 ws->buffer_wait(res->buf, RADEON_USAGE_WRITE);
975
976 int i;
977 for (i = 0; i < 256; i++) {
978 printf("%02X", map[i]);
979 if (i % 16 == 15)
980 printf("\n");
981 }
982 #endif
983
984 return &rscreen->b.b;
985 }