r600g: Fix build with --enable-opencl
[mesa.git] / src / gallium / drivers / r600 / r600_pipe.c
1 /*
2 * Copyright 2010 Jerome Glisse <glisse@freedesktop.org>
3 *
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * on the rights to use, copy, modify, merge, publish, distribute, sub
8 * license, and/or sell copies of the Software, and to permit persons to whom
9 * the Software is furnished to do so, subject to the following conditions:
10 *
11 * The above copyright notice and this permission notice (including the next
12 * paragraph) shall be included in all copies or substantial portions of the
13 * Software.
14 *
15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17 * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL
18 * THE AUTHOR(S) AND/OR THEIR SUPPLIERS BE LIABLE FOR ANY CLAIM,
19 * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
20 * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
21 * USE OR OTHER DEALINGS IN THE SOFTWARE.
22 */
23 #include "r600_pipe.h"
24 #include "r600_public.h"
25 #include "r600_isa.h"
26 #include "evergreen_compute.h"
27 #include "r600d.h"
28
29 #include <errno.h>
30 #include "pipe/p_shader_tokens.h"
31 #include "util/u_blitter.h"
32 #include "util/u_debug.h"
33 #include "util/u_format_s3tc.h"
34 #include "util/u_memory.h"
35 #include "util/u_simple_shaders.h"
36 #include "util/u_upload_mgr.h"
37 #include "util/u_math.h"
38 #include "vl/vl_decoder.h"
39 #include "vl/vl_video_buffer.h"
40 #include "radeon/radeon_uvd.h"
41 #include "os/os_time.h"
42
43 static const struct debug_named_value debug_options[] = {
44 /* logging */
45 { "texdepth", DBG_TEX_DEPTH, "Print texture depth info" },
46 { "compute", DBG_COMPUTE, "Print compute info" },
47 { "vm", DBG_VM, "Print virtual addresses when creating resources" },
48
49 /* shaders */
50 { "fs", DBG_FS, "Print fetch shaders" },
51 { "vs", DBG_VS, "Print vertex shaders" },
52 { "gs", DBG_GS, "Print geometry shaders" },
53 { "ps", DBG_PS, "Print pixel shaders" },
54 { "cs", DBG_CS, "Print compute shaders" },
55
56 /* features */
57 { "nohyperz", DBG_NO_HYPERZ, "Disable Hyper-Z" },
58 #if defined(R600_USE_LLVM)
59 { "nollvm", DBG_NO_LLVM, "Disable the LLVM shader compiler" },
60 #endif
61 { "nocpdma", DBG_NO_CP_DMA, "Disable CP DMA" },
62 { "nodma", DBG_NO_ASYNC_DMA, "Disable asynchronous DMA" },
63 /* GL uses the word INVALIDATE, gallium uses the word DISCARD */
64 { "noinvalrange", DBG_NO_DISCARD_RANGE, "Disable handling of INVALIDATE_RANGE map flags" },
65
66 DEBUG_NAMED_VALUE_END /* must be last */
67 };
68
69 /*
70 * pipe_context
71 */
72 static struct r600_fence *r600_create_fence(struct r600_context *rctx)
73 {
74 struct r600_screen *rscreen = rctx->screen;
75 struct r600_fence *fence = NULL;
76
77 pipe_mutex_lock(rscreen->fences.mutex);
78
79 if (!rscreen->fences.bo) {
80 /* Create the shared buffer object */
81 rscreen->fences.bo = (struct r600_resource*)
82 pipe_buffer_create(&rscreen->screen, PIPE_BIND_CUSTOM,
83 PIPE_USAGE_STAGING, 4096);
84 if (!rscreen->fences.bo) {
85 R600_ERR("r600: failed to create bo for fence objects\n");
86 goto out;
87 }
88 rscreen->fences.data = r600_buffer_mmap_sync_with_rings(rctx, rscreen->fences.bo, PIPE_TRANSFER_READ_WRITE);
89 }
90
91 if (!LIST_IS_EMPTY(&rscreen->fences.pool)) {
92 struct r600_fence *entry;
93
94 /* Try to find a freed fence that has been signalled */
95 LIST_FOR_EACH_ENTRY(entry, &rscreen->fences.pool, head) {
96 if (rscreen->fences.data[entry->index] != 0) {
97 LIST_DELINIT(&entry->head);
98 fence = entry;
99 break;
100 }
101 }
102 }
103
104 if (!fence) {
105 /* Allocate a new fence */
106 struct r600_fence_block *block;
107 unsigned index;
108
109 if ((rscreen->fences.next_index + 1) >= 1024) {
110 R600_ERR("r600: too many concurrent fences\n");
111 goto out;
112 }
113
114 index = rscreen->fences.next_index++;
115
116 if (!(index % FENCE_BLOCK_SIZE)) {
117 /* Allocate a new block */
118 block = CALLOC_STRUCT(r600_fence_block);
119 if (block == NULL)
120 goto out;
121
122 LIST_ADD(&block->head, &rscreen->fences.blocks);
123 } else {
124 block = LIST_ENTRY(struct r600_fence_block, rscreen->fences.blocks.next, head);
125 }
126
127 fence = &block->fences[index % FENCE_BLOCK_SIZE];
128 fence->index = index;
129 }
130
131 pipe_reference_init(&fence->reference, 1);
132
133 rscreen->fences.data[fence->index] = 0;
134 r600_context_emit_fence(rctx, rscreen->fences.bo, fence->index, 1);
135
136 /* Create a dummy BO so that fence_finish without a timeout can sleep waiting for completion */
137 fence->sleep_bo = (struct r600_resource*)
138 pipe_buffer_create(&rctx->screen->screen, PIPE_BIND_CUSTOM,
139 PIPE_USAGE_STAGING, 1);
140 /* Add the fence as a dummy relocation. */
141 r600_context_bo_reloc(rctx, &rctx->rings.gfx, fence->sleep_bo, RADEON_USAGE_READWRITE);
142
143 out:
144 pipe_mutex_unlock(rscreen->fences.mutex);
145 return fence;
146 }
147
148 static void r600_flush(struct pipe_context *ctx, unsigned flags)
149 {
150 struct r600_context *rctx = (struct r600_context *)ctx;
151 struct pipe_query *render_cond = NULL;
152 unsigned render_cond_mode = 0;
153
154 rctx->rings.gfx.flushing = true;
155 /* Disable render condition. */
156 if (rctx->current_render_cond) {
157 render_cond = rctx->current_render_cond;
158 render_cond_mode = rctx->current_render_cond_mode;
159 ctx->render_condition(ctx, NULL, 0);
160 }
161
162 r600_context_flush(rctx, flags);
163 rctx->rings.gfx.flushing = false;
164 r600_begin_new_cs(rctx);
165
166 /* Re-enable render condition. */
167 if (render_cond) {
168 ctx->render_condition(ctx, render_cond, render_cond_mode);
169 }
170 }
171
172 static void r600_flush_from_st(struct pipe_context *ctx,
173 struct pipe_fence_handle **fence,
174 enum pipe_flush_flags flags)
175 {
176 struct r600_context *rctx = (struct r600_context *)ctx;
177 struct r600_fence **rfence = (struct r600_fence**)fence;
178 unsigned fflags;
179
180 fflags = flags & PIPE_FLUSH_END_OF_FRAME ? RADEON_FLUSH_END_OF_FRAME : 0;
181 if (rfence) {
182 *rfence = r600_create_fence(rctx);
183 }
184 /* flush gfx & dma ring, order does not matter as only one can be live */
185 if (rctx->rings.dma.cs) {
186 rctx->rings.dma.flush(rctx, fflags);
187 }
188 rctx->rings.gfx.flush(rctx, fflags);
189 }
190
191 static void r600_flush_gfx_ring(void *ctx, unsigned flags)
192 {
193 r600_flush((struct pipe_context*)ctx, flags);
194 }
195
196 static void r600_flush_dma_ring(void *ctx, unsigned flags)
197 {
198 struct r600_context *rctx = (struct r600_context *)ctx;
199 struct radeon_winsys_cs *cs = rctx->rings.dma.cs;
200 unsigned padding_dw, i;
201
202 if (!cs->cdw) {
203 return;
204 }
205
206 /* Pad the DMA CS to a multiple of 8 dwords. */
207 padding_dw = 8 - cs->cdw % 8;
208 if (padding_dw < 8) {
209 for (i = 0; i < padding_dw; i++) {
210 cs->buf[cs->cdw++] = DMA_PACKET(DMA_PACKET_NOP, 0, 0, 0);
211 }
212 }
213
214 rctx->rings.dma.flushing = true;
215 rctx->ws->cs_flush(cs, flags);
216 rctx->rings.dma.flushing = false;
217 }
218
219 boolean r600_rings_is_buffer_referenced(struct r600_context *ctx,
220 struct radeon_winsys_cs_handle *buf,
221 enum radeon_bo_usage usage)
222 {
223 if (ctx->ws->cs_is_buffer_referenced(ctx->rings.gfx.cs, buf, usage)) {
224 return TRUE;
225 }
226 if (ctx->rings.dma.cs) {
227 if (ctx->ws->cs_is_buffer_referenced(ctx->rings.dma.cs, buf, usage)) {
228 return TRUE;
229 }
230 }
231 return FALSE;
232 }
233
234 void *r600_buffer_mmap_sync_with_rings(struct r600_context *ctx,
235 struct r600_resource *resource,
236 unsigned usage)
237 {
238 enum radeon_bo_usage rusage = RADEON_USAGE_READWRITE;
239 unsigned flags = 0;
240 bool sync_flush = TRUE;
241
242 if (usage & PIPE_TRANSFER_UNSYNCHRONIZED) {
243 return ctx->ws->buffer_map(resource->cs_buf, NULL, usage);
244 }
245
246 if (!(usage & PIPE_TRANSFER_WRITE)) {
247 /* have to wait for pending read */
248 rusage = RADEON_USAGE_WRITE;
249 }
250 if (usage & PIPE_TRANSFER_DONTBLOCK) {
251 flags |= RADEON_FLUSH_ASYNC;
252 }
253
254 if (ctx->ws->cs_is_buffer_referenced(ctx->rings.gfx.cs, resource->cs_buf, rusage) && ctx->rings.gfx.cs->cdw) {
255 ctx->rings.gfx.flush(ctx, flags);
256 if (usage & PIPE_TRANSFER_DONTBLOCK) {
257 return NULL;
258 }
259 }
260 if (ctx->rings.dma.cs) {
261 if (ctx->ws->cs_is_buffer_referenced(ctx->rings.dma.cs, resource->cs_buf, rusage) && ctx->rings.dma.cs->cdw) {
262 ctx->rings.dma.flush(ctx, flags);
263 if (usage & PIPE_TRANSFER_DONTBLOCK) {
264 return NULL;
265 }
266 }
267 }
268
269 if (usage & PIPE_TRANSFER_DONTBLOCK) {
270 if (ctx->ws->buffer_is_busy(resource->buf, rusage)) {
271 return NULL;
272 }
273 }
274 if (sync_flush) {
275 /* Try to avoid busy-waiting in radeon_bo_wait. */
276 ctx->ws->cs_sync_flush(ctx->rings.gfx.cs);
277 if (ctx->rings.dma.cs) {
278 ctx->ws->cs_sync_flush(ctx->rings.dma.cs);
279 }
280 }
281 ctx->ws->buffer_wait(resource->buf, rusage);
282
283 /* at this point everything is synchronized */
284 return ctx->ws->buffer_map(resource->cs_buf, NULL, usage | PIPE_TRANSFER_UNSYNCHRONIZED);
285 }
286
287 static void r600_flush_from_winsys(void *ctx, unsigned flags)
288 {
289 struct r600_context *rctx = (struct r600_context *)ctx;
290
291 rctx->rings.gfx.flush(rctx, flags);
292 }
293
294 static void r600_flush_dma_from_winsys(void *ctx, unsigned flags)
295 {
296 struct r600_context *rctx = (struct r600_context *)ctx;
297
298 rctx->rings.dma.flush(rctx, flags);
299 }
300
301 static void r600_destroy_context(struct pipe_context *context)
302 {
303 struct r600_context *rctx = (struct r600_context *)context;
304
305 r600_isa_destroy(rctx->isa);
306
307 pipe_resource_reference((struct pipe_resource**)&rctx->dummy_cmask, NULL);
308 pipe_resource_reference((struct pipe_resource**)&rctx->dummy_fmask, NULL);
309
310 if (rctx->dummy_pixel_shader) {
311 rctx->context.delete_fs_state(&rctx->context, rctx->dummy_pixel_shader);
312 }
313 if (rctx->custom_dsa_flush) {
314 rctx->context.delete_depth_stencil_alpha_state(&rctx->context, rctx->custom_dsa_flush);
315 }
316 if (rctx->custom_blend_resolve) {
317 rctx->context.delete_blend_state(&rctx->context, rctx->custom_blend_resolve);
318 }
319 if (rctx->custom_blend_decompress) {
320 rctx->context.delete_blend_state(&rctx->context, rctx->custom_blend_decompress);
321 }
322 if (rctx->custom_blend_fmask_decompress) {
323 rctx->context.delete_blend_state(&rctx->context, rctx->custom_blend_fmask_decompress);
324 }
325 util_unreference_framebuffer_state(&rctx->framebuffer.state);
326
327 if (rctx->blitter) {
328 util_blitter_destroy(rctx->blitter);
329 }
330 if (rctx->uploader) {
331 u_upload_destroy(rctx->uploader);
332 }
333 if (rctx->allocator_so_filled_size) {
334 u_suballocator_destroy(rctx->allocator_so_filled_size);
335 }
336 if (rctx->allocator_fetch_shader) {
337 u_suballocator_destroy(rctx->allocator_fetch_shader);
338 }
339 util_slab_destroy(&rctx->pool_transfers);
340
341 r600_release_command_buffer(&rctx->start_cs_cmd);
342
343 if (rctx->rings.gfx.cs) {
344 rctx->ws->cs_destroy(rctx->rings.gfx.cs);
345 }
346 if (rctx->rings.dma.cs) {
347 rctx->ws->cs_destroy(rctx->rings.dma.cs);
348 }
349
350 FREE(rctx);
351 }
352
353 static struct pipe_context *r600_create_context(struct pipe_screen *screen, void *priv)
354 {
355 struct r600_context *rctx = CALLOC_STRUCT(r600_context);
356 struct r600_screen* rscreen = (struct r600_screen *)screen;
357
358 if (rctx == NULL)
359 return NULL;
360
361 util_slab_create(&rctx->pool_transfers,
362 sizeof(struct r600_transfer), 64,
363 UTIL_SLAB_SINGLETHREADED);
364
365 rctx->context.screen = screen;
366 rctx->context.priv = priv;
367 rctx->context.destroy = r600_destroy_context;
368 rctx->context.flush = r600_flush_from_st;
369
370 /* Easy accessing of screen/winsys. */
371 rctx->screen = rscreen;
372 rctx->ws = rscreen->ws;
373 rctx->family = rscreen->family;
374 rctx->chip_class = rscreen->chip_class;
375 rctx->keep_tiling_flags = rscreen->info.drm_minor >= 12;
376
377 LIST_INITHEAD(&rctx->active_nontimer_queries);
378
379 r600_init_blit_functions(rctx);
380 r600_init_query_functions(rctx);
381 r600_init_context_resource_functions(rctx);
382 r600_init_surface_functions(rctx);
383
384 if (rscreen->info.has_uvd) {
385 rctx->context.create_video_decoder = r600_uvd_create_decoder;
386 rctx->context.create_video_buffer = r600_video_buffer_create;
387 } else {
388 rctx->context.create_video_decoder = vl_create_decoder;
389 rctx->context.create_video_buffer = vl_video_buffer_create;
390 }
391
392 r600_init_common_state_functions(rctx);
393
394 switch (rctx->chip_class) {
395 case R600:
396 case R700:
397 r600_init_state_functions(rctx);
398 r600_init_atom_start_cs(rctx);
399 rctx->max_db = 4;
400 rctx->custom_dsa_flush = r600_create_db_flush_dsa(rctx);
401 rctx->custom_blend_resolve = rctx->chip_class == R700 ? r700_create_resolve_blend(rctx)
402 : r600_create_resolve_blend(rctx);
403 rctx->custom_blend_decompress = r600_create_decompress_blend(rctx);
404 rctx->has_vertex_cache = !(rctx->family == CHIP_RV610 ||
405 rctx->family == CHIP_RV620 ||
406 rctx->family == CHIP_RS780 ||
407 rctx->family == CHIP_RS880 ||
408 rctx->family == CHIP_RV710);
409 break;
410 case EVERGREEN:
411 case CAYMAN:
412 evergreen_init_state_functions(rctx);
413 evergreen_init_atom_start_cs(rctx);
414 evergreen_init_atom_start_compute_cs(rctx);
415 rctx->max_db = 8;
416 rctx->custom_dsa_flush = evergreen_create_db_flush_dsa(rctx);
417 rctx->custom_blend_resolve = evergreen_create_resolve_blend(rctx);
418 rctx->custom_blend_decompress = evergreen_create_decompress_blend(rctx);
419 rctx->custom_blend_fmask_decompress = evergreen_create_fmask_decompress_blend(rctx);
420 rctx->has_vertex_cache = !(rctx->family == CHIP_CEDAR ||
421 rctx->family == CHIP_PALM ||
422 rctx->family == CHIP_SUMO ||
423 rctx->family == CHIP_SUMO2 ||
424 rctx->family == CHIP_CAICOS ||
425 rctx->family == CHIP_CAYMAN ||
426 rctx->family == CHIP_ARUBA);
427 break;
428 default:
429 R600_ERR("Unsupported chip class %d.\n", rctx->chip_class);
430 goto fail;
431 }
432
433 rctx->rings.gfx.cs = rctx->ws->cs_create(rctx->ws, RING_GFX);
434 rctx->rings.gfx.flush = r600_flush_gfx_ring;
435 rctx->ws->cs_set_flush_callback(rctx->rings.gfx.cs, r600_flush_from_winsys, rctx);
436 rctx->rings.gfx.flushing = false;
437
438 rctx->rings.dma.cs = NULL;
439 if (rscreen->info.r600_has_dma && !(rscreen->debug_flags & DBG_NO_ASYNC_DMA)) {
440 rctx->rings.dma.cs = rctx->ws->cs_create(rctx->ws, RING_DMA);
441 rctx->rings.dma.flush = r600_flush_dma_ring;
442 rctx->ws->cs_set_flush_callback(rctx->rings.dma.cs, r600_flush_dma_from_winsys, rctx);
443 rctx->rings.dma.flushing = false;
444 }
445
446 rctx->uploader = u_upload_create(&rctx->context, 1024 * 1024, 256,
447 PIPE_BIND_INDEX_BUFFER |
448 PIPE_BIND_CONSTANT_BUFFER);
449 if (!rctx->uploader)
450 goto fail;
451
452 rctx->allocator_fetch_shader = u_suballocator_create(&rctx->context, 64 * 1024, 256,
453 0, PIPE_USAGE_STATIC, FALSE);
454 if (!rctx->allocator_fetch_shader)
455 goto fail;
456
457 rctx->allocator_so_filled_size = u_suballocator_create(&rctx->context, 4096, 4,
458 0, PIPE_USAGE_STATIC, TRUE);
459 if (!rctx->allocator_so_filled_size)
460 goto fail;
461
462 rctx->isa = calloc(1, sizeof(struct r600_isa));
463 if (!rctx->isa || r600_isa_init(rctx, rctx->isa))
464 goto fail;
465
466 rctx->blitter = util_blitter_create(&rctx->context);
467 if (rctx->blitter == NULL)
468 goto fail;
469 util_blitter_set_texture_multisample(rctx->blitter, rscreen->has_msaa);
470 rctx->blitter->draw_rectangle = r600_draw_rectangle;
471
472 r600_begin_new_cs(rctx);
473 r600_get_backend_mask(rctx); /* this emits commands and must be last */
474
475 rctx->dummy_pixel_shader =
476 util_make_fragment_cloneinput_shader(&rctx->context, 0,
477 TGSI_SEMANTIC_GENERIC,
478 TGSI_INTERPOLATE_CONSTANT);
479 rctx->context.bind_fs_state(&rctx->context, rctx->dummy_pixel_shader);
480
481 return &rctx->context;
482
483 fail:
484 r600_destroy_context(&rctx->context);
485 return NULL;
486 }
487
488 /*
489 * pipe_screen
490 */
491 static const char* r600_get_vendor(struct pipe_screen* pscreen)
492 {
493 return "X.Org";
494 }
495
496 static const char *r600_get_family_name(enum radeon_family family)
497 {
498 switch(family) {
499 case CHIP_R600: return "AMD R600";
500 case CHIP_RV610: return "AMD RV610";
501 case CHIP_RV630: return "AMD RV630";
502 case CHIP_RV670: return "AMD RV670";
503 case CHIP_RV620: return "AMD RV620";
504 case CHIP_RV635: return "AMD RV635";
505 case CHIP_RS780: return "AMD RS780";
506 case CHIP_RS880: return "AMD RS880";
507 case CHIP_RV770: return "AMD RV770";
508 case CHIP_RV730: return "AMD RV730";
509 case CHIP_RV710: return "AMD RV710";
510 case CHIP_RV740: return "AMD RV740";
511 case CHIP_CEDAR: return "AMD CEDAR";
512 case CHIP_REDWOOD: return "AMD REDWOOD";
513 case CHIP_JUNIPER: return "AMD JUNIPER";
514 case CHIP_CYPRESS: return "AMD CYPRESS";
515 case CHIP_HEMLOCK: return "AMD HEMLOCK";
516 case CHIP_PALM: return "AMD PALM";
517 case CHIP_SUMO: return "AMD SUMO";
518 case CHIP_SUMO2: return "AMD SUMO2";
519 case CHIP_BARTS: return "AMD BARTS";
520 case CHIP_TURKS: return "AMD TURKS";
521 case CHIP_CAICOS: return "AMD CAICOS";
522 case CHIP_CAYMAN: return "AMD CAYMAN";
523 case CHIP_ARUBA: return "AMD ARUBA";
524 default: return "AMD unknown";
525 }
526 }
527
528 static const char* r600_get_name(struct pipe_screen* pscreen)
529 {
530 struct r600_screen *rscreen = (struct r600_screen *)pscreen;
531
532 return r600_get_family_name(rscreen->family);
533 }
534
535 static int r600_get_param(struct pipe_screen* pscreen, enum pipe_cap param)
536 {
537 struct r600_screen *rscreen = (struct r600_screen *)pscreen;
538 enum radeon_family family = rscreen->family;
539
540 switch (param) {
541 /* Supported features (boolean caps). */
542 case PIPE_CAP_NPOT_TEXTURES:
543 case PIPE_CAP_TWO_SIDED_STENCIL:
544 case PIPE_CAP_ANISOTROPIC_FILTER:
545 case PIPE_CAP_POINT_SPRITE:
546 case PIPE_CAP_OCCLUSION_QUERY:
547 case PIPE_CAP_TEXTURE_SHADOW_MAP:
548 case PIPE_CAP_TEXTURE_MIRROR_CLAMP:
549 case PIPE_CAP_BLEND_EQUATION_SEPARATE:
550 case PIPE_CAP_TEXTURE_SWIZZLE:
551 case PIPE_CAP_DEPTH_CLIP_DISABLE:
552 case PIPE_CAP_SHADER_STENCIL_EXPORT:
553 case PIPE_CAP_VERTEX_ELEMENT_INSTANCE_DIVISOR:
554 case PIPE_CAP_MIXED_COLORBUFFER_FORMATS:
555 case PIPE_CAP_TGSI_FS_COORD_ORIGIN_UPPER_LEFT:
556 case PIPE_CAP_TGSI_FS_COORD_PIXEL_CENTER_HALF_INTEGER:
557 case PIPE_CAP_SM3:
558 case PIPE_CAP_SEAMLESS_CUBE_MAP:
559 case PIPE_CAP_PRIMITIVE_RESTART:
560 case PIPE_CAP_CONDITIONAL_RENDER:
561 case PIPE_CAP_TEXTURE_BARRIER:
562 case PIPE_CAP_VERTEX_COLOR_UNCLAMPED:
563 case PIPE_CAP_QUADS_FOLLOW_PROVOKING_VERTEX_CONVENTION:
564 case PIPE_CAP_TGSI_INSTANCEID:
565 case PIPE_CAP_VERTEX_BUFFER_OFFSET_4BYTE_ALIGNED_ONLY:
566 case PIPE_CAP_VERTEX_BUFFER_STRIDE_4BYTE_ALIGNED_ONLY:
567 case PIPE_CAP_VERTEX_ELEMENT_SRC_OFFSET_4BYTE_ALIGNED_ONLY:
568 case PIPE_CAP_USER_INDEX_BUFFERS:
569 case PIPE_CAP_USER_CONSTANT_BUFFERS:
570 case PIPE_CAP_COMPUTE:
571 case PIPE_CAP_START_INSTANCE:
572 case PIPE_CAP_MAX_DUAL_SOURCE_RENDER_TARGETS:
573 case PIPE_CAP_TEXTURE_BUFFER_OBJECTS:
574 case PIPE_CAP_PREFER_BLIT_BASED_TEXTURE_TRANSFER:
575 case PIPE_CAP_QUERY_PIPELINE_STATISTICS:
576 return 1;
577 case PIPE_CAP_TGSI_TEXCOORD:
578 return 0;
579
580 case PIPE_CAP_MIN_MAP_BUFFER_ALIGNMENT:
581 return R600_MAP_BUFFER_ALIGNMENT;
582
583 case PIPE_CAP_CONSTANT_BUFFER_OFFSET_ALIGNMENT:
584 return 256;
585
586 case PIPE_CAP_TEXTURE_BUFFER_OFFSET_ALIGNMENT:
587 return 1;
588
589 case PIPE_CAP_GLSL_FEATURE_LEVEL:
590 return 140;
591
592 case PIPE_CAP_TEXTURE_MULTISAMPLE:
593 return rscreen->msaa_texture_support != MSAA_TEXTURE_SAMPLE_ZERO;
594
595 /* Supported except the original R600. */
596 case PIPE_CAP_INDEP_BLEND_ENABLE:
597 case PIPE_CAP_INDEP_BLEND_FUNC:
598 /* R600 doesn't support per-MRT blends */
599 return family == CHIP_R600 ? 0 : 1;
600
601 /* Supported on Evergreen. */
602 case PIPE_CAP_SEAMLESS_CUBE_MAP_PER_TEXTURE:
603 case PIPE_CAP_CUBE_MAP_ARRAY:
604 return family >= CHIP_CEDAR ? 1 : 0;
605
606 /* Unsupported features. */
607 case PIPE_CAP_TGSI_FS_COORD_ORIGIN_LOWER_LEFT:
608 case PIPE_CAP_TGSI_FS_COORD_PIXEL_CENTER_INTEGER:
609 case PIPE_CAP_SCALED_RESOLVE:
610 case PIPE_CAP_TGSI_CAN_COMPACT_VARYINGS:
611 case PIPE_CAP_TGSI_CAN_COMPACT_CONSTANTS:
612 case PIPE_CAP_FRAGMENT_COLOR_CLAMPED:
613 case PIPE_CAP_VERTEX_COLOR_CLAMPED:
614 case PIPE_CAP_USER_VERTEX_BUFFERS:
615 return 0;
616
617 /* Stream output. */
618 case PIPE_CAP_MAX_STREAM_OUTPUT_BUFFERS:
619 return rscreen->has_streamout ? 4 : 0;
620 case PIPE_CAP_STREAM_OUTPUT_PAUSE_RESUME:
621 return rscreen->has_streamout ? 1 : 0;
622 case PIPE_CAP_MAX_STREAM_OUTPUT_SEPARATE_COMPONENTS:
623 case PIPE_CAP_MAX_STREAM_OUTPUT_INTERLEAVED_COMPONENTS:
624 return 32*4;
625
626 /* Texturing. */
627 case PIPE_CAP_MAX_TEXTURE_2D_LEVELS:
628 case PIPE_CAP_MAX_TEXTURE_3D_LEVELS:
629 case PIPE_CAP_MAX_TEXTURE_CUBE_LEVELS:
630 if (family >= CHIP_CEDAR)
631 return 15;
632 else
633 return 14;
634 case PIPE_CAP_MAX_TEXTURE_ARRAY_LAYERS:
635 return rscreen->info.drm_minor >= 9 ?
636 (family >= CHIP_CEDAR ? 16384 : 8192) : 0;
637 case PIPE_CAP_MAX_COMBINED_SAMPLERS:
638 return 32;
639
640 /* Render targets. */
641 case PIPE_CAP_MAX_RENDER_TARGETS:
642 /* XXX some r6xx are buggy and can only do 4 */
643 return 8;
644
645 /* Timer queries, present when the clock frequency is non zero. */
646 case PIPE_CAP_QUERY_TIME_ELAPSED:
647 return rscreen->info.r600_clock_crystal_freq != 0;
648 case PIPE_CAP_QUERY_TIMESTAMP:
649 return rscreen->info.drm_minor >= 20 &&
650 rscreen->info.r600_clock_crystal_freq != 0;
651
652 case PIPE_CAP_MIN_TEXEL_OFFSET:
653 return -8;
654
655 case PIPE_CAP_MAX_TEXEL_OFFSET:
656 return 7;
657 }
658 return 0;
659 }
660
661 static float r600_get_paramf(struct pipe_screen* pscreen,
662 enum pipe_capf param)
663 {
664 struct r600_screen *rscreen = (struct r600_screen *)pscreen;
665 enum radeon_family family = rscreen->family;
666
667 switch (param) {
668 case PIPE_CAPF_MAX_LINE_WIDTH:
669 case PIPE_CAPF_MAX_LINE_WIDTH_AA:
670 case PIPE_CAPF_MAX_POINT_WIDTH:
671 case PIPE_CAPF_MAX_POINT_WIDTH_AA:
672 if (family >= CHIP_CEDAR)
673 return 16384.0f;
674 else
675 return 8192.0f;
676 case PIPE_CAPF_MAX_TEXTURE_ANISOTROPY:
677 return 16.0f;
678 case PIPE_CAPF_MAX_TEXTURE_LOD_BIAS:
679 return 16.0f;
680 case PIPE_CAPF_GUARD_BAND_LEFT:
681 case PIPE_CAPF_GUARD_BAND_TOP:
682 case PIPE_CAPF_GUARD_BAND_RIGHT:
683 case PIPE_CAPF_GUARD_BAND_BOTTOM:
684 return 0.0f;
685 }
686 return 0.0f;
687 }
688
689 static int r600_get_shader_param(struct pipe_screen* pscreen, unsigned shader, enum pipe_shader_cap param)
690 {
691 switch(shader)
692 {
693 case PIPE_SHADER_FRAGMENT:
694 case PIPE_SHADER_VERTEX:
695 case PIPE_SHADER_COMPUTE:
696 break;
697 case PIPE_SHADER_GEOMETRY:
698 /* XXX: support and enable geometry programs */
699 return 0;
700 default:
701 /* XXX: support tessellation on Evergreen */
702 return 0;
703 }
704
705 switch (param) {
706 case PIPE_SHADER_CAP_MAX_INSTRUCTIONS:
707 case PIPE_SHADER_CAP_MAX_ALU_INSTRUCTIONS:
708 case PIPE_SHADER_CAP_MAX_TEX_INSTRUCTIONS:
709 case PIPE_SHADER_CAP_MAX_TEX_INDIRECTIONS:
710 return 16384;
711 case PIPE_SHADER_CAP_MAX_CONTROL_FLOW_DEPTH:
712 return 32;
713 case PIPE_SHADER_CAP_MAX_INPUTS:
714 return 32;
715 case PIPE_SHADER_CAP_MAX_TEMPS:
716 return 256; /* Max native temporaries. */
717 case PIPE_SHADER_CAP_MAX_ADDRS:
718 /* XXX Isn't this equal to TEMPS? */
719 return 1; /* Max native address registers */
720 case PIPE_SHADER_CAP_MAX_CONSTS:
721 return R600_MAX_CONST_BUFFER_SIZE;
722 case PIPE_SHADER_CAP_MAX_CONST_BUFFERS:
723 return R600_MAX_USER_CONST_BUFFERS;
724 case PIPE_SHADER_CAP_MAX_PREDS:
725 return 0; /* nothing uses this */
726 case PIPE_SHADER_CAP_TGSI_CONT_SUPPORTED:
727 return 1;
728 case PIPE_SHADER_CAP_TGSI_SQRT_SUPPORTED:
729 return 0;
730 case PIPE_SHADER_CAP_INDIRECT_INPUT_ADDR:
731 case PIPE_SHADER_CAP_INDIRECT_OUTPUT_ADDR:
732 case PIPE_SHADER_CAP_INDIRECT_TEMP_ADDR:
733 case PIPE_SHADER_CAP_INDIRECT_CONST_ADDR:
734 return 1;
735 case PIPE_SHADER_CAP_SUBROUTINES:
736 return 0;
737 case PIPE_SHADER_CAP_INTEGERS:
738 return 1;
739 case PIPE_SHADER_CAP_MAX_TEXTURE_SAMPLERS:
740 return 16;
741 case PIPE_SHADER_CAP_PREFERRED_IR:
742 if (shader == PIPE_SHADER_COMPUTE) {
743 return PIPE_SHADER_IR_LLVM;
744 } else {
745 return PIPE_SHADER_IR_TGSI;
746 }
747 }
748 return 0;
749 }
750
751 static int r600_get_video_param(struct pipe_screen *screen,
752 enum pipe_video_profile profile,
753 enum pipe_video_cap param)
754 {
755 switch (param) {
756 case PIPE_VIDEO_CAP_SUPPORTED:
757 return vl_profile_supported(screen, profile);
758 case PIPE_VIDEO_CAP_NPOT_TEXTURES:
759 return 1;
760 case PIPE_VIDEO_CAP_MAX_WIDTH:
761 case PIPE_VIDEO_CAP_MAX_HEIGHT:
762 return vl_video_buffer_max_size(screen);
763 case PIPE_VIDEO_CAP_PREFERED_FORMAT:
764 return PIPE_FORMAT_NV12;
765 case PIPE_VIDEO_CAP_PREFERS_INTERLACED:
766 return false;
767 case PIPE_VIDEO_CAP_SUPPORTS_INTERLACED:
768 return false;
769 case PIPE_VIDEO_CAP_SUPPORTS_PROGRESSIVE:
770 return true;
771 default:
772 return 0;
773 }
774 }
775
776 const char * r600_llvm_gpu_string(enum radeon_family family)
777 {
778 const char * gpu_family;
779
780 switch (family) {
781 case CHIP_R600:
782 case CHIP_RV610:
783 case CHIP_RV630:
784 case CHIP_RV620:
785 case CHIP_RV635:
786 case CHIP_RV670:
787 case CHIP_RS780:
788 case CHIP_RS880:
789 gpu_family = "r600";
790 break;
791 case CHIP_RV710:
792 gpu_family = "rv710";
793 break;
794 case CHIP_RV730:
795 gpu_family = "rv730";
796 break;
797 case CHIP_RV740:
798 case CHIP_RV770:
799 gpu_family = "rv770";
800 break;
801 case CHIP_PALM:
802 case CHIP_CEDAR:
803 gpu_family = "cedar";
804 break;
805 case CHIP_SUMO:
806 case CHIP_SUMO2:
807 case CHIP_REDWOOD:
808 gpu_family = "redwood";
809 break;
810 case CHIP_JUNIPER:
811 gpu_family = "juniper";
812 break;
813 case CHIP_HEMLOCK:
814 case CHIP_CYPRESS:
815 gpu_family = "cypress";
816 break;
817 case CHIP_BARTS:
818 gpu_family = "barts";
819 break;
820 case CHIP_TURKS:
821 gpu_family = "turks";
822 break;
823 case CHIP_CAICOS:
824 gpu_family = "caicos";
825 break;
826 case CHIP_CAYMAN:
827 case CHIP_ARUBA:
828 gpu_family = "cayman";
829 break;
830 default:
831 gpu_family = "";
832 fprintf(stderr, "Chip not supported by r600 llvm "
833 "backend, please file a bug at " PACKAGE_BUGREPORT "\n");
834 break;
835 }
836 return gpu_family;
837 }
838
839
840 static int r600_get_compute_param(struct pipe_screen *screen,
841 enum pipe_compute_cap param,
842 void *ret)
843 {
844 struct r600_screen *rscreen = (struct r600_screen *)screen;
845 //TODO: select these params by asic
846 switch (param) {
847 case PIPE_COMPUTE_CAP_IR_TARGET: {
848 const char *gpu = r600_llvm_gpu_string(rscreen->family);
849 if (ret) {
850 sprintf(ret, "%s-r600--", gpu);
851 }
852 return (8 + strlen(gpu)) * sizeof(char);
853 }
854 case PIPE_COMPUTE_CAP_GRID_DIMENSION:
855 if (ret) {
856 uint64_t * grid_dimension = ret;
857 grid_dimension[0] = 3;
858 }
859 return 1 * sizeof(uint64_t);
860
861 case PIPE_COMPUTE_CAP_MAX_GRID_SIZE:
862 if (ret) {
863 uint64_t * grid_size = ret;
864 grid_size[0] = 65535;
865 grid_size[1] = 65535;
866 grid_size[2] = 1;
867 }
868 return 3 * sizeof(uint64_t) ;
869
870 case PIPE_COMPUTE_CAP_MAX_BLOCK_SIZE:
871 if (ret) {
872 uint64_t * block_size = ret;
873 block_size[0] = 256;
874 block_size[1] = 256;
875 block_size[2] = 256;
876 }
877 return 3 * sizeof(uint64_t);
878
879 case PIPE_COMPUTE_CAP_MAX_THREADS_PER_BLOCK:
880 if (ret) {
881 uint64_t * max_threads_per_block = ret;
882 *max_threads_per_block = 256;
883 }
884 return sizeof(uint64_t);
885
886 case PIPE_COMPUTE_CAP_MAX_GLOBAL_SIZE:
887 if (ret) {
888 uint64_t * max_global_size = ret;
889 /* XXX: This is what the proprietary driver reports, we
890 * may want to use a different value. */
891 *max_global_size = 201326592;
892 }
893 return sizeof(uint64_t);
894
895 case PIPE_COMPUTE_CAP_MAX_INPUT_SIZE:
896 if (ret) {
897 uint64_t * max_input_size = ret;
898 *max_input_size = 1024;
899 }
900 return sizeof(uint64_t);
901
902 case PIPE_COMPUTE_CAP_MAX_LOCAL_SIZE:
903 if (ret) {
904 uint64_t * max_local_size = ret;
905 /* XXX: This is what the proprietary driver reports, we
906 * may want to use a different value. */
907 *max_local_size = 32768;
908 }
909 return sizeof(uint64_t);
910
911 case PIPE_COMPUTE_CAP_MAX_MEM_ALLOC_SIZE:
912 if (ret) {
913 uint64_t max_global_size;
914 uint64_t * max_mem_alloc_size = ret;
915 r600_get_compute_param(screen,
916 PIPE_COMPUTE_CAP_MAX_GLOBAL_SIZE,
917 &max_global_size);
918 /* OpenCL requres this value be at least
919 * max(MAX_GLOBAL_SIZE / 4, 128 * 1024 *1024)
920 * I'm really not sure what value to report here, but
921 * MAX_GLOBAL_SIZE / 4 seems resonable.
922 */
923 *max_mem_alloc_size = max_global_size / 4;
924 }
925 return sizeof(uint64_t);
926
927 default:
928 fprintf(stderr, "unknown PIPE_COMPUTE_CAP %d\n", param);
929 return 0;
930 }
931 }
932
933 static void r600_destroy_screen(struct pipe_screen* pscreen)
934 {
935 struct r600_screen *rscreen = (struct r600_screen *)pscreen;
936
937 if (rscreen == NULL)
938 return;
939
940 if (rscreen->global_pool) {
941 compute_memory_pool_delete(rscreen->global_pool);
942 }
943
944 if (rscreen->fences.bo) {
945 struct r600_fence_block *entry, *tmp;
946
947 LIST_FOR_EACH_ENTRY_SAFE(entry, tmp, &rscreen->fences.blocks, head) {
948 LIST_DEL(&entry->head);
949 FREE(entry);
950 }
951
952 rscreen->ws->buffer_unmap(rscreen->fences.bo->cs_buf);
953 pipe_resource_reference((struct pipe_resource**)&rscreen->fences.bo, NULL);
954 }
955 #if R600_TRACE_CS
956 if (rscreen->trace_bo) {
957 rscreen->ws->buffer_unmap(rscreen->trace_bo->cs_buf);
958 pipe_resource_reference((struct pipe_resource**)&rscreen->trace_bo, NULL);
959 }
960 #endif
961 pipe_mutex_destroy(rscreen->fences.mutex);
962
963 rscreen->ws->destroy(rscreen->ws);
964 FREE(rscreen);
965 }
966
967 static void r600_fence_reference(struct pipe_screen *pscreen,
968 struct pipe_fence_handle **ptr,
969 struct pipe_fence_handle *fence)
970 {
971 struct r600_fence **oldf = (struct r600_fence**)ptr;
972 struct r600_fence *newf = (struct r600_fence*)fence;
973
974 if (pipe_reference(&(*oldf)->reference, &newf->reference)) {
975 struct r600_screen *rscreen = (struct r600_screen *)pscreen;
976 pipe_mutex_lock(rscreen->fences.mutex);
977 pipe_resource_reference((struct pipe_resource**)&(*oldf)->sleep_bo, NULL);
978 LIST_ADDTAIL(&(*oldf)->head, &rscreen->fences.pool);
979 pipe_mutex_unlock(rscreen->fences.mutex);
980 }
981
982 *ptr = fence;
983 }
984
985 static boolean r600_fence_signalled(struct pipe_screen *pscreen,
986 struct pipe_fence_handle *fence)
987 {
988 struct r600_screen *rscreen = (struct r600_screen *)pscreen;
989 struct r600_fence *rfence = (struct r600_fence*)fence;
990
991 return rscreen->fences.data[rfence->index] != 0;
992 }
993
994 static boolean r600_fence_finish(struct pipe_screen *pscreen,
995 struct pipe_fence_handle *fence,
996 uint64_t timeout)
997 {
998 struct r600_screen *rscreen = (struct r600_screen *)pscreen;
999 struct r600_fence *rfence = (struct r600_fence*)fence;
1000 int64_t start_time = 0;
1001 unsigned spins = 0;
1002
1003 if (timeout != PIPE_TIMEOUT_INFINITE) {
1004 start_time = os_time_get();
1005
1006 /* Convert to microseconds. */
1007 timeout /= 1000;
1008 }
1009
1010 while (rscreen->fences.data[rfence->index] == 0) {
1011 /* Special-case infinite timeout - wait for the dummy BO to become idle */
1012 if (timeout == PIPE_TIMEOUT_INFINITE) {
1013 rscreen->ws->buffer_wait(rfence->sleep_bo->buf, RADEON_USAGE_READWRITE);
1014 break;
1015 }
1016
1017 /* The dummy BO will be busy until the CS including the fence has completed, or
1018 * the GPU is reset. Don't bother continuing to spin when the BO is idle. */
1019 if (!rscreen->ws->buffer_is_busy(rfence->sleep_bo->buf, RADEON_USAGE_READWRITE))
1020 break;
1021
1022 if (++spins % 256)
1023 continue;
1024 #ifdef PIPE_OS_UNIX
1025 sched_yield();
1026 #else
1027 os_time_sleep(10);
1028 #endif
1029 if (timeout != PIPE_TIMEOUT_INFINITE &&
1030 os_time_get() - start_time >= timeout) {
1031 break;
1032 }
1033 }
1034
1035 return rscreen->fences.data[rfence->index] != 0;
1036 }
1037
1038 static int r600_interpret_tiling(struct r600_screen *rscreen, uint32_t tiling_config)
1039 {
1040 switch ((tiling_config & 0xe) >> 1) {
1041 case 0:
1042 rscreen->tiling_info.num_channels = 1;
1043 break;
1044 case 1:
1045 rscreen->tiling_info.num_channels = 2;
1046 break;
1047 case 2:
1048 rscreen->tiling_info.num_channels = 4;
1049 break;
1050 case 3:
1051 rscreen->tiling_info.num_channels = 8;
1052 break;
1053 default:
1054 return -EINVAL;
1055 }
1056
1057 switch ((tiling_config & 0x30) >> 4) {
1058 case 0:
1059 rscreen->tiling_info.num_banks = 4;
1060 break;
1061 case 1:
1062 rscreen->tiling_info.num_banks = 8;
1063 break;
1064 default:
1065 return -EINVAL;
1066
1067 }
1068 switch ((tiling_config & 0xc0) >> 6) {
1069 case 0:
1070 rscreen->tiling_info.group_bytes = 256;
1071 break;
1072 case 1:
1073 rscreen->tiling_info.group_bytes = 512;
1074 break;
1075 default:
1076 return -EINVAL;
1077 }
1078 return 0;
1079 }
1080
1081 static int evergreen_interpret_tiling(struct r600_screen *rscreen, uint32_t tiling_config)
1082 {
1083 switch (tiling_config & 0xf) {
1084 case 0:
1085 rscreen->tiling_info.num_channels = 1;
1086 break;
1087 case 1:
1088 rscreen->tiling_info.num_channels = 2;
1089 break;
1090 case 2:
1091 rscreen->tiling_info.num_channels = 4;
1092 break;
1093 case 3:
1094 rscreen->tiling_info.num_channels = 8;
1095 break;
1096 default:
1097 return -EINVAL;
1098 }
1099
1100 switch ((tiling_config & 0xf0) >> 4) {
1101 case 0:
1102 rscreen->tiling_info.num_banks = 4;
1103 break;
1104 case 1:
1105 rscreen->tiling_info.num_banks = 8;
1106 break;
1107 case 2:
1108 rscreen->tiling_info.num_banks = 16;
1109 break;
1110 default:
1111 return -EINVAL;
1112 }
1113
1114 switch ((tiling_config & 0xf00) >> 8) {
1115 case 0:
1116 rscreen->tiling_info.group_bytes = 256;
1117 break;
1118 case 1:
1119 rscreen->tiling_info.group_bytes = 512;
1120 break;
1121 default:
1122 return -EINVAL;
1123 }
1124 return 0;
1125 }
1126
1127 static int r600_init_tiling(struct r600_screen *rscreen)
1128 {
1129 uint32_t tiling_config = rscreen->info.r600_tiling_config;
1130
1131 /* set default group bytes, overridden by tiling info ioctl */
1132 if (rscreen->chip_class <= R700) {
1133 rscreen->tiling_info.group_bytes = 256;
1134 } else {
1135 rscreen->tiling_info.group_bytes = 512;
1136 }
1137
1138 if (!tiling_config)
1139 return 0;
1140
1141 if (rscreen->chip_class <= R700) {
1142 return r600_interpret_tiling(rscreen, tiling_config);
1143 } else {
1144 return evergreen_interpret_tiling(rscreen, tiling_config);
1145 }
1146 }
1147
1148 static uint64_t r600_get_timestamp(struct pipe_screen *screen)
1149 {
1150 struct r600_screen *rscreen = (struct r600_screen*)screen;
1151
1152 return 1000000 * rscreen->ws->query_value(rscreen->ws, RADEON_TIMESTAMP) /
1153 rscreen->info.r600_clock_crystal_freq;
1154 }
1155
1156 static int r600_get_driver_query_info(struct pipe_screen *screen,
1157 unsigned index,
1158 struct pipe_driver_query_info *info)
1159 {
1160 struct r600_screen *rscreen = (struct r600_screen*)screen;
1161 struct pipe_driver_query_info list[] = {
1162 {"draw-calls", R600_QUERY_DRAW_CALLS, 0},
1163 {"requested-VRAM", R600_QUERY_REQUESTED_VRAM, rscreen->info.vram_size, TRUE},
1164 {"requested-GTT", R600_QUERY_REQUESTED_GTT, rscreen->info.gart_size, TRUE},
1165 {"buffer-wait-time", R600_QUERY_BUFFER_WAIT_TIME, 0, FALSE}
1166 };
1167
1168 if (!info)
1169 return Elements(list);
1170
1171 if (index >= Elements(list))
1172 return 0;
1173
1174 *info = list[index];
1175 return 1;
1176 }
1177
1178 struct pipe_screen *r600_screen_create(struct radeon_winsys *ws)
1179 {
1180 struct r600_screen *rscreen = CALLOC_STRUCT(r600_screen);
1181
1182 if (rscreen == NULL) {
1183 return NULL;
1184 }
1185
1186 rscreen->ws = ws;
1187 ws->query_info(ws, &rscreen->info);
1188
1189 rscreen->debug_flags = debug_get_flags_option("R600_DEBUG", debug_options, 0);
1190 if (debug_get_bool_option("R600_DEBUG_COMPUTE", FALSE))
1191 rscreen->debug_flags |= DBG_COMPUTE;
1192 if (debug_get_bool_option("R600_DUMP_SHADERS", FALSE))
1193 rscreen->debug_flags |= DBG_FS | DBG_VS | DBG_GS | DBG_PS | DBG_CS;
1194 if (!debug_get_bool_option("R600_HYPERZ", TRUE))
1195 rscreen->debug_flags |= DBG_NO_HYPERZ;
1196 if (!debug_get_bool_option("R600_LLVM", TRUE))
1197 rscreen->debug_flags |= DBG_NO_LLVM;
1198 if (debug_get_bool_option("R600_PRINT_TEXDEPTH", FALSE))
1199 rscreen->debug_flags |= DBG_TEX_DEPTH;
1200 rscreen->family = rscreen->info.family;
1201 rscreen->chip_class = rscreen->info.chip_class;
1202
1203 if (rscreen->family == CHIP_UNKNOWN) {
1204 fprintf(stderr, "r600: Unknown chipset 0x%04X\n", rscreen->info.pci_id);
1205 FREE(rscreen);
1206 return NULL;
1207 }
1208
1209 /* Figure out streamout kernel support. */
1210 switch (rscreen->chip_class) {
1211 case R600:
1212 if (rscreen->family < CHIP_RS780) {
1213 rscreen->has_streamout = rscreen->info.drm_minor >= 14;
1214 } else {
1215 rscreen->has_streamout = rscreen->info.drm_minor >= 23;
1216 }
1217 break;
1218 case R700:
1219 rscreen->has_streamout = rscreen->info.drm_minor >= 17;
1220 break;
1221 case EVERGREEN:
1222 case CAYMAN:
1223 rscreen->has_streamout = rscreen->info.drm_minor >= 14;
1224 break;
1225 default:
1226 rscreen->has_streamout = FALSE;
1227 break;
1228 }
1229
1230 /* MSAA support. */
1231 switch (rscreen->chip_class) {
1232 case R600:
1233 case R700:
1234 rscreen->has_msaa = rscreen->info.drm_minor >= 22;
1235 rscreen->msaa_texture_support = MSAA_TEXTURE_DECOMPRESSED;
1236 break;
1237 case EVERGREEN:
1238 rscreen->has_msaa = rscreen->info.drm_minor >= 19;
1239 rscreen->msaa_texture_support =
1240 rscreen->info.drm_minor >= 24 ? MSAA_TEXTURE_COMPRESSED :
1241 MSAA_TEXTURE_DECOMPRESSED;
1242 break;
1243 case CAYMAN:
1244 rscreen->has_msaa = rscreen->info.drm_minor >= 19;
1245 /* We should be able to read compressed MSAA textures, but it doesn't work. */
1246 rscreen->msaa_texture_support = MSAA_TEXTURE_SAMPLE_ZERO;
1247 break;
1248 default:
1249 rscreen->has_msaa = FALSE;
1250 rscreen->msaa_texture_support = 0;
1251 break;
1252 }
1253
1254 rscreen->has_cp_dma = rscreen->info.drm_minor >= 27 &&
1255 !(rscreen->debug_flags & DBG_NO_CP_DMA);
1256
1257 if (r600_init_tiling(rscreen)) {
1258 FREE(rscreen);
1259 return NULL;
1260 }
1261
1262 rscreen->screen.destroy = r600_destroy_screen;
1263 rscreen->screen.get_name = r600_get_name;
1264 rscreen->screen.get_vendor = r600_get_vendor;
1265 rscreen->screen.get_param = r600_get_param;
1266 rscreen->screen.get_shader_param = r600_get_shader_param;
1267 rscreen->screen.get_paramf = r600_get_paramf;
1268 rscreen->screen.get_compute_param = r600_get_compute_param;
1269 rscreen->screen.get_timestamp = r600_get_timestamp;
1270
1271 if (rscreen->chip_class >= EVERGREEN) {
1272 rscreen->screen.is_format_supported = evergreen_is_format_supported;
1273 rscreen->dma_blit = &evergreen_dma_blit;
1274 } else {
1275 rscreen->screen.is_format_supported = r600_is_format_supported;
1276 rscreen->dma_blit = &r600_dma_blit;
1277 }
1278 rscreen->screen.context_create = r600_create_context;
1279 rscreen->screen.fence_reference = r600_fence_reference;
1280 rscreen->screen.fence_signalled = r600_fence_signalled;
1281 rscreen->screen.fence_finish = r600_fence_finish;
1282 rscreen->screen.get_driver_query_info = r600_get_driver_query_info;
1283
1284 if (rscreen->info.has_uvd) {
1285 rscreen->screen.get_video_param = ruvd_get_video_param;
1286 rscreen->screen.is_video_format_supported = ruvd_is_format_supported;
1287 } else {
1288 rscreen->screen.get_video_param = r600_get_video_param;
1289 rscreen->screen.is_video_format_supported = vl_video_buffer_is_format_supported;
1290 }
1291
1292 r600_init_screen_resource_functions(&rscreen->screen);
1293
1294 util_format_s3tc_init();
1295
1296 rscreen->fences.bo = NULL;
1297 rscreen->fences.data = NULL;
1298 rscreen->fences.next_index = 0;
1299 LIST_INITHEAD(&rscreen->fences.pool);
1300 LIST_INITHEAD(&rscreen->fences.blocks);
1301 pipe_mutex_init(rscreen->fences.mutex);
1302
1303 rscreen->global_pool = compute_memory_pool_new(rscreen);
1304
1305 #if R600_TRACE_CS
1306 rscreen->cs_count = 0;
1307 if (rscreen->info.drm_minor >= 28) {
1308 rscreen->trace_bo = (struct r600_resource*)pipe_buffer_create(&rscreen->screen,
1309 PIPE_BIND_CUSTOM,
1310 PIPE_USAGE_STAGING,
1311 4096);
1312 if (rscreen->trace_bo) {
1313 rscreen->trace_ptr = rscreen->ws->buffer_map(rscreen->trace_bo->cs_buf, NULL,
1314 PIPE_TRANSFER_UNSYNCHRONIZED);
1315 }
1316 }
1317 #endif
1318
1319 return &rscreen->screen;
1320 }