7f308f7d001e058bee4dba9622450a3c789c25ff
[mesa.git] / src / gallium / drivers / r600 / r600_pipe.c
1 /*
2 * Copyright 2010 Jerome Glisse <glisse@freedesktop.org>
3 *
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * on the rights to use, copy, modify, merge, publish, distribute, sub
8 * license, and/or sell copies of the Software, and to permit persons to whom
9 * the Software is furnished to do so, subject to the following conditions:
10 *
11 * The above copyright notice and this permission notice (including the next
12 * paragraph) shall be included in all copies or substantial portions of the
13 * Software.
14 *
15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17 * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL
18 * THE AUTHOR(S) AND/OR THEIR SUPPLIERS BE LIABLE FOR ANY CLAIM,
19 * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
20 * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
21 * USE OR OTHER DEALINGS IN THE SOFTWARE.
22 */
23 #include "r600_pipe.h"
24 #include "r600_public.h"
25 #include "r600_isa.h"
26 #include "evergreen_compute.h"
27 #include "r600d.h"
28
29 #include <errno.h>
30 #include "pipe/p_shader_tokens.h"
31 #include "util/u_blitter.h"
32 #include "util/u_debug.h"
33 #include "util/u_format_s3tc.h"
34 #include "util/u_memory.h"
35 #include "util/u_simple_shaders.h"
36 #include "util/u_upload_mgr.h"
37 #include "util/u_math.h"
38 #include "vl/vl_decoder.h"
39 #include "vl/vl_video_buffer.h"
40 #include "os/os_time.h"
41
42 static const struct debug_named_value debug_options[] = {
43 /* logging */
44 { "texdepth", DBG_TEX_DEPTH, "Print texture depth info" },
45 { "compute", DBG_COMPUTE, "Print compute info" },
46
47 /* shaders */
48 { "fs", DBG_FS, "Print fetch shaders" },
49 { "vs", DBG_VS, "Print vertex shaders" },
50 { "gs", DBG_GS, "Print geometry shaders" },
51 { "ps", DBG_PS, "Print pixel shaders" },
52 { "cs", DBG_CS, "Print compute shaders" },
53
54 /* features */
55 { "nohyperz", DBG_NO_HYPERZ, "Disable Hyper-Z" },
56 #if defined(R600_USE_LLVM)
57 { "nollvm", DBG_NO_LLVM, "Disable the LLVM shader compiler" },
58 #endif
59 { "nocpdma", DBG_NO_CP_DMA, "Disable CP DMA" },
60 { "nodma", DBG_NO_ASYNC_DMA, "Disable asynchronous DMA" },
61 /* GL uses the word INVALIDATE, gallium uses the word DISCARD */
62 { "noinvalrange", DBG_NO_DISCARD_RANGE, "Disable handling of INVALIDATE_RANGE map flags" },
63
64 DEBUG_NAMED_VALUE_END /* must be last */
65 };
66
67 /*
68 * pipe_context
69 */
70 static struct r600_fence *r600_create_fence(struct r600_context *rctx)
71 {
72 struct r600_screen *rscreen = rctx->screen;
73 struct r600_fence *fence = NULL;
74
75 pipe_mutex_lock(rscreen->fences.mutex);
76
77 if (!rscreen->fences.bo) {
78 /* Create the shared buffer object */
79 rscreen->fences.bo = (struct r600_resource*)
80 pipe_buffer_create(&rscreen->screen, PIPE_BIND_CUSTOM,
81 PIPE_USAGE_STAGING, 4096);
82 if (!rscreen->fences.bo) {
83 R600_ERR("r600: failed to create bo for fence objects\n");
84 goto out;
85 }
86 rscreen->fences.data = r600_buffer_mmap_sync_with_rings(rctx, rscreen->fences.bo, PIPE_TRANSFER_READ_WRITE);
87 }
88
89 if (!LIST_IS_EMPTY(&rscreen->fences.pool)) {
90 struct r600_fence *entry;
91
92 /* Try to find a freed fence that has been signalled */
93 LIST_FOR_EACH_ENTRY(entry, &rscreen->fences.pool, head) {
94 if (rscreen->fences.data[entry->index] != 0) {
95 LIST_DELINIT(&entry->head);
96 fence = entry;
97 break;
98 }
99 }
100 }
101
102 if (!fence) {
103 /* Allocate a new fence */
104 struct r600_fence_block *block;
105 unsigned index;
106
107 if ((rscreen->fences.next_index + 1) >= 1024) {
108 R600_ERR("r600: too many concurrent fences\n");
109 goto out;
110 }
111
112 index = rscreen->fences.next_index++;
113
114 if (!(index % FENCE_BLOCK_SIZE)) {
115 /* Allocate a new block */
116 block = CALLOC_STRUCT(r600_fence_block);
117 if (block == NULL)
118 goto out;
119
120 LIST_ADD(&block->head, &rscreen->fences.blocks);
121 } else {
122 block = LIST_ENTRY(struct r600_fence_block, rscreen->fences.blocks.next, head);
123 }
124
125 fence = &block->fences[index % FENCE_BLOCK_SIZE];
126 fence->index = index;
127 }
128
129 pipe_reference_init(&fence->reference, 1);
130
131 rscreen->fences.data[fence->index] = 0;
132 r600_context_emit_fence(rctx, rscreen->fences.bo, fence->index, 1);
133
134 /* Create a dummy BO so that fence_finish without a timeout can sleep waiting for completion */
135 fence->sleep_bo = (struct r600_resource*)
136 pipe_buffer_create(&rctx->screen->screen, PIPE_BIND_CUSTOM,
137 PIPE_USAGE_STAGING, 1);
138 /* Add the fence as a dummy relocation. */
139 r600_context_bo_reloc(rctx, &rctx->rings.gfx, fence->sleep_bo, RADEON_USAGE_READWRITE);
140
141 out:
142 pipe_mutex_unlock(rscreen->fences.mutex);
143 return fence;
144 }
145
146 static void r600_flush(struct pipe_context *ctx, unsigned flags)
147 {
148 struct r600_context *rctx = (struct r600_context *)ctx;
149 struct pipe_query *render_cond = NULL;
150 unsigned render_cond_mode = 0;
151
152 rctx->rings.gfx.flushing = true;
153 /* Disable render condition. */
154 if (rctx->current_render_cond) {
155 render_cond = rctx->current_render_cond;
156 render_cond_mode = rctx->current_render_cond_mode;
157 ctx->render_condition(ctx, NULL, 0);
158 }
159
160 r600_context_flush(rctx, flags);
161 rctx->rings.gfx.flushing = false;
162 r600_begin_new_cs(rctx);
163
164 /* Re-enable render condition. */
165 if (render_cond) {
166 ctx->render_condition(ctx, render_cond, render_cond_mode);
167 }
168 }
169
170 static void r600_flush_from_st(struct pipe_context *ctx,
171 struct pipe_fence_handle **fence,
172 enum pipe_flush_flags flags)
173 {
174 struct r600_context *rctx = (struct r600_context *)ctx;
175 struct r600_fence **rfence = (struct r600_fence**)fence;
176 unsigned fflags;
177
178 fflags = flags & PIPE_FLUSH_END_OF_FRAME ? RADEON_FLUSH_END_OF_FRAME : 0;
179 if (rfence) {
180 *rfence = r600_create_fence(rctx);
181 }
182 /* flush gfx & dma ring, order does not matter as only one can be live */
183 if (rctx->rings.dma.cs) {
184 rctx->rings.dma.flush(rctx, fflags);
185 }
186 rctx->rings.gfx.flush(rctx, fflags);
187 }
188
189 static void r600_flush_gfx_ring(void *ctx, unsigned flags)
190 {
191 r600_flush((struct pipe_context*)ctx, flags);
192 }
193
194 static void r600_flush_dma_ring(void *ctx, unsigned flags)
195 {
196 struct r600_context *rctx = (struct r600_context *)ctx;
197 struct radeon_winsys_cs *cs = rctx->rings.dma.cs;
198 unsigned padding_dw, i;
199
200 if (!cs->cdw) {
201 return;
202 }
203
204 /* Pad the DMA CS to a multiple of 8 dwords. */
205 padding_dw = 8 - cs->cdw % 8;
206 if (padding_dw < 8) {
207 for (i = 0; i < padding_dw; i++) {
208 cs->buf[cs->cdw++] = DMA_PACKET(DMA_PACKET_NOP, 0, 0, 0);
209 }
210 }
211
212 rctx->rings.dma.flushing = true;
213 rctx->ws->cs_flush(cs, flags);
214 rctx->rings.dma.flushing = false;
215 }
216
217 boolean r600_rings_is_buffer_referenced(struct r600_context *ctx,
218 struct radeon_winsys_cs_handle *buf,
219 enum radeon_bo_usage usage)
220 {
221 if (ctx->ws->cs_is_buffer_referenced(ctx->rings.gfx.cs, buf, usage)) {
222 return TRUE;
223 }
224 if (ctx->rings.dma.cs) {
225 if (ctx->ws->cs_is_buffer_referenced(ctx->rings.dma.cs, buf, usage)) {
226 return TRUE;
227 }
228 }
229 return FALSE;
230 }
231
232 void *r600_buffer_mmap_sync_with_rings(struct r600_context *ctx,
233 struct r600_resource *resource,
234 unsigned usage)
235 {
236 enum radeon_bo_usage rusage = RADEON_USAGE_READWRITE;
237 unsigned flags = 0;
238 bool sync_flush = TRUE;
239
240 if (usage & PIPE_TRANSFER_UNSYNCHRONIZED) {
241 return ctx->ws->buffer_map(resource->cs_buf, NULL, usage);
242 }
243
244 if (!(usage & PIPE_TRANSFER_WRITE)) {
245 /* have to wait for pending read */
246 rusage = RADEON_USAGE_WRITE;
247 }
248 if (usage & PIPE_TRANSFER_DONTBLOCK) {
249 flags |= RADEON_FLUSH_ASYNC;
250 }
251
252 if (ctx->ws->cs_is_buffer_referenced(ctx->rings.gfx.cs, resource->cs_buf, rusage) && ctx->rings.gfx.cs->cdw) {
253 ctx->rings.gfx.flush(ctx, flags);
254 if (usage & PIPE_TRANSFER_DONTBLOCK) {
255 return NULL;
256 }
257 }
258 if (ctx->rings.dma.cs) {
259 if (ctx->ws->cs_is_buffer_referenced(ctx->rings.dma.cs, resource->cs_buf, rusage) && ctx->rings.dma.cs->cdw) {
260 ctx->rings.dma.flush(ctx, flags);
261 if (usage & PIPE_TRANSFER_DONTBLOCK) {
262 return NULL;
263 }
264 }
265 }
266
267 if (usage & PIPE_TRANSFER_DONTBLOCK) {
268 if (ctx->ws->buffer_is_busy(resource->buf, rusage)) {
269 return NULL;
270 }
271 }
272 if (sync_flush) {
273 /* Try to avoid busy-waiting in radeon_bo_wait. */
274 ctx->ws->cs_sync_flush(ctx->rings.gfx.cs);
275 if (ctx->rings.dma.cs) {
276 ctx->ws->cs_sync_flush(ctx->rings.dma.cs);
277 }
278 }
279 ctx->ws->buffer_wait(resource->buf, rusage);
280
281 /* at this point everything is synchronized */
282 return ctx->ws->buffer_map(resource->cs_buf, NULL, usage | PIPE_TRANSFER_UNSYNCHRONIZED);
283 }
284
285 static void r600_flush_from_winsys(void *ctx, unsigned flags)
286 {
287 struct r600_context *rctx = (struct r600_context *)ctx;
288
289 rctx->rings.gfx.flush(rctx, flags);
290 }
291
292 static void r600_flush_dma_from_winsys(void *ctx, unsigned flags)
293 {
294 struct r600_context *rctx = (struct r600_context *)ctx;
295
296 rctx->rings.dma.flush(rctx, flags);
297 }
298
299 static void r600_destroy_context(struct pipe_context *context)
300 {
301 struct r600_context *rctx = (struct r600_context *)context;
302
303 r600_isa_destroy(rctx->isa);
304
305 pipe_resource_reference((struct pipe_resource**)&rctx->dummy_cmask, NULL);
306 pipe_resource_reference((struct pipe_resource**)&rctx->dummy_fmask, NULL);
307
308 if (rctx->dummy_pixel_shader) {
309 rctx->context.delete_fs_state(&rctx->context, rctx->dummy_pixel_shader);
310 }
311 if (rctx->custom_dsa_flush) {
312 rctx->context.delete_depth_stencil_alpha_state(&rctx->context, rctx->custom_dsa_flush);
313 }
314 if (rctx->custom_blend_resolve) {
315 rctx->context.delete_blend_state(&rctx->context, rctx->custom_blend_resolve);
316 }
317 if (rctx->custom_blend_decompress) {
318 rctx->context.delete_blend_state(&rctx->context, rctx->custom_blend_decompress);
319 }
320 if (rctx->custom_blend_fmask_decompress) {
321 rctx->context.delete_blend_state(&rctx->context, rctx->custom_blend_fmask_decompress);
322 }
323 util_unreference_framebuffer_state(&rctx->framebuffer.state);
324
325 if (rctx->blitter) {
326 util_blitter_destroy(rctx->blitter);
327 }
328 if (rctx->uploader) {
329 u_upload_destroy(rctx->uploader);
330 }
331 if (rctx->allocator_so_filled_size) {
332 u_suballocator_destroy(rctx->allocator_so_filled_size);
333 }
334 if (rctx->allocator_fetch_shader) {
335 u_suballocator_destroy(rctx->allocator_fetch_shader);
336 }
337 util_slab_destroy(&rctx->pool_transfers);
338
339 r600_release_command_buffer(&rctx->start_cs_cmd);
340
341 if (rctx->rings.gfx.cs) {
342 rctx->ws->cs_destroy(rctx->rings.gfx.cs);
343 }
344 if (rctx->rings.dma.cs) {
345 rctx->ws->cs_destroy(rctx->rings.dma.cs);
346 }
347
348 FREE(rctx);
349 }
350
351 static struct pipe_context *r600_create_context(struct pipe_screen *screen, void *priv)
352 {
353 struct r600_context *rctx = CALLOC_STRUCT(r600_context);
354 struct r600_screen* rscreen = (struct r600_screen *)screen;
355
356 if (rctx == NULL)
357 return NULL;
358
359 util_slab_create(&rctx->pool_transfers,
360 sizeof(struct r600_transfer), 64,
361 UTIL_SLAB_SINGLETHREADED);
362
363 rctx->context.screen = screen;
364 rctx->context.priv = priv;
365 rctx->context.destroy = r600_destroy_context;
366 rctx->context.flush = r600_flush_from_st;
367
368 /* Easy accessing of screen/winsys. */
369 rctx->screen = rscreen;
370 rctx->ws = rscreen->ws;
371 rctx->family = rscreen->family;
372 rctx->chip_class = rscreen->chip_class;
373 rctx->keep_tiling_flags = rscreen->info.drm_minor >= 12;
374
375 LIST_INITHEAD(&rctx->active_nontimer_queries);
376
377 r600_init_blit_functions(rctx);
378 r600_init_query_functions(rctx);
379 r600_init_context_resource_functions(rctx);
380 r600_init_surface_functions(rctx);
381
382 rctx->context.create_video_decoder = vl_create_decoder;
383 rctx->context.create_video_buffer = vl_video_buffer_create;
384
385 r600_init_common_state_functions(rctx);
386
387 switch (rctx->chip_class) {
388 case R600:
389 case R700:
390 r600_init_state_functions(rctx);
391 r600_init_atom_start_cs(rctx);
392 rctx->max_db = 4;
393 rctx->custom_dsa_flush = r600_create_db_flush_dsa(rctx);
394 rctx->custom_blend_resolve = rctx->chip_class == R700 ? r700_create_resolve_blend(rctx)
395 : r600_create_resolve_blend(rctx);
396 rctx->custom_blend_decompress = r600_create_decompress_blend(rctx);
397 rctx->has_vertex_cache = !(rctx->family == CHIP_RV610 ||
398 rctx->family == CHIP_RV620 ||
399 rctx->family == CHIP_RS780 ||
400 rctx->family == CHIP_RS880 ||
401 rctx->family == CHIP_RV710);
402 break;
403 case EVERGREEN:
404 case CAYMAN:
405 evergreen_init_state_functions(rctx);
406 evergreen_init_atom_start_cs(rctx);
407 evergreen_init_atom_start_compute_cs(rctx);
408 rctx->max_db = 8;
409 rctx->custom_dsa_flush = evergreen_create_db_flush_dsa(rctx);
410 rctx->custom_blend_resolve = evergreen_create_resolve_blend(rctx);
411 rctx->custom_blend_decompress = evergreen_create_decompress_blend(rctx);
412 rctx->custom_blend_fmask_decompress = evergreen_create_fmask_decompress_blend(rctx);
413 rctx->has_vertex_cache = !(rctx->family == CHIP_CEDAR ||
414 rctx->family == CHIP_PALM ||
415 rctx->family == CHIP_SUMO ||
416 rctx->family == CHIP_SUMO2 ||
417 rctx->family == CHIP_CAICOS ||
418 rctx->family == CHIP_CAYMAN ||
419 rctx->family == CHIP_ARUBA);
420 break;
421 default:
422 R600_ERR("Unsupported chip class %d.\n", rctx->chip_class);
423 goto fail;
424 }
425
426 rctx->rings.gfx.cs = rctx->ws->cs_create(rctx->ws, RING_GFX);
427 rctx->rings.gfx.flush = r600_flush_gfx_ring;
428 rctx->ws->cs_set_flush_callback(rctx->rings.gfx.cs, r600_flush_from_winsys, rctx);
429 rctx->rings.gfx.flushing = false;
430
431 rctx->rings.dma.cs = NULL;
432 if (rscreen->info.r600_has_dma && !(rscreen->debug_flags & DBG_NO_ASYNC_DMA)) {
433 rctx->rings.dma.cs = rctx->ws->cs_create(rctx->ws, RING_DMA);
434 rctx->rings.dma.flush = r600_flush_dma_ring;
435 rctx->ws->cs_set_flush_callback(rctx->rings.dma.cs, r600_flush_dma_from_winsys, rctx);
436 rctx->rings.dma.flushing = false;
437 }
438
439 rctx->uploader = u_upload_create(&rctx->context, 1024 * 1024, 256,
440 PIPE_BIND_INDEX_BUFFER |
441 PIPE_BIND_CONSTANT_BUFFER);
442 if (!rctx->uploader)
443 goto fail;
444
445 rctx->allocator_fetch_shader = u_suballocator_create(&rctx->context, 64 * 1024, 256,
446 0, PIPE_USAGE_STATIC, FALSE);
447 if (!rctx->allocator_fetch_shader)
448 goto fail;
449
450 rctx->allocator_so_filled_size = u_suballocator_create(&rctx->context, 4096, 4,
451 0, PIPE_USAGE_STATIC, TRUE);
452 if (!rctx->allocator_so_filled_size)
453 goto fail;
454
455 rctx->isa = calloc(1, sizeof(struct r600_isa));
456 if (!rctx->isa || r600_isa_init(rctx, rctx->isa))
457 goto fail;
458
459 rctx->blitter = util_blitter_create(&rctx->context);
460 if (rctx->blitter == NULL)
461 goto fail;
462 util_blitter_set_texture_multisample(rctx->blitter, rscreen->has_msaa);
463 rctx->blitter->draw_rectangle = r600_draw_rectangle;
464
465 r600_begin_new_cs(rctx);
466 r600_get_backend_mask(rctx); /* this emits commands and must be last */
467
468 rctx->dummy_pixel_shader =
469 util_make_fragment_cloneinput_shader(&rctx->context, 0,
470 TGSI_SEMANTIC_GENERIC,
471 TGSI_INTERPOLATE_CONSTANT);
472 rctx->context.bind_fs_state(&rctx->context, rctx->dummy_pixel_shader);
473
474 return &rctx->context;
475
476 fail:
477 r600_destroy_context(&rctx->context);
478 return NULL;
479 }
480
481 /*
482 * pipe_screen
483 */
484 static const char* r600_get_vendor(struct pipe_screen* pscreen)
485 {
486 return "X.Org";
487 }
488
489 static const char *r600_get_family_name(enum radeon_family family)
490 {
491 switch(family) {
492 case CHIP_R600: return "AMD R600";
493 case CHIP_RV610: return "AMD RV610";
494 case CHIP_RV630: return "AMD RV630";
495 case CHIP_RV670: return "AMD RV670";
496 case CHIP_RV620: return "AMD RV620";
497 case CHIP_RV635: return "AMD RV635";
498 case CHIP_RS780: return "AMD RS780";
499 case CHIP_RS880: return "AMD RS880";
500 case CHIP_RV770: return "AMD RV770";
501 case CHIP_RV730: return "AMD RV730";
502 case CHIP_RV710: return "AMD RV710";
503 case CHIP_RV740: return "AMD RV740";
504 case CHIP_CEDAR: return "AMD CEDAR";
505 case CHIP_REDWOOD: return "AMD REDWOOD";
506 case CHIP_JUNIPER: return "AMD JUNIPER";
507 case CHIP_CYPRESS: return "AMD CYPRESS";
508 case CHIP_HEMLOCK: return "AMD HEMLOCK";
509 case CHIP_PALM: return "AMD PALM";
510 case CHIP_SUMO: return "AMD SUMO";
511 case CHIP_SUMO2: return "AMD SUMO2";
512 case CHIP_BARTS: return "AMD BARTS";
513 case CHIP_TURKS: return "AMD TURKS";
514 case CHIP_CAICOS: return "AMD CAICOS";
515 case CHIP_CAYMAN: return "AMD CAYMAN";
516 case CHIP_ARUBA: return "AMD ARUBA";
517 default: return "AMD unknown";
518 }
519 }
520
521 static const char* r600_get_name(struct pipe_screen* pscreen)
522 {
523 struct r600_screen *rscreen = (struct r600_screen *)pscreen;
524
525 return r600_get_family_name(rscreen->family);
526 }
527
528 static int r600_get_param(struct pipe_screen* pscreen, enum pipe_cap param)
529 {
530 struct r600_screen *rscreen = (struct r600_screen *)pscreen;
531 enum radeon_family family = rscreen->family;
532
533 switch (param) {
534 /* Supported features (boolean caps). */
535 case PIPE_CAP_NPOT_TEXTURES:
536 case PIPE_CAP_TWO_SIDED_STENCIL:
537 case PIPE_CAP_ANISOTROPIC_FILTER:
538 case PIPE_CAP_POINT_SPRITE:
539 case PIPE_CAP_OCCLUSION_QUERY:
540 case PIPE_CAP_TEXTURE_SHADOW_MAP:
541 case PIPE_CAP_TEXTURE_MIRROR_CLAMP:
542 case PIPE_CAP_BLEND_EQUATION_SEPARATE:
543 case PIPE_CAP_TEXTURE_SWIZZLE:
544 case PIPE_CAP_DEPTH_CLIP_DISABLE:
545 case PIPE_CAP_SHADER_STENCIL_EXPORT:
546 case PIPE_CAP_VERTEX_ELEMENT_INSTANCE_DIVISOR:
547 case PIPE_CAP_MIXED_COLORBUFFER_FORMATS:
548 case PIPE_CAP_TGSI_FS_COORD_ORIGIN_UPPER_LEFT:
549 case PIPE_CAP_TGSI_FS_COORD_PIXEL_CENTER_HALF_INTEGER:
550 case PIPE_CAP_SM3:
551 case PIPE_CAP_SEAMLESS_CUBE_MAP:
552 case PIPE_CAP_PRIMITIVE_RESTART:
553 case PIPE_CAP_CONDITIONAL_RENDER:
554 case PIPE_CAP_TEXTURE_BARRIER:
555 case PIPE_CAP_VERTEX_COLOR_UNCLAMPED:
556 case PIPE_CAP_QUADS_FOLLOW_PROVOKING_VERTEX_CONVENTION:
557 case PIPE_CAP_TGSI_INSTANCEID:
558 case PIPE_CAP_VERTEX_BUFFER_OFFSET_4BYTE_ALIGNED_ONLY:
559 case PIPE_CAP_VERTEX_BUFFER_STRIDE_4BYTE_ALIGNED_ONLY:
560 case PIPE_CAP_VERTEX_ELEMENT_SRC_OFFSET_4BYTE_ALIGNED_ONLY:
561 case PIPE_CAP_USER_INDEX_BUFFERS:
562 case PIPE_CAP_USER_CONSTANT_BUFFERS:
563 case PIPE_CAP_COMPUTE:
564 case PIPE_CAP_START_INSTANCE:
565 case PIPE_CAP_MAX_DUAL_SOURCE_RENDER_TARGETS:
566 case PIPE_CAP_TEXTURE_BUFFER_OBJECTS:
567 case PIPE_CAP_PREFER_BLIT_BASED_TEXTURE_TRANSFER:
568 return 1;
569 case PIPE_CAP_TGSI_TEXCOORD:
570 return 0;
571
572 case PIPE_CAP_MIN_MAP_BUFFER_ALIGNMENT:
573 return R600_MAP_BUFFER_ALIGNMENT;
574
575 case PIPE_CAP_CONSTANT_BUFFER_OFFSET_ALIGNMENT:
576 return 256;
577
578 case PIPE_CAP_GLSL_FEATURE_LEVEL:
579 return 140;
580
581 case PIPE_CAP_TEXTURE_MULTISAMPLE:
582 return rscreen->msaa_texture_support != MSAA_TEXTURE_SAMPLE_ZERO;
583
584 /* Supported except the original R600. */
585 case PIPE_CAP_INDEP_BLEND_ENABLE:
586 case PIPE_CAP_INDEP_BLEND_FUNC:
587 /* R600 doesn't support per-MRT blends */
588 return family == CHIP_R600 ? 0 : 1;
589
590 /* Supported on Evergreen. */
591 case PIPE_CAP_SEAMLESS_CUBE_MAP_PER_TEXTURE:
592 case PIPE_CAP_CUBE_MAP_ARRAY:
593 return family >= CHIP_CEDAR ? 1 : 0;
594
595 /* Unsupported features. */
596 case PIPE_CAP_TGSI_FS_COORD_ORIGIN_LOWER_LEFT:
597 case PIPE_CAP_TGSI_FS_COORD_PIXEL_CENTER_INTEGER:
598 case PIPE_CAP_SCALED_RESOLVE:
599 case PIPE_CAP_TGSI_CAN_COMPACT_VARYINGS:
600 case PIPE_CAP_TGSI_CAN_COMPACT_CONSTANTS:
601 case PIPE_CAP_FRAGMENT_COLOR_CLAMPED:
602 case PIPE_CAP_VERTEX_COLOR_CLAMPED:
603 case PIPE_CAP_USER_VERTEX_BUFFERS:
604 case PIPE_CAP_TEXTURE_BUFFER_OFFSET_ALIGNMENT:
605 case PIPE_CAP_QUERY_PIPELINE_STATISTICS:
606 return 0;
607
608 /* Stream output. */
609 case PIPE_CAP_MAX_STREAM_OUTPUT_BUFFERS:
610 return rscreen->has_streamout ? 4 : 0;
611 case PIPE_CAP_STREAM_OUTPUT_PAUSE_RESUME:
612 return rscreen->has_streamout ? 1 : 0;
613 case PIPE_CAP_MAX_STREAM_OUTPUT_SEPARATE_COMPONENTS:
614 case PIPE_CAP_MAX_STREAM_OUTPUT_INTERLEAVED_COMPONENTS:
615 return 32*4;
616
617 /* Texturing. */
618 case PIPE_CAP_MAX_TEXTURE_2D_LEVELS:
619 case PIPE_CAP_MAX_TEXTURE_3D_LEVELS:
620 case PIPE_CAP_MAX_TEXTURE_CUBE_LEVELS:
621 if (family >= CHIP_CEDAR)
622 return 15;
623 else
624 return 14;
625 case PIPE_CAP_MAX_TEXTURE_ARRAY_LAYERS:
626 return rscreen->info.drm_minor >= 9 ?
627 (family >= CHIP_CEDAR ? 16384 : 8192) : 0;
628 case PIPE_CAP_MAX_COMBINED_SAMPLERS:
629 return 32;
630
631 /* Render targets. */
632 case PIPE_CAP_MAX_RENDER_TARGETS:
633 /* XXX some r6xx are buggy and can only do 4 */
634 return 8;
635
636 /* Timer queries, present when the clock frequency is non zero. */
637 case PIPE_CAP_QUERY_TIME_ELAPSED:
638 return rscreen->info.r600_clock_crystal_freq != 0;
639 case PIPE_CAP_QUERY_TIMESTAMP:
640 return rscreen->info.drm_minor >= 20 &&
641 rscreen->info.r600_clock_crystal_freq != 0;
642
643 case PIPE_CAP_MIN_TEXEL_OFFSET:
644 return -8;
645
646 case PIPE_CAP_MAX_TEXEL_OFFSET:
647 return 7;
648 }
649 return 0;
650 }
651
652 static float r600_get_paramf(struct pipe_screen* pscreen,
653 enum pipe_capf param)
654 {
655 struct r600_screen *rscreen = (struct r600_screen *)pscreen;
656 enum radeon_family family = rscreen->family;
657
658 switch (param) {
659 case PIPE_CAPF_MAX_LINE_WIDTH:
660 case PIPE_CAPF_MAX_LINE_WIDTH_AA:
661 case PIPE_CAPF_MAX_POINT_WIDTH:
662 case PIPE_CAPF_MAX_POINT_WIDTH_AA:
663 if (family >= CHIP_CEDAR)
664 return 16384.0f;
665 else
666 return 8192.0f;
667 case PIPE_CAPF_MAX_TEXTURE_ANISOTROPY:
668 return 16.0f;
669 case PIPE_CAPF_MAX_TEXTURE_LOD_BIAS:
670 return 16.0f;
671 case PIPE_CAPF_GUARD_BAND_LEFT:
672 case PIPE_CAPF_GUARD_BAND_TOP:
673 case PIPE_CAPF_GUARD_BAND_RIGHT:
674 case PIPE_CAPF_GUARD_BAND_BOTTOM:
675 return 0.0f;
676 }
677 return 0.0f;
678 }
679
680 static int r600_get_shader_param(struct pipe_screen* pscreen, unsigned shader, enum pipe_shader_cap param)
681 {
682 switch(shader)
683 {
684 case PIPE_SHADER_FRAGMENT:
685 case PIPE_SHADER_VERTEX:
686 case PIPE_SHADER_COMPUTE:
687 break;
688 case PIPE_SHADER_GEOMETRY:
689 /* XXX: support and enable geometry programs */
690 return 0;
691 default:
692 /* XXX: support tessellation on Evergreen */
693 return 0;
694 }
695
696 switch (param) {
697 case PIPE_SHADER_CAP_MAX_INSTRUCTIONS:
698 case PIPE_SHADER_CAP_MAX_ALU_INSTRUCTIONS:
699 case PIPE_SHADER_CAP_MAX_TEX_INSTRUCTIONS:
700 case PIPE_SHADER_CAP_MAX_TEX_INDIRECTIONS:
701 return 16384;
702 case PIPE_SHADER_CAP_MAX_CONTROL_FLOW_DEPTH:
703 return 32;
704 case PIPE_SHADER_CAP_MAX_INPUTS:
705 return 32;
706 case PIPE_SHADER_CAP_MAX_TEMPS:
707 return 256; /* Max native temporaries. */
708 case PIPE_SHADER_CAP_MAX_ADDRS:
709 /* XXX Isn't this equal to TEMPS? */
710 return 1; /* Max native address registers */
711 case PIPE_SHADER_CAP_MAX_CONSTS:
712 return R600_MAX_CONST_BUFFER_SIZE;
713 case PIPE_SHADER_CAP_MAX_CONST_BUFFERS:
714 return R600_MAX_USER_CONST_BUFFERS;
715 case PIPE_SHADER_CAP_MAX_PREDS:
716 return 0; /* nothing uses this */
717 case PIPE_SHADER_CAP_TGSI_CONT_SUPPORTED:
718 return 1;
719 case PIPE_SHADER_CAP_TGSI_SQRT_SUPPORTED:
720 return 0;
721 case PIPE_SHADER_CAP_INDIRECT_INPUT_ADDR:
722 case PIPE_SHADER_CAP_INDIRECT_OUTPUT_ADDR:
723 case PIPE_SHADER_CAP_INDIRECT_TEMP_ADDR:
724 case PIPE_SHADER_CAP_INDIRECT_CONST_ADDR:
725 return 1;
726 case PIPE_SHADER_CAP_SUBROUTINES:
727 return 0;
728 case PIPE_SHADER_CAP_INTEGERS:
729 return 1;
730 case PIPE_SHADER_CAP_MAX_TEXTURE_SAMPLERS:
731 return 16;
732 case PIPE_SHADER_CAP_PREFERRED_IR:
733 if (shader == PIPE_SHADER_COMPUTE) {
734 return PIPE_SHADER_IR_LLVM;
735 } else {
736 return PIPE_SHADER_IR_TGSI;
737 }
738 }
739 return 0;
740 }
741
742 static int r600_get_video_param(struct pipe_screen *screen,
743 enum pipe_video_profile profile,
744 enum pipe_video_cap param)
745 {
746 switch (param) {
747 case PIPE_VIDEO_CAP_SUPPORTED:
748 return vl_profile_supported(screen, profile);
749 case PIPE_VIDEO_CAP_NPOT_TEXTURES:
750 return 1;
751 case PIPE_VIDEO_CAP_MAX_WIDTH:
752 case PIPE_VIDEO_CAP_MAX_HEIGHT:
753 return vl_video_buffer_max_size(screen);
754 case PIPE_VIDEO_CAP_PREFERED_FORMAT:
755 return PIPE_FORMAT_NV12;
756 case PIPE_VIDEO_CAP_PREFERS_INTERLACED:
757 return false;
758 case PIPE_VIDEO_CAP_SUPPORTS_INTERLACED:
759 return false;
760 case PIPE_VIDEO_CAP_SUPPORTS_PROGRESSIVE:
761 return true;
762 default:
763 return 0;
764 }
765 }
766
767 const char * r600_llvm_gpu_string(enum radeon_family family)
768 {
769 const char * gpu_family;
770
771 switch (family) {
772 case CHIP_R600:
773 case CHIP_RV610:
774 case CHIP_RV630:
775 case CHIP_RV620:
776 case CHIP_RV635:
777 case CHIP_RV670:
778 case CHIP_RS780:
779 case CHIP_RS880:
780 gpu_family = "r600";
781 break;
782 case CHIP_RV710:
783 gpu_family = "rv710";
784 break;
785 case CHIP_RV730:
786 gpu_family = "rv730";
787 break;
788 case CHIP_RV740:
789 case CHIP_RV770:
790 gpu_family = "rv770";
791 break;
792 case CHIP_PALM:
793 case CHIP_CEDAR:
794 gpu_family = "cedar";
795 break;
796 case CHIP_SUMO:
797 case CHIP_SUMO2:
798 case CHIP_REDWOOD:
799 gpu_family = "redwood";
800 break;
801 case CHIP_JUNIPER:
802 gpu_family = "juniper";
803 break;
804 case CHIP_HEMLOCK:
805 case CHIP_CYPRESS:
806 gpu_family = "cypress";
807 break;
808 case CHIP_BARTS:
809 gpu_family = "barts";
810 break;
811 case CHIP_TURKS:
812 gpu_family = "turks";
813 break;
814 case CHIP_CAICOS:
815 gpu_family = "caicos";
816 break;
817 case CHIP_CAYMAN:
818 case CHIP_ARUBA:
819 gpu_family = "cayman";
820 break;
821 default:
822 gpu_family = "";
823 fprintf(stderr, "Chip not supported by r600 llvm "
824 "backend, please file a bug at " PACKAGE_BUGREPORT "\n");
825 break;
826 }
827 return gpu_family;
828 }
829
830
831 static int r600_get_compute_param(struct pipe_screen *screen,
832 enum pipe_compute_cap param,
833 void *ret)
834 {
835 struct r600_screen *rscreen = (struct r600_screen *)screen;
836 //TODO: select these params by asic
837 switch (param) {
838 case PIPE_COMPUTE_CAP_IR_TARGET: {
839 const char *gpu = r600_llvm_gpu_string(rscreen->family);
840 if (ret) {
841 sprintf(ret, "%s-r600--", gpu);
842 }
843 return (8 + strlen(gpu)) * sizeof(char);
844 }
845 case PIPE_COMPUTE_CAP_GRID_DIMENSION:
846 if (ret) {
847 uint64_t * grid_dimension = ret;
848 grid_dimension[0] = 3;
849 }
850 return 1 * sizeof(uint64_t);
851
852 case PIPE_COMPUTE_CAP_MAX_GRID_SIZE:
853 if (ret) {
854 uint64_t * grid_size = ret;
855 grid_size[0] = 65535;
856 grid_size[1] = 65535;
857 grid_size[2] = 1;
858 }
859 return 3 * sizeof(uint64_t) ;
860
861 case PIPE_COMPUTE_CAP_MAX_BLOCK_SIZE:
862 if (ret) {
863 uint64_t * block_size = ret;
864 block_size[0] = 256;
865 block_size[1] = 256;
866 block_size[2] = 256;
867 }
868 return 3 * sizeof(uint64_t);
869
870 case PIPE_COMPUTE_CAP_MAX_THREADS_PER_BLOCK:
871 if (ret) {
872 uint64_t * max_threads_per_block = ret;
873 *max_threads_per_block = 256;
874 }
875 return sizeof(uint64_t);
876
877 case PIPE_COMPUTE_CAP_MAX_GLOBAL_SIZE:
878 if (ret) {
879 uint64_t * max_global_size = ret;
880 /* XXX: This is what the proprietary driver reports, we
881 * may want to use a different value. */
882 *max_global_size = 201326592;
883 }
884 return sizeof(uint64_t);
885
886 case PIPE_COMPUTE_CAP_MAX_INPUT_SIZE:
887 if (ret) {
888 uint64_t * max_input_size = ret;
889 *max_input_size = 1024;
890 }
891 return sizeof(uint64_t);
892
893 case PIPE_COMPUTE_CAP_MAX_LOCAL_SIZE:
894 if (ret) {
895 uint64_t * max_local_size = ret;
896 /* XXX: This is what the proprietary driver reports, we
897 * may want to use a different value. */
898 *max_local_size = 32768;
899 }
900 return sizeof(uint64_t);
901
902 case PIPE_COMPUTE_CAP_MAX_MEM_ALLOC_SIZE:
903 if (ret) {
904 uint64_t max_global_size;
905 uint64_t * max_mem_alloc_size = ret;
906 r600_get_compute_param(screen,
907 PIPE_COMPUTE_CAP_MAX_GLOBAL_SIZE,
908 &max_global_size);
909 /* OpenCL requres this value be at least
910 * max(MAX_GLOBAL_SIZE / 4, 128 * 1024 *1024)
911 * I'm really not sure what value to report here, but
912 * MAX_GLOBAL_SIZE / 4 seems resonable.
913 */
914 *max_mem_alloc_size = max_global_size / 4;
915 }
916 return sizeof(uint64_t);
917
918 default:
919 fprintf(stderr, "unknown PIPE_COMPUTE_CAP %d\n", param);
920 return 0;
921 }
922 }
923
924 static void r600_destroy_screen(struct pipe_screen* pscreen)
925 {
926 struct r600_screen *rscreen = (struct r600_screen *)pscreen;
927
928 if (rscreen == NULL)
929 return;
930
931 if (rscreen->global_pool) {
932 compute_memory_pool_delete(rscreen->global_pool);
933 }
934
935 if (rscreen->fences.bo) {
936 struct r600_fence_block *entry, *tmp;
937
938 LIST_FOR_EACH_ENTRY_SAFE(entry, tmp, &rscreen->fences.blocks, head) {
939 LIST_DEL(&entry->head);
940 FREE(entry);
941 }
942
943 rscreen->ws->buffer_unmap(rscreen->fences.bo->cs_buf);
944 pipe_resource_reference((struct pipe_resource**)&rscreen->fences.bo, NULL);
945 }
946 #if R600_TRACE_CS
947 if (rscreen->trace_bo) {
948 rscreen->ws->buffer_unmap(rscreen->trace_bo->cs_buf);
949 pipe_resource_reference((struct pipe_resource**)&rscreen->trace_bo, NULL);
950 }
951 #endif
952 pipe_mutex_destroy(rscreen->fences.mutex);
953
954 rscreen->ws->destroy(rscreen->ws);
955 FREE(rscreen);
956 }
957
958 static void r600_fence_reference(struct pipe_screen *pscreen,
959 struct pipe_fence_handle **ptr,
960 struct pipe_fence_handle *fence)
961 {
962 struct r600_fence **oldf = (struct r600_fence**)ptr;
963 struct r600_fence *newf = (struct r600_fence*)fence;
964
965 if (pipe_reference(&(*oldf)->reference, &newf->reference)) {
966 struct r600_screen *rscreen = (struct r600_screen *)pscreen;
967 pipe_mutex_lock(rscreen->fences.mutex);
968 pipe_resource_reference((struct pipe_resource**)&(*oldf)->sleep_bo, NULL);
969 LIST_ADDTAIL(&(*oldf)->head, &rscreen->fences.pool);
970 pipe_mutex_unlock(rscreen->fences.mutex);
971 }
972
973 *ptr = fence;
974 }
975
976 static boolean r600_fence_signalled(struct pipe_screen *pscreen,
977 struct pipe_fence_handle *fence)
978 {
979 struct r600_screen *rscreen = (struct r600_screen *)pscreen;
980 struct r600_fence *rfence = (struct r600_fence*)fence;
981
982 return rscreen->fences.data[rfence->index] != 0;
983 }
984
985 static boolean r600_fence_finish(struct pipe_screen *pscreen,
986 struct pipe_fence_handle *fence,
987 uint64_t timeout)
988 {
989 struct r600_screen *rscreen = (struct r600_screen *)pscreen;
990 struct r600_fence *rfence = (struct r600_fence*)fence;
991 int64_t start_time = 0;
992 unsigned spins = 0;
993
994 if (timeout != PIPE_TIMEOUT_INFINITE) {
995 start_time = os_time_get();
996
997 /* Convert to microseconds. */
998 timeout /= 1000;
999 }
1000
1001 while (rscreen->fences.data[rfence->index] == 0) {
1002 /* Special-case infinite timeout - wait for the dummy BO to become idle */
1003 if (timeout == PIPE_TIMEOUT_INFINITE) {
1004 rscreen->ws->buffer_wait(rfence->sleep_bo->buf, RADEON_USAGE_READWRITE);
1005 break;
1006 }
1007
1008 /* The dummy BO will be busy until the CS including the fence has completed, or
1009 * the GPU is reset. Don't bother continuing to spin when the BO is idle. */
1010 if (!rscreen->ws->buffer_is_busy(rfence->sleep_bo->buf, RADEON_USAGE_READWRITE))
1011 break;
1012
1013 if (++spins % 256)
1014 continue;
1015 #ifdef PIPE_OS_UNIX
1016 sched_yield();
1017 #else
1018 os_time_sleep(10);
1019 #endif
1020 if (timeout != PIPE_TIMEOUT_INFINITE &&
1021 os_time_get() - start_time >= timeout) {
1022 break;
1023 }
1024 }
1025
1026 return rscreen->fences.data[rfence->index] != 0;
1027 }
1028
1029 static int r600_interpret_tiling(struct r600_screen *rscreen, uint32_t tiling_config)
1030 {
1031 switch ((tiling_config & 0xe) >> 1) {
1032 case 0:
1033 rscreen->tiling_info.num_channels = 1;
1034 break;
1035 case 1:
1036 rscreen->tiling_info.num_channels = 2;
1037 break;
1038 case 2:
1039 rscreen->tiling_info.num_channels = 4;
1040 break;
1041 case 3:
1042 rscreen->tiling_info.num_channels = 8;
1043 break;
1044 default:
1045 return -EINVAL;
1046 }
1047
1048 switch ((tiling_config & 0x30) >> 4) {
1049 case 0:
1050 rscreen->tiling_info.num_banks = 4;
1051 break;
1052 case 1:
1053 rscreen->tiling_info.num_banks = 8;
1054 break;
1055 default:
1056 return -EINVAL;
1057
1058 }
1059 switch ((tiling_config & 0xc0) >> 6) {
1060 case 0:
1061 rscreen->tiling_info.group_bytes = 256;
1062 break;
1063 case 1:
1064 rscreen->tiling_info.group_bytes = 512;
1065 break;
1066 default:
1067 return -EINVAL;
1068 }
1069 return 0;
1070 }
1071
1072 static int evergreen_interpret_tiling(struct r600_screen *rscreen, uint32_t tiling_config)
1073 {
1074 switch (tiling_config & 0xf) {
1075 case 0:
1076 rscreen->tiling_info.num_channels = 1;
1077 break;
1078 case 1:
1079 rscreen->tiling_info.num_channels = 2;
1080 break;
1081 case 2:
1082 rscreen->tiling_info.num_channels = 4;
1083 break;
1084 case 3:
1085 rscreen->tiling_info.num_channels = 8;
1086 break;
1087 default:
1088 return -EINVAL;
1089 }
1090
1091 switch ((tiling_config & 0xf0) >> 4) {
1092 case 0:
1093 rscreen->tiling_info.num_banks = 4;
1094 break;
1095 case 1:
1096 rscreen->tiling_info.num_banks = 8;
1097 break;
1098 case 2:
1099 rscreen->tiling_info.num_banks = 16;
1100 break;
1101 default:
1102 return -EINVAL;
1103 }
1104
1105 switch ((tiling_config & 0xf00) >> 8) {
1106 case 0:
1107 rscreen->tiling_info.group_bytes = 256;
1108 break;
1109 case 1:
1110 rscreen->tiling_info.group_bytes = 512;
1111 break;
1112 default:
1113 return -EINVAL;
1114 }
1115 return 0;
1116 }
1117
1118 static int r600_init_tiling(struct r600_screen *rscreen)
1119 {
1120 uint32_t tiling_config = rscreen->info.r600_tiling_config;
1121
1122 /* set default group bytes, overridden by tiling info ioctl */
1123 if (rscreen->chip_class <= R700) {
1124 rscreen->tiling_info.group_bytes = 256;
1125 } else {
1126 rscreen->tiling_info.group_bytes = 512;
1127 }
1128
1129 if (!tiling_config)
1130 return 0;
1131
1132 if (rscreen->chip_class <= R700) {
1133 return r600_interpret_tiling(rscreen, tiling_config);
1134 } else {
1135 return evergreen_interpret_tiling(rscreen, tiling_config);
1136 }
1137 }
1138
1139 static uint64_t r600_get_timestamp(struct pipe_screen *screen)
1140 {
1141 struct r600_screen *rscreen = (struct r600_screen*)screen;
1142
1143 return 1000000 * rscreen->ws->query_timestamp(rscreen->ws) /
1144 rscreen->info.r600_clock_crystal_freq;
1145 }
1146
1147 static int r600_get_driver_query_info(struct pipe_screen *screen,
1148 unsigned index,
1149 struct pipe_driver_query_info *info)
1150 {
1151 struct r600_screen *rscreen = (struct r600_screen*)screen;
1152 struct pipe_driver_query_info list[] = {
1153 {"draw-calls", R600_QUERY_DRAW_CALLS, 0},
1154 {"requested-VRAM", R600_QUERY_REQUESTED_VRAM, rscreen->info.vram_size, TRUE},
1155 {"requested-GTT", R600_QUERY_REQUESTED_GTT, rscreen->info.gart_size, TRUE},
1156 };
1157
1158 if (!info)
1159 return Elements(list);
1160
1161 if (index >= Elements(list))
1162 return 0;
1163
1164 *info = list[index];
1165 return 1;
1166 }
1167
1168 struct pipe_screen *r600_screen_create(struct radeon_winsys *ws)
1169 {
1170 struct r600_screen *rscreen = CALLOC_STRUCT(r600_screen);
1171
1172 if (rscreen == NULL) {
1173 return NULL;
1174 }
1175
1176 rscreen->ws = ws;
1177 ws->query_info(ws, &rscreen->info);
1178
1179 rscreen->debug_flags = debug_get_flags_option("R600_DEBUG", debug_options, 0);
1180 if (debug_get_bool_option("R600_DEBUG_COMPUTE", FALSE))
1181 rscreen->debug_flags |= DBG_COMPUTE;
1182 if (debug_get_bool_option("R600_DUMP_SHADERS", FALSE))
1183 rscreen->debug_flags |= DBG_FS | DBG_VS | DBG_GS | DBG_PS | DBG_CS;
1184 if (!debug_get_bool_option("R600_HYPERZ", TRUE))
1185 rscreen->debug_flags |= DBG_NO_HYPERZ;
1186 if (!debug_get_bool_option("R600_LLVM", TRUE))
1187 rscreen->debug_flags |= DBG_NO_LLVM;
1188 if (debug_get_bool_option("R600_PRINT_TEXDEPTH", FALSE))
1189 rscreen->debug_flags |= DBG_TEX_DEPTH;
1190 rscreen->family = rscreen->info.family;
1191 rscreen->chip_class = rscreen->info.chip_class;
1192
1193 if (rscreen->family == CHIP_UNKNOWN) {
1194 fprintf(stderr, "r600: Unknown chipset 0x%04X\n", rscreen->info.pci_id);
1195 FREE(rscreen);
1196 return NULL;
1197 }
1198
1199 /* Figure out streamout kernel support. */
1200 switch (rscreen->chip_class) {
1201 case R600:
1202 if (rscreen->family < CHIP_RS780) {
1203 rscreen->has_streamout = rscreen->info.drm_minor >= 14;
1204 } else {
1205 rscreen->has_streamout = rscreen->info.drm_minor >= 23;
1206 }
1207 break;
1208 case R700:
1209 rscreen->has_streamout = rscreen->info.drm_minor >= 17;
1210 break;
1211 case EVERGREEN:
1212 case CAYMAN:
1213 rscreen->has_streamout = rscreen->info.drm_minor >= 14;
1214 break;
1215 default:
1216 rscreen->has_streamout = FALSE;
1217 break;
1218 }
1219
1220 /* MSAA support. */
1221 switch (rscreen->chip_class) {
1222 case R600:
1223 case R700:
1224 rscreen->has_msaa = rscreen->info.drm_minor >= 22;
1225 rscreen->msaa_texture_support = MSAA_TEXTURE_DECOMPRESSED;
1226 break;
1227 case EVERGREEN:
1228 rscreen->has_msaa = rscreen->info.drm_minor >= 19;
1229 rscreen->msaa_texture_support =
1230 rscreen->info.drm_minor >= 24 ? MSAA_TEXTURE_COMPRESSED :
1231 MSAA_TEXTURE_DECOMPRESSED;
1232 break;
1233 case CAYMAN:
1234 rscreen->has_msaa = rscreen->info.drm_minor >= 19;
1235 /* We should be able to read compressed MSAA textures, but it doesn't work. */
1236 rscreen->msaa_texture_support = MSAA_TEXTURE_SAMPLE_ZERO;
1237 break;
1238 default:
1239 rscreen->has_msaa = FALSE;
1240 rscreen->msaa_texture_support = 0;
1241 break;
1242 }
1243
1244 rscreen->has_cp_dma = rscreen->info.drm_minor >= 27 &&
1245 !(rscreen->debug_flags & DBG_NO_CP_DMA);
1246
1247 if (r600_init_tiling(rscreen)) {
1248 FREE(rscreen);
1249 return NULL;
1250 }
1251
1252 rscreen->screen.destroy = r600_destroy_screen;
1253 rscreen->screen.get_name = r600_get_name;
1254 rscreen->screen.get_vendor = r600_get_vendor;
1255 rscreen->screen.get_param = r600_get_param;
1256 rscreen->screen.get_shader_param = r600_get_shader_param;
1257 rscreen->screen.get_paramf = r600_get_paramf;
1258 rscreen->screen.get_video_param = r600_get_video_param;
1259 rscreen->screen.get_compute_param = r600_get_compute_param;
1260 rscreen->screen.get_timestamp = r600_get_timestamp;
1261
1262 if (rscreen->chip_class >= EVERGREEN) {
1263 rscreen->screen.is_format_supported = evergreen_is_format_supported;
1264 rscreen->dma_blit = &evergreen_dma_blit;
1265 } else {
1266 rscreen->screen.is_format_supported = r600_is_format_supported;
1267 rscreen->dma_blit = &r600_dma_blit;
1268 }
1269 rscreen->screen.is_video_format_supported = vl_video_buffer_is_format_supported;
1270 rscreen->screen.context_create = r600_create_context;
1271 rscreen->screen.fence_reference = r600_fence_reference;
1272 rscreen->screen.fence_signalled = r600_fence_signalled;
1273 rscreen->screen.fence_finish = r600_fence_finish;
1274 rscreen->screen.get_driver_query_info = r600_get_driver_query_info;
1275 r600_init_screen_resource_functions(&rscreen->screen);
1276
1277 util_format_s3tc_init();
1278
1279 rscreen->fences.bo = NULL;
1280 rscreen->fences.data = NULL;
1281 rscreen->fences.next_index = 0;
1282 LIST_INITHEAD(&rscreen->fences.pool);
1283 LIST_INITHEAD(&rscreen->fences.blocks);
1284 pipe_mutex_init(rscreen->fences.mutex);
1285
1286 rscreen->global_pool = compute_memory_pool_new(rscreen);
1287
1288 #if R600_TRACE_CS
1289 rscreen->cs_count = 0;
1290 if (rscreen->info.drm_minor >= 28) {
1291 rscreen->trace_bo = (struct r600_resource*)pipe_buffer_create(&rscreen->screen,
1292 PIPE_BIND_CUSTOM,
1293 PIPE_USAGE_STAGING,
1294 4096);
1295 if (rscreen->trace_bo) {
1296 rscreen->trace_ptr = rscreen->ws->buffer_map(rscreen->trace_bo->cs_buf, NULL,
1297 PIPE_TRANSFER_UNSYNCHRONIZED);
1298 }
1299 }
1300 #endif
1301
1302 return &rscreen->screen;
1303 }