2 * Copyright 2010 Jerome Glisse <glisse@freedesktop.org>
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * on the rights to use, copy, modify, merge, publish, distribute, sub
8 * license, and/or sell copies of the Software, and to permit persons to whom
9 * the Software is furnished to do so, subject to the following conditions:
11 * The above copyright notice and this permission notice (including the next
12 * paragraph) shall be included in all copies or substantial portions of the
15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17 * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL
18 * THE AUTHOR(S) AND/OR THEIR SUPPLIERS BE LIABLE FOR ANY CLAIM,
19 * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
20 * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
21 * USE OR OTHER DEALINGS IN THE SOFTWARE.
23 #include "r600_pipe.h"
24 #include "r600_public.h"
26 #include "evergreen_compute.h"
30 #include "pipe/p_shader_tokens.h"
31 #include "util/u_blitter.h"
32 #include "util/u_debug.h"
33 #include "util/u_format_s3tc.h"
34 #include "util/u_memory.h"
35 #include "util/u_simple_shaders.h"
36 #include "util/u_upload_mgr.h"
37 #include "util/u_math.h"
38 #include "vl/vl_decoder.h"
39 #include "vl/vl_video_buffer.h"
40 #include "radeon/radeon_uvd.h"
41 #include "os/os_time.h"
43 static const struct debug_named_value debug_options
[] = {
45 { "texdepth", DBG_TEX_DEPTH
, "Print texture depth info" },
46 { "compute", DBG_COMPUTE
, "Print compute info" },
47 { "vm", DBG_VM
, "Print virtual addresses when creating resources" },
50 { "fs", DBG_FS
, "Print fetch shaders" },
51 { "vs", DBG_VS
, "Print vertex shaders" },
52 { "gs", DBG_GS
, "Print geometry shaders" },
53 { "ps", DBG_PS
, "Print pixel shaders" },
54 { "cs", DBG_CS
, "Print compute shaders" },
57 { "nohyperz", DBG_NO_HYPERZ
, "Disable Hyper-Z" },
58 #if defined(R600_USE_LLVM)
59 { "nollvm", DBG_NO_LLVM
, "Disable the LLVM shader compiler" },
61 { "nocpdma", DBG_NO_CP_DMA
, "Disable CP DMA" },
62 { "nodma", DBG_NO_ASYNC_DMA
, "Disable asynchronous DMA" },
63 /* GL uses the word INVALIDATE, gallium uses the word DISCARD */
64 { "noinvalrange", DBG_NO_DISCARD_RANGE
, "Disable handling of INVALIDATE_RANGE map flags" },
66 DEBUG_NAMED_VALUE_END
/* must be last */
72 static struct r600_fence
*r600_create_fence(struct r600_context
*rctx
)
74 struct r600_screen
*rscreen
= rctx
->screen
;
75 struct r600_fence
*fence
= NULL
;
77 pipe_mutex_lock(rscreen
->fences
.mutex
);
79 if (!rscreen
->fences
.bo
) {
80 /* Create the shared buffer object */
81 rscreen
->fences
.bo
= (struct r600_resource
*)
82 pipe_buffer_create(&rscreen
->screen
, PIPE_BIND_CUSTOM
,
83 PIPE_USAGE_STAGING
, 4096);
84 if (!rscreen
->fences
.bo
) {
85 R600_ERR("r600: failed to create bo for fence objects\n");
88 rscreen
->fences
.data
= r600_buffer_mmap_sync_with_rings(rctx
, rscreen
->fences
.bo
, PIPE_TRANSFER_READ_WRITE
);
91 if (!LIST_IS_EMPTY(&rscreen
->fences
.pool
)) {
92 struct r600_fence
*entry
;
94 /* Try to find a freed fence that has been signalled */
95 LIST_FOR_EACH_ENTRY(entry
, &rscreen
->fences
.pool
, head
) {
96 if (rscreen
->fences
.data
[entry
->index
] != 0) {
97 LIST_DELINIT(&entry
->head
);
105 /* Allocate a new fence */
106 struct r600_fence_block
*block
;
109 if ((rscreen
->fences
.next_index
+ 1) >= 1024) {
110 R600_ERR("r600: too many concurrent fences\n");
114 index
= rscreen
->fences
.next_index
++;
116 if (!(index
% FENCE_BLOCK_SIZE
)) {
117 /* Allocate a new block */
118 block
= CALLOC_STRUCT(r600_fence_block
);
122 LIST_ADD(&block
->head
, &rscreen
->fences
.blocks
);
124 block
= LIST_ENTRY(struct r600_fence_block
, rscreen
->fences
.blocks
.next
, head
);
127 fence
= &block
->fences
[index
% FENCE_BLOCK_SIZE
];
128 fence
->index
= index
;
131 pipe_reference_init(&fence
->reference
, 1);
133 rscreen
->fences
.data
[fence
->index
] = 0;
134 r600_context_emit_fence(rctx
, rscreen
->fences
.bo
, fence
->index
, 1);
136 /* Create a dummy BO so that fence_finish without a timeout can sleep waiting for completion */
137 fence
->sleep_bo
= (struct r600_resource
*)
138 pipe_buffer_create(&rctx
->screen
->screen
, PIPE_BIND_CUSTOM
,
139 PIPE_USAGE_STAGING
, 1);
140 /* Add the fence as a dummy relocation. */
141 r600_context_bo_reloc(rctx
, &rctx
->rings
.gfx
, fence
->sleep_bo
, RADEON_USAGE_READWRITE
);
144 pipe_mutex_unlock(rscreen
->fences
.mutex
);
148 static void r600_flush(struct pipe_context
*ctx
, unsigned flags
)
150 struct r600_context
*rctx
= (struct r600_context
*)ctx
;
151 struct pipe_query
*render_cond
= NULL
;
152 unsigned render_cond_mode
= 0;
154 rctx
->rings
.gfx
.flushing
= true;
155 /* Disable render condition. */
156 if (rctx
->current_render_cond
) {
157 render_cond
= rctx
->current_render_cond
;
158 render_cond_mode
= rctx
->current_render_cond_mode
;
159 ctx
->render_condition(ctx
, NULL
, 0);
162 r600_context_flush(rctx
, flags
);
163 rctx
->rings
.gfx
.flushing
= false;
164 r600_begin_new_cs(rctx
);
166 /* Re-enable render condition. */
168 ctx
->render_condition(ctx
, render_cond
, render_cond_mode
);
172 static void r600_flush_from_st(struct pipe_context
*ctx
,
173 struct pipe_fence_handle
**fence
,
174 enum pipe_flush_flags flags
)
176 struct r600_context
*rctx
= (struct r600_context
*)ctx
;
177 struct r600_fence
**rfence
= (struct r600_fence
**)fence
;
180 fflags
= flags
& PIPE_FLUSH_END_OF_FRAME
? RADEON_FLUSH_END_OF_FRAME
: 0;
182 *rfence
= r600_create_fence(rctx
);
184 /* flush gfx & dma ring, order does not matter as only one can be live */
185 if (rctx
->rings
.dma
.cs
) {
186 rctx
->rings
.dma
.flush(rctx
, fflags
);
188 rctx
->rings
.gfx
.flush(rctx
, fflags
);
191 static void r600_flush_gfx_ring(void *ctx
, unsigned flags
)
193 r600_flush((struct pipe_context
*)ctx
, flags
);
196 static void r600_flush_dma_ring(void *ctx
, unsigned flags
)
198 struct r600_context
*rctx
= (struct r600_context
*)ctx
;
199 struct radeon_winsys_cs
*cs
= rctx
->rings
.dma
.cs
;
200 unsigned padding_dw
, i
;
206 /* Pad the DMA CS to a multiple of 8 dwords. */
207 padding_dw
= 8 - cs
->cdw
% 8;
208 if (padding_dw
< 8) {
209 for (i
= 0; i
< padding_dw
; i
++) {
210 cs
->buf
[cs
->cdw
++] = DMA_PACKET(DMA_PACKET_NOP
, 0, 0, 0);
214 rctx
->rings
.dma
.flushing
= true;
215 rctx
->ws
->cs_flush(cs
, flags
);
216 rctx
->rings
.dma
.flushing
= false;
219 boolean
r600_rings_is_buffer_referenced(struct r600_context
*ctx
,
220 struct radeon_winsys_cs_handle
*buf
,
221 enum radeon_bo_usage usage
)
223 if (ctx
->ws
->cs_is_buffer_referenced(ctx
->rings
.gfx
.cs
, buf
, usage
)) {
226 if (ctx
->rings
.dma
.cs
) {
227 if (ctx
->ws
->cs_is_buffer_referenced(ctx
->rings
.dma
.cs
, buf
, usage
)) {
234 void *r600_buffer_mmap_sync_with_rings(struct r600_context
*ctx
,
235 struct r600_resource
*resource
,
238 enum radeon_bo_usage rusage
= RADEON_USAGE_READWRITE
;
240 bool sync_flush
= TRUE
;
242 if (usage
& PIPE_TRANSFER_UNSYNCHRONIZED
) {
243 return ctx
->ws
->buffer_map(resource
->cs_buf
, NULL
, usage
);
246 if (!(usage
& PIPE_TRANSFER_WRITE
)) {
247 /* have to wait for pending read */
248 rusage
= RADEON_USAGE_WRITE
;
250 if (usage
& PIPE_TRANSFER_DONTBLOCK
) {
251 flags
|= RADEON_FLUSH_ASYNC
;
254 if (ctx
->ws
->cs_is_buffer_referenced(ctx
->rings
.gfx
.cs
, resource
->cs_buf
, rusage
) && ctx
->rings
.gfx
.cs
->cdw
) {
255 ctx
->rings
.gfx
.flush(ctx
, flags
);
256 if (usage
& PIPE_TRANSFER_DONTBLOCK
) {
260 if (ctx
->rings
.dma
.cs
) {
261 if (ctx
->ws
->cs_is_buffer_referenced(ctx
->rings
.dma
.cs
, resource
->cs_buf
, rusage
) && ctx
->rings
.dma
.cs
->cdw
) {
262 ctx
->rings
.dma
.flush(ctx
, flags
);
263 if (usage
& PIPE_TRANSFER_DONTBLOCK
) {
269 if (usage
& PIPE_TRANSFER_DONTBLOCK
) {
270 if (ctx
->ws
->buffer_is_busy(resource
->buf
, rusage
)) {
275 /* Try to avoid busy-waiting in radeon_bo_wait. */
276 ctx
->ws
->cs_sync_flush(ctx
->rings
.gfx
.cs
);
277 if (ctx
->rings
.dma
.cs
) {
278 ctx
->ws
->cs_sync_flush(ctx
->rings
.dma
.cs
);
281 ctx
->ws
->buffer_wait(resource
->buf
, rusage
);
283 /* at this point everything is synchronized */
284 return ctx
->ws
->buffer_map(resource
->cs_buf
, NULL
, usage
| PIPE_TRANSFER_UNSYNCHRONIZED
);
287 static void r600_flush_from_winsys(void *ctx
, unsigned flags
)
289 struct r600_context
*rctx
= (struct r600_context
*)ctx
;
291 rctx
->rings
.gfx
.flush(rctx
, flags
);
294 static void r600_flush_dma_from_winsys(void *ctx
, unsigned flags
)
296 struct r600_context
*rctx
= (struct r600_context
*)ctx
;
298 rctx
->rings
.dma
.flush(rctx
, flags
);
301 static void r600_destroy_context(struct pipe_context
*context
)
303 struct r600_context
*rctx
= (struct r600_context
*)context
;
305 r600_isa_destroy(rctx
->isa
);
307 pipe_resource_reference((struct pipe_resource
**)&rctx
->dummy_cmask
, NULL
);
308 pipe_resource_reference((struct pipe_resource
**)&rctx
->dummy_fmask
, NULL
);
310 if (rctx
->dummy_pixel_shader
) {
311 rctx
->context
.delete_fs_state(&rctx
->context
, rctx
->dummy_pixel_shader
);
313 if (rctx
->custom_dsa_flush
) {
314 rctx
->context
.delete_depth_stencil_alpha_state(&rctx
->context
, rctx
->custom_dsa_flush
);
316 if (rctx
->custom_blend_resolve
) {
317 rctx
->context
.delete_blend_state(&rctx
->context
, rctx
->custom_blend_resolve
);
319 if (rctx
->custom_blend_decompress
) {
320 rctx
->context
.delete_blend_state(&rctx
->context
, rctx
->custom_blend_decompress
);
322 if (rctx
->custom_blend_fmask_decompress
) {
323 rctx
->context
.delete_blend_state(&rctx
->context
, rctx
->custom_blend_fmask_decompress
);
325 util_unreference_framebuffer_state(&rctx
->framebuffer
.state
);
328 util_blitter_destroy(rctx
->blitter
);
330 if (rctx
->uploader
) {
331 u_upload_destroy(rctx
->uploader
);
333 if (rctx
->allocator_so_filled_size
) {
334 u_suballocator_destroy(rctx
->allocator_so_filled_size
);
336 if (rctx
->allocator_fetch_shader
) {
337 u_suballocator_destroy(rctx
->allocator_fetch_shader
);
339 util_slab_destroy(&rctx
->pool_transfers
);
341 r600_release_command_buffer(&rctx
->start_cs_cmd
);
343 if (rctx
->rings
.gfx
.cs
) {
344 rctx
->ws
->cs_destroy(rctx
->rings
.gfx
.cs
);
346 if (rctx
->rings
.dma
.cs
) {
347 rctx
->ws
->cs_destroy(rctx
->rings
.dma
.cs
);
353 static struct pipe_context
*r600_create_context(struct pipe_screen
*screen
, void *priv
)
355 struct r600_context
*rctx
= CALLOC_STRUCT(r600_context
);
356 struct r600_screen
* rscreen
= (struct r600_screen
*)screen
;
361 util_slab_create(&rctx
->pool_transfers
,
362 sizeof(struct r600_transfer
), 64,
363 UTIL_SLAB_SINGLETHREADED
);
365 rctx
->context
.screen
= screen
;
366 rctx
->context
.priv
= priv
;
367 rctx
->context
.destroy
= r600_destroy_context
;
368 rctx
->context
.flush
= r600_flush_from_st
;
370 /* Easy accessing of screen/winsys. */
371 rctx
->screen
= rscreen
;
372 rctx
->ws
= rscreen
->ws
;
373 rctx
->family
= rscreen
->family
;
374 rctx
->chip_class
= rscreen
->chip_class
;
375 rctx
->keep_tiling_flags
= rscreen
->info
.drm_minor
>= 12;
377 LIST_INITHEAD(&rctx
->active_nontimer_queries
);
379 r600_init_blit_functions(rctx
);
380 r600_init_query_functions(rctx
);
381 r600_init_context_resource_functions(rctx
);
382 r600_init_surface_functions(rctx
);
384 if (rscreen
->info
.has_uvd
) {
385 rctx
->context
.create_video_decoder
= r600_uvd_create_decoder
;
386 rctx
->context
.create_video_buffer
= r600_video_buffer_create
;
388 rctx
->context
.create_video_decoder
= vl_create_decoder
;
389 rctx
->context
.create_video_buffer
= vl_video_buffer_create
;
392 r600_init_common_state_functions(rctx
);
394 switch (rctx
->chip_class
) {
397 r600_init_state_functions(rctx
);
398 r600_init_atom_start_cs(rctx
);
400 rctx
->custom_dsa_flush
= r600_create_db_flush_dsa(rctx
);
401 rctx
->custom_blend_resolve
= rctx
->chip_class
== R700
? r700_create_resolve_blend(rctx
)
402 : r600_create_resolve_blend(rctx
);
403 rctx
->custom_blend_decompress
= r600_create_decompress_blend(rctx
);
404 rctx
->has_vertex_cache
= !(rctx
->family
== CHIP_RV610
||
405 rctx
->family
== CHIP_RV620
||
406 rctx
->family
== CHIP_RS780
||
407 rctx
->family
== CHIP_RS880
||
408 rctx
->family
== CHIP_RV710
);
412 evergreen_init_state_functions(rctx
);
413 evergreen_init_atom_start_cs(rctx
);
414 evergreen_init_atom_start_compute_cs(rctx
);
416 rctx
->custom_dsa_flush
= evergreen_create_db_flush_dsa(rctx
);
417 rctx
->custom_blend_resolve
= evergreen_create_resolve_blend(rctx
);
418 rctx
->custom_blend_decompress
= evergreen_create_decompress_blend(rctx
);
419 rctx
->custom_blend_fmask_decompress
= evergreen_create_fmask_decompress_blend(rctx
);
420 rctx
->has_vertex_cache
= !(rctx
->family
== CHIP_CEDAR
||
421 rctx
->family
== CHIP_PALM
||
422 rctx
->family
== CHIP_SUMO
||
423 rctx
->family
== CHIP_SUMO2
||
424 rctx
->family
== CHIP_CAICOS
||
425 rctx
->family
== CHIP_CAYMAN
||
426 rctx
->family
== CHIP_ARUBA
);
429 R600_ERR("Unsupported chip class %d.\n", rctx
->chip_class
);
433 rctx
->rings
.gfx
.cs
= rctx
->ws
->cs_create(rctx
->ws
, RING_GFX
);
434 rctx
->rings
.gfx
.flush
= r600_flush_gfx_ring
;
435 rctx
->ws
->cs_set_flush_callback(rctx
->rings
.gfx
.cs
, r600_flush_from_winsys
, rctx
);
436 rctx
->rings
.gfx
.flushing
= false;
438 rctx
->rings
.dma
.cs
= NULL
;
439 if (rscreen
->info
.r600_has_dma
&& !(rscreen
->debug_flags
& DBG_NO_ASYNC_DMA
)) {
440 rctx
->rings
.dma
.cs
= rctx
->ws
->cs_create(rctx
->ws
, RING_DMA
);
441 rctx
->rings
.dma
.flush
= r600_flush_dma_ring
;
442 rctx
->ws
->cs_set_flush_callback(rctx
->rings
.dma
.cs
, r600_flush_dma_from_winsys
, rctx
);
443 rctx
->rings
.dma
.flushing
= false;
446 rctx
->uploader
= u_upload_create(&rctx
->context
, 1024 * 1024, 256,
447 PIPE_BIND_INDEX_BUFFER
|
448 PIPE_BIND_CONSTANT_BUFFER
);
452 rctx
->allocator_fetch_shader
= u_suballocator_create(&rctx
->context
, 64 * 1024, 256,
453 0, PIPE_USAGE_STATIC
, FALSE
);
454 if (!rctx
->allocator_fetch_shader
)
457 rctx
->allocator_so_filled_size
= u_suballocator_create(&rctx
->context
, 4096, 4,
458 0, PIPE_USAGE_STATIC
, TRUE
);
459 if (!rctx
->allocator_so_filled_size
)
462 rctx
->isa
= calloc(1, sizeof(struct r600_isa
));
463 if (!rctx
->isa
|| r600_isa_init(rctx
, rctx
->isa
))
466 rctx
->blitter
= util_blitter_create(&rctx
->context
);
467 if (rctx
->blitter
== NULL
)
469 util_blitter_set_texture_multisample(rctx
->blitter
, rscreen
->has_msaa
);
470 rctx
->blitter
->draw_rectangle
= r600_draw_rectangle
;
472 r600_begin_new_cs(rctx
);
473 r600_get_backend_mask(rctx
); /* this emits commands and must be last */
475 rctx
->dummy_pixel_shader
=
476 util_make_fragment_cloneinput_shader(&rctx
->context
, 0,
477 TGSI_SEMANTIC_GENERIC
,
478 TGSI_INTERPOLATE_CONSTANT
);
479 rctx
->context
.bind_fs_state(&rctx
->context
, rctx
->dummy_pixel_shader
);
481 return &rctx
->context
;
484 r600_destroy_context(&rctx
->context
);
491 static const char* r600_get_vendor(struct pipe_screen
* pscreen
)
496 static const char *r600_get_family_name(enum radeon_family family
)
499 case CHIP_R600
: return "AMD R600";
500 case CHIP_RV610
: return "AMD RV610";
501 case CHIP_RV630
: return "AMD RV630";
502 case CHIP_RV670
: return "AMD RV670";
503 case CHIP_RV620
: return "AMD RV620";
504 case CHIP_RV635
: return "AMD RV635";
505 case CHIP_RS780
: return "AMD RS780";
506 case CHIP_RS880
: return "AMD RS880";
507 case CHIP_RV770
: return "AMD RV770";
508 case CHIP_RV730
: return "AMD RV730";
509 case CHIP_RV710
: return "AMD RV710";
510 case CHIP_RV740
: return "AMD RV740";
511 case CHIP_CEDAR
: return "AMD CEDAR";
512 case CHIP_REDWOOD
: return "AMD REDWOOD";
513 case CHIP_JUNIPER
: return "AMD JUNIPER";
514 case CHIP_CYPRESS
: return "AMD CYPRESS";
515 case CHIP_HEMLOCK
: return "AMD HEMLOCK";
516 case CHIP_PALM
: return "AMD PALM";
517 case CHIP_SUMO
: return "AMD SUMO";
518 case CHIP_SUMO2
: return "AMD SUMO2";
519 case CHIP_BARTS
: return "AMD BARTS";
520 case CHIP_TURKS
: return "AMD TURKS";
521 case CHIP_CAICOS
: return "AMD CAICOS";
522 case CHIP_CAYMAN
: return "AMD CAYMAN";
523 case CHIP_ARUBA
: return "AMD ARUBA";
524 default: return "AMD unknown";
528 static const char* r600_get_name(struct pipe_screen
* pscreen
)
530 struct r600_screen
*rscreen
= (struct r600_screen
*)pscreen
;
532 return r600_get_family_name(rscreen
->family
);
535 static int r600_get_param(struct pipe_screen
* pscreen
, enum pipe_cap param
)
537 struct r600_screen
*rscreen
= (struct r600_screen
*)pscreen
;
538 enum radeon_family family
= rscreen
->family
;
541 /* Supported features (boolean caps). */
542 case PIPE_CAP_NPOT_TEXTURES
:
543 case PIPE_CAP_TWO_SIDED_STENCIL
:
544 case PIPE_CAP_ANISOTROPIC_FILTER
:
545 case PIPE_CAP_POINT_SPRITE
:
546 case PIPE_CAP_OCCLUSION_QUERY
:
547 case PIPE_CAP_TEXTURE_SHADOW_MAP
:
548 case PIPE_CAP_TEXTURE_MIRROR_CLAMP
:
549 case PIPE_CAP_BLEND_EQUATION_SEPARATE
:
550 case PIPE_CAP_TEXTURE_SWIZZLE
:
551 case PIPE_CAP_DEPTH_CLIP_DISABLE
:
552 case PIPE_CAP_SHADER_STENCIL_EXPORT
:
553 case PIPE_CAP_VERTEX_ELEMENT_INSTANCE_DIVISOR
:
554 case PIPE_CAP_MIXED_COLORBUFFER_FORMATS
:
555 case PIPE_CAP_TGSI_FS_COORD_ORIGIN_UPPER_LEFT
:
556 case PIPE_CAP_TGSI_FS_COORD_PIXEL_CENTER_HALF_INTEGER
:
558 case PIPE_CAP_SEAMLESS_CUBE_MAP
:
559 case PIPE_CAP_PRIMITIVE_RESTART
:
560 case PIPE_CAP_CONDITIONAL_RENDER
:
561 case PIPE_CAP_TEXTURE_BARRIER
:
562 case PIPE_CAP_VERTEX_COLOR_UNCLAMPED
:
563 case PIPE_CAP_QUADS_FOLLOW_PROVOKING_VERTEX_CONVENTION
:
564 case PIPE_CAP_TGSI_INSTANCEID
:
565 case PIPE_CAP_VERTEX_BUFFER_OFFSET_4BYTE_ALIGNED_ONLY
:
566 case PIPE_CAP_VERTEX_BUFFER_STRIDE_4BYTE_ALIGNED_ONLY
:
567 case PIPE_CAP_VERTEX_ELEMENT_SRC_OFFSET_4BYTE_ALIGNED_ONLY
:
568 case PIPE_CAP_USER_INDEX_BUFFERS
:
569 case PIPE_CAP_USER_CONSTANT_BUFFERS
:
570 case PIPE_CAP_COMPUTE
:
571 case PIPE_CAP_START_INSTANCE
:
572 case PIPE_CAP_MAX_DUAL_SOURCE_RENDER_TARGETS
:
573 case PIPE_CAP_TEXTURE_BUFFER_OBJECTS
:
574 case PIPE_CAP_PREFER_BLIT_BASED_TEXTURE_TRANSFER
:
575 case PIPE_CAP_QUERY_PIPELINE_STATISTICS
:
577 case PIPE_CAP_TGSI_TEXCOORD
:
580 case PIPE_CAP_MIN_MAP_BUFFER_ALIGNMENT
:
581 return R600_MAP_BUFFER_ALIGNMENT
;
583 case PIPE_CAP_CONSTANT_BUFFER_OFFSET_ALIGNMENT
:
586 case PIPE_CAP_TEXTURE_BUFFER_OFFSET_ALIGNMENT
:
589 case PIPE_CAP_GLSL_FEATURE_LEVEL
:
592 case PIPE_CAP_TEXTURE_MULTISAMPLE
:
593 return rscreen
->msaa_texture_support
!= MSAA_TEXTURE_SAMPLE_ZERO
;
595 /* Supported except the original R600. */
596 case PIPE_CAP_INDEP_BLEND_ENABLE
:
597 case PIPE_CAP_INDEP_BLEND_FUNC
:
598 /* R600 doesn't support per-MRT blends */
599 return family
== CHIP_R600
? 0 : 1;
601 /* Supported on Evergreen. */
602 case PIPE_CAP_SEAMLESS_CUBE_MAP_PER_TEXTURE
:
603 case PIPE_CAP_CUBE_MAP_ARRAY
:
604 return family
>= CHIP_CEDAR
? 1 : 0;
606 /* Unsupported features. */
607 case PIPE_CAP_TGSI_FS_COORD_ORIGIN_LOWER_LEFT
:
608 case PIPE_CAP_TGSI_FS_COORD_PIXEL_CENTER_INTEGER
:
609 case PIPE_CAP_SCALED_RESOLVE
:
610 case PIPE_CAP_TGSI_CAN_COMPACT_VARYINGS
:
611 case PIPE_CAP_TGSI_CAN_COMPACT_CONSTANTS
:
612 case PIPE_CAP_FRAGMENT_COLOR_CLAMPED
:
613 case PIPE_CAP_VERTEX_COLOR_CLAMPED
:
614 case PIPE_CAP_USER_VERTEX_BUFFERS
:
618 case PIPE_CAP_MAX_STREAM_OUTPUT_BUFFERS
:
619 return rscreen
->has_streamout
? 4 : 0;
620 case PIPE_CAP_STREAM_OUTPUT_PAUSE_RESUME
:
621 return rscreen
->has_streamout
? 1 : 0;
622 case PIPE_CAP_MAX_STREAM_OUTPUT_SEPARATE_COMPONENTS
:
623 case PIPE_CAP_MAX_STREAM_OUTPUT_INTERLEAVED_COMPONENTS
:
627 case PIPE_CAP_MAX_TEXTURE_2D_LEVELS
:
628 case PIPE_CAP_MAX_TEXTURE_3D_LEVELS
:
629 case PIPE_CAP_MAX_TEXTURE_CUBE_LEVELS
:
630 if (family
>= CHIP_CEDAR
)
634 case PIPE_CAP_MAX_TEXTURE_ARRAY_LAYERS
:
635 return rscreen
->info
.drm_minor
>= 9 ?
636 (family
>= CHIP_CEDAR
? 16384 : 8192) : 0;
637 case PIPE_CAP_MAX_COMBINED_SAMPLERS
:
640 /* Render targets. */
641 case PIPE_CAP_MAX_RENDER_TARGETS
:
642 /* XXX some r6xx are buggy and can only do 4 */
645 /* Timer queries, present when the clock frequency is non zero. */
646 case PIPE_CAP_QUERY_TIME_ELAPSED
:
647 return rscreen
->info
.r600_clock_crystal_freq
!= 0;
648 case PIPE_CAP_QUERY_TIMESTAMP
:
649 return rscreen
->info
.drm_minor
>= 20 &&
650 rscreen
->info
.r600_clock_crystal_freq
!= 0;
652 case PIPE_CAP_MIN_TEXEL_OFFSET
:
655 case PIPE_CAP_MAX_TEXEL_OFFSET
:
661 static float r600_get_paramf(struct pipe_screen
* pscreen
,
662 enum pipe_capf param
)
664 struct r600_screen
*rscreen
= (struct r600_screen
*)pscreen
;
665 enum radeon_family family
= rscreen
->family
;
668 case PIPE_CAPF_MAX_LINE_WIDTH
:
669 case PIPE_CAPF_MAX_LINE_WIDTH_AA
:
670 case PIPE_CAPF_MAX_POINT_WIDTH
:
671 case PIPE_CAPF_MAX_POINT_WIDTH_AA
:
672 if (family
>= CHIP_CEDAR
)
676 case PIPE_CAPF_MAX_TEXTURE_ANISOTROPY
:
678 case PIPE_CAPF_MAX_TEXTURE_LOD_BIAS
:
680 case PIPE_CAPF_GUARD_BAND_LEFT
:
681 case PIPE_CAPF_GUARD_BAND_TOP
:
682 case PIPE_CAPF_GUARD_BAND_RIGHT
:
683 case PIPE_CAPF_GUARD_BAND_BOTTOM
:
689 static int r600_get_shader_param(struct pipe_screen
* pscreen
, unsigned shader
, enum pipe_shader_cap param
)
693 case PIPE_SHADER_FRAGMENT
:
694 case PIPE_SHADER_VERTEX
:
695 case PIPE_SHADER_COMPUTE
:
697 case PIPE_SHADER_GEOMETRY
:
698 /* XXX: support and enable geometry programs */
701 /* XXX: support tessellation on Evergreen */
706 case PIPE_SHADER_CAP_MAX_INSTRUCTIONS
:
707 case PIPE_SHADER_CAP_MAX_ALU_INSTRUCTIONS
:
708 case PIPE_SHADER_CAP_MAX_TEX_INSTRUCTIONS
:
709 case PIPE_SHADER_CAP_MAX_TEX_INDIRECTIONS
:
711 case PIPE_SHADER_CAP_MAX_CONTROL_FLOW_DEPTH
:
713 case PIPE_SHADER_CAP_MAX_INPUTS
:
715 case PIPE_SHADER_CAP_MAX_TEMPS
:
716 return 256; /* Max native temporaries. */
717 case PIPE_SHADER_CAP_MAX_ADDRS
:
718 /* XXX Isn't this equal to TEMPS? */
719 return 1; /* Max native address registers */
720 case PIPE_SHADER_CAP_MAX_CONSTS
:
721 return R600_MAX_CONST_BUFFER_SIZE
;
722 case PIPE_SHADER_CAP_MAX_CONST_BUFFERS
:
723 return R600_MAX_USER_CONST_BUFFERS
;
724 case PIPE_SHADER_CAP_MAX_PREDS
:
725 return 0; /* nothing uses this */
726 case PIPE_SHADER_CAP_TGSI_CONT_SUPPORTED
:
728 case PIPE_SHADER_CAP_TGSI_SQRT_SUPPORTED
:
730 case PIPE_SHADER_CAP_INDIRECT_INPUT_ADDR
:
731 case PIPE_SHADER_CAP_INDIRECT_OUTPUT_ADDR
:
732 case PIPE_SHADER_CAP_INDIRECT_TEMP_ADDR
:
733 case PIPE_SHADER_CAP_INDIRECT_CONST_ADDR
:
735 case PIPE_SHADER_CAP_SUBROUTINES
:
737 case PIPE_SHADER_CAP_INTEGERS
:
739 case PIPE_SHADER_CAP_MAX_TEXTURE_SAMPLERS
:
741 case PIPE_SHADER_CAP_PREFERRED_IR
:
742 if (shader
== PIPE_SHADER_COMPUTE
) {
743 return PIPE_SHADER_IR_LLVM
;
745 return PIPE_SHADER_IR_TGSI
;
751 static int r600_get_video_param(struct pipe_screen
*screen
,
752 enum pipe_video_profile profile
,
753 enum pipe_video_cap param
)
756 case PIPE_VIDEO_CAP_SUPPORTED
:
757 return vl_profile_supported(screen
, profile
);
758 case PIPE_VIDEO_CAP_NPOT_TEXTURES
:
760 case PIPE_VIDEO_CAP_MAX_WIDTH
:
761 case PIPE_VIDEO_CAP_MAX_HEIGHT
:
762 return vl_video_buffer_max_size(screen
);
763 case PIPE_VIDEO_CAP_PREFERED_FORMAT
:
764 return PIPE_FORMAT_NV12
;
765 case PIPE_VIDEO_CAP_PREFERS_INTERLACED
:
767 case PIPE_VIDEO_CAP_SUPPORTS_INTERLACED
:
769 case PIPE_VIDEO_CAP_SUPPORTS_PROGRESSIVE
:
776 const char * r600_llvm_gpu_string(enum radeon_family family
)
778 const char * gpu_family
;
792 gpu_family
= "rv710";
795 gpu_family
= "rv730";
799 gpu_family
= "rv770";
803 gpu_family
= "cedar";
808 gpu_family
= "redwood";
811 gpu_family
= "juniper";
815 gpu_family
= "cypress";
818 gpu_family
= "barts";
821 gpu_family
= "turks";
824 gpu_family
= "caicos";
828 gpu_family
= "cayman";
832 fprintf(stderr
, "Chip not supported by r600 llvm "
833 "backend, please file a bug at " PACKAGE_BUGREPORT
"\n");
840 static int r600_get_compute_param(struct pipe_screen
*screen
,
841 enum pipe_compute_cap param
,
844 struct r600_screen
*rscreen
= (struct r600_screen
*)screen
;
845 //TODO: select these params by asic
847 case PIPE_COMPUTE_CAP_IR_TARGET
: {
848 const char *gpu
= r600_llvm_gpu_string(rscreen
->family
);
850 sprintf(ret
, "%s-r600--", gpu
);
852 return (8 + strlen(gpu
)) * sizeof(char);
854 case PIPE_COMPUTE_CAP_GRID_DIMENSION
:
856 uint64_t * grid_dimension
= ret
;
857 grid_dimension
[0] = 3;
859 return 1 * sizeof(uint64_t);
861 case PIPE_COMPUTE_CAP_MAX_GRID_SIZE
:
863 uint64_t * grid_size
= ret
;
864 grid_size
[0] = 65535;
865 grid_size
[1] = 65535;
868 return 3 * sizeof(uint64_t) ;
870 case PIPE_COMPUTE_CAP_MAX_BLOCK_SIZE
:
872 uint64_t * block_size
= ret
;
877 return 3 * sizeof(uint64_t);
879 case PIPE_COMPUTE_CAP_MAX_THREADS_PER_BLOCK
:
881 uint64_t * max_threads_per_block
= ret
;
882 *max_threads_per_block
= 256;
884 return sizeof(uint64_t);
886 case PIPE_COMPUTE_CAP_MAX_GLOBAL_SIZE
:
888 uint64_t * max_global_size
= ret
;
889 /* XXX: This is what the proprietary driver reports, we
890 * may want to use a different value. */
891 *max_global_size
= 201326592;
893 return sizeof(uint64_t);
895 case PIPE_COMPUTE_CAP_MAX_INPUT_SIZE
:
897 uint64_t * max_input_size
= ret
;
898 *max_input_size
= 1024;
900 return sizeof(uint64_t);
902 case PIPE_COMPUTE_CAP_MAX_LOCAL_SIZE
:
904 uint64_t * max_local_size
= ret
;
905 /* XXX: This is what the proprietary driver reports, we
906 * may want to use a different value. */
907 *max_local_size
= 32768;
909 return sizeof(uint64_t);
911 case PIPE_COMPUTE_CAP_MAX_MEM_ALLOC_SIZE
:
913 uint64_t max_global_size
;
914 uint64_t * max_mem_alloc_size
= ret
;
915 r600_get_compute_param(screen
,
916 PIPE_COMPUTE_CAP_MAX_GLOBAL_SIZE
,
918 /* OpenCL requres this value be at least
919 * max(MAX_GLOBAL_SIZE / 4, 128 * 1024 *1024)
920 * I'm really not sure what value to report here, but
921 * MAX_GLOBAL_SIZE / 4 seems resonable.
923 *max_mem_alloc_size
= max_global_size
/ 4;
925 return sizeof(uint64_t);
928 fprintf(stderr
, "unknown PIPE_COMPUTE_CAP %d\n", param
);
933 static void r600_destroy_screen(struct pipe_screen
* pscreen
)
935 struct r600_screen
*rscreen
= (struct r600_screen
*)pscreen
;
940 if (rscreen
->global_pool
) {
941 compute_memory_pool_delete(rscreen
->global_pool
);
944 if (rscreen
->fences
.bo
) {
945 struct r600_fence_block
*entry
, *tmp
;
947 LIST_FOR_EACH_ENTRY_SAFE(entry
, tmp
, &rscreen
->fences
.blocks
, head
) {
948 LIST_DEL(&entry
->head
);
952 rscreen
->ws
->buffer_unmap(rscreen
->fences
.bo
->cs_buf
);
953 pipe_resource_reference((struct pipe_resource
**)&rscreen
->fences
.bo
, NULL
);
956 if (rscreen
->trace_bo
) {
957 rscreen
->ws
->buffer_unmap(rscreen
->trace_bo
->cs_buf
);
958 pipe_resource_reference((struct pipe_resource
**)&rscreen
->trace_bo
, NULL
);
961 pipe_mutex_destroy(rscreen
->fences
.mutex
);
963 rscreen
->ws
->destroy(rscreen
->ws
);
967 static void r600_fence_reference(struct pipe_screen
*pscreen
,
968 struct pipe_fence_handle
**ptr
,
969 struct pipe_fence_handle
*fence
)
971 struct r600_fence
**oldf
= (struct r600_fence
**)ptr
;
972 struct r600_fence
*newf
= (struct r600_fence
*)fence
;
974 if (pipe_reference(&(*oldf
)->reference
, &newf
->reference
)) {
975 struct r600_screen
*rscreen
= (struct r600_screen
*)pscreen
;
976 pipe_mutex_lock(rscreen
->fences
.mutex
);
977 pipe_resource_reference((struct pipe_resource
**)&(*oldf
)->sleep_bo
, NULL
);
978 LIST_ADDTAIL(&(*oldf
)->head
, &rscreen
->fences
.pool
);
979 pipe_mutex_unlock(rscreen
->fences
.mutex
);
985 static boolean
r600_fence_signalled(struct pipe_screen
*pscreen
,
986 struct pipe_fence_handle
*fence
)
988 struct r600_screen
*rscreen
= (struct r600_screen
*)pscreen
;
989 struct r600_fence
*rfence
= (struct r600_fence
*)fence
;
991 return rscreen
->fences
.data
[rfence
->index
] != 0;
994 static boolean
r600_fence_finish(struct pipe_screen
*pscreen
,
995 struct pipe_fence_handle
*fence
,
998 struct r600_screen
*rscreen
= (struct r600_screen
*)pscreen
;
999 struct r600_fence
*rfence
= (struct r600_fence
*)fence
;
1000 int64_t start_time
= 0;
1003 if (timeout
!= PIPE_TIMEOUT_INFINITE
) {
1004 start_time
= os_time_get();
1006 /* Convert to microseconds. */
1010 while (rscreen
->fences
.data
[rfence
->index
] == 0) {
1011 /* Special-case infinite timeout - wait for the dummy BO to become idle */
1012 if (timeout
== PIPE_TIMEOUT_INFINITE
) {
1013 rscreen
->ws
->buffer_wait(rfence
->sleep_bo
->buf
, RADEON_USAGE_READWRITE
);
1017 /* The dummy BO will be busy until the CS including the fence has completed, or
1018 * the GPU is reset. Don't bother continuing to spin when the BO is idle. */
1019 if (!rscreen
->ws
->buffer_is_busy(rfence
->sleep_bo
->buf
, RADEON_USAGE_READWRITE
))
1029 if (timeout
!= PIPE_TIMEOUT_INFINITE
&&
1030 os_time_get() - start_time
>= timeout
) {
1035 return rscreen
->fences
.data
[rfence
->index
] != 0;
1038 static int r600_interpret_tiling(struct r600_screen
*rscreen
, uint32_t tiling_config
)
1040 switch ((tiling_config
& 0xe) >> 1) {
1042 rscreen
->tiling_info
.num_channels
= 1;
1045 rscreen
->tiling_info
.num_channels
= 2;
1048 rscreen
->tiling_info
.num_channels
= 4;
1051 rscreen
->tiling_info
.num_channels
= 8;
1057 switch ((tiling_config
& 0x30) >> 4) {
1059 rscreen
->tiling_info
.num_banks
= 4;
1062 rscreen
->tiling_info
.num_banks
= 8;
1068 switch ((tiling_config
& 0xc0) >> 6) {
1070 rscreen
->tiling_info
.group_bytes
= 256;
1073 rscreen
->tiling_info
.group_bytes
= 512;
1081 static int evergreen_interpret_tiling(struct r600_screen
*rscreen
, uint32_t tiling_config
)
1083 switch (tiling_config
& 0xf) {
1085 rscreen
->tiling_info
.num_channels
= 1;
1088 rscreen
->tiling_info
.num_channels
= 2;
1091 rscreen
->tiling_info
.num_channels
= 4;
1094 rscreen
->tiling_info
.num_channels
= 8;
1100 switch ((tiling_config
& 0xf0) >> 4) {
1102 rscreen
->tiling_info
.num_banks
= 4;
1105 rscreen
->tiling_info
.num_banks
= 8;
1108 rscreen
->tiling_info
.num_banks
= 16;
1114 switch ((tiling_config
& 0xf00) >> 8) {
1116 rscreen
->tiling_info
.group_bytes
= 256;
1119 rscreen
->tiling_info
.group_bytes
= 512;
1127 static int r600_init_tiling(struct r600_screen
*rscreen
)
1129 uint32_t tiling_config
= rscreen
->info
.r600_tiling_config
;
1131 /* set default group bytes, overridden by tiling info ioctl */
1132 if (rscreen
->chip_class
<= R700
) {
1133 rscreen
->tiling_info
.group_bytes
= 256;
1135 rscreen
->tiling_info
.group_bytes
= 512;
1141 if (rscreen
->chip_class
<= R700
) {
1142 return r600_interpret_tiling(rscreen
, tiling_config
);
1144 return evergreen_interpret_tiling(rscreen
, tiling_config
);
1148 static uint64_t r600_get_timestamp(struct pipe_screen
*screen
)
1150 struct r600_screen
*rscreen
= (struct r600_screen
*)screen
;
1152 return 1000000 * rscreen
->ws
->query_value(rscreen
->ws
, RADEON_TIMESTAMP
) /
1153 rscreen
->info
.r600_clock_crystal_freq
;
1156 static int r600_get_driver_query_info(struct pipe_screen
*screen
,
1158 struct pipe_driver_query_info
*info
)
1160 struct r600_screen
*rscreen
= (struct r600_screen
*)screen
;
1161 struct pipe_driver_query_info list
[] = {
1162 {"draw-calls", R600_QUERY_DRAW_CALLS
, 0},
1163 {"requested-VRAM", R600_QUERY_REQUESTED_VRAM
, rscreen
->info
.vram_size
, TRUE
},
1164 {"requested-GTT", R600_QUERY_REQUESTED_GTT
, rscreen
->info
.gart_size
, TRUE
},
1165 {"buffer-wait-time", R600_QUERY_BUFFER_WAIT_TIME
, 0, FALSE
}
1169 return Elements(list
);
1171 if (index
>= Elements(list
))
1174 *info
= list
[index
];
1178 struct pipe_screen
*r600_screen_create(struct radeon_winsys
*ws
)
1180 struct r600_screen
*rscreen
= CALLOC_STRUCT(r600_screen
);
1182 if (rscreen
== NULL
) {
1187 ws
->query_info(ws
, &rscreen
->info
);
1189 rscreen
->debug_flags
= debug_get_flags_option("R600_DEBUG", debug_options
, 0);
1190 if (debug_get_bool_option("R600_DEBUG_COMPUTE", FALSE
))
1191 rscreen
->debug_flags
|= DBG_COMPUTE
;
1192 if (debug_get_bool_option("R600_DUMP_SHADERS", FALSE
))
1193 rscreen
->debug_flags
|= DBG_FS
| DBG_VS
| DBG_GS
| DBG_PS
| DBG_CS
;
1194 if (!debug_get_bool_option("R600_HYPERZ", TRUE
))
1195 rscreen
->debug_flags
|= DBG_NO_HYPERZ
;
1196 if (!debug_get_bool_option("R600_LLVM", TRUE
))
1197 rscreen
->debug_flags
|= DBG_NO_LLVM
;
1198 if (debug_get_bool_option("R600_PRINT_TEXDEPTH", FALSE
))
1199 rscreen
->debug_flags
|= DBG_TEX_DEPTH
;
1200 rscreen
->family
= rscreen
->info
.family
;
1201 rscreen
->chip_class
= rscreen
->info
.chip_class
;
1203 if (rscreen
->family
== CHIP_UNKNOWN
) {
1204 fprintf(stderr
, "r600: Unknown chipset 0x%04X\n", rscreen
->info
.pci_id
);
1209 /* Figure out streamout kernel support. */
1210 switch (rscreen
->chip_class
) {
1212 if (rscreen
->family
< CHIP_RS780
) {
1213 rscreen
->has_streamout
= rscreen
->info
.drm_minor
>= 14;
1215 rscreen
->has_streamout
= rscreen
->info
.drm_minor
>= 23;
1219 rscreen
->has_streamout
= rscreen
->info
.drm_minor
>= 17;
1223 rscreen
->has_streamout
= rscreen
->info
.drm_minor
>= 14;
1226 rscreen
->has_streamout
= FALSE
;
1231 switch (rscreen
->chip_class
) {
1234 rscreen
->has_msaa
= rscreen
->info
.drm_minor
>= 22;
1235 rscreen
->msaa_texture_support
= MSAA_TEXTURE_DECOMPRESSED
;
1238 rscreen
->has_msaa
= rscreen
->info
.drm_minor
>= 19;
1239 rscreen
->msaa_texture_support
=
1240 rscreen
->info
.drm_minor
>= 24 ? MSAA_TEXTURE_COMPRESSED
:
1241 MSAA_TEXTURE_DECOMPRESSED
;
1244 rscreen
->has_msaa
= rscreen
->info
.drm_minor
>= 19;
1245 /* We should be able to read compressed MSAA textures, but it doesn't work. */
1246 rscreen
->msaa_texture_support
= MSAA_TEXTURE_SAMPLE_ZERO
;
1249 rscreen
->has_msaa
= FALSE
;
1250 rscreen
->msaa_texture_support
= 0;
1254 rscreen
->has_cp_dma
= rscreen
->info
.drm_minor
>= 27 &&
1255 !(rscreen
->debug_flags
& DBG_NO_CP_DMA
);
1257 if (r600_init_tiling(rscreen
)) {
1262 rscreen
->screen
.destroy
= r600_destroy_screen
;
1263 rscreen
->screen
.get_name
= r600_get_name
;
1264 rscreen
->screen
.get_vendor
= r600_get_vendor
;
1265 rscreen
->screen
.get_param
= r600_get_param
;
1266 rscreen
->screen
.get_shader_param
= r600_get_shader_param
;
1267 rscreen
->screen
.get_paramf
= r600_get_paramf
;
1268 rscreen
->screen
.get_compute_param
= r600_get_compute_param
;
1269 rscreen
->screen
.get_timestamp
= r600_get_timestamp
;
1271 if (rscreen
->chip_class
>= EVERGREEN
) {
1272 rscreen
->screen
.is_format_supported
= evergreen_is_format_supported
;
1273 rscreen
->dma_blit
= &evergreen_dma_blit
;
1275 rscreen
->screen
.is_format_supported
= r600_is_format_supported
;
1276 rscreen
->dma_blit
= &r600_dma_blit
;
1278 rscreen
->screen
.context_create
= r600_create_context
;
1279 rscreen
->screen
.fence_reference
= r600_fence_reference
;
1280 rscreen
->screen
.fence_signalled
= r600_fence_signalled
;
1281 rscreen
->screen
.fence_finish
= r600_fence_finish
;
1282 rscreen
->screen
.get_driver_query_info
= r600_get_driver_query_info
;
1284 if (rscreen
->info
.has_uvd
) {
1285 rscreen
->screen
.get_video_param
= ruvd_get_video_param
;
1286 rscreen
->screen
.is_video_format_supported
= ruvd_is_format_supported
;
1288 rscreen
->screen
.get_video_param
= r600_get_video_param
;
1289 rscreen
->screen
.is_video_format_supported
= vl_video_buffer_is_format_supported
;
1292 r600_init_screen_resource_functions(&rscreen
->screen
);
1294 util_format_s3tc_init();
1296 rscreen
->fences
.bo
= NULL
;
1297 rscreen
->fences
.data
= NULL
;
1298 rscreen
->fences
.next_index
= 0;
1299 LIST_INITHEAD(&rscreen
->fences
.pool
);
1300 LIST_INITHEAD(&rscreen
->fences
.blocks
);
1301 pipe_mutex_init(rscreen
->fences
.mutex
);
1303 rscreen
->global_pool
= compute_memory_pool_new(rscreen
);
1306 rscreen
->cs_count
= 0;
1307 if (rscreen
->info
.drm_minor
>= 28) {
1308 rscreen
->trace_bo
= (struct r600_resource
*)pipe_buffer_create(&rscreen
->screen
,
1312 if (rscreen
->trace_bo
) {
1313 rscreen
->trace_ptr
= rscreen
->ws
->buffer_map(rscreen
->trace_bo
->cs_buf
, NULL
,
1314 PIPE_TRANSFER_UNSYNCHRONIZED
);
1319 return &rscreen
->screen
;