2 * Copyright 2010 Christoph Bumiller
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
11 * The above copyright notice and this permission notice shall be included in
12 * all copies or substantial portions of the Software.
14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
17 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR
18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20 * OTHER DEALINGS IN THE SOFTWARE.
25 #include <nouveau_drm.h>
26 #include "util/u_format.h"
27 #include "util/u_format_s3tc.h"
28 #include "pipe/p_screen.h"
30 #include "nv50/nv50_context.h"
31 #include "nv50/nv50_screen.h"
33 #include "nouveau_vp3_video.h"
35 #include "nv_object.xml.h"
37 /* affected by LOCAL_WARPS_LOG_ALLOC / LOCAL_WARPS_NO_CLAMP */
38 #define LOCAL_WARPS_ALLOC 32
39 /* affected by STACK_WARPS_LOG_ALLOC / STACK_WARPS_NO_CLAMP */
40 #define STACK_WARPS_ALLOC 32
42 #define THREADS_IN_WARP 32
45 nv50_screen_is_format_supported(struct pipe_screen
*pscreen
,
46 enum pipe_format format
,
47 enum pipe_texture_target target
,
48 unsigned sample_count
,
53 if (!(0x117 & (1 << sample_count
))) /* 0, 1, 2, 4 or 8 */
55 if (sample_count
== 8 && util_format_get_blocksizebits(format
) >= 128)
58 if (!util_format_is_supported(format
, bindings
))
62 case PIPE_FORMAT_Z16_UNORM
:
63 if (nv50_screen(pscreen
)->tesla
->oclass
< NVA0_3D_CLASS
)
70 if (bindings
& PIPE_BIND_LINEAR
)
71 if (util_format_is_depth_or_stencil(format
) ||
72 (target
!= PIPE_TEXTURE_1D
&&
73 target
!= PIPE_TEXTURE_2D
&&
74 target
!= PIPE_TEXTURE_RECT
) ||
78 /* shared is always supported */
79 bindings
&= ~(PIPE_BIND_LINEAR
|
82 return (( nv50_format_table
[format
].usage
|
83 nv50_vertex_format
[format
].usage
) & bindings
) == bindings
;
87 nv50_screen_get_param(struct pipe_screen
*pscreen
, enum pipe_cap param
)
89 const uint16_t class_3d
= nouveau_screen(pscreen
)->class_3d
;
90 struct nouveau_device
*dev
= nouveau_screen(pscreen
)->device
;
93 /* non-boolean caps */
94 case PIPE_CAP_MAX_TEXTURE_2D_LEVELS
:
96 case PIPE_CAP_MAX_TEXTURE_3D_LEVELS
:
98 case PIPE_CAP_MAX_TEXTURE_CUBE_LEVELS
:
100 case PIPE_CAP_MAX_TEXTURE_ARRAY_LAYERS
:
102 case PIPE_CAP_MIN_TEXTURE_GATHER_OFFSET
:
103 case PIPE_CAP_MIN_TEXEL_OFFSET
:
105 case PIPE_CAP_MAX_TEXTURE_GATHER_OFFSET
:
106 case PIPE_CAP_MAX_TEXEL_OFFSET
:
108 case PIPE_CAP_MAX_TEXTURE_BUFFER_SIZE
:
109 return 128 * 1024 * 1024;
110 case PIPE_CAP_GLSL_FEATURE_LEVEL
:
112 case PIPE_CAP_MAX_RENDER_TARGETS
:
114 case PIPE_CAP_MAX_DUAL_SOURCE_RENDER_TARGETS
:
116 case PIPE_CAP_MAX_STREAM_OUTPUT_BUFFERS
:
118 case PIPE_CAP_MAX_STREAM_OUTPUT_INTERLEAVED_COMPONENTS
:
119 case PIPE_CAP_MAX_STREAM_OUTPUT_SEPARATE_COMPONENTS
:
121 case PIPE_CAP_MAX_GEOMETRY_OUTPUT_VERTICES
:
122 case PIPE_CAP_MAX_GEOMETRY_TOTAL_OUTPUT_COMPONENTS
:
124 case PIPE_CAP_MAX_VERTEX_STREAMS
:
126 case PIPE_CAP_MAX_VERTEX_ATTRIB_STRIDE
:
128 case PIPE_CAP_CONSTANT_BUFFER_OFFSET_ALIGNMENT
:
130 case PIPE_CAP_TEXTURE_BUFFER_OFFSET_ALIGNMENT
:
131 return 16; /* 256 for binding as RT, but that's not possible in GL */
132 case PIPE_CAP_MIN_MAP_BUFFER_ALIGNMENT
:
133 return NOUVEAU_MIN_BUFFER_MAP_ALIGN
;
134 case PIPE_CAP_MAX_VIEWPORTS
:
135 return NV50_MAX_VIEWPORTS
;
136 case PIPE_CAP_TEXTURE_BORDER_COLOR_QUIRK
:
137 return PIPE_QUIRK_TEXTURE_BORDER_COLOR_SWIZZLE_NV50
;
138 case PIPE_CAP_ENDIANNESS
:
139 return PIPE_ENDIAN_LITTLE
;
140 case PIPE_CAP_MAX_TEXTURE_GATHER_COMPONENTS
:
141 return (class_3d
>= NVA3_3D_CLASS
) ? 4 : 0;
142 case PIPE_CAP_MAX_WINDOW_RECTANGLES
:
143 return NV50_MAX_WINDOW_RECTANGLES
;
146 case PIPE_CAP_TEXTURE_MIRROR_CLAMP
:
147 case PIPE_CAP_TEXTURE_SWIZZLE
:
148 case PIPE_CAP_TEXTURE_SHADOW_MAP
:
149 case PIPE_CAP_NPOT_TEXTURES
:
150 case PIPE_CAP_MIXED_FRAMEBUFFER_SIZES
:
151 case PIPE_CAP_MIXED_COLOR_DEPTH_BITS
:
152 case PIPE_CAP_ANISOTROPIC_FILTER
:
153 case PIPE_CAP_TEXTURE_BUFFER_OBJECTS
:
154 case PIPE_CAP_BUFFER_MAP_PERSISTENT_COHERENT
:
155 case PIPE_CAP_TWO_SIDED_STENCIL
:
156 case PIPE_CAP_DEPTH_CLIP_DISABLE
:
157 case PIPE_CAP_POINT_SPRITE
:
159 case PIPE_CAP_FRAGMENT_COLOR_CLAMPED
:
160 case PIPE_CAP_VERTEX_COLOR_UNCLAMPED
:
161 case PIPE_CAP_VERTEX_COLOR_CLAMPED
:
162 case PIPE_CAP_QUERY_TIMESTAMP
:
163 case PIPE_CAP_QUERY_TIME_ELAPSED
:
164 case PIPE_CAP_OCCLUSION_QUERY
:
165 case PIPE_CAP_BLEND_EQUATION_SEPARATE
:
166 case PIPE_CAP_INDEP_BLEND_ENABLE
:
167 case PIPE_CAP_TGSI_FS_COORD_ORIGIN_UPPER_LEFT
:
168 case PIPE_CAP_TGSI_FS_COORD_PIXEL_CENTER_HALF_INTEGER
:
169 case PIPE_CAP_PRIMITIVE_RESTART
:
170 case PIPE_CAP_TGSI_INSTANCEID
:
171 case PIPE_CAP_VERTEX_ELEMENT_INSTANCE_DIVISOR
:
172 case PIPE_CAP_MIXED_COLORBUFFER_FORMATS
:
173 case PIPE_CAP_CONDITIONAL_RENDER
:
174 case PIPE_CAP_TEXTURE_BARRIER
:
175 case PIPE_CAP_QUADS_FOLLOW_PROVOKING_VERTEX_CONVENTION
:
176 case PIPE_CAP_START_INSTANCE
:
177 case PIPE_CAP_USER_CONSTANT_BUFFERS
:
178 case PIPE_CAP_USER_VERTEX_BUFFERS
:
179 case PIPE_CAP_TEXTURE_MULTISAMPLE
:
180 case PIPE_CAP_PREFER_BLIT_BASED_TEXTURE_TRANSFER
:
181 case PIPE_CAP_TGSI_FS_FINE_DERIVATIVE
:
182 case PIPE_CAP_SAMPLER_VIEW_TARGET
:
183 case PIPE_CAP_CONDITIONAL_RENDER_INVERTED
:
184 case PIPE_CAP_CLIP_HALFZ
:
185 case PIPE_CAP_POLYGON_OFFSET_CLAMP
:
186 case PIPE_CAP_QUERY_PIPELINE_STATISTICS
:
187 case PIPE_CAP_TEXTURE_FLOAT_LINEAR
:
188 case PIPE_CAP_TEXTURE_HALF_FLOAT_LINEAR
:
189 case PIPE_CAP_DEPTH_BOUNDS_TEST
:
190 case PIPE_CAP_TGSI_TXQS
:
191 case PIPE_CAP_COPY_BETWEEN_COMPRESSED_AND_PLAIN_FORMATS
:
192 case PIPE_CAP_SHAREABLE_SHADERS
:
193 case PIPE_CAP_CLEAR_TEXTURE
:
194 case PIPE_CAP_COMPUTE
:
195 case PIPE_CAP_TGSI_FS_FACE_IS_INTEGER_SYSVAL
:
196 case PIPE_CAP_INVALIDATE_BUFFER
:
197 case PIPE_CAP_STRING_MARKER
:
198 case PIPE_CAP_CULL_DISTANCE
:
199 case PIPE_CAP_TGSI_ARRAY_COMPONENTS
:
200 case PIPE_CAP_TGSI_MUL_ZERO_WINS
:
201 case PIPE_CAP_TGSI_TEX_TXF_LZ
:
202 case PIPE_CAP_TGSI_CLOCK
:
203 case PIPE_CAP_CAN_BIND_CONST_BUFFER_AS_VERTEX
:
204 case PIPE_CAP_ALLOW_MAPPED_BUFFERS_DURING_EXECUTION
:
206 case PIPE_CAP_SEAMLESS_CUBE_MAP
:
207 return 1; /* class_3d >= NVA0_3D_CLASS; */
208 /* supported on nva0+ */
209 case PIPE_CAP_STREAM_OUTPUT_PAUSE_RESUME
:
210 return class_3d
>= NVA0_3D_CLASS
;
211 /* supported on nva3+ */
212 case PIPE_CAP_CUBE_MAP_ARRAY
:
213 case PIPE_CAP_INDEP_BLEND_FUNC
:
214 case PIPE_CAP_TEXTURE_QUERY_LOD
:
215 case PIPE_CAP_SAMPLE_SHADING
:
216 case PIPE_CAP_FORCE_PERSAMPLE_INTERP
:
217 return class_3d
>= NVA3_3D_CLASS
;
219 /* unsupported caps */
220 case PIPE_CAP_SEAMLESS_CUBE_MAP_PER_TEXTURE
:
221 case PIPE_CAP_TGSI_FS_COORD_ORIGIN_LOWER_LEFT
:
222 case PIPE_CAP_TGSI_FS_COORD_PIXEL_CENTER_INTEGER
:
223 case PIPE_CAP_SHADER_STENCIL_EXPORT
:
224 case PIPE_CAP_TGSI_CAN_COMPACT_CONSTANTS
:
225 case PIPE_CAP_VERTEX_BUFFER_OFFSET_4BYTE_ALIGNED_ONLY
:
226 case PIPE_CAP_VERTEX_BUFFER_STRIDE_4BYTE_ALIGNED_ONLY
:
227 case PIPE_CAP_VERTEX_ELEMENT_SRC_OFFSET_4BYTE_ALIGNED_ONLY
:
228 case PIPE_CAP_TGSI_TEXCOORD
:
229 case PIPE_CAP_TGSI_VS_LAYER_VIEWPORT
:
230 case PIPE_CAP_TEXTURE_GATHER_SM5
:
231 case PIPE_CAP_FAKE_SW_MSAA
:
232 case PIPE_CAP_TEXTURE_GATHER_OFFSETS
:
233 case PIPE_CAP_TGSI_VS_WINDOW_SPACE_POSITION
:
234 case PIPE_CAP_DRAW_INDIRECT
:
235 case PIPE_CAP_MULTI_DRAW_INDIRECT
:
236 case PIPE_CAP_MULTI_DRAW_INDIRECT_PARAMS
:
237 case PIPE_CAP_VERTEXID_NOBASE
:
238 case PIPE_CAP_MULTISAMPLE_Z_RESOLVE
: /* potentially supported on some hw */
239 case PIPE_CAP_RESOURCE_FROM_USER_MEMORY
:
240 case PIPE_CAP_DEVICE_RESET_STATUS_QUERY
:
241 case PIPE_CAP_MAX_SHADER_PATCH_VARYINGS
:
242 case PIPE_CAP_DRAW_PARAMETERS
:
243 case PIPE_CAP_TGSI_PACK_HALF_FLOAT
:
244 case PIPE_CAP_TGSI_FS_POSITION_IS_SYSVAL
:
245 case PIPE_CAP_SHADER_BUFFER_OFFSET_ALIGNMENT
:
246 case PIPE_CAP_GENERATE_MIPMAP
:
247 case PIPE_CAP_BUFFER_SAMPLER_VIEW_RGBA_ONLY
:
248 case PIPE_CAP_SURFACE_REINTERPRET_BLOCKS
:
249 case PIPE_CAP_QUERY_BUFFER_OBJECT
:
250 case PIPE_CAP_QUERY_MEMORY_INFO
:
251 case PIPE_CAP_PCI_GROUP
:
252 case PIPE_CAP_PCI_BUS
:
253 case PIPE_CAP_PCI_DEVICE
:
254 case PIPE_CAP_PCI_FUNCTION
:
255 case PIPE_CAP_FRAMEBUFFER_NO_ATTACHMENT
:
256 case PIPE_CAP_ROBUST_BUFFER_ACCESS_BEHAVIOR
:
257 case PIPE_CAP_PRIMITIVE_RESTART_FOR_PATCHES
:
258 case PIPE_CAP_TGSI_VOTE
:
259 case PIPE_CAP_POLYGON_OFFSET_UNITS_UNSCALED
:
260 case PIPE_CAP_VIEWPORT_SUBPIXEL_BITS
:
261 case PIPE_CAP_STREAM_OUTPUT_INTERLEAVE_BUFFERS
:
262 case PIPE_CAP_TGSI_CAN_READ_OUTPUTS
:
263 case PIPE_CAP_NATIVE_FENCE_FD
:
264 case PIPE_CAP_GLSL_OPTIMIZE_CONSERVATIVELY
:
265 case PIPE_CAP_TGSI_FS_FBFETCH
:
266 case PIPE_CAP_DOUBLES
:
268 case PIPE_CAP_INT64_DIVMOD
:
269 case PIPE_CAP_POLYGON_MODE_FILL_RECTANGLE
:
270 case PIPE_CAP_SPARSE_BUFFER_PAGE_SIZE
:
271 case PIPE_CAP_TGSI_BALLOT
:
272 case PIPE_CAP_TGSI_TES_LAYER_VIEWPORT
:
273 case PIPE_CAP_POST_DEPTH_COVERAGE
:
274 case PIPE_CAP_BINDLESS_TEXTURE
:
275 case PIPE_CAP_NIR_SAMPLERS_AS_DEREF
:
276 case PIPE_CAP_QUERY_SO_OVERFLOW
:
277 case PIPE_CAP_MEMOBJ
:
278 case PIPE_CAP_LOAD_CONSTBUF
:
279 case PIPE_CAP_TGSI_ANY_REG_AS_ADDRESS
:
280 case PIPE_CAP_TILE_RASTER_ORDER
:
281 case PIPE_CAP_MAX_COMBINED_SHADER_OUTPUT_RESOURCES
:
282 case PIPE_CAP_SIGNED_VERTEX_BUFFER_OFFSET
:
283 case PIPE_CAP_CONTEXT_PRIORITY_MASK
:
286 case PIPE_CAP_VENDOR_ID
:
288 case PIPE_CAP_DEVICE_ID
: {
290 if (nouveau_getparam(dev
, NOUVEAU_GETPARAM_PCI_DEVICE
, &device_id
)) {
291 NOUVEAU_ERR("NOUVEAU_GETPARAM_PCI_DEVICE failed.\n");
296 case PIPE_CAP_ACCELERATED
:
298 case PIPE_CAP_VIDEO_MEMORY
:
299 return dev
->vram_size
>> 20;
304 NOUVEAU_ERR("unknown PIPE_CAP %d\n", param
);
309 nv50_screen_get_shader_param(struct pipe_screen
*pscreen
,
310 enum pipe_shader_type shader
,
311 enum pipe_shader_cap param
)
314 case PIPE_SHADER_VERTEX
:
315 case PIPE_SHADER_GEOMETRY
:
316 case PIPE_SHADER_FRAGMENT
:
318 case PIPE_SHADER_COMPUTE
:
324 case PIPE_SHADER_CAP_MAX_INSTRUCTIONS
:
325 case PIPE_SHADER_CAP_MAX_ALU_INSTRUCTIONS
:
326 case PIPE_SHADER_CAP_MAX_TEX_INSTRUCTIONS
:
327 case PIPE_SHADER_CAP_MAX_TEX_INDIRECTIONS
:
329 case PIPE_SHADER_CAP_MAX_CONTROL_FLOW_DEPTH
:
331 case PIPE_SHADER_CAP_MAX_INPUTS
:
332 if (shader
== PIPE_SHADER_VERTEX
)
335 case PIPE_SHADER_CAP_MAX_OUTPUTS
:
337 case PIPE_SHADER_CAP_MAX_CONST_BUFFER_SIZE
:
339 case PIPE_SHADER_CAP_MAX_CONST_BUFFERS
:
340 return NV50_MAX_PIPE_CONSTBUFS
;
341 case PIPE_SHADER_CAP_INDIRECT_OUTPUT_ADDR
:
342 return shader
!= PIPE_SHADER_FRAGMENT
;
343 case PIPE_SHADER_CAP_INDIRECT_INPUT_ADDR
:
344 case PIPE_SHADER_CAP_INDIRECT_TEMP_ADDR
:
345 case PIPE_SHADER_CAP_INDIRECT_CONST_ADDR
:
347 case PIPE_SHADER_CAP_MAX_TEMPS
:
348 return nv50_screen(pscreen
)->max_tls_space
/ ONE_TEMP_SIZE
;
349 case PIPE_SHADER_CAP_TGSI_CONT_SUPPORTED
:
351 case PIPE_SHADER_CAP_TGSI_SQRT_SUPPORTED
:
353 case PIPE_SHADER_CAP_INT64_ATOMICS
:
354 case PIPE_SHADER_CAP_FP16
:
355 case PIPE_SHADER_CAP_SUBROUTINES
:
356 return 0; /* please inline, or provide function declarations */
357 case PIPE_SHADER_CAP_INTEGERS
:
359 case PIPE_SHADER_CAP_TGSI_SKIP_MERGE_REGISTERS
:
361 case PIPE_SHADER_CAP_MAX_TEXTURE_SAMPLERS
:
362 /* The chip could handle more sampler views than samplers */
363 case PIPE_SHADER_CAP_MAX_SAMPLER_VIEWS
:
364 return MIN2(16, PIPE_MAX_SAMPLERS
);
365 case PIPE_SHADER_CAP_PREFERRED_IR
:
366 return PIPE_SHADER_IR_TGSI
;
367 case PIPE_SHADER_CAP_MAX_UNROLL_ITERATIONS_HINT
:
369 case PIPE_SHADER_CAP_TGSI_DROUND_SUPPORTED
:
370 case PIPE_SHADER_CAP_TGSI_DFRACEXP_DLDEXP_SUPPORTED
:
371 case PIPE_SHADER_CAP_TGSI_LDEXP_SUPPORTED
:
372 case PIPE_SHADER_CAP_TGSI_FMA_SUPPORTED
:
373 case PIPE_SHADER_CAP_TGSI_ANY_INOUT_DECL_RANGE
:
374 case PIPE_SHADER_CAP_MAX_SHADER_BUFFERS
:
375 case PIPE_SHADER_CAP_SUPPORTED_IRS
:
376 case PIPE_SHADER_CAP_MAX_SHADER_IMAGES
:
377 case PIPE_SHADER_CAP_LOWER_IF_THRESHOLD
:
378 case PIPE_SHADER_CAP_MAX_HW_ATOMIC_COUNTERS
:
379 case PIPE_SHADER_CAP_MAX_HW_ATOMIC_COUNTER_BUFFERS
:
382 NOUVEAU_ERR("unknown PIPE_SHADER_CAP %d\n", param
);
388 nv50_screen_get_paramf(struct pipe_screen
*pscreen
, enum pipe_capf param
)
391 case PIPE_CAPF_MAX_LINE_WIDTH
:
392 case PIPE_CAPF_MAX_LINE_WIDTH_AA
:
394 case PIPE_CAPF_MAX_POINT_WIDTH
:
395 case PIPE_CAPF_MAX_POINT_WIDTH_AA
:
397 case PIPE_CAPF_MAX_TEXTURE_ANISOTROPY
:
399 case PIPE_CAPF_MAX_TEXTURE_LOD_BIAS
:
401 case PIPE_CAPF_GUARD_BAND_LEFT
:
402 case PIPE_CAPF_GUARD_BAND_TOP
:
404 case PIPE_CAPF_GUARD_BAND_RIGHT
:
405 case PIPE_CAPF_GUARD_BAND_BOTTOM
:
406 return 0.0f
; /* that or infinity */
409 NOUVEAU_ERR("unknown PIPE_CAPF %d\n", param
);
414 nv50_screen_get_compute_param(struct pipe_screen
*pscreen
,
415 enum pipe_shader_ir ir_type
,
416 enum pipe_compute_cap param
, void *data
)
418 struct nv50_screen
*screen
= nv50_screen(pscreen
);
420 #define RET(x) do { \
422 memcpy(data, x, sizeof(x)); \
427 case PIPE_COMPUTE_CAP_GRID_DIMENSION
:
428 RET((uint64_t []) { 2 });
429 case PIPE_COMPUTE_CAP_MAX_GRID_SIZE
:
430 RET(((uint64_t []) { 65535, 65535 }));
431 case PIPE_COMPUTE_CAP_MAX_BLOCK_SIZE
:
432 RET(((uint64_t []) { 512, 512, 64 }));
433 case PIPE_COMPUTE_CAP_MAX_THREADS_PER_BLOCK
:
434 RET((uint64_t []) { 512 });
435 case PIPE_COMPUTE_CAP_MAX_GLOBAL_SIZE
: /* g0-15[] */
436 RET((uint64_t []) { 1ULL << 32 });
437 case PIPE_COMPUTE_CAP_MAX_LOCAL_SIZE
: /* s[] */
438 RET((uint64_t []) { 16 << 10 });
439 case PIPE_COMPUTE_CAP_MAX_PRIVATE_SIZE
: /* l[] */
440 RET((uint64_t []) { 16 << 10 });
441 case PIPE_COMPUTE_CAP_MAX_INPUT_SIZE
: /* c[], arbitrary limit */
442 RET((uint64_t []) { 4096 });
443 case PIPE_COMPUTE_CAP_SUBGROUP_SIZE
:
444 RET((uint32_t []) { 32 });
445 case PIPE_COMPUTE_CAP_MAX_MEM_ALLOC_SIZE
:
446 RET((uint64_t []) { 1ULL << 40 });
447 case PIPE_COMPUTE_CAP_IMAGES_SUPPORTED
:
448 RET((uint32_t []) { 0 });
449 case PIPE_COMPUTE_CAP_MAX_COMPUTE_UNITS
:
450 RET((uint32_t []) { screen
->mp_count
});
451 case PIPE_COMPUTE_CAP_MAX_CLOCK_FREQUENCY
:
452 RET((uint32_t []) { 512 }); /* FIXME: arbitrary limit */
453 case PIPE_COMPUTE_CAP_ADDRESS_BITS
:
454 RET((uint32_t []) { 32 });
455 case PIPE_COMPUTE_CAP_MAX_VARIABLE_THREADS_PER_BLOCK
:
456 RET((uint64_t []) { 0 });
465 nv50_screen_destroy(struct pipe_screen
*pscreen
)
467 struct nv50_screen
*screen
= nv50_screen(pscreen
);
469 if (!nouveau_drm_screen_unref(&screen
->base
))
472 if (screen
->base
.fence
.current
) {
473 struct nouveau_fence
*current
= NULL
;
475 /* nouveau_fence_wait will create a new current fence, so wait on the
476 * _current_ one, and remove both.
478 nouveau_fence_ref(screen
->base
.fence
.current
, ¤t
);
479 nouveau_fence_wait(current
, NULL
);
480 nouveau_fence_ref(NULL
, ¤t
);
481 nouveau_fence_ref(NULL
, &screen
->base
.fence
.current
);
483 if (screen
->base
.pushbuf
)
484 screen
->base
.pushbuf
->user_priv
= NULL
;
487 nv50_blitter_destroy(screen
);
488 if (screen
->pm
.prog
) {
489 screen
->pm
.prog
->code
= NULL
; /* hardcoded, don't FREE */
490 nv50_program_destroy(NULL
, screen
->pm
.prog
);
491 FREE(screen
->pm
.prog
);
494 nouveau_bo_ref(NULL
, &screen
->code
);
495 nouveau_bo_ref(NULL
, &screen
->tls_bo
);
496 nouveau_bo_ref(NULL
, &screen
->stack_bo
);
497 nouveau_bo_ref(NULL
, &screen
->txc
);
498 nouveau_bo_ref(NULL
, &screen
->uniforms
);
499 nouveau_bo_ref(NULL
, &screen
->fence
.bo
);
501 nouveau_heap_destroy(&screen
->vp_code_heap
);
502 nouveau_heap_destroy(&screen
->gp_code_heap
);
503 nouveau_heap_destroy(&screen
->fp_code_heap
);
505 FREE(screen
->tic
.entries
);
507 nouveau_object_del(&screen
->tesla
);
508 nouveau_object_del(&screen
->eng2d
);
509 nouveau_object_del(&screen
->m2mf
);
510 nouveau_object_del(&screen
->compute
);
511 nouveau_object_del(&screen
->sync
);
513 nouveau_screen_fini(&screen
->base
);
519 nv50_screen_fence_emit(struct pipe_screen
*pscreen
, u32
*sequence
)
521 struct nv50_screen
*screen
= nv50_screen(pscreen
);
522 struct nouveau_pushbuf
*push
= screen
->base
.pushbuf
;
524 /* we need to do it after possible flush in MARK_RING */
525 *sequence
= ++screen
->base
.fence
.sequence
;
527 assert(PUSH_AVAIL(push
) + push
->rsvd_kick
>= 5);
528 PUSH_DATA (push
, NV50_FIFO_PKHDR(NV50_3D(QUERY_ADDRESS_HIGH
), 4));
529 PUSH_DATAh(push
, screen
->fence
.bo
->offset
);
530 PUSH_DATA (push
, screen
->fence
.bo
->offset
);
531 PUSH_DATA (push
, *sequence
);
532 PUSH_DATA (push
, NV50_3D_QUERY_GET_MODE_WRITE_UNK0
|
533 NV50_3D_QUERY_GET_UNK4
|
534 NV50_3D_QUERY_GET_UNIT_CROP
|
535 NV50_3D_QUERY_GET_TYPE_QUERY
|
536 NV50_3D_QUERY_GET_QUERY_SELECT_ZERO
|
537 NV50_3D_QUERY_GET_SHORT
);
541 nv50_screen_fence_update(struct pipe_screen
*pscreen
)
543 return nv50_screen(pscreen
)->fence
.map
[0];
547 nv50_screen_init_hwctx(struct nv50_screen
*screen
)
549 struct nouveau_pushbuf
*push
= screen
->base
.pushbuf
;
550 struct nv04_fifo
*fifo
;
553 fifo
= (struct nv04_fifo
*)screen
->base
.channel
->data
;
555 BEGIN_NV04(push
, SUBC_M2MF(NV01_SUBCHAN_OBJECT
), 1);
556 PUSH_DATA (push
, screen
->m2mf
->handle
);
557 BEGIN_NV04(push
, SUBC_M2MF(NV03_M2MF_DMA_NOTIFY
), 3);
558 PUSH_DATA (push
, screen
->sync
->handle
);
559 PUSH_DATA (push
, fifo
->vram
);
560 PUSH_DATA (push
, fifo
->vram
);
562 BEGIN_NV04(push
, SUBC_2D(NV01_SUBCHAN_OBJECT
), 1);
563 PUSH_DATA (push
, screen
->eng2d
->handle
);
564 BEGIN_NV04(push
, NV50_2D(DMA_NOTIFY
), 4);
565 PUSH_DATA (push
, screen
->sync
->handle
);
566 PUSH_DATA (push
, fifo
->vram
);
567 PUSH_DATA (push
, fifo
->vram
);
568 PUSH_DATA (push
, fifo
->vram
);
569 BEGIN_NV04(push
, NV50_2D(OPERATION
), 1);
570 PUSH_DATA (push
, NV50_2D_OPERATION_SRCCOPY
);
571 BEGIN_NV04(push
, NV50_2D(CLIP_ENABLE
), 1);
573 BEGIN_NV04(push
, NV50_2D(COLOR_KEY_ENABLE
), 1);
575 BEGIN_NV04(push
, SUBC_2D(0x0888), 1);
577 BEGIN_NV04(push
, NV50_2D(COND_MODE
), 1);
578 PUSH_DATA (push
, NV50_2D_COND_MODE_ALWAYS
);
580 BEGIN_NV04(push
, SUBC_3D(NV01_SUBCHAN_OBJECT
), 1);
581 PUSH_DATA (push
, screen
->tesla
->handle
);
583 BEGIN_NV04(push
, NV50_3D(COND_MODE
), 1);
584 PUSH_DATA (push
, NV50_3D_COND_MODE_ALWAYS
);
586 BEGIN_NV04(push
, NV50_3D(DMA_NOTIFY
), 1);
587 PUSH_DATA (push
, screen
->sync
->handle
);
588 BEGIN_NV04(push
, NV50_3D(DMA_ZETA
), 11);
589 for (i
= 0; i
< 11; ++i
)
590 PUSH_DATA(push
, fifo
->vram
);
591 BEGIN_NV04(push
, NV50_3D(DMA_COLOR(0)), NV50_3D_DMA_COLOR__LEN
);
592 for (i
= 0; i
< NV50_3D_DMA_COLOR__LEN
; ++i
)
593 PUSH_DATA(push
, fifo
->vram
);
595 BEGIN_NV04(push
, NV50_3D(REG_MODE
), 1);
596 PUSH_DATA (push
, NV50_3D_REG_MODE_STRIPED
);
597 BEGIN_NV04(push
, NV50_3D(UNK1400_LANES
), 1);
598 PUSH_DATA (push
, 0xf);
600 if (debug_get_bool_option("NOUVEAU_SHADER_WATCHDOG", true)) {
601 BEGIN_NV04(push
, NV50_3D(WATCHDOG_TIMER
), 1);
602 PUSH_DATA (push
, 0x18);
605 BEGIN_NV04(push
, NV50_3D(ZETA_COMP_ENABLE
), 1);
606 PUSH_DATA(push
, screen
->base
.drm
->version
>= 0x01000101);
608 BEGIN_NV04(push
, NV50_3D(RT_COMP_ENABLE(0)), 8);
609 for (i
= 0; i
< 8; ++i
)
610 PUSH_DATA(push
, screen
->base
.drm
->version
>= 0x01000101);
612 BEGIN_NV04(push
, NV50_3D(RT_CONTROL
), 1);
615 BEGIN_NV04(push
, NV50_3D(CSAA_ENABLE
), 1);
617 BEGIN_NV04(push
, NV50_3D(MULTISAMPLE_ENABLE
), 1);
619 BEGIN_NV04(push
, NV50_3D(MULTISAMPLE_MODE
), 1);
620 PUSH_DATA (push
, NV50_3D_MULTISAMPLE_MODE_MS1
);
621 BEGIN_NV04(push
, NV50_3D(MULTISAMPLE_CTRL
), 1);
623 BEGIN_NV04(push
, NV50_3D(PRIM_RESTART_WITH_DRAW_ARRAYS
), 1);
625 BEGIN_NV04(push
, NV50_3D(BLEND_SEPARATE_ALPHA
), 1);
628 if (screen
->tesla
->oclass
>= NVA0_3D_CLASS
) {
629 BEGIN_NV04(push
, SUBC_3D(NVA0_3D_TEX_MISC
), 1);
633 BEGIN_NV04(push
, NV50_3D(SCREEN_Y_CONTROL
), 1);
635 BEGIN_NV04(push
, NV50_3D(WINDOW_OFFSET_X
), 2);
638 BEGIN_NV04(push
, NV50_3D(ZCULL_REGION
), 1);
639 PUSH_DATA (push
, 0x3f);
641 BEGIN_NV04(push
, NV50_3D(VP_ADDRESS_HIGH
), 2);
642 PUSH_DATAh(push
, screen
->code
->offset
+ (0 << NV50_CODE_BO_SIZE_LOG2
));
643 PUSH_DATA (push
, screen
->code
->offset
+ (0 << NV50_CODE_BO_SIZE_LOG2
));
645 BEGIN_NV04(push
, NV50_3D(FP_ADDRESS_HIGH
), 2);
646 PUSH_DATAh(push
, screen
->code
->offset
+ (1 << NV50_CODE_BO_SIZE_LOG2
));
647 PUSH_DATA (push
, screen
->code
->offset
+ (1 << NV50_CODE_BO_SIZE_LOG2
));
649 BEGIN_NV04(push
, NV50_3D(GP_ADDRESS_HIGH
), 2);
650 PUSH_DATAh(push
, screen
->code
->offset
+ (2 << NV50_CODE_BO_SIZE_LOG2
));
651 PUSH_DATA (push
, screen
->code
->offset
+ (2 << NV50_CODE_BO_SIZE_LOG2
));
653 BEGIN_NV04(push
, NV50_3D(LOCAL_ADDRESS_HIGH
), 3);
654 PUSH_DATAh(push
, screen
->tls_bo
->offset
);
655 PUSH_DATA (push
, screen
->tls_bo
->offset
);
656 PUSH_DATA (push
, util_logbase2(screen
->cur_tls_space
/ 8));
658 BEGIN_NV04(push
, NV50_3D(STACK_ADDRESS_HIGH
), 3);
659 PUSH_DATAh(push
, screen
->stack_bo
->offset
);
660 PUSH_DATA (push
, screen
->stack_bo
->offset
);
663 BEGIN_NV04(push
, NV50_3D(CB_DEF_ADDRESS_HIGH
), 3);
664 PUSH_DATAh(push
, screen
->uniforms
->offset
+ (0 << 16));
665 PUSH_DATA (push
, screen
->uniforms
->offset
+ (0 << 16));
666 PUSH_DATA (push
, (NV50_CB_PVP
<< 16) | 0x0000);
668 BEGIN_NV04(push
, NV50_3D(CB_DEF_ADDRESS_HIGH
), 3);
669 PUSH_DATAh(push
, screen
->uniforms
->offset
+ (1 << 16));
670 PUSH_DATA (push
, screen
->uniforms
->offset
+ (1 << 16));
671 PUSH_DATA (push
, (NV50_CB_PGP
<< 16) | 0x0000);
673 BEGIN_NV04(push
, NV50_3D(CB_DEF_ADDRESS_HIGH
), 3);
674 PUSH_DATAh(push
, screen
->uniforms
->offset
+ (2 << 16));
675 PUSH_DATA (push
, screen
->uniforms
->offset
+ (2 << 16));
676 PUSH_DATA (push
, (NV50_CB_PFP
<< 16) | 0x0000);
678 BEGIN_NV04(push
, NV50_3D(CB_DEF_ADDRESS_HIGH
), 3);
679 PUSH_DATAh(push
, screen
->uniforms
->offset
+ (3 << 16));
680 PUSH_DATA (push
, screen
->uniforms
->offset
+ (3 << 16));
681 PUSH_DATA (push
, (NV50_CB_AUX
<< 16) | (NV50_CB_AUX_SIZE
& 0xffff));
683 BEGIN_NI04(push
, NV50_3D(SET_PROGRAM_CB
), 3);
684 PUSH_DATA (push
, (NV50_CB_AUX
<< 12) | 0xf01);
685 PUSH_DATA (push
, (NV50_CB_AUX
<< 12) | 0xf21);
686 PUSH_DATA (push
, (NV50_CB_AUX
<< 12) | 0xf31);
688 /* return { 0.0, 0.0, 0.0, 0.0 } on out-of-bounds vtxbuf access */
689 BEGIN_NV04(push
, NV50_3D(CB_ADDR
), 1);
690 PUSH_DATA (push
, (NV50_CB_AUX_RUNOUT_OFFSET
<< (8 - 2)) | NV50_CB_AUX
);
691 BEGIN_NI04(push
, NV50_3D(CB_DATA(0)), 4);
692 PUSH_DATAf(push
, 0.0f
);
693 PUSH_DATAf(push
, 0.0f
);
694 PUSH_DATAf(push
, 0.0f
);
695 PUSH_DATAf(push
, 0.0f
);
696 BEGIN_NV04(push
, NV50_3D(VERTEX_RUNOUT_ADDRESS_HIGH
), 2);
697 PUSH_DATAh(push
, screen
->uniforms
->offset
+ (3 << 16) + NV50_CB_AUX_RUNOUT_OFFSET
);
698 PUSH_DATA (push
, screen
->uniforms
->offset
+ (3 << 16) + NV50_CB_AUX_RUNOUT_OFFSET
);
700 nv50_upload_ms_info(push
);
702 /* max TIC (bits 4:8) & TSC bindings, per program type */
703 for (i
= 0; i
< 3; ++i
) {
704 BEGIN_NV04(push
, NV50_3D(TEX_LIMITS(i
)), 1);
705 PUSH_DATA (push
, 0x54);
708 BEGIN_NV04(push
, NV50_3D(TIC_ADDRESS_HIGH
), 3);
709 PUSH_DATAh(push
, screen
->txc
->offset
);
710 PUSH_DATA (push
, screen
->txc
->offset
);
711 PUSH_DATA (push
, NV50_TIC_MAX_ENTRIES
- 1);
713 BEGIN_NV04(push
, NV50_3D(TSC_ADDRESS_HIGH
), 3);
714 PUSH_DATAh(push
, screen
->txc
->offset
+ 65536);
715 PUSH_DATA (push
, screen
->txc
->offset
+ 65536);
716 PUSH_DATA (push
, NV50_TSC_MAX_ENTRIES
- 1);
718 BEGIN_NV04(push
, NV50_3D(LINKED_TSC
), 1);
721 BEGIN_NV04(push
, NV50_3D(CLIP_RECTS_EN
), 1);
723 BEGIN_NV04(push
, NV50_3D(CLIP_RECTS_MODE
), 1);
724 PUSH_DATA (push
, NV50_3D_CLIP_RECTS_MODE_INSIDE_ANY
);
725 BEGIN_NV04(push
, NV50_3D(CLIP_RECT_HORIZ(0)), 8 * 2);
726 for (i
= 0; i
< 8 * 2; ++i
)
728 BEGIN_NV04(push
, NV50_3D(CLIPID_ENABLE
), 1);
731 BEGIN_NV04(push
, NV50_3D(VIEWPORT_TRANSFORM_EN
), 1);
733 for (i
= 0; i
< NV50_MAX_VIEWPORTS
; i
++) {
734 BEGIN_NV04(push
, NV50_3D(DEPTH_RANGE_NEAR(i
)), 2);
735 PUSH_DATAf(push
, 0.0f
);
736 PUSH_DATAf(push
, 1.0f
);
737 BEGIN_NV04(push
, NV50_3D(VIEWPORT_HORIZ(i
)), 2);
738 PUSH_DATA (push
, 8192 << 16);
739 PUSH_DATA (push
, 8192 << 16);
742 BEGIN_NV04(push
, NV50_3D(VIEW_VOLUME_CLIP_CTRL
), 1);
743 #ifdef NV50_SCISSORS_CLIPPING
744 PUSH_DATA (push
, 0x0000);
746 PUSH_DATA (push
, 0x1080);
749 BEGIN_NV04(push
, NV50_3D(CLEAR_FLAGS
), 1);
750 PUSH_DATA (push
, NV50_3D_CLEAR_FLAGS_CLEAR_RECT_VIEWPORT
);
752 /* We use scissors instead of exact view volume clipping,
753 * so they're always enabled.
755 for (i
= 0; i
< NV50_MAX_VIEWPORTS
; i
++) {
756 BEGIN_NV04(push
, NV50_3D(SCISSOR_ENABLE(i
)), 3);
758 PUSH_DATA (push
, 8192 << 16);
759 PUSH_DATA (push
, 8192 << 16);
762 BEGIN_NV04(push
, NV50_3D(RASTERIZE_ENABLE
), 1);
764 BEGIN_NV04(push
, NV50_3D(POINT_RASTER_RULES
), 1);
765 PUSH_DATA (push
, NV50_3D_POINT_RASTER_RULES_OGL
);
766 BEGIN_NV04(push
, NV50_3D(FRAG_COLOR_CLAMP_EN
), 1);
767 PUSH_DATA (push
, 0x11111111);
768 BEGIN_NV04(push
, NV50_3D(EDGEFLAG
), 1);
771 BEGIN_NV04(push
, NV50_3D(VB_ELEMENT_BASE
), 1);
773 if (screen
->base
.class_3d
>= NV84_3D_CLASS
) {
774 BEGIN_NV04(push
, NV84_3D(VERTEX_ID_BASE
), 1);
778 BEGIN_NV04(push
, NV50_3D(UNK0FDC
), 1);
780 BEGIN_NV04(push
, NV50_3D(UNK19C0
), 1);
786 static int nv50_tls_alloc(struct nv50_screen
*screen
, unsigned tls_space
,
789 struct nouveau_device
*dev
= screen
->base
.device
;
792 screen
->cur_tls_space
= util_next_power_of_two(tls_space
/ ONE_TEMP_SIZE
) *
794 if (nouveau_mesa_debug
)
795 debug_printf("allocating space for %u temps\n",
796 util_next_power_of_two(tls_space
/ ONE_TEMP_SIZE
));
797 *tls_size
= screen
->cur_tls_space
* util_next_power_of_two(screen
->TPs
) *
798 screen
->MPsInTP
* LOCAL_WARPS_ALLOC
* THREADS_IN_WARP
;
800 ret
= nouveau_bo_new(dev
, NOUVEAU_BO_VRAM
, 1 << 16,
801 *tls_size
, NULL
, &screen
->tls_bo
);
803 NOUVEAU_ERR("Failed to allocate local bo: %d\n", ret
);
810 int nv50_tls_realloc(struct nv50_screen
*screen
, unsigned tls_space
)
812 struct nouveau_pushbuf
*push
= screen
->base
.pushbuf
;
816 if (tls_space
< screen
->cur_tls_space
)
818 if (tls_space
> screen
->max_tls_space
) {
819 /* fixable by limiting number of warps (LOCAL_WARPS_LOG_ALLOC /
820 * LOCAL_WARPS_NO_CLAMP) */
821 NOUVEAU_ERR("Unsupported number of temporaries (%u > %u). Fixable if someone cares.\n",
822 (unsigned)(tls_space
/ ONE_TEMP_SIZE
),
823 (unsigned)(screen
->max_tls_space
/ ONE_TEMP_SIZE
));
827 nouveau_bo_ref(NULL
, &screen
->tls_bo
);
828 ret
= nv50_tls_alloc(screen
, tls_space
, &tls_size
);
832 BEGIN_NV04(push
, NV50_3D(LOCAL_ADDRESS_HIGH
), 3);
833 PUSH_DATAh(push
, screen
->tls_bo
->offset
);
834 PUSH_DATA (push
, screen
->tls_bo
->offset
);
835 PUSH_DATA (push
, util_logbase2(screen
->cur_tls_space
/ 8));
840 struct nouveau_screen
*
841 nv50_screen_create(struct nouveau_device
*dev
)
843 struct nv50_screen
*screen
;
844 struct pipe_screen
*pscreen
;
845 struct nouveau_object
*chan
;
847 uint32_t tesla_class
;
851 screen
= CALLOC_STRUCT(nv50_screen
);
854 pscreen
= &screen
->base
.base
;
855 pscreen
->destroy
= nv50_screen_destroy
;
857 ret
= nouveau_screen_init(&screen
->base
, dev
);
859 NOUVEAU_ERR("nouveau_screen_init failed: %d\n", ret
);
863 /* TODO: Prevent FIFO prefetch before transfer of index buffers and
864 * admit them to VRAM.
866 screen
->base
.vidmem_bindings
|= PIPE_BIND_CONSTANT_BUFFER
|
867 PIPE_BIND_VERTEX_BUFFER
;
868 screen
->base
.sysmem_bindings
|=
869 PIPE_BIND_VERTEX_BUFFER
| PIPE_BIND_INDEX_BUFFER
;
871 screen
->base
.pushbuf
->user_priv
= screen
;
872 screen
->base
.pushbuf
->rsvd_kick
= 5;
874 chan
= screen
->base
.channel
;
876 pscreen
->context_create
= nv50_create
;
877 pscreen
->is_format_supported
= nv50_screen_is_format_supported
;
878 pscreen
->get_param
= nv50_screen_get_param
;
879 pscreen
->get_shader_param
= nv50_screen_get_shader_param
;
880 pscreen
->get_paramf
= nv50_screen_get_paramf
;
881 pscreen
->get_compute_param
= nv50_screen_get_compute_param
;
882 pscreen
->get_driver_query_info
= nv50_screen_get_driver_query_info
;
883 pscreen
->get_driver_query_group_info
= nv50_screen_get_driver_query_group_info
;
885 nv50_screen_init_resource_functions(pscreen
);
887 if (screen
->base
.device
->chipset
< 0x84 ||
888 debug_get_bool_option("NOUVEAU_PMPEG", false)) {
890 nouveau_screen_init_vdec(&screen
->base
);
891 } else if (screen
->base
.device
->chipset
< 0x98 ||
892 screen
->base
.device
->chipset
== 0xa0) {
894 screen
->base
.base
.get_video_param
= nv84_screen_get_video_param
;
895 screen
->base
.base
.is_video_format_supported
= nv84_screen_video_supported
;
898 screen
->base
.base
.get_video_param
= nouveau_vp3_screen_get_video_param
;
899 screen
->base
.base
.is_video_format_supported
= nouveau_vp3_screen_video_supported
;
902 ret
= nouveau_bo_new(dev
, NOUVEAU_BO_GART
| NOUVEAU_BO_MAP
, 0, 4096,
903 NULL
, &screen
->fence
.bo
);
905 NOUVEAU_ERR("Failed to allocate fence bo: %d\n", ret
);
909 nouveau_bo_map(screen
->fence
.bo
, 0, NULL
);
910 screen
->fence
.map
= screen
->fence
.bo
->map
;
911 screen
->base
.fence
.emit
= nv50_screen_fence_emit
;
912 screen
->base
.fence
.update
= nv50_screen_fence_update
;
914 ret
= nouveau_object_new(chan
, 0xbeef0301, NOUVEAU_NOTIFIER_CLASS
,
915 &(struct nv04_notify
){ .length
= 32 },
916 sizeof(struct nv04_notify
), &screen
->sync
);
918 NOUVEAU_ERR("Failed to allocate notifier: %d\n", ret
);
922 ret
= nouveau_object_new(chan
, 0xbeef5039, NV50_M2MF_CLASS
,
923 NULL
, 0, &screen
->m2mf
);
925 NOUVEAU_ERR("Failed to allocate PGRAPH context for M2MF: %d\n", ret
);
929 ret
= nouveau_object_new(chan
, 0xbeef502d, NV50_2D_CLASS
,
930 NULL
, 0, &screen
->eng2d
);
932 NOUVEAU_ERR("Failed to allocate PGRAPH context for 2D: %d\n", ret
);
936 switch (dev
->chipset
& 0xf0) {
938 tesla_class
= NV50_3D_CLASS
;
942 tesla_class
= NV84_3D_CLASS
;
945 switch (dev
->chipset
) {
949 tesla_class
= NVA0_3D_CLASS
;
952 tesla_class
= NVAF_3D_CLASS
;
955 tesla_class
= NVA3_3D_CLASS
;
960 NOUVEAU_ERR("Not a known NV50 chipset: NV%02x\n", dev
->chipset
);
963 screen
->base
.class_3d
= tesla_class
;
965 ret
= nouveau_object_new(chan
, 0xbeef5097, tesla_class
,
966 NULL
, 0, &screen
->tesla
);
968 NOUVEAU_ERR("Failed to allocate PGRAPH context for 3D: %d\n", ret
);
972 /* This over-allocates by a page. The GP, which would execute at the end of
973 * the last page, would trigger faults. The going theory is that it
974 * prefetches up to a certain amount.
976 ret
= nouveau_bo_new(dev
, NOUVEAU_BO_VRAM
, 1 << 16,
977 (3 << NV50_CODE_BO_SIZE_LOG2
) + 0x1000,
978 NULL
, &screen
->code
);
980 NOUVEAU_ERR("Failed to allocate code bo: %d\n", ret
);
984 nouveau_heap_init(&screen
->vp_code_heap
, 0, 1 << NV50_CODE_BO_SIZE_LOG2
);
985 nouveau_heap_init(&screen
->gp_code_heap
, 0, 1 << NV50_CODE_BO_SIZE_LOG2
);
986 nouveau_heap_init(&screen
->fp_code_heap
, 0, 1 << NV50_CODE_BO_SIZE_LOG2
);
988 nouveau_getparam(dev
, NOUVEAU_GETPARAM_GRAPH_UNITS
, &value
);
990 screen
->TPs
= util_bitcount(value
& 0xffff);
991 screen
->MPsInTP
= util_bitcount((value
>> 24) & 0xf);
993 screen
->mp_count
= screen
->TPs
* screen
->MPsInTP
;
995 stack_size
= util_next_power_of_two(screen
->TPs
) * screen
->MPsInTP
*
996 STACK_WARPS_ALLOC
* 64 * 8;
998 ret
= nouveau_bo_new(dev
, NOUVEAU_BO_VRAM
, 1 << 16, stack_size
, NULL
,
1001 NOUVEAU_ERR("Failed to allocate stack bo: %d\n", ret
);
1005 uint64_t size_of_one_temp
= util_next_power_of_two(screen
->TPs
) *
1006 screen
->MPsInTP
* LOCAL_WARPS_ALLOC
* THREADS_IN_WARP
*
1008 screen
->max_tls_space
= dev
->vram_size
/ size_of_one_temp
* ONE_TEMP_SIZE
;
1009 screen
->max_tls_space
/= 2; /* half of vram */
1011 /* hw can address max 64 KiB */
1012 screen
->max_tls_space
= MIN2(screen
->max_tls_space
, 64 << 10);
1015 unsigned tls_space
= 4/*temps*/ * ONE_TEMP_SIZE
;
1016 ret
= nv50_tls_alloc(screen
, tls_space
, &tls_size
);
1020 if (nouveau_mesa_debug
)
1021 debug_printf("TPs = %u, MPsInTP = %u, VRAM = %"PRIu64
" MiB, tls_size = %"PRIu64
" KiB\n",
1022 screen
->TPs
, screen
->MPsInTP
, dev
->vram_size
>> 20, tls_size
>> 10);
1024 ret
= nouveau_bo_new(dev
, NOUVEAU_BO_VRAM
, 1 << 16, 4 << 16, NULL
,
1027 NOUVEAU_ERR("Failed to allocate uniforms bo: %d\n", ret
);
1031 ret
= nouveau_bo_new(dev
, NOUVEAU_BO_VRAM
, 1 << 16, 3 << 16, NULL
,
1034 NOUVEAU_ERR("Failed to allocate TIC/TSC bo: %d\n", ret
);
1038 screen
->tic
.entries
= CALLOC(4096, sizeof(void *));
1039 screen
->tsc
.entries
= screen
->tic
.entries
+ 2048;
1041 if (!nv50_blitter_create(screen
))
1044 nv50_screen_init_hwctx(screen
);
1046 ret
= nv50_screen_compute_setup(screen
, screen
->base
.pushbuf
);
1048 NOUVEAU_ERR("Failed to init compute context: %d\n", ret
);
1052 nouveau_fence_new(&screen
->base
, &screen
->base
.fence
.current
);
1054 return &screen
->base
;
1057 screen
->base
.base
.context_create
= NULL
;
1058 return &screen
->base
;
1062 nv50_screen_tic_alloc(struct nv50_screen
*screen
, void *entry
)
1064 int i
= screen
->tic
.next
;
1066 while (screen
->tic
.lock
[i
/ 32] & (1 << (i
% 32)))
1067 i
= (i
+ 1) & (NV50_TIC_MAX_ENTRIES
- 1);
1069 screen
->tic
.next
= (i
+ 1) & (NV50_TIC_MAX_ENTRIES
- 1);
1071 if (screen
->tic
.entries
[i
])
1072 nv50_tic_entry(screen
->tic
.entries
[i
])->id
= -1;
1074 screen
->tic
.entries
[i
] = entry
;
1079 nv50_screen_tsc_alloc(struct nv50_screen
*screen
, void *entry
)
1081 int i
= screen
->tsc
.next
;
1083 while (screen
->tsc
.lock
[i
/ 32] & (1 << (i
% 32)))
1084 i
= (i
+ 1) & (NV50_TSC_MAX_ENTRIES
- 1);
1086 screen
->tsc
.next
= (i
+ 1) & (NV50_TSC_MAX_ENTRIES
- 1);
1088 if (screen
->tsc
.entries
[i
])
1089 nv50_tsc_entry(screen
->tsc
.entries
[i
])->id
= -1;
1091 screen
->tsc
.entries
[i
] = entry
;