gallium: add start_slot parameter to set_vertex_buffers
[mesa.git] / src / gallium / auxiliary / util / u_inlines.h
1 /**************************************************************************
2 *
3 * Copyright 2007 Tungsten Graphics, Inc., Cedar Park, Texas.
4 * All Rights Reserved.
5 *
6 * Permission is hereby granted, free of charge, to any person obtaining a
7 * copy of this software and associated documentation files (the
8 * "Software"), to deal in the Software without restriction, including
9 * without limitation the rights to use, copy, modify, merge, publish,
10 * distribute, sub license, and/or sell copies of the Software, and to
11 * permit persons to whom the Software is furnished to do so, subject to
12 * the following conditions:
13 *
14 * The above copyright notice and this permission notice (including the
15 * next paragraph) shall be included in all copies or substantial portions
16 * of the Software.
17 *
18 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
19 * OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
20 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT.
21 * IN NO EVENT SHALL TUNGSTEN GRAPHICS AND/OR ITS SUPPLIERS BE LIABLE FOR
22 * ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,
23 * TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
24 * SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
25 *
26 **************************************************************************/
27
28 #ifndef U_INLINES_H
29 #define U_INLINES_H
30
31 #include "pipe/p_context.h"
32 #include "pipe/p_defines.h"
33 #include "pipe/p_shader_tokens.h"
34 #include "pipe/p_state.h"
35 #include "pipe/p_screen.h"
36 #include "util/u_debug.h"
37 #include "util/u_debug_describe.h"
38 #include "util/u_debug_refcnt.h"
39 #include "util/u_atomic.h"
40 #include "util/u_box.h"
41 #include "util/u_math.h"
42
43
44 #ifdef __cplusplus
45 extern "C" {
46 #endif
47
48
49 /*
50 * Reference counting helper functions.
51 */
52
53
54 static INLINE void
55 pipe_reference_init(struct pipe_reference *reference, unsigned count)
56 {
57 p_atomic_set(&reference->count, count);
58 }
59
60 static INLINE boolean
61 pipe_is_referenced(struct pipe_reference *reference)
62 {
63 return p_atomic_read(&reference->count) != 0;
64 }
65
66 /**
67 * Update reference counting.
68 * The old thing pointed to, if any, will be unreferenced.
69 * Both 'ptr' and 'reference' may be NULL.
70 * \return TRUE if the object's refcount hits zero and should be destroyed.
71 */
72 static INLINE boolean
73 pipe_reference_described(struct pipe_reference *ptr,
74 struct pipe_reference *reference,
75 debug_reference_descriptor get_desc)
76 {
77 boolean destroy = FALSE;
78
79 if(ptr != reference) {
80 /* bump the reference.count first */
81 if (reference) {
82 assert(pipe_is_referenced(reference));
83 p_atomic_inc(&reference->count);
84 debug_reference(reference, get_desc, 1);
85 }
86
87 if (ptr) {
88 assert(pipe_is_referenced(ptr));
89 if (p_atomic_dec_zero(&ptr->count)) {
90 destroy = TRUE;
91 }
92 debug_reference(ptr, get_desc, -1);
93 }
94 }
95
96 return destroy;
97 }
98
99 static INLINE boolean
100 pipe_reference(struct pipe_reference *ptr, struct pipe_reference *reference)
101 {
102 return pipe_reference_described(ptr, reference,
103 (debug_reference_descriptor)debug_describe_reference);
104 }
105
106 static INLINE void
107 pipe_surface_reference(struct pipe_surface **ptr, struct pipe_surface *surf)
108 {
109 struct pipe_surface *old_surf = *ptr;
110
111 if (pipe_reference_described(&(*ptr)->reference, &surf->reference,
112 (debug_reference_descriptor)debug_describe_surface))
113 old_surf->context->surface_destroy(old_surf->context, old_surf);
114 *ptr = surf;
115 }
116
117 static INLINE void
118 pipe_resource_reference(struct pipe_resource **ptr, struct pipe_resource *tex)
119 {
120 struct pipe_resource *old_tex = *ptr;
121
122 if (pipe_reference_described(&(*ptr)->reference, &tex->reference,
123 (debug_reference_descriptor)debug_describe_resource))
124 old_tex->screen->resource_destroy(old_tex->screen, old_tex);
125 *ptr = tex;
126 }
127
128 static INLINE void
129 pipe_sampler_view_reference(struct pipe_sampler_view **ptr, struct pipe_sampler_view *view)
130 {
131 struct pipe_sampler_view *old_view = *ptr;
132
133 if (pipe_reference_described(&(*ptr)->reference, &view->reference,
134 (debug_reference_descriptor)debug_describe_sampler_view))
135 old_view->context->sampler_view_destroy(old_view->context, old_view);
136 *ptr = view;
137 }
138
139 /**
140 * Similar to pipe_sampler_view_reference() but always set the pointer to
141 * NULL and pass in an explicit context. Passing an explicit context is a
142 * work-around for fixing a dangling context pointer problem when textures
143 * are shared by multiple contexts. XXX fix this someday.
144 */
145 static INLINE void
146 pipe_sampler_view_release(struct pipe_context *ctx,
147 struct pipe_sampler_view **ptr)
148 {
149 struct pipe_sampler_view *old_view = *ptr;
150 if (*ptr && (*ptr)->context != ctx) {
151 debug_printf_once(("context mis-match in pipe_sampler_view_release()\n"));
152 }
153 if (pipe_reference_described(&(*ptr)->reference, NULL,
154 (debug_reference_descriptor)debug_describe_sampler_view)) {
155 ctx->sampler_view_destroy(ctx, old_view);
156 }
157 *ptr = NULL;
158 }
159
160
161 static INLINE void
162 pipe_so_target_reference(struct pipe_stream_output_target **ptr,
163 struct pipe_stream_output_target *target)
164 {
165 struct pipe_stream_output_target *old = *ptr;
166
167 if (pipe_reference_described(&(*ptr)->reference, &target->reference,
168 (debug_reference_descriptor)debug_describe_so_target))
169 old->context->stream_output_target_destroy(old->context, old);
170 *ptr = target;
171 }
172
173 static INLINE void
174 pipe_surface_reset(struct pipe_context *ctx, struct pipe_surface* ps,
175 struct pipe_resource *pt, unsigned level, unsigned layer,
176 unsigned flags)
177 {
178 pipe_resource_reference(&ps->texture, pt);
179 ps->format = pt->format;
180 ps->width = u_minify(pt->width0, level);
181 ps->height = u_minify(pt->height0, level);
182 ps->usage = flags;
183 ps->u.tex.level = level;
184 ps->u.tex.first_layer = ps->u.tex.last_layer = layer;
185 ps->context = ctx;
186 }
187
188 static INLINE void
189 pipe_surface_init(struct pipe_context *ctx, struct pipe_surface* ps,
190 struct pipe_resource *pt, unsigned level, unsigned layer,
191 unsigned flags)
192 {
193 ps->texture = 0;
194 pipe_reference_init(&ps->reference, 1);
195 pipe_surface_reset(ctx, ps, pt, level, layer, flags);
196 }
197
198 /* Return true if the surfaces are equal. */
199 static INLINE boolean
200 pipe_surface_equal(struct pipe_surface *s1, struct pipe_surface *s2)
201 {
202 return s1->texture == s2->texture &&
203 s1->format == s2->format &&
204 (s1->texture->target != PIPE_BUFFER ||
205 (s1->u.buf.first_element == s2->u.buf.first_element &&
206 s1->u.buf.last_element == s2->u.buf.last_element)) &&
207 (s1->texture->target == PIPE_BUFFER ||
208 (s1->u.tex.level == s2->u.tex.level &&
209 s1->u.tex.first_layer == s2->u.tex.first_layer &&
210 s1->u.tex.last_layer == s2->u.tex.last_layer));
211 }
212
213 /*
214 * Convenience wrappers for screen buffer functions.
215 */
216
217 static INLINE struct pipe_resource *
218 pipe_buffer_create( struct pipe_screen *screen,
219 unsigned bind,
220 unsigned usage,
221 unsigned size )
222 {
223 struct pipe_resource buffer;
224 memset(&buffer, 0, sizeof buffer);
225 buffer.target = PIPE_BUFFER;
226 buffer.format = PIPE_FORMAT_R8_UNORM; /* want TYPELESS or similar */
227 buffer.bind = bind;
228 buffer.usage = usage;
229 buffer.flags = 0;
230 buffer.width0 = size;
231 buffer.height0 = 1;
232 buffer.depth0 = 1;
233 buffer.array_size = 1;
234 return screen->resource_create(screen, &buffer);
235 }
236
237 static INLINE void *
238 pipe_buffer_map_range(struct pipe_context *pipe,
239 struct pipe_resource *buffer,
240 unsigned offset,
241 unsigned length,
242 unsigned usage,
243 struct pipe_transfer **transfer)
244 {
245 struct pipe_box box;
246 void *map;
247
248 assert(offset < buffer->width0);
249 assert(offset + length <= buffer->width0);
250 assert(length);
251
252 u_box_1d(offset, length, &box);
253
254 map = pipe->transfer_map(pipe, buffer, 0, usage, &box, transfer);
255 if (map == NULL) {
256 return NULL;
257 }
258
259 return map;
260 }
261
262
263 static INLINE void *
264 pipe_buffer_map(struct pipe_context *pipe,
265 struct pipe_resource *buffer,
266 unsigned usage,
267 struct pipe_transfer **transfer)
268 {
269 return pipe_buffer_map_range(pipe, buffer, 0, buffer->width0, usage, transfer);
270 }
271
272
273 static INLINE void
274 pipe_buffer_unmap(struct pipe_context *pipe,
275 struct pipe_transfer *transfer)
276 {
277 pipe->transfer_unmap(pipe, transfer);
278 }
279
280 static INLINE void
281 pipe_buffer_flush_mapped_range(struct pipe_context *pipe,
282 struct pipe_transfer *transfer,
283 unsigned offset,
284 unsigned length)
285 {
286 struct pipe_box box;
287 int transfer_offset;
288
289 assert(length);
290 assert(transfer->box.x <= (int) offset);
291 assert((int) (offset + length) <= transfer->box.x + transfer->box.width);
292
293 /* Match old screen->buffer_flush_mapped_range() behaviour, where
294 * offset parameter is relative to the start of the buffer, not the
295 * mapped range.
296 */
297 transfer_offset = offset - transfer->box.x;
298
299 u_box_1d(transfer_offset, length, &box);
300
301 pipe->transfer_flush_region(pipe, transfer, &box);
302 }
303
304 static INLINE void
305 pipe_buffer_write(struct pipe_context *pipe,
306 struct pipe_resource *buf,
307 unsigned offset,
308 unsigned size,
309 const void *data)
310 {
311 struct pipe_box box;
312 unsigned usage = PIPE_TRANSFER_WRITE;
313
314 if (offset == 0 && size == buf->width0) {
315 usage |= PIPE_TRANSFER_DISCARD_WHOLE_RESOURCE;
316 } else {
317 usage |= PIPE_TRANSFER_DISCARD_RANGE;
318 }
319
320 u_box_1d(offset, size, &box);
321
322 pipe->transfer_inline_write( pipe,
323 buf,
324 0,
325 usage,
326 &box,
327 data,
328 size,
329 0);
330 }
331
332 /**
333 * Special case for writing non-overlapping ranges.
334 *
335 * We can avoid GPU/CPU synchronization when writing range that has never
336 * been written before.
337 */
338 static INLINE void
339 pipe_buffer_write_nooverlap(struct pipe_context *pipe,
340 struct pipe_resource *buf,
341 unsigned offset, unsigned size,
342 const void *data)
343 {
344 struct pipe_box box;
345
346 u_box_1d(offset, size, &box);
347
348 pipe->transfer_inline_write(pipe,
349 buf,
350 0,
351 (PIPE_TRANSFER_WRITE |
352 PIPE_TRANSFER_UNSYNCHRONIZED),
353 &box,
354 data,
355 0, 0);
356 }
357
358 static INLINE struct pipe_resource *
359 pipe_buffer_create_with_data(struct pipe_context *pipe,
360 unsigned bind,
361 unsigned usage,
362 unsigned size,
363 void *ptr)
364 {
365 struct pipe_resource *res = pipe_buffer_create(pipe->screen,
366 bind, usage, size);
367 pipe_buffer_write_nooverlap(pipe, res, 0, size, ptr);
368 return res;
369 }
370
371 static INLINE void
372 pipe_buffer_read(struct pipe_context *pipe,
373 struct pipe_resource *buf,
374 unsigned offset,
375 unsigned size,
376 void *data)
377 {
378 struct pipe_transfer *src_transfer;
379 ubyte *map;
380
381 map = (ubyte *) pipe_buffer_map_range(pipe,
382 buf,
383 offset, size,
384 PIPE_TRANSFER_READ,
385 &src_transfer);
386 if (!map)
387 return;
388
389 memcpy(data, map, size);
390 pipe_buffer_unmap(pipe, src_transfer);
391 }
392
393 static INLINE void *
394 pipe_transfer_map(struct pipe_context *context,
395 struct pipe_resource *resource,
396 unsigned level, unsigned layer,
397 enum pipe_transfer_usage usage,
398 unsigned x, unsigned y,
399 unsigned w, unsigned h,
400 struct pipe_transfer **transfer)
401 {
402 struct pipe_box box;
403 u_box_2d_zslice(x, y, layer, w, h, &box);
404 return context->transfer_map(context,
405 resource,
406 level,
407 usage,
408 &box, transfer);
409 }
410
411 static INLINE void
412 pipe_transfer_unmap( struct pipe_context *context,
413 struct pipe_transfer *transfer )
414 {
415 context->transfer_unmap( context, transfer );
416 }
417
418 static INLINE void
419 pipe_set_constant_buffer(struct pipe_context *pipe, uint shader, uint index,
420 struct pipe_resource *buf)
421 {
422 if (buf) {
423 struct pipe_constant_buffer cb;
424 cb.buffer = buf;
425 cb.buffer_offset = 0;
426 cb.buffer_size = buf->width0;
427 cb.user_buffer = NULL;
428 pipe->set_constant_buffer(pipe, shader, index, &cb);
429 } else {
430 pipe->set_constant_buffer(pipe, shader, index, NULL);
431 }
432 }
433
434
435 static INLINE boolean util_get_offset(
436 const struct pipe_rasterizer_state *templ,
437 unsigned fill_mode)
438 {
439 switch(fill_mode) {
440 case PIPE_POLYGON_MODE_POINT:
441 return templ->offset_point;
442 case PIPE_POLYGON_MODE_LINE:
443 return templ->offset_line;
444 case PIPE_POLYGON_MODE_FILL:
445 return templ->offset_tri;
446 default:
447 assert(0);
448 return FALSE;
449 }
450 }
451
452 static INLINE float
453 util_get_min_point_size(const struct pipe_rasterizer_state *state)
454 {
455 /* The point size should be clamped to this value at the rasterizer stage.
456 */
457 return state->gl_rasterization_rules &&
458 !state->point_quad_rasterization &&
459 !state->point_smooth &&
460 !state->multisample ? 1.0f : 0.0f;
461 }
462
463 static INLINE void
464 util_query_clear_result(union pipe_query_result *result, unsigned type)
465 {
466 switch (type) {
467 case PIPE_QUERY_OCCLUSION_PREDICATE:
468 case PIPE_QUERY_SO_OVERFLOW_PREDICATE:
469 case PIPE_QUERY_GPU_FINISHED:
470 result->b = FALSE;
471 break;
472 case PIPE_QUERY_OCCLUSION_COUNTER:
473 case PIPE_QUERY_TIMESTAMP:
474 case PIPE_QUERY_TIME_ELAPSED:
475 case PIPE_QUERY_PRIMITIVES_GENERATED:
476 case PIPE_QUERY_PRIMITIVES_EMITTED:
477 result->u64 = 0;
478 break;
479 case PIPE_QUERY_SO_STATISTICS:
480 memset(&result->so_statistics, 0, sizeof(result->so_statistics));
481 break;
482 case PIPE_QUERY_TIMESTAMP_DISJOINT:
483 memset(&result->timestamp_disjoint, 0, sizeof(result->timestamp_disjoint));
484 break;
485 case PIPE_QUERY_PIPELINE_STATISTICS:
486 memset(&result->pipeline_statistics, 0, sizeof(result->pipeline_statistics));
487 break;
488 default:
489 assert(0);
490 }
491 }
492
493 /** Convert PIPE_TEXTURE_x to TGSI_TEXTURE_x */
494 static INLINE unsigned
495 util_pipe_tex_to_tgsi_tex(enum pipe_texture_target pipe_tex_target,
496 unsigned nr_samples)
497 {
498 switch (pipe_tex_target) {
499 case PIPE_TEXTURE_1D:
500 assert(nr_samples <= 1);
501 return TGSI_TEXTURE_1D;
502
503 case PIPE_TEXTURE_2D:
504 return nr_samples > 1 ? TGSI_TEXTURE_2D_MSAA : TGSI_TEXTURE_2D;
505
506 case PIPE_TEXTURE_RECT:
507 assert(nr_samples <= 1);
508 return TGSI_TEXTURE_RECT;
509
510 case PIPE_TEXTURE_3D:
511 assert(nr_samples <= 1);
512 return TGSI_TEXTURE_3D;
513
514 case PIPE_TEXTURE_CUBE:
515 assert(nr_samples <= 1);
516 return TGSI_TEXTURE_CUBE;
517
518 case PIPE_TEXTURE_1D_ARRAY:
519 assert(nr_samples <= 1);
520 return TGSI_TEXTURE_1D_ARRAY;
521
522 case PIPE_TEXTURE_2D_ARRAY:
523 return nr_samples > 1 ? TGSI_TEXTURE_2D_ARRAY_MSAA :
524 TGSI_TEXTURE_2D_ARRAY;
525
526 default:
527 assert(0 && "unexpected texture target");
528 return TGSI_TEXTURE_UNKNOWN;
529 }
530 }
531
532 #ifdef __cplusplus
533 }
534 #endif
535
536 #endif /* U_INLINES_H */