1 /**************************************************************************
3 * Copyright 2007 Tungsten Graphics, Inc., Cedar Park, Texas.
6 * Permission is hereby granted, free of charge, to any person obtaining a
7 * copy of this software and associated documentation files (the
8 * "Software"), to deal in the Software without restriction, including
9 * without limitation the rights to use, copy, modify, merge, publish,
10 * distribute, sub license, and/or sell copies of the Software, and to
11 * permit persons to whom the Software is furnished to do so, subject to
12 * the following conditions:
14 * The above copyright notice and this permission notice (including the
15 * next paragraph) shall be included in all copies or substantial portions
18 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
19 * OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
20 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT.
21 * IN NO EVENT SHALL TUNGSTEN GRAPHICS AND/OR ITS SUPPLIERS BE LIABLE FOR
22 * ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,
23 * TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
24 * SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
26 **************************************************************************/
31 #include "pipe/p_context.h"
32 #include "pipe/p_defines.h"
33 #include "pipe/p_state.h"
34 #include "pipe/p_screen.h"
35 #include "util/u_debug.h"
36 #include "util/u_atomic.h"
45 * Reference counting helper functions.
50 pipe_reference_init(struct pipe_reference
*reference
, unsigned count
)
52 p_atomic_set(&reference
->count
, count
);
56 pipe_is_referenced(struct pipe_reference
*reference
)
58 return p_atomic_read(&reference
->count
) != 0;
62 * Update reference counting.
63 * The old thing pointed to, if any, will be unreferenced.
64 * Both 'ptr' and 'reference' may be NULL.
65 * \return TRUE if the object's refcount hits zero and should be destroyed.
68 pipe_reference(struct pipe_reference
*ptr
, struct pipe_reference
*reference
)
70 boolean destroy
= FALSE
;
72 if(ptr
!= reference
) {
73 /* bump the reference.count first */
75 assert(pipe_is_referenced(reference
));
76 p_atomic_inc(&reference
->count
);
80 assert(pipe_is_referenced(ptr
));
81 if (p_atomic_dec_zero(&ptr
->count
)) {
91 pipe_buffer_reference(struct pipe_buffer
**ptr
, struct pipe_buffer
*buf
)
93 struct pipe_buffer
*old_buf
;
98 if (pipe_reference(&(*ptr
)->reference
, &buf
->reference
))
99 old_buf
->screen
->buffer_destroy(old_buf
);
104 pipe_surface_reference(struct pipe_surface
**ptr
, struct pipe_surface
*surf
)
106 struct pipe_surface
*old_surf
= *ptr
;
108 if (pipe_reference(&(*ptr
)->reference
, &surf
->reference
))
109 old_surf
->texture
->screen
->tex_surface_destroy(old_surf
);
114 pipe_texture_reference(struct pipe_texture
**ptr
, struct pipe_texture
*tex
)
116 struct pipe_texture
*old_tex
= *ptr
;
118 if (pipe_reference(&(*ptr
)->reference
, &tex
->reference
))
119 old_tex
->screen
->texture_destroy(old_tex
);
125 * Convenience wrappers for screen buffer functions.
128 static INLINE
struct pipe_buffer
*
129 pipe_buffer_create( struct pipe_screen
*screen
,
130 unsigned alignment
, unsigned usage
, unsigned size
)
132 return screen
->buffer_create(screen
, alignment
, usage
, size
);
135 static INLINE
struct pipe_buffer
*
136 pipe_user_buffer_create( struct pipe_screen
*screen
, void *ptr
, unsigned size
)
138 return screen
->user_buffer_create(screen
, ptr
, size
);
142 pipe_buffer_map(struct pipe_screen
*screen
,
143 struct pipe_buffer
*buf
,
146 if(screen
->buffer_map_range
) {
148 unsigned length
= buf
->size
;
149 return screen
->buffer_map_range(screen
, buf
, offset
, length
, usage
);
152 return screen
->buffer_map(screen
, buf
, usage
);
156 pipe_buffer_unmap(struct pipe_screen
*screen
,
157 struct pipe_buffer
*buf
)
159 screen
->buffer_unmap(screen
, buf
);
163 pipe_buffer_map_range(struct pipe_screen
*screen
,
164 struct pipe_buffer
*buf
,
169 assert(offset
< buf
->size
);
170 assert(offset
+ length
<= buf
->size
);
172 if(screen
->buffer_map_range
)
173 return screen
->buffer_map_range(screen
, buf
, offset
, length
, usage
);
175 return screen
->buffer_map(screen
, buf
, usage
);
179 pipe_buffer_flush_mapped_range(struct pipe_screen
*screen
,
180 struct pipe_buffer
*buf
,
184 assert(offset
< buf
->size
);
185 assert(offset
+ length
<= buf
->size
);
187 if(screen
->buffer_flush_mapped_range
)
188 screen
->buffer_flush_mapped_range(screen
, buf
, offset
, length
);
192 pipe_buffer_write(struct pipe_screen
*screen
,
193 struct pipe_buffer
*buf
,
194 unsigned offset
, unsigned size
,
199 assert(offset
< buf
->size
);
200 assert(offset
+ size
<= buf
->size
);
203 map
= pipe_buffer_map_range(screen
, buf
, offset
, size
,
204 PIPE_BUFFER_USAGE_CPU_WRITE
|
205 PIPE_BUFFER_USAGE_FLUSH_EXPLICIT
|
206 PIPE_BUFFER_USAGE_DISCARD
);
209 memcpy((uint8_t *)map
+ offset
, data
, size
);
210 pipe_buffer_flush_mapped_range(screen
, buf
, offset
, size
);
211 pipe_buffer_unmap(screen
, buf
);
216 * Special case for writing non-overlapping ranges.
218 * We can avoid GPU/CPU synchronization when writing range that has never
219 * been written before.
222 pipe_buffer_write_nooverlap(struct pipe_screen
*screen
,
223 struct pipe_buffer
*buf
,
224 unsigned offset
, unsigned size
,
229 assert(offset
< buf
->size
);
230 assert(offset
+ size
<= buf
->size
);
233 map
= pipe_buffer_map_range(screen
, buf
, offset
, size
,
234 PIPE_BUFFER_USAGE_CPU_WRITE
|
235 PIPE_BUFFER_USAGE_FLUSH_EXPLICIT
|
236 PIPE_BUFFER_USAGE_DISCARD
|
237 PIPE_BUFFER_USAGE_UNSYNCHRONIZED
);
240 memcpy((uint8_t *)map
+ offset
, data
, size
);
241 pipe_buffer_flush_mapped_range(screen
, buf
, offset
, size
);
242 pipe_buffer_unmap(screen
, buf
);
247 pipe_buffer_read(struct pipe_screen
*screen
,
248 struct pipe_buffer
*buf
,
249 unsigned offset
, unsigned size
,
254 assert(offset
< buf
->size
);
255 assert(offset
+ size
<= buf
->size
);
258 map
= pipe_buffer_map_range(screen
, buf
, offset
, size
, PIPE_BUFFER_USAGE_CPU_READ
);
261 memcpy(data
, (const uint8_t *)map
+ offset
, size
);
262 pipe_buffer_unmap(screen
, buf
);
267 pipe_transfer_map( struct pipe_context
*context
,
268 struct pipe_transfer
*transf
)
270 return context
->transfer_map(context
, transf
);
274 pipe_transfer_unmap( struct pipe_context
*context
,
275 struct pipe_transfer
*transf
)
277 context
->transfer_unmap(context
, transf
);
281 pipe_transfer_destroy( struct pipe_context
*context
,
282 struct pipe_transfer
*transfer
)
284 context
->tex_transfer_destroy(context
, transfer
);
287 static INLINE
unsigned
288 pipe_transfer_buffer_flags( struct pipe_transfer
*transf
)
290 switch (transf
->usage
& PIPE_TRANSFER_READ_WRITE
) {
291 case PIPE_TRANSFER_READ_WRITE
:
292 return PIPE_BUFFER_USAGE_CPU_READ
| PIPE_BUFFER_USAGE_CPU_WRITE
;
293 case PIPE_TRANSFER_READ
:
294 return PIPE_BUFFER_USAGE_CPU_READ
;
295 case PIPE_TRANSFER_WRITE
:
296 return PIPE_BUFFER_USAGE_CPU_WRITE
;
307 #endif /* U_INLINES_H */