1 /**************************************************************************
3 * Copyright 2007 Tungsten Graphics, Inc., Cedar Park, Texas.
6 * Permission is hereby granted, free of charge, to any person obtaining a
7 * copy of this software and associated documentation files (the
8 * "Software"), to deal in the Software without restriction, including
9 * without limitation the rights to use, copy, modify, merge, publish,
10 * distribute, sub license, and/or sell copies of the Software, and to
11 * permit persons to whom the Software is furnished to do so, subject to
12 * the following conditions:
14 * The above copyright notice and this permission notice (including the
15 * next paragraph) shall be included in all copies or substantial portions
18 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
19 * OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
20 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT.
21 * IN NO EVENT SHALL TUNGSTEN GRAPHICS AND/OR ITS SUPPLIERS BE LIABLE FOR
22 * ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,
23 * TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
24 * SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
26 **************************************************************************/
31 #include "pipe/p_context.h"
32 #include "pipe/p_defines.h"
33 #include "pipe/p_state.h"
34 #include "pipe/p_screen.h"
35 #include "util/u_debug.h"
36 #include "util/u_atomic.h"
45 * Reference counting helper functions.
50 pipe_reference_init(struct pipe_reference
*reference
, unsigned count
)
52 p_atomic_set(&reference
->count
, count
);
56 pipe_is_referenced(struct pipe_reference
*reference
)
58 return p_atomic_read(&reference
->count
) != 0;
62 * Update reference counting.
63 * The old thing pointed to, if any, will be unreferenced.
64 * Both 'ptr' and 'reference' may be NULL.
65 * \return TRUE if the object's refcount hits zero and should be destroyed.
68 pipe_reference(struct pipe_reference
*ptr
, struct pipe_reference
*reference
)
70 boolean destroy
= FALSE
;
72 if(ptr
!= reference
) {
73 /* bump the reference.count first */
75 assert(pipe_is_referenced(reference
));
76 p_atomic_inc(&reference
->count
);
80 assert(pipe_is_referenced(ptr
));
81 if (p_atomic_dec_zero(&ptr
->count
)) {
91 pipe_buffer_reference(struct pipe_buffer
**ptr
, struct pipe_buffer
*buf
)
93 struct pipe_buffer
*old_buf
= *ptr
;
95 if (pipe_reference(&(*ptr
)->reference
, &buf
->reference
))
96 old_buf
->screen
->buffer_destroy(old_buf
);
101 pipe_surface_reference(struct pipe_surface
**ptr
, struct pipe_surface
*surf
)
103 struct pipe_surface
*old_surf
= *ptr
;
105 if (pipe_reference(&(*ptr
)->reference
, &surf
->reference
))
106 old_surf
->texture
->screen
->tex_surface_destroy(old_surf
);
111 pipe_texture_reference(struct pipe_texture
**ptr
, struct pipe_texture
*tex
)
113 struct pipe_texture
*old_tex
= *ptr
;
115 if (pipe_reference(&(*ptr
)->reference
, &tex
->reference
))
116 old_tex
->screen
->texture_destroy(old_tex
);
121 pipe_sampler_view_reference(struct pipe_sampler_view
**ptr
, struct pipe_sampler_view
*view
)
123 struct pipe_sampler_view
*old_view
= *ptr
;
125 if (pipe_reference(&(*ptr
)->reference
, &view
->reference
))
126 old_view
->context
->sampler_view_destroy(old_view
->context
, old_view
);
132 * Convenience wrappers for screen buffer functions.
135 static INLINE
struct pipe_buffer
*
136 pipe_buffer_create( struct pipe_screen
*screen
,
137 unsigned alignment
, unsigned usage
, unsigned size
)
139 return screen
->buffer_create(screen
, alignment
, usage
, size
);
142 static INLINE
struct pipe_buffer
*
143 pipe_user_buffer_create( struct pipe_screen
*screen
, void *ptr
, unsigned size
)
145 return screen
->user_buffer_create(screen
, ptr
, size
);
149 pipe_buffer_map(struct pipe_screen
*screen
,
150 struct pipe_buffer
*buf
,
153 if(screen
->buffer_map_range
) {
155 unsigned length
= buf
->size
;
156 return screen
->buffer_map_range(screen
, buf
, offset
, length
, usage
);
159 return screen
->buffer_map(screen
, buf
, usage
);
163 pipe_buffer_unmap(struct pipe_screen
*screen
,
164 struct pipe_buffer
*buf
)
166 screen
->buffer_unmap(screen
, buf
);
170 pipe_buffer_map_range(struct pipe_screen
*screen
,
171 struct pipe_buffer
*buf
,
176 assert(offset
< buf
->size
);
177 assert(offset
+ length
<= buf
->size
);
179 if(screen
->buffer_map_range
)
180 return screen
->buffer_map_range(screen
, buf
, offset
, length
, usage
);
182 return screen
->buffer_map(screen
, buf
, usage
);
186 pipe_buffer_flush_mapped_range(struct pipe_screen
*screen
,
187 struct pipe_buffer
*buf
,
191 assert(offset
< buf
->size
);
192 assert(offset
+ length
<= buf
->size
);
194 if(screen
->buffer_flush_mapped_range
)
195 screen
->buffer_flush_mapped_range(screen
, buf
, offset
, length
);
199 pipe_buffer_write(struct pipe_screen
*screen
,
200 struct pipe_buffer
*buf
,
201 unsigned offset
, unsigned size
,
206 assert(offset
< buf
->size
);
207 assert(offset
+ size
<= buf
->size
);
210 map
= pipe_buffer_map_range(screen
, buf
, offset
, size
,
211 PIPE_BUFFER_USAGE_CPU_WRITE
|
212 PIPE_BUFFER_USAGE_FLUSH_EXPLICIT
|
213 PIPE_BUFFER_USAGE_DISCARD
);
216 memcpy((uint8_t *)map
+ offset
, data
, size
);
217 pipe_buffer_flush_mapped_range(screen
, buf
, offset
, size
);
218 pipe_buffer_unmap(screen
, buf
);
223 * Special case for writing non-overlapping ranges.
225 * We can avoid GPU/CPU synchronization when writing range that has never
226 * been written before.
229 pipe_buffer_write_nooverlap(struct pipe_screen
*screen
,
230 struct pipe_buffer
*buf
,
231 unsigned offset
, unsigned size
,
236 assert(offset
< buf
->size
);
237 assert(offset
+ size
<= buf
->size
);
240 map
= pipe_buffer_map_range(screen
, buf
, offset
, size
,
241 PIPE_BUFFER_USAGE_CPU_WRITE
|
242 PIPE_BUFFER_USAGE_FLUSH_EXPLICIT
|
243 PIPE_BUFFER_USAGE_DISCARD
|
244 PIPE_BUFFER_USAGE_UNSYNCHRONIZED
);
247 memcpy((uint8_t *)map
+ offset
, data
, size
);
248 pipe_buffer_flush_mapped_range(screen
, buf
, offset
, size
);
249 pipe_buffer_unmap(screen
, buf
);
254 pipe_buffer_read(struct pipe_screen
*screen
,
255 struct pipe_buffer
*buf
,
256 unsigned offset
, unsigned size
,
261 assert(offset
< buf
->size
);
262 assert(offset
+ size
<= buf
->size
);
265 map
= pipe_buffer_map_range(screen
, buf
, offset
, size
, PIPE_BUFFER_USAGE_CPU_READ
);
268 memcpy(data
, (const uint8_t *)map
+ offset
, size
);
269 pipe_buffer_unmap(screen
, buf
);
274 pipe_transfer_map( struct pipe_transfer
*transf
)
276 struct pipe_screen
*screen
= transf
->texture
->screen
;
277 return screen
->transfer_map(screen
, transf
);
281 pipe_transfer_unmap( struct pipe_transfer
*transf
)
283 struct pipe_screen
*screen
= transf
->texture
->screen
;
284 screen
->transfer_unmap(screen
, transf
);
288 pipe_transfer_destroy( struct pipe_transfer
*transf
)
290 struct pipe_screen
*screen
= transf
->texture
->screen
;
291 screen
->tex_transfer_destroy(transf
);
294 static INLINE
unsigned
295 pipe_transfer_buffer_flags( struct pipe_transfer
*transf
)
297 switch (transf
->usage
& PIPE_TRANSFER_READ_WRITE
) {
298 case PIPE_TRANSFER_READ_WRITE
:
299 return PIPE_BUFFER_USAGE_CPU_READ
| PIPE_BUFFER_USAGE_CPU_WRITE
;
300 case PIPE_TRANSFER_READ
:
301 return PIPE_BUFFER_USAGE_CPU_READ
;
302 case PIPE_TRANSFER_WRITE
:
303 return PIPE_BUFFER_USAGE_CPU_WRITE
;
314 #endif /* U_INLINES_H */