1 /**************************************************************************
3 * Copyright 2007 VMware, Inc.
6 * Permission is hereby granted, free of charge, to any person obtaining a
7 * copy of this software and associated documentation files (the
8 * "Software"), to deal in the Software without restriction, including
9 * without limitation the rights to use, copy, modify, merge, publish,
10 * distribute, sub license, and/or sell copies of the Software, and to
11 * permit persons to whom the Software is furnished to do so, subject to
12 * the following conditions:
14 * The above copyright notice and this permission notice (including the
15 * next paragraph) shall be included in all copies or substantial portions
18 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
19 * OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
20 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT.
21 * IN NO EVENT SHALL VMWARE AND/OR ITS SUPPLIERS BE LIABLE FOR
22 * ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,
23 * TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
24 * SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
26 **************************************************************************/
30 * Functions for pixel buffer objects and vertex/element buffer objects.
34 #include <inttypes.h> /* for PRId64 macro */
36 #include "main/imports.h"
37 #include "main/mtypes.h"
38 #include "main/arrayobj.h"
39 #include "main/bufferobj.h"
41 #include "st_context.h"
42 #include "st_cb_bufferobjects.h"
45 #include "pipe/p_context.h"
46 #include "pipe/p_defines.h"
47 #include "util/u_inlines.h"
51 * There is some duplication between mesa's bufferobjects and our
52 * bufmgr buffers. Both have an integer handle and a hashtable to
53 * lookup an opaque structure. It would be nice if the handles and
54 * internal structure where somehow shared.
56 static struct gl_buffer_object
*
57 st_bufferobj_alloc(struct gl_context
*ctx
, GLuint name
)
59 struct st_buffer_object
*st_obj
= ST_CALLOC_STRUCT(st_buffer_object
);
64 _mesa_initialize_buffer_object(ctx
, &st_obj
->Base
, name
);
72 * Deallocate/free a vertex/pixel buffer object.
73 * Called via glDeleteBuffersARB().
76 st_bufferobj_free(struct gl_context
*ctx
, struct gl_buffer_object
*obj
)
78 struct st_buffer_object
*st_obj
= st_buffer_object(obj
);
80 assert(obj
->RefCount
== 0);
81 _mesa_buffer_unmap_all_mappings(ctx
, obj
);
84 pipe_resource_reference(&st_obj
->buffer
, NULL
);
86 _mesa_delete_buffer_object(ctx
, obj
);
92 * Replace data in a subrange of buffer object. If the data range
93 * specified by size + offset extends beyond the end of the buffer or
94 * if data is NULL, no copy is performed.
95 * Called via glBufferSubDataARB().
98 st_bufferobj_subdata(struct gl_context
*ctx
,
101 const void * data
, struct gl_buffer_object
*obj
)
103 struct st_buffer_object
*st_obj
= st_buffer_object(obj
);
105 /* we may be called from VBO code, so double-check params here */
108 assert(offset
+ size
<= obj
->Size
);
114 * According to ARB_vertex_buffer_object specification, if data is null,
115 * then the contents of the buffer object's data store is undefined. We just
116 * ignore, and leave it unchanged.
121 if (!st_obj
->buffer
) {
122 /* we probably ran out of memory during buffer allocation */
126 /* Now that transfers are per-context, we don't have to figure out
127 * flushing here. Usually drivers won't need to flush in this case
128 * even if the buffer is currently referenced by hardware - they
129 * just queue the upload as dma rather than mapping the underlying
132 pipe_buffer_write(st_context(ctx
)->pipe
,
139 * Called via glGetBufferSubDataARB().
142 st_bufferobj_get_subdata(struct gl_context
*ctx
,
145 void * data
, struct gl_buffer_object
*obj
)
147 struct st_buffer_object
*st_obj
= st_buffer_object(obj
);
149 /* we may be called from VBO code, so double-check params here */
152 assert(offset
+ size
<= obj
->Size
);
157 if (!st_obj
->buffer
) {
158 /* we probably ran out of memory during buffer allocation */
162 pipe_buffer_read(st_context(ctx
)->pipe
, st_obj
->buffer
,
168 * Allocate space for and store data in a buffer object. Any data that was
169 * previously stored in the buffer object is lost. If data is NULL,
170 * memory will be allocated, but no copy will occur.
171 * Called via ctx->Driver.BufferData().
172 * \return GL_TRUE for success, GL_FALSE if out of memory
175 st_bufferobj_data(struct gl_context
*ctx
,
180 GLbitfield storageFlags
,
181 struct gl_buffer_object
*obj
)
183 struct st_context
*st
= st_context(ctx
);
184 struct pipe_context
*pipe
= st
->pipe
;
185 struct pipe_screen
*screen
= pipe
->screen
;
186 struct st_buffer_object
*st_obj
= st_buffer_object(obj
);
187 unsigned bind
, pipe_usage
, pipe_flags
= 0;
189 if (target
!= GL_EXTERNAL_VIRTUAL_MEMORY_BUFFER_AMD
&&
190 size
&& st_obj
->buffer
&&
191 st_obj
->Base
.Size
== size
&&
192 st_obj
->Base
.Usage
== usage
&&
193 st_obj
->Base
.StorageFlags
== storageFlags
) {
195 /* Just discard the old contents and write new data.
196 * This should be the same as creating a new buffer, but we avoid
197 * a lot of validation in Mesa.
201 u_box_1d(0, size
, &box
);
202 pipe
->transfer_inline_write(pipe
, st_obj
->buffer
, 0,
203 PIPE_TRANSFER_DISCARD_WHOLE_RESOURCE
,
206 } else if (screen
->get_param(screen
, PIPE_CAP_INVALIDATE_BUFFER
)) {
207 pipe
->invalidate_resource(pipe
, st_obj
->buffer
);
212 st_obj
->Base
.Size
= size
;
213 st_obj
->Base
.Usage
= usage
;
214 st_obj
->Base
.StorageFlags
= storageFlags
;
217 case GL_PIXEL_PACK_BUFFER_ARB
:
218 case GL_PIXEL_UNPACK_BUFFER_ARB
:
219 bind
= PIPE_BIND_RENDER_TARGET
| PIPE_BIND_SAMPLER_VIEW
;
221 case GL_ARRAY_BUFFER_ARB
:
222 bind
= PIPE_BIND_VERTEX_BUFFER
;
224 case GL_ELEMENT_ARRAY_BUFFER_ARB
:
225 bind
= PIPE_BIND_INDEX_BUFFER
;
227 case GL_TEXTURE_BUFFER
:
228 bind
= PIPE_BIND_SAMPLER_VIEW
;
230 case GL_TRANSFORM_FEEDBACK_BUFFER
:
231 bind
= PIPE_BIND_STREAM_OUTPUT
;
233 case GL_UNIFORM_BUFFER
:
234 bind
= PIPE_BIND_CONSTANT_BUFFER
;
236 case GL_DRAW_INDIRECT_BUFFER
:
237 case GL_PARAMETER_BUFFER_ARB
:
238 bind
= PIPE_BIND_COMMAND_ARGS_BUFFER
;
240 case GL_ATOMIC_COUNTER_BUFFER
:
241 case GL_SHADER_STORAGE_BUFFER
:
242 bind
= PIPE_BIND_SHADER_BUFFER
;
244 case GL_QUERY_BUFFER
:
245 bind
= PIPE_BIND_QUERY_BUFFER
;
252 if (st_obj
->Base
.Immutable
) {
254 if (storageFlags
& GL_CLIENT_STORAGE_BIT
)
255 pipe_usage
= PIPE_USAGE_STAGING
;
257 pipe_usage
= PIPE_USAGE_DEFAULT
;
265 pipe_usage
= PIPE_USAGE_DEFAULT
;
267 case GL_DYNAMIC_DRAW
:
268 case GL_DYNAMIC_COPY
:
269 pipe_usage
= PIPE_USAGE_DYNAMIC
;
273 /* XXX: Remove this test and fall-through when we have PBO unpacking
274 * acceleration. Right now, PBO unpacking is done by the CPU, so we
275 * have to make sure CPU reads are fast.
277 if (target
!= GL_PIXEL_UNPACK_BUFFER_ARB
) {
278 pipe_usage
= PIPE_USAGE_STREAM
;
283 case GL_DYNAMIC_READ
:
285 pipe_usage
= PIPE_USAGE_STAGING
;
291 if (storageFlags
& GL_MAP_PERSISTENT_BIT
)
292 pipe_flags
|= PIPE_RESOURCE_FLAG_MAP_PERSISTENT
;
293 if (storageFlags
& GL_MAP_COHERENT_BIT
)
294 pipe_flags
|= PIPE_RESOURCE_FLAG_MAP_COHERENT
;
296 pipe_resource_reference( &st_obj
->buffer
, NULL
);
298 if (ST_DEBUG
& DEBUG_BUFFER
) {
299 debug_printf("Create buffer size %" PRId64
" bind 0x%x\n",
300 (int64_t) size
, bind
);
304 struct pipe_resource buffer
;
306 memset(&buffer
, 0, sizeof buffer
);
307 buffer
.target
= PIPE_BUFFER
;
308 buffer
.format
= PIPE_FORMAT_R8_UNORM
; /* want TYPELESS or similar */
310 buffer
.usage
= pipe_usage
;
311 buffer
.flags
= pipe_flags
;
312 buffer
.width0
= size
;
315 buffer
.array_size
= 1;
317 if (target
== GL_EXTERNAL_VIRTUAL_MEMORY_BUFFER_AMD
) {
319 screen
->resource_from_user_memory(screen
, &buffer
, (void*)data
);
322 st_obj
->buffer
= screen
->resource_create(screen
, &buffer
);
324 if (st_obj
->buffer
&& data
)
325 pipe_buffer_write(pipe
, st_obj
->buffer
, 0, size
, data
);
328 if (!st_obj
->buffer
) {
330 st_obj
->Base
.Size
= 0;
335 /* The current buffer may be bound, so we have to revalidate all atoms that
338 /* TODO: Add arrays to usage history */
339 st
->dirty
.st
|= ST_NEW_VERTEX_ARRAYS
;
340 if (st_obj
->Base
.UsageHistory
& USAGE_UNIFORM_BUFFER
)
341 st
->dirty
.st
|= ST_NEW_UNIFORM_BUFFER
;
342 if (st_obj
->Base
.UsageHistory
& USAGE_SHADER_STORAGE_BUFFER
)
343 st
->dirty
.st
|= ST_NEW_STORAGE_BUFFER
;
344 if (st_obj
->Base
.UsageHistory
& USAGE_TEXTURE_BUFFER
)
345 st
->dirty
.st
|= ST_NEW_SAMPLER_VIEWS
| ST_NEW_IMAGE_UNITS
;
346 if (st_obj
->Base
.UsageHistory
& USAGE_ATOMIC_COUNTER_BUFFER
)
347 st
->dirty
.st
|= ST_NEW_ATOMIC_BUFFER
;
354 * Called via glInvalidateBuffer(Sub)Data.
357 st_bufferobj_invalidate(struct gl_context
*ctx
,
358 struct gl_buffer_object
*obj
,
362 struct st_context
*st
= st_context(ctx
);
363 struct pipe_context
*pipe
= st
->pipe
;
364 struct st_buffer_object
*st_obj
= st_buffer_object(obj
);
366 /* We ignore partial invalidates. */
367 if (offset
!= 0 || size
!= obj
->Size
)
370 /* Nothing to invalidate. */
374 pipe
->invalidate_resource(pipe
, st_obj
->buffer
);
379 * Called via glMapBufferRange().
382 st_bufferobj_map_range(struct gl_context
*ctx
,
383 GLintptr offset
, GLsizeiptr length
, GLbitfield access
,
384 struct gl_buffer_object
*obj
,
385 gl_map_buffer_index index
)
387 struct pipe_context
*pipe
= st_context(ctx
)->pipe
;
388 struct st_buffer_object
*st_obj
= st_buffer_object(obj
);
389 enum pipe_transfer_usage flags
= 0x0;
391 if (access
& GL_MAP_WRITE_BIT
)
392 flags
|= PIPE_TRANSFER_WRITE
;
394 if (access
& GL_MAP_READ_BIT
)
395 flags
|= PIPE_TRANSFER_READ
;
397 if (access
& GL_MAP_FLUSH_EXPLICIT_BIT
)
398 flags
|= PIPE_TRANSFER_FLUSH_EXPLICIT
;
400 if (access
& GL_MAP_INVALIDATE_BUFFER_BIT
) {
401 flags
|= PIPE_TRANSFER_DISCARD_WHOLE_RESOURCE
;
403 else if (access
& GL_MAP_INVALIDATE_RANGE_BIT
) {
404 if (offset
== 0 && length
== obj
->Size
)
405 flags
|= PIPE_TRANSFER_DISCARD_WHOLE_RESOURCE
;
407 flags
|= PIPE_TRANSFER_DISCARD_RANGE
;
410 if (access
& GL_MAP_UNSYNCHRONIZED_BIT
)
411 flags
|= PIPE_TRANSFER_UNSYNCHRONIZED
;
413 if (access
& GL_MAP_PERSISTENT_BIT
)
414 flags
|= PIPE_TRANSFER_PERSISTENT
;
416 if (access
& GL_MAP_COHERENT_BIT
)
417 flags
|= PIPE_TRANSFER_COHERENT
;
419 /* ... other flags ...
422 if (access
& MESA_MAP_NOWAIT_BIT
)
423 flags
|= PIPE_TRANSFER_DONTBLOCK
;
427 assert(offset
< obj
->Size
);
428 assert(offset
+ length
<= obj
->Size
);
430 obj
->Mappings
[index
].Pointer
= pipe_buffer_map_range(pipe
,
434 &st_obj
->transfer
[index
]);
435 if (obj
->Mappings
[index
].Pointer
) {
436 obj
->Mappings
[index
].Offset
= offset
;
437 obj
->Mappings
[index
].Length
= length
;
438 obj
->Mappings
[index
].AccessFlags
= access
;
441 st_obj
->transfer
[index
] = NULL
;
444 return obj
->Mappings
[index
].Pointer
;
449 st_bufferobj_flush_mapped_range(struct gl_context
*ctx
,
450 GLintptr offset
, GLsizeiptr length
,
451 struct gl_buffer_object
*obj
,
452 gl_map_buffer_index index
)
454 struct pipe_context
*pipe
= st_context(ctx
)->pipe
;
455 struct st_buffer_object
*st_obj
= st_buffer_object(obj
);
457 /* Subrange is relative to mapped range */
460 assert(offset
+ length
<= obj
->Mappings
[index
].Length
);
461 assert(obj
->Mappings
[index
].Pointer
);
466 pipe_buffer_flush_mapped_range(pipe
, st_obj
->transfer
[index
],
467 obj
->Mappings
[index
].Offset
+ offset
,
473 * Called via glUnmapBufferARB().
476 st_bufferobj_unmap(struct gl_context
*ctx
, struct gl_buffer_object
*obj
,
477 gl_map_buffer_index index
)
479 struct pipe_context
*pipe
= st_context(ctx
)->pipe
;
480 struct st_buffer_object
*st_obj
= st_buffer_object(obj
);
482 if (obj
->Mappings
[index
].Length
)
483 pipe_buffer_unmap(pipe
, st_obj
->transfer
[index
]);
485 st_obj
->transfer
[index
] = NULL
;
486 obj
->Mappings
[index
].Pointer
= NULL
;
487 obj
->Mappings
[index
].Offset
= 0;
488 obj
->Mappings
[index
].Length
= 0;
494 * Called via glCopyBufferSubData().
497 st_copy_buffer_subdata(struct gl_context
*ctx
,
498 struct gl_buffer_object
*src
,
499 struct gl_buffer_object
*dst
,
500 GLintptr readOffset
, GLintptr writeOffset
,
503 struct pipe_context
*pipe
= st_context(ctx
)->pipe
;
504 struct st_buffer_object
*srcObj
= st_buffer_object(src
);
505 struct st_buffer_object
*dstObj
= st_buffer_object(dst
);
511 /* buffer should not already be mapped */
512 assert(!_mesa_check_disallowed_mapping(src
));
513 assert(!_mesa_check_disallowed_mapping(dst
));
515 u_box_1d(readOffset
, size
, &box
);
517 pipe
->resource_copy_region(pipe
, dstObj
->buffer
, 0, writeOffset
, 0, 0,
518 srcObj
->buffer
, 0, &box
);
522 * Called via glClearBufferSubData().
525 st_clear_buffer_subdata(struct gl_context
*ctx
,
526 GLintptr offset
, GLsizeiptr size
,
527 const void *clearValue
,
528 GLsizeiptr clearValueSize
,
529 struct gl_buffer_object
*bufObj
)
531 struct pipe_context
*pipe
= st_context(ctx
)->pipe
;
532 struct st_buffer_object
*buf
= st_buffer_object(bufObj
);
533 static const char zeros
[16] = {0};
535 if (!pipe
->clear_buffer
) {
536 _mesa_ClearBufferSubData_sw(ctx
, offset
, size
,
537 clearValue
, clearValueSize
, bufObj
);
544 pipe
->clear_buffer(pipe
, buf
->buffer
, offset
, size
,
545 clearValue
, clearValueSize
);
549 /* TODO: if buffer wasn't created with appropriate usage flags, need
550 * to recreate it now and copy contents -- or possibly create a
551 * gallium entrypoint to extend the usage flags and let the driver
552 * decide if a copy is necessary.
555 st_bufferobj_validate_usage(struct st_context
*st
,
556 struct st_buffer_object
*obj
,
563 st_init_bufferobject_functions(struct pipe_screen
*screen
,
564 struct dd_function_table
*functions
)
566 /* plug in default driver fallbacks (such as for ClearBufferSubData) */
567 _mesa_init_buffer_object_functions(functions
);
569 functions
->NewBufferObject
= st_bufferobj_alloc
;
570 functions
->DeleteBuffer
= st_bufferobj_free
;
571 functions
->BufferData
= st_bufferobj_data
;
572 functions
->BufferSubData
= st_bufferobj_subdata
;
573 functions
->GetBufferSubData
= st_bufferobj_get_subdata
;
574 functions
->MapBufferRange
= st_bufferobj_map_range
;
575 functions
->FlushMappedBufferRange
= st_bufferobj_flush_mapped_range
;
576 functions
->UnmapBuffer
= st_bufferobj_unmap
;
577 functions
->CopyBufferSubData
= st_copy_buffer_subdata
;
578 functions
->ClearBufferSubData
= st_clear_buffer_subdata
;
580 if (screen
->get_param(screen
, PIPE_CAP_INVALIDATE_BUFFER
))
581 functions
->InvalidateBufferSubData
= st_bufferobj_invalidate
;