X-Git-Url: https://git.libre-soc.org/?a=blobdiff_plain;f=src%2Fgallium%2Fdrivers%2Fsvga%2Fsvga_resource_buffer.c;h=6629a8cc14d96083cdb2d9a0b1a8b7d3f797a758;hb=8c9b9aac7d09e65195dca6681d59c10e4ef713d9;hp=3aa6644a4a3bea7bad506b14f1b3bf01521f19bc;hpb=2aa783318dd40d9ba82b541c1af144f3103b0888;p=mesa.git diff --git a/src/gallium/drivers/svga/svga_resource_buffer.c b/src/gallium/drivers/svga/svga_resource_buffer.c index 3aa6644a4a3..6629a8cc14d 100644 --- a/src/gallium/drivers/svga/svga_resource_buffer.c +++ b/src/gallium/drivers/svga/svga_resource_buffer.c @@ -31,6 +31,7 @@ #include "os/os_thread.h" #include "util/u_math.h" #include "util/u_memory.h" +#include "util/u_resource.h" #include "svga_context.h" #include "svga_screen.h" @@ -41,15 +42,40 @@ /** - * Vertex and index buffers need hardware backing. Constant buffers - * do not. No other types of buffers currently supported. + * Determine what buffers eventually need hardware backing. + * + * Vertex- and index buffers need hardware backing. Constant buffers + * do on vgpu10. Staging texture-upload buffers do when they are + * supported. */ -static INLINE boolean -svga_buffer_needs_hw_storage(unsigned usage) +static inline boolean +svga_buffer_needs_hw_storage(const struct svga_screen *ss, + const struct pipe_resource *template) { - return usage & (PIPE_BIND_VERTEX_BUFFER | PIPE_BIND_INDEX_BUFFER); -} + unsigned bind_mask = (PIPE_BIND_VERTEX_BUFFER | PIPE_BIND_INDEX_BUFFER | + PIPE_BIND_SAMPLER_VIEW | PIPE_BIND_STREAM_OUTPUT); + + if (ss->sws->have_vgpu10) { + /* + * Driver-created upload const0- and staging texture upload buffers + * tagged with PIPE_BIND_CUSTOM + */ + bind_mask |= PIPE_BIND_CUSTOM; + /* Uniform buffer objects. + * Make sure we don't create hardware storage for gallium frontend + * const0 buffers, because we frequently map them for reading. + * They are distinguished by having PIPE_USAGE_STREAM, but not + * PIPE_BIND_CUSTOM. + */ + if (template->usage != PIPE_USAGE_STREAM) + bind_mask |= PIPE_BIND_CONSTANT_BUFFER; + } + if (template->flags & PIPE_RESOURCE_FLAG_MAP_PERSISTENT) + return TRUE; + + return !!(template->bind & bind_mask); +} /** * Create a buffer transfer. @@ -74,20 +100,72 @@ svga_buffer_transfer_map(struct pipe_context *pipe, struct svga_screen *ss = svga_screen(pipe->screen); struct svga_buffer *sbuf = svga_buffer(resource); struct pipe_transfer *transfer; - uint8_t *map; + uint8_t *map = NULL; + int64_t begin = svga_get_time(svga); - transfer = CALLOC_STRUCT(pipe_transfer); - if (transfer == NULL) { - return NULL; + SVGA_STATS_TIME_PUSH(svga_sws(svga), SVGA_STATS_TIME_BUFFERTRANSFERMAP); + + assert(box->y == 0); + assert(box->z == 0); + assert(box->height == 1); + assert(box->depth == 1); + + transfer = MALLOC_STRUCT(pipe_transfer); + if (!transfer) { + goto done; } transfer->resource = resource; transfer->level = level; transfer->usage = usage; transfer->box = *box; + transfer->stride = 0; + transfer->layer_stride = 0; if (usage & PIPE_TRANSFER_WRITE) { - if (usage & PIPE_TRANSFER_DISCARD_WHOLE_RESOURCE) { + /* If we write to the buffer for any reason, free any saved translated + * vertices. + */ + pipe_resource_reference(&sbuf->translated_indices.buffer, NULL); + } + + if ((usage & PIPE_TRANSFER_READ) && sbuf->dirty && + !sbuf->key.coherent && !svga->swc->force_coherent) { + enum pipe_error ret; + + /* Host-side buffers can only be dirtied with vgpu10 features + * (streamout and buffer copy). + */ + assert(svga_have_vgpu10(svga)); + + if (!sbuf->user) { + (void) svga_buffer_handle(svga, resource, sbuf->bind_flags); + } + + if (sbuf->dma.pending) { + svga_buffer_upload_flush(svga, sbuf); + svga_context_finish(svga); + } + + assert(sbuf->handle); + + ret = SVGA3D_vgpu10_ReadbackSubResource(svga->swc, sbuf->handle, 0); + if (ret != PIPE_OK) { + svga_context_flush(svga, NULL); + ret = SVGA3D_vgpu10_ReadbackSubResource(svga->swc, sbuf->handle, 0); + assert(ret == PIPE_OK); + } + + svga->hud.num_readbacks++; + + svga_context_finish(svga); + + sbuf->dirty = FALSE; + } + + if (usage & PIPE_TRANSFER_WRITE) { + if ((usage & PIPE_TRANSFER_DISCARD_WHOLE_RESOURCE) && + !(resource->flags & PIPE_RESOURCE_FLAG_MAP_PERSISTENT)) { /* * Flush any pending primitives, finish writing any pending DMA * commands, and tell the host to discard the buffer contents on @@ -102,9 +180,13 @@ svga_buffer_transfer_map(struct pipe_context *pipe, /* * Instead of flushing the context command buffer, simply discard * the current hwbuf, and start a new one. + * With GB objects, the map operation takes care of this + * if passed the PIPE_TRANSFER_DISCARD_WHOLE_RESOURCE flag, + * and the old backing store is busy. */ - svga_buffer_destroy_hw_storage(ss, sbuf); + if (!svga_have_gb_objects(svga)) + svga_buffer_destroy_hw_storage(ss, sbuf); } sbuf->map.num_ranges = 0; @@ -131,11 +213,11 @@ svga_buffer_transfer_map(struct pipe_context *pipe, if (sbuf->dma.pending) { svga_buffer_upload_flush(svga, sbuf); - if (sbuf->hwbuf) { + if (svga_buffer_has_hw_storage(sbuf)) { /* * We have a pending DMA upload from a hardware buffer, therefore * we need to ensure that the host finishes processing that DMA - * command before the state tracker can start overwriting the + * command before the gallium frontend can start overwriting the * hardware buffer. * * XXX: This could be avoided by tying the hardware buffer to @@ -156,7 +238,7 @@ svga_buffer_transfer_map(struct pipe_context *pipe, */ FREE(transfer); - return NULL; + goto done; } svga_context_flush(svga, NULL); @@ -167,8 +249,8 @@ svga_buffer_transfer_map(struct pipe_context *pipe, } } - if (!sbuf->swbuf && !sbuf->hwbuf) { - if (svga_buffer_create_hw_storage(ss, sbuf) != PIPE_OK) { + if (!sbuf->swbuf && !svga_buffer_has_hw_storage(sbuf)) { + if (svga_buffer_create_hw_storage(ss, sbuf, sbuf->bind_flags) != PIPE_OK) { /* * We can't create a hardware buffer big enough, so create a malloc * buffer instead. @@ -183,7 +265,7 @@ svga_buffer_transfer_map(struct pipe_context *pipe, sbuf->swbuf = align_malloc(sbuf->b.b.width0, 16); if (!sbuf->swbuf) { FREE(transfer); - return NULL; + goto done; } } } @@ -192,11 +274,19 @@ svga_buffer_transfer_map(struct pipe_context *pipe, /* User/malloc buffer */ map = sbuf->swbuf; } - else if (sbuf->hwbuf) { - struct svga_screen *ss = svga_screen(pipe->screen); - struct svga_winsys_screen *sws = ss->sws; + else if (svga_buffer_has_hw_storage(sbuf)) { + boolean retry; - map = sws->buffer_map(sws, sbuf->hwbuf, transfer->usage); + map = svga_buffer_hw_storage_map(svga, sbuf, transfer->usage, &retry); + if (map == NULL && retry) { + /* + * At this point, svga_buffer_get_transfer() has already + * hit the DISCARD_WHOLE_RESOURCE path and flushed HWTNL + * for this buffer. + */ + svga_context_flush(svga, NULL); + map = svga_buffer_hw_storage_map(svga, sbuf, transfer->usage, &retry); + } } else { map = NULL; @@ -209,48 +299,59 @@ svga_buffer_transfer_map(struct pipe_context *pipe, } else { FREE(transfer); } - + + svga->hud.map_buffer_time += (svga_get_time(svga) - begin); + +done: + SVGA_STATS_TIME_POP(svga_sws(svga)); return map; } static void -svga_buffer_transfer_flush_region( struct pipe_context *pipe, - struct pipe_transfer *transfer, - const struct pipe_box *box) +svga_buffer_transfer_flush_region(struct pipe_context *pipe, + struct pipe_transfer *transfer, + const struct pipe_box *box) { struct svga_screen *ss = svga_screen(pipe->screen); struct svga_buffer *sbuf = svga_buffer(transfer->resource); - + struct svga_context *svga = svga_context(pipe); unsigned offset = transfer->box.x + box->x; unsigned length = box->width; assert(transfer->usage & PIPE_TRANSFER_WRITE); assert(transfer->usage & PIPE_TRANSFER_FLUSH_EXPLICIT); - pipe_mutex_lock(ss->swc_mutex); - svga_buffer_add_range(sbuf, offset, offset + length); - pipe_mutex_unlock(ss->swc_mutex); + if (!(svga->swc->force_coherent || sbuf->key.coherent) || sbuf->swbuf) { + mtx_lock(&ss->swc_mutex); + svga_buffer_add_range(sbuf, offset, offset + length); + mtx_unlock(&ss->swc_mutex); + } } static void -svga_buffer_transfer_unmap( struct pipe_context *pipe, - struct pipe_transfer *transfer ) +svga_buffer_transfer_unmap(struct pipe_context *pipe, + struct pipe_transfer *transfer) { struct svga_screen *ss = svga_screen(pipe->screen); - struct svga_winsys_screen *sws = ss->sws; + struct svga_context *svga = svga_context(pipe); struct svga_buffer *sbuf = svga_buffer(transfer->resource); - - pipe_mutex_lock(ss->swc_mutex); - + + SVGA_STATS_TIME_PUSH(svga_sws(svga), SVGA_STATS_TIME_BUFFERTRANSFERUNMAP); + + mtx_lock(&ss->swc_mutex); + assert(sbuf->map.count); if (sbuf->map.count) { --sbuf->map.count; } - if (sbuf->hwbuf) { - sws->buffer_unmap(sws, sbuf->hwbuf); + if (svga_buffer_has_hw_storage(sbuf)) { + /* Note: we may wind up flushing here and unmapping other buffers + * which leads to recursively locking ss->swc_mutex. + */ + svga_buffer_hw_storage_unmap(svga, sbuf); } if (transfer->usage & PIPE_TRANSFER_WRITE) { @@ -262,106 +363,165 @@ svga_buffer_transfer_unmap( struct pipe_context *pipe, */ SVGA_DBG(DEBUG_DMA, "flushing the whole buffer\n"); - + sbuf->dma.flags.discard = TRUE; - svga_buffer_add_range(sbuf, 0, sbuf->b.b.width0); + if (!(svga->swc->force_coherent || sbuf->key.coherent) || sbuf->swbuf) + svga_buffer_add_range(sbuf, 0, sbuf->b.b.width0); } } - pipe_mutex_unlock(ss->swc_mutex); + mtx_unlock(&ss->swc_mutex); FREE(transfer); + SVGA_STATS_TIME_POP(svga_sws(svga)); } static void -svga_buffer_destroy( struct pipe_screen *screen, - struct pipe_resource *buf ) +svga_buffer_destroy(struct pipe_screen *screen, + struct pipe_resource *buf) { - struct svga_screen *ss = svga_screen(screen); - struct svga_buffer *sbuf = svga_buffer( buf ); + struct svga_screen *ss = svga_screen(screen); + struct svga_buffer *sbuf = svga_buffer(buf); assert(!p_atomic_read(&buf->reference.count)); - + assert(!sbuf->dma.pending); - if(sbuf->handle) + if (sbuf->handle) svga_buffer_destroy_host_surface(ss, sbuf); - - if(sbuf->uploaded.buffer) + + if (sbuf->uploaded.buffer) pipe_resource_reference(&sbuf->uploaded.buffer, NULL); - if(sbuf->hwbuf) + if (sbuf->hwbuf) svga_buffer_destroy_hw_storage(ss, sbuf); - - if(sbuf->swbuf && !sbuf->user) + + if (sbuf->swbuf && !sbuf->user) align_free(sbuf->swbuf); - + + pipe_resource_reference(&sbuf->translated_indices.buffer, NULL); + + ss->hud.total_resource_bytes -= sbuf->size; + assert(ss->hud.num_resources > 0); + if (ss->hud.num_resources > 0) + ss->hud.num_resources--; + FREE(sbuf); } -struct u_resource_vtbl svga_buffer_vtbl = +struct u_resource_vtbl svga_buffer_vtbl = { u_default_resource_get_handle, /* get_handle */ svga_buffer_destroy, /* resource_destroy */ svga_buffer_transfer_map, /* transfer_map */ svga_buffer_transfer_flush_region, /* transfer_flush_region */ svga_buffer_transfer_unmap, /* transfer_unmap */ - u_default_transfer_inline_write /* transfer_inline_write */ }; struct pipe_resource * svga_buffer_create(struct pipe_screen *screen, - const struct pipe_resource *template) + const struct pipe_resource *template) { struct svga_screen *ss = svga_screen(screen); struct svga_buffer *sbuf; - + unsigned bind_flags; + + SVGA_STATS_TIME_PUSH(ss->sws, SVGA_STATS_TIME_CREATEBUFFER); + sbuf = CALLOC_STRUCT(svga_buffer); - if(!sbuf) + if (!sbuf) goto error1; - + sbuf->b.b = *template; sbuf->b.vtbl = &svga_buffer_vtbl; pipe_reference_init(&sbuf->b.b.reference, 1); sbuf->b.b.screen = screen; + bind_flags = template->bind & ~PIPE_BIND_CUSTOM; + + list_inithead(&sbuf->surfaces); + + if (bind_flags & PIPE_BIND_CONSTANT_BUFFER) { + /* Constant buffers can only have the PIPE_BIND_CONSTANT_BUFFER + * flag set. + */ + if (ss->sws->have_vgpu10) { + bind_flags = PIPE_BIND_CONSTANT_BUFFER; + } + } + + /* Although svga device only requires constant buffer size to be + * in multiples of 16, in order to allow bind_flags promotion, + * we are mandating all buffer size to be in multiples of 16. + */ + sbuf->b.b.width0 = align(sbuf->b.b.width0, 16); + + if (svga_buffer_needs_hw_storage(ss, template)) { + + /* If the buffer is not used for constant buffer, set + * the vertex/index bind flags as well so that the buffer will be + * accepted for those uses. + * Note that the PIPE_BIND_ flags we get from the gallium frontend are + * just a hint about how the buffer may be used. And OpenGL buffer + * object may be used for many different things. + * Also note that we do not unconditionally set the streamout + * bind flag since streamout buffer is an output buffer and + * might have performance implication. + */ + if (!(template->bind & PIPE_BIND_CONSTANT_BUFFER) && + !(template->bind & PIPE_BIND_CUSTOM)) { + /* Not a constant- or staging buffer. + * The buffer may be used for vertex data or indexes. + */ + bind_flags |= (PIPE_BIND_VERTEX_BUFFER | + PIPE_BIND_INDEX_BUFFER); + } - if(svga_buffer_needs_hw_storage(template->bind)) { - if(svga_buffer_create_host_surface(ss, sbuf) != PIPE_OK) + if (svga_buffer_create_host_surface(ss, sbuf, bind_flags) != PIPE_OK) goto error2; } else { - sbuf->swbuf = align_malloc(template->width0, 64); - if(!sbuf->swbuf) + sbuf->swbuf = align_malloc(sbuf->b.b.width0, 64); + if (!sbuf->swbuf) goto error2; } - + debug_reference(&sbuf->b.b.reference, (debug_reference_descriptor)debug_describe_resource, 0); - return &sbuf->b.b; + sbuf->bind_flags = bind_flags; + sbuf->size = util_resource_size(&sbuf->b.b); + ss->hud.total_resource_bytes += sbuf->size; + + ss->hud.num_resources++; + SVGA_STATS_TIME_POP(ss->sws); + + return &sbuf->b.b; error2: FREE(sbuf); error1: + SVGA_STATS_TIME_POP(ss->sws); return NULL; } + struct pipe_resource * svga_user_buffer_create(struct pipe_screen *screen, void *ptr, unsigned bytes, - unsigned bind) + unsigned bind) { struct svga_buffer *sbuf; - + struct svga_screen *ss = svga_screen(screen); + sbuf = CALLOC_STRUCT(svga_buffer); - if(!sbuf) + if (!sbuf) goto no_sbuf; - + pipe_reference_init(&sbuf->b.b.reference, 1); sbuf->b.vtbl = &svga_buffer_vtbl; sbuf->b.b.screen = screen; @@ -373,17 +533,17 @@ svga_user_buffer_create(struct pipe_screen *screen, sbuf->b.b.depth0 = 1; sbuf->b.b.array_size = 1; + sbuf->bind_flags = bind; sbuf->swbuf = ptr; sbuf->user = TRUE; debug_reference(&sbuf->b.b.reference, (debug_reference_descriptor)debug_describe_resource, 0); - - return &sbuf->b.b; + + ss->hud.num_resources++; + + return &sbuf->b.b; no_sbuf: return NULL; } - - -