X-Git-Url: https://git.libre-soc.org/?a=blobdiff_plain;f=src%2Fgallium%2Fdrivers%2Fr600%2Fr600_buffer.c;h=88281b07106776cb8448862c2233043a58593808;hb=d5b23dfc1c07f98afe749053b9cb4b69829fe3d4;hp=28d8c6af1cb75c01084e75821867b41a63785215;hpb=2ce783d8ddec1b1fcadc0798af0ebb045bba1cc4;p=mesa.git diff --git a/src/gallium/drivers/r600/r600_buffer.c b/src/gallium/drivers/r600/r600_buffer.c index 28d8c6af1cb..88281b07106 100644 --- a/src/gallium/drivers/r600/r600_buffer.c +++ b/src/gallium/drivers/r600/r600_buffer.c @@ -24,269 +24,289 @@ * Jerome Glisse * Corbin Simpson */ -#include - -#include -#include -#include -#include -#include -#include "util/u_upload_mgr.h" - -#include -#include "radeon_drm.h" - -#include "r600.h" #include "r600_pipe.h" +#include "util/u_upload_mgr.h" +#include "util/u_memory.h" +#include "util/u_surface.h" static void r600_buffer_destroy(struct pipe_screen *screen, struct pipe_resource *buf) { - struct r600_screen *rscreen = (struct r600_screen*)screen; - struct r600_resource_buffer *rbuffer = r600_buffer(buf); + struct r600_resource *rbuffer = r600_resource(buf); + + util_range_destroy(&rbuffer->valid_buffer_range); + pb_reference(&rbuffer->buf, NULL); + FREE(rbuffer); +} - if (rbuffer->r.bo) { - r600_bo_reference(rscreen->radeon, &rbuffer->r.bo, NULL); +static void r600_set_constants_dirty_if_bound(struct r600_context *rctx, + struct r600_resource *rbuffer) +{ + unsigned shader; + + for (shader = 0; shader < PIPE_SHADER_TYPES; shader++) { + struct r600_constbuf_state *state = &rctx->constbuf_state[shader]; + bool found = false; + uint32_t mask = state->enabled_mask; + + while (mask) { + unsigned i = u_bit_scan(&mask); + if (state->cb[i].buffer == &rbuffer->b.b) { + found = true; + state->dirty_mask |= 1 << i; + } + } + if (found) { + r600_constant_buffers_dirty(rctx, state); + } } - rbuffer->r.bo = NULL; - util_slab_free(&rscreen->pool_buffers, rbuffer); } -static struct pipe_transfer *r600_get_transfer(struct pipe_context *ctx, - struct pipe_resource *resource, - unsigned level, - unsigned usage, - const struct pipe_box *box) +static void *r600_buffer_get_transfer(struct pipe_context *ctx, + struct pipe_resource *resource, + unsigned level, + unsigned usage, + const struct pipe_box *box, + struct pipe_transfer **ptransfer, + void *data, struct r600_resource *staging, + unsigned offset) { - struct r600_pipe_context *rctx = (struct r600_pipe_context*)ctx; - struct pipe_transfer *transfer = util_slab_alloc(&rctx->pool_transfers); - - transfer->resource = resource; - transfer->level = level; - transfer->usage = usage; - transfer->box = *box; - transfer->stride = 0; - transfer->layer_stride = 0; - transfer->data = NULL; - - /* Note strides are zero, this is ok for buffers, but not for - * textures 2d & higher at least. - */ - return transfer; + struct r600_context *rctx = (struct r600_context*)ctx; + struct r600_transfer *transfer = util_slab_alloc(&rctx->pool_transfers); + + transfer->transfer.resource = resource; + transfer->transfer.level = level; + transfer->transfer.usage = usage; + transfer->transfer.box = *box; + transfer->transfer.stride = 0; + transfer->transfer.layer_stride = 0; + transfer->offset = offset; + transfer->staging = staging; + *ptransfer = &transfer->transfer; + return data; } -static void *r600_buffer_transfer_map(struct pipe_context *pipe, - struct pipe_transfer *transfer) +static void *r600_buffer_transfer_map(struct pipe_context *ctx, + struct pipe_resource *resource, + unsigned level, + unsigned usage, + const struct pipe_box *box, + struct pipe_transfer **ptransfer) { - struct r600_resource_buffer *rbuffer = r600_buffer(transfer->resource); - struct r600_pipe_context *rctx = (struct r600_pipe_context*)pipe; + struct r600_context *rctx = (struct r600_context*)ctx; + struct r600_resource *rbuffer = r600_resource(resource); uint8_t *data; - if (rbuffer->r.b.user_ptr) - return (uint8_t*)rbuffer->r.b.user_ptr + transfer->box.x; + assert(box->x + box->width <= resource->width0); + + /* See if the buffer range being mapped has never been initialized, + * in which case it can be mapped unsynchronized. */ + if (!(usage & PIPE_TRANSFER_UNSYNCHRONIZED) && + usage & PIPE_TRANSFER_WRITE && + !util_ranges_intersect(&rbuffer->valid_buffer_range, box->x, box->x + box->width)) { + usage |= PIPE_TRANSFER_UNSYNCHRONIZED; + } + + if (usage & PIPE_TRANSFER_DISCARD_WHOLE_RESOURCE && + !(usage & PIPE_TRANSFER_UNSYNCHRONIZED)) { + assert(usage & PIPE_TRANSFER_WRITE); + + /* Check if mapping this buffer would cause waiting for the GPU. */ + if (r600_rings_is_buffer_referenced(rctx, rbuffer->cs_buf, RADEON_USAGE_READWRITE) || + rctx->b.ws->buffer_is_busy(rbuffer->buf, RADEON_USAGE_READWRITE)) { + unsigned i, mask; + + /* Discard the buffer. */ + pb_reference(&rbuffer->buf, NULL); + + /* Create a new one in the same pipe_resource. */ + /* XXX We probably want a different alignment for buffers and textures. */ + r600_init_resource(rctx->screen, rbuffer, rbuffer->b.b.width0, 4096, + TRUE, rbuffer->b.b.usage); + + /* We changed the buffer, now we need to bind it where the old one was bound. */ + /* Vertex buffers. */ + mask = rctx->vertex_buffer_state.enabled_mask; + while (mask) { + i = u_bit_scan(&mask); + if (rctx->vertex_buffer_state.vb[i].buffer == &rbuffer->b.b) { + rctx->vertex_buffer_state.dirty_mask |= 1 << i; + r600_vertex_buffers_dirty(rctx); + } + } + /* Streamout buffers. */ + for (i = 0; i < rctx->b.streamout.num_targets; i++) { + if (rctx->b.streamout.targets[i]->b.buffer == &rbuffer->b.b) { + if (rctx->b.streamout.begin_emitted) { + r600_emit_streamout_end(&rctx->b); + } + rctx->b.streamout.append_bitmask = rctx->b.streamout.enabled_mask; + r600_streamout_buffers_dirty(&rctx->b); + } + } + /* Constant buffers. */ + r600_set_constants_dirty_if_bound(rctx, rbuffer); + } + } + else if ((usage & PIPE_TRANSFER_DISCARD_RANGE) && + !(usage & PIPE_TRANSFER_UNSYNCHRONIZED) && + !(rctx->screen->debug_flags & DBG_NO_DISCARD_RANGE) && + (rctx->screen->has_cp_dma || + (rctx->screen->has_streamout && + /* The buffer range must be aligned to 4 with streamout. */ + box->x % 4 == 0 && box->width % 4 == 0))) { + assert(usage & PIPE_TRANSFER_WRITE); + + /* Check if mapping this buffer would cause waiting for the GPU. */ + if (r600_rings_is_buffer_referenced(rctx, rbuffer->cs_buf, RADEON_USAGE_READWRITE) || + rctx->b.ws->buffer_is_busy(rbuffer->buf, RADEON_USAGE_READWRITE)) { + /* Do a wait-free write-only transfer using a temporary buffer. */ + unsigned offset; + struct r600_resource *staging = NULL; + + u_upload_alloc(rctx->uploader, 0, box->width + (box->x % R600_MAP_BUFFER_ALIGNMENT), + &offset, (struct pipe_resource**)&staging, (void**)&data); + + if (staging) { + data += box->x % R600_MAP_BUFFER_ALIGNMENT; + return r600_buffer_get_transfer(ctx, resource, level, usage, box, + ptransfer, data, staging, offset); + } + } + } - data = r600_bo_map(rctx->screen->radeon, rbuffer->r.bo, transfer->usage, pipe); - if (!data) + /* mmap and synchronize with rings */ + data = r600_buffer_mmap_sync_with_rings(rctx, rbuffer, usage); + if (!data) { return NULL; + } + data += box->x; - return (uint8_t*)data + transfer->box.x; + return r600_buffer_get_transfer(ctx, resource, level, usage, box, + ptransfer, data, NULL, 0); } static void r600_buffer_transfer_unmap(struct pipe_context *pipe, struct pipe_transfer *transfer) { - struct r600_resource_buffer *rbuffer = r600_buffer(transfer->resource); - struct r600_pipe_context *rctx = (struct r600_pipe_context*)pipe; - - if (rbuffer->r.b.user_ptr) - return; - - if (rbuffer->r.bo) - r600_bo_unmap(rctx->screen->radeon, rbuffer->r.bo); -} + struct r600_context *rctx = (struct r600_context*)pipe; + struct r600_transfer *rtransfer = (struct r600_transfer*)transfer; + struct r600_resource *rbuffer = r600_resource(transfer->resource); + + if (rtransfer->staging) { + struct pipe_resource *dst, *src; + unsigned soffset, doffset, size; + + dst = transfer->resource; + src = &rtransfer->staging->b.b; + size = transfer->box.width; + doffset = transfer->box.x; + soffset = rtransfer->offset + transfer->box.x % R600_MAP_BUFFER_ALIGNMENT; + /* Copy the staging buffer into the original one. */ + if (rctx->b.rings.dma.cs && !(size % 4) && !(doffset % 4) && !(soffset % 4)) { + if (rctx->screen->b.chip_class >= EVERGREEN) { + evergreen_dma_copy(rctx, dst, src, doffset, soffset, size); + } else { + r600_dma_copy(rctx, dst, src, doffset, soffset, size); + } + } else { + struct pipe_box box; -static void r600_buffer_transfer_flush_region(struct pipe_context *pipe, - struct pipe_transfer *transfer, - const struct pipe_box *box) -{ -} + u_box_1d(soffset, size, &box); + r600_copy_buffer(pipe, dst, doffset, src, &box); + } + pipe_resource_reference((struct pipe_resource**)&rtransfer->staging, NULL); + } -static void r600_transfer_destroy(struct pipe_context *ctx, - struct pipe_transfer *transfer) -{ - struct r600_pipe_context *rctx = (struct r600_pipe_context*)ctx; + if (transfer->usage & PIPE_TRANSFER_WRITE) { + util_range_add(&rbuffer->valid_buffer_range, transfer->box.x, + transfer->box.x + transfer->box.width); + } util_slab_free(&rctx->pool_transfers, transfer); } -static void r600_buffer_transfer_inline_write(struct pipe_context *pipe, - struct pipe_resource *resource, - unsigned level, - unsigned usage, - const struct pipe_box *box, - const void *data, - unsigned stride, - unsigned layer_stride) -{ - struct r600_pipe_context *rctx = (struct r600_pipe_context*)pipe; - struct radeon *radeon = rctx->screen->radeon; - struct r600_resource_buffer *rbuffer = r600_buffer(resource); - uint8_t *map = NULL; - - assert(rbuffer->r.b.user_ptr == NULL); - - map = r600_bo_map(radeon, rbuffer->r.bo, - PIPE_TRANSFER_WRITE | PIPE_TRANSFER_DISCARD | usage, - pipe); - - memcpy(map + box->x, data, box->width); - - if (rbuffer->r.bo) - r600_bo_unmap(radeon, rbuffer->r.bo); -} - static const struct u_resource_vtbl r600_buffer_vtbl = { u_default_resource_get_handle, /* get_handle */ r600_buffer_destroy, /* resource_destroy */ - r600_get_transfer, /* get_transfer */ - r600_transfer_destroy, /* transfer_destroy */ r600_buffer_transfer_map, /* transfer_map */ - r600_buffer_transfer_flush_region, /* transfer_flush_region */ + NULL, /* transfer_flush_region */ r600_buffer_transfer_unmap, /* transfer_unmap */ - r600_buffer_transfer_inline_write /* transfer_inline_write */ + NULL /* transfer_inline_write */ }; -struct pipe_resource *r600_buffer_create(struct pipe_screen *screen, - const struct pipe_resource *templ) -{ - struct r600_screen *rscreen = (struct r600_screen*)screen; - struct r600_resource_buffer *rbuffer; - struct r600_bo *bo; - /* XXX We probably want a different alignment for buffers and textures. */ - unsigned alignment = 4096; - - rbuffer = util_slab_alloc(&rscreen->pool_buffers); - - rbuffer->magic = R600_BUFFER_MAGIC; - rbuffer->r.b.b.b = *templ; - pipe_reference_init(&rbuffer->r.b.b.b.reference, 1); - rbuffer->r.b.b.b.screen = screen; - rbuffer->r.b.b.vtbl = &r600_buffer_vtbl; - rbuffer->r.b.user_ptr = NULL; - rbuffer->r.size = rbuffer->r.b.b.b.width0; - rbuffer->r.bo_size = rbuffer->r.size; - - bo = r600_bo(rscreen->radeon, - rbuffer->r.b.b.b.width0, - alignment, rbuffer->r.b.b.b.bind, - rbuffer->r.b.b.b.usage); - - if (bo == NULL) { - FREE(rbuffer); - return NULL; - } - rbuffer->r.bo = bo; - return &rbuffer->r.b.b.b; -} - -struct pipe_resource *r600_user_buffer_create(struct pipe_screen *screen, - void *ptr, unsigned bytes, - unsigned bind) +bool r600_init_resource(struct r600_screen *rscreen, + struct r600_resource *res, + unsigned size, unsigned alignment, + bool use_reusable_pool, unsigned usage) { - struct r600_screen *rscreen = (struct r600_screen*)screen; - struct r600_resource_buffer *rbuffer; - - rbuffer = util_slab_alloc(&rscreen->pool_buffers); - - rbuffer->magic = R600_BUFFER_MAGIC; - pipe_reference_init(&rbuffer->r.b.b.b.reference, 1); - rbuffer->r.b.b.vtbl = &r600_buffer_vtbl; - rbuffer->r.b.b.b.screen = screen; - rbuffer->r.b.b.b.target = PIPE_BUFFER; - rbuffer->r.b.b.b.format = PIPE_FORMAT_R8_UNORM; - rbuffer->r.b.b.b.usage = PIPE_USAGE_IMMUTABLE; - rbuffer->r.b.b.b.bind = bind; - rbuffer->r.b.b.b.width0 = bytes; - rbuffer->r.b.b.b.height0 = 1; - rbuffer->r.b.b.b.depth0 = 1; - rbuffer->r.b.b.b.array_size = 1; - rbuffer->r.b.b.b.flags = 0; - rbuffer->r.b.user_ptr = ptr; - rbuffer->r.bo = NULL; - rbuffer->r.bo_size = 0; - return &rbuffer->r.b.b.b; -} - -struct pipe_resource *r600_buffer_from_handle(struct pipe_screen *screen, - struct winsys_handle *whandle) -{ - struct radeon *rw = ((struct r600_screen*)screen)->radeon; - struct r600_resource *rbuffer; - struct r600_bo *bo = NULL; - - bo = r600_bo_handle(rw, whandle, NULL, NULL); - if (bo == NULL) { - return NULL; + uint32_t initial_domain, domains; + + switch(usage) { + case PIPE_USAGE_STAGING: + /* Staging resources participate in transfers, i.e. are used + * for uploads and downloads from regular resources. + * We generate them internally for some transfers. + */ + initial_domain = RADEON_DOMAIN_GTT; + domains = RADEON_DOMAIN_GTT; + break; + case PIPE_USAGE_DYNAMIC: + case PIPE_USAGE_STREAM: + /* Default to GTT, but allow the memory manager to move it to VRAM. */ + initial_domain = RADEON_DOMAIN_GTT; + domains = RADEON_DOMAIN_GTT | RADEON_DOMAIN_VRAM; + break; + case PIPE_USAGE_DEFAULT: + case PIPE_USAGE_STATIC: + case PIPE_USAGE_IMMUTABLE: + default: + /* Don't list GTT here, because the memory manager would put some + * resources to GTT no matter what the initial domain is. + * Not listing GTT in the domains improves performance a lot. */ + initial_domain = RADEON_DOMAIN_VRAM; + domains = RADEON_DOMAIN_VRAM; + break; } - rbuffer = CALLOC_STRUCT(r600_resource); - if (rbuffer == NULL) { - r600_bo_reference(rw, &bo, NULL); - return NULL; + res->buf = rscreen->b.ws->buffer_create(rscreen->b.ws, size, alignment, + use_reusable_pool, + initial_domain); + if (!res->buf) { + return false; } - pipe_reference_init(&rbuffer->b.b.b.reference, 1); - rbuffer->b.b.b.target = PIPE_BUFFER; - rbuffer->b.b.b.screen = screen; - rbuffer->b.b.vtbl = &r600_buffer_vtbl; - rbuffer->bo = bo; - return &rbuffer->b.b.b; -} + res->cs_buf = rscreen->b.ws->buffer_get_cs_handle(res->buf); + res->domains = domains; + util_range_set_empty(&res->valid_buffer_range); -void r600_upload_index_buffer(struct r600_pipe_context *rctx, struct r600_drawl *draw) -{ - struct r600_resource_buffer *rbuffer = r600_buffer(draw->index_buffer); - boolean flushed; - - u_upload_data(rctx->vbuf_mgr->uploader, 0, - draw->info.count * draw->index_size, - rbuffer->r.b.user_ptr, - &draw->index_buffer_offset, - &draw->index_buffer, &flushed); + if (rscreen->debug_flags & DBG_VM && res->b.b.target == PIPE_BUFFER) { + fprintf(stderr, "VM start=0x%llX end=0x%llX | Buffer %u bytes\n", + r600_resource_va(&rscreen->b.b, &res->b.b), + r600_resource_va(&rscreen->b.b, &res->b.b) + res->buf->size, + res->buf->size); + } + return true; } -void r600_upload_const_buffer(struct r600_pipe_context *rctx, struct r600_resource_buffer **rbuffer, - uint32_t *const_offset) +struct pipe_resource *r600_buffer_create(struct pipe_screen *screen, + const struct pipe_resource *templ, + unsigned alignment) { - if ((*rbuffer)->r.b.user_ptr) { - uint8_t *ptr = (*rbuffer)->r.b.user_ptr; - unsigned size = (*rbuffer)->r.b.b.b.width0; - boolean flushed; - - *rbuffer = NULL; - - if (R600_BIG_ENDIAN) { - uint32_t *tmpPtr; - unsigned i; - - if (!(tmpPtr = malloc(size))) { - R600_ERR("Failed to allocate BE swap buffer.\n"); - return; - } + struct r600_screen *rscreen = (struct r600_screen*)screen; + struct r600_resource *rbuffer; - for (i = 0; i < size / 4; ++i) { - tmpPtr[i] = bswap_32(((uint32_t *)ptr)[i]); - } + rbuffer = MALLOC_STRUCT(r600_resource); - u_upload_data(rctx->vbuf_mgr->uploader, 0, size, tmpPtr, const_offset, - (struct pipe_resource**)rbuffer, &flushed); + rbuffer->b.b = *templ; + pipe_reference_init(&rbuffer->b.b.reference, 1); + rbuffer->b.b.screen = screen; + rbuffer->b.vtbl = &r600_buffer_vtbl; + util_range_init(&rbuffer->valid_buffer_range); - free(tmpPtr); - } else { - u_upload_data(rctx->vbuf_mgr->uploader, 0, size, ptr, const_offset, - (struct pipe_resource**)rbuffer, &flushed); - } - } else { - *const_offset = 0; + if (!r600_init_resource(rscreen, rbuffer, templ->width0, alignment, TRUE, templ->usage)) { + FREE(rbuffer); + return NULL; } + return &rbuffer->b.b; }