nv50,nvc0: implement blit
[mesa.git] / src / gallium / drivers / r600 / r600_buffer.c
index 0ca6ff114b400a763c1b5ba89faa6615fefb45ce..0b0ac3460e1bfb3fec66028a2fa3d7ea9e2fe5d9 100644 (file)
  */
 #include "r600_pipe.h"
 #include "util/u_upload_mgr.h"
+#include "util/u_memory.h"
 
 static void r600_buffer_destroy(struct pipe_screen *screen,
                                struct pipe_resource *buf)
 {
-       struct r600_screen *rscreen = (struct r600_screen*)screen;
        struct r600_resource *rbuffer = r600_resource(buf);
 
        pb_reference(&rbuffer->buf, NULL);
-       util_slab_free(&rscreen->pool_buffers, rbuffer);
+       FREE(rbuffer);
 }
 
 static struct pipe_transfer *r600_get_transfer(struct pipe_context *ctx,
@@ -44,38 +44,46 @@ static struct pipe_transfer *r600_get_transfer(struct pipe_context *ctx,
                                               const struct pipe_box *box)
 {
        struct r600_context *rctx = (struct r600_context*)ctx;
-       struct pipe_transfer *transfer = util_slab_alloc(&rctx->pool_transfers);
+       struct r600_transfer *transfer = util_slab_alloc(&rctx->pool_transfers);
+
+       assert(box->x + box->width <= resource->width0);
 
-       transfer->resource = resource;
-       transfer->level = level;
-       transfer->usage = usage;
-       transfer->box = *box;
-       transfer->stride = 0;
-       transfer->layer_stride = 0;
-       transfer->data = NULL;
+       transfer->transfer.resource = resource;
+       transfer->transfer.level = level;
+       transfer->transfer.usage = usage;
+       transfer->transfer.box = *box;
+       transfer->transfer.stride = 0;
+       transfer->transfer.layer_stride = 0;
+       transfer->transfer.data = NULL;
+       transfer->staging = NULL;
+       transfer->offset = 0;
 
        /* Note strides are zero, this is ok for buffers, but not for
         * textures 2d & higher at least.
         */
-       return transfer;
+       return &transfer->transfer;
 }
 
 static void r600_set_constants_dirty_if_bound(struct r600_context *rctx,
-                                             struct r600_constbuf_state *state,
                                              struct r600_resource *rbuffer)
 {
-       bool found = false;
-       uint32_t mask = state->enabled_mask;
-
-       while (mask) {
-               unsigned i = u_bit_scan(&mask);
-               if (state->cb[i].buffer == &rbuffer->b.b) {
-                       found = true;
-                       state->dirty_mask |= 1 << i;
+       unsigned shader;
+
+       for (shader = 0; shader < PIPE_SHADER_TYPES; shader++) {
+               struct r600_constbuf_state *state = &rctx->constbuf_state[shader];
+               bool found = false;
+               uint32_t mask = state->enabled_mask;
+
+               while (mask) {
+                       unsigned i = u_bit_scan(&mask);
+                       if (state->cb[i].buffer == &rbuffer->b.b) {
+                               found = true;
+                               state->dirty_mask |= 1 << i;
+                       }
+               }
+               if (found) {
+                       r600_constant_buffers_dirty(rctx, state);
                }
-       }
-       if (found) {
-               r600_constant_buffers_dirty(rctx, state);
        }
 }
 
@@ -86,15 +94,14 @@ static void *r600_buffer_transfer_map(struct pipe_context *pipe,
        struct r600_context *rctx = (struct r600_context*)pipe;
        uint8_t *data;
 
-       if (transfer->usage & PIPE_TRANSFER_DISCARD_WHOLE_RESOURCE) {
-               /* When mapping for read, we only need to check if the GPU is writing to it. */
-               enum radeon_bo_usage rusage = transfer->usage & PIPE_TRANSFER_WRITE ?
-                       RADEON_USAGE_READWRITE : RADEON_USAGE_WRITE;
+       if (transfer->usage & PIPE_TRANSFER_DISCARD_WHOLE_RESOURCE &&
+           !(transfer->usage & PIPE_TRANSFER_UNSYNCHRONIZED)) {
+               assert(transfer->usage & PIPE_TRANSFER_WRITE);
 
                /* Check if mapping this buffer would cause waiting for the GPU. */
-               if (rctx->ws->cs_is_buffer_referenced(rctx->cs, rbuffer->cs_buf, rusage) ||
-                   rctx->ws->buffer_is_busy(rbuffer->buf, rusage)) {
-                       unsigned i;
+               if (rctx->ws->cs_is_buffer_referenced(rctx->cs, rbuffer->cs_buf, RADEON_USAGE_READWRITE) ||
+                   rctx->ws->buffer_is_busy(rbuffer->buf, RADEON_USAGE_READWRITE)) {
+                       unsigned i, mask;
 
                        /* Discard the buffer. */
                        pb_reference(&rbuffer->buf, NULL);
@@ -106,10 +113,12 @@ static void *r600_buffer_transfer_map(struct pipe_context *pipe,
 
                        /* We changed the buffer, now we need to bind it where the old one was bound. */
                        /* Vertex buffers. */
-                       for (i = 0; i < rctx->nr_vertex_buffers; i++) {
-                               if (rctx->vertex_buffer[i].buffer == &rbuffer->b.b) {
-                                       r600_inval_vertex_cache(rctx);
-                                       r600_atom_dirty(rctx, &rctx->vertex_buffer_state);
+                       mask = rctx->vertex_buffer_state.enabled_mask;
+                       while (mask) {
+                               i = u_bit_scan(&mask);
+                               if (rctx->vertex_buffer_state.vb[i].buffer == &rbuffer->b.b) {
+                                       rctx->vertex_buffer_state.dirty_mask |= 1 << i;
+                                       r600_vertex_buffers_dirty(rctx);
                                }
                        }
                        /* Streamout buffers. */
@@ -121,10 +130,30 @@ static void *r600_buffer_transfer_map(struct pipe_context *pipe,
                                }
                        }
                        /* Constant buffers. */
-                       r600_set_constants_dirty_if_bound(rctx, &rctx->vs_constbuf_state, rbuffer);
-                       r600_set_constants_dirty_if_bound(rctx, &rctx->ps_constbuf_state, rbuffer);
+                       r600_set_constants_dirty_if_bound(rctx, rbuffer);
                }
        }
+#if 0 /* this is broken (see Bug 53130) */
+       else if ((transfer->usage & PIPE_TRANSFER_DISCARD_RANGE) &&
+                !(transfer->usage & PIPE_TRANSFER_UNSYNCHRONIZED) &&
+                rctx->screen->has_streamout &&
+                /* The buffer range must be aligned to 4. */
+                transfer->box.x % 4 == 0 && transfer->box.width % 4 == 0) {
+               assert(transfer->usage & PIPE_TRANSFER_WRITE);
+
+               /* Check if mapping this buffer would cause waiting for the GPU. */
+               if (rctx->ws->cs_is_buffer_referenced(rctx->cs, rbuffer->cs_buf, RADEON_USAGE_READWRITE) ||
+                   rctx->ws->buffer_is_busy(rbuffer->buf, RADEON_USAGE_READWRITE)) {
+                       /* Do a wait-free write-only transfer using a temporary buffer. */
+                       struct r600_transfer *rtransfer = (struct r600_transfer*)transfer;
+
+                       rtransfer->staging = (struct r600_resource*)
+                               pipe_buffer_create(pipe->screen, PIPE_BIND_VERTEX_BUFFER,
+                                                  PIPE_USAGE_STAGING, transfer->box.width);
+                       return rctx->ws->buffer_map(rtransfer->staging->cs_buf, rctx->cs, PIPE_TRANSFER_WRITE);
+               }
+       }
+#endif
 
        data = rctx->ws->buffer_map(rbuffer->cs_buf, rctx->cs, transfer->usage);
        if (!data)
@@ -136,7 +165,17 @@ static void *r600_buffer_transfer_map(struct pipe_context *pipe,
 static void r600_buffer_transfer_unmap(struct pipe_context *pipe,
                                        struct pipe_transfer *transfer)
 {
-       /* no-op */
+       struct r600_transfer *rtransfer = (struct r600_transfer*)transfer;
+
+       if (rtransfer->staging) {
+               struct pipe_box box;
+               u_box_1d(0, transfer->box.width, &box);
+
+               /* Copy the staging buffer into the original one. */
+               r600_copy_buffer(pipe, transfer->resource, transfer->box.x,
+                                &rtransfer->staging->b.b, &box);
+               pipe_resource_reference((struct pipe_resource**)&rtransfer->staging, NULL);
+       }
 }
 
 static void r600_transfer_destroy(struct pipe_context *ctx,
@@ -201,14 +240,13 @@ bool r600_init_resource(struct r600_screen *rscreen,
 }
 
 struct pipe_resource *r600_buffer_create(struct pipe_screen *screen,
-                                        const struct pipe_resource *templ)
+                                        const struct pipe_resource *templ,
+                                        unsigned alignment)
 {
        struct r600_screen *rscreen = (struct r600_screen*)screen;
        struct r600_resource *rbuffer;
-       /* XXX We probably want a different alignment for buffers and textures. */
-       unsigned alignment = 4096;
 
-       rbuffer = util_slab_alloc(&rscreen->pool_buffers);
+       rbuffer = MALLOC_STRUCT(r600_resource);
 
        rbuffer->b.b = *templ;
        pipe_reference_init(&rbuffer->b.b.reference, 1);
@@ -216,33 +254,8 @@ struct pipe_resource *r600_buffer_create(struct pipe_screen *screen,
        rbuffer->b.vtbl = &r600_buffer_vtbl;
 
        if (!r600_init_resource(rscreen, rbuffer, templ->width0, alignment, templ->bind, templ->usage)) {
-               util_slab_free(&rscreen->pool_buffers, rbuffer);
+               FREE(rbuffer);
                return NULL;
        }
        return &rbuffer->b.b;
 }
-
-struct pipe_resource *r600_user_buffer_create(struct pipe_screen *screen,
-                                             void *ptr, unsigned bytes,
-                                             unsigned bind)
-{
-       struct r600_screen *rscreen = (struct r600_screen*)screen;
-       struct r600_resource *rbuffer;
-
-       rbuffer = util_slab_alloc(&rscreen->pool_buffers);
-
-       pipe_reference_init(&rbuffer->b.b.reference, 1);
-       rbuffer->b.vtbl = &r600_buffer_vtbl;
-       rbuffer->b.b.screen = screen;
-       rbuffer->b.b.target = PIPE_BUFFER;
-       rbuffer->b.b.format = PIPE_FORMAT_R8_UNORM;
-       rbuffer->b.b.usage = PIPE_USAGE_IMMUTABLE;
-       rbuffer->b.b.bind = bind;
-       rbuffer->b.b.width0 = bytes;
-       rbuffer->b.b.height0 = 1;
-       rbuffer->b.b.depth0 = 1;
-       rbuffer->b.b.array_size = 1;
-       rbuffer->b.b.flags = 0;
-       rbuffer->buf = NULL;
-       return &rbuffer->b.b;
-}