X-Git-Url: https://git.libre-soc.org/?a=blobdiff_plain;f=src%2Fgallium%2Fdrivers%2Fsvga%2Fsvga_resource_texture.c;h=2aa4e52faa75ae3f4a6b7d74379b3fa21bc3ff4e;hb=8c9b9aac7d09e65195dca6681d59c10e4ef713d9;hp=85e1805c546ffb0d9027fb8cbc9a93836200c811;hpb=a9c4a861d5d6bb88c9e6c375403c3985187fd453;p=mesa.git diff --git a/src/gallium/drivers/svga/svga_resource_texture.c b/src/gallium/drivers/svga/svga_resource_texture.c index 85e1805c546..2aa4e52faa7 100644 --- a/src/gallium/drivers/svga/svga_resource_texture.c +++ b/src/gallium/drivers/svga/svga_resource_texture.c @@ -29,11 +29,12 @@ #include "pipe/p_state.h" #include "pipe/p_defines.h" #include "os/os_thread.h" -#include "util/u_format.h" +#include "util/format/u_format.h" #include "util/u_inlines.h" #include "util/u_math.h" #include "util/u_memory.h" #include "util/u_resource.h" +#include "util/u_upload_mgr.h" #include "svga_cmd.h" #include "svga_format.h" @@ -115,42 +116,43 @@ svga_transfer_dma(struct svga_context *svga, /* Ensure any pending operations on host surfaces are queued on the command * buffer first. */ - svga_surfaces_flush( svga ); + svga_surfaces_flush(svga); if (!st->swbuf) { /* Do the DMA transfer in a single go */ svga_transfer_dma_band(svga, st, transfer, - st->base.box.x, st->base.box.y, st->base.box.z, - st->base.box.width, st->base.box.height, st->base.box.depth, + st->box.x, st->box.y, st->box.z, + st->box.w, st->box.h, st->box.d, 0, 0, 0, flags); if (transfer == SVGA3D_READ_HOST_VRAM) { svga_context_flush(svga, &fence); - sws->fence_finish(sws, fence, 0); + sws->fence_finish(sws, fence, PIPE_TIMEOUT_INFINITE, 0); sws->fence_reference(sws, &fence, NULL); } } else { - int y, h, srcy; + int y, h, y_max; unsigned blockheight = util_format_get_blockheight(st->base.resource->format); h = st->hw_nblocksy * blockheight; - srcy = 0; + y_max = st->box.y + st->box.h; - for (y = 0; y < st->base.box.height; y += h) { + for (y = st->box.y; y < y_max; y += h) { unsigned offset, length; void *hw, *sw; - if (y + h > st->base.box.height) - h = st->base.box.height - y; + if (y + h > y_max) + h = y_max - y; /* Transfer band must be aligned to pixel block boundaries */ assert(y % blockheight == 0); assert(h % blockheight == 0); - offset = y * st->base.stride / blockheight; + /* First band starts at the top of the SW buffer. */ + offset = (y - st->box.y) * st->base.stride / blockheight; length = h * st->base.stride / blockheight; sw = (uint8_t *) st->swbuf + offset; @@ -158,9 +160,9 @@ svga_transfer_dma(struct svga_context *svga, if (transfer == SVGA3D_WRITE_HOST_VRAM) { unsigned usage = PIPE_TRANSFER_WRITE; - /* Wait for the previous DMAs to complete */ - /* TODO: keep one DMA (at half the size) in the background */ - if (y) { + /* Don't write to an in-flight DMA buffer. Synchronize or + * discard in-flight storage. */ + if (y != st->box.y) { svga_context_flush(svga, NULL); usage |= PIPE_TRANSFER_DISCARD_WHOLE_RESOURCE; } @@ -174,9 +176,9 @@ svga_transfer_dma(struct svga_context *svga, } svga_transfer_dma_band(svga, st, transfer, - st->base.box.x, y, st->base.box.z, - st->base.box.width, h, st->base.box.depth, - 0, srcy, 0, flags); + st->box.x, y, st->box.z, + st->box.w, h, st->box.d, + 0, 0, 0, flags); /* * Prevent the texture contents to be discarded on the next band @@ -186,7 +188,7 @@ svga_transfer_dma(struct svga_context *svga, if (transfer == SVGA3D_READ_HOST_VRAM) { svga_context_flush(svga, &fence); - sws->fence_finish(sws, fence, 0); + sws->fence_finish(sws, fence, PIPE_TIMEOUT_INFINITE, 0); hw = sws->buffer_map(sws, st->hwbuf, PIPE_TRANSFER_READ); assert(hw); @@ -201,7 +203,7 @@ svga_transfer_dma(struct svga_context *svga, -static boolean +static bool svga_texture_get_handle(struct pipe_screen *screen, struct pipe_resource *texture, struct winsys_handle *whandle) @@ -222,7 +224,7 @@ svga_texture_get_handle(struct pipe_screen *screen, static void svga_texture_destroy(struct pipe_screen *screen, - struct pipe_resource *pt) + struct pipe_resource *pt) { struct svga_screen *ss = svga_screen(screen); struct svga_texture *tex = svga_texture(pt); @@ -237,6 +239,10 @@ svga_texture_destroy(struct pipe_screen *screen, SVGA_DBG(DEBUG_DMA, "unref sid %p (texture)\n", tex->handle); svga_screen_surface_destroy(ss, &tex->key, &tex->handle); + /* Destroy the backed surface handle if exists */ + if (tex->backed_handle) + svga_screen_surface_destroy(ss, &tex->backed_key, &tex->backed_handle); + ss->hud.total_resource_bytes -= tex->size; FREE(tex->defined); @@ -253,28 +259,16 @@ svga_texture_destroy(struct pipe_screen *screen, /** * Determine if we need to read back a texture image before mapping it. */ -static boolean -need_tex_readback(struct pipe_transfer *transfer) +static inline boolean +need_tex_readback(struct svga_transfer *st) { - struct svga_texture *t = svga_texture(transfer->resource); - - if (transfer->usage & PIPE_TRANSFER_READ) + if (st->base.usage & PIPE_TRANSFER_READ) return TRUE; - if ((transfer->usage & PIPE_TRANSFER_WRITE) && - ((transfer->usage & PIPE_TRANSFER_DISCARD_WHOLE_RESOURCE) == 0)) { - unsigned face; - - if (transfer->resource->target == PIPE_TEXTURE_CUBE) { - assert(transfer->box.depth == 1); - face = transfer->box.z; - } - else { - face = 0; - } - if (svga_was_texture_rendered_to(t, face, transfer->level)) { - return TRUE; - } + if ((st->base.usage & PIPE_TRANSFER_WRITE) && + ((st->base.usage & PIPE_TRANSFER_DISCARD_WHOLE_RESOURCE) == 0)) { + return svga_was_texture_rendered_to(svga_texture(st->base.resource), + st->slice, st->base.level); } return FALSE; @@ -332,9 +326,9 @@ svga_texture_transfer_map_dma(struct svga_context *svga, unsigned usage = st->base.usage; /* we'll put the data into a tightly packed buffer */ - nblocksx = util_format_get_nblocksx(texture->format, st->base.box.width); - nblocksy = util_format_get_nblocksy(texture->format, st->base.box.height); - d = st->base.box.depth; + nblocksx = util_format_get_nblocksx(texture->format, st->box.w); + nblocksy = util_format_get_nblocksy(texture->format, st->box.h); + d = st->box.d; st->base.stride = nblocksx*util_format_get_blocksize(texture->format); st->base.layer_stride = st->base.stride * nblocksy; @@ -400,50 +394,31 @@ svga_texture_transfer_map_direct(struct svga_context *svga, struct svga_texture *tex = svga_texture(texture); struct svga_winsys_surface *surf = tex->handle; unsigned level = st->base.level; - unsigned w, h, nblocksx, nblocksy; + unsigned w, h, nblocksx, nblocksy, i; unsigned usage = st->base.usage; - if (!surf) { - FREE(st); - return NULL; - } - - /* we'll directly access the guest-backed surface */ - w = u_minify(texture->width0, level); - h = u_minify(texture->height0, level); - nblocksx = util_format_get_nblocksx(texture->format, w); - nblocksy = util_format_get_nblocksy(texture->format, h); - st->hw_nblocksy = nblocksy; - st->base.stride = nblocksx*util_format_get_blocksize(texture->format); - st->base.layer_stride = st->base.stride * nblocksy; - - /* If this is the first time mapping to the surface in this - * command buffer, clear the dirty masks of this surface. - */ - if (sws->surface_is_flushed(sws, surf)) { - svga_clear_texture_dirty(tex); - } - - if (need_tex_readback(transfer)) { + if (need_tex_readback(st)) { enum pipe_error ret; svga_surfaces_flush(svga); - if (svga_have_vgpu10(svga)) { - ret = readback_image_vgpu10(svga, surf, st->slice, level, - tex->b.b.last_level + 1); - } else { - ret = readback_image_vgpu9(svga, surf, st->slice, level); - } - - svga->hud.num_readbacks++; - SVGA_STATS_COUNT_INC(sws, SVGA_STATS_COUNT_TEXREADBACK); - - assert(ret == PIPE_OK); - (void) ret; + if (!svga->swc->force_coherent || tex->imported) { + for (i = 0; i < st->box.d; i++) { + if (svga_have_vgpu10(svga)) { + ret = readback_image_vgpu10(svga, surf, st->slice + i, level, + tex->b.b.last_level + 1); + } else { + ret = readback_image_vgpu9(svga, surf, st->slice + i, level); + } + } + svga->hud.num_readbacks++; + SVGA_STATS_COUNT_INC(sws, SVGA_STATS_COUNT_TEXREADBACK); - svga_context_flush(svga, NULL); + assert(ret == PIPE_OK); + (void) ret; + svga_context_flush(svga, NULL); + } /* * Note: if PIPE_TRANSFER_DISCARD_WHOLE_RESOURCE were specified * we could potentially clear the flag for all faces/layers/mips. @@ -468,16 +443,30 @@ svga_texture_transfer_map_direct(struct svga_context *svga, } } + /* we'll directly access the guest-backed surface */ + w = u_minify(texture->width0, level); + h = u_minify(texture->height0, level); + nblocksx = util_format_get_nblocksx(texture->format, w); + nblocksy = util_format_get_nblocksy(texture->format, h); + st->hw_nblocksy = nblocksy; + st->base.stride = nblocksx*util_format_get_blocksize(texture->format); + st->base.layer_stride = st->base.stride * nblocksy; + /* * Begin mapping code */ { SVGA3dSize baseLevelSize; uint8_t *map; - boolean retry; + boolean retry, rebind; unsigned offset, mip_width, mip_height; + struct svga_winsys_context *swc = svga->swc; + + if (swc->force_coherent) { + usage |= PIPE_TRANSFER_PERSISTENT | PIPE_TRANSFER_COHERENT; + } - map = svga->swc->surface_map(svga->swc, surf, usage, &retry); + map = swc->surface_map(swc, surf, usage, &retry, &rebind); if (map == NULL && retry) { /* * At this point, the svga_surfaces_flush() should already have @@ -485,7 +474,18 @@ svga_texture_transfer_map_direct(struct svga_context *svga, */ svga->hud.surface_write_flushes++; svga_context_flush(svga, NULL); - map = svga->swc->surface_map(svga->swc, surf, usage, &retry); + map = swc->surface_map(swc, surf, usage, &retry, &rebind); + } + if (map && rebind) { + enum pipe_error ret; + + ret = SVGA3D_BindGBSurface(swc, surf); + if (ret != PIPE_OK) { + svga_context_flush(svga, NULL); + ret = SVGA3D_BindGBSurface(swc, surf); + assert(ret == PIPE_OK); + } + svga_context_flush(svga, NULL); } /* @@ -503,7 +503,8 @@ svga_texture_transfer_map_direct(struct svga_context *svga, baseLevelSize.depth = tex->b.b.depth0; if ((tex->b.b.target == PIPE_TEXTURE_1D_ARRAY) || - (tex->b.b.target == PIPE_TEXTURE_2D_ARRAY)) { + (tex->b.b.target == PIPE_TEXTURE_2D_ARRAY) || + (tex->b.b.target == PIPE_TEXTURE_CUBE_ARRAY)) { st->base.layer_stride = svga3dsurface_get_image_offset(tex->key.format, baseLevelSize, tex->b.b.last_level + 1, 1, 0); @@ -521,14 +522,9 @@ svga_texture_transfer_map_direct(struct svga_context *svga, offset += svga3dsurface_get_pixel_offset(tex->key.format, mip_width, mip_height, - st->base.box.x, - st->base.box.y, - st->base.box.z); - - if (usage & PIPE_TRANSFER_WRITE) { - /* mark this texture level as dirty */ - svga_set_texture_dirty(tex, st->slice, level); - } + st->box.x, + st->box.y, + st->box.z); return (void *) (map + offset); } @@ -550,13 +546,17 @@ svga_texture_transfer_map(struct pipe_context *pipe, struct svga_winsys_screen *sws = svga_screen(pipe->screen)->sws; struct svga_texture *tex = svga_texture(texture); struct svga_transfer *st; + struct svga_winsys_surface *surf = tex->handle; boolean use_direct_map = svga_have_gb_objects(svga) && - !svga_have_gb_dma(svga); - void *returnVal = NULL; + (!svga_have_gb_dma(svga) || (usage & PIPE_TRANSFER_WRITE)); + void *map = NULL; int64_t begin = svga_get_time(svga); SVGA_STATS_TIME_PUSH(sws, SVGA_STATS_TIME_TEXTRANSFERMAP); + if (!surf) + goto done; + /* We can't map texture storage directly unless we have GB objects */ if (usage & PIPE_TRANSFER_MAP_DIRECTLY) { if (svga_have_gb_objects(svga)) @@ -573,15 +573,26 @@ svga_texture_transfer_map(struct pipe_context *pipe, st->base.usage = usage; st->base.box = *box; + /* The modified transfer map box with the array index removed from z. + * The array index is specified in slice. + */ + st->box.x = box->x; + st->box.y = box->y; + st->box.z = box->z; + st->box.w = box->width; + st->box.h = box->height; + st->box.d = box->depth; + switch (tex->b.b.target) { case PIPE_TEXTURE_CUBE: st->slice = st->base.box.z; - st->base.box.z = 0; /* so we don't apply double offsets below */ + st->box.z = 0; /* so we don't apply double offsets below */ break; - case PIPE_TEXTURE_2D_ARRAY: case PIPE_TEXTURE_1D_ARRAY: + case PIPE_TEXTURE_2D_ARRAY: + case PIPE_TEXTURE_CUBE_ARRAY: st->slice = st->base.box.z; - st->base.box.z = 0; /* so we don't apply double offsets below */ + st->box.z = 0; /* so we don't apply double offsets below */ /* Force direct map for transfering multiple slices */ if (st->base.box.depth > 1) @@ -593,17 +604,69 @@ svga_texture_transfer_map(struct pipe_context *pipe, break; } + /* Force direct map for multisample surface */ + if (texture->nr_samples > 1) { + assert(svga_have_gb_objects(svga)); + assert(sws->have_sm4_1); + use_direct_map = TRUE; + } + st->use_direct_map = use_direct_map; pipe_resource_reference(&st->base.resource, texture); - if (use_direct_map) { - returnVal = svga_texture_transfer_map_direct(svga, st); + /* If this is the first time mapping to the surface in this + * command buffer, clear the dirty masks of this surface. + */ + if (sws->surface_is_flushed(sws, surf)) { + svga_clear_texture_dirty(tex); + } + + if (!use_direct_map) { + /* upload to the DMA buffer */ + map = svga_texture_transfer_map_dma(svga, st); } else { - returnVal = svga_texture_transfer_map_dma(svga, st); + boolean can_use_upload = tex->can_use_upload && + !(st->base.usage & PIPE_TRANSFER_READ); + boolean was_rendered_to = + svga_was_texture_rendered_to(svga_texture(texture), + st->slice, st->base.level); + + /* If the texture was already rendered to and upload buffer + * is supported, then we will use upload buffer to + * avoid the need to read back the texture content; otherwise, + * we'll first try to map directly to the GB surface, if it is blocked, + * then we'll try the upload buffer. + */ + if (was_rendered_to && can_use_upload) { + map = svga_texture_transfer_map_upload(svga, st); + } + else { + unsigned orig_usage = st->base.usage; + + /* First try directly map to the GB surface */ + if (can_use_upload) + st->base.usage |= PIPE_TRANSFER_DONTBLOCK; + map = svga_texture_transfer_map_direct(svga, st); + st->base.usage = orig_usage; + + if (!map && can_use_upload) { + /* if direct map with DONTBLOCK fails, then try upload to the + * texture upload buffer. + */ + map = svga_texture_transfer_map_upload(svga, st); + } + } + + /* If upload fails, then try direct map again without forcing it + * to DONTBLOCK. + */ + if (!map) { + map = svga_texture_transfer_map_direct(svga, st); + } } - if (!returnVal) { + if (!map) { FREE(st); } else { @@ -612,14 +675,19 @@ svga_texture_transfer_map(struct pipe_context *pipe, if (usage & PIPE_TRANSFER_WRITE) { /* record texture upload for HUD */ svga->hud.num_bytes_uploaded += - st->base.layer_stride * st->base.box.depth; + st->base.layer_stride * st->box.d; + + /* mark this texture level as dirty */ + svga_set_texture_dirty(tex, st->slice, level); } } done: svga->hud.map_buffer_time += (svga_get_time(svga) - begin); SVGA_STATS_TIME_POP(sws); - return returnVal; + (void) sws; + + return map; } /** @@ -679,6 +747,7 @@ update_image_vgpu10(struct svga_context *svga, unsigned subResource; subResource = slice * numMipLevels + level; + ret = SVGA3D_vgpu10_UpdateSubResource(svga->swc, surf, box, subResource); if (ret != PIPE_OK) { svga_context_flush(svga, NULL); @@ -693,16 +762,19 @@ update_image_vgpu10(struct svga_context *svga, */ static void svga_texture_transfer_unmap_dma(struct svga_context *svga, - struct svga_transfer *st) + struct svga_transfer *st) { struct svga_winsys_screen *sws = svga_screen(svga->pipe.screen)->sws; - if (st->hwbuf) + if (!st->swbuf) sws->buffer_unmap(sws, st->hwbuf); if (st->base.usage & PIPE_TRANSFER_WRITE) { /* Use DMA to transfer texture data */ SVGA3dSurfaceDMAFlags flags; + struct pipe_resource *texture = st->base.resource; + struct svga_texture *tex = svga_texture(texture); + memset(&flags, 0, sizeof flags); if (st->base.usage & PIPE_TRANSFER_DISCARD_WHOLE_RESOURCE) { @@ -713,6 +785,7 @@ svga_texture_transfer_unmap_dma(struct svga_context *svga, } svga_transfer_dma(svga, st, SVGA3D_WRITE_HOST_VRAM, flags); + svga_set_texture_rendered_to(tex, st->slice, st->base.level); } FREE(st->swbuf); @@ -725,72 +798,68 @@ svga_texture_transfer_unmap_dma(struct svga_context *svga, */ static void svga_texture_transfer_unmap_direct(struct svga_context *svga, - struct svga_transfer *st) + struct svga_transfer *st) { struct pipe_transfer *transfer = &st->base; struct svga_texture *tex = svga_texture(transfer->resource); svga_texture_surface_unmap(svga, transfer); + /* Now send an update command to update the content in the backend. */ if (st->base.usage & PIPE_TRANSFER_WRITE) { struct svga_winsys_surface *surf = tex->handle; - SVGA3dBox box; enum pipe_error ret; - unsigned nlayers = 1; assert(svga_have_gb_objects(svga)); /* update the effected region */ - box.x = transfer->box.x; - box.y = transfer->box.y; - box.w = transfer->box.width; - box.h = transfer->box.height; - box.d = transfer->box.depth; + SVGA3dBox box = st->box; + unsigned nlayers; switch (tex->b.b.target) { - case PIPE_TEXTURE_CUBE: - box.z = 0; - break; case PIPE_TEXTURE_2D_ARRAY: - nlayers = box.d; - box.z = 0; - box.d = 1; - break; + case PIPE_TEXTURE_CUBE_ARRAY: case PIPE_TEXTURE_1D_ARRAY: nlayers = box.d; - box.y = box.z = 0; box.d = 1; break; default: - box.z = transfer->box.z; + nlayers = 1; break; } + if (0) debug_printf("%s %d, %d, %d %d x %d x %d\n", __FUNCTION__, box.x, box.y, box.z, box.w, box.h, box.d); - if (svga_have_vgpu10(svga)) { - unsigned i; - for (i = 0; i < nlayers; i++) { - ret = update_image_vgpu10(svga, surf, &box, - st->slice + i, transfer->level, - tex->b.b.last_level + 1); + if (!svga->swc->force_coherent || tex->imported) { + if (svga_have_vgpu10(svga)) { + unsigned i; + + for (i = 0; i < nlayers; i++) { + ret = update_image_vgpu10(svga, surf, &box, + st->slice + i, transfer->level, + tex->b.b.last_level + 1); + assert(ret == PIPE_OK); + } + } else { + assert(nlayers == 1); + ret = update_image_vgpu9(svga, surf, &box, st->slice, + transfer->level); assert(ret == PIPE_OK); } - } else { - assert(nlayers == 1); - ret = update_image_vgpu9(svga, surf, &box, st->slice, transfer->level); - assert(ret == PIPE_OK); } + (void) ret; } } + static void svga_texture_transfer_unmap(struct pipe_context *pipe, - struct pipe_transfer *transfer) + struct pipe_transfer *transfer) { struct svga_context *svga = svga_context(pipe); struct svga_screen *ss = svga_screen(pipe->screen); @@ -800,16 +869,20 @@ svga_texture_transfer_unmap(struct pipe_context *pipe, SVGA_STATS_TIME_PUSH(sws, SVGA_STATS_TIME_TEXTRANSFERUNMAP); - if (st->use_direct_map) { - svga_texture_transfer_unmap_direct(svga, st); + if (!st->use_direct_map) { + svga_texture_transfer_unmap_dma(svga, st); + } + else if (st->upload.buf) { + svga_texture_transfer_unmap_upload(svga, st); } else { - svga_texture_transfer_unmap_dma(svga, st); + svga_texture_transfer_unmap_direct(svga, st); } if (st->base.usage & PIPE_TRANSFER_WRITE) { svga->hud.num_resource_updates++; + /* Mark the texture level as dirty */ ss->texture_timestamp++; svga_age_texture_view(tex, transfer->level); if (transfer->resource->target == PIPE_TEXTURE_CUBE) @@ -821,6 +894,7 @@ svga_texture_transfer_unmap(struct pipe_context *pipe, pipe_resource_reference(&st->base.resource, NULL); FREE(st); SVGA_STATS_TIME_POP(sws); + (void) sws; } @@ -861,6 +935,39 @@ svga_texture_create(struct pipe_screen *screen, goto fail_notex; } + /* Verify the number of mipmap levels isn't impossibly large. For example, + * if the base 2D image is 16x16, we can't have 8 mipmap levels. + * the gallium frontend should never ask us to create a resource with invalid + * parameters. + */ + { + unsigned max_dim = template->width0; + + switch (template->target) { + case PIPE_TEXTURE_1D: + case PIPE_TEXTURE_1D_ARRAY: + // nothing + break; + case PIPE_TEXTURE_2D: + case PIPE_TEXTURE_CUBE: + case PIPE_TEXTURE_CUBE_ARRAY: + case PIPE_TEXTURE_2D_ARRAY: + max_dim = MAX2(max_dim, template->height0); + break; + case PIPE_TEXTURE_3D: + max_dim = MAX3(max_dim, template->height0, template->depth0); + break; + case PIPE_TEXTURE_RECT: + case PIPE_BUFFER: + assert(template->last_level == 0); + /* the assertion below should always pass */ + break; + default: + debug_printf("Unexpected texture target type\n"); + } + assert(1 << template->last_level <= max_dim); + } + tex = CALLOC_STRUCT(svga_texture); if (!tex) { goto fail_notex; @@ -897,13 +1004,17 @@ svga_texture_create(struct pipe_screen *screen, tex->key.arraySize = 1; tex->key.numFaces = 1; - /* single sample texture can be treated as non-multisamples texture */ - tex->key.sampleCount = template->nr_samples > 1 ? template->nr_samples : 0; - - if (template->nr_samples > 1) { - tex->key.flags |= SVGA3D_SURFACE_MASKABLE_ANTIALIAS; + /* nr_samples=1 must be treated as a non-multisample texture */ + if (tex->b.b.nr_samples == 1) { + tex->b.b.nr_samples = 0; + } + else if (tex->b.b.nr_samples > 1) { + assert(svgascreen->sws->have_sm4_1); + tex->key.flags |= SVGA3D_SURFACE_MULTISAMPLE; } + tex->key.sampleCount = tex->b.b.nr_samples; + if (svgascreen->sws->have_vgpu10) { switch (template->target) { case PIPE_TEXTURE_1D: @@ -923,6 +1034,12 @@ svga_texture_create(struct pipe_screen *screen, tex->key.flags |= (SVGA3D_SURFACE_CUBEMAP | SVGA3D_SURFACE_ARRAY); tex->key.numFaces = 6; break; + case PIPE_TEXTURE_CUBE_ARRAY: + assert(svgascreen->sws->have_sm4_1); + tex->key.flags |= (SVGA3D_SURFACE_CUBEMAP | SVGA3D_SURFACE_ARRAY); + tex->key.numFaces = 1; // arraySize already includes the 6 faces + tex->key.arraySize = template->array_size; + break; default: break; } @@ -949,6 +1066,7 @@ svga_texture_create(struct pipe_screen *screen, if (screen->is_format_supported(screen, template->format, template->target, template->nr_samples, + template->nr_storage_samples, PIPE_BIND_SAMPLER_VIEW)) { bindings |= PIPE_BIND_SAMPLER_VIEW; } @@ -959,14 +1077,26 @@ svga_texture_create(struct pipe_screen *screen, tex->key.flags |= SVGA3D_SURFACE_BIND_SHADER_RESOURCE; if (!(bindings & PIPE_BIND_RENDER_TARGET)) { - /* Also check if the format is renderable */ + /* Also check if the format is color renderable */ if (screen->is_format_supported(screen, template->format, template->target, template->nr_samples, + template->nr_storage_samples, PIPE_BIND_RENDER_TARGET)) { bindings |= PIPE_BIND_RENDER_TARGET; } } + + if (!(bindings & PIPE_BIND_DEPTH_STENCIL)) { + /* Also check if the format is depth/stencil renderable */ + if (screen->is_format_supported(screen, template->format, + template->target, + template->nr_samples, + template->nr_storage_samples, + PIPE_BIND_DEPTH_STENCIL)) { + bindings |= PIPE_BIND_DEPTH_STENCIL; + } + } } if (bindings & PIPE_BIND_DISPLAY_TARGET) { @@ -989,7 +1119,7 @@ svga_texture_create(struct pipe_screen *screen, * and it always requests PIPE_BIND_RENDER_TARGET, therefore * passing the SVGA3D_SURFACE_HINT_RENDERTARGET here defeats its purpose. * - * However, this was changed since other state trackers + * However, this was changed since other gallium frontends * (XA for example) uses it accurately and certain device versions * relies on it in certain situations to render correctly. */ @@ -1012,17 +1142,14 @@ svga_texture_create(struct pipe_screen *screen, goto fail; } - /* The actual allocation is done with a typeless format. Typeless + /* Use typeless formats for sRGB and depth resources. Typeless * formats can be reinterpreted as other formats. For example, * SVGA3D_R8G8B8A8_UNORM_TYPELESS can be interpreted as * SVGA3D_R8G8B8A8_UNORM_SRGB or SVGA3D_R8G8B8A8_UNORM. - * Do not use typeless formats for SHARED, DISPLAY_TARGET or SCANOUT - * buffers. */ - if (svgascreen->sws->have_vgpu10 - && ((bindings & (PIPE_BIND_SHARED | - PIPE_BIND_DISPLAY_TARGET | - PIPE_BIND_SCANOUT)) == 0)) { + if (svgascreen->sws->have_vgpu10 && + (util_format_is_srgb(template->format) || + format_has_depth(template->format))) { SVGA3dSurfaceFormat typeless = svga_typeless_format(tex->key.format); if (0) { debug_printf("Convert resource type %s -> %s (bind 0x%x)\n", @@ -1045,7 +1172,8 @@ svga_texture_create(struct pipe_screen *screen, SVGA_DBG(DEBUG_DMA, "surface_create for texture\n", tex->handle); tex->handle = svga_screen_surface_create(svgascreen, bindings, - tex->b.b.usage, &tex->key); + tex->b.b.usage, + &tex->validated, &tex->key); if (!tex->handle) { goto fail; } @@ -1056,6 +1184,14 @@ svga_texture_create(struct pipe_screen *screen, (debug_reference_descriptor)debug_describe_resource, 0); tex->size = util_resource_size(template); + + /* Determine if texture upload buffer can be used to upload this texture */ + tex->can_use_upload = svga_texture_transfer_map_can_upload(svgascreen, + &tex->b.b); + + /* Initialize the backing resource cache */ + tex->backed_handle = NULL; + svgascreen->hud.total_resource_bytes += tex->size; svgascreen->hud.num_resources++; @@ -1079,8 +1215,8 @@ fail_notex: struct pipe_resource * svga_texture_from_handle(struct pipe_screen *screen, - const struct pipe_resource *template, - struct winsys_handle *whandle) + const struct pipe_resource *template, + struct winsys_handle *whandle) { struct svga_winsys_screen *sws = svga_winsys_screen(screen); struct svga_screen *ss = svga_screen(screen); @@ -1102,42 +1238,18 @@ svga_texture_from_handle(struct pipe_screen *screen, if (!srf) return NULL; - if (svga_translate_format(svga_screen(screen), template->format, - template->bind) != format) { - unsigned f1 = svga_translate_format(svga_screen(screen), - template->format, template->bind); - unsigned f2 = format; - - /* It's okay for XRGB and ARGB or depth with/out stencil to get mixed up. - */ - if (f1 == SVGA3D_B8G8R8A8_UNORM) - f1 = SVGA3D_A8R8G8B8; - if (f1 == SVGA3D_B8G8R8X8_UNORM) - f1 = SVGA3D_X8R8G8B8; - - if ( !( (f1 == f2) || - (f1 == SVGA3D_X8R8G8B8 && f2 == SVGA3D_A8R8G8B8) || - (f1 == SVGA3D_X8R8G8B8 && f2 == SVGA3D_B8G8R8X8_UNORM) || - (f1 == SVGA3D_A8R8G8B8 && f2 == SVGA3D_X8R8G8B8) || - (f1 == SVGA3D_A8R8G8B8 && f2 == SVGA3D_B8G8R8A8_UNORM) || - (f1 == SVGA3D_Z_D24X8 && f2 == SVGA3D_Z_D24S8) || - (f1 == SVGA3D_Z_DF24 && f2 == SVGA3D_Z_D24S8_INT) ) ) { - debug_printf("%s wrong format %s != %s\n", __FUNCTION__, - svga_format_name(f1), svga_format_name(f2)); - return NULL; - } - } + if (!svga_format_is_shareable(ss, template->format, format, + template->bind, true)) + goto out_unref; tex = CALLOC_STRUCT(svga_texture); if (!tex) - return NULL; + goto out_unref; tex->defined = CALLOC(template->depth0 * template->array_size, sizeof(tex->defined[0])); - if (!tex->defined) { - FREE(tex); - return NULL; - } + if (!tex->defined) + goto out_no_defined; tex->b.b = *template; tex->b.vtbl = &svga_texture_vtbl; @@ -1152,11 +1264,11 @@ svga_texture_from_handle(struct pipe_screen *screen, tex->rendered_to = CALLOC(1, sizeof(tex->rendered_to[0])); if (!tex->rendered_to) - goto fail; + goto out_no_rendered_to; tex->dirty = CALLOC(1, sizeof(tex->dirty[0])); if (!tex->dirty) - goto fail; + goto out_no_dirty; tex->imported = TRUE; @@ -1164,18 +1276,18 @@ svga_texture_from_handle(struct pipe_screen *screen, return &tex->b.b; -fail: - if (tex->defined) - FREE(tex->defined); - if (tex->rendered_to) - FREE(tex->rendered_to); - if (tex->dirty) - FREE(tex->dirty); +out_no_dirty: + FREE(tex->rendered_to); +out_no_rendered_to: + FREE(tex->defined); +out_no_defined: FREE(tex); +out_unref: + sws->surface_reference(sws, &srf, NULL); return NULL; } -boolean +bool svga_texture_generate_mipmap(struct pipe_context *pipe, struct pipe_resource *pt, enum pipe_format format, @@ -1194,19 +1306,19 @@ svga_texture_generate_mipmap(struct pipe_context *pipe, /* Only support 2D texture for now */ if (pt->target != PIPE_TEXTURE_2D) - return FALSE; + return false; /* Fallback to the mipmap generation utility for those formats that * do not support hw generate mipmap */ if (!svga_format_support_gen_mips(format)) - return FALSE; + return false; /* Make sure the texture surface was created with * SVGA3D_SURFACE_BIND_RENDER_TARGET */ if (!tex->handle || !(tex->key.flags & SVGA3D_SURFACE_BIND_RENDER_TARGET)) - return FALSE; + return false; templ.format = format; templ.u.tex.first_layer = first_layer; @@ -1216,7 +1328,7 @@ svga_texture_generate_mipmap(struct pipe_context *pipe, psv = pipe->create_sampler_view(pipe, pt, &templ); if (psv == NULL) - return FALSE; + return false; sv = svga_pipe_sampler_view(psv); ret = svga_validate_pipe_sampler_view(svga, sv); @@ -1235,5 +1347,239 @@ svga_texture_generate_mipmap(struct pipe_context *pipe, svga->hud.num_generate_mipmap++; + return true; +} + + +/* texture upload buffer default size in bytes */ +#define TEX_UPLOAD_DEFAULT_SIZE (1024 * 1024) + +/** + * Create a texture upload buffer + */ +boolean +svga_texture_transfer_map_upload_create(struct svga_context *svga) +{ + svga->tex_upload = u_upload_create(&svga->pipe, TEX_UPLOAD_DEFAULT_SIZE, + PIPE_BIND_CUSTOM, PIPE_USAGE_STAGING, 0); + if (svga->tex_upload) + u_upload_disable_persistent(svga->tex_upload); + + return svga->tex_upload != NULL; +} + + +/** + * Destroy the texture upload buffer + */ +void +svga_texture_transfer_map_upload_destroy(struct svga_context *svga) +{ + u_upload_destroy(svga->tex_upload); +} + + +/** + * Returns true if this transfer map request can use the upload buffer. + */ +boolean +svga_texture_transfer_map_can_upload(const struct svga_screen *svgascreen, + const struct pipe_resource *texture) +{ + if (svgascreen->sws->have_transfer_from_buffer_cmd == FALSE) + return FALSE; + + /* TransferFromBuffer command is not well supported with multi-samples surface */ + if (texture->nr_samples > 1) + return FALSE; + + if (util_format_is_compressed(texture->format)) { + /* XXX Need to take a closer look to see why texture upload + * with 3D texture with compressed format fails + */ + if (texture->target == PIPE_TEXTURE_3D) + return FALSE; + } + else if (texture->format == PIPE_FORMAT_R9G9B9E5_FLOAT) { + return FALSE; + } + return TRUE; } + + +/** + * Use upload buffer for the transfer map request. + */ +void * +svga_texture_transfer_map_upload(struct svga_context *svga, + struct svga_transfer *st) +{ + struct pipe_resource *texture = st->base.resource; + struct pipe_resource *tex_buffer = NULL; + void *tex_map; + unsigned nblocksx, nblocksy; + unsigned offset; + unsigned upload_size; + + assert(svga->tex_upload); + + st->upload.box.x = st->base.box.x; + st->upload.box.y = st->base.box.y; + st->upload.box.z = st->base.box.z; + st->upload.box.w = st->base.box.width; + st->upload.box.h = st->base.box.height; + st->upload.box.d = st->base.box.depth; + st->upload.nlayers = 1; + + switch (texture->target) { + case PIPE_TEXTURE_CUBE: + st->upload.box.z = 0; + break; + case PIPE_TEXTURE_2D_ARRAY: + case PIPE_TEXTURE_CUBE_ARRAY: + st->upload.nlayers = st->base.box.depth; + st->upload.box.z = 0; + st->upload.box.d = 1; + break; + case PIPE_TEXTURE_1D_ARRAY: + st->upload.nlayers = st->base.box.depth; + st->upload.box.y = st->upload.box.z = 0; + st->upload.box.d = 1; + break; + default: + break; + } + + nblocksx = util_format_get_nblocksx(texture->format, st->base.box.width); + nblocksy = util_format_get_nblocksy(texture->format, st->base.box.height); + + st->base.stride = nblocksx * util_format_get_blocksize(texture->format); + st->base.layer_stride = st->base.stride * nblocksy; + + /* In order to use the TransferFromBuffer command to update the + * texture content from the buffer, the layer stride for a multi-layers + * surface needs to be in multiples of 16 bytes. + */ + if (st->upload.nlayers > 1 && st->base.layer_stride & 15) + return NULL; + + upload_size = st->base.layer_stride * st->base.box.depth; + upload_size = align(upload_size, 16); + +#ifdef DEBUG + if (util_format_is_compressed(texture->format)) { + struct svga_texture *tex = svga_texture(texture); + unsigned blockw, blockh, bytesPerBlock; + + svga_format_size(tex->key.format, &blockw, &blockh, &bytesPerBlock); + + /* dest box must start on block boundary */ + assert((st->base.box.x % blockw) == 0); + assert((st->base.box.y % blockh) == 0); + } +#endif + + /* If the upload size exceeds the default buffer size, the + * upload buffer manager code will try to allocate a new buffer + * with the new buffer size. + */ + u_upload_alloc(svga->tex_upload, 0, upload_size, 16, + &offset, &tex_buffer, &tex_map); + + if (!tex_map) { + return NULL; + } + + st->upload.buf = tex_buffer; + st->upload.map = tex_map; + st->upload.offset = offset; + + return tex_map; +} + + +/** + * Unmap upload map transfer request + */ +void +svga_texture_transfer_unmap_upload(struct svga_context *svga, + struct svga_transfer *st) +{ + struct svga_winsys_surface *srcsurf; + struct svga_winsys_surface *dstsurf; + struct pipe_resource *texture = st->base.resource; + struct svga_texture *tex = svga_texture(texture); + enum pipe_error ret; + unsigned subResource; + unsigned numMipLevels; + unsigned i, layer; + unsigned offset = st->upload.offset; + + assert(svga->tex_upload); + assert(st->upload.buf); + + /* unmap the texture upload buffer */ + u_upload_unmap(svga->tex_upload); + + srcsurf = svga_buffer_handle(svga, st->upload.buf, 0); + dstsurf = svga_texture(texture)->handle; + assert(dstsurf); + + numMipLevels = texture->last_level + 1; + + for (i = 0, layer = st->slice; i < st->upload.nlayers; i++, layer++) { + subResource = layer * numMipLevels + st->base.level; + + /* send a transferFromBuffer command to update the host texture surface */ + assert((offset & 15) == 0); + + ret = SVGA3D_vgpu10_TransferFromBuffer(svga->swc, srcsurf, + offset, + st->base.stride, + st->base.layer_stride, + dstsurf, subResource, + &st->upload.box); + if (ret != PIPE_OK) { + svga_context_flush(svga, NULL); + ret = SVGA3D_vgpu10_TransferFromBuffer(svga->swc, srcsurf, + offset, + st->base.stride, + st->base.layer_stride, + dstsurf, subResource, + &st->upload.box); + assert(ret == PIPE_OK); + } + offset += st->base.layer_stride; + + /* Set rendered-to flag */ + svga_set_texture_rendered_to(tex, layer, st->base.level); + } + + pipe_resource_reference(&st->upload.buf, NULL); +} + +/** + * Does the device format backing this surface have an + * alpha channel? + * + * \param texture[in] The texture whose format we're querying + * \return TRUE if the format has an alpha channel, FALSE otherwise + * + * For locally created textures, the device (svga) format is typically + * identical to svga_format(texture->format), and we can use the gallium + * format tests to determine whether the device format has an alpha channel + * or not. However, for textures backed by imported svga surfaces that is + * not always true, and we have to look at the SVGA3D utilities. + */ +boolean +svga_texture_device_format_has_alpha(struct pipe_resource *texture) +{ + /* the svga_texture() call below is invalid for PIPE_BUFFER resources */ + assert(texture->target != PIPE_BUFFER); + + enum svga3d_block_desc block_desc = + svga3dsurface_get_desc(svga_texture(texture)->key.format)->block_desc; + + return !!(block_desc & SVGA3DBLOCKDESC_ALPHA); +}