X-Git-Url: https://git.libre-soc.org/?a=blobdiff_plain;f=src%2Famd%2Fvulkan%2Fradv_meta_resolve.c;h=f3e088b10c4869c4dd29486170eacd33b22601be;hb=0defc5554780a444c9e2009178dc88f97740a174;hp=23ca569d801b76600c12acc82bcff90a277c02af;hpb=6d5d6dad20901fbffd65239937562b80c075e476;p=mesa.git diff --git a/src/amd/vulkan/radv_meta_resolve.c b/src/amd/vulkan/radv_meta_resolve.c index 23ca569d801..f3e088b10c4 100644 --- a/src/amd/vulkan/radv_meta_resolve.c +++ b/src/amd/vulkan/radv_meta_resolve.c @@ -26,42 +26,11 @@ #include "radv_meta.h" #include "radv_private.h" +#include "vk_format.h" #include "nir/nir_builder.h" #include "sid.h" -/** - * Vertex attributes used by all pipelines. - */ -struct vertex_attrs { - float position[2]; /**< 3DPRIM_RECTLIST */ -}; - -/* passthrough vertex shader */ -static nir_shader * -build_nir_vs(void) -{ - const struct glsl_type *vec4 = glsl_vec4_type(); - - nir_builder b; - nir_variable *a_position; - nir_variable *v_position; - - nir_builder_init_simple_shader(&b, NULL, MESA_SHADER_VERTEX, NULL); - b.shader->info->name = ralloc_strdup(b.shader, "meta_resolve_vs"); - - a_position = nir_variable_create(b.shader, nir_var_shader_in, vec4, - "a_position"); - a_position->data.location = VERT_ATTRIB_GENERIC0; - - v_position = nir_variable_create(b.shader, nir_var_shader_out, vec4, - "gl_Position"); - v_position->data.location = VARYING_SLOT_POS; - - nir_copy_var(&b, v_position, a_position); - - return b.shader; -} -/* simple passthrough shader */ +/* emit 0, 0, 0, 1 */ static nir_shader * build_nir_fs(void) { @@ -70,7 +39,7 @@ build_nir_fs(void) nir_variable *f_color; /* vec4, fragment output color */ nir_builder_init_simple_shader(&b, NULL, MESA_SHADER_FRAGMENT, NULL); - b.shader->info->name = ralloc_asprintf(b.shader, + b.shader->info.name = ralloc_asprintf(b.shader, "meta_resolve_fs"); f_color = nir_variable_create(b.shader, nir_var_shader_out, vec4, @@ -82,7 +51,7 @@ build_nir_fs(void) } static VkResult -create_pass(struct radv_device *device) +create_pass(struct radv_device *device, VkFormat vk_format, VkRenderPass *pass) { VkResult result; VkDevice device_h = radv_device_to_handle(device); @@ -91,7 +60,7 @@ create_pass(struct radv_device *device) int i; for (i = 0; i < 2; i++) { - attachments[i].format = VK_FORMAT_UNDEFINED; + attachments[i].format = vk_format; attachments[i].samples = 1; attachments[i].loadOp = VK_ATTACHMENT_LOAD_OP_LOAD; attachments[i].storeOp = VK_ATTACHMENT_STORE_OP_STORE; @@ -131,14 +100,16 @@ create_pass(struct radv_device *device) .dependencyCount = 0, }, alloc, - &device->meta_state.resolve.pass); + pass); return result; } static VkResult create_pipeline(struct radv_device *device, - VkShaderModule vs_module_h) + VkShaderModule vs_module_h, + VkPipeline *pipeline, + VkRenderPass pass) { VkResult result; VkDevice device_h = radv_device_to_handle(device); @@ -153,6 +124,23 @@ create_pipeline(struct radv_device *device, goto cleanup; } + VkPipelineLayoutCreateInfo pl_create_info = { + .sType = VK_STRUCTURE_TYPE_PIPELINE_LAYOUT_CREATE_INFO, + .setLayoutCount = 0, + .pSetLayouts = NULL, + .pushConstantRangeCount = 0, + .pPushConstantRanges = NULL, + }; + + if (!device->meta_state.resolve.p_layout) { + result = radv_CreatePipelineLayout(radv_device_to_handle(device), + &pl_create_info, + &device->meta_state.alloc, + &device->meta_state.resolve.p_layout); + if (result != VK_SUCCESS) + goto cleanup; + } + result = radv_graphics_pipeline_create(device_h, radv_pipeline_cache_to_handle(&device->meta_state.cache), &(VkGraphicsPipelineCreateInfo) { @@ -174,24 +162,8 @@ create_pipeline(struct radv_device *device, }, .pVertexInputState = &(VkPipelineVertexInputStateCreateInfo) { .sType = VK_STRUCTURE_TYPE_PIPELINE_VERTEX_INPUT_STATE_CREATE_INFO, - .vertexBindingDescriptionCount = 1, - .pVertexBindingDescriptions = (VkVertexInputBindingDescription[]) { - { - .binding = 0, - .stride = sizeof(struct vertex_attrs), - .inputRate = VK_VERTEX_INPUT_RATE_VERTEX - }, - }, - .vertexAttributeDescriptionCount = 1, - .pVertexAttributeDescriptions = (VkVertexInputAttributeDescription[]) { - { - /* Position */ - .location = 0, - .binding = 0, - .format = VK_FORMAT_R32G32_SFLOAT, - .offset = offsetof(struct vertex_attrs, position), - }, - }, + .vertexBindingDescriptionCount = 0, + .vertexAttributeDescriptionCount = 0, }, .pInputAssemblyState = &(VkPipelineInputAssemblyStateCreateInfo) { .sType = VK_STRUCTURE_TYPE_PIPELINE_INPUT_ASSEMBLY_STATE_CREATE_INFO, @@ -200,8 +172,8 @@ create_pipeline(struct radv_device *device, }, .pViewportState = &(VkPipelineViewportStateCreateInfo) { .sType = VK_STRUCTURE_TYPE_PIPELINE_VIEWPORT_STATE_CREATE_INFO, - .viewportCount = 0, - .scissorCount = 0, + .viewportCount = 1, + .scissorCount = 1, }, .pRasterizationState = &(VkPipelineRasterizationStateCreateInfo) { .sType = VK_STRUCTURE_TYPE_PIPELINE_RASTERIZATION_STATE_CREATE_INFO, @@ -236,16 +208,23 @@ create_pipeline(struct radv_device *device, } }, }, - .pDynamicState = NULL, - .renderPass = device->meta_state.resolve.pass, + .pDynamicState = &(VkPipelineDynamicStateCreateInfo) { + .sType = VK_STRUCTURE_TYPE_PIPELINE_DYNAMIC_STATE_CREATE_INFO, + .dynamicStateCount = 2, + .pDynamicStates = (VkDynamicState[]) { + VK_DYNAMIC_STATE_VIEWPORT, + VK_DYNAMIC_STATE_SCISSOR, + }, + }, + .layout = device->meta_state.resolve.p_layout, + .renderPass = pass, .subpass = 0, }, &(struct radv_graphics_pipeline_create_info) { .use_rectlist = true, .custom_blend_mode = V_028808_CB_RESOLVE, }, - &device->meta_state.alloc, - &device->meta_state.resolve.pipeline); + &device->meta_state.alloc, pipeline); if (result != VK_SUCCESS) goto cleanup; @@ -260,42 +239,58 @@ void radv_device_finish_meta_resolve_state(struct radv_device *device) { struct radv_meta_state *state = &device->meta_state; - VkDevice device_h = radv_device_to_handle(device); - VkRenderPass pass_h = device->meta_state.resolve.pass; - const VkAllocationCallbacks *alloc = &device->meta_state.alloc; - - if (pass_h) - radv_DestroyRenderPass(device_h, pass_h, - &device->meta_state.alloc); - VkPipeline pipeline_h = state->resolve.pipeline; - if (pipeline_h) { - radv_DestroyPipeline(device_h, pipeline_h, alloc); + for (uint32_t j = 0; j < NUM_META_FS_KEYS; j++) { + radv_DestroyRenderPass(radv_device_to_handle(device), + state->resolve.pass[j], &state->alloc); + radv_DestroyPipeline(radv_device_to_handle(device), + state->resolve.pipeline[j], &state->alloc); } + radv_DestroyPipelineLayout(radv_device_to_handle(device), + state->resolve.p_layout, &state->alloc); + } +static VkFormat pipeline_formats[] = { + VK_FORMAT_R8G8B8A8_UNORM, + VK_FORMAT_R8G8B8A8_UINT, + VK_FORMAT_R8G8B8A8_SINT, + VK_FORMAT_A2R10G10B10_UINT_PACK32, + VK_FORMAT_A2R10G10B10_SINT_PACK32, + VK_FORMAT_R16G16B16A16_UNORM, + VK_FORMAT_R16G16B16A16_SNORM, + VK_FORMAT_R16G16B16A16_UINT, + VK_FORMAT_R16G16B16A16_SINT, + VK_FORMAT_R32_SFLOAT, + VK_FORMAT_R32G32_SFLOAT, + VK_FORMAT_R32G32B32A32_SFLOAT +}; + VkResult radv_device_init_meta_resolve_state(struct radv_device *device) { VkResult res = VK_SUCCESS; - - zero(device->meta_state.resolve); - - struct radv_shader_module vs_module = { .nir = build_nir_vs() }; + struct radv_meta_state *state = &device->meta_state; + struct radv_shader_module vs_module = { .nir = radv_meta_build_nir_vs_generate_vertices() }; if (!vs_module.nir) { /* XXX: Need more accurate error */ res = VK_ERROR_OUT_OF_HOST_MEMORY; goto fail; } - res = create_pass(device); - if (res != VK_SUCCESS) - goto fail; - - VkShaderModule vs_module_h = radv_shader_module_to_handle(&vs_module); - res = create_pipeline(device, vs_module_h); - if (res != VK_SUCCESS) - goto fail; + for (uint32_t i = 0; i < ARRAY_SIZE(pipeline_formats); ++i) { + VkFormat format = pipeline_formats[i]; + unsigned fs_key = radv_format_meta_fs_key(format); + res = create_pass(device, format, &state->resolve.pass[fs_key]); + if (res != VK_SUCCESS) + goto fail; + + VkShaderModule vs_module_h = radv_shader_module_to_handle(&vs_module); + res = create_pipeline(device, vs_module_h, + &state->resolve.pipeline[fs_key], state->resolve.pass[fs_key]); + if (res != VK_SUCCESS) + goto fail; + } goto cleanup; @@ -310,61 +305,65 @@ cleanup: static void emit_resolve(struct radv_cmd_buffer *cmd_buffer, + VkFormat vk_format, const VkOffset2D *dest_offset, const VkExtent2D *resolve_extent) { struct radv_device *device = cmd_buffer->device; VkCommandBuffer cmd_buffer_h = radv_cmd_buffer_to_handle(cmd_buffer); - uint32_t offset; - const struct vertex_attrs vertex_data[3] = { - { - .position = { - dest_offset->x, - dest_offset->y, - }, - }, - { - .position = { - dest_offset->x, - dest_offset->y + resolve_extent->height, - }, - }, - { - .position = { - dest_offset->x + resolve_extent->width, - dest_offset->y, - }, - }, - }; + unsigned fs_key = radv_format_meta_fs_key(vk_format); cmd_buffer->state.flush_bits |= RADV_CMD_FLAG_FLUSH_AND_INV_CB; - radv_cmd_buffer_upload_data(cmd_buffer, sizeof(vertex_data), 16, vertex_data, &offset); - struct radv_buffer vertex_buffer = { - .device = device, - .size = sizeof(vertex_data), - .bo = cmd_buffer->upload.upload_bo, - .offset = offset, - }; - VkBuffer vertex_buffer_h = radv_buffer_to_handle(&vertex_buffer); + radv_CmdBindPipeline(cmd_buffer_h, VK_PIPELINE_BIND_POINT_GRAPHICS, + device->meta_state.resolve.pipeline[fs_key]); - radv_CmdBindVertexBuffers(cmd_buffer_h, - /*firstBinding*/ 0, - /*bindingCount*/ 1, - (VkBuffer[]) { vertex_buffer_h }, - (VkDeviceSize[]) { 0 }); + radv_CmdSetViewport(radv_cmd_buffer_to_handle(cmd_buffer), 0, 1, &(VkViewport) { + .x = dest_offset->x, + .y = dest_offset->y, + .width = resolve_extent->width, + .height = resolve_extent->height, + .minDepth = 0.0f, + .maxDepth = 1.0f + }); - VkPipeline pipeline_h = device->meta_state.resolve.pipeline; - RADV_FROM_HANDLE(radv_pipeline, pipeline, pipeline_h); - - if (cmd_buffer->state.pipeline != pipeline) { - radv_CmdBindPipeline(cmd_buffer_h, VK_PIPELINE_BIND_POINT_GRAPHICS, - pipeline_h); - } + radv_CmdSetScissor(radv_cmd_buffer_to_handle(cmd_buffer), 0, 1, &(VkRect2D) { + .offset = *dest_offset, + .extent = *resolve_extent, + }); radv_CmdDraw(cmd_buffer_h, 3, 1, 0, 0); cmd_buffer->state.flush_bits |= RADV_CMD_FLAG_FLUSH_AND_INV_CB; - si_emit_cache_flush(cmd_buffer); +} + +enum radv_resolve_method { + RESOLVE_HW, + RESOLVE_COMPUTE, + RESOLVE_FRAGMENT, +}; + +static void radv_pick_resolve_method_images(struct radv_image *src_image, + struct radv_image *dest_image, + VkImageLayout dest_image_layout, + struct radv_cmd_buffer *cmd_buffer, + enum radv_resolve_method *method) + +{ + uint32_t queue_mask = radv_image_queue_family_mask(dest_image, + cmd_buffer->queue_family_index, + cmd_buffer->queue_family_index); + + if (src_image->vk_format == VK_FORMAT_R16G16_UNORM || + src_image->vk_format == VK_FORMAT_R16G16_SNORM) + *method = RESOLVE_COMPUTE; + else if (vk_format_is_int(src_image->vk_format)) + *method = RESOLVE_COMPUTE; + + if (radv_layout_dcc_compressed(dest_image, dest_image_layout, queue_mask)) { + *method = RESOLVE_FRAGMENT; + } else if (dest_image->surface.micro_tile_mode != src_image->surface.micro_tile_mode) { + *method = RESOLVE_COMPUTE; + } } void radv_CmdResolveImage( @@ -382,29 +381,40 @@ void radv_CmdResolveImage( struct radv_device *device = cmd_buffer->device; struct radv_meta_saved_state saved_state; VkDevice device_h = radv_device_to_handle(device); - bool use_compute_resolve = false; - + enum radv_resolve_method resolve_method = RESOLVE_HW; /* we can use the hw resolve only for single full resolves */ if (region_count == 1) { if (regions[0].srcOffset.x || regions[0].srcOffset.y || regions[0].srcOffset.z) - use_compute_resolve = true; + resolve_method = RESOLVE_COMPUTE; if (regions[0].dstOffset.x || regions[0].dstOffset.y || regions[0].dstOffset.z) - use_compute_resolve = true; + resolve_method = RESOLVE_COMPUTE; - if (regions[0].extent.width != src_image->extent.width || - regions[0].extent.height != src_image->extent.height || - regions[0].extent.depth != src_image->extent.depth) - use_compute_resolve = true; + if (regions[0].extent.width != src_image->info.width || + regions[0].extent.height != src_image->info.height || + regions[0].extent.depth != src_image->info.depth) + resolve_method = RESOLVE_COMPUTE; } else - use_compute_resolve = true; - - if (use_compute_resolve) { + resolve_method = RESOLVE_COMPUTE; + + radv_pick_resolve_method_images(src_image, dest_image, + dest_image_layout, cmd_buffer, + &resolve_method); + + if (resolve_method == RESOLVE_FRAGMENT) { + radv_meta_resolve_fragment_image(cmd_buffer, + src_image, + src_image_layout, + dest_image, + dest_image_layout, + region_count, regions); + return; + } - radv_fast_clear_flush_image_inplace(cmd_buffer, src_image); + if (resolve_method == RESOLVE_COMPUTE) { radv_meta_resolve_compute_image(cmd_buffer, src_image, src_image_layout, @@ -414,12 +424,18 @@ void radv_CmdResolveImage( return; } - radv_meta_save_graphics_reset_vport_scissor(&saved_state, cmd_buffer); + radv_meta_save(&saved_state, cmd_buffer, + RADV_META_SAVE_GRAPHICS_PIPELINE); - assert(src_image->samples > 1); - assert(dest_image->samples == 1); + assert(src_image->info.samples > 1); + if (src_image->info.samples <= 1) { + /* this causes GPU hangs if we get past here */ + fprintf(stderr, "radv: Illegal resolve operation (src not multisampled), will hang GPU."); + return; + } + assert(dest_image->info.samples == 1); - if (src_image->samples >= 16) { + if (src_image->info.samples >= 16) { /* See commit aa3f9aaf31e9056a255f9e0472ebdfdaa60abe54 for the * glBlitFramebuffer workaround for samples >= 16. */ @@ -427,12 +443,13 @@ void radv_CmdResolveImage( "samples >= 16"); } - if (src_image->array_size > 1) + if (src_image->info.array_size > 1) radv_finishme("vkCmdResolveImage: multisample array images"); - if (dest_image->surface.dcc_size) { + if (radv_image_has_dcc(dest_image)) { radv_initialize_dcc(cmd_buffer, dest_image, 0xffffffff); } + unsigned fs_key = radv_format_meta_fs_key(dest_image->vk_format); for (uint32_t r = 0; r < region_count; ++r) { const VkImageResolve *region = ®ions[r]; @@ -493,8 +510,7 @@ void radv_CmdResolveImage( .baseArrayLayer = src_base_layer + layer, .layerCount = 1, }, - }, - cmd_buffer, VK_IMAGE_USAGE_SAMPLED_BIT); + }); struct radv_image_view dest_iview; radv_image_view_init(&dest_iview, cmd_buffer->device, @@ -510,8 +526,7 @@ void radv_CmdResolveImage( .baseArrayLayer = dest_base_layer + layer, .layerCount = 1, }, - }, - cmd_buffer, VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT); + }); VkFramebuffer fb_h; radv_CreateFramebuffer(device_h, @@ -522,9 +537,9 @@ void radv_CmdResolveImage( radv_image_view_to_handle(&src_iview), radv_image_view_to_handle(&dest_iview), }, - .width = radv_minify(dest_image->extent.width, + .width = radv_minify(dest_image->info.width, region->dstSubresource.mipLevel), - .height = radv_minify(dest_image->extent.height, + .height = radv_minify(dest_image->info.height, region->dstSubresource.mipLevel), .layers = 1 }, @@ -534,7 +549,7 @@ void radv_CmdResolveImage( radv_CmdBeginRenderPass(cmd_buffer_h, &(VkRenderPassBeginInfo) { .sType = VK_STRUCTURE_TYPE_RENDER_PASS_BEGIN_INFO, - .renderPass = device->meta_state.resolve.pass, + .renderPass = device->meta_state.resolve.pass[fs_key], .framebuffer = fb_h, .renderArea = { .offset = { @@ -552,6 +567,7 @@ void radv_CmdResolveImage( VK_SUBPASS_CONTENTS_INLINE); emit_resolve(cmd_buffer, + dest_iview.vk_format, &(VkOffset2D) { .x = dstOffset.x, .y = dstOffset.y, @@ -580,6 +596,7 @@ radv_cmd_buffer_resolve_subpass(struct radv_cmd_buffer *cmd_buffer) struct radv_framebuffer *fb = cmd_buffer->state.framebuffer; const struct radv_subpass *subpass = cmd_buffer->state.subpass; struct radv_meta_saved_state saved_state; + enum radv_resolve_method resolve_method = RESOLVE_HW; /* FINISHME(perf): Skip clears for resolve attachments. * @@ -593,16 +610,45 @@ radv_cmd_buffer_resolve_subpass(struct radv_cmd_buffer *cmd_buffer) if (!subpass->has_resolve) return; - radv_meta_save_graphics_reset_vport_scissor(&saved_state, cmd_buffer); - for (uint32_t i = 0; i < subpass->color_count; ++i) { VkAttachmentReference src_att = subpass->color_attachments[i]; VkAttachmentReference dest_att = subpass->resolve_attachments[i]; + + if (src_att.attachment == VK_ATTACHMENT_UNUSED || + dest_att.attachment == VK_ATTACHMENT_UNUSED) + continue; + struct radv_image *dst_img = cmd_buffer->state.framebuffer->attachments[dest_att.attachment].attachment->image; - if (dest_att.attachment == VK_ATTACHMENT_UNUSED) + struct radv_image *src_img = cmd_buffer->state.framebuffer->attachments[src_att.attachment].attachment->image; + + radv_pick_resolve_method_images(src_img, dst_img, dest_att.layout, cmd_buffer, &resolve_method); + if (resolve_method == RESOLVE_FRAGMENT) { + break; + } + } + + if (resolve_method == RESOLVE_COMPUTE) { + radv_cmd_buffer_resolve_subpass_cs(cmd_buffer); + return; + } else if (resolve_method == RESOLVE_FRAGMENT) { + radv_cmd_buffer_resolve_subpass_fs(cmd_buffer); + return; + } + + radv_meta_save(&saved_state, cmd_buffer, + RADV_META_SAVE_GRAPHICS_PIPELINE); + + for (uint32_t i = 0; i < subpass->color_count; ++i) { + VkAttachmentReference src_att = subpass->color_attachments[i]; + VkAttachmentReference dest_att = subpass->resolve_attachments[i]; + + if (src_att.attachment == VK_ATTACHMENT_UNUSED || + dest_att.attachment == VK_ATTACHMENT_UNUSED) continue; - if (dst_img->surface.dcc_size) { + struct radv_image *dst_img = cmd_buffer->state.framebuffer->attachments[dest_att.attachment].attachment->image; + + if (radv_image_has_dcc(dst_img)) { radv_initialize_dcc(cmd_buffer, dst_img, 0xffffffff); cmd_buffer->state.attachments[dest_att.attachment].current_layout = VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL; } @@ -615,14 +661,8 @@ radv_cmd_buffer_resolve_subpass(struct radv_cmd_buffer *cmd_buffer) radv_cmd_buffer_set_subpass(cmd_buffer, &resolve_subpass, false); - /* Subpass resolves must respect the render area. We can ignore the - * render area here because vkCmdBeginRenderPass set the render area - * with 3DSTATE_DRAWING_RECTANGLE. - * - * XXX(chadv): Does the hardware really respect - * 3DSTATE_DRAWING_RECTANGLE when draing a 3DPRIM_RECTLIST? - */ emit_resolve(cmd_buffer, + dst_img->vk_format, &(VkOffset2D) { 0, 0 }, &(VkExtent2D) { fb->width, fb->height }); } @@ -630,3 +670,71 @@ radv_cmd_buffer_resolve_subpass(struct radv_cmd_buffer *cmd_buffer) cmd_buffer->state.subpass = subpass; radv_meta_restore(&saved_state, cmd_buffer); } + +/** + * Decompress CMask/FMask before resolving a multisampled source image inside a + * subpass. + */ +void +radv_decompress_resolve_subpass_src(struct radv_cmd_buffer *cmd_buffer) +{ + const struct radv_subpass *subpass = cmd_buffer->state.subpass; + struct radv_framebuffer *fb = cmd_buffer->state.framebuffer; + + for (uint32_t i = 0; i < subpass->color_count; ++i) { + VkAttachmentReference src_att = subpass->color_attachments[i]; + VkAttachmentReference dest_att = subpass->resolve_attachments[i]; + + if (src_att.attachment == VK_ATTACHMENT_UNUSED || + dest_att.attachment == VK_ATTACHMENT_UNUSED) + continue; + + struct radv_image *src_image = + fb->attachments[src_att.attachment].attachment->image; + + VkImageResolve region = {}; + region.srcSubresource.baseArrayLayer = 0; + region.srcSubresource.mipLevel = 0; + region.srcSubresource.layerCount = 1; + + radv_decompress_resolve_src(cmd_buffer, src_image, + src_att.layout, 1, ®ion); + } +} + +/** + * Decompress CMask/FMask before resolving a multisampled source image. + */ +void +radv_decompress_resolve_src(struct radv_cmd_buffer *cmd_buffer, + struct radv_image *src_image, + VkImageLayout src_image_layout, + uint32_t region_count, + const VkImageResolve *regions) +{ + for (uint32_t r = 0; r < region_count; ++r) { + const VkImageResolve *region = ®ions[r]; + const uint32_t src_base_layer = + radv_meta_get_iview_layer(src_image, ®ion->srcSubresource, + ®ion->srcOffset); + VkImageSubresourceRange range; + range.aspectMask = VK_IMAGE_ASPECT_COLOR_BIT; + range.baseMipLevel = region->srcSubresource.mipLevel; + range.levelCount = 1; + range.baseArrayLayer = src_base_layer; + range.layerCount = region->srcSubresource.layerCount; + + uint32_t queue_mask = + radv_image_queue_family_mask(src_image, + cmd_buffer->queue_family_index, + cmd_buffer->queue_family_index); + + if (radv_layout_dcc_compressed(src_image, src_image_layout, + queue_mask)) { + radv_decompress_dcc(cmd_buffer, src_image, &range); + } else { + radv_fast_clear_flush_image_inplace(cmd_buffer, + src_image, &range); + } + } +}