radv: use a linked list for physical devices
[mesa.git] / src / amd / vulkan / radv_cmd_buffer.c
index 21ef5caa8e565d3b5dcc74387ef7f972f2d92cce..d96b5649976bc35d090a8040dfc62ec149b0a091 100644 (file)
@@ -92,6 +92,10 @@ const struct radv_dynamic_state default_dynamic_state = {
                .front = 0u,
                .back = 0u,
        },
+       .line_stipple = {
+               .factor = 0u,
+               .pattern = 0u,
+       },
 };
 
 static void
@@ -212,6 +216,14 @@ radv_bind_dynamic_state(struct radv_cmd_buffer *cmd_buffer,
                }
        }
 
+       if (copy_mask & RADV_DYNAMIC_LINE_STIPPLE) {
+               if (memcmp(&dest->line_stipple, &src->line_stipple,
+                          sizeof(src->line_stipple))) {
+                       dest->line_stipple = src->line_stipple;
+                       dest_mask |= RADV_DYNAMIC_LINE_STIPPLE;
+               }
+       }
+
        cmd_buffer->state.dirty |= dest_mask;
 }
 
@@ -313,7 +325,7 @@ radv_cmd_buffer_destroy(struct radv_cmd_buffer *cmd_buffer)
                cmd_buffer->device->ws->buffer_destroy(cmd_buffer->upload.upload_bo);
        cmd_buffer->device->ws->cs_destroy(cmd_buffer->cs);
 
-       for (unsigned i = 0; i < VK_PIPELINE_BIND_POINT_RANGE_SIZE; i++)
+       for (unsigned i = 0; i < MAX_BIND_POINTS; i++)
                free(cmd_buffer->descriptors[i].push_set.set.mapped_ptr);
 
        vk_free(&cmd_buffer->pool->alloc, cmd_buffer);
@@ -352,7 +364,7 @@ radv_reset_cmd_buffer(struct radv_cmd_buffer *cmd_buffer)
 
        memset(cmd_buffer->vertex_bindings, 0, sizeof(cmd_buffer->vertex_bindings));
 
-       for (unsigned i = 0; i < VK_PIPELINE_BIND_POINT_RANGE_SIZE; i++) {
+       for (unsigned i = 0; i < MAX_BIND_POINTS; i++) {
                cmd_buffer->descriptors[i].dirty = 0;
                cmd_buffer->descriptors[i].valid = 0;
                cmd_buffer->descriptors[i].push_dirty = false;
@@ -519,6 +531,11 @@ static void
 radv_cmd_buffer_after_draw(struct radv_cmd_buffer *cmd_buffer,
                           enum radv_cmd_flush_bits flags)
 {
+       if (unlikely(cmd_buffer->device->thread_trace_bo)) {
+               radeon_emit(cmd_buffer->cs, PKT3(PKT3_EVENT_WRITE, 0, 0));
+               radeon_emit(cmd_buffer->cs, EVENT_TYPE(V_028A90_THREAD_TRACE_MARKER) | EVENT_INDEX(0));
+       }
+
        if (cmd_buffer->device->instance->debug_flags & RADV_DEBUG_SYNC_SHADERS) {
                assert(flags & (RADV_CMD_FLAG_PS_PARTIAL_FLUSH |
                                RADV_CMD_FLAG_CS_PARTIAL_FLUSH));
@@ -683,8 +700,8 @@ radv_convert_user_sample_locs(struct radv_sample_locations_state *state,
                float shifted_pos_x = user_locs[i].x - 0.5;
                float shifted_pos_y = user_locs[i].y - 0.5;
 
-               int32_t scaled_pos_x = floor(shifted_pos_x * 16);
-               int32_t scaled_pos_y = floor(shifted_pos_y * 16);
+               int32_t scaled_pos_x = floorf(shifted_pos_x * 16);
+               int32_t scaled_pos_y = floorf(shifted_pos_y * 16);
 
                sample_locs[i].x = CLAMP(scaled_pos_x, -8, 7);
                sample_locs[i].y = CLAMP(scaled_pos_y, -8, 7);
@@ -761,8 +778,6 @@ radv_compute_centroid_priority(struct radv_cmd_buffer *cmd_buffer,
 static void
 radv_emit_sample_locations(struct radv_cmd_buffer *cmd_buffer)
 {
-       struct radv_pipeline *pipeline = cmd_buffer->state.pipeline;
-       struct radv_multisample_state *ms = &pipeline->graphics.ms;
        struct radv_sample_locations_state *sample_location =
                &cmd_buffer->state.dynamic.sample_location;
        uint32_t num_samples = (uint32_t)sample_location->per_pixel;
@@ -793,10 +808,12 @@ radv_emit_sample_locations(struct radv_cmd_buffer *cmd_buffer)
                                               num_samples);
 
        /* Compute the maximum sample distance from the specified locations. */
-       for (uint32_t i = 0; i < num_samples; i++) {
-               VkOffset2D offset = sample_locs[0][i];
-               max_sample_dist = MAX2(max_sample_dist,
-                                      MAX2(abs(offset.x), abs(offset.y)));
+       for (unsigned i = 0; i < 4; ++i) {
+               for (uint32_t j = 0; j < num_samples; j++) {
+                       VkOffset2D offset = sample_locs[i][j];
+                       max_sample_dist = MAX2(max_sample_dist,
+                                              MAX2(abs(offset.x), abs(offset.y)));
+               }
        }
 
        /* Emit the specified user sample locations. */
@@ -823,13 +840,9 @@ radv_emit_sample_locations(struct radv_cmd_buffer *cmd_buffer)
        }
 
        /* Emit the maximum sample distance and the centroid priority. */
-       uint32_t pa_sc_aa_config = ms->pa_sc_aa_config;
-
-       pa_sc_aa_config &= C_028BE0_MAX_SAMPLE_DIST;
-       pa_sc_aa_config |= S_028BE0_MAX_SAMPLE_DIST(max_sample_dist);
-
-       radeon_set_context_reg_seq(cs, R_028BE0_PA_SC_AA_CONFIG, 1);
-       radeon_emit(cs, pa_sc_aa_config);
+       radeon_set_context_reg_rmw(cs, R_028BE0_PA_SC_AA_CONFIG,
+                                  S_028BE0_MAX_SAMPLE_DIST(max_sample_dist),
+                                  ~C_028BE0_MAX_SAMPLE_DIST);
 
        radeon_set_context_reg_seq(cs, R_028BD4_PA_SC_CENTROID_PRIORITY_0, 2);
        radeon_emit(cs, centroid_priority);
@@ -996,8 +1009,9 @@ radv_emit_rbplus_state(struct radv_cmd_buffer *cmd_buffer)
 
        for (unsigned i = 0; i < subpass->color_count; ++i) {
                if (subpass->color_attachments[i].attachment == VK_ATTACHMENT_UNUSED) {
-                       sx_blend_opt_control |= S_02875C_MRT0_COLOR_OPT_DISABLE(1) << (i * 4);
-                       sx_blend_opt_control |= S_02875C_MRT0_ALPHA_OPT_DISABLE(1) << (i * 4);
+                       /* We don't set the DISABLE bits, because the HW can't have holes,
+                        * so the SPI color format is set to 32-bit 1-component. */
+                       sx_ps_downconvert |= V_028754_SX_RT_EXPORT_32_R << (i * 4);
                        continue;
                }
 
@@ -1113,17 +1127,25 @@ radv_emit_rbplus_state(struct radv_cmd_buffer *cmd_buffer)
                }
        }
 
-       for (unsigned i = subpass->color_count; i < 8; ++i) {
-               sx_blend_opt_control |= S_02875C_MRT0_COLOR_OPT_DISABLE(1) << (i * 4);
-               sx_blend_opt_control |= S_02875C_MRT0_ALPHA_OPT_DISABLE(1) << (i * 4);
-       }
-       /* TODO: avoid redundantly setting context registers */
+       /* Do not set the DISABLE bits for the unused attachments, as that
+        * breaks dual source blending in SkQP and does not seem to improve
+        * performance. */
+
+       if (sx_ps_downconvert == cmd_buffer->state.last_sx_ps_downconvert &&
+           sx_blend_opt_epsilon == cmd_buffer->state.last_sx_blend_opt_epsilon &&
+           sx_blend_opt_control == cmd_buffer->state.last_sx_blend_opt_control)
+               return;
+
        radeon_set_context_reg_seq(cmd_buffer->cs, R_028754_SX_PS_DOWNCONVERT, 3);
        radeon_emit(cmd_buffer->cs, sx_ps_downconvert);
        radeon_emit(cmd_buffer->cs, sx_blend_opt_epsilon);
        radeon_emit(cmd_buffer->cs, sx_blend_opt_control);
 
        cmd_buffer->state.context_roll_without_scissor_emitted = true;
+
+       cmd_buffer->state.last_sx_ps_downconvert = sx_ps_downconvert;
+       cmd_buffer->state.last_sx_blend_opt_epsilon = sx_blend_opt_epsilon;
+       cmd_buffer->state.last_sx_blend_opt_control = sx_blend_opt_control;
 }
 
 static void
@@ -1308,6 +1330,22 @@ radv_emit_depth_bias(struct radv_cmd_buffer *cmd_buffer)
        radeon_emit(cmd_buffer->cs, bias); /* BACK OFFSET */
 }
 
+static void
+radv_emit_line_stipple(struct radv_cmd_buffer *cmd_buffer)
+{
+       struct radv_dynamic_state *d = &cmd_buffer->state.dynamic;
+       struct radv_pipeline *pipeline = cmd_buffer->state.pipeline;
+       uint32_t auto_reset_cntl = 1;
+
+       if (pipeline->graphics.topology == VK_PRIMITIVE_TOPOLOGY_LINE_STRIP)
+               auto_reset_cntl = 2;
+
+       radeon_set_context_reg(cmd_buffer->cs, R_028A0C_PA_SC_LINE_STIPPLE,
+                              S_028A0C_LINE_PATTERN(d->line_stipple.pattern) |
+                              S_028A0C_REPEAT_COUNT(d->line_stipple.factor - 1) |
+                              S_028A0C_AUTO_RESET_CNTL(auto_reset_cntl));
+}
+
 static void
 radv_emit_fb_color_state(struct radv_cmd_buffer *cmd_buffer,
                         int index,
@@ -1433,10 +1471,10 @@ radv_update_zrange_precision(struct radv_cmd_buffer *cmd_buffer,
            !radv_image_is_tc_compat_htile(image))
                return;
 
-       if (!radv_layout_has_htile(image, layout, in_render_loop,
-                                  radv_image_queue_family_mask(image,
-                                                               cmd_buffer->queue_family_index,
-                                                               cmd_buffer->queue_family_index))) {
+       if (!radv_layout_is_htile_compressed(image, layout, in_render_loop,
+                                            radv_image_queue_family_mask(image,
+                                                                         cmd_buffer->queue_family_index,
+                                                                         cmd_buffer->queue_family_index))) {
                db_z_info &= C_028040_TILE_SURFACE_ENABLE;
        }
 
@@ -1476,10 +1514,10 @@ radv_emit_fb_ds_state(struct radv_cmd_buffer *cmd_buffer,
        uint32_t db_z_info = ds->db_z_info;
        uint32_t db_stencil_info = ds->db_stencil_info;
 
-       if (!radv_layout_has_htile(image, layout, in_render_loop,
-                                  radv_image_queue_family_mask(image,
-                                                               cmd_buffer->queue_family_index,
-                                                               cmd_buffer->queue_family_index))) {
+       if (!radv_layout_is_htile_compressed(image, layout, in_render_loop,
+                                            radv_image_queue_family_mask(image,
+                                                                         cmd_buffer->queue_family_index,
+                                                                         cmd_buffer->queue_family_index))) {
                db_z_info &= C_028040_TILE_SURFACE_ENABLE;
                db_stencil_info |= S_028044_TILE_STENCIL_DISABLE(1);
        }
@@ -2011,14 +2049,7 @@ radv_emit_framebuffer_state(struct radv_cmd_buffer *cmd_buffer)
                VkImageLayout layout = subpass->depth_stencil_attachment->layout;
                bool in_render_loop = subpass->depth_stencil_attachment->in_render_loop;
                struct radv_image_view *iview = cmd_buffer->state.attachments[idx].iview;
-               struct radv_image *image = iview->image;
                radv_cs_add_buffer(cmd_buffer->device->ws, cmd_buffer->cs, cmd_buffer->state.attachments[idx].iview->bo);
-               ASSERTED uint32_t queue_mask = radv_image_queue_family_mask(image,
-                                                                               cmd_buffer->queue_family_index,
-                                                                               cmd_buffer->queue_family_index);
-               /* We currently don't support writing decompressed HTILE */
-               assert(radv_layout_has_htile(image, layout, in_render_loop, queue_mask) ==
-                      radv_layout_is_htile_compressed(image, layout, in_render_loop, queue_mask));
 
                radv_emit_fb_ds_state(cmd_buffer, &cmd_buffer->state.attachments[idx].ds, iview, layout, in_render_loop);
 
@@ -2062,7 +2093,7 @@ radv_emit_framebuffer_state(struct radv_cmd_buffer *cmd_buffer)
 }
 
 static void
-radv_emit_index_buffer(struct radv_cmd_buffer *cmd_buffer)
+radv_emit_index_buffer(struct radv_cmd_buffer *cmd_buffer, bool indirect)
 {
        struct radeon_cmdbuf *cs = cmd_buffer->cs;
        struct radv_cmd_state *state = &cmd_buffer->state;
@@ -2080,6 +2111,11 @@ radv_emit_index_buffer(struct radv_cmd_buffer *cmd_buffer)
                state->last_index_type = state->index_type;
        }
 
+       /* For the direct indexed draws we use DRAW_INDEX_2, which includes
+        * the index_va and max_index_count already. */
+       if (!indirect)
+               return;
+
        radeon_emit(cs, PKT3(PKT3_INDEX_BASE, 1, 0));
        radeon_emit(cs, state->index_va);
        radeon_emit(cs, state->index_va >> 32);
@@ -2188,6 +2224,9 @@ radv_cmd_buffer_flush_dynamic_state(struct radv_cmd_buffer *cmd_buffer)
        if (states & RADV_CMD_DIRTY_DYNAMIC_SAMPLE_LOCATIONS)
                radv_emit_sample_locations(cmd_buffer);
 
+       if (states & RADV_CMD_DIRTY_DYNAMIC_LINE_STIPPLE)
+               radv_emit_line_stipple(cmd_buffer);
+
        cmd_buffer->state.dirty &= ~states;
 }
 
@@ -3362,6 +3401,9 @@ VkResult radv_BeginCommandBuffer(
        cmd_buffer->state.last_vertex_offset = -1;
        cmd_buffer->state.last_first_instance = -1;
        cmd_buffer->state.predication_type = -1;
+       cmd_buffer->state.last_sx_ps_downconvert = -1;
+       cmd_buffer->state.last_sx_blend_opt_epsilon = -1;
+       cmd_buffer->state.last_sx_blend_opt_control = -1;
        cmd_buffer->usage_flags = pBeginInfo->flags;
 
        if (cmd_buffer->level == VK_COMMAND_BUFFER_LEVEL_SECONDARY &&
@@ -3385,14 +3427,10 @@ VkResult radv_BeginCommandBuffer(
                radv_cmd_buffer_set_subpass(cmd_buffer, subpass);
        }
 
-       if (unlikely(cmd_buffer->device->trace_bo)) {
-               struct radv_device *device = cmd_buffer->device;
-
-               radv_cs_add_buffer(device->ws, cmd_buffer->cs,
-                                  device->trace_bo);
-
+       if (unlikely(cmd_buffer->device->trace_bo))
                radv_cmd_buffer_trace_emit(cmd_buffer);
-       }
+
+       radv_describe_begin_cmd_buffer(cmd_buffer);
 
        cmd_buffer->status = RADV_CMD_BUFFER_STATUS_RECORDING;
 
@@ -3415,19 +3453,22 @@ void radv_CmdBindVertexBuffers(
 
        assert(firstBinding + bindingCount <= MAX_VBS);
        for (uint32_t i = 0; i < bindingCount; i++) {
+               RADV_FROM_HANDLE(radv_buffer, buffer, pBuffers[i]);
                uint32_t idx = firstBinding + i;
 
                if (!changed &&
-                   (vb[idx].buffer != radv_buffer_from_handle(pBuffers[i]) ||
+                   (vb[idx].buffer != buffer ||
                     vb[idx].offset != pOffsets[i])) {
                        changed = true;
                }
 
-               vb[idx].buffer = radv_buffer_from_handle(pBuffers[i]);
+               vb[idx].buffer = buffer;
                vb[idx].offset = pOffsets[i];
 
-               radv_cs_add_buffer(cmd_buffer->device->ws, cmd_buffer->cs,
-                                  vb[idx].buffer->bo);
+               if (buffer) {
+                       radv_cs_add_buffer(cmd_buffer->device->ws,
+                                          cmd_buffer->cs, vb[idx].buffer->bo);
+               }
        }
 
        if (!changed) {
@@ -3510,7 +3551,7 @@ radv_bind_descriptor_set(struct radv_cmd_buffer *cmd_buffer,
        assert(!(set->layout->flags & VK_DESCRIPTOR_SET_LAYOUT_CREATE_PUSH_DESCRIPTOR_BIT_KHR));
 
        if (!cmd_buffer->device->use_global_bo_list) {
-               for (unsigned j = 0; j < set->layout->buffer_count; ++j)
+               for (unsigned j = 0; j < set->buffer_count; ++j)
                        if (set->descriptors[j])
                                radv_cs_add_buffer(ws, cmd_buffer->cs, set->descriptors[j]);
        }
@@ -3747,6 +3788,8 @@ VkResult radv_EndCommandBuffer(
         */
        si_cp_dma_wait_for_idle(cmd_buffer);
 
+       radv_describe_end_cmd_buffer(cmd_buffer);
+
        vk_free(&cmd_buffer->pool->alloc, cmd_buffer->state.attachments);
        vk_free(&cmd_buffer->pool->alloc, cmd_buffer->state.subpass_sample_locs);
 
@@ -3830,9 +3873,7 @@ void radv_CmdBindPipeline(
                /* Prefetch all pipeline shaders at first draw time. */
                cmd_buffer->state.prefetch_L2_mask |= RADV_PREFETCH_SHADERS;
 
-               if ((cmd_buffer->device->physical_device->rad_info.family == CHIP_NAVI10 ||
-                    cmd_buffer->device->physical_device->rad_info.family == CHIP_NAVI12 ||
-                    cmd_buffer->device->physical_device->rad_info.family == CHIP_NAVI14) &&
+               if (cmd_buffer->device->physical_device->rad_info.chip_class == GFX10 &&
                    cmd_buffer->state.emitted_pipeline &&
                    radv_pipeline_has_ngg(cmd_buffer->state.emitted_pipeline) &&
                    !radv_pipeline_has_ngg(cmd_buffer->state.pipeline)) {
@@ -4090,6 +4131,20 @@ void radv_CmdSetSampleLocationsEXT(
        state->dirty |= RADV_CMD_DIRTY_DYNAMIC_SAMPLE_LOCATIONS;
 }
 
+void radv_CmdSetLineStippleEXT(
+       VkCommandBuffer                             commandBuffer,
+       uint32_t                                    lineStippleFactor,
+       uint16_t                                    lineStipplePattern)
+{
+       RADV_FROM_HANDLE(radv_cmd_buffer, cmd_buffer, commandBuffer);
+       struct radv_cmd_state *state = &cmd_buffer->state;
+
+       state->dynamic.line_stipple.factor = lineStippleFactor;
+       state->dynamic.line_stipple.pattern = lineStipplePattern;
+
+       state->dirty |= RADV_CMD_DIRTY_DYNAMIC_LINE_STIPPLE;
+}
+
 void radv_CmdExecuteCommands(
        VkCommandBuffer                             commandBuffer,
        uint32_t                                    commandBufferCount,
@@ -4172,6 +4227,9 @@ void radv_CmdExecuteCommands(
                primary->state.last_first_instance = secondary->state.last_first_instance;
                primary->state.last_num_instances = secondary->state.last_num_instances;
                primary->state.last_vertex_offset = secondary->state.last_vertex_offset;
+               primary->state.last_sx_ps_downconvert = secondary->state.last_sx_ps_downconvert;
+               primary->state.last_sx_blend_opt_epsilon = secondary->state.last_sx_blend_opt_epsilon;
+               primary->state.last_sx_blend_opt_control = secondary->state.last_sx_blend_opt_control;
 
                if (secondary->state.last_index_type != -1) {
                        primary->state.last_index_type =
@@ -4291,6 +4349,8 @@ radv_cmd_buffer_begin_subpass(struct radv_cmd_buffer *cmd_buffer,
 
        radv_cmd_buffer_set_subpass(cmd_buffer, subpass);
 
+       radv_describe_barrier_start(cmd_buffer, RGP_BARRIER_EXTERNAL_RENDER_PASS_SYNC);
+
        for (uint32_t i = 0; i < subpass->attachment_count; ++i) {
                const uint32_t a = subpass->attachments[i].attachment;
                if (a == VK_ATTACHMENT_UNUSED)
@@ -4301,6 +4361,8 @@ radv_cmd_buffer_begin_subpass(struct radv_cmd_buffer *cmd_buffer,
                                                     true);
        }
 
+       radv_describe_barrier_end(cmd_buffer);
+
        radv_cmd_buffer_clear_subpass(cmd_buffer);
 
        assert(cmd_buffer->cs->cdw <= cdw_max);
@@ -4315,6 +4377,8 @@ radv_cmd_buffer_end_subpass(struct radv_cmd_buffer *cmd_buffer)
 
        radv_cmd_buffer_resolve_subpass(cmd_buffer);
 
+       radv_describe_barrier_start(cmd_buffer, RGP_BARRIER_EXTERNAL_RENDER_PASS_SYNC);
+
        for (uint32_t i = 0; i < subpass->attachment_count; ++i) {
                const uint32_t a = subpass->attachments[i].attachment;
                if (a == VK_ATTACHMENT_UNUSED)
@@ -4328,14 +4392,14 @@ radv_cmd_buffer_end_subpass(struct radv_cmd_buffer *cmd_buffer)
                struct radv_subpass_attachment att = { a, layout, stencil_layout };
                radv_handle_subpass_image_transition(cmd_buffer, att, false);
        }
+
+       radv_describe_barrier_end(cmd_buffer);
 }
 
-void radv_CmdBeginRenderPass(
-       VkCommandBuffer                             commandBuffer,
-       const VkRenderPassBeginInfo*                pRenderPassBegin,
-       VkSubpassContents                           contents)
+void
+radv_cmd_buffer_begin_render_pass(struct radv_cmd_buffer *cmd_buffer,
+                                 const VkRenderPassBeginInfo *pRenderPassBegin)
 {
-       RADV_FROM_HANDLE(radv_cmd_buffer, cmd_buffer, commandBuffer);
        RADV_FROM_HANDLE(radv_render_pass, pass, pRenderPassBegin->renderPass);
        RADV_FROM_HANDLE(radv_framebuffer, framebuffer, pRenderPassBegin->framebuffer);
        VkResult result;
@@ -4351,6 +4415,16 @@ void radv_CmdBeginRenderPass(
        result = radv_cmd_state_setup_sample_locations(cmd_buffer, pass, pRenderPassBegin);
        if (result != VK_SUCCESS)
                return;
+}
+
+void radv_CmdBeginRenderPass(
+       VkCommandBuffer                             commandBuffer,
+       const VkRenderPassBeginInfo*                pRenderPassBegin,
+       VkSubpassContents                           contents)
+{
+       RADV_FROM_HANDLE(radv_cmd_buffer, cmd_buffer, commandBuffer);
+
+       radv_cmd_buffer_begin_render_pass(cmd_buffer, pRenderPassBegin);
 
        radv_cmd_buffer_begin_subpass(cmd_buffer, 0);
 }
@@ -4664,7 +4738,7 @@ radv_emit_all_graphics_states(struct radv_cmd_buffer *cmd_buffer,
 
        if (info->indexed) {
                if (cmd_buffer->state.dirty & RADV_CMD_DIRTY_INDEX_BUFFER)
-                       radv_emit_index_buffer(cmd_buffer);
+                       radv_emit_index_buffer(cmd_buffer, info->indirect);
        } else {
                /* On GFX7 and later, non-indexed draws overwrite VGT_INDEX_TYPE,
                 * so the state must be re-emitted before the next indexed
@@ -4713,6 +4787,8 @@ radv_draw(struct radv_cmd_buffer *cmd_buffer,
                        return;
        }
 
+       radv_describe_draw(cmd_buffer);
+
        /* Use optimal packet order based on whether we need to sync the
         * pipeline.
         */
@@ -5088,6 +5164,8 @@ radv_dispatch(struct radv_cmd_buffer *cmd_buffer,
        bool pipeline_is_dirty = pipeline &&
                                 pipeline != cmd_buffer->state.emitted_compute_pipeline;
 
+       radv_describe_dispatch(cmd_buffer, 8, 8, 8);
+
        if (cmd_buffer->state.flush_bits & (RADV_CMD_FLAG_FLUSH_AND_INV_CB |
                                            RADV_CMD_FLAG_FLUSH_AND_INV_DB |
                                            RADV_CMD_FLAG_PS_PARTIAL_FLUSH |
@@ -5198,15 +5276,9 @@ void radv_unaligned_dispatch(
        radv_dispatch(cmd_buffer, &info);
 }
 
-void radv_CmdEndRenderPass(
-       VkCommandBuffer                             commandBuffer)
+void
+radv_cmd_buffer_end_render_pass(struct radv_cmd_buffer *cmd_buffer)
 {
-       RADV_FROM_HANDLE(radv_cmd_buffer, cmd_buffer, commandBuffer);
-
-       radv_subpass_barrier(cmd_buffer, &cmd_buffer->state.pass->end_barrier);
-
-       radv_cmd_buffer_end_subpass(cmd_buffer);
-
        vk_free(&cmd_buffer->pool->alloc, cmd_buffer->state.attachments);
        vk_free(&cmd_buffer->pool->alloc, cmd_buffer->state.subpass_sample_locs);
 
@@ -5217,6 +5289,18 @@ void radv_CmdEndRenderPass(
        cmd_buffer->state.subpass_sample_locs = NULL;
 }
 
+void radv_CmdEndRenderPass(
+       VkCommandBuffer                             commandBuffer)
+{
+       RADV_FROM_HANDLE(radv_cmd_buffer, cmd_buffer, commandBuffer);
+
+       radv_subpass_barrier(cmd_buffer, &cmd_buffer->state.pass->end_barrier);
+
+       radv_cmd_buffer_end_subpass(cmd_buffer);
+
+       radv_cmd_buffer_end_render_pass(cmd_buffer);
+}
+
 void radv_CmdEndRenderPass2(
     VkCommandBuffer                             commandBuffer,
     const VkSubpassEndInfo*                     pSubpassEndInfo)
@@ -5241,10 +5325,14 @@ static void radv_initialize_htile(struct radv_cmd_buffer *cmd_buffer,
        struct radv_cmd_state *state = &cmd_buffer->state;
        uint32_t htile_value = vk_format_is_stencil(image->vk_format) ? 0xfffff30f : 0xfffc000f;
        VkClearDepthStencilValue value = {};
+       struct radv_barrier_data barrier = {};
 
        state->flush_bits |= RADV_CMD_FLAG_FLUSH_AND_INV_DB |
                             RADV_CMD_FLAG_FLUSH_AND_INV_DB_META;
 
+       barrier.layout_transitions.init_mask_ram = 1;
+       radv_describe_layout_transition(cmd_buffer, &barrier);
+
        state->flush_bits |= radv_clear_htile(cmd_buffer, image, range, htile_value);
 
        state->flush_bits |= RADV_CMD_FLAG_FLUSH_AND_INV_DB_META;
@@ -5288,8 +5376,8 @@ static void radv_handle_depth_image_transition(struct radv_cmd_buffer *cmd_buffe
                cmd_buffer->state.flush_bits |= RADV_CMD_FLAG_FLUSH_AND_INV_DB |
                                                RADV_CMD_FLAG_FLUSH_AND_INV_DB_META;
 
-               radv_decompress_depth_image_inplace(cmd_buffer, image, range,
-                                                   sample_locs);
+               radv_decompress_depth_stencil(cmd_buffer, image, range,
+                                             sample_locs);
 
                cmd_buffer->state.flush_bits |= RADV_CMD_FLAG_FLUSH_AND_INV_DB |
                                                RADV_CMD_FLAG_FLUSH_AND_INV_DB_META;
@@ -5302,10 +5390,14 @@ static void radv_initialise_cmask(struct radv_cmd_buffer *cmd_buffer,
                                  uint32_t value)
 {
        struct radv_cmd_state *state = &cmd_buffer->state;
+       struct radv_barrier_data barrier = {};
 
        state->flush_bits |= RADV_CMD_FLAG_FLUSH_AND_INV_CB |
                            RADV_CMD_FLAG_FLUSH_AND_INV_CB_META;
 
+       barrier.layout_transitions.init_mask_ram = 1;
+       radv_describe_layout_transition(cmd_buffer, &barrier);
+
        state->flush_bits |= radv_clear_cmask(cmd_buffer, image, range, value);
 
        state->flush_bits |= RADV_CMD_FLAG_FLUSH_AND_INV_CB_META;
@@ -5324,10 +5416,14 @@ void radv_initialize_fmask(struct radv_cmd_buffer *cmd_buffer,
        };
        uint32_t log2_samples = util_logbase2(image->info.samples);
        uint32_t value = fmask_clear_values[log2_samples];
+       struct radv_barrier_data barrier = {};
 
        state->flush_bits |= RADV_CMD_FLAG_FLUSH_AND_INV_CB |
                             RADV_CMD_FLAG_FLUSH_AND_INV_CB_META;
 
+       barrier.layout_transitions.init_mask_ram = 1;
+       radv_describe_layout_transition(cmd_buffer, &barrier);
+
        state->flush_bits |= radv_clear_fmask(cmd_buffer, image, range, value);
 
        state->flush_bits |= RADV_CMD_FLAG_FLUSH_AND_INV_CB_META;
@@ -5338,11 +5434,15 @@ void radv_initialize_dcc(struct radv_cmd_buffer *cmd_buffer,
                         const VkImageSubresourceRange *range, uint32_t value)
 {
        struct radv_cmd_state *state = &cmd_buffer->state;
+       struct radv_barrier_data barrier = {};
        unsigned size = 0;
 
        state->flush_bits |= RADV_CMD_FLAG_FLUSH_AND_INV_CB |
                             RADV_CMD_FLAG_FLUSH_AND_INV_CB_META;
 
+       barrier.layout_transitions.init_mask_ram = 1;
+       radv_describe_layout_transition(cmd_buffer, &barrier);
+
        state->flush_bits |= radv_clear_dcc(cmd_buffer, image, range, value);
 
        if (cmd_buffer->device->physical_device->rad_info.chip_class == GFX8) {
@@ -5483,8 +5583,13 @@ static void radv_handle_color_image_transition(struct radv_cmd_buffer *cmd_buffe
                if (fce_eliminate || fmask_expand)
                        radv_fast_clear_flush_image_inplace(cmd_buffer, image, range);
 
-               if (fmask_expand)
+               if (fmask_expand) {
+                       struct radv_barrier_data barrier = {};
+                       barrier.layout_transitions.fmask_color_expand = 1;
+                       radv_describe_layout_transition(cmd_buffer, &barrier);
+
                        radv_expand_fmask_image_inplace(cmd_buffer, image, range);
+               }
        }
 }
 
@@ -5546,6 +5651,7 @@ static void radv_handle_image_transition(struct radv_cmd_buffer *cmd_buffer,
 }
 
 struct radv_barrier_info {
+       enum rgp_barrier_reason reason;
        uint32_t eventCount;
        const VkEvent *pEvents;
        VkPipelineStageFlags srcStageMask;
@@ -5566,6 +5672,8 @@ radv_barrier(struct radv_cmd_buffer *cmd_buffer,
        enum radv_cmd_flush_bits src_flush_bits = 0;
        enum radv_cmd_flush_bits dst_flush_bits = 0;
 
+       radv_describe_barrier_start(cmd_buffer, info->reason);
+
        for (unsigned i = 0; i < info->eventCount; ++i) {
                RADV_FROM_HANDLE(radv_event, event, info->pEvents[i]);
                uint64_t va = radv_buffer_get_va(event->bo);
@@ -5653,6 +5761,8 @@ radv_barrier(struct radv_cmd_buffer *cmd_buffer,
                si_cp_dma_wait_for_idle(cmd_buffer);
 
        cmd_buffer->state.flush_bits |= dst_flush_bits;
+
+       radv_describe_barrier_end(cmd_buffer);
 }
 
 void radv_CmdPipelineBarrier(
@@ -5670,6 +5780,7 @@ void radv_CmdPipelineBarrier(
        RADV_FROM_HANDLE(radv_cmd_buffer, cmd_buffer, commandBuffer);
        struct radv_barrier_info info;
 
+       info.reason = RGP_BARRIER_EXTERNAL_CMD_PIPELINE_BARRIER;
        info.eventCount = 0;
        info.pEvents = NULL;
        info.srcStageMask = srcStageMask;
@@ -5781,6 +5892,7 @@ void radv_CmdWaitEvents(VkCommandBuffer commandBuffer,
        RADV_FROM_HANDLE(radv_cmd_buffer, cmd_buffer, commandBuffer);
        struct radv_barrier_info info;
 
+       info.reason = RGP_BARRIER_EXTERNAL_CMD_WAIT_EVENTS;
        info.eventCount = eventCount;
        info.pEvents = pEvents;
        info.srcStageMask = 0;
@@ -5904,7 +6016,12 @@ void radv_CmdBindTransformFeedbackBuffersEXT(
 
                sb[idx].buffer = radv_buffer_from_handle(pBuffers[i]);
                sb[idx].offset = pOffsets[i];
-               sb[idx].size = pSizes[i];
+
+               if (!pSizes || pSizes[i] == VK_WHOLE_SIZE) {
+                       sb[idx].size = sb[idx].buffer->size - sb[idx].offset;
+               } else {
+                       sb[idx].size = pSizes[i];
+               }
 
                radv_cs_add_buffer(cmd_buffer->device->ws, cmd_buffer->cs,
                                   sb[idx].buffer->bo);