radv: implement missing VK_ACCESS_MEMORY_{READ,WRITE}_BIT
[mesa.git] / src / amd / vulkan / radv_cmd_buffer.c
index 549b7e8843d8b1cdebdbc99e456af3dc45442973..df0191f28d628a1908efa0296f84d7f87b4a0d9d 100644 (file)
@@ -92,6 +92,10 @@ const struct radv_dynamic_state default_dynamic_state = {
                .front = 0u,
                .back = 0u,
        },
+       .line_stipple = {
+               .factor = 0u,
+               .pattern = 0u,
+       },
 };
 
 static void
@@ -212,6 +216,14 @@ radv_bind_dynamic_state(struct radv_cmd_buffer *cmd_buffer,
                }
        }
 
+       if (copy_mask & RADV_DYNAMIC_LINE_STIPPLE) {
+               if (memcmp(&dest->line_stipple, &src->line_stipple,
+                          sizeof(src->line_stipple))) {
+                       dest->line_stipple = src->line_stipple;
+                       dest_mask |= RADV_DYNAMIC_LINE_STIPPLE;
+               }
+       }
+
        cmd_buffer->state.dirty |= dest_mask;
 }
 
@@ -265,22 +277,15 @@ static VkResult radv_create_cmd_buffer(
        if (cmd_buffer == NULL)
                return vk_error(device->instance, VK_ERROR_OUT_OF_HOST_MEMORY);
 
-       cmd_buffer->_loader_data.loaderMagic = ICD_LOADER_MAGIC;
+       vk_object_base_init(&device->vk, &cmd_buffer->base,
+                           VK_OBJECT_TYPE_COMMAND_BUFFER);
+
        cmd_buffer->device = device;
        cmd_buffer->pool = pool;
        cmd_buffer->level = level;
 
-       if (pool) {
-               list_addtail(&cmd_buffer->pool_link, &pool->cmd_buffers);
-               cmd_buffer->queue_family_index = pool->queue_family_index;
-
-       } else {
-               /* Init the pool_link so we can safely call list_del when we destroy
-                * the command buffer
-                */
-               list_inithead(&cmd_buffer->pool_link);
-               cmd_buffer->queue_family_index = RADV_QUEUE_GENERAL;
-       }
+       list_addtail(&cmd_buffer->pool_link, &pool->cmd_buffers);
+       cmd_buffer->queue_family_index = pool->queue_family_index;
 
        ring = radv_queue_family_to_ring(cmd_buffer->queue_family_index);
 
@@ -313,9 +318,11 @@ radv_cmd_buffer_destroy(struct radv_cmd_buffer *cmd_buffer)
                cmd_buffer->device->ws->buffer_destroy(cmd_buffer->upload.upload_bo);
        cmd_buffer->device->ws->cs_destroy(cmd_buffer->cs);
 
-       for (unsigned i = 0; i < VK_PIPELINE_BIND_POINT_RANGE_SIZE; i++)
+       for (unsigned i = 0; i < MAX_BIND_POINTS; i++)
                free(cmd_buffer->descriptors[i].push_set.set.mapped_ptr);
 
+       vk_object_base_finish(&cmd_buffer->base);
+
        vk_free(&cmd_buffer->pool->alloc, cmd_buffer);
 }
 
@@ -352,7 +359,7 @@ radv_reset_cmd_buffer(struct radv_cmd_buffer *cmd_buffer)
 
        memset(cmd_buffer->vertex_bindings, 0, sizeof(cmd_buffer->vertex_bindings));
 
-       for (unsigned i = 0; i < VK_PIPELINE_BIND_POINT_RANGE_SIZE; i++) {
+       for (unsigned i = 0; i < MAX_BIND_POINTS; i++) {
                cmd_buffer->descriptors[i].dirty = 0;
                cmd_buffer->descriptors[i].valid = 0;
                cmd_buffer->descriptors[i].push_dirty = false;
@@ -519,6 +526,11 @@ static void
 radv_cmd_buffer_after_draw(struct radv_cmd_buffer *cmd_buffer,
                           enum radv_cmd_flush_bits flags)
 {
+       if (unlikely(cmd_buffer->device->thread_trace_bo)) {
+               radeon_emit(cmd_buffer->cs, PKT3(PKT3_EVENT_WRITE, 0, 0));
+               radeon_emit(cmd_buffer->cs, EVENT_TYPE(V_028A90_THREAD_TRACE_MARKER) | EVENT_INDEX(0));
+       }
+
        if (cmd_buffer->device->instance->debug_flags & RADV_DEBUG_SYNC_SHADERS) {
                assert(flags & (RADV_CMD_FLAG_PS_PARTIAL_FLUSH |
                                RADV_CMD_FLAG_CS_PARTIAL_FLUSH));
@@ -683,8 +695,8 @@ radv_convert_user_sample_locs(struct radv_sample_locations_state *state,
                float shifted_pos_x = user_locs[i].x - 0.5;
                float shifted_pos_y = user_locs[i].y - 0.5;
 
-               int32_t scaled_pos_x = floor(shifted_pos_x * 16);
-               int32_t scaled_pos_y = floor(shifted_pos_y * 16);
+               int32_t scaled_pos_x = floorf(shifted_pos_x * 16);
+               int32_t scaled_pos_y = floorf(shifted_pos_y * 16);
 
                sample_locs[i].x = CLAMP(scaled_pos_x, -8, 7);
                sample_locs[i].y = CLAMP(scaled_pos_y, -8, 7);
@@ -761,8 +773,6 @@ radv_compute_centroid_priority(struct radv_cmd_buffer *cmd_buffer,
 static void
 radv_emit_sample_locations(struct radv_cmd_buffer *cmd_buffer)
 {
-       struct radv_pipeline *pipeline = cmd_buffer->state.pipeline;
-       struct radv_multisample_state *ms = &pipeline->graphics.ms;
        struct radv_sample_locations_state *sample_location =
                &cmd_buffer->state.dynamic.sample_location;
        uint32_t num_samples = (uint32_t)sample_location->per_pixel;
@@ -793,10 +803,12 @@ radv_emit_sample_locations(struct radv_cmd_buffer *cmd_buffer)
                                               num_samples);
 
        /* Compute the maximum sample distance from the specified locations. */
-       for (uint32_t i = 0; i < num_samples; i++) {
-               VkOffset2D offset = sample_locs[0][i];
-               max_sample_dist = MAX2(max_sample_dist,
-                                      MAX2(abs(offset.x), abs(offset.y)));
+       for (unsigned i = 0; i < 4; ++i) {
+               for (uint32_t j = 0; j < num_samples; j++) {
+                       VkOffset2D offset = sample_locs[i][j];
+                       max_sample_dist = MAX2(max_sample_dist,
+                                              MAX2(abs(offset.x), abs(offset.y)));
+               }
        }
 
        /* Emit the specified user sample locations. */
@@ -823,13 +835,9 @@ radv_emit_sample_locations(struct radv_cmd_buffer *cmd_buffer)
        }
 
        /* Emit the maximum sample distance and the centroid priority. */
-       uint32_t pa_sc_aa_config = ms->pa_sc_aa_config;
-
-       pa_sc_aa_config &= C_028BE0_MAX_SAMPLE_DIST;
-       pa_sc_aa_config |= S_028BE0_MAX_SAMPLE_DIST(max_sample_dist);
-
-       radeon_set_context_reg_seq(cs, R_028BE0_PA_SC_AA_CONFIG, 1);
-       radeon_emit(cs, pa_sc_aa_config);
+       radeon_set_context_reg_rmw(cs, R_028BE0_PA_SC_AA_CONFIG,
+                                  S_028BE0_MAX_SAMPLE_DIST(max_sample_dist),
+                                  ~C_028BE0_MAX_SAMPLE_DIST);
 
        radeon_set_context_reg_seq(cs, R_028BD4_PA_SC_CENTROID_PRIORITY_0, 2);
        radeon_emit(cs, centroid_priority);
@@ -1118,13 +1126,21 @@ radv_emit_rbplus_state(struct radv_cmd_buffer *cmd_buffer)
         * breaks dual source blending in SkQP and does not seem to improve
         * performance. */
 
-       /* TODO: avoid redundantly setting context registers */
+       if (sx_ps_downconvert == cmd_buffer->state.last_sx_ps_downconvert &&
+           sx_blend_opt_epsilon == cmd_buffer->state.last_sx_blend_opt_epsilon &&
+           sx_blend_opt_control == cmd_buffer->state.last_sx_blend_opt_control)
+               return;
+
        radeon_set_context_reg_seq(cmd_buffer->cs, R_028754_SX_PS_DOWNCONVERT, 3);
        radeon_emit(cmd_buffer->cs, sx_ps_downconvert);
        radeon_emit(cmd_buffer->cs, sx_blend_opt_epsilon);
        radeon_emit(cmd_buffer->cs, sx_blend_opt_control);
 
        cmd_buffer->state.context_roll_without_scissor_emitted = true;
+
+       cmd_buffer->state.last_sx_ps_downconvert = sx_ps_downconvert;
+       cmd_buffer->state.last_sx_blend_opt_epsilon = sx_blend_opt_epsilon;
+       cmd_buffer->state.last_sx_blend_opt_control = sx_blend_opt_control;
 }
 
 static void
@@ -1250,7 +1266,7 @@ radv_emit_line_width(struct radv_cmd_buffer *cmd_buffer)
        unsigned width = cmd_buffer->state.dynamic.line_width * 8;
 
        radeon_set_context_reg(cmd_buffer->cs, R_028A08_PA_SU_LINE_CNTL,
-                              S_028A08_WIDTH(CLAMP(width, 0, 0xFFF)));
+                              S_028A08_WIDTH(CLAMP(width, 0, 0xFFFF)));
 }
 
 static void
@@ -1309,6 +1325,22 @@ radv_emit_depth_bias(struct radv_cmd_buffer *cmd_buffer)
        radeon_emit(cmd_buffer->cs, bias); /* BACK OFFSET */
 }
 
+static void
+radv_emit_line_stipple(struct radv_cmd_buffer *cmd_buffer)
+{
+       struct radv_dynamic_state *d = &cmd_buffer->state.dynamic;
+       struct radv_pipeline *pipeline = cmd_buffer->state.pipeline;
+       uint32_t auto_reset_cntl = 1;
+
+       if (pipeline->graphics.topology == VK_PRIMITIVE_TOPOLOGY_LINE_STRIP)
+               auto_reset_cntl = 2;
+
+       radeon_set_context_reg(cmd_buffer->cs, R_028A0C_PA_SC_LINE_STIPPLE,
+                              S_028A0C_LINE_PATTERN(d->line_stipple.pattern) |
+                              S_028A0C_REPEAT_COUNT(d->line_stipple.factor - 1) |
+                              S_028A0C_AUTO_RESET_CNTL(auto_reset_cntl));
+}
+
 static void
 radv_emit_fb_color_state(struct radv_cmd_buffer *cmd_buffer,
                         int index,
@@ -1328,6 +1360,13 @@ radv_emit_fb_color_state(struct radv_cmd_buffer *cmd_buffer,
                cb_color_info &= C_028C70_DCC_ENABLE;
        }
 
+       if (!radv_layout_can_fast_clear(image, layout, in_render_loop,
+                                       radv_image_queue_family_mask(image,
+                                                                    cmd_buffer->queue_family_index,
+                                                                    cmd_buffer->queue_family_index))) {
+               cb_color_info &= C_028C70_COMPRESSION;
+       }
+
        if (radv_image_is_tc_compat_cmask(image) &&
            (radv_is_fmask_decompress_pipeline(cmd_buffer) ||
             radv_is_dcc_decompress_pipeline(cmd_buffer))) {
@@ -1337,6 +1376,19 @@ radv_emit_fb_color_state(struct radv_cmd_buffer *cmd_buffer,
                cb_color_info &= C_028C70_FMASK_COMPRESS_1FRAG_ONLY;
        }
 
+       if (radv_image_has_fmask(image) &&
+           (radv_is_fmask_decompress_pipeline(cmd_buffer) ||
+            radv_is_hw_resolve_pipeline(cmd_buffer))) {
+               /* Make sure FMASK is enabled if it has been cleared because:
+                *
+                * 1) it's required for FMASK_DECOMPRESS operations to avoid
+                * GPU hangs
+                * 2) it's necessary for CB_RESOLVE which can read compressed
+                * FMASK data anyways.
+                */
+               cb_color_info |= S_028C70_COMPRESSION(1);
+       }
+
        if (cmd_buffer->device->physical_device->rad_info.chip_class >= GFX10) {
                        radeon_set_context_reg_seq(cmd_buffer->cs, R_028C60_CB_COLOR0_BASE + index * 0x3c, 11);
                        radeon_emit(cmd_buffer->cs, cb->cb_color_base);
@@ -1434,10 +1486,10 @@ radv_update_zrange_precision(struct radv_cmd_buffer *cmd_buffer,
            !radv_image_is_tc_compat_htile(image))
                return;
 
-       if (!radv_layout_has_htile(image, layout, in_render_loop,
-                                  radv_image_queue_family_mask(image,
-                                                               cmd_buffer->queue_family_index,
-                                                               cmd_buffer->queue_family_index))) {
+       if (!radv_layout_is_htile_compressed(image, layout, in_render_loop,
+                                            radv_image_queue_family_mask(image,
+                                                                         cmd_buffer->queue_family_index,
+                                                                         cmd_buffer->queue_family_index))) {
                db_z_info &= C_028040_TILE_SURFACE_ENABLE;
        }
 
@@ -1477,10 +1529,10 @@ radv_emit_fb_ds_state(struct radv_cmd_buffer *cmd_buffer,
        uint32_t db_z_info = ds->db_z_info;
        uint32_t db_stencil_info = ds->db_stencil_info;
 
-       if (!radv_layout_has_htile(image, layout, in_render_loop,
-                                  radv_image_queue_family_mask(image,
-                                                               cmd_buffer->queue_family_index,
-                                                               cmd_buffer->queue_family_index))) {
+       if (!radv_layout_is_htile_compressed(image, layout, in_render_loop,
+                                            radv_image_queue_family_mask(image,
+                                                                         cmd_buffer->queue_family_index,
+                                                                         cmd_buffer->queue_family_index))) {
                db_z_info &= C_028040_TILE_SURFACE_ENABLE;
                db_stencil_info |= S_028044_TILE_STENCIL_DISABLE(1);
        }
@@ -1771,7 +1823,7 @@ radv_load_ds_clear_metadata(struct radv_cmd_buffer *cmd_buffer,
        uint32_t reg = R_028028_DB_STENCIL_CLEAR + 4 * reg_offset;
 
        if (cmd_buffer->device->physical_device->rad_info.has_load_ctx_reg_pkt) {
-               radeon_emit(cs, PKT3(PKT3_LOAD_CONTEXT_REG, 3, 0));
+               radeon_emit(cs, PKT3(PKT3_LOAD_CONTEXT_REG_INDEX, 3, 0));
                radeon_emit(cs, va);
                radeon_emit(cs, va >> 32);
                radeon_emit(cs, (reg - SI_CONTEXT_REG_OFFSET) >> 2);
@@ -1955,7 +2007,7 @@ radv_load_color_clear_metadata(struct radv_cmd_buffer *cmd_buffer,
        uint32_t reg = R_028C8C_CB_COLOR0_CLEAR_WORD0 + cb_idx * 0x3c;
 
        if (cmd_buffer->device->physical_device->rad_info.has_load_ctx_reg_pkt) {
-               radeon_emit(cs, PKT3(PKT3_LOAD_CONTEXT_REG, 3, cmd_buffer->state.predicating));
+               radeon_emit(cs, PKT3(PKT3_LOAD_CONTEXT_REG_INDEX, 3, cmd_buffer->state.predicating));
                radeon_emit(cs, va);
                radeon_emit(cs, va >> 32);
                radeon_emit(cs, (reg - SI_CONTEXT_REG_OFFSET) >> 2);
@@ -2012,14 +2064,7 @@ radv_emit_framebuffer_state(struct radv_cmd_buffer *cmd_buffer)
                VkImageLayout layout = subpass->depth_stencil_attachment->layout;
                bool in_render_loop = subpass->depth_stencil_attachment->in_render_loop;
                struct radv_image_view *iview = cmd_buffer->state.attachments[idx].iview;
-               struct radv_image *image = iview->image;
                radv_cs_add_buffer(cmd_buffer->device->ws, cmd_buffer->cs, cmd_buffer->state.attachments[idx].iview->bo);
-               ASSERTED uint32_t queue_mask = radv_image_queue_family_mask(image,
-                                                                               cmd_buffer->queue_family_index,
-                                                                               cmd_buffer->queue_family_index);
-               /* We currently don't support writing decompressed HTILE */
-               assert(radv_layout_has_htile(image, layout, in_render_loop, queue_mask) ==
-                      radv_layout_is_htile_compressed(image, layout, in_render_loop, queue_mask));
 
                radv_emit_fb_ds_state(cmd_buffer, &cmd_buffer->state.attachments[idx].ds, iview, layout, in_render_loop);
 
@@ -2063,7 +2108,7 @@ radv_emit_framebuffer_state(struct radv_cmd_buffer *cmd_buffer)
 }
 
 static void
-radv_emit_index_buffer(struct radv_cmd_buffer *cmd_buffer)
+radv_emit_index_buffer(struct radv_cmd_buffer *cmd_buffer, bool indirect)
 {
        struct radeon_cmdbuf *cs = cmd_buffer->cs;
        struct radv_cmd_state *state = &cmd_buffer->state;
@@ -2081,6 +2126,11 @@ radv_emit_index_buffer(struct radv_cmd_buffer *cmd_buffer)
                state->last_index_type = state->index_type;
        }
 
+       /* For the direct indexed draws we use DRAW_INDEX_2, which includes
+        * the index_va and max_index_count already. */
+       if (!indirect)
+               return;
+
        radeon_emit(cs, PKT3(PKT3_INDEX_BASE, 1, 0));
        radeon_emit(cs, state->index_va);
        radeon_emit(cs, state->index_va >> 32);
@@ -2121,8 +2171,11 @@ void radv_set_db_count_control(struct radv_cmd_buffer *cmd_buffer)
                bool gfx10_perfect = cmd_buffer->device->physical_device->rad_info.chip_class >= GFX10 && has_perfect_queries;
 
                if (cmd_buffer->device->physical_device->rad_info.chip_class >= GFX7) {
+                       /* Always enable PERFECT_ZPASS_COUNTS due to issues with partially
+                        * covered tiles, discards, and early depth testing. For more details,
+                        * see https://gitlab.freedesktop.org/mesa/mesa/-/issues/3218 */
                        db_count_control =
-                               S_028004_PERFECT_ZPASS_COUNTS(has_perfect_queries) |
+                               S_028004_PERFECT_ZPASS_COUNTS(1) |
                                S_028004_DISABLE_CONSERVATIVE_ZPASS_COUNTS(gfx10_perfect) |
                                S_028004_SAMPLE_RATE(sample_rate) |
                                S_028004_ZPASS_ENABLE(1) |
@@ -2189,6 +2242,9 @@ radv_cmd_buffer_flush_dynamic_state(struct radv_cmd_buffer *cmd_buffer)
        if (states & RADV_CMD_DIRTY_DYNAMIC_SAMPLE_LOCATIONS)
                radv_emit_sample_locations(cmd_buffer);
 
+       if (states & RADV_CMD_DIRTY_DYNAMIC_LINE_STIPPLE)
+               radv_emit_line_stipple(cmd_buffer);
+
        cmd_buffer->state.dirty &= ~states;
 }
 
@@ -2851,6 +2907,17 @@ radv_src_access_flush(struct radv_cmd_buffer *cmd_buffer,
                                      RADV_CMD_FLAG_FLUSH_AND_INV_DB |
                                      RADV_CMD_FLAG_INV_L2;
 
+                       if (flush_CB_meta)
+                               flush_bits |= RADV_CMD_FLAG_FLUSH_AND_INV_CB_META;
+                       if (flush_DB_meta)
+                               flush_bits |= RADV_CMD_FLAG_FLUSH_AND_INV_DB_META;
+                       break;
+               case VK_ACCESS_MEMORY_WRITE_BIT:
+                       flush_bits |= RADV_CMD_FLAG_INV_L2 |
+                                     RADV_CMD_FLAG_WB_L2 |
+                                     RADV_CMD_FLAG_FLUSH_AND_INV_CB |
+                                     RADV_CMD_FLAG_FLUSH_AND_INV_DB;
+
                        if (flush_CB_meta)
                                flush_bits |= RADV_CMD_FLAG_FLUSH_AND_INV_CB_META;
                        if (flush_DB_meta)
@@ -2920,8 +2987,7 @@ radv_dst_access_flush(struct radv_cmd_buffer *cmd_buffer,
                        flush_bits |= RADV_CMD_FLAG_INV_VCACHE;
                        /* Unlike LLVM, ACO uses SMEM for SSBOs and we have to
                         * invalidate the scalar cache. */
-                       if (cmd_buffer->device->physical_device->use_aco &&
-                           cmd_buffer->device->physical_device->rad_info.chip_class >= GFX8)
+                       if (!cmd_buffer->device->physical_device->use_llvm)
                                flush_bits |= RADV_CMD_FLAG_INV_SCACHE;
 
                        if (!image_is_coherent)
@@ -2939,6 +3005,19 @@ radv_dst_access_flush(struct radv_cmd_buffer *cmd_buffer,
                        if (flush_DB_meta)
                                flush_bits |= RADV_CMD_FLAG_FLUSH_AND_INV_DB_META;
                        break;
+               case VK_ACCESS_MEMORY_READ_BIT:
+                       flush_bits |= RADV_CMD_FLAG_INV_VCACHE |
+                                     RADV_CMD_FLAG_INV_SCACHE |
+                                     RADV_CMD_FLAG_INV_L2;
+                       if (flush_CB)
+                               flush_bits |= RADV_CMD_FLAG_FLUSH_AND_INV_CB;
+                       if (flush_CB_meta)
+                               flush_bits |= RADV_CMD_FLAG_FLUSH_AND_INV_CB_META;
+                       if (flush_DB)
+                               flush_bits |= RADV_CMD_FLAG_FLUSH_AND_INV_DB;
+                       if (flush_DB_meta)
+                               flush_bits |= RADV_CMD_FLAG_FLUSH_AND_INV_DB_META;
+                       break;
                default:
                        break;
                }
@@ -3280,7 +3359,6 @@ VkResult radv_AllocateCommandBuffers(
                        list_addtail(&cmd_buffer->pool_link, &pool->cmd_buffers);
 
                        result = radv_reset_cmd_buffer(cmd_buffer);
-                       cmd_buffer->_loader_data.loaderMagic = ICD_LOADER_MAGIC;
                        cmd_buffer->level = pAllocateInfo->level;
 
                        pCommandBuffers[i] = radv_cmd_buffer_to_handle(cmd_buffer);
@@ -3363,6 +3441,9 @@ VkResult radv_BeginCommandBuffer(
        cmd_buffer->state.last_vertex_offset = -1;
        cmd_buffer->state.last_first_instance = -1;
        cmd_buffer->state.predication_type = -1;
+       cmd_buffer->state.last_sx_ps_downconvert = -1;
+       cmd_buffer->state.last_sx_blend_opt_epsilon = -1;
+       cmd_buffer->state.last_sx_blend_opt_control = -1;
        cmd_buffer->usage_flags = pBeginInfo->flags;
 
        if (cmd_buffer->level == VK_COMMAND_BUFFER_LEVEL_SECONDARY &&
@@ -3386,14 +3467,10 @@ VkResult radv_BeginCommandBuffer(
                radv_cmd_buffer_set_subpass(cmd_buffer, subpass);
        }
 
-       if (unlikely(cmd_buffer->device->trace_bo)) {
-               struct radv_device *device = cmd_buffer->device;
-
-               radv_cs_add_buffer(device->ws, cmd_buffer->cs,
-                                  device->trace_bo);
-
+       if (unlikely(cmd_buffer->device->trace_bo))
                radv_cmd_buffer_trace_emit(cmd_buffer);
-       }
+
+       radv_describe_begin_cmd_buffer(cmd_buffer);
 
        cmd_buffer->status = RADV_CMD_BUFFER_STATUS_RECORDING;
 
@@ -3416,19 +3493,22 @@ void radv_CmdBindVertexBuffers(
 
        assert(firstBinding + bindingCount <= MAX_VBS);
        for (uint32_t i = 0; i < bindingCount; i++) {
+               RADV_FROM_HANDLE(radv_buffer, buffer, pBuffers[i]);
                uint32_t idx = firstBinding + i;
 
                if (!changed &&
-                   (vb[idx].buffer != radv_buffer_from_handle(pBuffers[i]) ||
+                   (vb[idx].buffer != buffer ||
                     vb[idx].offset != pOffsets[i])) {
                        changed = true;
                }
 
-               vb[idx].buffer = radv_buffer_from_handle(pBuffers[i]);
+               vb[idx].buffer = buffer;
                vb[idx].offset = pOffsets[i];
 
-               radv_cs_add_buffer(cmd_buffer->device->ws, cmd_buffer->cs,
-                                  vb[idx].buffer->bo);
+               if (buffer) {
+                       radv_cs_add_buffer(cmd_buffer->device->ws,
+                                          cmd_buffer->cs, vb[idx].buffer->bo);
+               }
        }
 
        if (!changed) {
@@ -3511,7 +3591,7 @@ radv_bind_descriptor_set(struct radv_cmd_buffer *cmd_buffer,
        assert(!(set->layout->flags & VK_DESCRIPTOR_SET_LAYOUT_CREATE_PUSH_DESCRIPTOR_BIT_KHR));
 
        if (!cmd_buffer->device->use_global_bo_list) {
-               for (unsigned j = 0; j < set->layout->buffer_count; ++j)
+               for (unsigned j = 0; j < set->buffer_count; ++j)
                        if (set->descriptors[j])
                                radv_cs_add_buffer(ws, cmd_buffer->cs, set->descriptors[j]);
        }
@@ -3748,11 +3828,14 @@ VkResult radv_EndCommandBuffer(
         */
        si_cp_dma_wait_for_idle(cmd_buffer);
 
+       radv_describe_end_cmd_buffer(cmd_buffer);
+
        vk_free(&cmd_buffer->pool->alloc, cmd_buffer->state.attachments);
        vk_free(&cmd_buffer->pool->alloc, cmd_buffer->state.subpass_sample_locs);
 
-       if (!cmd_buffer->device->ws->cs_finalize(cmd_buffer->cs))
-               return vk_error(cmd_buffer->device->instance, VK_ERROR_OUT_OF_DEVICE_MEMORY);
+       VkResult result = cmd_buffer->device->ws->cs_finalize(cmd_buffer->cs);
+       if (result != VK_SUCCESS)
+               return vk_error(cmd_buffer->device->instance, result);
 
        cmd_buffer->status = RADV_CMD_BUFFER_STATUS_EXECUTABLE;
 
@@ -3831,9 +3914,7 @@ void radv_CmdBindPipeline(
                /* Prefetch all pipeline shaders at first draw time. */
                cmd_buffer->state.prefetch_L2_mask |= RADV_PREFETCH_SHADERS;
 
-               if ((cmd_buffer->device->physical_device->rad_info.family == CHIP_NAVI10 ||
-                    cmd_buffer->device->physical_device->rad_info.family == CHIP_NAVI12 ||
-                    cmd_buffer->device->physical_device->rad_info.family == CHIP_NAVI14) &&
+               if (cmd_buffer->device->physical_device->rad_info.chip_class == GFX10 &&
                    cmd_buffer->state.emitted_pipeline &&
                    radv_pipeline_has_ngg(cmd_buffer->state.emitted_pipeline) &&
                    !radv_pipeline_has_ngg(cmd_buffer->state.pipeline)) {
@@ -4091,6 +4172,20 @@ void radv_CmdSetSampleLocationsEXT(
        state->dirty |= RADV_CMD_DIRTY_DYNAMIC_SAMPLE_LOCATIONS;
 }
 
+void radv_CmdSetLineStippleEXT(
+       VkCommandBuffer                             commandBuffer,
+       uint32_t                                    lineStippleFactor,
+       uint16_t                                    lineStipplePattern)
+{
+       RADV_FROM_HANDLE(radv_cmd_buffer, cmd_buffer, commandBuffer);
+       struct radv_cmd_state *state = &cmd_buffer->state;
+
+       state->dynamic.line_stipple.factor = lineStippleFactor;
+       state->dynamic.line_stipple.pattern = lineStipplePattern;
+
+       state->dirty |= RADV_CMD_DIRTY_DYNAMIC_LINE_STIPPLE;
+}
+
 void radv_CmdExecuteCommands(
        VkCommandBuffer                             commandBuffer,
        uint32_t                                    commandBufferCount,
@@ -4173,6 +4268,9 @@ void radv_CmdExecuteCommands(
                primary->state.last_first_instance = secondary->state.last_first_instance;
                primary->state.last_num_instances = secondary->state.last_num_instances;
                primary->state.last_vertex_offset = secondary->state.last_vertex_offset;
+               primary->state.last_sx_ps_downconvert = secondary->state.last_sx_ps_downconvert;
+               primary->state.last_sx_blend_opt_epsilon = secondary->state.last_sx_blend_opt_epsilon;
+               primary->state.last_sx_blend_opt_control = secondary->state.last_sx_blend_opt_control;
 
                if (secondary->state.last_index_type != -1) {
                        primary->state.last_index_type =
@@ -4199,15 +4297,18 @@ VkResult radv_CreateCommandPool(
        RADV_FROM_HANDLE(radv_device, device, _device);
        struct radv_cmd_pool *pool;
 
-       pool = vk_alloc2(&device->alloc, pAllocator, sizeof(*pool), 8,
+       pool = vk_alloc2(&device->vk.alloc, pAllocator, sizeof(*pool), 8,
                           VK_SYSTEM_ALLOCATION_SCOPE_OBJECT);
        if (pool == NULL)
                return vk_error(device->instance, VK_ERROR_OUT_OF_HOST_MEMORY);
 
+       vk_object_base_init(&device->vk, &pool->base,
+                           VK_OBJECT_TYPE_COMMAND_POOL);
+
        if (pAllocator)
                pool->alloc = *pAllocator;
        else
-               pool->alloc = device->alloc;
+               pool->alloc = device->vk.alloc;
 
        list_inithead(&pool->cmd_buffers);
        list_inithead(&pool->free_cmd_buffers);
@@ -4241,7 +4342,8 @@ void radv_DestroyCommandPool(
                radv_cmd_buffer_destroy(cmd_buffer);
        }
 
-       vk_free2(&device->alloc, pAllocator, pool);
+       vk_object_base_finish(&pool->base);
+       vk_free2(&device->vk.alloc, pAllocator, pool);
 }
 
 VkResult radv_ResetCommandPool(
@@ -4292,6 +4394,8 @@ radv_cmd_buffer_begin_subpass(struct radv_cmd_buffer *cmd_buffer,
 
        radv_cmd_buffer_set_subpass(cmd_buffer, subpass);
 
+       radv_describe_barrier_start(cmd_buffer, RGP_BARRIER_EXTERNAL_RENDER_PASS_SYNC);
+
        for (uint32_t i = 0; i < subpass->attachment_count; ++i) {
                const uint32_t a = subpass->attachments[i].attachment;
                if (a == VK_ATTACHMENT_UNUSED)
@@ -4302,6 +4406,8 @@ radv_cmd_buffer_begin_subpass(struct radv_cmd_buffer *cmd_buffer,
                                                     true);
        }
 
+       radv_describe_barrier_end(cmd_buffer);
+
        radv_cmd_buffer_clear_subpass(cmd_buffer);
 
        assert(cmd_buffer->cs->cdw <= cdw_max);
@@ -4316,6 +4422,8 @@ radv_cmd_buffer_end_subpass(struct radv_cmd_buffer *cmd_buffer)
 
        radv_cmd_buffer_resolve_subpass(cmd_buffer);
 
+       radv_describe_barrier_start(cmd_buffer, RGP_BARRIER_EXTERNAL_RENDER_PASS_SYNC);
+
        for (uint32_t i = 0; i < subpass->attachment_count; ++i) {
                const uint32_t a = subpass->attachments[i].attachment;
                if (a == VK_ATTACHMENT_UNUSED)
@@ -4329,14 +4437,14 @@ radv_cmd_buffer_end_subpass(struct radv_cmd_buffer *cmd_buffer)
                struct radv_subpass_attachment att = { a, layout, stencil_layout };
                radv_handle_subpass_image_transition(cmd_buffer, att, false);
        }
+
+       radv_describe_barrier_end(cmd_buffer);
 }
 
-void radv_CmdBeginRenderPass(
-       VkCommandBuffer                             commandBuffer,
-       const VkRenderPassBeginInfo*                pRenderPassBegin,
-       VkSubpassContents                           contents)
+void
+radv_cmd_buffer_begin_render_pass(struct radv_cmd_buffer *cmd_buffer,
+                                 const VkRenderPassBeginInfo *pRenderPassBegin)
 {
-       RADV_FROM_HANDLE(radv_cmd_buffer, cmd_buffer, commandBuffer);
        RADV_FROM_HANDLE(radv_render_pass, pass, pRenderPassBegin->renderPass);
        RADV_FROM_HANDLE(radv_framebuffer, framebuffer, pRenderPassBegin->framebuffer);
        VkResult result;
@@ -4352,6 +4460,16 @@ void radv_CmdBeginRenderPass(
        result = radv_cmd_state_setup_sample_locations(cmd_buffer, pass, pRenderPassBegin);
        if (result != VK_SUCCESS)
                return;
+}
+
+void radv_CmdBeginRenderPass(
+       VkCommandBuffer                             commandBuffer,
+       const VkRenderPassBeginInfo*                pRenderPassBegin,
+       VkSubpassContents                           contents)
+{
+       RADV_FROM_HANDLE(radv_cmd_buffer, cmd_buffer, commandBuffer);
+
+       radv_cmd_buffer_begin_render_pass(cmd_buffer, pRenderPassBegin);
 
        radv_cmd_buffer_begin_subpass(cmd_buffer, 0);
 }
@@ -4665,7 +4783,7 @@ radv_emit_all_graphics_states(struct radv_cmd_buffer *cmd_buffer,
 
        if (info->indexed) {
                if (cmd_buffer->state.dirty & RADV_CMD_DIRTY_INDEX_BUFFER)
-                       radv_emit_index_buffer(cmd_buffer);
+                       radv_emit_index_buffer(cmd_buffer, info->indirect);
        } else {
                /* On GFX7 and later, non-indexed draws overwrite VGT_INDEX_TYPE,
                 * so the state must be re-emitted before the next indexed
@@ -4714,6 +4832,8 @@ radv_draw(struct radv_cmd_buffer *cmd_buffer,
                        return;
        }
 
+       radv_describe_draw(cmd_buffer);
+
        /* Use optimal packet order based on whether we need to sync the
         * pipeline.
         */
@@ -5089,6 +5209,8 @@ radv_dispatch(struct radv_cmd_buffer *cmd_buffer,
        bool pipeline_is_dirty = pipeline &&
                                 pipeline != cmd_buffer->state.emitted_compute_pipeline;
 
+       radv_describe_dispatch(cmd_buffer, 8, 8, 8);
+
        if (cmd_buffer->state.flush_bits & (RADV_CMD_FLAG_FLUSH_AND_INV_CB |
                                            RADV_CMD_FLAG_FLUSH_AND_INV_DB |
                                            RADV_CMD_FLAG_PS_PARTIAL_FLUSH |
@@ -5199,15 +5321,9 @@ void radv_unaligned_dispatch(
        radv_dispatch(cmd_buffer, &info);
 }
 
-void radv_CmdEndRenderPass(
-       VkCommandBuffer                             commandBuffer)
+void
+radv_cmd_buffer_end_render_pass(struct radv_cmd_buffer *cmd_buffer)
 {
-       RADV_FROM_HANDLE(radv_cmd_buffer, cmd_buffer, commandBuffer);
-
-       radv_subpass_barrier(cmd_buffer, &cmd_buffer->state.pass->end_barrier);
-
-       radv_cmd_buffer_end_subpass(cmd_buffer);
-
        vk_free(&cmd_buffer->pool->alloc, cmd_buffer->state.attachments);
        vk_free(&cmd_buffer->pool->alloc, cmd_buffer->state.subpass_sample_locs);
 
@@ -5218,6 +5334,18 @@ void radv_CmdEndRenderPass(
        cmd_buffer->state.subpass_sample_locs = NULL;
 }
 
+void radv_CmdEndRenderPass(
+       VkCommandBuffer                             commandBuffer)
+{
+       RADV_FROM_HANDLE(radv_cmd_buffer, cmd_buffer, commandBuffer);
+
+       radv_subpass_barrier(cmd_buffer, &cmd_buffer->state.pass->end_barrier);
+
+       radv_cmd_buffer_end_subpass(cmd_buffer);
+
+       radv_cmd_buffer_end_render_pass(cmd_buffer);
+}
+
 void radv_CmdEndRenderPass2(
     VkCommandBuffer                             commandBuffer,
     const VkSubpassEndInfo*                     pSubpassEndInfo)
@@ -5242,10 +5370,14 @@ static void radv_initialize_htile(struct radv_cmd_buffer *cmd_buffer,
        struct radv_cmd_state *state = &cmd_buffer->state;
        uint32_t htile_value = vk_format_is_stencil(image->vk_format) ? 0xfffff30f : 0xfffc000f;
        VkClearDepthStencilValue value = {};
+       struct radv_barrier_data barrier = {};
 
        state->flush_bits |= RADV_CMD_FLAG_FLUSH_AND_INV_DB |
                             RADV_CMD_FLAG_FLUSH_AND_INV_DB_META;
 
+       barrier.layout_transitions.init_mask_ram = 1;
+       radv_describe_layout_transition(cmd_buffer, &barrier);
+
        state->flush_bits |= radv_clear_htile(cmd_buffer, image, range, htile_value);
 
        state->flush_bits |= RADV_CMD_FLAG_FLUSH_AND_INV_DB_META;
@@ -5289,8 +5421,8 @@ static void radv_handle_depth_image_transition(struct radv_cmd_buffer *cmd_buffe
                cmd_buffer->state.flush_bits |= RADV_CMD_FLAG_FLUSH_AND_INV_DB |
                                                RADV_CMD_FLAG_FLUSH_AND_INV_DB_META;
 
-               radv_decompress_depth_image_inplace(cmd_buffer, image, range,
-                                                   sample_locs);
+               radv_decompress_depth_stencil(cmd_buffer, image, range,
+                                             sample_locs);
 
                cmd_buffer->state.flush_bits |= RADV_CMD_FLAG_FLUSH_AND_INV_DB |
                                                RADV_CMD_FLAG_FLUSH_AND_INV_DB_META;
@@ -5303,10 +5435,14 @@ static void radv_initialise_cmask(struct radv_cmd_buffer *cmd_buffer,
                                  uint32_t value)
 {
        struct radv_cmd_state *state = &cmd_buffer->state;
+       struct radv_barrier_data barrier = {};
 
        state->flush_bits |= RADV_CMD_FLAG_FLUSH_AND_INV_CB |
                            RADV_CMD_FLAG_FLUSH_AND_INV_CB_META;
 
+       barrier.layout_transitions.init_mask_ram = 1;
+       radv_describe_layout_transition(cmd_buffer, &barrier);
+
        state->flush_bits |= radv_clear_cmask(cmd_buffer, image, range, value);
 
        state->flush_bits |= RADV_CMD_FLAG_FLUSH_AND_INV_CB_META;
@@ -5325,10 +5461,14 @@ void radv_initialize_fmask(struct radv_cmd_buffer *cmd_buffer,
        };
        uint32_t log2_samples = util_logbase2(image->info.samples);
        uint32_t value = fmask_clear_values[log2_samples];
+       struct radv_barrier_data barrier = {};
 
        state->flush_bits |= RADV_CMD_FLAG_FLUSH_AND_INV_CB |
                             RADV_CMD_FLAG_FLUSH_AND_INV_CB_META;
 
+       barrier.layout_transitions.init_mask_ram = 1;
+       radv_describe_layout_transition(cmd_buffer, &barrier);
+
        state->flush_bits |= radv_clear_fmask(cmd_buffer, image, range, value);
 
        state->flush_bits |= RADV_CMD_FLAG_FLUSH_AND_INV_CB_META;
@@ -5339,11 +5479,15 @@ void radv_initialize_dcc(struct radv_cmd_buffer *cmd_buffer,
                         const VkImageSubresourceRange *range, uint32_t value)
 {
        struct radv_cmd_state *state = &cmd_buffer->state;
+       struct radv_barrier_data barrier = {};
        unsigned size = 0;
 
        state->flush_bits |= RADV_CMD_FLAG_FLUSH_AND_INV_CB |
                             RADV_CMD_FLAG_FLUSH_AND_INV_CB_META;
 
+       barrier.layout_transitions.init_mask_ram = 1;
+       radv_describe_layout_transition(cmd_buffer, &barrier);
+
        state->flush_bits |= radv_clear_dcc(cmd_buffer, image, range, value);
 
        if (cmd_buffer->device->physical_device->rad_info.chip_class == GFX8) {
@@ -5368,7 +5512,7 @@ void radv_initialize_dcc(struct radv_cmd_buffer *cmd_buffer,
                if (size != image->planes[0].surface.dcc_size) {
                        state->flush_bits |=
                                radv_fill_buffer(cmd_buffer, image->bo,
-                                                image->offset + image->dcc_offset + size,
+                                                image->offset + image->planes[0].surface.dcc_offset + size,
                                                 image->planes[0].surface.dcc_size - size,
                                                 0xffffffff);
                }
@@ -5484,8 +5628,13 @@ static void radv_handle_color_image_transition(struct radv_cmd_buffer *cmd_buffe
                if (fce_eliminate || fmask_expand)
                        radv_fast_clear_flush_image_inplace(cmd_buffer, image, range);
 
-               if (fmask_expand)
+               if (fmask_expand) {
+                       struct radv_barrier_data barrier = {};
+                       barrier.layout_transitions.fmask_color_expand = 1;
+                       radv_describe_layout_transition(cmd_buffer, &barrier);
+
                        radv_expand_fmask_image_inplace(cmd_buffer, image, range);
+               }
        }
 }
 
@@ -5547,6 +5696,7 @@ static void radv_handle_image_transition(struct radv_cmd_buffer *cmd_buffer,
 }
 
 struct radv_barrier_info {
+       enum rgp_barrier_reason reason;
        uint32_t eventCount;
        const VkEvent *pEvents;
        VkPipelineStageFlags srcStageMask;
@@ -5567,6 +5717,8 @@ radv_barrier(struct radv_cmd_buffer *cmd_buffer,
        enum radv_cmd_flush_bits src_flush_bits = 0;
        enum radv_cmd_flush_bits dst_flush_bits = 0;
 
+       radv_describe_barrier_start(cmd_buffer, info->reason);
+
        for (unsigned i = 0; i < info->eventCount; ++i) {
                RADV_FROM_HANDLE(radv_event, event, info->pEvents[i]);
                uint64_t va = radv_buffer_get_va(event->bo);
@@ -5654,6 +5806,8 @@ radv_barrier(struct radv_cmd_buffer *cmd_buffer,
                si_cp_dma_wait_for_idle(cmd_buffer);
 
        cmd_buffer->state.flush_bits |= dst_flush_bits;
+
+       radv_describe_barrier_end(cmd_buffer);
 }
 
 void radv_CmdPipelineBarrier(
@@ -5671,6 +5825,7 @@ void radv_CmdPipelineBarrier(
        RADV_FROM_HANDLE(radv_cmd_buffer, cmd_buffer, commandBuffer);
        struct radv_barrier_info info;
 
+       info.reason = RGP_BARRIER_EXTERNAL_CMD_PIPELINE_BARRIER;
        info.eventCount = 0;
        info.pEvents = NULL;
        info.srcStageMask = srcStageMask;
@@ -5782,6 +5937,7 @@ void radv_CmdWaitEvents(VkCommandBuffer commandBuffer,
        RADV_FROM_HANDLE(radv_cmd_buffer, cmd_buffer, commandBuffer);
        struct radv_barrier_info info;
 
+       info.reason = RGP_BARRIER_EXTERNAL_CMD_WAIT_EVENTS;
        info.eventCount = eventCount;
        info.pEvents = pEvents;
        info.srcStageMask = 0;
@@ -5905,7 +6061,12 @@ void radv_CmdBindTransformFeedbackBuffersEXT(
 
                sb[idx].buffer = radv_buffer_from_handle(pBuffers[i]);
                sb[idx].offset = pOffsets[i];
-               sb[idx].size = pSizes[i];
+
+               if (!pSizes || pSizes[i] == VK_WHOLE_SIZE) {
+                       sb[idx].size = sb[idx].buffer->size - sb[idx].offset;
+               } else {
+                       sb[idx].size = pSizes[i];
+               }
 
                radv_cs_add_buffer(cmd_buffer->device->ws, cmd_buffer->cs,
                                   sb[idx].buffer->bo);