radeonsi: remove Constant Engine support
authorMarek Olšák <marek.olsak@amd.com>
Sat, 19 Aug 2017 16:56:36 +0000 (18:56 +0200)
committerMarek Olšák <marek.olsak@amd.com>
Tue, 22 Aug 2017 11:29:47 +0000 (13:29 +0200)
We have come to the conclusion that it doesn't improve performance.

Tested-by: Dieter Nützel <Dieter@nuetzel-hh.de>
Reviewed-by: Samuel Pitoiset <samuel.pitoiset@gmail.com>
Reviewed-by: Nicolai Hähnle <nicolai.haehnle@amd.com>
16 files changed:
src/gallium/drivers/radeon/r600_gpu_load.c
src/gallium/drivers/radeon/r600_pipe_common.c
src/gallium/drivers/radeon/r600_pipe_common.h
src/gallium/drivers/radeon/r600_query.c
src/gallium/drivers/radeon/r600_query.h
src/gallium/drivers/radeon/radeon_winsys.h
src/gallium/drivers/radeonsi/si_compute.c
src/gallium/drivers/radeonsi/si_debug.c
src/gallium/drivers/radeonsi/si_descriptors.c
src/gallium/drivers/radeonsi/si_hw_context.c
src/gallium/drivers/radeonsi/si_pipe.c
src/gallium/drivers/radeonsi/si_pipe.h
src/gallium/drivers/radeonsi/si_state.h
src/gallium/drivers/radeonsi/si_state_draw.c
src/gallium/winsys/amdgpu/drm/amdgpu_cs.c
src/gallium/winsys/amdgpu/drm/amdgpu_cs.h

index 4e9f7ea37cd1adc00e09a5fa7f6bf0040569f569..d35be4f327aa2ba8bb72bb7484fb116c44045a86 100644 (file)
@@ -68,7 +68,6 @@
 #define SURFACE_SYNC_BUSY(x)   (((x) >> 21) & 0x1)
 #define DMA_BUSY(x)            (((x) >> 22) & 0x1)
 #define SCRATCH_RAM_BUSY(x)    (((x) >> 24) & 0x1)
-#define CE_BUSY(x)             (((x) >> 26) & 0x1)
 
 #define IDENTITY(x) x
 
@@ -123,7 +122,6 @@ static void r600_update_mmio_counters(struct r600_common_screen *rscreen,
                UPDATE_COUNTER(surf_sync, SURFACE_SYNC_BUSY);
                UPDATE_COUNTER(cp_dma, DMA_BUSY);
                UPDATE_COUNTER(scratch_ram, SCRATCH_RAM_BUSY);
-               UPDATE_COUNTER(ce, CE_BUSY);
        }
 
        value = gui_busy || sdma_busy;
@@ -266,8 +264,6 @@ static unsigned busy_index_from_type(struct r600_common_screen *rscreen,
                return BUSY_INDEX(rscreen, cp_dma);
        case R600_QUERY_GPU_SCRATCH_RAM_BUSY:
                return BUSY_INDEX(rscreen, scratch_ram);
-       case R600_QUERY_GPU_CE_BUSY:
-               return BUSY_INDEX(rscreen, ce);
        default:
                unreachable("invalid query type");
        }
index dc54b5e5b79117da21678a726f835525417ff7b0..37c12dea3740d781a4959aea248b4d654e5ee3e0 100644 (file)
@@ -815,8 +815,6 @@ static const struct debug_named_value common_debug_options[] = {
        { "norbplus", DBG_NO_RB_PLUS, "Disable RB+." },
        { "sisched", DBG_SI_SCHED, "Enable LLVM SI Machine Instruction Scheduler." },
        { "mono", DBG_MONOLITHIC_SHADERS, "Use old-style monolithic shaders compiled on demand" },
-       { "ce", DBG_CE, "Force enable the constant engine" },
-       { "noce", DBG_NO_CE, "Disable the constant engine"},
        { "unsafemath", DBG_UNSAFE_MATH, "Enable unsafe math shader optimizations" },
        { "nodccfb", DBG_NO_DCC_FB, "Disable separate DCC on the main framebuffer" },
 
index abe9c4ca57b7213b7dca7cfa377e903ed72aeb7c..c10cf18219ea59220a53017a9a33648583ae47bb 100644 (file)
@@ -72,7 +72,7 @@ struct u_log_context;
 #define DBG_NIR                        (1 << 1)
 #define DBG_COMPUTE            (1 << 2)
 #define DBG_VM                 (1 << 3)
-#define DBG_CE                 (1 << 4)
+/* gap */
 /* shader logging */
 #define DBG_FS                 (1 << 5)
 #define DBG_VS                 (1 << 6)
@@ -108,7 +108,7 @@ struct u_log_context;
 #define DBG_NO_RB_PLUS         (1ull << 45)
 #define DBG_SI_SCHED           (1ull << 46)
 #define DBG_MONOLITHIC_SHADERS (1ull << 47)
-#define DBG_NO_CE              (1ull << 48)
+/* gap */
 #define DBG_UNSAFE_MATH                (1ull << 49)
 #define DBG_NO_DCC_FB          (1ull << 50)
 #define DBG_TEST_VMFAULT_CP    (1ull << 51)
@@ -375,7 +375,6 @@ union r600_mmio_counters {
                struct r600_mmio_counter surf_sync;
                struct r600_mmio_counter cp_dma;
                struct r600_mmio_counter scratch_ram;
-               struct r600_mmio_counter ce;
        } named;
        unsigned array[0];
 };
index bccfe7f94f83b05c30b54935e35751b9f236887c..98bdd80d7398d9d68156556c264f3670f50088fb 100644 (file)
@@ -219,7 +219,6 @@ static bool r600_query_sw_begin(struct r600_common_context *rctx,
        case R600_QUERY_GPU_SURF_SYNC_BUSY:
        case R600_QUERY_GPU_CP_DMA_BUSY:
        case R600_QUERY_GPU_SCRATCH_RAM_BUSY:
-       case R600_QUERY_GPU_CE_BUSY:
                query->begin_result = r600_begin_counter(rctx->screen,
                                                         query->b.type);
                break;
@@ -376,7 +375,6 @@ static bool r600_query_sw_end(struct r600_common_context *rctx,
        case R600_QUERY_GPU_SURF_SYNC_BUSY:
        case R600_QUERY_GPU_CP_DMA_BUSY:
        case R600_QUERY_GPU_SCRATCH_RAM_BUSY:
-       case R600_QUERY_GPU_CE_BUSY:
                query->end_result = r600_end_counter(rctx->screen,
                                                     query->b.type,
                                                     query->begin_result);
@@ -2075,7 +2073,6 @@ static struct pipe_driver_query_info r600_driver_query_list[] = {
        X("GPU-surf-sync-busy",         GPU_SURF_SYNC_BUSY,     UINT64, AVERAGE),
        X("GPU-cp-dma-busy",            GPU_CP_DMA_BUSY,        UINT64, AVERAGE),
        X("GPU-scratch-ram-busy",       GPU_SCRATCH_RAM_BUSY,   UINT64, AVERAGE),
-       X("GPU-ce-busy",                GPU_CE_BUSY,            UINT64, AVERAGE),
 };
 
 #undef X
index 815dc7fc3c2e5f2f15691e85af475dac8caf7383..7455c8e63a88db2fa9fddb74d358dce41ed10b07 100644 (file)
@@ -102,7 +102,6 @@ enum {
        R600_QUERY_GPU_SURF_SYNC_BUSY,
        R600_QUERY_GPU_CP_DMA_BUSY,
        R600_QUERY_GPU_SCRATCH_RAM_BUSY,
-       R600_QUERY_GPU_CE_BUSY,
        R600_QUERY_NUM_COMPILATIONS,
        R600_QUERY_NUM_SHADERS_CREATED,
        R600_QUERY_BACK_BUFFER_PS_DRAW_RATIO,
index 351edcd76f98a0dcda7123bdf5d45a616286ee74..b00b1443e7442846f2a48315b964c535651cddb1 100644 (file)
@@ -173,8 +173,7 @@ struct radeon_winsys_cs {
     unsigned                      max_prev; /* Space in array pointed to by prev. */
     unsigned                      prev_dw; /* Total number of dwords in previous chunks. */
 
-    /* Memory usage of the buffer list. These are always 0 for CE and preamble
-     * IBs. */
+    /* Memory usage of the buffer list. These are always 0 for preamble IBs. */
     uint64_t                      used_vram;
     uint64_t                      used_gart;
 };
@@ -456,36 +455,6 @@ struct radeon_winsys {
                                                        struct pipe_fence_handle **fence),
                                           void *flush_ctx);
 
-    /**
-     * Add a constant engine IB to a graphics CS. This makes the graphics CS
-     * from "cs_create" a group of two IBs that share a buffer list and are
-     * flushed together.
-     *
-     * The returned constant CS is only a stream for writing packets to the new
-     * IB. Calling other winsys functions with it is not allowed, not even
-     * "cs_destroy".
-     *
-     * In order to add buffers and check memory usage, use the graphics CS.
-     * In order to flush it, use the graphics CS, which will flush both IBs.
-     * Destroying the graphics CS will destroy both of them.
-     *
-     * \param cs  The graphics CS from "cs_create" that will hold the buffer
-     *            list and will be used for flushing.
-     */
-    struct radeon_winsys_cs *(*cs_add_const_ib)(struct radeon_winsys_cs *cs);
-
-     /**
-     * Add a constant engine preamble IB to a graphics CS. This add an extra IB
-     * in similar manner to cs_add_const_ib. This should always be called after
-     * cs_add_const_ib.
-     *
-     * The returned IB is a constant engine IB that only gets flushed if the
-     * context changed.
-     *
-     * \param cs  The graphics CS from "cs_create" that will hold the buffer
-     *            list and will be used for flushing.
-     */
-    struct radeon_winsys_cs *(*cs_add_const_preamble_ib)(struct radeon_winsys_cs *cs);
     /**
      * Destroy a command stream.
      *
index d0e481a3f15313cfae229fe26e41962bf4b4a54f..3ebd22c3c16c11719a6215d4babb2716534e09c8 100644 (file)
@@ -844,12 +844,8 @@ static void si_launch_grid(
        if (program->ir_type == PIPE_SHADER_IR_TGSI)
                si_setup_tgsi_grid(sctx, info);
 
-       si_ce_pre_draw_synchronization(sctx);
-
        si_emit_dispatch_packets(sctx, info);
 
-       si_ce_post_draw_synchronization(sctx);
-
        if (unlikely(sctx->current_saved_cs))
                si_trace_emit(sctx);
 
index b6bddc521608423210f6c063649a219a628023cc..c2242a6deabbbc5903131f99a4d1dfb825f9d8aa 100644 (file)
@@ -274,7 +274,6 @@ struct si_log_chunk_cs {
        struct si_saved_cs *cs;
        bool dump_bo_list;
        unsigned gfx_begin, gfx_end;
-       unsigned ce_begin, ce_end;
 };
 
 static void si_log_chunk_type_cs_destroy(void *data)
@@ -331,7 +330,6 @@ static void si_log_chunk_type_cs_print(void *data, FILE *f)
        struct si_context *ctx = chunk->ctx;
        struct si_saved_cs *scs = chunk->cs;
        int last_trace_id = -1;
-       int last_ce_trace_id = -1;
 
        /* We are expecting that the ddebug pipe has already
         * waited for the context, so this buffer should be idle.
@@ -341,10 +339,8 @@ static void si_log_chunk_type_cs_print(void *data, FILE *f)
                                              NULL,
                                              PIPE_TRANSFER_UNSYNCHRONIZED |
                                              PIPE_TRANSFER_READ);
-       if (map) {
+       if (map)
                last_trace_id = map[0];
-               last_ce_trace_id = map[1];
-       }
 
        if (chunk->gfx_end != chunk->gfx_begin) {
                if (chunk->gfx_begin == 0) {
@@ -372,21 +368,6 @@ static void si_log_chunk_type_cs_print(void *data, FILE *f)
                }
        }
 
-       if (chunk->ce_end != chunk->ce_begin) {
-               assert(ctx->ce_ib);
-
-               if (scs->flushed) {
-                       ac_parse_ib(f, scs->ce.ib + chunk->ce_begin,
-                                   chunk->ce_end - chunk->ce_begin,
-                                   last_ce_trace_id, "CE IB", ctx->b.chip_class,
-                                   NULL, NULL);
-               } else {
-                       si_parse_current_ib(f, ctx->ce_ib, chunk->ce_begin,
-                                           chunk->ce_end, last_ce_trace_id, "CE IB",
-                                           ctx->b.chip_class);
-               }
-       }
-
        if (chunk->dump_bo_list) {
                fprintf(f, "Flushing.\n\n");
                si_dump_bo_list(ctx, &scs->gfx, f);
@@ -405,14 +386,9 @@ static void si_log_cs(struct si_context *ctx, struct u_log_context *log,
 
        struct si_saved_cs *scs = ctx->current_saved_cs;
        unsigned gfx_cur = ctx->b.gfx.cs->prev_dw + ctx->b.gfx.cs->current.cdw;
-       unsigned ce_cur = 0;
-
-       if (ctx->ce_ib)
-               ce_cur = ctx->ce_ib->prev_dw + ctx->ce_ib->current.cdw;
 
        if (!dump_bo_list &&
-           gfx_cur == scs->gfx_last_dw &&
-           ce_cur == scs->ce_last_dw)
+           gfx_cur == scs->gfx_last_dw)
                return;
 
        struct si_log_chunk_cs *chunk = calloc(1, sizeof(*chunk));
@@ -425,10 +401,6 @@ static void si_log_cs(struct si_context *ctx, struct u_log_context *log,
        chunk->gfx_end = gfx_cur;
        scs->gfx_last_dw = gfx_cur;
 
-       chunk->ce_begin = scs->ce_last_dw;
-       chunk->ce_end = ce_cur;
-       scs->ce_last_dw = ce_cur;
-
        u_log_chunk(log, &si_log_chunk_type_cs, chunk);
 }
 
index ab399a5fb0d419ee901e68552f946e90c44a8516..646a9ec2570fcb7637ad3adef5adf5714b065365 100644 (file)
@@ -97,11 +97,6 @@ static uint32_t null_image_descriptor[8] = {
         * descriptor */
 };
 
-static uint16_t si_ce_ram_size(struct si_context *sctx)
-{
-       return sctx->b.chip_class >= GFX9 ? 4096 : 32768;
-}
-
 static void si_init_descriptor_list(uint32_t *desc_list,
                                    unsigned element_dw_size,
                                    unsigned num_elements,
@@ -117,32 +112,15 @@ static void si_init_descriptor_list(uint32_t *desc_list,
        }
 }
 
-static void si_init_descriptors(struct si_context *sctx,
-                               struct si_descriptors *desc,
+static void si_init_descriptors(struct si_descriptors *desc,
                                unsigned shader_userdata_index,
                                unsigned element_dw_size,
-                               unsigned num_elements,
-                               unsigned first_ce_slot,
-                               unsigned num_ce_slots,
-                               unsigned *ce_offset)
+                               unsigned num_elements)
 {
        desc->list = CALLOC(num_elements, element_dw_size * 4);
        desc->element_dw_size = element_dw_size;
        desc->num_elements = num_elements;
-       desc->first_ce_slot = sctx->ce_ib ? first_ce_slot : 0;
-       desc->num_ce_slots = sctx->ce_ib ? num_ce_slots : 0;
-       desc->dirty_mask = 0;
        desc->shader_userdata_offset = shader_userdata_index * 4;
-
-       if (desc->num_ce_slots) {
-               assert(num_elements <= sizeof(desc->dirty_mask)*8);
-
-               desc->uses_ce = true;
-               desc->ce_offset = *ce_offset;
-               desc->dirty_mask = u_bit_consecutive64(0, num_elements);
-
-               *ce_offset += element_dw_size * desc->num_ce_slots * 4;
-       }
 }
 
 static void si_release_descriptors(struct si_descriptors *desc)
@@ -151,80 +129,6 @@ static void si_release_descriptors(struct si_descriptors *desc)
        FREE(desc->list);
 }
 
-static bool si_ce_upload(struct si_context *sctx, unsigned ce_offset, unsigned size,
-                        unsigned *out_offset, struct r600_resource **out_buf)
-{
-       uint64_t va;
-       unsigned cache_line_size = sctx->screen->b.info.tcc_cache_line_size;
-
-       /* The base and size should be aligned to the L2 cache line size
-        * for optimal performance. (all dumps should rewrite whole lines)
-        */
-       size = align(size, cache_line_size);
-
-       (void)si_ce_ram_size; /* silence an "unused" warning */
-       assert(ce_offset + size <= si_ce_ram_size(sctx));
-
-       u_suballocator_alloc(sctx->ce_suballocator, size, cache_line_size,
-                            out_offset, (struct pipe_resource**)out_buf);
-       if (!out_buf)
-                       return false;
-
-       va = (*out_buf)->gpu_address + *out_offset;
-
-       radeon_emit(sctx->ce_ib, PKT3(PKT3_DUMP_CONST_RAM, 3, 0));
-       radeon_emit(sctx->ce_ib, ce_offset);
-       radeon_emit(sctx->ce_ib, size / 4);
-       radeon_emit(sctx->ce_ib, va);
-       radeon_emit(sctx->ce_ib, va >> 32);
-
-       radeon_add_to_buffer_list(&sctx->b, &sctx->b.gfx, *out_buf,
-                              RADEON_USAGE_READWRITE, RADEON_PRIO_DESCRIPTORS);
-
-       sctx->ce_need_synchronization = true;
-       return true;
-}
-
-void si_ce_save_all_descriptors_at_ib_end(struct si_context* sctx)
-{
-       bool success = si_ce_upload(sctx, 0, sctx->total_ce_ram_allocated,
-                                   &sctx->ce_ram_saved_offset,
-                                   &sctx->ce_ram_saved_buffer);
-       (void)success;
-       assert(success);
-}
-
-void si_ce_restore_all_descriptors_at_ib_start(struct si_context *sctx)
-{
-       if (!sctx->ce_ram_saved_buffer)
-               return;
-
-       struct radeon_winsys_cs *ib = sctx->ce_preamble_ib;
-       if (!ib)
-               ib = sctx->ce_ib;
-
-       uint64_t va = sctx->ce_ram_saved_buffer->gpu_address +
-                     sctx->ce_ram_saved_offset;
-
-       radeon_emit(ib, PKT3(PKT3_LOAD_CONST_RAM, 3, 0));
-       radeon_emit(ib, va);
-       radeon_emit(ib, va >> 32);
-       radeon_emit(ib, sctx->total_ce_ram_allocated / 4);
-       radeon_emit(ib, 0);
-
-       radeon_add_to_buffer_list(&sctx->b, &sctx->b.gfx,
-                                 sctx->ce_ram_saved_buffer,
-                                 RADEON_USAGE_READ, RADEON_PRIO_DESCRIPTORS);
-}
-
-void si_ce_enable_loads(struct radeon_winsys_cs *ib)
-{
-       radeon_emit(ib, PKT3(PKT3_CONTEXT_CONTROL, 1, 0));
-       radeon_emit(ib, CONTEXT_CONTROL_LOAD_ENABLE(1) |
-                       CONTEXT_CONTROL_LOAD_CE_RAM(1));
-       radeon_emit(ib, CONTEXT_CONTROL_SHADOW_ENABLE(1));
-}
-
 static bool si_upload_descriptors(struct si_context *sctx,
                                  struct si_descriptors *desc,
                                  struct r600_atom * atom)
@@ -240,56 +144,25 @@ static bool si_upload_descriptors(struct si_context *sctx,
        if (!upload_size)
                return true;
 
-       if (desc->uses_ce) {
-               const uint32_t *list = desc->list +
-                                      desc->first_ce_slot * desc->element_dw_size;
-               uint64_t mask = (desc->dirty_mask >> desc->first_ce_slot) &
-                               u_bit_consecutive64(0, desc->num_ce_slots);
-
-
-               while (mask) {
-                       int begin, count;
-                       u_bit_scan_consecutive_range64(&mask, &begin, &count);
-
-                       begin *= desc->element_dw_size;
-                       count *= desc->element_dw_size;
-
-                       radeon_emit(sctx->ce_ib,
-                                   PKT3(PKT3_WRITE_CONST_RAM, count, 0));
-                       radeon_emit(sctx->ce_ib, desc->ce_offset + begin * 4);
-                       radeon_emit_array(sctx->ce_ib, list + begin, count);
-               }
-
-               if (!si_ce_upload(sctx,
-                                 desc->ce_offset +
-                                 (first_slot_offset - desc->first_ce_slot * slot_size),
-                                 upload_size, (unsigned*)&desc->buffer_offset,
-                                 &desc->buffer))
-                       return false;
-       } else {
-               uint32_t *ptr;
-
-               u_upload_alloc(sctx->b.b.const_uploader, 0, upload_size,
-                              si_optimal_tcc_alignment(sctx, upload_size),
-                              (unsigned*)&desc->buffer_offset,
-                              (struct pipe_resource**)&desc->buffer,
-                              (void**)&ptr);
-               if (!desc->buffer)
-                       return false; /* skip the draw call */
+       uint32_t *ptr;
+       u_upload_alloc(sctx->b.b.const_uploader, 0, upload_size,
+                      si_optimal_tcc_alignment(sctx, upload_size),
+                      (unsigned*)&desc->buffer_offset,
+                      (struct pipe_resource**)&desc->buffer,
+                      (void**)&ptr);
+       if (!desc->buffer)
+               return false; /* skip the draw call */
 
-               util_memcpy_cpu_to_le32(ptr, (char*)desc->list + first_slot_offset,
-                                       upload_size);
-               desc->gpu_list = ptr - first_slot_offset / 4;
+       util_memcpy_cpu_to_le32(ptr, (char*)desc->list + first_slot_offset,
+                               upload_size);
+       desc->gpu_list = ptr - first_slot_offset / 4;
 
-               radeon_add_to_buffer_list(&sctx->b, &sctx->b.gfx, desc->buffer,
-                                   RADEON_USAGE_READ, RADEON_PRIO_DESCRIPTORS);
-       }
+       radeon_add_to_buffer_list(&sctx->b, &sctx->b.gfx, desc->buffer,
+                            RADEON_USAGE_READ, RADEON_PRIO_DESCRIPTORS);
 
        /* The shader pointer should point to slot 0. */
        desc->buffer_offset -= first_slot_offset;
 
-       desc->dirty_mask = 0;
-
        if (atom)
                si_mark_atom_dirty(sctx, atom);
 
@@ -598,7 +471,6 @@ static void si_set_sampler_view(struct si_context *sctx,
                views->enabled_mask &= ~(1u << slot);
        }
 
-       descs->dirty_mask |= 1ull << desc_slot;
        sctx->descriptors_dirty |= 1u << si_sampler_and_image_descriptors_idx(shader);
 }
 
@@ -750,8 +622,6 @@ si_disable_shader_image(struct si_context *ctx, unsigned shader, unsigned slot)
 
                memcpy(descs->list + desc_slot*8, null_image_descriptor, 8*4);
                images->enabled_mask &= ~(1u << slot);
-               /* two 8-byte images share one 16-byte slot */
-               descs->dirty_mask |= 1u << (desc_slot / 2);
                ctx->descriptors_dirty |= 1u << si_sampler_and_image_descriptors_idx(shader);
        }
 }
@@ -887,8 +757,6 @@ static void si_set_shader_image(struct si_context *ctx,
        }
 
        images->enabled_mask |= 1u << slot;
-       /* two 8-byte images share one 16-byte slot */
-       descs->dirty_mask |= 1u << (desc_slot / 2);
        ctx->descriptors_dirty |= 1u << si_sampler_and_image_descriptors_idx(shader);
 
        /* Since this can flush, it must be done after enabled_mask is updated. */
@@ -984,25 +852,20 @@ static void si_bind_sampler_states(struct pipe_context *ctx,
                        continue;
 
                memcpy(desc->list + desc_slot * 16 + 12, sstates[i]->val, 4*4);
-               desc->dirty_mask |= 1ull << desc_slot;
                sctx->descriptors_dirty |= 1u << si_sampler_and_image_descriptors_idx(shader);
        }
 }
 
 /* BUFFER RESOURCES */
 
-static void si_init_buffer_resources(struct si_context *sctx,
-                                    struct si_buffer_resources *buffers,
+static void si_init_buffer_resources(struct si_buffer_resources *buffers,
                                     struct si_descriptors *descs,
                                     unsigned num_buffers,
-                                    unsigned first_ce_slot,
-                                    unsigned num_ce_slots,
                                     unsigned shader_userdata_index,
                                     enum radeon_bo_usage shader_usage,
                                     enum radeon_bo_usage shader_usage_constbuf,
                                     enum radeon_bo_priority priority,
-                                    enum radeon_bo_priority priority_constbuf,
-                                    unsigned *ce_offset)
+                                    enum radeon_bo_priority priority_constbuf)
 {
        buffers->shader_usage = shader_usage;
        buffers->shader_usage_constbuf = shader_usage_constbuf;
@@ -1010,8 +873,7 @@ static void si_init_buffer_resources(struct si_context *sctx,
        buffers->priority_constbuf = priority_constbuf;
        buffers->buffers = CALLOC(num_buffers, sizeof(struct pipe_resource*));
 
-       si_init_descriptors(sctx, descs, shader_userdata_index, 4, num_buffers,
-                           first_ce_slot, num_ce_slots, ce_offset);
+       si_init_descriptors(descs, shader_userdata_index, 4, num_buffers);
 }
 
 static void si_release_buffer_resources(struct si_buffer_resources *buffers,
@@ -1277,7 +1139,6 @@ static void si_set_constant_buffer(struct si_context *sctx,
                buffers->enabled_mask &= ~(1u << slot);
        }
 
-       descs->dirty_mask |= 1u << slot;
        sctx->descriptors_dirty |= 1u << descriptors_idx;
 }
 
@@ -1339,7 +1200,6 @@ static void si_set_shader_buffers(struct pipe_context *ctx,
                        pipe_resource_reference(&buffers->buffers[slot], NULL);
                        memset(desc, 0, sizeof(uint32_t) * 4);
                        buffers->enabled_mask &= ~(1u << slot);
-                       descs->dirty_mask |= 1u << slot;
                        sctx->descriptors_dirty |=
                                1u << si_const_and_shader_buffer_descriptors_idx(shader);
                        continue;
@@ -1366,7 +1226,6 @@ static void si_set_shader_buffers(struct pipe_context *ctx,
                buf->bind_history |= PIPE_BIND_SHADER_BUFFER;
 
                buffers->enabled_mask |= 1u << slot;
-               descs->dirty_mask |= 1u << slot;
                sctx->descriptors_dirty |=
                        1u << si_const_and_shader_buffer_descriptors_idx(shader);
 
@@ -1486,7 +1345,6 @@ void si_set_ring_buffer(struct pipe_context *ctx, uint slot,
                buffers->enabled_mask &= ~(1u << slot);
        }
 
-       descs->dirty_mask |= 1u << slot;
        sctx->descriptors_dirty |= 1u << SI_DESCS_RW_BUFFERS;
 }
 
@@ -1591,7 +1449,6 @@ static void si_set_streamout_targets(struct pipe_context *ctx,
                                                NULL);
                        buffers->enabled_mask &= ~(1u << bufidx);
                }
-               descs->dirty_mask |= 1u << bufidx;
        }
        for (; i < old_num_targets; i++) {
                bufidx = SI_VS_STREAMOUT_BUF0 + i;
@@ -1599,7 +1456,6 @@ static void si_set_streamout_targets(struct pipe_context *ctx,
                memset(descs->list + bufidx*4, 0, sizeof(uint32_t) * 4);
                pipe_resource_reference(&buffers->buffers[bufidx], NULL);
                buffers->enabled_mask &= ~(1u << bufidx);
-               descs->dirty_mask |= 1u << bufidx;
        }
 
        sctx->descriptors_dirty |= 1u << SI_DESCS_RW_BUFFERS;
@@ -1718,7 +1574,6 @@ static void si_reset_buffer_resources(struct si_context *sctx,
                        si_desc_reset_buffer_offset(&sctx->b.b,
                                                    descs->list + i*4,
                                                    old_va, buf);
-                       descs->dirty_mask |= 1u << i;
                        sctx->descriptors_dirty |= 1u << descriptors_idx;
 
                        radeon_add_to_buffer_list_check_mem(&sctx->b, &sctx->b.gfx,
@@ -1772,7 +1627,6 @@ static void si_rebind_buffer(struct pipe_context *ctx, struct pipe_resource *buf
 
                        si_desc_reset_buffer_offset(ctx, descs->list + i*4,
                                                    old_va, buf);
-                       descs->dirty_mask |= 1u << i;
                        sctx->descriptors_dirty |= 1u << SI_DESCS_RW_BUFFERS;
 
                        radeon_add_to_buffer_list_check_mem(&sctx->b, &sctx->b.gfx,
@@ -1827,7 +1681,6 @@ static void si_rebind_buffer(struct pipe_context *ctx, struct pipe_resource *buf
                                                                    descs->list +
                                                                    desc_slot * 16 + 4,
                                                                    old_va, buf);
-                                       descs->dirty_mask |= 1ull << desc_slot;
                                        sctx->descriptors_dirty |=
                                                1u << si_sampler_and_image_descriptors_idx(shader);
 
@@ -1860,8 +1713,6 @@ static void si_rebind_buffer(struct pipe_context *ctx, struct pipe_resource *buf
                                        si_desc_reset_buffer_offset(
                                                ctx, descs->list + desc_slot * 8 + 4,
                                                old_va, buf);
-                                       /* two 8-byte images share one 16-byte slot */
-                                       descs->dirty_mask |= 1u << (desc_slot / 2);
                                        sctx->descriptors_dirty |=
                                                1u << si_sampler_and_image_descriptors_idx(shader);
 
@@ -2328,8 +2179,7 @@ static void si_init_bindless_descriptors(struct si_context *sctx,
                                         unsigned shader_userdata_index,
                                         unsigned num_elements)
 {
-       si_init_descriptors(sctx, desc, shader_userdata_index, 16, num_elements,
-                           0, 0, NULL);
+       si_init_descriptors(desc, shader_userdata_index, 16, num_elements);
        sctx->bindless_descriptors.num_active_slots = num_elements;
 
        /* The first bindless descriptor is stored at slot 1, because 0 is not
@@ -2753,52 +2603,9 @@ void si_all_resident_buffers_begin_new_cs(struct si_context *sctx)
 
 /* INIT/DEINIT/UPLOAD */
 
-/* GFX9 has only 4KB of CE, while previous chips had 32KB. In order
- * to make CE RAM as useful as possible, this defines limits
- * for the number slots that can be in CE RAM on GFX9. If a shader
- * is using more, descriptors will be uploaded to memory directly and
- * CE won't be used.
- *
- * These numbers are based on shader-db.
- */
-static unsigned gfx9_max_ce_samplers[SI_NUM_SHADERS] = {
-       [PIPE_SHADER_VERTEX] = 0,
-       [PIPE_SHADER_TESS_CTRL] = 0,
-       [PIPE_SHADER_TESS_EVAL] = 1,
-       [PIPE_SHADER_GEOMETRY] = 0,
-       [PIPE_SHADER_FRAGMENT] = 24,
-       [PIPE_SHADER_COMPUTE] = 16,
-};
-static unsigned gfx9_max_ce_images[SI_NUM_SHADERS] = {
-       /* these must be even due to slot alignment */
-       [PIPE_SHADER_VERTEX] = 0,
-       [PIPE_SHADER_TESS_CTRL] = 0,
-       [PIPE_SHADER_TESS_EVAL] = 0,
-       [PIPE_SHADER_GEOMETRY] = 0,
-       [PIPE_SHADER_FRAGMENT] = 2,
-       [PIPE_SHADER_COMPUTE] = 8,
-};
-static unsigned gfx9_max_ce_const_buffers[SI_NUM_SHADERS] = {
-       [PIPE_SHADER_VERTEX] = 9,
-       [PIPE_SHADER_TESS_CTRL] = 3,
-       [PIPE_SHADER_TESS_EVAL] = 5,
-       [PIPE_SHADER_GEOMETRY] = 0,
-       [PIPE_SHADER_FRAGMENT] = 8,
-       [PIPE_SHADER_COMPUTE] = 6,
-};
-static unsigned gfx9_max_ce_shader_buffers[SI_NUM_SHADERS] = {
-       [PIPE_SHADER_VERTEX] = 0,
-       [PIPE_SHADER_TESS_CTRL] = 0,
-       [PIPE_SHADER_TESS_EVAL] = 0,
-       [PIPE_SHADER_GEOMETRY] = 0,
-       [PIPE_SHADER_FRAGMENT] = 12,
-       [PIPE_SHADER_COMPUTE] = 13,
-};
-
 void si_init_all_descriptors(struct si_context *sctx)
 {
        int i;
-       unsigned ce_offset = 0;
 
        STATIC_ASSERT(GFX9_SGPR_TCS_CONST_AND_SHADER_BUFFERS % 2 == 0);
        STATIC_ASSERT(GFX9_SGPR_GS_CONST_AND_SHADER_BUFFERS % 2 == 0);
@@ -2809,49 +2616,28 @@ void si_init_all_descriptors(struct si_context *sctx)
                unsigned num_sampler_slots = SI_NUM_IMAGES / 2 + SI_NUM_SAMPLERS;
                unsigned num_buffer_slots = SI_NUM_SHADER_BUFFERS + SI_NUM_CONST_BUFFERS;
 
-               unsigned first_sampler_ce_slot = 0;
-               unsigned num_sampler_ce_slots = num_sampler_slots;
-
-               unsigned first_buffer_ce_slot = 0;
-               unsigned num_buffer_ce_slots = num_buffer_slots;
-
-               /* Adjust CE slot ranges based on GFX9 CE RAM limits. */
                if (sctx->b.chip_class >= GFX9) {
                        gfx9_tcs = i == PIPE_SHADER_TESS_CTRL;
                        gfx9_gs = i == PIPE_SHADER_GEOMETRY;
-
-                       first_sampler_ce_slot =
-                               si_get_image_slot(gfx9_max_ce_images[i] - 1) / 2;
-                       num_sampler_ce_slots = gfx9_max_ce_images[i] / 2 +
-                                              gfx9_max_ce_samplers[i];
-
-                       first_buffer_ce_slot =
-                               si_get_shaderbuf_slot(gfx9_max_ce_shader_buffers[i] - 1);
-                       num_buffer_ce_slots = gfx9_max_ce_shader_buffers[i] +
-                                             gfx9_max_ce_const_buffers[i];
                }
 
-               si_init_buffer_resources(sctx, &sctx->const_and_shader_buffers[i],
+               si_init_buffer_resources(&sctx->const_and_shader_buffers[i],
                                         si_const_and_shader_buffer_descriptors(sctx, i),
                                         num_buffer_slots,
-                                        first_buffer_ce_slot, num_buffer_ce_slots,
                                         gfx9_tcs ? GFX9_SGPR_TCS_CONST_AND_SHADER_BUFFERS :
                                         gfx9_gs ? GFX9_SGPR_GS_CONST_AND_SHADER_BUFFERS :
                                                   SI_SGPR_CONST_AND_SHADER_BUFFERS,
                                         RADEON_USAGE_READWRITE,
                                         RADEON_USAGE_READ,
                                         RADEON_PRIO_SHADER_RW_BUFFER,
-                                        RADEON_PRIO_CONST_BUFFER,
-                                        &ce_offset);
+                                        RADEON_PRIO_CONST_BUFFER);
 
                struct si_descriptors *desc = si_sampler_and_image_descriptors(sctx, i);
-               si_init_descriptors(sctx, desc,
+               si_init_descriptors(desc,
                                    gfx9_tcs ? GFX9_SGPR_TCS_SAMPLERS_AND_IMAGES :
                                    gfx9_gs ? GFX9_SGPR_GS_SAMPLERS_AND_IMAGES :
                                              SI_SGPR_SAMPLERS_AND_IMAGES,
-                                   16, num_sampler_slots,
-                                   first_sampler_ce_slot, num_sampler_ce_slots,
-                                   &ce_offset);
+                                   16, num_sampler_slots);
 
                int j;
                for (j = 0; j < SI_NUM_IMAGES; j++)
@@ -2860,19 +2646,17 @@ void si_init_all_descriptors(struct si_context *sctx)
                        memcpy(desc->list + j * 8, null_texture_descriptor, 8 * 4);
        }
 
-       si_init_buffer_resources(sctx, &sctx->rw_buffers,
+       si_init_buffer_resources(&sctx->rw_buffers,
                                 &sctx->descriptors[SI_DESCS_RW_BUFFERS],
-                                SI_NUM_RW_BUFFERS, 0, SI_NUM_RW_BUFFERS,
-                                SI_SGPR_RW_BUFFERS,
+                                SI_NUM_RW_BUFFERS, SI_SGPR_RW_BUFFERS,
                                 /* The second set of usage/priority is used by
                                  * const buffers in RW buffer slots. */
                                 RADEON_USAGE_READWRITE, RADEON_USAGE_READ,
-                                RADEON_PRIO_SHADER_RINGS, RADEON_PRIO_CONST_BUFFER,
-                                &ce_offset);
+                                RADEON_PRIO_SHADER_RINGS, RADEON_PRIO_CONST_BUFFER);
        sctx->descriptors[SI_DESCS_RW_BUFFERS].num_active_slots = SI_NUM_RW_BUFFERS;
 
-       si_init_descriptors(sctx, &sctx->vertex_buffers, SI_SGPR_VERTEX_BUFFERS,
-                           4, SI_NUM_VERTEX_BUFFERS, 0, 0, NULL);
+       si_init_descriptors(&sctx->vertex_buffers, SI_SGPR_VERTEX_BUFFERS,
+                           4, SI_NUM_VERTEX_BUFFERS);
        FREE(sctx->vertex_buffers.list); /* not used */
        sctx->vertex_buffers.list = NULL;
 
@@ -2884,9 +2668,6 @@ void si_init_all_descriptors(struct si_context *sctx)
                                     1024);
 
        sctx->descriptors_dirty = u_bit_consecutive(0, SI_NUM_DESCS);
-       sctx->total_ce_ram_allocated = ce_offset;
-
-       assert(ce_offset <= si_ce_ram_size(sctx));
 
        /* Set pipe_context functions. */
        sctx->b.b.bind_sampler_states = si_bind_sampler_states;
@@ -3026,26 +2807,6 @@ void si_set_active_descriptors(struct si_context *sctx, unsigned desc_idx,
            first + count > desc->first_active_slot + desc->num_active_slots)
                sctx->descriptors_dirty |= 1u << desc_idx;
 
-       /* Enable or disable CE for this descriptor array. */
-       bool used_ce = desc->uses_ce;
-       desc->uses_ce = desc->first_ce_slot <= first &&
-                       desc->first_ce_slot + desc->num_ce_slots >= first + count;
-
-       if (desc->uses_ce != used_ce) {
-               /* Upload or dump descriptors if we're disabling or enabling CE,
-                * respectively. */
-               sctx->descriptors_dirty |= 1u << desc_idx;
-
-               /* If we're enabling CE, re-upload all descriptors to CE RAM.
-                * When CE was disabled, uploads to CE RAM stopped.
-                */
-               if (desc->uses_ce) {
-                       desc->dirty_mask |=
-                               u_bit_consecutive64(desc->first_ce_slot,
-                                                   desc->num_ce_slots);
-               }
-       }
-
        desc->first_active_slot = first;
        desc->num_active_slots = count;
 }
index 7481d013f3c36c48b4acf10afbbe40b3e9489b98..f3af8dcb4460fe0525cfc1ae7dd4c2cab48fb064 100644 (file)
 #include "si_pipe.h"
 #include "radeon/r600_cs.h"
 
-static unsigned si_descriptor_list_cs_space(unsigned count, unsigned element_size)
-{
-       /* Ensure we have enough space to start a new range in a hole */
-       assert(element_size >= 3);
-
-       /* 5 dwords for write to L2 + 3 bytes for the packet header of
-        * every disjoint range written to CE RAM.
-        */
-       return 5 + (3 * count / 2) + count * element_size;
-}
-
-static unsigned si_ce_needed_cs_space(void)
-{
-       unsigned space = 0;
-
-       space += si_descriptor_list_cs_space(SI_NUM_SHADER_BUFFERS +
-                                            SI_NUM_CONST_BUFFERS, 4);
-       /* two 8-byte images share one 16-byte slot */
-       space += si_descriptor_list_cs_space(SI_NUM_IMAGES / 2 +
-                                            SI_NUM_SAMPLERS, 16);
-       space *= SI_NUM_SHADERS;
-
-       space += si_descriptor_list_cs_space(SI_NUM_RW_BUFFERS, 4);
-
-       /* Increment CE counter packet */
-       space += 2;
-
-       return space;
-}
-
 void si_destroy_saved_cs(struct si_saved_cs *scs)
 {
        radeon_clear_saved_cs(&scs->gfx);
-       radeon_clear_saved_cs(&scs->ce);
        r600_resource_reference(&scs->trace_buf, NULL);
        free(scs);
 }
@@ -69,7 +38,6 @@ void si_destroy_saved_cs(struct si_saved_cs *scs)
 void si_need_cs_space(struct si_context *ctx)
 {
        struct radeon_winsys_cs *cs = ctx->b.gfx.cs;
-       struct radeon_winsys_cs *ce_ib = ctx->ce_ib;
 
        /* There is no need to flush the DMA IB here, because
         * r600_need_dma_space always flushes the GFX IB if there is
@@ -95,8 +63,7 @@ void si_need_cs_space(struct si_context *ctx)
        /* If the CS is sufficiently large, don't count the space needed
         * and just flush if there is not enough space left.
         */
-       if (!ctx->b.ws->cs_check_space(cs, 2048) ||
-           (ce_ib && !ctx->b.ws->cs_check_space(ce_ib, si_ce_needed_cs_space())))
+       if (!ctx->b.ws->cs_check_space(cs, 2048))
                ctx->b.gfx.flush(ctx, RADEON_FLUSH_ASYNC, NULL);
 }
 
@@ -131,10 +98,6 @@ void si_context_gfx_flush(void *context, unsigned flags,
 
        ctx->gfx_flush_in_progress = true;
 
-       /* This CE dump should be done in parallel with the last draw. */
-       if (ctx->ce_ib)
-               si_ce_save_all_descriptors_at_ib_end(ctx);
-
        r600_preflush_suspend_features(&ctx->b);
 
        ctx->b.flags |= SI_CONTEXT_CS_PARTIAL_FLUSH |
@@ -153,8 +116,6 @@ void si_context_gfx_flush(void *context, unsigned flags,
 
                /* Save the IB for debug contexts. */
                radeon_save_cs(ws, cs, &ctx->current_saved_cs->gfx, true);
-               if (ctx->ce_ib)
-                       radeon_save_cs(ws, ctx->ce_ib, &ctx->current_saved_cs->ce, false);
                ctx->current_saved_cs->flushed = true;
        }
 
@@ -183,7 +144,7 @@ void si_context_gfx_flush(void *context, unsigned flags,
 
 static void si_begin_cs_debug(struct si_context *ctx)
 {
-       static const uint32_t zeros[2];
+       static const uint32_t zeros[1];
        assert(!ctx->current_saved_cs);
 
        ctx->current_saved_cs = calloc(1, sizeof(*ctx->current_saved_cs));
@@ -233,14 +194,6 @@ void si_begin_new_cs(struct si_context *ctx)
        if (ctx->init_config_gs_rings)
                si_pm4_emit(ctx, ctx->init_config_gs_rings);
 
-       if (ctx->ce_preamble_ib)
-               si_ce_enable_loads(ctx->ce_preamble_ib);
-       else if (ctx->ce_ib)
-               si_ce_enable_loads(ctx->ce_ib);
-
-       if (ctx->ce_ib)
-               si_ce_restore_all_descriptors_at_ib_start(ctx);
-
        if (ctx->queued.named.ls)
                ctx->prefetch_L2_mask |= SI_PREFETCH_LS;
        if (ctx->queued.named.hs)
index 1523eaed94185fbf03d7319780592452056ca187..3ceaaac165a9e8e49075da277a825f741afb1a19 100644 (file)
@@ -55,10 +55,6 @@ static void si_destroy_context(struct pipe_context *context)
 
        si_release_all_descriptors(sctx);
 
-       if (sctx->ce_suballocator)
-               u_suballocator_destroy(sctx->ce_suballocator);
-
-       r600_resource_reference(&sctx->ce_ram_saved_buffer, NULL);
        pipe_resource_reference(&sctx->esgs_ring, NULL);
        pipe_resource_reference(&sctx->gsvs_ring, NULL);
        pipe_resource_reference(&sctx->tf_ring, NULL);
@@ -210,45 +206,6 @@ static struct pipe_context *si_create_context(struct pipe_screen *screen,
 
        sctx->b.gfx.cs = ws->cs_create(sctx->b.ctx, RING_GFX,
                                       si_context_gfx_flush, sctx);
-
-       bool enable_ce = sscreen->b.chip_class != SI && /* SI hangs */
-                        /* These can't use CE due to a power gating bug in the kernel. */
-                        sscreen->b.family != CHIP_CARRIZO &&
-                        sscreen->b.family != CHIP_STONEY;
-
-       /* CE is currently disabled by default, because it makes s_load latency
-        * worse, because CE IB doesn't run in lockstep with DE.
-        * Remove this line after that performance issue has been resolved.
-        */
-       enable_ce = false;
-
-       /* Apply CE overrides. */
-       if (sscreen->b.debug_flags & DBG_NO_CE)
-               enable_ce = false;
-       else if (sscreen->b.debug_flags & DBG_CE)
-               enable_ce = true;
-
-       if (ws->cs_add_const_ib && enable_ce) {
-               sctx->ce_ib = ws->cs_add_const_ib(sctx->b.gfx.cs);
-               if (!sctx->ce_ib)
-                       goto fail;
-
-               if (ws->cs_add_const_preamble_ib) {
-                       sctx->ce_preamble_ib =
-                                  ws->cs_add_const_preamble_ib(sctx->b.gfx.cs);
-
-                       if (!sctx->ce_preamble_ib)
-                               goto fail;
-               }
-
-               sctx->ce_suballocator =
-                       u_suballocator_create(&sctx->b.b, 1024 * 1024, 0,
-                                             PIPE_USAGE_DEFAULT,
-                                             R600_RESOURCE_FLAG_UNMAPPABLE, false);
-               if (!sctx->ce_suballocator)
-                       goto fail;
-       }
-
        sctx->b.gfx.flush = si_context_gfx_flush;
 
        /* Border colors. */
index fd99c975ad40888f3babf7b49c647594d49d69f6..69a35ea1945af3ccf4ef44162a390eaca099fcc4 100644 (file)
@@ -266,12 +266,10 @@ struct si_saved_cs {
        struct pipe_reference   reference;
        struct si_context       *ctx;
        struct radeon_saved_cs  gfx;
-       struct radeon_saved_cs  ce;
        struct r600_resource    *trace_buf;
        unsigned                trace_id;
 
        unsigned                gfx_last_dw;
-       unsigned                ce_last_dw;
        bool                    flushed;
 };
 
@@ -288,15 +286,7 @@ struct si_context {
        struct si_shader_ctx_state      fixed_func_tcs_shader;
        struct r600_resource            *wait_mem_scratch;
        unsigned                        wait_mem_number;
-
-       struct radeon_winsys_cs         *ce_ib;
-       struct radeon_winsys_cs         *ce_preamble_ib;
-       struct r600_resource            *ce_ram_saved_buffer;
-       struct u_suballocator           *ce_suballocator;
-       unsigned                        ce_ram_saved_offset;
-       uint16_t                        total_ce_ram_allocated;
        uint16_t                        prefetch_L2_mask;
-       bool                            ce_need_synchronization:1;
 
        bool                            gfx_flush_in_progress:1;
        bool                            compute_is_busy:1;
index 2b3c37fa16d867760cccf7994fc5eb87ab65ae99..ca701658d0b66ac30425d05ba901da88c35715ea 100644 (file)
@@ -224,8 +224,6 @@ struct si_descriptors {
        uint32_t *list;
        /* The list in mapped GPU memory. */
        uint32_t *gpu_list;
-       /* Slots that have been changed and need to be uploaded. */
-       uint64_t dirty_mask;
 
        /* The buffer where the descriptors have been uploaded. */
        struct r600_resource *buffer;
@@ -236,27 +234,12 @@ struct si_descriptors {
        /* The maximum number of descriptors. */
        uint32_t num_elements;
 
-       /* Offset in CE RAM */
-       uint16_t ce_offset;
-
-       /* Slots allocated in CE RAM. If we get active slots outside of this
-        * range, direct uploads to memory will be used instead. This basically
-        * governs switching between onchip (CE) and offchip (upload) modes.
-        */
-       uint32_t first_ce_slot;
-       uint32_t num_ce_slots;
-
        /* Slots that are used by currently-bound shaders.
-        * With CE: It determines which slots are dumped to L2.
-        *          It doesn't skip uploads to CE RAM.
-        * Without CE: It determines which slots are uploaded.
+        * It determines which slots are uploaded.
         */
        uint32_t first_active_slot;
        uint32_t num_active_slots;
 
-       /* Whether CE is used to upload this descriptor array. */
-       bool uses_ce;
-
        /* The SGPR index where the 64-bit pointer to the descriptor array will
         * be stored. */
        ubyte shader_userdata_offset;
@@ -307,9 +290,6 @@ struct si_buffer_resources {
        } while(0)
 
 /* si_descriptors.c */
-void si_ce_save_all_descriptors_at_ib_end(struct si_context* sctx);
-void si_ce_restore_all_descriptors_at_ib_start(struct si_context *sctx);
-void si_ce_enable_loads(struct radeon_winsys_cs *ib);
 void si_set_mutable_tex_desc_fields(struct si_screen *sscreen,
                                    struct r600_texture *tex,
                                    const struct legacy_surf_level *base_level_info,
@@ -401,8 +381,6 @@ void si_get_active_slot_masks(const struct tgsi_shader_info *info,
 /* si_state_draw.c */
 void si_init_ia_multi_vgt_param_table(struct si_context *sctx);
 void si_emit_cache_flush(struct si_context *sctx);
-void si_ce_pre_draw_synchronization(struct si_context *sctx);
-void si_ce_post_draw_synchronization(struct si_context *sctx);
 void si_draw_vbo(struct pipe_context *ctx, const struct pipe_draw_info *dinfo);
 void si_trace_emit(struct si_context *sctx);
 
index cb9bbd20805bf2007f3747664b57817d1c1bdfe3..f2b889677a5e99ad7ae72cef4681012e981928f2 100644 (file)
@@ -1145,27 +1145,6 @@ static void si_get_draw_start_count(struct si_context *sctx,
        }
 }
 
-void si_ce_pre_draw_synchronization(struct si_context *sctx)
-{
-       if (sctx->ce_need_synchronization) {
-               radeon_emit(sctx->ce_ib, PKT3(PKT3_INCREMENT_CE_COUNTER, 0, 0));
-               radeon_emit(sctx->ce_ib, 1); /* 1 = increment CE counter */
-
-               radeon_emit(sctx->b.gfx.cs, PKT3(PKT3_WAIT_ON_CE_COUNTER, 0, 0));
-               radeon_emit(sctx->b.gfx.cs, 0); /* 0 = don't flush sL1 conditionally */
-       }
-}
-
-void si_ce_post_draw_synchronization(struct si_context *sctx)
-{
-       if (sctx->ce_need_synchronization) {
-               radeon_emit(sctx->b.gfx.cs, PKT3(PKT3_INCREMENT_DE_COUNTER, 0, 0));
-               radeon_emit(sctx->b.gfx.cs, 0); /* unused */
-
-               sctx->ce_need_synchronization = false;
-       }
-}
-
 static void si_emit_all_states(struct si_context *sctx, const struct pipe_draw_info *info,
                               unsigned skip_atom_mask)
 {
@@ -1413,7 +1392,6 @@ void si_draw_vbo(struct pipe_context *ctx, const struct pipe_draw_info *info)
                        sctx->dirty_atoms = 0;
                }
 
-               si_ce_pre_draw_synchronization(sctx);
                si_emit_draw_packets(sctx, info, indexbuf, index_size, index_offset);
                /* <-- CUs are busy here. */
 
@@ -1436,12 +1414,9 @@ void si_draw_vbo(struct pipe_context *ctx, const struct pipe_draw_info *info)
                        return;
 
                si_emit_all_states(sctx, info, 0);
-               si_ce_pre_draw_synchronization(sctx);
                si_emit_draw_packets(sctx, info, indexbuf, index_size, index_offset);
        }
 
-       si_ce_post_draw_synchronization(sctx);
-
        if (unlikely(sctx->current_saved_cs))
                si_trace_emit(sctx);
 
@@ -1485,20 +1460,6 @@ void si_trace_emit(struct si_context *sctx)
        radeon_emit(cs, PKT3(PKT3_NOP, 0, 0));
        radeon_emit(cs, AC_ENCODE_TRACE_POINT(trace_id));
 
-       if (sctx->ce_ib) {
-               struct radeon_winsys_cs *ce = sctx->ce_ib;
-
-               radeon_emit(ce, PKT3(PKT3_WRITE_DATA, 3, 0));
-               radeon_emit(ce, S_370_DST_SEL(V_370_MEM_ASYNC) |
-                           S_370_WR_CONFIRM(1) |
-                           S_370_ENGINE_SEL(V_370_CE));
-               radeon_emit(ce, va + 4);
-               radeon_emit(ce, (va + 4) >> 32);
-               radeon_emit(ce, trace_id);
-               radeon_emit(ce, PKT3(PKT3_NOP, 0, 0));
-               radeon_emit(ce, AC_ENCODE_TRACE_POINT(trace_id));
-       }
-
        if (sctx->b.log)
                u_log_flush(sctx->b.log);
 }
index d2662538865b49683d6dd121453381bec37f7329..9cadfc4298dff56d752690566aeee47fc13a8ad9 100644 (file)
@@ -566,12 +566,6 @@ static bool amdgpu_ib_new_buffer(struct amdgpu_winsys *ws, struct amdgpu_ib *ib)
    buffer_size = MIN2(buffer_size, 4 * 512 * 1024);
 
    switch (ib->ib_type) {
-   case IB_CONST_PREAMBLE:
-      buffer_size = MAX2(buffer_size, 4 * 1024);
-      break;
-   case IB_CONST:
-      buffer_size = MAX2(buffer_size, 16 * 1024 * 4);
-      break;
    case IB_MAIN:
       buffer_size = MAX2(buffer_size, 8 * 1024 * 4);
       break;
@@ -609,13 +603,6 @@ static unsigned amdgpu_ib_max_submit_dwords(enum ib_type ib_type)
        *   http://www.phoronix.com/scan.php?page=article&item=mesa-111-si&num=1
        */
       return 20 * 1024;
-   case IB_CONST_PREAMBLE:
-   case IB_CONST:
-      /* There isn't really any reason to limit CE IB size beyond the natural
-       * limit implied by the main IB, except perhaps GTT size. Just return
-       * an extremely large value that we never get anywhere close to.
-       */
-      return 16 * 1024 * 1024;
    default:
       unreachable("bad ib_type");
    }
@@ -634,14 +621,6 @@ static bool amdgpu_get_new_ib(struct radeon_winsys *ws, struct amdgpu_cs *cs,
    unsigned ib_size = 0;
 
    switch (ib_type) {
-   case IB_CONST_PREAMBLE:
-      ib = &cs->const_preamble_ib;
-      ib_size = 256 * 4;
-      break;
-   case IB_CONST:
-      ib = &cs->const_ib;
-      ib_size = 8 * 1024 * 4;
-      break;
    case IB_MAIN:
       ib = &cs->main;
       ib_size = 4 * 1024 * 4;
@@ -728,10 +707,6 @@ static bool amdgpu_init_cs_context(struct amdgpu_cs_context *cs,
    cs->request.number_of_ibs = 1;
    cs->request.ibs = &cs->ib[IB_MAIN];
 
-   cs->ib[IB_CONST].flags = AMDGPU_IB_FLAG_CE;
-   cs->ib[IB_CONST_PREAMBLE].flags = AMDGPU_IB_FLAG_CE |
-                                     AMDGPU_IB_FLAG_PREAMBLE;
-
    return true;
 }
 
@@ -799,8 +774,6 @@ amdgpu_cs_create(struct radeon_winsys_ctx *rwctx,
    cs->ring_type = ring_type;
 
    cs->main.ib_type = IB_MAIN;
-   cs->const_ib.ib_type = IB_CONST;
-   cs->const_preamble_ib.ib_type = IB_CONST_PREAMBLE;
 
    if (!amdgpu_init_cs_context(&cs->csc1, ring_type)) {
       FREE(cs);
@@ -828,52 +801,6 @@ amdgpu_cs_create(struct radeon_winsys_ctx *rwctx,
    return &cs->main.base;
 }
 
-static struct radeon_winsys_cs *
-amdgpu_cs_add_const_ib(struct radeon_winsys_cs *rcs)
-{
-   struct amdgpu_cs *cs = (struct amdgpu_cs*)rcs;
-   struct amdgpu_winsys *ws = cs->ctx->ws;
-
-   /* only one const IB can be added */
-   if (cs->ring_type != RING_GFX || cs->const_ib.ib_mapped)
-      return NULL;
-
-   if (!amdgpu_get_new_ib(&ws->base, cs, IB_CONST))
-      return NULL;
-
-   cs->csc->request.number_of_ibs = 2;
-   cs->csc->request.ibs = &cs->csc->ib[IB_CONST];
-
-   cs->cst->request.number_of_ibs = 2;
-   cs->cst->request.ibs = &cs->cst->ib[IB_CONST];
-
-   return &cs->const_ib.base;
-}
-
-static struct radeon_winsys_cs *
-amdgpu_cs_add_const_preamble_ib(struct radeon_winsys_cs *rcs)
-{
-   struct amdgpu_cs *cs = (struct amdgpu_cs*)rcs;
-   struct amdgpu_winsys *ws = cs->ctx->ws;
-
-   /* only one const preamble IB can be added and only when the const IB has
-    * also been mapped */
-   if (cs->ring_type != RING_GFX || !cs->const_ib.ib_mapped ||
-       cs->const_preamble_ib.ib_mapped)
-      return NULL;
-
-   if (!amdgpu_get_new_ib(&ws->base, cs, IB_CONST_PREAMBLE))
-      return NULL;
-
-   cs->csc->request.number_of_ibs = 3;
-   cs->csc->request.ibs = &cs->csc->ib[IB_CONST_PREAMBLE];
-
-   cs->cst->request.number_of_ibs = 3;
-   cs->cst->request.ibs = &cs->cst->ib[IB_CONST_PREAMBLE];
-
-   return &cs->const_preamble_ib.base;
-}
-
 static bool amdgpu_cs_validate(struct radeon_winsys_cs *rcs)
 {
    return true;
@@ -1323,15 +1250,6 @@ static int amdgpu_cs_flush(struct radeon_winsys_cs *rcs,
          while (rcs->current.cdw & 7)
             radeon_emit(rcs, 0xffff1000); /* type3 nop packet */
       }
-
-      /* Also pad the const IB. */
-      if (cs->const_ib.ib_mapped)
-         while (!cs->const_ib.base.current.cdw || (cs->const_ib.base.current.cdw & 7))
-            radeon_emit(&cs->const_ib.base, 0xffff1000); /* type3 nop packet */
-
-      if (cs->const_preamble_ib.ib_mapped)
-         while (!cs->const_preamble_ib.base.current.cdw || (cs->const_preamble_ib.base.current.cdw & 7))
-            radeon_emit(&cs->const_preamble_ib.base, 0xffff1000);
       break;
    case RING_UVD:
       while (rcs->current.cdw & 15)
@@ -1358,12 +1276,6 @@ static int amdgpu_cs_flush(struct radeon_winsys_cs *rcs,
       /* Set IB sizes. */
       amdgpu_ib_finalize(&cs->main);
 
-      if (cs->const_ib.ib_mapped)
-         amdgpu_ib_finalize(&cs->const_ib);
-
-      if (cs->const_preamble_ib.ib_mapped)
-         amdgpu_ib_finalize(&cs->const_preamble_ib);
-
       /* Create a fence. */
       amdgpu_fence_reference(&cur->fence, NULL);
       if (cs->next_fence) {
@@ -1409,10 +1321,6 @@ static int amdgpu_cs_flush(struct radeon_winsys_cs *rcs,
    }
 
    amdgpu_get_new_ib(&ws->base, cs, IB_MAIN);
-   if (cs->const_ib.ib_mapped)
-      amdgpu_get_new_ib(&ws->base, cs, IB_CONST);
-   if (cs->const_preamble_ib.ib_mapped)
-      amdgpu_get_new_ib(&ws->base, cs, IB_CONST_PREAMBLE);
 
    cs->main.base.used_gart = 0;
    cs->main.base.used_vram = 0;
@@ -1434,10 +1342,6 @@ static void amdgpu_cs_destroy(struct radeon_winsys_cs *rcs)
    p_atomic_dec(&cs->ctx->ws->num_cs);
    pb_reference(&cs->main.big_ib_buffer, NULL);
    FREE(cs->main.base.prev);
-   pb_reference(&cs->const_ib.big_ib_buffer, NULL);
-   FREE(cs->const_ib.base.prev);
-   pb_reference(&cs->const_preamble_ib.big_ib_buffer, NULL);
-   FREE(cs->const_preamble_ib.base.prev);
    amdgpu_destroy_cs_context(&cs->csc1);
    amdgpu_destroy_cs_context(&cs->csc2);
    amdgpu_fence_reference(&cs->next_fence, NULL);
@@ -1460,8 +1364,6 @@ void amdgpu_cs_init_functions(struct amdgpu_winsys *ws)
    ws->base.ctx_destroy = amdgpu_ctx_destroy;
    ws->base.ctx_query_reset_status = amdgpu_ctx_query_reset_status;
    ws->base.cs_create = amdgpu_cs_create;
-   ws->base.cs_add_const_ib = amdgpu_cs_add_const_ib;
-   ws->base.cs_add_const_preamble_ib = amdgpu_cs_add_const_preamble_ib;
    ws->base.cs_destroy = amdgpu_cs_destroy;
    ws->base.cs_add_buffer = amdgpu_cs_add_buffer;
    ws->base.cs_validate = amdgpu_cs_validate;
index d83c1e0fe19c87afdb15fc925c4d7967c2009a38..8f5c33678eb133080ef57f908c16080313c79784 100644 (file)
@@ -59,10 +59,8 @@ struct amdgpu_cs_buffer {
 };
 
 enum ib_type {
-   IB_CONST_PREAMBLE = 0,
-   IB_CONST = 1, /* the const IB must be first */
-   IB_MAIN = 2,
-   IB_NUM
+   IB_MAIN,
+   IB_NUM,
 };
 
 struct amdgpu_ib {
@@ -117,8 +115,6 @@ struct amdgpu_cs_context {
 
 struct amdgpu_cs {
    struct amdgpu_ib main; /* must be first because this is inherited */
-   struct amdgpu_ib const_ib; /* optional constant engine IB */
-   struct amdgpu_ib const_preamble_ib;
    struct amdgpu_ctx *ctx;
    enum ring_type ring_type;
 
@@ -199,10 +195,6 @@ amdgpu_cs_from_ib(struct amdgpu_ib *ib)
    switch (ib->ib_type) {
    case IB_MAIN:
       return get_container(ib, struct amdgpu_cs, main);
-   case IB_CONST:
-      return get_container(ib, struct amdgpu_cs, const_ib);
-   case IB_CONST_PREAMBLE:
-      return get_container(ib, struct amdgpu_cs, const_preamble_ib);
    default:
       unreachable("bad ib_type");
    }