radeonsi: prepare for driver-specific driconf options
[mesa.git] / src / gallium / drivers / radeonsi / si_pipe.c
index c1480d4bf277767986c8950d7c75fa15c7fae56d..d276b885c112591d91d508f2b63a6e670725924f 100644 (file)
 #include "sid.h"
 
 #include "radeon/radeon_uvd.h"
+#include "util/hash_table.h"
 #include "util/u_memory.h"
 #include "util/u_suballoc.h"
+#include "util/u_tests.h"
 #include "vl/vl_decoder.h"
 #include "../ddebug/dd_util.h"
 
-#define SI_LLVM_DEFAULT_FEATURES \
-       "+DumpCode,+vgpr-spilling,-fp32-denormals,-xnack"
+#include "compiler/nir/nir.h"
 
 /*
  * pipe_context
@@ -47,13 +48,15 @@ static void si_destroy_context(struct pipe_context *context)
         * properly.
         */
        struct pipe_framebuffer_state fb = {};
-       context->set_framebuffer_state(context, &fb);
+       if (context->set_framebuffer_state)
+               context->set_framebuffer_state(context, &fb);
 
        si_release_all_descriptors(sctx);
 
        if (sctx->ce_suballocator)
                u_suballocator_destroy(sctx->ce_suballocator);
 
+       r600_resource_reference(&sctx->ce_ram_saved_buffer, NULL);
        pipe_resource_reference(&sctx->esgs_ring, NULL);
        pipe_resource_reference(&sctx->gsvs_ring, NULL);
        pipe_resource_reference(&sctx->tf_ring, NULL);
@@ -63,6 +66,7 @@ static void si_destroy_context(struct pipe_context *context)
        free(sctx->border_color_table);
        r600_resource_reference(&sctx->scratch_buffer, NULL);
        r600_resource_reference(&sctx->compute_scratch_buffer, NULL);
+       r600_resource_reference(&sctx->wait_mem_scratch, NULL);
 
        si_pm4_free_state(sctx, sctx->init_config, ~0);
        if (sctx->init_config_gs_rings)
@@ -76,10 +80,10 @@ static void si_destroy_context(struct pipe_context *context)
                sctx->b.b.delete_depth_stencil_alpha_state(&sctx->b.b, sctx->custom_dsa_flush);
        if (sctx->custom_blend_resolve)
                sctx->b.b.delete_blend_state(&sctx->b.b, sctx->custom_blend_resolve);
-       if (sctx->custom_blend_decompress)
-               sctx->b.b.delete_blend_state(&sctx->b.b, sctx->custom_blend_decompress);
-       if (sctx->custom_blend_fastclear)
-               sctx->b.b.delete_blend_state(&sctx->b.b, sctx->custom_blend_fastclear);
+       if (sctx->custom_blend_fmask_decompress)
+               sctx->b.b.delete_blend_state(&sctx->b.b, sctx->custom_blend_fmask_decompress);
+       if (sctx->custom_blend_eliminate_fastclear)
+               sctx->b.b.delete_blend_state(&sctx->b.b, sctx->custom_blend_eliminate_fastclear);
        if (sctx->custom_blend_dcc_decompress)
                sctx->b.b.delete_blend_state(&sctx->b.b, sctx->custom_blend_dcc_decompress);
 
@@ -94,6 +98,17 @@ static void si_destroy_context(struct pipe_context *context)
        r600_resource_reference(&sctx->last_trace_buf, NULL);
        radeon_clear_saved_cs(&sctx->last_gfx);
 
+       pb_slabs_deinit(&sctx->bindless_descriptor_slabs);
+       util_dynarray_fini(&sctx->bindless_descriptors);
+
+       _mesa_hash_table_destroy(sctx->tex_handles, NULL);
+       _mesa_hash_table_destroy(sctx->img_handles, NULL);
+
+       util_dynarray_fini(&sctx->resident_tex_handles);
+       util_dynarray_fini(&sctx->resident_img_handles);
+       util_dynarray_fini(&sctx->resident_tex_needs_color_decompress);
+       util_dynarray_fini(&sctx->resident_img_needs_color_decompress);
+       util_dynarray_fini(&sctx->resident_tex_needs_depth_decompress);
        FREE(sctx);
 }
 
@@ -125,21 +140,24 @@ static LLVMTargetMachineRef
 si_create_llvm_target_machine(struct si_screen *sscreen)
 {
        const char *triple = "amdgcn--";
+       char features[256];
 
-       return LLVMCreateTargetMachine(si_llvm_get_amdgpu_target(triple), triple,
+       snprintf(features, sizeof(features),
+                "+DumpCode,+vgpr-spilling,-fp32-denormals,+fp64-denormals%s%s%s",
+                sscreen->b.chip_class >= GFX9 ? ",+xnack" : ",-xnack",
+                sscreen->llvm_has_working_vgpr_indexing ? "" : ",-promote-alloca",
+                sscreen->b.debug_flags & DBG_SI_SCHED ? ",+si-scheduler" : "");
+
+       return LLVMCreateTargetMachine(ac_get_llvm_target(triple), triple,
                                       r600_get_llvm_processor_name(sscreen->b.family),
-#if HAVE_LLVM >= 0x0308
-                                      sscreen->b.debug_flags & DBG_SI_SCHED ?
-                                              SI_LLVM_DEFAULT_FEATURES ",+si-scheduler" :
-#endif
-                                              SI_LLVM_DEFAULT_FEATURES,
+                                      features,
                                       LLVMCodeGenLevelDefault,
                                       LLVMRelocDefault,
                                       LLVMCodeModelDefault);
 }
 
 static struct pipe_context *si_create_context(struct pipe_screen *screen,
-                                              void *priv, unsigned flags)
+                                              unsigned flags)
 {
        struct si_context *sctx = CALLOC_STRUCT(si_context);
        struct si_screen* sscreen = (struct si_screen *)screen;
@@ -149,14 +167,11 @@ static struct pipe_context *si_create_context(struct pipe_screen *screen,
        if (!sctx)
                return NULL;
 
-       if (sscreen->b.debug_flags & DBG_CHECK_VM)
-               flags |= PIPE_CONTEXT_DEBUG;
-
        if (flags & PIPE_CONTEXT_DEBUG)
                sscreen->record_llvm_ir = true; /* racy but not critical */
 
        sctx->b.b.screen = screen; /* this must be set first */
-       sctx->b.b.priv = priv;
+       sctx->b.b.priv = NULL;
        sctx->b.b.destroy = si_destroy_context;
        sctx->b.b.emit_string_marker = si_emit_string_marker;
        sctx->b.set_atom_dirty = (void *)si_set_atom_dirty;
@@ -174,7 +189,7 @@ static struct pipe_context *si_create_context(struct pipe_screen *screen,
        si_init_cp_dma_functions(sctx);
        si_init_debug_functions(sctx);
 
-       if (sscreen->b.info.has_uvd) {
+       if (sscreen->b.info.has_hw_decode) {
                sctx->b.b.create_video_codec = si_uvd_create_decoder;
                sctx->b.b.create_video_buffer = si_video_buffer_create;
        } else {
@@ -204,8 +219,9 @@ static struct pipe_context *si_create_context(struct pipe_screen *screen,
                }
 
                sctx->ce_suballocator =
-                               u_suballocator_create(&sctx->b.b, 1024 * 1024,
-                                                     0, PIPE_USAGE_DEFAULT, false);
+                       u_suballocator_create(&sctx->b.b, 1024 * 1024, 0,
+                                             PIPE_USAGE_DEFAULT,
+                                             R600_RESOURCE_FLAG_UNMAPPABLE, false);
                if (!sctx->ce_suballocator)
                        goto fail;
        }
@@ -234,6 +250,7 @@ static struct pipe_context *si_create_context(struct pipe_screen *screen,
        si_init_all_descriptors(sctx);
        si_init_state_functions(sctx);
        si_init_shader_functions(sctx);
+       si_init_ia_multi_vgt_param_table(sctx);
 
        if (sctx->b.chip_class >= CIK)
                cik_init_sdma_functions(sctx);
@@ -252,13 +269,32 @@ static struct pipe_context *si_create_context(struct pipe_screen *screen,
 
        /* these must be last */
        si_begin_new_cs(sctx);
-       r600_query_init_backend_mask(&sctx->b); /* this emits commands and must be last */
+
+       if (sctx->b.chip_class >= GFX9) {
+               sctx->wait_mem_scratch = (struct r600_resource*)
+                       pipe_buffer_create(screen, 0, PIPE_USAGE_DEFAULT, 4);
+               if (!sctx->wait_mem_scratch)
+                       goto fail;
+
+               /* Initialize the memory. */
+               struct radeon_winsys_cs *cs = sctx->b.gfx.cs;
+               radeon_emit(cs, PKT3(PKT3_WRITE_DATA, 3, 0));
+               radeon_emit(cs, S_370_DST_SEL(V_370_MEMORY_SYNC) |
+                           S_370_WR_CONFIRM(1) |
+                           S_370_ENGINE_SEL(V_370_ME));
+               radeon_emit(cs, sctx->wait_mem_scratch->gpu_address);
+               radeon_emit(cs, sctx->wait_mem_scratch->gpu_address >> 32);
+               radeon_emit(cs, sctx->wait_mem_number);
+       }
 
        /* CIK cannot unbind a constant buffer (S_BUFFER_LOAD doesn't skip loads
         * if NUM_RECORDS == 0). We need to use a dummy buffer instead. */
        if (sctx->b.chip_class == CIK) {
-               sctx->null_const_buf.buffer = pipe_buffer_create(screen, PIPE_BIND_CONSTANT_BUFFER,
-                                                                PIPE_USAGE_DEFAULT, 16);
+               sctx->null_const_buf.buffer =
+                       r600_aligned_buffer_create(screen,
+                                                  R600_RESOURCE_FLAG_UNMAPPABLE,
+                                                  PIPE_USAGE_DEFAULT, 16,
+                                                  sctx->screen->b.info.tcc_cache_line_size);
                if (!sctx->null_const_buf.buffer)
                        goto fail;
                sctx->null_const_buf.buffer_size = sctx->null_const_buf.buffer->width0;
@@ -272,6 +308,8 @@ static struct pipe_context *si_create_context(struct pipe_screen *screen,
 
                si_set_rw_buffer(sctx, SI_HS_CONST_DEFAULT_TESS_LEVELS,
                                 &sctx->null_const_buf);
+               si_set_rw_buffer(sctx, SI_VS_CONST_INSTANCE_DIVISORS,
+                                &sctx->null_const_buf);
                si_set_rw_buffer(sctx, SI_VS_CONST_CLIP_PLANES,
                                 &sctx->null_const_buf);
                si_set_rw_buffer(sctx, SI_PS_CONST_POLY_STIPPLE,
@@ -307,6 +345,27 @@ static struct pipe_context *si_create_context(struct pipe_screen *screen,
 
        sctx->tm = si_create_llvm_target_machine(sscreen);
 
+       /* Create a slab allocator for all bindless descriptors. */
+       if (!pb_slabs_init(&sctx->bindless_descriptor_slabs, 6, 6, 1, sctx,
+                          si_bindless_descriptor_can_reclaim_slab,
+                          si_bindless_descriptor_slab_alloc,
+                          si_bindless_descriptor_slab_free))
+               goto fail;
+
+       util_dynarray_init(&sctx->bindless_descriptors, NULL);
+
+       /* Bindless handles. */
+       sctx->tex_handles = _mesa_hash_table_create(NULL, _mesa_hash_pointer,
+                                                   _mesa_key_pointer_equal);
+       sctx->img_handles = _mesa_hash_table_create(NULL, _mesa_hash_pointer,
+                                                   _mesa_key_pointer_equal);
+
+       util_dynarray_init(&sctx->resident_tex_handles, NULL);
+       util_dynarray_init(&sctx->resident_img_handles, NULL);
+       util_dynarray_init(&sctx->resident_tex_needs_color_decompress, NULL);
+       util_dynarray_init(&sctx->resident_img_needs_color_decompress, NULL);
+       util_dynarray_init(&sctx->resident_tex_needs_depth_decompress, NULL);
+
        return &sctx->b.b;
 fail:
        fprintf(stderr, "radeonsi: Failed to create a context.\n");
@@ -314,6 +373,42 @@ fail:
        return NULL;
 }
 
+static struct pipe_context *si_pipe_create_context(struct pipe_screen *screen,
+                                                  void *priv, unsigned flags)
+{
+       struct si_screen *sscreen = (struct si_screen *)screen;
+       struct pipe_context *ctx;
+
+       if (sscreen->b.debug_flags & DBG_CHECK_VM)
+               flags |= PIPE_CONTEXT_DEBUG;
+
+       ctx = si_create_context(screen, flags);
+
+       if (!(flags & PIPE_CONTEXT_PREFER_THREADED))
+               return ctx;
+
+       /* Clover (compute-only) is unsupported.
+        *
+        * Since the threaded context creates shader states from the non-driver
+        * thread, asynchronous compilation is required for create_{shader}_-
+        * state not to use pipe_context. Debug contexts (ddebug) disable
+        * asynchronous compilation, so don't use the threaded context with
+        * those.
+        */
+       if (flags & (PIPE_CONTEXT_COMPUTE_ONLY | PIPE_CONTEXT_DEBUG))
+               return ctx;
+
+       /* When shaders are logged to stderr, asynchronous compilation is
+        * disabled too. */
+       if (sscreen->b.debug_flags & (DBG_VS | DBG_TCS | DBG_TES | DBG_GS |
+                                     DBG_PS | DBG_CS))
+               return ctx;
+
+       return threaded_context_create(ctx, &sscreen->b.pool_transfers,
+                                      r600_replace_buffer_storage,
+                                      &((struct si_context*)ctx)->b.tc);
+}
+
 /*
  * pipe_screen
  */
@@ -321,8 +416,7 @@ static bool si_have_tgsi_compute(struct si_screen *sscreen)
 {
        /* Old kernels disallowed some register writes for SI
         * that are used for indirect dispatches. */
-       return HAVE_LLVM >= 0x309 &&
-              (sscreen->b.chip_class >= CIK ||
+       return (sscreen->b.chip_class >= CIK ||
                sscreen->b.info.drm_major == 3 ||
                (sscreen->b.info.drm_major == 2 &&
                 sscreen->b.info.drm_minor >= 45));
@@ -334,6 +428,7 @@ static int si_get_param(struct pipe_screen* pscreen, enum pipe_cap param)
 
        switch (param) {
        /* Supported features (boolean caps). */
+       case PIPE_CAP_ACCELERATED:
        case PIPE_CAP_TWO_SIDED_STENCIL:
        case PIPE_CAP_MAX_DUAL_SOURCE_RENDER_TARGETS:
        case PIPE_CAP_ANISOTROPIC_FILTER:
@@ -359,10 +454,6 @@ static int si_get_param(struct pipe_screen* pscreen, enum pipe_cap param)
        case PIPE_CAP_INDEP_BLEND_FUNC:
        case PIPE_CAP_SEAMLESS_CUBE_MAP_PER_TEXTURE:
        case PIPE_CAP_VERTEX_COLOR_UNCLAMPED:
-       case PIPE_CAP_VERTEX_BUFFER_OFFSET_4BYTE_ALIGNED_ONLY:
-       case PIPE_CAP_VERTEX_BUFFER_STRIDE_4BYTE_ALIGNED_ONLY:
-       case PIPE_CAP_VERTEX_ELEMENT_SRC_OFFSET_4BYTE_ALIGNED_ONLY:
-       case PIPE_CAP_USER_INDEX_BUFFERS:
        case PIPE_CAP_USER_CONSTANT_BUFFERS:
        case PIPE_CAP_START_INSTANCE:
        case PIPE_CAP_NPOT_TEXTURES:
@@ -414,9 +505,31 @@ static int si_get_param(struct pipe_screen* pscreen, enum pipe_cap param)
        case PIPE_CAP_TGSI_ARRAY_COMPONENTS:
        case PIPE_CAP_TGSI_CAN_READ_OUTPUTS:
        case PIPE_CAP_GLSL_OPTIMIZE_CONSERVATIVELY:
+       case PIPE_CAP_STREAM_OUTPUT_PAUSE_RESUME:
+       case PIPE_CAP_STREAM_OUTPUT_INTERLEAVE_BUFFERS:
+       case PIPE_CAP_DOUBLES:
+       case PIPE_CAP_TGSI_TEX_TXF_LZ:
+       case PIPE_CAP_TGSI_TES_LAYER_VIEWPORT:
+       case PIPE_CAP_BINDLESS_TEXTURE:
+       case PIPE_CAP_QUERY_TIMESTAMP:
+       case PIPE_CAP_QUERY_TIME_ELAPSED:
+       case PIPE_CAP_NIR_SAMPLERS_AS_DEREF:
+       case PIPE_CAP_QUERY_SO_OVERFLOW:
+               return 1;
+
        case PIPE_CAP_INT64:
+       case PIPE_CAP_INT64_DIVMOD:
+       case PIPE_CAP_TGSI_CLOCK:
+       case PIPE_CAP_CAN_BIND_CONST_BUFFER_AS_VERTEX:
+       case PIPE_CAP_ALLOW_MAPPED_BUFFERS_DURING_EXECUTION:
                return 1;
 
+       case PIPE_CAP_TGSI_VOTE:
+               return HAVE_LLVM >= 0x0400;
+
+       case PIPE_CAP_TGSI_BALLOT:
+               return HAVE_LLVM >= 0x0500;
+
        case PIPE_CAP_RESOURCE_FROM_USER_MEMORY:
                return !SI_BIG_ENDIAN && sscreen->b.info.has_userptr;
 
@@ -438,23 +551,45 @@ static int si_get_param(struct pipe_screen* pscreen, enum pipe_cap param)
        case PIPE_CAP_CONSTANT_BUFFER_OFFSET_ALIGNMENT:
        case PIPE_CAP_TEXTURE_BUFFER_OFFSET_ALIGNMENT:
        case PIPE_CAP_MAX_TEXTURE_GATHER_COMPONENTS:
-               return 4;
+       case PIPE_CAP_MAX_STREAM_OUTPUT_BUFFERS:
+       case PIPE_CAP_MAX_VERTEX_STREAMS:
        case PIPE_CAP_SHADER_BUFFER_OFFSET_ALIGNMENT:
-               return HAVE_LLVM >= 0x0309 ? 4 : 0;
+               return 4;
 
        case PIPE_CAP_GLSL_FEATURE_LEVEL:
+               if (sscreen->b.debug_flags & DBG_NIR)
+                       return 140; /* no geometry and tessellation shaders yet */
                if (si_have_tgsi_compute(sscreen))
                        return 450;
-               return HAVE_LLVM >= 0x0309 ? 420 :
-                      HAVE_LLVM >= 0x0307 ? 410 : 330;
+               return 420;
 
        case PIPE_CAP_MAX_TEXTURE_BUFFER_SIZE:
                return MIN2(sscreen->b.info.max_alloc_size, INT_MAX);
 
-       case PIPE_CAP_BUFFER_SAMPLER_VIEW_RGBA_ONLY:
+       case PIPE_CAP_VERTEX_BUFFER_OFFSET_4BYTE_ALIGNED_ONLY:
+       case PIPE_CAP_VERTEX_BUFFER_STRIDE_4BYTE_ALIGNED_ONLY:
+       case PIPE_CAP_VERTEX_ELEMENT_SRC_OFFSET_4BYTE_ALIGNED_ONLY:
+               /* SI doesn't support unaligned loads.
+                * CIK needs DRM 2.50.0 on radeon. */
+               return sscreen->b.chip_class == SI ||
+                      (sscreen->b.info.drm_major == 2 &&
+                       sscreen->b.info.drm_minor < 50);
+
+       case PIPE_CAP_SPARSE_BUFFER_PAGE_SIZE:
+               /* TODO: GFX9 hangs. */
+               if (sscreen->b.chip_class >= GFX9)
+                       return 0;
+               /* Disable on SI due to VM faults in CP DMA. Enable once these
+                * faults are mitigated in software.
+                */
+               if (sscreen->b.chip_class >= CIK &&
+                   sscreen->b.info.drm_major == 3 &&
+                   sscreen->b.info.drm_minor >= 13)
+                       return RADEON_SPARSE_PAGE_SIZE;
                return 0;
 
        /* Unsupported features. */
+       case PIPE_CAP_BUFFER_SAMPLER_VIEW_RGBA_ONLY:
        case PIPE_CAP_TGSI_FS_COORD_ORIGIN_LOWER_LEFT:
        case PIPE_CAP_TGSI_CAN_COMPACT_CONSTANTS:
        case PIPE_CAP_USER_VERTEX_BUFFERS:
@@ -462,11 +597,13 @@ static int si_get_param(struct pipe_screen* pscreen, enum pipe_cap param)
        case PIPE_CAP_TEXTURE_GATHER_OFFSETS:
        case PIPE_CAP_VERTEXID_NOBASE:
        case PIPE_CAP_PRIMITIVE_RESTART_FOR_PATCHES:
-       case PIPE_CAP_TGSI_VOTE:
        case PIPE_CAP_MAX_WINDOW_RECTANGLES:
        case PIPE_CAP_NATIVE_FENCE_FD:
        case PIPE_CAP_TGSI_FS_FBFETCH:
        case PIPE_CAP_TGSI_MUL_ZERO_WINS:
+       case PIPE_CAP_UMA:
+       case PIPE_CAP_POLYGON_MODE_FILL_RECTANGLE:
+       case PIPE_CAP_POST_DEPTH_COVERAGE:
                return 0;
 
        case PIPE_CAP_QUERY_BUFFER_OBJECT:
@@ -481,25 +618,19 @@ static int si_get_param(struct pipe_screen* pscreen, enum pipe_cap param)
                return 30;
 
        case PIPE_CAP_TEXTURE_BORDER_COLOR_QUIRK:
-               return PIPE_QUIRK_TEXTURE_BORDER_COLOR_SWIZZLE_R600;
+               return sscreen->b.chip_class <= VI ?
+                       PIPE_QUIRK_TEXTURE_BORDER_COLOR_SWIZZLE_R600 : 0;
 
        /* Stream output. */
-       case PIPE_CAP_MAX_STREAM_OUTPUT_BUFFERS:
-               return sscreen->b.has_streamout ? 4 : 0;
-       case PIPE_CAP_STREAM_OUTPUT_PAUSE_RESUME:
-       case PIPE_CAP_STREAM_OUTPUT_INTERLEAVE_BUFFERS:
-               return sscreen->b.has_streamout ? 1 : 0;
        case PIPE_CAP_MAX_STREAM_OUTPUT_SEPARATE_COMPONENTS:
        case PIPE_CAP_MAX_STREAM_OUTPUT_INTERLEAVED_COMPONENTS:
-               return sscreen->b.has_streamout ? 32*4 : 0;
+               return 32*4;
 
        /* Geometry shader output. */
        case PIPE_CAP_MAX_GEOMETRY_OUTPUT_VERTICES:
                return 1024;
        case PIPE_CAP_MAX_GEOMETRY_TOTAL_OUTPUT_COMPONENTS:
                return 4095;
-       case PIPE_CAP_MAX_VERTEX_STREAMS:
-               return 4;
 
        case PIPE_CAP_MAX_VERTEX_ATTRIB_STRIDE:
                return 2048;
@@ -515,20 +646,13 @@ static int si_get_param(struct pipe_screen* pscreen, enum pipe_cap param)
                /* textures support 8192, but layered rendering supports 2048 */
                return 2048;
 
-       /* Render targets. */
-       case PIPE_CAP_MAX_RENDER_TARGETS:
-               return 8;
-
+       /* Viewports and render targets. */
        case PIPE_CAP_MAX_VIEWPORTS:
                return R600_MAX_VIEWPORTS;
        case PIPE_CAP_VIEWPORT_SUBPIXEL_BITS:
+       case PIPE_CAP_MAX_RENDER_TARGETS:
                return 8;
 
-       /* Timer queries, present when the clock frequency is non zero. */
-       case PIPE_CAP_QUERY_TIMESTAMP:
-       case PIPE_CAP_QUERY_TIME_ELAPSED:
-               return sscreen->b.info.clock_crystal_freq != 0;
-
        case PIPE_CAP_MIN_TEXTURE_GATHER_OFFSET:
        case PIPE_CAP_MIN_TEXEL_OFFSET:
                return -32;
@@ -544,12 +668,8 @@ static int si_get_param(struct pipe_screen* pscreen, enum pipe_cap param)
                return ATI_VENDOR_ID;
        case PIPE_CAP_DEVICE_ID:
                return sscreen->b.info.pci_id;
-       case PIPE_CAP_ACCELERATED:
-               return 1;
        case PIPE_CAP_VIDEO_MEMORY:
                return sscreen->b.info.vram_size >> 20;
-       case PIPE_CAP_UMA:
-               return 0;
        case PIPE_CAP_PCI_GROUP:
                return sscreen->b.info.pci_domain;
        case PIPE_CAP_PCI_BUS:
@@ -562,7 +682,9 @@ static int si_get_param(struct pipe_screen* pscreen, enum pipe_cap param)
        return 0;
 }
 
-static int si_get_shader_param(struct pipe_screen* pscreen, unsigned shader, enum pipe_shader_cap param)
+static int si_get_shader_param(struct pipe_screen* pscreen,
+                              enum pipe_shader_type shader,
+                              enum pipe_shader_cap param)
 {
        struct si_screen *sscreen = (struct si_screen *)pscreen;
 
@@ -571,12 +693,8 @@ static int si_get_shader_param(struct pipe_screen* pscreen, unsigned shader, enu
        case PIPE_SHADER_FRAGMENT:
        case PIPE_SHADER_VERTEX:
        case PIPE_SHADER_GEOMETRY:
-               break;
        case PIPE_SHADER_TESS_CTRL:
        case PIPE_SHADER_TESS_EVAL:
-               /* LLVM 3.6.2 is required for tessellation because of bug fixes there */
-               if (HAVE_LLVM == 0x0306 && MESA_LLVM_VERSION_PATCH < 2)
-                       return 0;
                break;
        case PIPE_SHADER_COMPUTE:
                switch (param) {
@@ -591,8 +709,6 @@ static int si_get_shader_param(struct pipe_screen* pscreen, unsigned shader, enu
 
                        return ir;
                }
-               case PIPE_SHADER_CAP_DOUBLES:
-                       return HAVE_LLVM >= 0x0307;
 
                case PIPE_SHADER_CAP_MAX_CONST_BUFFER_SIZE: {
                        uint64_t max_const_buffer_size;
@@ -621,7 +737,7 @@ static int si_get_shader_param(struct pipe_screen* pscreen, unsigned shader, enu
        case PIPE_SHADER_CAP_MAX_CONTROL_FLOW_DEPTH:
                return 16384;
        case PIPE_SHADER_CAP_MAX_INPUTS:
-               return shader == PIPE_SHADER_VERTEX ? SI_NUM_VERTEX_BUFFERS : 32;
+               return shader == PIPE_SHADER_VERTEX ? SI_MAX_ATTRIBS : 32;
        case PIPE_SHADER_CAP_MAX_OUTPUTS:
                return shader == PIPE_SHADER_FRAGMENT ? 8 : 32;
        case PIPE_SHADER_CAP_MAX_TEMPS:
@@ -634,12 +750,16 @@ static int si_get_shader_param(struct pipe_screen* pscreen, unsigned shader, enu
        case PIPE_SHADER_CAP_MAX_SAMPLER_VIEWS:
                return SI_NUM_SAMPLERS;
        case PIPE_SHADER_CAP_MAX_SHADER_BUFFERS:
-               return HAVE_LLVM >= 0x0309 ? SI_NUM_SHADER_BUFFERS : 0;
+               return SI_NUM_SHADER_BUFFERS;
        case PIPE_SHADER_CAP_MAX_SHADER_IMAGES:
-               return HAVE_LLVM >= 0x0309 ? SI_NUM_IMAGES : 0;
+               return SI_NUM_IMAGES;
        case PIPE_SHADER_CAP_MAX_UNROLL_ITERATIONS_HINT:
                return 32;
        case PIPE_SHADER_CAP_PREFERRED_IR:
+               if (sscreen->b.debug_flags & DBG_NIR &&
+                   (shader == PIPE_SHADER_VERTEX ||
+                    shader == PIPE_SHADER_FRAGMENT))
+                       return PIPE_SHADER_IR_NIR;
                return PIPE_SHADER_IR_TGSI;
        case PIPE_SHADER_CAP_LOWER_IF_THRESHOLD:
                return 3;
@@ -647,25 +767,29 @@ static int si_get_shader_param(struct pipe_screen* pscreen, unsigned shader, enu
        /* Supported boolean features. */
        case PIPE_SHADER_CAP_TGSI_CONT_SUPPORTED:
        case PIPE_SHADER_CAP_TGSI_SQRT_SUPPORTED:
-       case PIPE_SHADER_CAP_INDIRECT_OUTPUT_ADDR:
        case PIPE_SHADER_CAP_INDIRECT_TEMP_ADDR:
        case PIPE_SHADER_CAP_INDIRECT_CONST_ADDR:
        case PIPE_SHADER_CAP_INTEGERS:
        case PIPE_SHADER_CAP_TGSI_FMA_SUPPORTED:
        case PIPE_SHADER_CAP_TGSI_ANY_INOUT_DECL_RANGE:
+       case PIPE_SHADER_CAP_TGSI_SKIP_MERGE_REGISTERS:
                return 1;
 
-       case PIPE_SHADER_CAP_DOUBLES:
-               return HAVE_LLVM >= 0x0307;
-
        case PIPE_SHADER_CAP_INDIRECT_INPUT_ADDR:
-               /* TODO: Indirection of geometry shader input dimension is not
-                * handled yet
-                */
-               return shader != PIPE_SHADER_GEOMETRY;
+               /* TODO: Indirect indexing of GS inputs is unimplemented. */
+               return shader != PIPE_SHADER_GEOMETRY &&
+                      (sscreen->llvm_has_working_vgpr_indexing ||
+                       /* TCS and TES load inputs directly from LDS or
+                        * offchip memory, so indirect indexing is trivial. */
+                       shader == PIPE_SHADER_TESS_CTRL ||
+                       shader == PIPE_SHADER_TESS_EVAL);
+
+       case PIPE_SHADER_CAP_INDIRECT_OUTPUT_ADDR:
+               return sscreen->llvm_has_working_vgpr_indexing ||
+                      /* TCS stores outputs directly to memory. */
+                      shader == PIPE_SHADER_TESS_CTRL;
 
        /* Unsupported boolean features. */
-       case PIPE_SHADER_CAP_MAX_PREDS:
        case PIPE_SHADER_CAP_SUBROUTINES:
        case PIPE_SHADER_CAP_SUPPORTED_IRS:
        case PIPE_SHADER_CAP_TGSI_DROUND_SUPPORTED:
@@ -675,12 +799,41 @@ static int si_get_shader_param(struct pipe_screen* pscreen, unsigned shader, enu
        return 0;
 }
 
+static const struct nir_shader_compiler_options nir_options = {
+       .vertex_id_zero_based = true,
+       .lower_scmp = true,
+       .lower_flrp32 = true,
+       .lower_fsat = true,
+       .lower_fdiv = true,
+       .lower_sub = true,
+       .lower_pack_snorm_2x16 = true,
+       .lower_pack_snorm_4x8 = true,
+       .lower_pack_unorm_2x16 = true,
+       .lower_pack_unorm_4x8 = true,
+       .lower_unpack_snorm_2x16 = true,
+       .lower_unpack_snorm_4x8 = true,
+       .lower_unpack_unorm_2x16 = true,
+       .lower_unpack_unorm_4x8 = true,
+       .lower_extract_byte = true,
+       .lower_extract_word = true,
+       .max_unroll_iterations = 32,
+       .native_integers = true,
+};
+
+static const void *
+si_get_compiler_options(struct pipe_screen *screen,
+                       enum pipe_shader_ir ir,
+                       enum pipe_shader_type shader)
+{
+       assert(ir == PIPE_SHADER_IR_NIR);
+       return &nir_options;
+}
+
 static void si_destroy_screen(struct pipe_screen* pscreen)
 {
        struct si_screen *sscreen = (struct si_screen *)pscreen;
        struct si_shader_part *parts[] = {
                sscreen->vs_prologs,
-               sscreen->vs_epilogs,
                sscreen->tcs_epilogs,
                sscreen->gs_prologs,
                sscreen->ps_prologs,
@@ -688,19 +841,20 @@ static void si_destroy_screen(struct pipe_screen* pscreen)
        };
        unsigned i;
 
-       if (!sscreen)
-               return;
-
        if (!sscreen->b.ws->unref(sscreen->b.ws))
                return;
 
-       if (util_queue_is_initialized(&sscreen->shader_compiler_queue))
-               util_queue_destroy(&sscreen->shader_compiler_queue);
+       util_queue_destroy(&sscreen->shader_compiler_queue);
+       util_queue_destroy(&sscreen->shader_compiler_queue_low_priority);
 
        for (i = 0; i < ARRAY_SIZE(sscreen->tm); i++)
                if (sscreen->tm[i])
                        LLVMDisposeTargetMachine(sscreen->tm[i]);
 
+       for (i = 0; i < ARRAY_SIZE(sscreen->tm_low_priority); i++)
+               if (sscreen->tm_low_priority[i])
+                       LLVMDisposeTargetMachine(sscreen->tm_low_priority[i]);
+
        /* Free shader parts. */
        for (i = 0; i < ARRAY_SIZE(parts); i++) {
                while (parts[i]) {
@@ -711,7 +865,7 @@ static void si_destroy_screen(struct pipe_screen* pscreen)
                        FREE(part);
                }
        }
-       pipe_mutex_destroy(sscreen->shader_parts_mutex);
+       mtx_destroy(&sscreen->shader_parts_mutex);
        si_destroy_shader_cache(sscreen);
        r600_destroy_common_screen(&sscreen->b);
 }
@@ -739,6 +893,8 @@ static bool si_init_gs_info(struct si_screen *sscreen)
        case CHIP_POLARIS10:
        case CHIP_POLARIS11:
        case CHIP_POLARIS12:
+       case CHIP_VEGA10:
+       case CHIP_RAVEN:
                sscreen->gs_table_depth = 32;
                return true;
        default:
@@ -759,7 +915,9 @@ static void si_handle_env_var_force_family(struct si_screen *sscreen)
                        /* Override family and chip_class. */
                        sscreen->b.family = sscreen->b.info.family = i;
 
-                       if (i >= CHIP_TONGA)
+                       if (i >= CHIP_VEGA10)
+                               sscreen->b.chip_class = sscreen->b.info.chip_class = GFX9;
+                       else if (i >= CHIP_TONGA)
                                sscreen->b.chip_class = sscreen->b.info.chip_class = VI;
                        else if (i >= CHIP_BONAIRE)
                                sscreen->b.chip_class = sscreen->b.info.chip_class = CIK;
@@ -776,31 +934,91 @@ static void si_handle_env_var_force_family(struct si_screen *sscreen)
        exit(1);
 }
 
-struct pipe_screen *radeonsi_screen_create(struct radeon_winsys *ws)
+static void si_test_vmfault(struct si_screen *sscreen)
+{
+       struct pipe_context *ctx = sscreen->b.aux_context;
+       struct si_context *sctx = (struct si_context *)ctx;
+       struct pipe_resource *buf =
+               pipe_buffer_create(&sscreen->b.b, 0, PIPE_USAGE_DEFAULT, 64);
+
+       if (!buf) {
+               puts("Buffer allocation failed.");
+               exit(1);
+       }
+
+       r600_resource(buf)->gpu_address = 0; /* cause a VM fault */
+
+       if (sscreen->b.debug_flags & DBG_TEST_VMFAULT_CP) {
+               si_copy_buffer(sctx, buf, buf, 0, 4, 4, 0);
+               ctx->flush(ctx, NULL, 0);
+               puts("VM fault test: CP - done.");
+       }
+       if (sscreen->b.debug_flags & DBG_TEST_VMFAULT_SDMA) {
+               sctx->b.dma_clear_buffer(ctx, buf, 0, 4, 0);
+               ctx->flush(ctx, NULL, 0);
+               puts("VM fault test: SDMA - done.");
+       }
+       if (sscreen->b.debug_flags & DBG_TEST_VMFAULT_SHADER) {
+               util_test_constant_buffer(ctx, buf);
+               puts("VM fault test: Shader - done.");
+       }
+       exit(0);
+}
+
+struct pipe_screen *radeonsi_screen_create(struct radeon_winsys *ws,
+                                          const struct pipe_screen_config *config)
 {
        struct si_screen *sscreen = CALLOC_STRUCT(si_screen);
-       unsigned num_cpus, num_compiler_threads, i;
+       unsigned num_threads, num_compiler_threads, num_compiler_threads_lowprio, i;
 
        if (!sscreen) {
                return NULL;
        }
 
        /* Set functions first. */
-       sscreen->b.b.context_create = si_create_context;
+       sscreen->b.b.context_create = si_pipe_create_context;
        sscreen->b.b.destroy = si_destroy_screen;
        sscreen->b.b.get_param = si_get_param;
        sscreen->b.b.get_shader_param = si_get_shader_param;
+       sscreen->b.b.get_compiler_options = si_get_compiler_options;
        sscreen->b.b.resource_create = r600_resource_create_common;
 
        si_init_screen_state_functions(sscreen);
 
-       if (!r600_common_screen_init(&sscreen->b, ws) ||
+       if (!r600_common_screen_init(&sscreen->b, ws, config->flags) ||
            !si_init_gs_info(sscreen) ||
            !si_init_shader_cache(sscreen)) {
                FREE(sscreen);
                return NULL;
        }
 
+       /* Only enable as many threads as we have target machines, but at most
+        * the number of CPUs - 1 if there is more than one.
+        */
+       num_threads = sysconf(_SC_NPROCESSORS_ONLN);
+       num_threads = MAX2(1, num_threads - 1);
+       num_compiler_threads = MIN2(num_threads, ARRAY_SIZE(sscreen->tm));
+       num_compiler_threads_lowprio =
+               MIN2(num_threads, ARRAY_SIZE(sscreen->tm_low_priority));
+
+       if (!util_queue_init(&sscreen->shader_compiler_queue, "si_shader",
+                            32, num_compiler_threads,
+                            UTIL_QUEUE_INIT_RESIZE_IF_FULL)) {
+               si_destroy_shader_cache(sscreen);
+               FREE(sscreen);
+               return NULL;
+       }
+
+       if (!util_queue_init(&sscreen->shader_compiler_queue_low_priority,
+                            "si_shader_low",
+                            32, num_compiler_threads_lowprio,
+                            UTIL_QUEUE_INIT_RESIZE_IF_FULL |
+                            UTIL_QUEUE_INIT_USE_MINIMUM_PRIORITY)) {
+              si_destroy_shader_cache(sscreen);
+              FREE(sscreen);
+              return NULL;
+       }
+
        si_handle_env_var_force_family(sscreen);
 
        if (!debug_get_bool_option("RADEON_DISABLE_PERFCOUNTERS", false))
@@ -825,42 +1043,65 @@ struct pipe_screen *radeonsi_screen_create(struct radeon_winsys *ws)
                 sscreen->b.info.pfp_fw_version >= 211 &&
                 sscreen->b.info.me_fw_version >= 173) ||
                (sscreen->b.chip_class == SI &&
-                sscreen->b.info.pfp_fw_version >= 121 &&
-                sscreen->b.info.me_fw_version >= 87);
-
-       sscreen->has_ds_bpermute = HAVE_LLVM >= 0x0309 &&
-                                  sscreen->b.chip_class >= VI;
+                sscreen->b.info.pfp_fw_version >= 79 &&
+                sscreen->b.info.me_fw_version >= 142);
+
+       sscreen->has_ds_bpermute = sscreen->b.chip_class >= VI;
+       sscreen->has_msaa_sample_loc_bug = (sscreen->b.family >= CHIP_POLARIS10 &&
+                                           sscreen->b.family <= CHIP_POLARIS12) ||
+                                          sscreen->b.family == CHIP_VEGA10 ||
+                                          sscreen->b.family == CHIP_RAVEN;
+       /* While it would be nice not to have this flag, we are constrained
+        * by the reality that LLVM 5.0 doesn't have working VGPR indexing
+        * on GFX9.
+        */
+       sscreen->llvm_has_working_vgpr_indexing = sscreen->b.chip_class <= VI;
 
        sscreen->b.has_cp_dma = true;
        sscreen->b.has_streamout = true;
-       pipe_mutex_init(sscreen->shader_parts_mutex);
+
+       /* Some chips have RB+ registers, but don't support RB+. Those must
+        * always disable it.
+        */
+       if (sscreen->b.family == CHIP_STONEY ||
+           sscreen->b.chip_class >= GFX9) {
+               sscreen->b.has_rbplus = true;
+
+               sscreen->b.rbplus_allowed =
+                       !(sscreen->b.debug_flags & DBG_NO_RB_PLUS) &&
+                       (sscreen->b.family == CHIP_STONEY ||
+                        sscreen->b.family == CHIP_RAVEN);
+       }
+
+       (void) mtx_init(&sscreen->shader_parts_mutex, mtx_plain);
        sscreen->use_monolithic_shaders =
-               HAVE_LLVM < 0x0308 ||
                (sscreen->b.debug_flags & DBG_MONOLITHIC_SHADERS) != 0;
 
        sscreen->b.barrier_flags.cp_to_L2 = SI_CONTEXT_INV_SMEM_L1 |
-                                           SI_CONTEXT_INV_VMEM_L1 |
-                                           SI_CONTEXT_INV_GLOBAL_L2;
+                                           SI_CONTEXT_INV_VMEM_L1;
+       if (sscreen->b.chip_class <= VI)
+               sscreen->b.barrier_flags.cp_to_L2 |= SI_CONTEXT_INV_GLOBAL_L2;
+
        sscreen->b.barrier_flags.compute_to_L2 = SI_CONTEXT_CS_PARTIAL_FLUSH;
 
        if (debug_get_bool_option("RADEON_DUMP_SHADERS", false))
                sscreen->b.debug_flags |= DBG_FS | DBG_VS | DBG_GS | DBG_PS | DBG_CS;
 
-       /* Only enable as many threads as we have target machines and CPUs. */
-       num_cpus = sysconf(_SC_NPROCESSORS_ONLN);
-       num_compiler_threads = MIN2(num_cpus, ARRAY_SIZE(sscreen->tm));
-
        for (i = 0; i < num_compiler_threads; i++)
                sscreen->tm[i] = si_create_llvm_target_machine(sscreen);
-
-       util_queue_init(&sscreen->shader_compiler_queue, "si_shader",
-                        32, num_compiler_threads);
+       for (i = 0; i < num_compiler_threads_lowprio; i++)
+               sscreen->tm_low_priority[i] = si_create_llvm_target_machine(sscreen);
 
        /* Create the auxiliary context. This must be done last. */
-       sscreen->b.aux_context = sscreen->b.b.context_create(&sscreen->b.b, NULL, 0);
+       sscreen->b.aux_context = si_create_context(&sscreen->b.b, 0);
 
        if (sscreen->b.debug_flags & DBG_TEST_DMA)
                r600_test_dma(&sscreen->b);
 
+       if (sscreen->b.debug_flags & (DBG_TEST_VMFAULT_CP |
+                                     DBG_TEST_VMFAULT_SDMA |
+                                     DBG_TEST_VMFAULT_SHADER))
+               si_test_vmfault(sscreen);
+
        return &sscreen->b.b;
 }