iris: Tidy exporting the flink handle
[mesa.git] / src / gallium / drivers / iris / iris_state.c
index 09a7bcc34de19919023afc022b80f1e7899cc481..a21fcc17dca28e3d10dadcc8db5fe532d0efcc4d 100644 (file)
@@ -485,22 +485,77 @@ _iris_emit_lri(struct iris_batch *batch, uint32_t reg, uint32_t val)
 }
 #define iris_emit_lri(b, r, v) _iris_emit_lri(b, GENX(r##_num), v)
 
-/**
- * Upload the initial GPU state for a render context.
- *
- * This sets some invariant state that needs to be programmed a particular
- * way, but we never actually change.
- */
 static void
-iris_init_render_context(struct iris_screen *screen,
-                         struct iris_batch *batch,
-                         struct iris_vtable *vtbl,
-                         struct pipe_debug_callback *dbg)
+emit_pipeline_select(struct iris_batch *batch, uint32_t pipeline)
 {
-   uint32_t reg_val;
+#if GEN_GEN >= 8 && GEN_GEN < 10
+   /* From the Broadwell PRM, Volume 2a: Instructions, PIPELINE_SELECT:
+    *
+    *   Software must clear the COLOR_CALC_STATE Valid field in
+    *   3DSTATE_CC_STATE_POINTERS command prior to send a PIPELINE_SELECT
+    *   with Pipeline Select set to GPGPU.
+    *
+    * The internal hardware docs recommend the same workaround for Gen9
+    * hardware too.
+    */
+   if (pipeline == GPGPU)
+      iris_emit_cmd(batch, GENX(3DSTATE_CC_STATE_POINTERS), t);
+#endif
+
 
-   iris_init_batch(batch, screen, vtbl, dbg, I915_EXEC_RENDER);
+   /* From "BXML » GT » MI » vol1a GPU Overview » [Instruction]
+    * PIPELINE_SELECT [DevBWR+]":
+    *
+    *    "Project: DEVSNB+
+    *
+    *     Software must ensure all the write caches are flushed through a
+    *     stalling PIPE_CONTROL command followed by another PIPE_CONTROL
+    *     command to invalidate read only caches prior to programming
+    *     MI_PIPELINE_SELECT command to change the Pipeline Select Mode."
+    */
+    iris_emit_pipe_control_flush(batch,
+                                 PIPE_CONTROL_RENDER_TARGET_FLUSH |
+                                 PIPE_CONTROL_DEPTH_CACHE_FLUSH |
+                                 PIPE_CONTROL_DATA_CACHE_FLUSH |
+                                 PIPE_CONTROL_CS_STALL);
+
+    iris_emit_pipe_control_flush(batch,
+                                 PIPE_CONTROL_TEXTURE_CACHE_INVALIDATE |
+                                 PIPE_CONTROL_CONST_CACHE_INVALIDATE |
+                                 PIPE_CONTROL_STATE_CACHE_INVALIDATE |
+                                 PIPE_CONTROL_INSTRUCTION_INVALIDATE);
+
+   iris_emit_cmd(batch, GENX(PIPELINE_SELECT), sel) {
+#if GEN_GEN >= 9
+      sel.MaskBits = 3;
+#endif
+      sel.PipelineSelection = pipeline;
+   }
+}
+
+UNUSED static void
+init_glk_barrier_mode(struct iris_batch *batch, uint32_t value)
+{
+#if GEN_GEN == 9
+   /* Project: DevGLK
+    *
+    *    "This chicken bit works around a hardware issue with barrier
+    *     logic encountered when switching between GPGPU and 3D pipelines.
+    *     To workaround the issue, this mode bit should be set after a
+    *     pipeline is selected."
+    */
+   uint32_t reg_val;
+   iris_pack_state(GENX(SLICE_COMMON_ECO_CHICKEN1), &reg_val, reg) {
+      reg.GLKBarrierMode = value;
+      reg.GLKBarrierModeMask = 1;
+   }
+   iris_emit_lri(batch, SLICE_COMMON_ECO_CHICKEN1, reg_val);
+#endif
+}
 
+static void
+init_state_base_address(struct iris_batch *batch)
+{
    flush_for_state_base_change(batch);
 
    /* We program most base addresses once at context initialization time.
@@ -539,6 +594,26 @@ iris_init_render_context(struct iris_screen *screen,
       sba.InstructionBufferSize    = 0xfffff;
       sba.DynamicStateBufferSize   = 0xfffff;
    }
+}
+
+/**
+ * Upload the initial GPU state for a render context.
+ *
+ * This sets some invariant state that needs to be programmed a particular
+ * way, but we never actually change.
+ */
+static void
+iris_init_render_context(struct iris_screen *screen,
+                         struct iris_batch *batch,
+                         struct iris_vtable *vtbl,
+                         struct pipe_debug_callback *dbg)
+{
+   UNUSED const struct gen_device_info *devinfo = &screen->devinfo;
+   uint32_t reg_val;
+
+   emit_pipeline_select(batch, _3D);
+
+   init_state_base_address(batch);
 
    // XXX: INSTPM on Gen8
    iris_pack_state(GENX(CS_DEBUG_MODE2), &reg_val, reg) {
@@ -555,6 +630,9 @@ iris_init_render_context(struct iris_screen *screen,
       reg.PartialResolveDisableInVCMask = true;
    }
    iris_emit_lri(batch, CACHE_MODE_1, reg_val);
+
+   if (devinfo->is_geminilake)
+      init_glk_barrier_mode(batch, GLK_BARRIER_MODE_3D_HULL);
 #endif
 
 #if GEN_GEN == 11
@@ -616,49 +694,40 @@ iris_init_compute_context(struct iris_screen *screen,
                           struct iris_vtable *vtbl,
                           struct pipe_debug_callback *dbg)
 {
-   iris_init_batch(batch, screen, vtbl, dbg, I915_EXEC_RENDER);
+   UNUSED const struct gen_device_info *devinfo = &screen->devinfo;
 
-   /* XXX: PIPE_CONTROLs */
+   emit_pipeline_select(batch, GPGPU);
 
-   iris_emit_cmd(batch, GENX(PIPELINE_SELECT), sel) {
-#if GEN_GEN >= 9
-      sel.MaskBits = 3;
-#endif
-      sel.PipelineSelection = GPGPU;
-   }
+   const bool has_slm = true;
+   const bool wants_dc_cache = true;
 
-   iris_emit_cmd(batch, GENX(STATE_BASE_ADDRESS), sba) {
-   #if 0
-   // XXX: MOCS is stupid for this.
-      sba.GeneralStateMemoryObjectControlState            = MOCS_WB;
-      sba.StatelessDataPortAccessMemoryObjectControlState = MOCS_WB;
-      sba.SurfaceStateMemoryObjectControlState            = MOCS_WB;
-      sba.DynamicStateMemoryObjectControlState            = MOCS_WB;
-      sba.IndirectObjectMemoryObjectControlState          = MOCS_WB;
-      sba.InstructionMemoryObjectControlState             = MOCS_WB;
-      sba.BindlessSurfaceStateMemoryObjectControlState    = MOCS_WB;
-   #endif
+   const struct gen_l3_weights w =
+      gen_get_default_l3_weights(devinfo, wants_dc_cache, has_slm);
+   const struct gen_l3_config *cfg = gen_get_l3_config(devinfo, w);
 
-      sba.GeneralStateBaseAddressModifyEnable   = true;
-      sba.SurfaceStateBaseAddressModifyEnable   = true;
-      sba.DynamicStateBaseAddressModifyEnable   = true;
-      sba.IndirectObjectBaseAddressModifyEnable = true;
-      sba.InstructionBaseAddressModifyEnable    = true;
-      sba.GeneralStateBufferSizeModifyEnable    = true;
-      sba.DynamicStateBufferSizeModifyEnable    = true;
-      sba.BindlessSurfaceStateBaseAddressModifyEnable = true;
-      sba.IndirectObjectBufferSizeModifyEnable  = true;
-      sba.InstructionBuffersizeModifyEnable     = true;
+   uint32_t reg_val;
+   iris_pack_state(GENX(L3CNTLREG), &reg_val, reg) {
+      reg.SLMEnable = has_slm;
+#if GEN_GEN == 11
+      /* WA_1406697149: Bit 9 "Error Detection Behavior Control" must be set
+       * in L3CNTLREG register. The default setting of the bit is not the
+       * desirable behavior.
+       */
+      reg.ErrorDetectionBehaviorControl = true;
+#endif
+      reg.URBAllocation = cfg->n[GEN_L3P_URB];
+      reg.ROAllocation = cfg->n[GEN_L3P_RO];
+      reg.DCAllocation = cfg->n[GEN_L3P_DC];
+      reg.AllAllocation = cfg->n[GEN_L3P_ALL];
+   }
+   iris_emit_lri(batch, L3CNTLREG, reg_val);
 
-      sba.InstructionBaseAddress  = ro_bo(NULL, IRIS_MEMZONE_SHADER_START);
-      sba.SurfaceStateBaseAddress = ro_bo(NULL, IRIS_MEMZONE_SURFACE_START);
-      sba.DynamicStateBaseAddress = ro_bo(NULL, IRIS_MEMZONE_DYNAMIC_START);
+   init_state_base_address(batch);
 
-      sba.GeneralStateBufferSize   = 0xfffff;
-      sba.IndirectObjectBufferSize = 0xfffff;
-      sba.InstructionBufferSize    = 0xfffff;
-      sba.DynamicStateBufferSize   = 0xfffff;
-   }
+#if GEN_GEN == 9
+   if (devinfo->is_geminilake)
+      init_glk_barrier_mode(batch, GLK_BARRIER_MODE_GPGPU);
+#endif
 }
 
 struct iris_vertex_buffer_state {
@@ -737,61 +806,70 @@ iris_create_blend_state(struct pipe_context *ctx,
                         const struct pipe_blend_state *state)
 {
    struct iris_blend_state *cso = malloc(sizeof(struct iris_blend_state));
-   uint32_t *blend_state = cso->blend_state;
+   uint32_t *blend_entry = cso->blend_state + GENX(BLEND_STATE_length);
 
    cso->alpha_to_coverage = state->alpha_to_coverage;
 
+   bool indep_alpha_blend = false;
+
+   for (int i = 0; i < BRW_MAX_DRAW_BUFFERS; i++) {
+      const struct pipe_rt_blend_state *rt =
+         &state->rt[state->independent_blend_enable ? i : 0];
+
+      if (rt->rgb_func != rt->alpha_func ||
+          rt->rgb_src_factor != rt->alpha_src_factor ||
+          rt->rgb_dst_factor != rt->alpha_dst_factor)
+         indep_alpha_blend = true;
+
+      iris_pack_state(GENX(BLEND_STATE_ENTRY), blend_entry, be) {
+         be.LogicOpEnable = state->logicop_enable;
+         be.LogicOpFunction = state->logicop_func;
+
+         be.PreBlendSourceOnlyClampEnable = false;
+         be.ColorClampRange = COLORCLAMP_RTFORMAT;
+         be.PreBlendColorClampEnable = true;
+         be.PostBlendColorClampEnable = true;
+
+         be.ColorBufferBlendEnable = rt->blend_enable;
+
+         be.ColorBlendFunction          = rt->rgb_func;
+         be.AlphaBlendFunction          = rt->alpha_func;
+         be.SourceBlendFactor           = rt->rgb_src_factor;
+         be.SourceAlphaBlendFactor      = rt->alpha_src_factor;
+         be.DestinationBlendFactor      = rt->rgb_dst_factor;
+         be.DestinationAlphaBlendFactor = rt->alpha_dst_factor;
+
+         be.WriteDisableRed   = !(rt->colormask & PIPE_MASK_R);
+         be.WriteDisableGreen = !(rt->colormask & PIPE_MASK_G);
+         be.WriteDisableBlue  = !(rt->colormask & PIPE_MASK_B);
+         be.WriteDisableAlpha = !(rt->colormask & PIPE_MASK_A);
+      }
+      blend_entry += GENX(BLEND_STATE_ENTRY_length);
+   }
+
    iris_pack_command(GENX(3DSTATE_PS_BLEND), cso->ps_blend, pb) {
       /* pb.HasWriteableRT is filled in at draw time. */
       /* pb.AlphaTestEnable is filled in at draw time. */
       pb.AlphaToCoverageEnable = state->alpha_to_coverage;
-      pb.IndependentAlphaBlendEnable = state->independent_blend_enable;
+      pb.IndependentAlphaBlendEnable = indep_alpha_blend;
 
       pb.ColorBufferBlendEnable = state->rt[0].blend_enable;
 
       pb.SourceBlendFactor           = state->rt[0].rgb_src_factor;
-      pb.SourceAlphaBlendFactor      = state->rt[0].alpha_func;
+      pb.SourceAlphaBlendFactor      = state->rt[0].alpha_src_factor;
       pb.DestinationBlendFactor      = state->rt[0].rgb_dst_factor;
       pb.DestinationAlphaBlendFactor = state->rt[0].alpha_dst_factor;
    }
 
-   iris_pack_state(GENX(BLEND_STATE), blend_state, bs) {
+   iris_pack_state(GENX(BLEND_STATE), cso->blend_state, bs) {
       bs.AlphaToCoverageEnable = state->alpha_to_coverage;
-      bs.IndependentAlphaBlendEnable = state->independent_blend_enable;
+      bs.IndependentAlphaBlendEnable = indep_alpha_blend;
       bs.AlphaToOneEnable = state->alpha_to_one;
       bs.AlphaToCoverageDitherEnable = state->alpha_to_coverage;
       bs.ColorDitherEnable = state->dither;
       /* bl.AlphaTestEnable and bs.AlphaTestFunction are filled in later. */
    }
 
-   blend_state += GENX(BLEND_STATE_length);
-
-   for (int i = 0; i < BRW_MAX_DRAW_BUFFERS; i++) {
-      iris_pack_state(GENX(BLEND_STATE_ENTRY), blend_state, be) {
-         be.LogicOpEnable = state->logicop_enable;
-         be.LogicOpFunction = state->logicop_func;
-
-         be.PreBlendSourceOnlyClampEnable = false;
-         be.ColorClampRange = COLORCLAMP_RTFORMAT;
-         be.PreBlendColorClampEnable = true;
-         be.PostBlendColorClampEnable = true;
-
-         be.ColorBufferBlendEnable = state->rt[i].blend_enable;
-
-         be.ColorBlendFunction          = state->rt[i].rgb_func;
-         be.AlphaBlendFunction          = state->rt[i].alpha_func;
-         be.SourceBlendFactor           = state->rt[i].rgb_src_factor;
-         be.SourceAlphaBlendFactor      = state->rt[i].alpha_func;
-         be.DestinationBlendFactor      = state->rt[i].rgb_dst_factor;
-         be.DestinationAlphaBlendFactor = state->rt[i].alpha_dst_factor;
-
-         be.WriteDisableRed   = !(state->rt[i].colormask & PIPE_MASK_R);
-         be.WriteDisableGreen = !(state->rt[i].colormask & PIPE_MASK_G);
-         be.WriteDisableBlue  = !(state->rt[i].colormask & PIPE_MASK_B);
-         be.WriteDisableAlpha = !(state->rt[i].colormask & PIPE_MASK_A);
-      }
-      blend_state += GENX(BLEND_STATE_ENTRY_length);
-   }
 
    return cso;
 }
@@ -921,6 +999,7 @@ struct iris_rasterizer_state {
    uint32_t wm[GENX(3DSTATE_WM_length)];
    uint32_t line_stipple[GENX(3DSTATE_LINE_STIPPLE_length)];
 
+   uint8_t num_clip_plane_consts;
    bool clip_halfz; /* for CC_VIEWPORT */
    bool depth_clip_near; /* for CC_VIEWPORT */
    bool depth_clip_far; /* for CC_VIEWPORT */
@@ -1010,6 +1089,11 @@ iris_create_rasterizer_state(struct pipe_context *ctx,
    cso->line_stipple_enable = state->line_stipple_enable;
    cso->poly_stipple_enable = state->poly_stipple_enable;
 
+   if (state->clip_plane_enable != 0)
+      cso->num_clip_plane_consts = util_logbase2(state->clip_plane_enable) + 1;
+   else
+      cso->num_clip_plane_consts = 0;
+
    float line_width = get_line_width(state);
 
    iris_pack_command(GENX(3DSTATE_SF), cso->sf, sf) {
@@ -1020,7 +1104,7 @@ iris_create_rasterizer_state(struct pipe_context *ctx,
          state->line_smooth ? _10pixels : _05pixels;
       sf.LastPixelEnable = state->line_last_pixel;
       sf.LineWidth = line_width;
-      sf.SmoothPointEnable = state->point_smooth;
+      sf.SmoothPointEnable = state->point_smooth || state->multisample;
       sf.PointWidthSource = state->point_size_per_vertex ? Vertex : State;
       sf.PointWidth = state->point_size;
 
@@ -1045,7 +1129,7 @@ iris_create_rasterizer_state(struct pipe_context *ctx,
       rr.GlobalDepthOffsetConstant = state->offset_units * 2;
       rr.GlobalDepthOffsetScale = state->offset_scale;
       rr.GlobalDepthOffsetClamp = state->offset_clamp;
-      rr.SmoothPointEnable = state->point_smooth;
+      rr.SmoothPointEnable = state->point_smooth || state->multisample;
       rr.AntialiasingEnable = state->line_smooth;
       rr.ScissorRectangleEnable = state->scissor;
       rr.ViewportZNearClipTestEnable = state->depth_clip_near;
@@ -1085,7 +1169,6 @@ iris_create_rasterizer_state(struct pipe_context *ctx,
       wm.LineAntialiasingRegionWidth = _10pixels;
       wm.LineEndCapAntialiasingRegionWidth = _05pixels;
       wm.PointRasterizationRule = RASTRULE_UPPER_RIGHT;
-      wm.StatisticsEnable = true;
       wm.LineStippleEnable = state->line_stipple_enable;
       wm.PolygonStippleEnable = state->poly_stipple_enable;
    }
@@ -1348,6 +1431,44 @@ fmt_swizzle(const struct iris_format_info *fmt, enum pipe_swizzle swz)
    }
 }
 
+static void
+fill_buffer_surface_state(struct isl_device *isl_dev,
+                          struct iris_bo *bo,
+                          void *map,
+                          enum isl_format format,
+                          unsigned offset,
+                          unsigned size)
+{
+   const struct isl_format_layout *fmtl = isl_format_get_layout(format);
+   const unsigned cpp = fmtl->bpb / 8;
+
+   /* The ARB_texture_buffer_specification says:
+    *
+    *    "The number of texels in the buffer texture's texel array is given by
+    *
+    *       floor(<buffer_size> / (<components> * sizeof(<base_type>)),
+    *
+    *     where <buffer_size> is the size of the buffer object, in basic
+    *     machine units and <components> and <base_type> are the element count
+    *     and base data type for elements, as specified in Table X.1.  The
+    *     number of texels in the texel array is then clamped to the
+    *     implementation-dependent limit MAX_TEXTURE_BUFFER_SIZE_ARB."
+    *
+    * We need to clamp the size in bytes to MAX_TEXTURE_BUFFER_SIZE * stride,
+    * so that when ISL divides by stride to obtain the number of texels, that
+    * texel count is clamped to MAX_TEXTURE_BUFFER_SIZE.
+    */
+   unsigned final_size =
+      MIN3(size, bo->size - offset, IRIS_MAX_TEXTURE_BUFFER_SIZE * cpp);
+
+   isl_buffer_fill_state(isl_dev, map,
+                         .address = bo->gtt_offset + offset,
+                         .size_B = final_size,
+                         .format = format,
+                         .stride_B = cpp,
+                         .mocs = MOCS_WB);
+}
+
 /**
  * The pipe->create_sampler_view() driver hook.
  */
@@ -1413,6 +1534,7 @@ iris_create_sampler_view(struct pipe_context *ctx,
    if (tmpl->target != PIPE_BUFFER) {
       isv->view.base_level = tmpl->u.tex.first_level;
       isv->view.levels = tmpl->u.tex.last_level - tmpl->u.tex.first_level + 1;
+      // XXX: do I need to port f9fd0cf4790cb2a530e75d1a2206dbb9d8af7cb2?
       isv->view.base_array_layer = tmpl->u.tex.first_layer;
       isv->view.array_len =
          tmpl->u.tex.last_layer - tmpl->u.tex.first_layer + 1;
@@ -1424,19 +1546,9 @@ iris_create_sampler_view(struct pipe_context *ctx,
                           // .aux_surf =
                           // .clear_color = clear_color,
    } else {
-      // XXX: what to do about isv->view?  other drivers don't use it for bufs
-      const struct isl_format_layout *fmtl =
-         isl_format_get_layout(isv->view.format);
-      const unsigned cpp = fmtl->bpb / 8;
-
-      isl_buffer_fill_state(&screen->isl_dev, map,
-                            .address = isv->res->bo->gtt_offset +
-                                       tmpl->u.buf.offset,
-                            // XXX: buffer_texture_range_size from i965?
-                            .size_B = tmpl->u.buf.size,
-                            .format = isv->view.format,
-                            .stride_B = cpp,
-                            .mocs = MOCS_WB);
+      fill_buffer_surface_state(&screen->isl_dev, isv->res->bo, map,
+                                isv->view.format, tmpl->u.buf.offset,
+                                tmpl->u.buf.size);
    }
 
    return &isv->base;
@@ -1602,18 +1714,9 @@ iris_set_shader_images(struct pipe_context *ctx,
                                 // .aux_surf =
                                 // .clear_color = clear_color,
          } else {
-            // XXX: what to do about view?  other drivers don't use it for bufs
-            const struct isl_format_layout *fmtl =
-               isl_format_get_layout(isl_format);
-            const unsigned cpp = fmtl->bpb / 8;
-
-            isl_buffer_fill_state(&screen->isl_dev, map,
-                                  .address = res->bo->gtt_offset,
-                                  // XXX: buffer_texture_range_size from i965?
-                                  .size_B = res->base.width0,
-                                  .format = isl_format,
-                                  .stride_B = cpp,
-                                  .mocs = MOCS_WB);
+            fill_buffer_surface_state(&screen->isl_dev, res->bo, map,
+                                      isl_format, img->u.buf.offset,
+                                      img->u.buf.size);
          }
       } else {
          pipe_resource_reference(&shs->image[start_slot + i].res, NULL);
@@ -1679,11 +1782,17 @@ iris_surface_destroy(struct pipe_context *ctx, struct pipe_surface *p_surf)
    free(surf);
 }
 
-// XXX: actually implement user clip planes
 static void
 iris_set_clip_state(struct pipe_context *ctx,
                     const struct pipe_clip_state *state)
 {
+   struct iris_context *ice = (struct iris_context *) ctx;
+   struct iris_shader_state *shs = &ice->state.shaders[MESA_SHADER_VERTEX];
+
+   memcpy(&ice->state.clip_planes, state, sizeof(*state));
+
+   ice->state.dirty |= IRIS_DIRTY_CONSTANTS_VS;
+   shs->cbuf0_needs_upload = true;
 }
 
 /**
@@ -1723,12 +1832,27 @@ static void
 iris_set_scissor_states(struct pipe_context *ctx,
                         unsigned start_slot,
                         unsigned num_scissors,
-                        const struct pipe_scissor_state *states)
+                        const struct pipe_scissor_state *rects)
 {
    struct iris_context *ice = (struct iris_context *) ctx;
 
    for (unsigned i = 0; i < num_scissors; i++) {
-      ice->state.scissors[start_slot + i] = states[i];
+      if (rects[i].minx == rects[i].maxx || rects[i].miny == rects[i].maxy) {
+         /* If the scissor was out of bounds and got clamped to 0 width/height
+          * at the bounds, the subtraction of 1 from maximums could produce a
+          * negative number and thus not clip anything.  Instead, just provide
+          * a min > max scissor inside the bounds, which produces the expected
+          * no rendering.
+          */
+         ice->state.scissors[start_slot + i] = (struct pipe_scissor_state) {
+            .minx = 1, .maxx = 0, .miny = 1, .maxy = 0,
+         };
+      } else {
+         ice->state.scissors[start_slot + i] = (struct pipe_scissor_state) {
+            .minx = rects[i].minx,     .miny = rects[i].miny,
+            .maxx = rects[i].maxx - 1, .maxy = rects[i].maxy - 1,
+         };
+      }
    }
 
    ice->state.dirty |= IRIS_DIRTY_SCISSOR_RECT;
@@ -1851,7 +1975,8 @@ iris_set_viewport_states(struct pipe_context *ctx,
 {
    struct iris_context *ice = (struct iris_context *) ctx;
    struct iris_genx_state *genx = ice->state.genx;
-   uint32_t *vp_map = &genx->sf_cl_vp[start_slot];
+   uint32_t *vp_map =
+      &genx->sf_cl_vp[start_slot * GENX(SF_CLIP_VIEWPORT_length)];
 
    for (unsigned i = 0; i < count; i++) {
       const struct pipe_viewport_state *state = &states[i];
@@ -1971,7 +2096,10 @@ iris_set_framebuffer_state(struct pipe_context *ctx,
    void *null_surf_map =
       upload_state(ice->state.surface_uploader, &ice->state.null_fb,
                    4 * GENX(RENDER_SURFACE_STATE_length), 64);
-   isl_null_fill_state(&screen->isl_dev, null_surf_map, isl_extent3d(cso->width, cso->height, cso->layers ? cso->layers : 1));
+   isl_null_fill_state(&screen->isl_dev, null_surf_map,
+                       isl_extent3d(MAX2(cso->width, 1),
+                                    MAX2(cso->height, 1),
+                                    cso->layers ? cso->layers : 1));
    ice->state.null_fb.offset +=
       iris_bo_offset_from_base_address(iris_resource_bo(ice->state.null_fb.res));
 
@@ -2001,6 +2129,36 @@ iris_set_framebuffer_state(struct pipe_context *ctx,
 #endif
 }
 
+static void
+upload_ubo_surf_state(struct iris_context *ice,
+                      struct iris_const_buffer *cbuf,
+                      unsigned buffer_size)
+{
+   struct pipe_context *ctx = &ice->ctx;
+   struct iris_screen *screen = (struct iris_screen *) ctx->screen;
+
+   // XXX: these are not retained forever, use a separate uploader?
+   void *map =
+      upload_state(ice->state.surface_uploader, &cbuf->surface_state,
+                   4 * GENX(RENDER_SURFACE_STATE_length), 64);
+   if (!unlikely(map)) {
+      pipe_resource_reference(&cbuf->data.res, NULL);
+      return;
+   }
+
+   struct iris_resource *res = (void *) cbuf->data.res;
+   struct iris_bo *surf_bo = iris_resource_bo(cbuf->surface_state.res);
+   cbuf->surface_state.offset += iris_bo_offset_from_base_address(surf_bo);
+
+   isl_buffer_fill_state(&screen->isl_dev, map,
+                         .address = res->bo->gtt_offset + cbuf->data.offset,
+                         .size_B = MIN2(buffer_size,
+                                        res->bo->size - cbuf->data.offset),
+                         .format = ISL_FORMAT_R32G32B32A32_FLOAT,
+                         .stride_B = 1,
+                         .mocs = MOCS_WB)
+}
+
 /**
  * The pipe->set_constant_buffer() driver hook.
  *
@@ -2013,44 +2171,31 @@ iris_set_constant_buffer(struct pipe_context *ctx,
                          const struct pipe_constant_buffer *input)
 {
    struct iris_context *ice = (struct iris_context *) ctx;
-   struct iris_screen *screen = (struct iris_screen *)ctx->screen;
    gl_shader_stage stage = stage_from_pipe(p_stage);
    struct iris_shader_state *shs = &ice->state.shaders[stage];
    struct iris_const_buffer *cbuf = &shs->constbuf[index];
 
-   if (input && (input->buffer || input->user_buffer)) {
-      if (input->user_buffer) {
-         u_upload_data(ctx->const_uploader, 0, input->buffer_size, 32,
-                       input->user_buffer, &cbuf->data.offset,
-                       &cbuf->data.res);
-      } else {
-         pipe_resource_reference(&cbuf->data.res, input->buffer);
-      }
-
-      // XXX: these are not retained forever, use a separate uploader?
-      void *map =
-         upload_state(ice->state.surface_uploader, &cbuf->surface_state,
-                      4 * GENX(RENDER_SURFACE_STATE_length), 64);
-      if (!unlikely(map)) {
-         pipe_resource_reference(&cbuf->data.res, NULL);
-         return;
-      }
+   if (input && input->buffer) {
+      assert(index > 0);
 
-      struct iris_resource *res = (void *) cbuf->data.res;
-      struct iris_bo *surf_bo = iris_resource_bo(cbuf->surface_state.res);
-      cbuf->surface_state.offset += iris_bo_offset_from_base_address(surf_bo);
+      pipe_resource_reference(&cbuf->data.res, input->buffer);
+      cbuf->data.offset = input->buffer_offset;
 
-      isl_buffer_fill_state(&screen->isl_dev, map,
-                            .address = res->bo->gtt_offset + cbuf->data.offset,
-                            .size_B = input->buffer_size,
-                            .format = ISL_FORMAT_R32G32B32A32_FLOAT,
-                            .stride_B = 1,
-                            .mocs = MOCS_WB)
+      upload_ubo_surf_state(ice, cbuf, input->buffer_size);
    } else {
       pipe_resource_reference(&cbuf->data.res, NULL);
       pipe_resource_reference(&cbuf->surface_state.res, NULL);
    }
 
+   if (index == 0) {
+      if (input)
+         memcpy(&shs->cbuf0, input, sizeof(shs->cbuf0));
+      else
+         memset(&shs->cbuf0, 0, sizeof(shs->cbuf0));
+
+      shs->cbuf0_needs_upload = true;
+   }
+
    ice->state.dirty |= IRIS_DIRTY_CONSTANTS_VS << stage;
    // XXX: maybe not necessary all the time...?
    // XXX: we need 3DS_BTP to commit these changes, and if we fell back to
@@ -2058,6 +2203,45 @@ iris_set_constant_buffer(struct pipe_context *ctx,
    ice->state.dirty |= IRIS_DIRTY_BINDINGS_VS << stage;
 }
 
+static void
+upload_uniforms(struct iris_context *ice,
+                gl_shader_stage stage)
+{
+   struct iris_shader_state *shs = &ice->state.shaders[stage];
+   struct iris_const_buffer *cbuf = &shs->constbuf[0];
+   struct iris_compiled_shader *shader = ice->shaders.prog[stage];
+
+   unsigned upload_size = shader->num_system_values * sizeof(uint32_t) +
+                          shs->cbuf0.buffer_size;
+
+   if (upload_size == 0)
+      return;
+
+   uint32_t *map =
+      upload_state(ice->ctx.const_uploader, &cbuf->data, upload_size, 64);
+
+   for (int i = 0; i < shader->num_system_values; i++) {
+      uint32_t sysval = shader->system_values[i];
+      uint32_t value = 0;
+
+      if (BRW_PARAM_BUILTIN_IS_CLIP_PLANE(sysval)) {
+         int plane = BRW_PARAM_BUILTIN_CLIP_PLANE_IDX(sysval);
+         int comp  = BRW_PARAM_BUILTIN_CLIP_PLANE_COMP(sysval);
+         value = fui(ice->state.clip_planes.ucp[plane][comp]);
+      } else {
+         assert(!"unhandled system value");
+      }
+
+      *map++ = value;
+   }
+
+   if (shs->cbuf0.user_buffer) {
+      memcpy(map, shs->cbuf0.user_buffer, shs->cbuf0.buffer_size);
+   }
+
+   upload_ubo_surf_state(ice, cbuf, upload_size);
+}
+
 /**
  * The pipe->set_shader_buffers() driver hook.
  *
@@ -2099,7 +2283,9 @@ iris_set_shader_buffers(struct pipe_context *ctx,
          isl_buffer_fill_state(&screen->isl_dev, map,
                                .address =
                                   res->bo->gtt_offset + buffer->buffer_offset,
-                               .size_B = buffer->buffer_size,
+                               .size_B =
+                                  MIN2(buffer->buffer_size,
+                                       res->bo->size - buffer->buffer_offset),
                                .format = ISL_FORMAT_RAW,
                                .stride_B = 1,
                                .mocs = MOCS_WB);
@@ -2171,9 +2357,13 @@ iris_set_vertex_buffers(struct pipe_context *ctx,
          vb.MOCS = MOCS_WB;
          vb.AddressModifyEnable = true;
          vb.BufferPitch = buffers[i].stride;
-         vb.BufferSize = res->bo->size;
-         vb.BufferStartingAddress =
-            ro_bo(NULL, res->bo->gtt_offset + buffers[i].buffer_offset);
+         if (res) {
+            vb.BufferSize = res->bo->size;
+            vb.BufferStartingAddress =
+               ro_bo(NULL, res->bo->gtt_offset + buffers[i].buffer_offset);
+         } else {
+            vb.NullVertexBuffer = true;
+         }
       }
 
       vb_pack_dest += GENX(VERTEX_BUFFER_STATE_length);
@@ -2384,6 +2574,14 @@ iris_set_stream_output_targets(struct pipe_context *ctx,
    if (ice->state.streamout_active != active) {
       ice->state.streamout_active = active;
       ice->state.dirty |= IRIS_DIRTY_STREAMOUT;
+
+      /* We only emit 3DSTATE_SO_DECL_LIST when streamout is active, because
+       * it's a non-pipelined command.  If we're switching streamout on, we
+       * may have missed emitting it earlier, so do so now.  (We're already
+       * taking a stall to update 3DSTATE_SO_BUFFERS anyway...)
+       */
+      if (active)
+         ice->state.dirty |= IRIS_DIRTY_SO_DECL_LIST;
    }
 
    for (int i = 0; i < 4; i++) {
@@ -2800,9 +2998,16 @@ iris_populate_sampler_key(const struct iris_context *ice,
  */
 static void
 iris_populate_vs_key(const struct iris_context *ice,
+                     const struct shader_info *info,
                      struct brw_vs_prog_key *key)
 {
+   const struct iris_rasterizer_state *cso_rast = ice->state.cso_rast;
+
    iris_populate_sampler_key(ice, &key->tex);
+
+   if (info->clip_distance_array_size == 0 &&
+       (info->outputs_written & (VARYING_BIT_POS | VARYING_BIT_CLIP_VERTEX)))
+      key->nr_userclip_plane_consts = cso_rast->num_clip_plane_consts;
 }
 
 /**
@@ -2882,8 +3087,6 @@ iris_populate_cs_key(const struct iris_context *ice,
    // XXX: these need to go in INIT_THREAD_DISPATCH_FIELDS
    pkt.SamplerCount =                                                     \
       DIV_ROUND_UP(CLAMP(stage_state->sampler_count, 0, 16), 4);          \
-   pkt.PerThreadScratchSpace = prog_data->total_scratch == 0 ? 0 :        \
-      ffs(stage_state->per_thread_scratch) - 11;                          \
 
 #endif
 
@@ -2898,7 +3101,7 @@ KSP(const struct iris_compiled_shader *shader)
 // prefetching of binding tables in A0 and B0 steppings.  XXX: Revisit
 // this WA on C0 stepping.
 
-#define INIT_THREAD_DISPATCH_FIELDS(pkt, prefix)                          \
+#define INIT_THREAD_DISPATCH_FIELDS(pkt, prefix, stage)                   \
    pkt.KernelStartPointer = KSP(shader);                                  \
    pkt.BindingTableEntryCount = GEN_GEN == 11 ? 0 :                       \
       prog_data->binding_table.size_bytes / 4;                            \
@@ -2910,20 +3113,28 @@ KSP(const struct iris_compiled_shader *shader)
    pkt.prefix##URBEntryReadOffset = 0;                                    \
                                                                           \
    pkt.StatisticsEnable = true;                                           \
-   pkt.Enable           = true;
+   pkt.Enable           = true;                                           \
+                                                                          \
+   if (prog_data->total_scratch) {                                        \
+      uint32_t scratch_addr =                                             \
+         iris_get_scratch_space(ice, prog_data->total_scratch, stage);    \
+      pkt.PerThreadScratchSpace = ffs(prog_data->total_scratch) - 11;     \
+      pkt.ScratchSpaceBasePointer = rw_bo(NULL, scratch_addr);            \
+   }
 
 /**
  * Encode most of 3DSTATE_VS based on the compiled shader.
  */
 static void
-iris_store_vs_state(const struct gen_device_info *devinfo,
+iris_store_vs_state(struct iris_context *ice,
+                    const struct gen_device_info *devinfo,
                     struct iris_compiled_shader *shader)
 {
    struct brw_stage_prog_data *prog_data = shader->prog_data;
    struct brw_vue_prog_data *vue_prog_data = (void *) prog_data;
 
    iris_pack_command(GENX(3DSTATE_VS), shader->derived_data, vs) {
-      INIT_THREAD_DISPATCH_FIELDS(vs, Vertex);
+      INIT_THREAD_DISPATCH_FIELDS(vs, Vertex, MESA_SHADER_VERTEX);
       vs.MaximumNumberofThreads = devinfo->max_vs_threads - 1;
       vs.SIMD8DispatchEnable = true;
       vs.UserClipDistanceCullTestEnableBitmask =
@@ -2935,7 +3146,8 @@ iris_store_vs_state(const struct gen_device_info *devinfo,
  * Encode most of 3DSTATE_HS based on the compiled shader.
  */
 static void
-iris_store_tcs_state(const struct gen_device_info *devinfo,
+iris_store_tcs_state(struct iris_context *ice,
+                     const struct gen_device_info *devinfo,
                      struct iris_compiled_shader *shader)
 {
    struct brw_stage_prog_data *prog_data = shader->prog_data;
@@ -2943,7 +3155,7 @@ iris_store_tcs_state(const struct gen_device_info *devinfo,
    struct brw_tcs_prog_data *tcs_prog_data = (void *) prog_data;
 
    iris_pack_command(GENX(3DSTATE_HS), shader->derived_data, hs) {
-      INIT_THREAD_DISPATCH_FIELDS(hs, Vertex);
+      INIT_THREAD_DISPATCH_FIELDS(hs, Vertex, MESA_SHADER_TESS_CTRL);
 
       hs.InstanceCount = tcs_prog_data->instances - 1;
       hs.MaximumNumberofThreads = devinfo->max_tcs_threads - 1;
@@ -2955,7 +3167,8 @@ iris_store_tcs_state(const struct gen_device_info *devinfo,
  * Encode 3DSTATE_TE and most of 3DSTATE_DS based on the compiled shader.
  */
 static void
-iris_store_tes_state(const struct gen_device_info *devinfo,
+iris_store_tes_state(struct iris_context *ice,
+                     const struct gen_device_info *devinfo,
                      struct iris_compiled_shader *shader)
 {
    struct brw_stage_prog_data *prog_data = shader->prog_data;
@@ -2975,7 +3188,7 @@ iris_store_tes_state(const struct gen_device_info *devinfo,
    }
 
    iris_pack_command(GENX(3DSTATE_DS), ds_state, ds) {
-      INIT_THREAD_DISPATCH_FIELDS(ds, Patch);
+      INIT_THREAD_DISPATCH_FIELDS(ds, Patch, MESA_SHADER_TESS_EVAL);
 
       ds.DispatchMode = DISPATCH_MODE_SIMD8_SINGLE_PATCH;
       ds.MaximumNumberofThreads = devinfo->max_tes_threads - 1;
@@ -2992,7 +3205,8 @@ iris_store_tes_state(const struct gen_device_info *devinfo,
  * Encode most of 3DSTATE_GS based on the compiled shader.
  */
 static void
-iris_store_gs_state(const struct gen_device_info *devinfo,
+iris_store_gs_state(struct iris_context *ice,
+                    const struct gen_device_info *devinfo,
                     struct iris_compiled_shader *shader)
 {
    struct brw_stage_prog_data *prog_data = shader->prog_data;
@@ -3000,7 +3214,7 @@ iris_store_gs_state(const struct gen_device_info *devinfo,
    struct brw_gs_prog_data *gs_prog_data = (void *) prog_data;
 
    iris_pack_command(GENX(3DSTATE_GS), shader->derived_data, gs) {
-      INIT_THREAD_DISPATCH_FIELDS(gs, Vertex);
+      INIT_THREAD_DISPATCH_FIELDS(gs, Vertex, MESA_SHADER_GEOMETRY);
 
       gs.OutputVertexSize = gs_prog_data->output_vertex_size_hwords * 2 - 1;
       gs.OutputTopology = gs_prog_data->output_topology;
@@ -3039,7 +3253,8 @@ iris_store_gs_state(const struct gen_device_info *devinfo,
  * Encode most of 3DSTATE_PS and 3DSTATE_PS_EXTRA based on the shader.
  */
 static void
-iris_store_fs_state(const struct gen_device_info *devinfo,
+iris_store_fs_state(struct iris_context *ice,
+                    const struct gen_device_info *devinfo,
                     struct iris_compiled_shader *shader)
 {
    struct brw_stage_prog_data *prog_data = shader->prog_data;
@@ -3057,7 +3272,7 @@ iris_store_fs_state(const struct gen_device_info *devinfo,
       ps.FloatingPointMode = prog_data->use_alt_mode;
       ps.MaximumNumberofThreadsPerPSD = 64 - (GEN_GEN == 8 ? 2 : 1);
 
-      ps.PushConstantEnable = prog_data->nr_params > 0 ||
+      ps.PushConstantEnable = shader->num_system_values > 0 ||
                               prog_data->ubo_ranges[0].length > 0;
 
       /* From the documentation for this packet:
@@ -3094,6 +3309,14 @@ iris_store_fs_state(const struct gen_device_info *devinfo,
          KSP(shader) + brw_wm_prog_data_prog_offset(wm_prog_data, ps, 1);
       ps.KernelStartPointer2 =
          KSP(shader) + brw_wm_prog_data_prog_offset(wm_prog_data, ps, 2);
+
+      if (prog_data->total_scratch) {
+         uint32_t scratch_addr =
+            iris_get_scratch_space(ice, prog_data->total_scratch,
+                                   MESA_SHADER_FRAGMENT);
+         ps.PerThreadScratchSpace = ffs(prog_data->total_scratch) - 11;
+         ps.ScratchSpaceBasePointer = rw_bo(NULL, scratch_addr);
+      }
    }
 
    iris_pack_command(GENX(3DSTATE_PS_EXTRA), psx_state, psx) {
@@ -3127,7 +3350,8 @@ iris_store_fs_state(const struct gen_device_info *devinfo,
  * This must match the data written by the iris_store_xs_state() functions.
  */
 static void
-iris_store_cs_state(const struct gen_device_info *devinfo,
+iris_store_cs_state(struct iris_context *ice,
+                    const struct gen_device_info *devinfo,
                     struct iris_compiled_shader *shader)
 {
    struct brw_stage_prog_data *prog_data = shader->prog_data;
@@ -3172,28 +3396,31 @@ iris_derived_program_state_size(enum iris_program_cache_id cache_id)
  * get most of the state packet without having to reconstruct it.
  */
 static void
-iris_store_derived_program_state(const struct gen_device_info *devinfo,
+iris_store_derived_program_state(struct iris_context *ice,
                                  enum iris_program_cache_id cache_id,
                                  struct iris_compiled_shader *shader)
 {
+   struct iris_screen *screen = (void *) ice->ctx.screen;
+   const struct gen_device_info *devinfo = &screen->devinfo;
+
    switch (cache_id) {
    case IRIS_CACHE_VS:
-      iris_store_vs_state(devinfo, shader);
+      iris_store_vs_state(ice, devinfo, shader);
       break;
    case IRIS_CACHE_TCS:
-      iris_store_tcs_state(devinfo, shader);
+      iris_store_tcs_state(ice, devinfo, shader);
       break;
    case IRIS_CACHE_TES:
-      iris_store_tes_state(devinfo, shader);
+      iris_store_tes_state(ice, devinfo, shader);
       break;
    case IRIS_CACHE_GS:
-      iris_store_gs_state(devinfo, shader);
+      iris_store_gs_state(ice, devinfo, shader);
       break;
    case IRIS_CACHE_FS:
-      iris_store_fs_state(devinfo, shader);
+      iris_store_fs_state(ice, devinfo, shader);
       break;
    case IRIS_CACHE_CS:
-      iris_store_cs_state(devinfo, shader);
+      iris_store_cs_state(ice, devinfo, shader);
    case IRIS_CACHE_BLORP:
       break;
    default:
@@ -3266,6 +3493,10 @@ use_null_surface(struct iris_batch *batch, struct iris_context *ice)
 static uint32_t
 use_null_fb_surface(struct iris_batch *batch, struct iris_context *ice)
 {
+   /* If set_framebuffer_state() was never called, fall back to 1x1x1 */
+   if (!ice->state.null_fb.res)
+      return use_null_surface(batch, ice);
+
    struct iris_bo *state_bo = iris_resource_bo(ice->state.null_fb.res);
 
    iris_use_pinned_bo(batch, state_bo, false);
@@ -3361,8 +3592,7 @@ static void
 iris_populate_binding_table(struct iris_context *ice,
                             struct iris_batch *batch,
                             gl_shader_stage stage,
-                            bool pin_only,
-                            struct iris_state_ref *grid_size_surf)
+                            bool pin_only)
 {
    const struct iris_binder *binder = &ice->state.binder;
    struct iris_compiled_shader *shader = ice->shaders.prog[stage];
@@ -3385,12 +3615,11 @@ iris_populate_binding_table(struct iris_context *ice,
 
    if (stage == MESA_SHADER_COMPUTE) {
       /* surface for gl_NumWorkGroups */
-      assert(grid_size_surf || pin_only);
-      if (grid_size_surf) {
-         struct iris_bo *bo = iris_resource_bo(grid_size_surf->res);
-         iris_use_pinned_bo(batch, bo, false);
-         push_bt_entry(grid_size_surf->offset);
-      }
+      struct iris_state_ref *grid_data = &ice->state.grid_size;
+      struct iris_state_ref *grid_state = &ice->state.grid_surf_state;
+      iris_use_pinned_bo(batch, iris_resource_bo(grid_data->res),  false);
+      iris_use_pinned_bo(batch, iris_resource_bo(grid_state->res), false);
+      push_bt_entry(grid_state->offset);
    }
 
    if (stage == MESA_SHADER_FRAGMENT) {
@@ -3536,7 +3765,7 @@ iris_restore_render_saved_bos(struct iris_context *ice,
    for (int stage = 0; stage <= MESA_SHADER_FRAGMENT; stage++) {
       if (clean & (IRIS_DIRTY_BINDINGS_VS << stage)) {
          /* Re-pin any buffers referred to by the binding table. */
-         iris_populate_binding_table(ice, batch, stage, true, NULL);
+         iris_populate_binding_table(ice, batch, stage, true);
       }
    }
 
@@ -3622,7 +3851,7 @@ iris_restore_compute_saved_bos(struct iris_context *ice,
 
    if (clean & IRIS_DIRTY_BINDINGS_CS) {
       /* Re-pin any buffers referred to by the binding table. */
-      iris_populate_binding_table(ice, batch, stage, true, NULL);
+      iris_populate_binding_table(ice, batch, stage, true);
    }
 
    struct pipe_resource *sampler_res = shs->sampler_table.res;
@@ -3668,7 +3897,7 @@ iris_upload_dirty_render_state(struct iris_context *ice,
 {
    const uint64_t dirty = ice->state.dirty;
 
-   if (!dirty)
+   if (!(dirty & IRIS_ALL_DIRTY_FOR_RENDER))
       return;
 
    struct iris_genx_state *genx = ice->state.genx;
@@ -3821,6 +4050,9 @@ iris_upload_dirty_render_state(struct iris_context *ice,
       if (!shader)
          continue;
 
+      if (shs->cbuf0_needs_upload)
+         upload_uniforms(ice, stage);
+
       struct brw_stage_prog_data *prog_data = (void *) shader->prog_data;
 
       iris_emit_cmd(batch, GENX(3DSTATE_CONSTANT_VS), pkt) {
@@ -3871,7 +4103,7 @@ iris_upload_dirty_render_state(struct iris_context *ice,
 
    for (int stage = 0; stage <= MESA_SHADER_FRAGMENT; stage++) {
       if (dirty & (IRIS_DIRTY_BINDINGS_VS << stage)) {
-         iris_populate_binding_table(ice, batch, stage, false, NULL);
+         iris_populate_binding_table(ice, batch, stage, false);
       }
    }
 
@@ -4007,6 +4239,8 @@ iris_upload_dirty_render_state(struct iris_context *ice,
       uint32_t dynamic_wm[GENX(3DSTATE_WM_length)];
 
       iris_pack_command(GENX(3DSTATE_WM), &dynamic_wm, wm) {
+         wm.StatisticsEnable = ice->state.statistics_counters_enabled;
+
          wm.BarycentricInterpolationMode =
             wm_prog_data->barycentric_interp_modes;
 
@@ -4067,9 +4301,14 @@ iris_upload_dirty_render_state(struct iris_context *ice,
       iris_batch_emit(batch, cso_z->packets, sizeof(cso_z->packets));
 
       if (cso_fb->zsbuf) {
-         struct iris_resource *zres = (void *) cso_fb->zsbuf->texture;
-         // XXX: depth might not be writable...
-         iris_use_pinned_bo(batch, zres->bo, true);
+         struct iris_resource *zres, *sres;
+         iris_get_depth_stencil_resources(cso_fb->zsbuf->texture,
+                                          &zres, &sres);
+         // XXX: might not be writable...
+         if (zres)
+            iris_use_pinned_bo(batch, zres->bo, true);
+         if (sres)
+            iris_use_pinned_bo(batch, sres->bo, true);
       }
    }
 
@@ -4103,7 +4342,8 @@ iris_upload_dirty_render_state(struct iris_context *ice,
 
          for (unsigned i = 0; i < cso->num_buffers; i++) {
             struct iris_resource *res = (void *) cso->resources[i];
-            iris_use_pinned_bo(batch, res->bo, false);
+            if (res)
+               iris_use_pinned_bo(batch, res->bo, false);
          }
       }
    }
@@ -4274,44 +4514,11 @@ iris_upload_compute_state(struct iris_context *ice,
    struct brw_stage_prog_data *prog_data = shader->prog_data;
    struct brw_cs_prog_data *cs_prog_data = (void *) prog_data;
 
-   struct pipe_resource *grid_size_res = NULL;
-   uint32_t grid_size_offset;
-   if (grid->indirect) {
-      grid_size_res = grid->indirect;
-      grid_size_offset = grid->indirect_offset;
-   } else {
-      uint32_t *grid_size_map =
-         stream_state(batch, ice->state.surface_uploader, &grid_size_res, 12, 4,
-                      &grid_size_offset);
-      grid_size_map[0] = grid->grid[0];
-      grid_size_map[1] = grid->grid[1];
-      grid_size_map[2] = grid->grid[2];
-      struct iris_bo *grid_size_bo = iris_resource_bo(grid_size_res);
-      grid_size_offset -= iris_bo_offset_from_base_address(grid_size_bo);
-   }
-
-   struct iris_state_ref grid_size_surf;
-   memset(&grid_size_surf, 0, sizeof(grid_size_surf));
-   void *grid_surf_state_map =
-      upload_state(ice->state.surface_uploader,
-                   &grid_size_surf,
-                   4 * GENX(RENDER_SURFACE_STATE_length), 64);
-   assert(grid_surf_state_map);
-   struct iris_bo *grid_size_bo = iris_resource_bo(grid_size_res);
-   iris_use_pinned_bo(batch, grid_size_bo, false);
-   grid_size_surf.offset +=
-      iris_bo_offset_from_base_address(iris_resource_bo(grid_size_surf.res));
-   isl_buffer_fill_state(&screen->isl_dev, grid_surf_state_map,
-                         .address =
-                            grid_size_bo->gtt_offset + grid_size_offset,
-                         .size_B = 12,
-                         .format = ISL_FORMAT_RAW,
-                         .stride_B = 1,
-                         .mocs = MOCS_WB);
+   if ((dirty & IRIS_DIRTY_CONSTANTS_CS) && shs->cbuf0_needs_upload)
+      upload_uniforms(ice, MESA_SHADER_COMPUTE);
 
-   if (dirty & IRIS_DIRTY_BINDINGS_CS || grid_size_res)
-      iris_populate_binding_table(ice, batch, MESA_SHADER_COMPUTE, false,
-                                  &grid_size_surf);
+   if (dirty & IRIS_DIRTY_BINDINGS_CS)
+      iris_populate_binding_table(ice, batch, MESA_SHADER_COMPUTE, false);
 
    iris_use_optional_res(batch, shs->sampler_table.res, false);
    iris_use_pinned_bo(batch, iris_resource_bo(shader->assembly.res), false);
@@ -4319,43 +4526,45 @@ iris_upload_compute_state(struct iris_context *ice,
    if (ice->state.need_border_colors)
       iris_use_pinned_bo(batch, ice->state.border_color_pool.bo, false);
 
-   /* The MEDIA_VFE_STATE documentation for Gen8+ says:
-    *
-    *   "A stalling PIPE_CONTROL is required before MEDIA_VFE_STATE unless
-    *    the only bits that are changed are scoreboard related: Scoreboard
-    *    Enable, Scoreboard Type, Scoreboard Mask, Scoreboard * Delta. For
-    *    these scoreboard related states, a MEDIA_STATE_FLUSH is sufficient."
-    */
-   iris_emit_pipe_control_flush(batch, PIPE_CONTROL_CS_STALL);
-
-   iris_emit_cmd(batch, GENX(MEDIA_VFE_STATE), vfe) {
-      if (prog_data->total_scratch) {
-         /* Per Thread Scratch Space is in the range [0, 11] where
-          * 0 = 1k, 1 = 2k, 2 = 4k, ..., 11 = 2M.
-          */
-         // XXX: vfe.ScratchSpaceBasePointer
-         //vfe.PerThreadScratchSpace =
-            //ffs(stage_state->per_thread_scratch) - 11;
-      }
+   if (dirty & IRIS_DIRTY_CS) {
+      /* The MEDIA_VFE_STATE documentation for Gen8+ says:
+       *
+       *   "A stalling PIPE_CONTROL is required before MEDIA_VFE_STATE unless
+       *    the only bits that are changed are scoreboard related: Scoreboard
+       *    Enable, Scoreboard Type, Scoreboard Mask, Scoreboard Delta.  For
+       *    these scoreboard related states, a MEDIA_STATE_FLUSH is
+       *    sufficient."
+       */
+      iris_emit_pipe_control_flush(batch, PIPE_CONTROL_CS_STALL);
+
+      iris_emit_cmd(batch, GENX(MEDIA_VFE_STATE), vfe) {
+         if (prog_data->total_scratch) {
+            uint32_t scratch_addr =
+               iris_get_scratch_space(ice, prog_data->total_scratch,
+                                      MESA_SHADER_COMPUTE);
+            vfe.PerThreadScratchSpace = ffs(prog_data->total_scratch) - 11;
+            vfe.ScratchSpaceBasePointer = rw_bo(NULL, scratch_addr);
+         }
 
-      vfe.MaximumNumberofThreads =
-         devinfo->max_cs_threads * screen->subslice_total - 1;
+         vfe.MaximumNumberofThreads =
+            devinfo->max_cs_threads * screen->subslice_total - 1;
 #if GEN_GEN < 11
-      vfe.ResetGatewayTimer =
-         Resettingrelativetimerandlatchingtheglobaltimestamp;
+         vfe.ResetGatewayTimer =
+            Resettingrelativetimerandlatchingtheglobaltimestamp;
 #endif
 
-      vfe.NumberofURBEntries = 2;
-      vfe.URBEntryAllocationSize = 2;
+         vfe.NumberofURBEntries = 2;
+         vfe.URBEntryAllocationSize = 2;
 
-      // XXX: Use Indirect Payload Storage?
-      vfe.CURBEAllocationSize =
-         ALIGN(cs_prog_data->push.per_thread.regs * cs_prog_data->threads +
-               cs_prog_data->push.cross_thread.regs, 2);
+         // XXX: Use Indirect Payload Storage?
+         vfe.CURBEAllocationSize =
+            ALIGN(cs_prog_data->push.per_thread.regs * cs_prog_data->threads +
+                  cs_prog_data->push.cross_thread.regs, 2);
+      }
    }
 
-   // XXX: hack iris_set_constant_buffers to upload compute shader constants
-   // XXX: differently...?
+   // XXX: hack iris_set_constant_buffers to upload these thread counts
+   // XXX: along with regular uniforms for compute shaders, somehow.
 
    uint32_t curbe_data_offset = 0;
    // TODO: Move subgroup-id into uniforms ubo so we can push uniforms
@@ -4370,37 +4579,40 @@ iris_upload_compute_state(struct iris_context *ice,
    assert(curbe_data_map);
    memset(curbe_data_map, 0x5a, ALIGN(cs_prog_data->push.total.size, 64));
    iris_fill_cs_push_const_buffer(cs_prog_data, curbe_data_map);
-   iris_emit_cmd(batch, GENX(MEDIA_CURBE_LOAD), curbe) {
-     curbe.CURBETotalDataLength =
-        ALIGN(cs_prog_data->push.total.size, 64);
-     curbe.CURBEDataStartAddress = curbe_data_offset;
+
+   if (dirty & IRIS_DIRTY_CONSTANTS_CS) {
+      iris_emit_cmd(batch, GENX(MEDIA_CURBE_LOAD), curbe) {
+         curbe.CURBETotalDataLength =
+            ALIGN(cs_prog_data->push.total.size, 64);
+         curbe.CURBEDataStartAddress = curbe_data_offset;
+      }
    }
 
-   struct pipe_resource *desc_res = NULL;
-   uint32_t desc[GENX(INTERFACE_DESCRIPTOR_DATA_length)];
+   if (dirty & (IRIS_DIRTY_SAMPLER_STATES_CS |
+                IRIS_DIRTY_BINDINGS_CS |
+                IRIS_DIRTY_CONSTANTS_CS |
+                IRIS_DIRTY_CS)) {
+      struct pipe_resource *desc_res = NULL;
+      uint32_t desc[GENX(INTERFACE_DESCRIPTOR_DATA_length)];
 
-   iris_pack_state(GENX(INTERFACE_DESCRIPTOR_DATA), desc, idd) {
-      idd.SamplerStatePointer = shs->sampler_table.offset;
-      idd.BindingTablePointer = binder->bt_offset[MESA_SHADER_COMPUTE];
-      idd.ConstantURBEntryReadLength = cs_prog_data->push.per_thread.regs;
-#if GEN_GEN >= 8 || GEN_IS_HASWELL
-      idd.CrossThreadConstantDataReadLength =
-         cs_prog_data->push.cross_thread.regs;
-#endif
-   }
+      iris_pack_state(GENX(INTERFACE_DESCRIPTOR_DATA), desc, idd) {
+         idd.SamplerStatePointer = shs->sampler_table.offset;
+         idd.BindingTablePointer = binder->bt_offset[MESA_SHADER_COMPUTE];
+      }
 
-   for (int i = 0; i < GENX(INTERFACE_DESCRIPTOR_DATA_length); i++)
-      desc[i] |= ((uint32_t *) shader->derived_data)[i];
+      for (int i = 0; i < GENX(INTERFACE_DESCRIPTOR_DATA_length); i++)
+         desc[i] |= ((uint32_t *) shader->derived_data)[i];
 
-   iris_emit_cmd(batch, GENX(MEDIA_INTERFACE_DESCRIPTOR_LOAD), load) {
-      load.InterfaceDescriptorTotalLength =
-         GENX(INTERFACE_DESCRIPTOR_DATA_length) * sizeof(uint32_t);
-      load.InterfaceDescriptorDataStartAddress =
-         emit_state(batch, ice->state.dynamic_uploader,
-                    &desc_res, desc, sizeof(desc), 32);
-   }
+      iris_emit_cmd(batch, GENX(MEDIA_INTERFACE_DESCRIPTOR_LOAD), load) {
+         load.InterfaceDescriptorTotalLength =
+            GENX(INTERFACE_DESCRIPTOR_DATA_length) * sizeof(uint32_t);
+         load.InterfaceDescriptorDataStartAddress =
+            emit_state(batch, ice->state.dynamic_uploader,
+                       &desc_res, desc, sizeof(desc), 32);
+      }
 
-   pipe_resource_reference(&desc_res, NULL);
+      pipe_resource_reference(&desc_res, NULL);
+   }
 
    uint32_t group_size = grid->block[0] * grid->block[1] * grid->block[2];
    uint32_t remainder = group_size & (cs_prog_data->simd_size - 1);
@@ -4416,18 +4628,19 @@ iris_upload_compute_state(struct iris_context *ice,
 #define GPGPU_DISPATCHDIMZ 0x2508
 
    if (grid->indirect) {
-      struct iris_bo *bo = iris_resource_bo(grid_size_res);
+      struct iris_state_ref *grid_size = &ice->state.grid_size;
+      struct iris_bo *bo = iris_resource_bo(grid_size->res);
       iris_emit_cmd(batch, GENX(MI_LOAD_REGISTER_MEM), lrm) {
          lrm.RegisterAddress = GPGPU_DISPATCHDIMX;
-         lrm.MemoryAddress = ro_bo(bo, grid_size_offset + 0);
+         lrm.MemoryAddress = ro_bo(bo, grid_size->offset + 0);
       }
       iris_emit_cmd(batch, GENX(MI_LOAD_REGISTER_MEM), lrm) {
          lrm.RegisterAddress = GPGPU_DISPATCHDIMY;
-         lrm.MemoryAddress = ro_bo(bo, grid_size_offset + 4);
+         lrm.MemoryAddress = ro_bo(bo, grid_size->offset + 4);
       }
       iris_emit_cmd(batch, GENX(MI_LOAD_REGISTER_MEM), lrm) {
          lrm.RegisterAddress = GPGPU_DISPATCHDIMZ;
-         lrm.MemoryAddress = ro_bo(bo, grid_size_offset + 8);
+         lrm.MemoryAddress = ro_bo(bo, grid_size->offset + 8);
       }
    }
 
@@ -4697,7 +4910,7 @@ iris_emit_raw_pipe_control(struct iris_batch *batch, uint32_t flags,
     * We do these now because they may add post-sync operations or CS stalls.
     */
 
-   if (flags & PIPE_CONTROL_VF_CACHE_INVALIDATE) {
+   if (GEN_GEN < 11 && flags & PIPE_CONTROL_VF_CACHE_INVALIDATE) {
       /* Project: BDW, SKL+ (stopping at CNL) / Argument: VF Invalidate
        *
        * "'Post Sync Operation' must be enabled to 'Write Immediate Data' or
@@ -5081,6 +5294,8 @@ genX(init_state)(struct iris_context *ice)
 
    ice->state.dirty = ~0ull;
 
+   ice->state.statistics_counters_enabled = true;
+
    ice->state.sample_mask = 0xffff;
    ice->state.num_viewports = 1;
    ice->state.genx = calloc(1, sizeof(struct iris_genx_state));
@@ -5092,4 +5307,11 @@ genX(init_state)(struct iris_context *ice)
    isl_null_fill_state(&screen->isl_dev, null_surf_map, isl_extent3d(1, 1, 1));
    ice->state.unbound_tex.offset +=
       iris_bo_offset_from_base_address(iris_resource_bo(ice->state.unbound_tex.res));
+
+   /* Default all scissor rectangles to be empty regions. */
+   for (int i = 0; i < IRIS_MAX_VIEWPORTS; i++) {
+      ice->state.scissors[i] = (struct pipe_scissor_state) {
+         .minx = 1, .maxx = 0, .miny = 1, .maxy = 0,
+      };
+   }
 }