radv: add radv_init_dcc_control_reg() helper
[mesa.git] / src / amd / vulkan / radv_image.c
index 46b62052a0ce40800a9a10ba59eb80eb16e64d49..acb569203d4cf3a131e0a700dc2f409e88497b72 100644 (file)
@@ -63,6 +63,111 @@ radv_choose_tiling(struct radv_device *device,
 
        return RADEON_SURF_MODE_2D;
 }
+
+static bool
+radv_use_tc_compat_htile_for_image(struct radv_device *device,
+                                  const VkImageCreateInfo *pCreateInfo)
+{
+       /* TC-compat HTILE is only available for GFX8+. */
+       if (device->physical_device->rad_info.chip_class < VI)
+               return false;
+
+       if (pCreateInfo->usage & VK_IMAGE_USAGE_STORAGE_BIT)
+               return false;
+
+       if (pCreateInfo->flags & (VK_IMAGE_CREATE_MUTABLE_FORMAT_BIT |
+                                 VK_IMAGE_CREATE_EXTENDED_USAGE_BIT_KHR))
+               return false;
+
+       if (pCreateInfo->tiling == VK_IMAGE_TILING_LINEAR)
+               return false;
+
+       if (pCreateInfo->mipLevels > 1)
+               return false;
+
+       /* FIXME: for some reason TC compat with 2/4/8 samples breaks some cts
+        * tests - disable for now */
+       if (pCreateInfo->samples >= 2 &&
+           pCreateInfo->format == VK_FORMAT_D32_SFLOAT_S8_UINT)
+               return false;
+
+       /* GFX9 supports both 32-bit and 16-bit depth surfaces, while GFX8 only
+        * supports 32-bit. Though, it's possible to enable TC-compat for
+        * 16-bit depth surfaces if no Z planes are compressed.
+        */
+       if (pCreateInfo->format != VK_FORMAT_D32_SFLOAT_S8_UINT &&
+           pCreateInfo->format != VK_FORMAT_D32_SFLOAT &&
+           pCreateInfo->format != VK_FORMAT_D16_UNORM)
+               return false;
+
+       return true;
+}
+
+static bool
+radv_use_dcc_for_image(struct radv_device *device,
+                      const struct radv_image_create_info *create_info,
+                      const VkImageCreateInfo *pCreateInfo)
+{
+       bool dcc_compatible_formats;
+       bool blendable;
+
+       /* DCC (Delta Color Compression) is only available for GFX8+. */
+       if (device->physical_device->rad_info.chip_class < VI)
+               return false;
+
+       if (device->instance->debug_flags & RADV_DEBUG_NO_DCC)
+               return false;
+
+       /* TODO: Enable DCC for storage images. */
+       if ((pCreateInfo->usage & VK_IMAGE_USAGE_STORAGE_BIT) ||
+           (pCreateInfo->flags & VK_IMAGE_CREATE_EXTENDED_USAGE_BIT_KHR))
+               return false;
+
+       if (pCreateInfo->tiling == VK_IMAGE_TILING_LINEAR)
+               return false;
+
+       /* TODO: Enable DCC for mipmaps and array layers. */
+       if (pCreateInfo->mipLevels > 1 || pCreateInfo->arrayLayers > 1)
+               return false;
+
+       if (create_info->scanout)
+               return false;
+
+       /* TODO: Enable DCC for MSAA textures. */
+       if (pCreateInfo->samples >= 2)
+               return false;
+
+       /* Determine if the formats are DCC compatible. */
+       dcc_compatible_formats =
+               radv_is_colorbuffer_format_supported(pCreateInfo->format,
+                                                    &blendable);
+
+       if (pCreateInfo->flags & VK_IMAGE_CREATE_MUTABLE_FORMAT_BIT) {
+               const struct VkImageFormatListCreateInfoKHR *format_list =
+                       (const struct  VkImageFormatListCreateInfoKHR *)
+                               vk_find_struct_const(pCreateInfo->pNext,
+                                                    IMAGE_FORMAT_LIST_CREATE_INFO_KHR);
+
+               /* We have to ignore the existence of the list if viewFormatCount = 0 */
+               if (format_list && format_list->viewFormatCount) {
+                       /* compatibility is transitive, so we only need to check
+                        * one format with everything else. */
+                       for (unsigned i = 0; i < format_list->viewFormatCount; ++i) {
+                               if (!radv_dcc_formats_compatible(pCreateInfo->format,
+                                                                format_list->pViewFormats[i]))
+                                       dcc_compatible_formats = false;
+                       }
+               } else {
+                       dcc_compatible_formats = false;
+               }
+       }
+
+       if (!dcc_compatible_formats)
+               return false;
+
+       return true;
+}
+
 static int
 radv_init_surface(struct radv_device *device,
                  struct radeon_surf *surface,
@@ -72,7 +177,7 @@ radv_init_surface(struct radv_device *device,
        unsigned array_mode = radv_choose_tiling(device, create_info);
        const struct vk_format_description *desc =
                vk_format_description(pCreateInfo->format);
-       bool is_depth, is_stencil, blendable;
+       bool is_depth, is_stencil;
 
        is_depth = vk_format_has_depth(desc);
        is_stencil = vk_format_has_stencil(desc);
@@ -109,6 +214,8 @@ radv_init_surface(struct radv_device *device,
 
        if (is_depth) {
                surface->flags |= RADEON_SURF_ZBUFFER;
+               if (radv_use_tc_compat_htile_for_image(device, pCreateInfo))
+                       surface->flags |= RADEON_SURF_TC_COMPATIBLE_HTILE;
        }
 
        if (is_stencil)
@@ -116,20 +223,14 @@ radv_init_surface(struct radv_device *device,
 
        surface->flags |= RADEON_SURF_OPTIMIZE_FOR_SPACE;
 
-       if ((pCreateInfo->usage & (VK_IMAGE_USAGE_TRANSFER_SRC_BIT |
-                                  VK_IMAGE_USAGE_STORAGE_BIT)) ||
-           (pCreateInfo->flags & VK_IMAGE_CREATE_MUTABLE_FORMAT_BIT) ||
-            (pCreateInfo->tiling == VK_IMAGE_TILING_LINEAR) ||
-            pCreateInfo->mipLevels > 1 || pCreateInfo->arrayLayers > 1 ||
-            device->physical_device->rad_info.chip_class < VI ||
-            create_info->scanout || (device->debug_flags & RADV_DEBUG_NO_DCC) ||
-            !radv_is_colorbuffer_format_supported(pCreateInfo->format, &blendable))
+       if (!radv_use_dcc_for_image(device, create_info, pCreateInfo))
                surface->flags |= RADEON_SURF_DISABLE_DCC;
+
        if (create_info->scanout)
                surface->flags |= RADEON_SURF_SCANOUT;
        return 0;
 }
-#define ATI_VENDOR_ID 0x1002
+
 static uint32_t si_get_bo_metadata_word1(struct radv_device *device)
 {
        return (ATI_VENDOR_ID << 16) | device->physical_device->rad_info.pci_id;
@@ -172,7 +273,7 @@ radv_make_buffer_descriptor(struct radv_device *device,
 {
        const struct vk_format_description *desc;
        unsigned stride;
-       uint64_t gpu_address = device->ws->buffer_get_va(buffer->bo);
+       uint64_t gpu_address = radv_buffer_get_va(buffer->bo);
        uint64_t va = gpu_address + buffer->offset;
        unsigned num_format, data_format;
        int first_non_void;
@@ -207,9 +308,9 @@ si_set_mutable_tex_desc_fields(struct radv_device *device,
                               const struct legacy_surf_level *base_level_info,
                               unsigned base_level, unsigned first_level,
                               unsigned block_width, bool is_stencil,
-                              uint32_t *state)
+                              bool is_storage_image, uint32_t *state)
 {
-       uint64_t gpu_address = image->bo ? device->ws->buffer_get_va(image->bo) + image->offset : 0;
+       uint64_t gpu_address = image->bo ? radv_buffer_get_va(image->bo) + image->offset : 0;
        uint64_t va = gpu_address;
        enum chip_class chip_class = device->physical_device->rad_info.chip_class;
        uint64_t meta_va = 0;
@@ -231,10 +332,16 @@ si_set_mutable_tex_desc_fields(struct radv_device *device,
        if (chip_class >= VI) {
                state[6] &= C_008F28_COMPRESSION_EN;
                state[7] = 0;
-               if (image->surface.dcc_size && first_level < image->surface.num_dcc_levels) {
+               if (!is_storage_image && radv_dcc_enabled(image, first_level)) {
                        meta_va = gpu_address + image->dcc_offset;
                        if (chip_class <= VI)
                                meta_va += base_level_info->dcc_offset;
+               } else if (!is_storage_image &&
+                          radv_image_is_tc_compat_htile(image)) {
+                       meta_va = gpu_address + image->htile_offset;
+               }
+
+               if (meta_va) {
                        state[6] |= S_008F28_COMPRESSION_EN(1);
                        state[7] = meta_va >> 8;
                        state[7] |= image->surface.tile_swizzle;
@@ -307,7 +414,7 @@ static unsigned radv_tex_dim(VkImageType image_type, VkImageViewType view_type,
        }
 }
 
-static unsigned gfx9_border_color_swizzle(const unsigned char swizzle[4])
+static unsigned gfx9_border_color_swizzle(const enum vk_swizzle swizzle[4])
 {
        unsigned bc_swizzle = V_008F20_BC_SWIZZLE_XYZW;
 
@@ -378,6 +485,15 @@ si_make_texture_descriptor(struct radv_device *device,
                data_format = 0;
        }
 
+       /* S8 with either Z16 or Z32 HTILE need a special format. */
+       if (device->physical_device->rad_info.chip_class >= GFX9 &&
+           vk_format == VK_FORMAT_S8_UINT &&
+           radv_image_is_tc_compat_htile(image)) {
+               if (image->vk_format == VK_FORMAT_D32_SFLOAT_S8_UINT)
+                       data_format = V_008F14_IMG_DATA_FORMAT_S8_32;
+               else if (image->vk_format == VK_FORMAT_D16_UNORM_S8_UINT)
+                       data_format = V_008F14_IMG_DATA_FORMAT_S8_16;
+       }
        type = radv_tex_dim(image->type, view_type, image->info.array_size, image->info.samples,
                            is_storage_image, device->physical_device->rad_info.chip_class >= GFX9);
        if (type == V_008F1C_SQ_RSRC_IMG_1D_ARRAY) {
@@ -412,7 +528,7 @@ si_make_texture_descriptor(struct radv_device *device,
        state[7] = 0;
 
        if (device->physical_device->rad_info.chip_class >= GFX9) {
-               unsigned bc_swizzle = gfx9_border_color_swizzle(desc->swizzle);
+               unsigned bc_swizzle = gfx9_border_color_swizzle(swizzle);
 
                /* Depth is the the last accessible layer on Gfx9.
                 * The hw doesn't need to know the total number of layers.
@@ -448,9 +564,9 @@ si_make_texture_descriptor(struct radv_device *device,
        }
 
        /* Initialize the sampler view for FMASK. */
-       if (image->fmask.size) {
+       if (radv_image_has_fmask(image)) {
                uint32_t fmask_format, num_format;
-               uint64_t gpu_address = device->ws->buffer_get_va(image->bo);
+               uint64_t gpu_address = radv_buffer_get_va(image->bo);
                uint64_t va;
 
                va = gpu_address + image->offset + image->fmask.offset;
@@ -553,7 +669,7 @@ radv_query_opaque_metadata(struct radv_device *device,
                                   desc, NULL);
 
        si_set_mutable_tex_desc_fields(device, image, &image->surface.u.legacy.level[0], 0, 0,
-                                      image->surface.blk_w, false, desc);
+                                      image->surface.blk_w, false, false, desc);
 
        /* Clear the base address and set the relative DCC offset. */
        desc[0] = 0;
@@ -744,8 +860,7 @@ radv_image_alloc_cmask(struct radv_device *device,
 }
 
 static void
-radv_image_alloc_dcc(struct radv_device *device,
-                      struct radv_image *image)
+radv_image_alloc_dcc(struct radv_image *image)
 {
        image->dcc_offset = align64(image->size, image->surface.dcc_alignment);
        /* + 16 for storing the clear values + dcc pred */
@@ -756,14 +871,8 @@ radv_image_alloc_dcc(struct radv_device *device,
 }
 
 static void
-radv_image_alloc_htile(struct radv_device *device,
-                      struct radv_image *image)
+radv_image_alloc_htile(struct radv_image *image)
 {
-       if ((device->debug_flags & RADV_DEBUG_NO_HIZ) || image->info.levels > 1) {
-               image->surface.htile_size = 0;
-               return;
-       }
-
        image->htile_offset = align64(image->size, image->surface.htile_alignment);
 
        /* + 8 for storing the clear values */
@@ -772,6 +881,59 @@ radv_image_alloc_htile(struct radv_device *device,
        image->alignment = align64(image->alignment, image->surface.htile_alignment);
 }
 
+static inline bool
+radv_image_can_enable_dcc_or_cmask(struct radv_image *image)
+{
+       if (image->info.samples <= 1 &&
+           image->info.width * image->info.height <= 512 * 512) {
+               /* Do not enable CMASK or DCC for small surfaces where the cost
+                * of the eliminate pass can be higher than the benefit of fast
+                * clear. RadeonSI does this, but the image threshold is
+                * different.
+                */
+               return false;
+       }
+
+       return image->usage & VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT &&
+              (image->exclusive || image->queue_family_mask == 1);
+}
+
+static inline bool
+radv_image_can_enable_dcc(struct radv_image *image)
+{
+       return radv_image_can_enable_dcc_or_cmask(image) &&
+              radv_image_has_dcc(image);
+}
+
+static inline bool
+radv_image_can_enable_cmask(struct radv_image *image)
+{
+       if (image->surface.bpe > 8 && image->info.samples == 1) {
+               /* Do not enable CMASK for non-MSAA images (fast color clear)
+                * because 128 bit formats are not supported, but FMASK might
+                * still be used.
+                */
+               return false;
+       }
+
+       return radv_image_can_enable_dcc_or_cmask(image) &&
+              image->info.levels == 1 &&
+              image->info.depth == 1 &&
+              !image->surface.is_linear;
+}
+
+static inline bool
+radv_image_can_enable_fmask(struct radv_image *image)
+{
+       return image->info.samples > 1 && vk_format_is_color(image->vk_format);
+}
+
+static inline bool
+radv_image_can_enable_htile(struct radv_image *image)
+{
+       return image->info.levels == 1 && vk_format_is_depth(image->vk_format);
+}
+
 VkResult
 radv_image_create(VkDevice _device,
                  const struct radv_image_create_info *create_info,
@@ -781,7 +943,6 @@ radv_image_create(VkDevice _device,
        RADV_FROM_HANDLE(radv_device, device, _device);
        const VkImageCreateInfo *pCreateInfo = create_info->vk_info;
        struct radv_image *image = NULL;
-       bool can_cmask_dcc = false;
        assert(pCreateInfo->sType == VK_STRUCTURE_TYPE_IMAGE_CREATE_INFO);
 
        radv_assert(pCreateInfo->mipLevels > 0);
@@ -791,12 +952,11 @@ radv_image_create(VkDevice _device,
        radv_assert(pCreateInfo->extent.height > 0);
        radv_assert(pCreateInfo->extent.depth > 0);
 
-       image = vk_alloc2(&device->alloc, alloc, sizeof(*image), 8,
-                           VK_SYSTEM_ALLOCATION_SCOPE_OBJECT);
+       image = vk_zalloc2(&device->alloc, alloc, sizeof(*image), 8,
+                          VK_SYSTEM_ALLOCATION_SCOPE_OBJECT);
        if (!image)
                return vk_error(VK_ERROR_OUT_OF_HOST_MEMORY);
 
-       memset(image, 0, sizeof(*image));
        image->type = pCreateInfo->imageType;
        image->info.width = pCreateInfo->extent.width;
        image->info.height = pCreateInfo->extent.height;
@@ -804,6 +964,7 @@ radv_image_create(VkDevice _device,
        image->info.samples = pCreateInfo->samples;
        image->info.array_size = pCreateInfo->arrayLayers;
        image->info.levels = pCreateInfo->mipLevels;
+       image->info.num_channels = 4; /* TODO: set this correctly */
 
        image->vk_format = pCreateInfo->format;
        image->tiling = pCreateInfo->tiling;
@@ -832,26 +993,34 @@ radv_image_create(VkDevice _device,
        image->size = image->surface.surf_size;
        image->alignment = image->surface.surf_alignment;
 
-       if (image->exclusive || image->queue_family_mask == 1)
-               can_cmask_dcc = true;
+       if (!create_info->no_metadata_planes) {
+               /* Try to enable DCC first. */
+               if (radv_image_can_enable_dcc(image)) {
+                       radv_image_alloc_dcc(image);
+               } else {
+                       /* When DCC cannot be enabled, try CMASK. */
+                       image->surface.dcc_size = 0;
+                       if (radv_image_can_enable_cmask(image)) {
+                               radv_image_alloc_cmask(device, image);
+                       }
+               }
 
-       if ((pCreateInfo->usage & VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT) &&
-           image->surface.dcc_size && can_cmask_dcc)
-               radv_image_alloc_dcc(device, image);
-       else
+               /* Try to enable FMASK for multisampled images. */
+               if (radv_image_can_enable_fmask(image)) {
+                       radv_image_alloc_fmask(device, image);
+               } else {
+                       /* Otherwise, try to enable HTILE for depth surfaces. */
+                       if (radv_image_can_enable_htile(image) &&
+                           !(device->instance->debug_flags & RADV_DEBUG_NO_HIZ)) {
+                               radv_image_alloc_htile(image);
+                               image->tc_compatible_htile = image->surface.flags & RADEON_SURF_TC_COMPATIBLE_HTILE;
+                       } else {
+                               image->surface.htile_size = 0;
+                       }
+               }
+       } else {
                image->surface.dcc_size = 0;
-
-       if ((pCreateInfo->usage & VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT) &&
-           pCreateInfo->mipLevels == 1 &&
-           !image->surface.dcc_size && image->info.depth == 1 && can_cmask_dcc &&
-           !image->surface.is_linear)
-               radv_image_alloc_cmask(device, image);
-
-       if (image->info.samples > 1 && vk_format_is_color(pCreateInfo->format)) {
-               radv_image_alloc_fmask(device, image);
-       } else if (vk_format_is_depth(pCreateInfo->format)) {
-
-               radv_image_alloc_htile(device, image);
+               image->surface.htile_size = 0;
        }
 
        if (pCreateInfo->flags & VK_IMAGE_CREATE_SPARSE_BINDING_BIT) {
@@ -882,15 +1051,12 @@ radv_image_view_make_descriptor(struct radv_image_view *iview,
        bool is_stencil = iview->aspect_mask == VK_IMAGE_ASPECT_STENCIL_BIT;
        uint32_t blk_w;
        uint32_t *descriptor;
-       uint32_t *fmask_descriptor;
        uint32_t hw_level = 0;
 
        if (is_storage_image) {
                descriptor = iview->storage_descriptor;
-               fmask_descriptor = iview->storage_fmask_descriptor;
        } else {
                descriptor = iview->descriptor;
-               fmask_descriptor = iview->fmask_descriptor;
        }
 
        assert(image->surface.blk_w % vk_format_get_blockwidth(image->vk_format) == 0);
@@ -909,7 +1075,7 @@ radv_image_view_make_descriptor(struct radv_image_view *iview,
                                   iview->extent.height,
                                   iview->extent.depth,
                                   descriptor,
-                                  fmask_descriptor);
+                                  descriptor + 8);
 
        const struct legacy_surf_level *base_level_info = NULL;
        if (device->physical_device->rad_info.chip_class <= GFX9) {
@@ -922,7 +1088,7 @@ radv_image_view_make_descriptor(struct radv_image_view *iview,
                                       base_level_info,
                                       iview->base_mip,
                                       iview->base_mip,
-                                      blk_w, is_stencil, descriptor);
+                                      blk_w, is_stencil, is_storage_image, descriptor);
 }
 
 void
@@ -972,10 +1138,55 @@ radv_image_view_init(struct radv_image_view *iview,
        }
 
        if (iview->vk_format != image->vk_format) {
-               iview->extent.width = round_up_u32(iview->extent.width * vk_format_get_blockwidth(iview->vk_format),
-                                                  vk_format_get_blockwidth(image->vk_format));
-               iview->extent.height = round_up_u32(iview->extent.height * vk_format_get_blockheight(iview->vk_format),
-                                                   vk_format_get_blockheight(image->vk_format));
+               unsigned view_bw = vk_format_get_blockwidth(iview->vk_format);
+               unsigned view_bh = vk_format_get_blockheight(iview->vk_format);
+               unsigned img_bw = vk_format_get_blockwidth(image->vk_format);
+               unsigned img_bh = vk_format_get_blockheight(image->vk_format);
+
+               iview->extent.width = round_up_u32(iview->extent.width * view_bw, img_bw);
+               iview->extent.height = round_up_u32(iview->extent.height * view_bh, img_bh);
+
+               /* Comment ported from amdvlk -
+                * If we have the following image:
+                *              Uncompressed pixels   Compressed block sizes (4x4)
+                *      mip0:       22 x 22                   6 x 6
+                *      mip1:       11 x 11                   3 x 3
+                *      mip2:        5 x  5                   2 x 2
+                *      mip3:        2 x  2                   1 x 1
+                *      mip4:        1 x  1                   1 x 1
+                *
+                * On GFX9 the descriptor is always programmed with the WIDTH and HEIGHT of the base level and the HW is
+                * calculating the degradation of the block sizes down the mip-chain as follows (straight-up
+                * divide-by-two integer math):
+                *      mip0:  6x6
+                *      mip1:  3x3
+                *      mip2:  1x1
+                *      mip3:  1x1
+                *
+                * This means that mip2 will be missing texels.
+                *
+                * Fix this by calculating the base mip's width and height, then convert that, and round it
+                * back up to get the level 0 size.
+                * Clamp the converted size between the original values, and next power of two, which
+                * means we don't oversize the image.
+                */
+                if (device->physical_device->rad_info.chip_class >= GFX9 &&
+                    vk_format_is_compressed(image->vk_format) &&
+                    !vk_format_is_compressed(iview->vk_format)) {
+                        unsigned rounded_img_w = util_next_power_of_two(iview->extent.width);
+                        unsigned rounded_img_h = util_next_power_of_two(iview->extent.height);
+                        unsigned lvl_width  = radv_minify(image->info.width , range->baseMipLevel);
+                        unsigned lvl_height = radv_minify(image->info.height, range->baseMipLevel);
+
+                        lvl_width = round_up_u32(lvl_width * view_bw, img_bw);
+                        lvl_height = round_up_u32(lvl_height * view_bh, img_bh);
+
+                        lvl_width <<= range->baseMipLevel;
+                        lvl_height <<= range->baseMipLevel;
+
+                        iview->extent.width = CLAMP(lvl_width, iview->extent.width, rounded_img_w);
+                        iview->extent.height = CLAMP(lvl_height, iview->extent.height, rounded_img_h);
+                }
        }
 
        iview->base_layer = range->baseArrayLayer;
@@ -991,7 +1202,10 @@ bool radv_layout_has_htile(const struct radv_image *image,
                            VkImageLayout layout,
                            unsigned queue_mask)
 {
-       return image->surface.htile_size &&
+       if (radv_image_is_tc_compat_htile(image))
+               return layout != VK_IMAGE_LAYOUT_GENERAL;
+
+       return radv_image_has_htile(image) &&
               (layout == VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL ||
                layout == VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL) &&
               queue_mask == (1u << RADV_QUEUE_GENERAL);
@@ -1001,7 +1215,10 @@ bool radv_layout_is_htile_compressed(const struct radv_image *image,
                                      VkImageLayout layout,
                                      unsigned queue_mask)
 {
-       return image->surface.htile_size &&
+       if (radv_image_is_tc_compat_htile(image))
+               return layout != VK_IMAGE_LAYOUT_GENERAL;
+
+       return radv_image_has_htile(image) &&
               (layout == VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL ||
                layout == VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL) &&
               queue_mask == (1u << RADV_QUEUE_GENERAL);
@@ -1015,6 +1232,18 @@ bool radv_layout_can_fast_clear(const struct radv_image *image,
                queue_mask == (1u << RADV_QUEUE_GENERAL);
 }
 
+bool radv_layout_dcc_compressed(const struct radv_image *image,
+                               VkImageLayout layout,
+                               unsigned queue_mask)
+{
+       /* Don't compress compute transfer dst, as image stores are not supported. */
+       if (layout == VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL &&
+           (queue_mask & (1u << RADV_QUEUE_COMPUTE)))
+               return false;
+
+       return image->surface.num_dcc_levels > 0 && layout != VK_IMAGE_LAYOUT_GENERAL;
+}
+
 
 unsigned radv_image_queue_family_mask(const struct radv_image *image, uint32_t family, uint32_t queue_family)
 {
@@ -1033,11 +1262,24 @@ radv_CreateImage(VkDevice device,
                 const VkAllocationCallbacks *pAllocator,
                 VkImage *pImage)
 {
+#ifdef ANDROID
+       const VkNativeBufferANDROID *gralloc_info =
+               vk_find_struct_const(pCreateInfo->pNext, NATIVE_BUFFER_ANDROID);
+
+       if (gralloc_info)
+               return radv_image_from_gralloc(device, pCreateInfo, gralloc_info,
+                                             pAllocator, pImage);
+#endif
+
+       const struct wsi_image_create_info *wsi_info =
+               vk_find_struct_const(pCreateInfo->pNext, WSI_IMAGE_CREATE_INFO_MESA);
+       bool scanout = wsi_info && wsi_info->scanout;
+
        return radv_image_create(device,
                                 &(struct radv_image_create_info) {
                                         .vk_info = pCreateInfo,
-                                                .scanout = false,
-                                                },
+                                        .scanout = scanout,
+                                },
                                 pAllocator,
                                 pImage);
 }
@@ -1055,6 +1297,9 @@ radv_DestroyImage(VkDevice _device, VkImage _image,
        if (image->flags & VK_IMAGE_CREATE_SPARSE_BINDING_BIT)
                device->ws->buffer_destroy(image->bo);
 
+       if (image->owned_memory != VK_NULL_HANDLE)
+               radv_FreeMemory(_device, image->owned_memory, pAllocator);
+
        vk_free2(&device->alloc, pAllocator, image);
 }
 
@@ -1079,11 +1324,11 @@ void radv_GetImageSubresourceLayout(
                if (image->type == VK_IMAGE_TYPE_3D)
                        pLayout->size *= u_minify(image->info.depth, level);
        } else {
-               pLayout->offset = surface->u.legacy.level[level].offset + surface->u.legacy.level[level].slice_size * layer;
+               pLayout->offset = surface->u.legacy.level[level].offset + (uint64_t)surface->u.legacy.level[level].slice_size_dw * 4 * layer;
                pLayout->rowPitch = surface->u.legacy.level[level].nblk_x * surface->bpe;
-               pLayout->arrayPitch = surface->u.legacy.level[level].slice_size;
-               pLayout->depthPitch = surface->u.legacy.level[level].slice_size;
-               pLayout->size = surface->u.legacy.level[level].slice_size;
+               pLayout->arrayPitch = (uint64_t)surface->u.legacy.level[level].slice_size_dw * 4;
+               pLayout->depthPitch = (uint64_t)surface->u.legacy.level[level].slice_size_dw * 4;
+               pLayout->size = (uint64_t)surface->u.legacy.level[level].slice_size_dw * 4;
                if (image->type == VK_IMAGE_TYPE_3D)
                        pLayout->size *= u_minify(image->info.depth, level);
        }
@@ -1125,8 +1370,7 @@ radv_DestroyImageView(VkDevice _device, VkImageView _iview,
 
 void radv_buffer_view_init(struct radv_buffer_view *view,
                           struct radv_device *device,
-                          const VkBufferViewCreateInfo* pCreateInfo,
-                          struct radv_cmd_buffer *cmd_buffer)
+                          const VkBufferViewCreateInfo* pCreateInfo)
 {
        RADV_FROM_HANDLE(radv_buffer, buffer, pCreateInfo->buffer);
 
@@ -1153,7 +1397,7 @@ radv_CreateBufferView(VkDevice _device,
        if (!view)
                return vk_error(VK_ERROR_OUT_OF_HOST_MEMORY);
 
-       radv_buffer_view_init(view, device, pCreateInfo, NULL);
+       radv_buffer_view_init(view, device, pCreateInfo);
 
        *pView = radv_buffer_view_to_handle(view);