radv: Add on-demand compilation of built-in shaders.
[mesa.git] / src / amd / vulkan / radv_meta_resolve.c
1 /*
2 * Copyright © 2016 Intel Corporation
3 *
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
10 *
11 * The above copyright notice and this permission notice (including the next
12 * paragraph) shall be included in all copies or substantial portions of the
13 * Software.
14 *
15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
18 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
19 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
20 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
21 * IN THE SOFTWARE.
22 */
23
24 #include <assert.h>
25 #include <stdbool.h>
26
27 #include "radv_meta.h"
28 #include "radv_private.h"
29 #include "vk_format.h"
30 #include "nir/nir_builder.h"
31 #include "sid.h"
32
33 /* emit 0, 0, 0, 1 */
34 static nir_shader *
35 build_nir_fs(void)
36 {
37 const struct glsl_type *vec4 = glsl_vec4_type();
38 nir_builder b;
39 nir_variable *f_color; /* vec4, fragment output color */
40
41 nir_builder_init_simple_shader(&b, NULL, MESA_SHADER_FRAGMENT, NULL);
42 b.shader->info.name = ralloc_asprintf(b.shader,
43 "meta_resolve_fs");
44
45 f_color = nir_variable_create(b.shader, nir_var_shader_out, vec4,
46 "f_color");
47 f_color->data.location = FRAG_RESULT_DATA0;
48 nir_store_var(&b, f_color, nir_imm_vec4(&b, 0.0, 0.0, 0.0, 1.0), 0xf);
49
50 return b.shader;
51 }
52
53 static VkResult
54 create_pass(struct radv_device *device, VkFormat vk_format, VkRenderPass *pass)
55 {
56 VkResult result;
57 VkDevice device_h = radv_device_to_handle(device);
58 const VkAllocationCallbacks *alloc = &device->meta_state.alloc;
59 VkAttachmentDescription attachments[2];
60 int i;
61
62 for (i = 0; i < 2; i++) {
63 attachments[i].format = vk_format;
64 attachments[i].samples = 1;
65 attachments[i].loadOp = VK_ATTACHMENT_LOAD_OP_LOAD;
66 attachments[i].storeOp = VK_ATTACHMENT_STORE_OP_STORE;
67 }
68 attachments[0].initialLayout = VK_IMAGE_LAYOUT_GENERAL;
69 attachments[0].finalLayout = VK_IMAGE_LAYOUT_GENERAL;
70 attachments[1].initialLayout = VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
71 attachments[1].finalLayout = VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
72
73 result = radv_CreateRenderPass(device_h,
74 &(VkRenderPassCreateInfo) {
75 .sType = VK_STRUCTURE_TYPE_RENDER_PASS_CREATE_INFO,
76 .attachmentCount = 2,
77 .pAttachments = attachments,
78 .subpassCount = 1,
79 .pSubpasses = &(VkSubpassDescription) {
80 .pipelineBindPoint = VK_PIPELINE_BIND_POINT_GRAPHICS,
81 .inputAttachmentCount = 0,
82 .colorAttachmentCount = 2,
83 .pColorAttachments = (VkAttachmentReference[]) {
84 {
85 .attachment = 0,
86 .layout = VK_IMAGE_LAYOUT_GENERAL,
87 },
88 {
89 .attachment = 1,
90 .layout = VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL,
91 },
92 },
93 .pResolveAttachments = NULL,
94 .pDepthStencilAttachment = &(VkAttachmentReference) {
95 .attachment = VK_ATTACHMENT_UNUSED,
96 },
97 .preserveAttachmentCount = 0,
98 .pPreserveAttachments = NULL,
99 },
100 .dependencyCount = 0,
101 },
102 alloc,
103 pass);
104
105 return result;
106 }
107
108 static VkResult
109 create_pipeline(struct radv_device *device,
110 VkShaderModule vs_module_h,
111 VkPipeline *pipeline,
112 VkRenderPass pass)
113 {
114 VkResult result;
115 VkDevice device_h = radv_device_to_handle(device);
116
117 struct radv_shader_module fs_module = {
118 .nir = build_nir_fs(),
119 };
120
121 if (!fs_module.nir) {
122 /* XXX: Need more accurate error */
123 result = VK_ERROR_OUT_OF_HOST_MEMORY;
124 goto cleanup;
125 }
126
127 VkPipelineLayoutCreateInfo pl_create_info = {
128 .sType = VK_STRUCTURE_TYPE_PIPELINE_LAYOUT_CREATE_INFO,
129 .setLayoutCount = 0,
130 .pSetLayouts = NULL,
131 .pushConstantRangeCount = 0,
132 .pPushConstantRanges = NULL,
133 };
134
135 if (!device->meta_state.resolve.p_layout) {
136 result = radv_CreatePipelineLayout(radv_device_to_handle(device),
137 &pl_create_info,
138 &device->meta_state.alloc,
139 &device->meta_state.resolve.p_layout);
140 if (result != VK_SUCCESS)
141 goto cleanup;
142 }
143
144 result = radv_graphics_pipeline_create(device_h,
145 radv_pipeline_cache_to_handle(&device->meta_state.cache),
146 &(VkGraphicsPipelineCreateInfo) {
147 .sType = VK_STRUCTURE_TYPE_GRAPHICS_PIPELINE_CREATE_INFO,
148 .stageCount = 2,
149 .pStages = (VkPipelineShaderStageCreateInfo[]) {
150 {
151 .sType = VK_STRUCTURE_TYPE_PIPELINE_SHADER_STAGE_CREATE_INFO,
152 .stage = VK_SHADER_STAGE_VERTEX_BIT,
153 .module = vs_module_h,
154 .pName = "main",
155 },
156 {
157 .sType = VK_STRUCTURE_TYPE_PIPELINE_SHADER_STAGE_CREATE_INFO,
158 .stage = VK_SHADER_STAGE_FRAGMENT_BIT,
159 .module = radv_shader_module_to_handle(&fs_module),
160 .pName = "main",
161 },
162 },
163 .pVertexInputState = &(VkPipelineVertexInputStateCreateInfo) {
164 .sType = VK_STRUCTURE_TYPE_PIPELINE_VERTEX_INPUT_STATE_CREATE_INFO,
165 .vertexBindingDescriptionCount = 0,
166 .vertexAttributeDescriptionCount = 0,
167 },
168 .pInputAssemblyState = &(VkPipelineInputAssemblyStateCreateInfo) {
169 .sType = VK_STRUCTURE_TYPE_PIPELINE_INPUT_ASSEMBLY_STATE_CREATE_INFO,
170 .topology = VK_PRIMITIVE_TOPOLOGY_TRIANGLE_STRIP,
171 .primitiveRestartEnable = false,
172 },
173 .pViewportState = &(VkPipelineViewportStateCreateInfo) {
174 .sType = VK_STRUCTURE_TYPE_PIPELINE_VIEWPORT_STATE_CREATE_INFO,
175 .viewportCount = 1,
176 .scissorCount = 1,
177 },
178 .pRasterizationState = &(VkPipelineRasterizationStateCreateInfo) {
179 .sType = VK_STRUCTURE_TYPE_PIPELINE_RASTERIZATION_STATE_CREATE_INFO,
180 .depthClampEnable = false,
181 .rasterizerDiscardEnable = false,
182 .polygonMode = VK_POLYGON_MODE_FILL,
183 .cullMode = VK_CULL_MODE_NONE,
184 .frontFace = VK_FRONT_FACE_COUNTER_CLOCKWISE,
185 },
186 .pMultisampleState = &(VkPipelineMultisampleStateCreateInfo) {
187 .sType = VK_STRUCTURE_TYPE_PIPELINE_MULTISAMPLE_STATE_CREATE_INFO,
188 .rasterizationSamples = 1,
189 .sampleShadingEnable = false,
190 .pSampleMask = NULL,
191 .alphaToCoverageEnable = false,
192 .alphaToOneEnable = false,
193 },
194 .pColorBlendState = &(VkPipelineColorBlendStateCreateInfo) {
195 .sType = VK_STRUCTURE_TYPE_PIPELINE_COLOR_BLEND_STATE_CREATE_INFO,
196 .logicOpEnable = false,
197 .attachmentCount = 2,
198 .pAttachments = (VkPipelineColorBlendAttachmentState []) {
199 {
200 .colorWriteMask = VK_COLOR_COMPONENT_R_BIT |
201 VK_COLOR_COMPONENT_G_BIT |
202 VK_COLOR_COMPONENT_B_BIT |
203 VK_COLOR_COMPONENT_A_BIT,
204 },
205 {
206 .colorWriteMask = 0,
207
208 }
209 },
210 },
211 .pDynamicState = &(VkPipelineDynamicStateCreateInfo) {
212 .sType = VK_STRUCTURE_TYPE_PIPELINE_DYNAMIC_STATE_CREATE_INFO,
213 .dynamicStateCount = 2,
214 .pDynamicStates = (VkDynamicState[]) {
215 VK_DYNAMIC_STATE_VIEWPORT,
216 VK_DYNAMIC_STATE_SCISSOR,
217 },
218 },
219 .layout = device->meta_state.resolve.p_layout,
220 .renderPass = pass,
221 .subpass = 0,
222 },
223 &(struct radv_graphics_pipeline_create_info) {
224 .use_rectlist = true,
225 .custom_blend_mode = V_028808_CB_RESOLVE,
226 },
227 &device->meta_state.alloc, pipeline);
228 if (result != VK_SUCCESS)
229 goto cleanup;
230
231 goto cleanup;
232
233 cleanup:
234 ralloc_free(fs_module.nir);
235 return result;
236 }
237
238 void
239 radv_device_finish_meta_resolve_state(struct radv_device *device)
240 {
241 struct radv_meta_state *state = &device->meta_state;
242
243 for (uint32_t j = 0; j < NUM_META_FS_KEYS; j++) {
244 radv_DestroyRenderPass(radv_device_to_handle(device),
245 state->resolve.pass[j], &state->alloc);
246 radv_DestroyPipeline(radv_device_to_handle(device),
247 state->resolve.pipeline[j], &state->alloc);
248 }
249 radv_DestroyPipelineLayout(radv_device_to_handle(device),
250 state->resolve.p_layout, &state->alloc);
251
252 }
253
254 VkResult
255 radv_device_init_meta_resolve_state(struct radv_device *device, bool on_demand)
256 {
257 if (on_demand)
258 return VK_SUCCESS;
259
260 VkResult res = VK_SUCCESS;
261 struct radv_meta_state *state = &device->meta_state;
262 struct radv_shader_module vs_module = { .nir = radv_meta_build_nir_vs_generate_vertices() };
263 if (!vs_module.nir) {
264 /* XXX: Need more accurate error */
265 res = VK_ERROR_OUT_OF_HOST_MEMORY;
266 goto fail;
267 }
268
269 for (uint32_t i = 0; i < NUM_META_FS_KEYS; ++i) {
270 VkFormat format = radv_fs_key_format_exemplars[i];
271 unsigned fs_key = radv_format_meta_fs_key(format);
272 res = create_pass(device, format, &state->resolve.pass[fs_key]);
273 if (res != VK_SUCCESS)
274 goto fail;
275
276 VkShaderModule vs_module_h = radv_shader_module_to_handle(&vs_module);
277 res = create_pipeline(device, vs_module_h,
278 &state->resolve.pipeline[fs_key], state->resolve.pass[fs_key]);
279 if (res != VK_SUCCESS)
280 goto fail;
281 }
282
283 goto cleanup;
284
285 fail:
286 radv_device_finish_meta_resolve_state(device);
287
288 cleanup:
289 ralloc_free(vs_module.nir);
290
291 return res;
292 }
293
294 static void
295 emit_resolve(struct radv_cmd_buffer *cmd_buffer,
296 VkFormat vk_format,
297 const VkOffset2D *dest_offset,
298 const VkExtent2D *resolve_extent)
299 {
300 struct radv_device *device = cmd_buffer->device;
301 VkCommandBuffer cmd_buffer_h = radv_cmd_buffer_to_handle(cmd_buffer);
302 unsigned fs_key = radv_format_meta_fs_key(vk_format);
303
304 cmd_buffer->state.flush_bits |= RADV_CMD_FLAG_FLUSH_AND_INV_CB;
305
306 radv_CmdBindPipeline(cmd_buffer_h, VK_PIPELINE_BIND_POINT_GRAPHICS,
307 device->meta_state.resolve.pipeline[fs_key]);
308
309 radv_CmdSetViewport(radv_cmd_buffer_to_handle(cmd_buffer), 0, 1, &(VkViewport) {
310 .x = dest_offset->x,
311 .y = dest_offset->y,
312 .width = resolve_extent->width,
313 .height = resolve_extent->height,
314 .minDepth = 0.0f,
315 .maxDepth = 1.0f
316 });
317
318 radv_CmdSetScissor(radv_cmd_buffer_to_handle(cmd_buffer), 0, 1, &(VkRect2D) {
319 .offset = *dest_offset,
320 .extent = *resolve_extent,
321 });
322
323 radv_CmdDraw(cmd_buffer_h, 3, 1, 0, 0);
324 cmd_buffer->state.flush_bits |= RADV_CMD_FLAG_FLUSH_AND_INV_CB;
325 }
326
327 enum radv_resolve_method {
328 RESOLVE_HW,
329 RESOLVE_COMPUTE,
330 RESOLVE_FRAGMENT,
331 };
332
333 static void radv_pick_resolve_method_images(struct radv_image *src_image,
334 struct radv_image *dest_image,
335 VkImageLayout dest_image_layout,
336 struct radv_cmd_buffer *cmd_buffer,
337 enum radv_resolve_method *method)
338
339 {
340 uint32_t queue_mask = radv_image_queue_family_mask(dest_image,
341 cmd_buffer->queue_family_index,
342 cmd_buffer->queue_family_index);
343
344 if (src_image->vk_format == VK_FORMAT_R16G16_UNORM ||
345 src_image->vk_format == VK_FORMAT_R16G16_SNORM)
346 *method = RESOLVE_COMPUTE;
347 else if (vk_format_is_int(src_image->vk_format))
348 *method = RESOLVE_COMPUTE;
349 else if (src_image->info.array_size > 1)
350 *method = RESOLVE_COMPUTE;
351
352 if (radv_layout_dcc_compressed(dest_image, dest_image_layout, queue_mask)) {
353 *method = RESOLVE_FRAGMENT;
354 } else if (dest_image->surface.micro_tile_mode != src_image->surface.micro_tile_mode) {
355 *method = RESOLVE_COMPUTE;
356 }
357 }
358
359 static VkResult
360 build_resolve_pipeline(struct radv_device *device,
361 unsigned fs_key)
362 {
363 VkResult result = VK_SUCCESS;
364
365 if (device->meta_state.resolve.pipeline[fs_key])
366 return result;
367
368 mtx_lock(&device->meta_state.mtx);
369 if (device->meta_state.resolve.pipeline[fs_key]) {
370 mtx_unlock(&device->meta_state.mtx);
371 return result;
372 }
373
374 struct radv_shader_module vs_module = { .nir = radv_meta_build_nir_vs_generate_vertices() };
375
376 result = create_pass(device, radv_fs_key_format_exemplars[fs_key], &device->meta_state.resolve.pass[fs_key]);
377 if (result != VK_SUCCESS)
378 goto fail;
379
380 VkShaderModule vs_module_h = radv_shader_module_to_handle(&vs_module);
381 result = create_pipeline(device, vs_module_h, &device->meta_state.resolve.pipeline[fs_key], device->meta_state.resolve.pass[fs_key]);
382
383 fail:
384 ralloc_free(vs_module.nir);
385 mtx_unlock(&device->meta_state.mtx);
386 return result;
387 }
388
389 void radv_CmdResolveImage(
390 VkCommandBuffer cmd_buffer_h,
391 VkImage src_image_h,
392 VkImageLayout src_image_layout,
393 VkImage dest_image_h,
394 VkImageLayout dest_image_layout,
395 uint32_t region_count,
396 const VkImageResolve* regions)
397 {
398 RADV_FROM_HANDLE(radv_cmd_buffer, cmd_buffer, cmd_buffer_h);
399 RADV_FROM_HANDLE(radv_image, src_image, src_image_h);
400 RADV_FROM_HANDLE(radv_image, dest_image, dest_image_h);
401 struct radv_device *device = cmd_buffer->device;
402 struct radv_meta_saved_state saved_state;
403 VkDevice device_h = radv_device_to_handle(device);
404 enum radv_resolve_method resolve_method = RESOLVE_HW;
405 /* we can use the hw resolve only for single full resolves */
406 if (region_count == 1) {
407 if (regions[0].srcOffset.x ||
408 regions[0].srcOffset.y ||
409 regions[0].srcOffset.z)
410 resolve_method = RESOLVE_COMPUTE;
411 if (regions[0].dstOffset.x ||
412 regions[0].dstOffset.y ||
413 regions[0].dstOffset.z)
414 resolve_method = RESOLVE_COMPUTE;
415
416 if (regions[0].extent.width != src_image->info.width ||
417 regions[0].extent.height != src_image->info.height ||
418 regions[0].extent.depth != src_image->info.depth)
419 resolve_method = RESOLVE_COMPUTE;
420 } else
421 resolve_method = RESOLVE_COMPUTE;
422
423 radv_pick_resolve_method_images(src_image, dest_image,
424 dest_image_layout, cmd_buffer,
425 &resolve_method);
426
427 if (resolve_method == RESOLVE_FRAGMENT) {
428 radv_meta_resolve_fragment_image(cmd_buffer,
429 src_image,
430 src_image_layout,
431 dest_image,
432 dest_image_layout,
433 region_count, regions);
434 return;
435 }
436
437 if (resolve_method == RESOLVE_COMPUTE) {
438 radv_meta_resolve_compute_image(cmd_buffer,
439 src_image,
440 src_image_layout,
441 dest_image,
442 dest_image_layout,
443 region_count, regions);
444 return;
445 }
446
447 radv_meta_save(&saved_state, cmd_buffer,
448 RADV_META_SAVE_GRAPHICS_PIPELINE);
449
450 assert(src_image->info.samples > 1);
451 if (src_image->info.samples <= 1) {
452 /* this causes GPU hangs if we get past here */
453 fprintf(stderr, "radv: Illegal resolve operation (src not multisampled), will hang GPU.");
454 return;
455 }
456 assert(dest_image->info.samples == 1);
457
458 if (src_image->info.samples >= 16) {
459 /* See commit aa3f9aaf31e9056a255f9e0472ebdfdaa60abe54 for the
460 * glBlitFramebuffer workaround for samples >= 16.
461 */
462 radv_finishme("vkCmdResolveImage: need interpolation workaround when "
463 "samples >= 16");
464 }
465
466 if (src_image->info.array_size > 1)
467 radv_finishme("vkCmdResolveImage: multisample array images");
468
469 if (radv_image_has_dcc(dest_image)) {
470 radv_initialize_dcc(cmd_buffer, dest_image, 0xffffffff);
471 }
472 unsigned fs_key = radv_format_meta_fs_key(dest_image->vk_format);
473 for (uint32_t r = 0; r < region_count; ++r) {
474 const VkImageResolve *region = &regions[r];
475
476 /* From the Vulkan 1.0 spec:
477 *
478 * - The aspectMask member of srcSubresource and dstSubresource must
479 * only contain VK_IMAGE_ASPECT_COLOR_BIT
480 *
481 * - The layerCount member of srcSubresource and dstSubresource must
482 * match
483 */
484 assert(region->srcSubresource.aspectMask == VK_IMAGE_ASPECT_COLOR_BIT);
485 assert(region->dstSubresource.aspectMask == VK_IMAGE_ASPECT_COLOR_BIT);
486 assert(region->srcSubresource.layerCount ==
487 region->dstSubresource.layerCount);
488
489 const uint32_t src_base_layer =
490 radv_meta_get_iview_layer(src_image, &region->srcSubresource,
491 &region->srcOffset);
492
493 const uint32_t dest_base_layer =
494 radv_meta_get_iview_layer(dest_image, &region->dstSubresource,
495 &region->dstOffset);
496
497 /**
498 * From Vulkan 1.0.6 spec: 18.6 Resolving Multisample Images
499 *
500 * extent is the size in texels of the source image to resolve in width,
501 * height and depth. 1D images use only x and width. 2D images use x, y,
502 * width and height. 3D images use x, y, z, width, height and depth.
503 *
504 * srcOffset and dstOffset select the initial x, y, and z offsets in
505 * texels of the sub-regions of the source and destination image data.
506 * extent is the size in texels of the source image to resolve in width,
507 * height and depth. 1D images use only x and width. 2D images use x, y,
508 * width and height. 3D images use x, y, z, width, height and depth.
509 */
510 const struct VkExtent3D extent =
511 radv_sanitize_image_extent(src_image->type, region->extent);
512 const struct VkOffset3D dstOffset =
513 radv_sanitize_image_offset(dest_image->type, region->dstOffset);
514
515
516 for (uint32_t layer = 0; layer < region->srcSubresource.layerCount;
517 ++layer) {
518
519 VkResult ret = build_resolve_pipeline(device, fs_key);
520 if (ret != VK_SUCCESS) {
521 cmd_buffer->record_result = ret;
522 break;
523 }
524
525 struct radv_image_view src_iview;
526 radv_image_view_init(&src_iview, cmd_buffer->device,
527 &(VkImageViewCreateInfo) {
528 .sType = VK_STRUCTURE_TYPE_IMAGE_VIEW_CREATE_INFO,
529 .image = src_image_h,
530 .viewType = radv_meta_get_view_type(src_image),
531 .format = src_image->vk_format,
532 .subresourceRange = {
533 .aspectMask = VK_IMAGE_ASPECT_COLOR_BIT,
534 .baseMipLevel = region->srcSubresource.mipLevel,
535 .levelCount = 1,
536 .baseArrayLayer = src_base_layer + layer,
537 .layerCount = 1,
538 },
539 });
540
541 struct radv_image_view dest_iview;
542 radv_image_view_init(&dest_iview, cmd_buffer->device,
543 &(VkImageViewCreateInfo) {
544 .sType = VK_STRUCTURE_TYPE_IMAGE_VIEW_CREATE_INFO,
545 .image = dest_image_h,
546 .viewType = radv_meta_get_view_type(dest_image),
547 .format = dest_image->vk_format,
548 .subresourceRange = {
549 .aspectMask = VK_IMAGE_ASPECT_COLOR_BIT,
550 .baseMipLevel = region->dstSubresource.mipLevel,
551 .levelCount = 1,
552 .baseArrayLayer = dest_base_layer + layer,
553 .layerCount = 1,
554 },
555 });
556
557 VkFramebuffer fb_h;
558 radv_CreateFramebuffer(device_h,
559 &(VkFramebufferCreateInfo) {
560 .sType = VK_STRUCTURE_TYPE_FRAMEBUFFER_CREATE_INFO,
561 .attachmentCount = 2,
562 .pAttachments = (VkImageView[]) {
563 radv_image_view_to_handle(&src_iview),
564 radv_image_view_to_handle(&dest_iview),
565 },
566 .width = radv_minify(dest_image->info.width,
567 region->dstSubresource.mipLevel),
568 .height = radv_minify(dest_image->info.height,
569 region->dstSubresource.mipLevel),
570 .layers = 1
571 },
572 &cmd_buffer->pool->alloc,
573 &fb_h);
574
575 radv_CmdBeginRenderPass(cmd_buffer_h,
576 &(VkRenderPassBeginInfo) {
577 .sType = VK_STRUCTURE_TYPE_RENDER_PASS_BEGIN_INFO,
578 .renderPass = device->meta_state.resolve.pass[fs_key],
579 .framebuffer = fb_h,
580 .renderArea = {
581 .offset = {
582 dstOffset.x,
583 dstOffset.y,
584 },
585 .extent = {
586 extent.width,
587 extent.height,
588 }
589 },
590 .clearValueCount = 0,
591 .pClearValues = NULL,
592 },
593 VK_SUBPASS_CONTENTS_INLINE);
594
595 emit_resolve(cmd_buffer,
596 dest_iview.vk_format,
597 &(VkOffset2D) {
598 .x = dstOffset.x,
599 .y = dstOffset.y,
600 },
601 &(VkExtent2D) {
602 .width = extent.width,
603 .height = extent.height,
604 });
605
606 radv_CmdEndRenderPass(cmd_buffer_h);
607
608 radv_DestroyFramebuffer(device_h, fb_h,
609 &cmd_buffer->pool->alloc);
610 }
611 }
612
613 radv_meta_restore(&saved_state, cmd_buffer);
614 }
615
616 /**
617 * Emit any needed resolves for the current subpass.
618 */
619 void
620 radv_cmd_buffer_resolve_subpass(struct radv_cmd_buffer *cmd_buffer)
621 {
622 struct radv_framebuffer *fb = cmd_buffer->state.framebuffer;
623 const struct radv_subpass *subpass = cmd_buffer->state.subpass;
624 struct radv_meta_saved_state saved_state;
625 enum radv_resolve_method resolve_method = RESOLVE_HW;
626
627 /* FINISHME(perf): Skip clears for resolve attachments.
628 *
629 * From the Vulkan 1.0 spec:
630 *
631 * If the first use of an attachment in a render pass is as a resolve
632 * attachment, then the loadOp is effectively ignored as the resolve is
633 * guaranteed to overwrite all pixels in the render area.
634 */
635
636 if (!subpass->has_resolve)
637 return;
638
639 for (uint32_t i = 0; i < subpass->color_count; ++i) {
640 struct radv_subpass_attachment src_att = subpass->color_attachments[i];
641 struct radv_subpass_attachment dest_att = subpass->resolve_attachments[i];
642
643 if (src_att.attachment == VK_ATTACHMENT_UNUSED ||
644 dest_att.attachment == VK_ATTACHMENT_UNUSED)
645 continue;
646
647 struct radv_image *dst_img = cmd_buffer->state.framebuffer->attachments[dest_att.attachment].attachment->image;
648 struct radv_image *src_img = cmd_buffer->state.framebuffer->attachments[src_att.attachment].attachment->image;
649
650 radv_pick_resolve_method_images(src_img, dst_img, dest_att.layout, cmd_buffer, &resolve_method);
651 if (resolve_method == RESOLVE_FRAGMENT) {
652 break;
653 }
654 }
655
656 if (resolve_method == RESOLVE_COMPUTE) {
657 radv_cmd_buffer_resolve_subpass_cs(cmd_buffer);
658 return;
659 } else if (resolve_method == RESOLVE_FRAGMENT) {
660 radv_cmd_buffer_resolve_subpass_fs(cmd_buffer);
661 return;
662 }
663
664 radv_meta_save(&saved_state, cmd_buffer,
665 RADV_META_SAVE_GRAPHICS_PIPELINE);
666
667 for (uint32_t i = 0; i < subpass->color_count; ++i) {
668 struct radv_subpass_attachment src_att = subpass->color_attachments[i];
669 struct radv_subpass_attachment dest_att = subpass->resolve_attachments[i];
670
671 if (src_att.attachment == VK_ATTACHMENT_UNUSED ||
672 dest_att.attachment == VK_ATTACHMENT_UNUSED)
673 continue;
674
675 struct radv_image *dst_img = cmd_buffer->state.framebuffer->attachments[dest_att.attachment].attachment->image;
676
677 if (radv_image_has_dcc(dst_img)) {
678 radv_initialize_dcc(cmd_buffer, dst_img, 0xffffffff);
679 cmd_buffer->state.attachments[dest_att.attachment].current_layout = VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
680 }
681
682 struct radv_subpass resolve_subpass = {
683 .color_count = 2,
684 .color_attachments = (struct radv_subpass_attachment[]) { src_att, dest_att },
685 .depth_stencil_attachment = { .attachment = VK_ATTACHMENT_UNUSED },
686 };
687
688 radv_cmd_buffer_set_subpass(cmd_buffer, &resolve_subpass, false);
689
690 VkResult ret = build_resolve_pipeline(cmd_buffer->device, radv_format_meta_fs_key(dst_img->vk_format));
691 if (ret != VK_SUCCESS) {
692 cmd_buffer->record_result = ret;
693 continue;
694 }
695
696 emit_resolve(cmd_buffer,
697 dst_img->vk_format,
698 &(VkOffset2D) { 0, 0 },
699 &(VkExtent2D) { fb->width, fb->height });
700 }
701
702 cmd_buffer->state.subpass = subpass;
703 radv_meta_restore(&saved_state, cmd_buffer);
704 }
705
706 /**
707 * Decompress CMask/FMask before resolving a multisampled source image inside a
708 * subpass.
709 */
710 void
711 radv_decompress_resolve_subpass_src(struct radv_cmd_buffer *cmd_buffer)
712 {
713 const struct radv_subpass *subpass = cmd_buffer->state.subpass;
714 struct radv_framebuffer *fb = cmd_buffer->state.framebuffer;
715
716 for (uint32_t i = 0; i < subpass->color_count; ++i) {
717 struct radv_subpass_attachment src_att = subpass->color_attachments[i];
718 struct radv_subpass_attachment dest_att = subpass->resolve_attachments[i];
719
720 if (src_att.attachment == VK_ATTACHMENT_UNUSED ||
721 dest_att.attachment == VK_ATTACHMENT_UNUSED)
722 continue;
723
724 struct radv_image *src_image =
725 fb->attachments[src_att.attachment].attachment->image;
726
727 VkImageResolve region = {};
728 region.srcSubresource.baseArrayLayer = 0;
729 region.srcSubresource.mipLevel = 0;
730 region.srcSubresource.layerCount = src_image->info.array_size;
731
732 radv_decompress_resolve_src(cmd_buffer, src_image,
733 src_att.layout, 1, &region);
734 }
735 }
736
737 /**
738 * Decompress CMask/FMask before resolving a multisampled source image.
739 */
740 void
741 radv_decompress_resolve_src(struct radv_cmd_buffer *cmd_buffer,
742 struct radv_image *src_image,
743 VkImageLayout src_image_layout,
744 uint32_t region_count,
745 const VkImageResolve *regions)
746 {
747 for (uint32_t r = 0; r < region_count; ++r) {
748 const VkImageResolve *region = &regions[r];
749 const uint32_t src_base_layer =
750 radv_meta_get_iview_layer(src_image, &region->srcSubresource,
751 &region->srcOffset);
752 VkImageSubresourceRange range;
753 range.aspectMask = VK_IMAGE_ASPECT_COLOR_BIT;
754 range.baseMipLevel = region->srcSubresource.mipLevel;
755 range.levelCount = 1;
756 range.baseArrayLayer = src_base_layer;
757 range.layerCount = region->srcSubresource.layerCount;
758
759 uint32_t queue_mask =
760 radv_image_queue_family_mask(src_image,
761 cmd_buffer->queue_family_index,
762 cmd_buffer->queue_family_index);
763
764 if (radv_layout_dcc_compressed(src_image, src_image_layout,
765 queue_mask)) {
766 radv_decompress_dcc(cmd_buffer, src_image, &range);
767 } else {
768 radv_fast_clear_flush_image_inplace(cmd_buffer,
769 src_image, &range);
770 }
771 }
772 }