2 * Copyright © 2015 Intel Corporation
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
11 * The above copyright notice and this permission notice (including the next
12 * paragraph) shall be included in all copies or substantial portions of the
15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
18 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
19 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
20 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
24 #include "anv_private.h"
26 #include "genxml/gen_macros.h"
27 #include "genxml/genX_pack.h"
30 compute_pipeline_create(
32 struct anv_pipeline_cache
* cache
,
33 const VkComputePipelineCreateInfo
* pCreateInfo
,
34 const VkAllocationCallbacks
* pAllocator
,
35 VkPipeline
* pPipeline
)
37 ANV_FROM_HANDLE(anv_device
, device
, _device
);
38 const struct anv_physical_device
*physical_device
=
39 &device
->instance
->physicalDevice
;
40 const struct gen_device_info
*devinfo
= &physical_device
->info
;
41 struct anv_pipeline
*pipeline
;
44 assert(pCreateInfo
->sType
== VK_STRUCTURE_TYPE_COMPUTE_PIPELINE_CREATE_INFO
);
46 pipeline
= vk_alloc2(&device
->alloc
, pAllocator
, sizeof(*pipeline
), 8,
47 VK_SYSTEM_ALLOCATION_SCOPE_OBJECT
);
49 return vk_error(VK_ERROR_OUT_OF_HOST_MEMORY
);
51 pipeline
->device
= device
;
52 pipeline
->layout
= anv_pipeline_layout_from_handle(pCreateInfo
->layout
);
54 pipeline
->blend_state
.map
= NULL
;
56 result
= anv_reloc_list_init(&pipeline
->batch_relocs
,
57 pAllocator
? pAllocator
: &device
->alloc
);
58 if (result
!= VK_SUCCESS
) {
59 vk_free2(&device
->alloc
, pAllocator
, pipeline
);
62 pipeline
->batch
.next
= pipeline
->batch
.start
= pipeline
->batch_data
;
63 pipeline
->batch
.end
= pipeline
->batch
.start
+ sizeof(pipeline
->batch_data
);
64 pipeline
->batch
.relocs
= &pipeline
->batch_relocs
;
66 /* When we free the pipeline, we detect stages based on the NULL status
67 * of various prog_data pointers. Make them NULL by default.
69 memset(pipeline
->shaders
, 0, sizeof(pipeline
->shaders
));
71 pipeline
->active_stages
= 0;
73 pipeline
->needs_data_cache
= false;
75 assert(pCreateInfo
->stage
.stage
== VK_SHADER_STAGE_COMPUTE_BIT
);
76 ANV_FROM_HANDLE(anv_shader_module
, module
, pCreateInfo
->stage
.module
);
77 result
= anv_pipeline_compile_cs(pipeline
, cache
, pCreateInfo
, module
,
78 pCreateInfo
->stage
.pName
,
79 pCreateInfo
->stage
.pSpecializationInfo
);
80 if (result
!= VK_SUCCESS
) {
81 vk_free2(&device
->alloc
, pAllocator
, pipeline
);
85 const struct brw_cs_prog_data
*cs_prog_data
= get_cs_prog_data(pipeline
);
87 anv_pipeline_setup_l3_config(pipeline
, cs_prog_data
->base
.total_shared
> 0);
89 uint32_t group_size
= cs_prog_data
->local_size
[0] *
90 cs_prog_data
->local_size
[1] * cs_prog_data
->local_size
[2];
91 uint32_t remainder
= group_size
& (cs_prog_data
->simd_size
- 1);
94 pipeline
->cs_right_mask
= ~0u >> (32 - remainder
);
96 pipeline
->cs_right_mask
= ~0u >> (32 - cs_prog_data
->simd_size
);
98 const uint32_t vfe_curbe_allocation
=
99 ALIGN(cs_prog_data
->push
.per_thread
.regs
* cs_prog_data
->threads
+
100 cs_prog_data
->push
.cross_thread
.regs
, 2);
102 const uint32_t subslices
= MAX2(physical_device
->subslice_total
, 1);
104 anv_batch_emit(&pipeline
->batch
, GENX(MEDIA_VFE_STATE
), vfe
) {
105 vfe
.ScratchSpaceBasePointer
= (struct anv_address
) {
106 .bo
= anv_scratch_pool_alloc(device
, &device
->scratch_pool
,
108 cs_prog_data
->base
.total_scratch
),
111 vfe
.PerThreadScratchSpace
= ffs(cs_prog_data
->base
.total_scratch
/ 2048);
115 vfe
.GPGPUMode
= true;
117 vfe
.MaximumNumberofThreads
=
118 devinfo
->max_cs_threads
* subslices
- 1;
119 vfe
.NumberofURBEntries
= GEN_GEN
<= 7 ? 0 : 2;
120 vfe
.ResetGatewayTimer
= true;
122 vfe
.BypassGatewayControl
= true;
124 vfe
.URBEntryAllocationSize
= GEN_GEN
<= 7 ? 0 : 2;
125 vfe
.CURBEAllocationSize
= vfe_curbe_allocation
;
128 *pPipeline
= anv_pipeline_to_handle(pipeline
);
133 VkResult
genX(CreateGraphicsPipelines
)(
135 VkPipelineCache pipelineCache
,
137 const VkGraphicsPipelineCreateInfo
* pCreateInfos
,
138 const VkAllocationCallbacks
* pAllocator
,
139 VkPipeline
* pPipelines
)
141 ANV_FROM_HANDLE(anv_pipeline_cache
, pipeline_cache
, pipelineCache
);
143 VkResult result
= VK_SUCCESS
;
146 for (; i
< count
; i
++) {
147 result
= genX(graphics_pipeline_create
)(_device
,
150 pAllocator
, &pPipelines
[i
]);
151 if (result
!= VK_SUCCESS
) {
152 for (unsigned j
= 0; j
< i
; j
++) {
153 anv_DestroyPipeline(_device
, pPipelines
[j
], pAllocator
);
163 VkResult
genX(CreateComputePipelines
)(
165 VkPipelineCache pipelineCache
,
167 const VkComputePipelineCreateInfo
* pCreateInfos
,
168 const VkAllocationCallbacks
* pAllocator
,
169 VkPipeline
* pPipelines
)
171 ANV_FROM_HANDLE(anv_pipeline_cache
, pipeline_cache
, pipelineCache
);
173 VkResult result
= VK_SUCCESS
;
176 for (; i
< count
; i
++) {
177 result
= compute_pipeline_create(_device
, pipeline_cache
,
179 pAllocator
, &pPipelines
[i
]);
180 if (result
!= VK_SUCCESS
) {
181 for (unsigned j
= 0; j
< i
; j
++) {
182 anv_DestroyPipeline(_device
, pPipelines
[j
], pAllocator
);