{
VkResult result;
struct radv_shader_module vs = {0};
- zero(device->meta_state.blit);
VkDescriptorSetLayoutCreateInfo ds_layout_info = {
.sType = VK_STRUCTURE_TYPE_DESCRIPTOR_SET_LAYOUT_CREATE_INFO,
{
VkResult result;
- zero(device->meta_state.blit2d);
-
const VkPushConstantRange push_constant_ranges[] = {
{VK_SHADER_STAGE_VERTEX_BIT, 0, 16},
{VK_SHADER_STAGE_FRAGMENT_BIT, 16, 4},
struct radv_shader_module fill_cs = { .nir = NULL };
struct radv_shader_module copy_cs = { .nir = NULL };
- zero(device->meta_state.buffer);
-
fill_cs.nir = build_buffer_fill_shader(device);
copy_cs.nir = build_buffer_copy_shader(device);
VkResult result;
struct radv_shader_module cs = { .nir = NULL };
- zero(device->meta_state.itob);
-
cs.nir = build_nir_itob_compute_shader(device);
/*
VkResult result;
struct radv_shader_module cs = { .nir = NULL };
- zero(device->meta_state.btoi);
-
cs.nir = build_nir_btoi_compute_shader(device);
/*
VkResult result;
struct radv_shader_module cs = { .nir = NULL };
- zero(device->meta_state.itoi);
-
cs.nir = build_nir_itoi_compute_shader(device);
/*
VkResult result;
struct radv_shader_module cs = { .nir = NULL };
- zero(device->meta_state.cleari);
-
cs.nir = build_nir_cleari_compute_shader(device);
/*
VkResult res;
struct radv_meta_state *state = &device->meta_state;
- memset(&device->meta_state.clear, 0, sizeof(device->meta_state.clear));
-
VkPipelineLayoutCreateInfo pl_color_create_info = {
.sType = VK_STRUCTURE_TYPE_PIPELINE_LAYOUT_CREATE_INFO,
.setLayoutCount = 0,
struct radv_meta_state *state = &device->meta_state;
VkResult res = VK_SUCCESS;
- zero(state->depth_decomp);
-
struct radv_shader_module vs_module = { .nir = radv_meta_build_nir_vs_generate_vertices() };
if (!vs_module.nir) {
/* XXX: Need more accurate error */
{
VkResult res = VK_SUCCESS;
- zero(device->meta_state.fast_clear_flush);
-
struct radv_shader_module vs_module = { .nir = radv_meta_build_nir_vs_generate_vertices() };
if (!vs_module.nir) {
/* XXX: Need more accurate error */
{
VkResult res = VK_SUCCESS;
- zero(device->meta_state.resolve);
-
struct radv_shader_module vs_module = { .nir = radv_meta_build_nir_vs_generate_vertices() };
if (!vs_module.nir) {
/* XXX: Need more accurate error */
{
struct radv_meta_state *state = &device->meta_state;
VkResult res;
- memset(&device->meta_state.resolve_compute, 0, sizeof(device->meta_state.resolve_compute));
res = create_layout(device);
if (res != VK_SUCCESS)
VkResult
radv_device_init_meta_resolve_fragment_state(struct radv_device *device)
{
- struct radv_meta_state *state = &device->meta_state;
VkResult res;
- memset(&state->resolve_fragment, 0, sizeof(state->resolve_fragment));
res = create_layout(device);
if (res != VK_SUCCESS)
memcpy((dest), (src), (count) * sizeof(*(src))); \
})
-#define zero(x) (memset(&(x), 0, sizeof(x)))
-
/* Whenever we generate an error, pass it through this function. Useful for
* debugging, where we can break on it. Only call at error site, not when
* propagating errors. Might be useful to plug in a stack trace here.
struct radv_shader_module occlusion_cs = { .nir = NULL };
struct radv_shader_module pipeline_statistics_cs = { .nir = NULL };
- zero(device->meta_state.query);
-
occlusion_cs.nir = build_occlusion_query_shader(device);
pipeline_statistics_cs.nir = build_pipeline_statistics_query_shader(device);