if (pipeline->gs_copy_shader) {
pipeline->gs_copy_shader =
- radv_pipeline_cache_insert_shader(cache,
+ radv_pipeline_cache_insert_shader(pipeline->device,
+ cache,
gs_copy_sha1,
pipeline->gs_copy_shader,
gs_copy_code,
ralloc_free(nir);
if (variant)
- variant = radv_pipeline_cache_insert_shader(cache, sha1, variant,
- code, code_size);
+ variant = radv_pipeline_cache_insert_shader(pipeline->device,
+ cache, sha1,
+ variant, code,
+ code_size);
if (code)
free(code);
ralloc_free(tcs_nir);
if (tes_variant)
- tes_variant = radv_pipeline_cache_insert_shader(cache, tes_sha1, tes_variant,
+ tes_variant = radv_pipeline_cache_insert_shader(pipeline->device, cache, tes_sha1, tes_variant,
tes_code, tes_code_size);
if (tcs_variant)
- tcs_variant = radv_pipeline_cache_insert_shader(cache, tcs_sha1, tcs_variant,
+ tcs_variant = radv_pipeline_cache_insert_shader(pipeline->device, cache, tcs_sha1, tcs_variant,
tcs_code, tcs_code_size);
if (tes_code)
if (cache)
entry = radv_pipeline_cache_search(cache, sha1);
+ else
+ entry = radv_pipeline_cache_search(device->mem_cache, sha1);
if (!entry)
return NULL;
}
struct radv_shader_variant *
-radv_pipeline_cache_insert_shader(struct radv_pipeline_cache *cache,
+radv_pipeline_cache_insert_shader(struct radv_device *device,
+ struct radv_pipeline_cache *cache,
const unsigned char *sha1,
struct radv_shader_variant *variant,
const void *code, unsigned code_size)
{
if (!cache)
- return variant;
+ cache = device->mem_cache;
pthread_mutex_lock(&cache->mutex);
struct cache_entry *entry = radv_pipeline_cache_search_unlocked(cache, sha1);
const unsigned char *sha1);
struct radv_shader_variant *
-radv_pipeline_cache_insert_shader(struct radv_pipeline_cache *cache,
+radv_pipeline_cache_insert_shader(struct radv_device *device,
+ struct radv_pipeline_cache *cache,
const unsigned char *sha1,
struct radv_shader_variant *variant,
const void *code, unsigned code_size);