#include "pan_resource.h"
#include "pan_public.h"
#include "pan_util.h"
-#include "pandecode/decode.h"
+#include "decode.h"
#include "pan_context.h"
#include "midgard/midgard_compile.h"
+#include "bifrost/bifrost_compile.h"
#include "panfrost-quirks.h"
static const struct debug_named_value debug_options[] = {
{"msgs", PAN_DBG_MSGS, "Print debug messages"},
{"trace", PAN_DBG_TRACE, "Trace the command stream"},
{"deqp", PAN_DBG_DEQP, "Hacks for dEQP"},
- {"afbc", PAN_DBG_AFBC, "Enable non-conformant AFBC impl"},
+ {"afbc", PAN_DBG_AFBC, "Enable AFBC buffer sharing"},
{"sync", PAN_DBG_SYNC, "Wait for each job's completion and check for any GPU fault"},
{"precompile", PAN_DBG_PRECOMPILE, "Precompile shaders for shader-db"},
- {"gles3", PAN_DBG_GLES3, "Enable experimental GLES3 implementation"},
+ {"nofp16", PAN_DBG_NOFP16, "Disable 16-bit support"},
+ {"bifrost", PAN_DBG_BIFROST, "Enable experimental Mali G31 and G52 support"},
+ {"gl3", PAN_DBG_GL3, "Enable experimental GL 3.x implementation, up to 3.3"},
DEBUG_NAMED_VALUE_END
};
-DEBUG_GET_ONCE_FLAGS_OPTION(pan_debug, "PAN_MESA_DEBUG", debug_options, 0)
-
-int pan_debug = 0;
-
static const char *
panfrost_get_name(struct pipe_screen *screen)
{
- return panfrost_model_name(pan_screen(screen)->gpu_id);
+ return panfrost_model_name(pan_device(screen)->gpu_id);
}
static const char *
panfrost_get_param(struct pipe_screen *screen, enum pipe_cap param)
{
/* We expose in-dev stuff for dEQP that we don't want apps to use yet */
- bool is_deqp = pan_debug & PAN_DBG_DEQP;
+ struct panfrost_device *dev = pan_device(screen);
+ bool is_deqp = dev->debug & PAN_DBG_DEQP;
+
+ /* Our GL 3.x implementation is WIP */
+ bool is_gl3 = dev->debug & PAN_DBG_GL3;
+ is_gl3 |= is_deqp;
- /* Our GLES3 implementation is WIP */
- bool is_gles3 = pan_debug & PAN_DBG_GLES3;
- is_gles3 |= is_deqp;
+ /* Don't expose MRT related CAPs on GPUs that don't implement them */
+ bool has_mrt = !(dev->quirks & MIDGARD_SFBD);
+
+ /* Bifrost is WIP. No MRT support yet. */
+ bool is_bifrost = (dev->quirks & IS_BIFROST);
+ has_mrt &= !is_bifrost;
switch (param) {
case PIPE_CAP_NPOT_TEXTURES:
- case PIPE_CAP_MIXED_FRAMEBUFFER_SIZES:
case PIPE_CAP_MIXED_COLOR_DEPTH_BITS:
case PIPE_CAP_FRAGMENT_SHADER_TEXTURE_LOD:
case PIPE_CAP_VERTEX_SHADER_SATURATE:
+ case PIPE_CAP_VERTEX_COLOR_UNCLAMPED:
case PIPE_CAP_POINT_SPRITE:
+ case PIPE_CAP_DEPTH_CLIP_DISABLE:
+ case PIPE_CAP_MIXED_COLORBUFFER_FORMATS:
+ case PIPE_CAP_MIXED_FRAMEBUFFER_SIZES:
+ case PIPE_CAP_DEPTH_CLIP_DISABLE_SEPARATE:
return 1;
case PIPE_CAP_MAX_RENDER_TARGETS:
- return is_gles3 ? 4 : 1;
+ case PIPE_CAP_FBFETCH:
+ case PIPE_CAP_FBFETCH_COHERENT:
+ return has_mrt ? 8 : 1;
+
+ case PIPE_CAP_MAX_DUAL_SOURCE_RENDER_TARGETS:
+ return 1;
+
+ case PIPE_CAP_SAMPLE_SHADING:
+ /* WIP */
+ return is_gl3 ? 1 : 0;
- /* Throttling frames breaks pipelining */
- case PIPE_CAP_THROTTLE:
- return 0;
+ /* ES3 features unsupported on Bifrost */
case PIPE_CAP_OCCLUSION_QUERY:
- return 1;
- case PIPE_CAP_QUERY_TIME_ELAPSED:
- case PIPE_CAP_QUERY_PIPELINE_STATISTICS:
- case PIPE_CAP_QUERY_TIMESTAMP:
- case PIPE_CAP_QUERY_SO_OVERFLOW:
- return 0;
+ case PIPE_CAP_TGSI_INSTANCEID:
+ case PIPE_CAP_TEXTURE_MULTISAMPLE:
+ case PIPE_CAP_SURFACE_SAMPLE_COUNT:
+ case PIPE_CAP_PRIMITIVE_RESTART:
+ case PIPE_CAP_PRIMITIVE_RESTART_FIXED_INDEX:
+ return !is_bifrost;
+ case PIPE_CAP_SAMPLER_VIEW_TARGET:
case PIPE_CAP_TEXTURE_SWIZZLE:
- return 1;
-
case PIPE_CAP_TEXTURE_MIRROR_CLAMP:
case PIPE_CAP_TEXTURE_MIRROR_CLAMP_TO_EDGE:
- return 1;
-
- case PIPE_CAP_TGSI_INSTANCEID:
case PIPE_CAP_VERTEX_ELEMENT_INSTANCE_DIVISOR:
- case PIPE_CAP_PRIMITIVE_RESTART:
+ case PIPE_CAP_BLEND_EQUATION_SEPARATE:
+ case PIPE_CAP_INDEP_BLEND_ENABLE:
+ case PIPE_CAP_INDEP_BLEND_FUNC:
+ case PIPE_CAP_GENERATE_MIPMAP:
+ case PIPE_CAP_ACCELERATED:
+ case PIPE_CAP_UMA:
+ case PIPE_CAP_TEXTURE_FLOAT_LINEAR:
+ case PIPE_CAP_TEXTURE_HALF_FLOAT_LINEAR:
+ case PIPE_CAP_COPY_BETWEEN_COMPRESSED_AND_PLAIN_FORMATS:
+ case PIPE_CAP_TGSI_ARRAY_COMPONENTS:
return 1;
case PIPE_CAP_MAX_STREAM_OUTPUT_BUFFERS:
- return is_gles3 ? 4 : 0;
+ return is_bifrost ? 0 : 4;
case PIPE_CAP_MAX_STREAM_OUTPUT_SEPARATE_COMPONENTS:
case PIPE_CAP_MAX_STREAM_OUTPUT_INTERLEAVED_COMPONENTS:
- return is_gles3 ? 64 : 0;
+ return is_bifrost ? 0 : 64;
case PIPE_CAP_STREAM_OUTPUT_INTERLEAVE_BUFFERS:
- return 1;
+ return is_bifrost ? 0 : 1;
case PIPE_CAP_MAX_TEXTURE_ARRAY_LAYERS:
- return 256;
+ return is_bifrost ? 0 : 256;
case PIPE_CAP_GLSL_FEATURE_LEVEL:
case PIPE_CAP_GLSL_FEATURE_LEVEL_COMPATIBILITY:
- return is_gles3 ? 140 : 120;
+ return is_gl3 ? 330 : (is_bifrost ? 120 : 140);
case PIPE_CAP_ESSL_FEATURE_LEVEL:
- return is_gles3 ? 300 : 120;
+ return is_bifrost ? 120 : 300;
case PIPE_CAP_CONSTANT_BUFFER_OFFSET_ALIGNMENT:
return 16;
- return is_deqp;
-
- case PIPE_CAP_TEXTURE_MULTISAMPLE:
- return is_gles3;
-
/* For faking GLES 3.1 for dEQP-GLES31 */
case PIPE_CAP_MAX_COMBINED_HW_ATOMIC_COUNTERS:
case PIPE_CAP_MAX_COMBINED_HW_ATOMIC_COUNTER_BUFFERS:
case PIPE_CAP_IMAGE_LOAD_FORMATTED:
case PIPE_CAP_CUBE_MAP_ARRAY:
- return is_deqp;
-
- /* For faking compute shaders */
case PIPE_CAP_COMPUTE:
return is_deqp;
+ case PIPE_CAP_MAX_TEXTURE_BUFFER_SIZE:
+ return is_deqp ? 65536 : 0;
+
+ case PIPE_CAP_TEXTURE_BUFFER_OBJECTS:
+ case PIPE_CAP_QUERY_TIMESTAMP:
+ case PIPE_CAP_CONDITIONAL_RENDER:
+ return is_gl3;
/* TODO: Where does this req come from in practice? */
case PIPE_CAP_VERTEX_BUFFER_STRIDE_4BYTE_ALIGNED_ONLY:
case PIPE_CAP_MAX_TEXTURE_2D_SIZE:
return 4096;
case PIPE_CAP_MAX_TEXTURE_3D_LEVELS:
+ return is_bifrost ? 0 : 13;
case PIPE_CAP_MAX_TEXTURE_CUBE_LEVELS:
return 13;
- case PIPE_CAP_BLEND_EQUATION_SEPARATE:
- case PIPE_CAP_INDEP_BLEND_ENABLE:
- case PIPE_CAP_INDEP_BLEND_FUNC:
- return 1;
-
case PIPE_CAP_TGSI_FS_COORD_ORIGIN_LOWER_LEFT:
/* Hardware is natively upper left */
return 0;
case PIPE_CAP_TGSI_FS_COORD_ORIGIN_UPPER_LEFT:
case PIPE_CAP_TGSI_FS_COORD_PIXEL_CENTER_HALF_INTEGER:
case PIPE_CAP_TGSI_FS_COORD_PIXEL_CENTER_INTEGER:
- case PIPE_CAP_GENERATE_MIPMAP:
+ case PIPE_CAP_TGSI_TEXCOORD:
return 1;
- /* We would prefer varyings */
+ /* We would prefer varyings on Midgard, but proper sysvals on Bifrost */
case PIPE_CAP_TGSI_FS_FACE_IS_INTEGER_SYSVAL:
case PIPE_CAP_TGSI_FS_POSITION_IS_SYSVAL:
- return 0;
-
- /* I really don't want to set this CAP but let's not swim against the
- * tide.. */
- case PIPE_CAP_TGSI_TEXCOORD:
- return 1;
+ case PIPE_CAP_TGSI_FS_POINT_IS_SYSVAL:
+ return is_bifrost;
case PIPE_CAP_SEAMLESS_CUBE_MAP:
case PIPE_CAP_SEAMLESS_CUBE_MAP_PER_TEXTURE:
- return 1;
+ return !is_bifrost;
case PIPE_CAP_MAX_VERTEX_ELEMENT_SRC_OFFSET:
return 0xffff;
- case PIPE_CAP_TEXTURE_BUFFER_OBJECTS:
- return 1;
-
- case PIPE_CAP_MAX_TEXTURE_BUFFER_SIZE:
- return 65536;
-
case PIPE_CAP_PREFER_BLIT_BASED_TEXTURE_TRANSFER:
return 0;
case PIPE_CAP_ENDIANNESS:
return PIPE_ENDIAN_NATIVE;
- case PIPE_CAP_SAMPLER_VIEW_TARGET:
return 1;
case PIPE_CAP_MIN_TEXTURE_GATHER_OFFSET:
case PIPE_CAP_MAX_TEXTURE_GATHER_OFFSET:
return 7;
- case PIPE_CAP_VENDOR_ID:
- case PIPE_CAP_DEVICE_ID:
- return 0xFFFFFFFF;
-
- case PIPE_CAP_ACCELERATED:
- case PIPE_CAP_UMA:
- case PIPE_CAP_TEXTURE_FLOAT_LINEAR:
- case PIPE_CAP_TEXTURE_HALF_FLOAT_LINEAR:
- case PIPE_CAP_COPY_BETWEEN_COMPRESSED_AND_PLAIN_FORMATS:
- case PIPE_CAP_TGSI_ARRAY_COMPONENTS:
- return 1;
-
case PIPE_CAP_VIDEO_MEMORY: {
uint64_t system_memory;
}
case PIPE_CAP_SHADER_STENCIL_EXPORT:
- return 1;
+ return !is_bifrost;
case PIPE_CAP_SHADER_BUFFER_OFFSET_ALIGNMENT:
return 4;
case PIPE_CAP_CLIP_PLANES:
return 0;
+ case PIPE_CAP_PACKED_STREAM_OUTPUT:
+ return 0;
+
+ case PIPE_CAP_VIEWPORT_TRANSFORM_LOWERED:
+ case PIPE_CAP_PSIZ_CLAMPED:
+ return 1;
+
default:
return u_pipe_screen_get_param_defaults(screen, param);
}
enum pipe_shader_type shader,
enum pipe_shader_cap param)
{
- bool is_deqp = pan_debug & PAN_DBG_DEQP;
+ struct panfrost_device *dev = pan_device(screen);
+ bool is_deqp = dev->debug & PAN_DBG_DEQP;
+ bool is_nofp16 = dev->debug & PAN_DBG_NOFP16;
+ bool is_bifrost = dev->quirks & IS_BIFROST;
if (shader != PIPE_SHADER_VERTEX &&
shader != PIPE_SHADER_FRAGMENT &&
return 16;
case PIPE_SHADER_CAP_MAX_OUTPUTS:
- return shader == PIPE_SHADER_FRAGMENT ? 4 : 16;
+ return shader == PIPE_SHADER_FRAGMENT ? 8 : 16;
case PIPE_SHADER_CAP_MAX_TEMPS:
return 256; /* GL_MAX_PROGRAM_TEMPORARIES_ARB */
return 0;
case PIPE_SHADER_CAP_INDIRECT_INPUT_ADDR:
- return 1;
+ return is_bifrost ? 0 : 1;
case PIPE_SHADER_CAP_INDIRECT_OUTPUT_ADDR:
return 0;
return 0;
case PIPE_SHADER_CAP_INDIRECT_CONST_ADDR:
- return 1;
+ return is_bifrost ? 0 : 1;
case PIPE_SHADER_CAP_SUBROUTINES:
return 0;
case PIPE_SHADER_CAP_INTEGERS:
return 1;
- case PIPE_SHADER_CAP_INT64_ATOMICS:
case PIPE_SHADER_CAP_FP16:
+ case PIPE_SHADER_CAP_GLSL_16BIT_CONSTS:
+ return !is_nofp16;
+
+ case PIPE_SHADER_CAP_FP16_DERIVATIVES:
+ case PIPE_SHADER_CAP_INT16:
+ case PIPE_SHADER_CAP_INT64_ATOMICS:
case PIPE_SHADER_CAP_TGSI_DROUND_SUPPORTED:
case PIPE_SHADER_CAP_TGSI_DFRACEXP_DLDEXP_SUPPORTED:
case PIPE_SHADER_CAP_TGSI_LDEXP_SUPPORTED:
return 0;
default:
- DBG("unknown shader param %d\n", param);
+ /* Other params are unknown */
return 0;
}
unsigned storage_sample_count,
unsigned bind)
{
+ struct panfrost_device *dev = pan_device(screen);
const struct util_format_description *format_desc;
assert(target == PIPE_BUFFER ||
return false;
/* MSAA 4x supported, but no more. Technically some revisions of the
- * hardware can go up to 16x but we don't support higher modes yet. */
+ * hardware can go up to 16x but we don't support higher modes yet.
+ * MSAA 2x is notably not supported and gets rounded up to MSAA 4x. */
- if (sample_count > 1 && !(pan_debug & PAN_DBG_DEQP))
- return false;
-
- if (sample_count > 4)
+ if (!(sample_count == 0 || sample_count == 1 || sample_count == 4))
return false;
if (MAX2(sample_count, 1) != MAX2(storage_sample_count, 1))
return false;
- /* Format wishlist */
- if (format == PIPE_FORMAT_X8Z24_UNORM)
- return false;
-
- if (format == PIPE_FORMAT_A1B5G5R5_UNORM || format == PIPE_FORMAT_X1B5G5R5_UNORM)
- return false;
-
- /* TODO */
- if (format == PIPE_FORMAT_B5G5R5A1_UNORM)
- return FALSE;
-
/* Don't confuse poorly written apps (workaround dEQP bug) that expect
* more alpha than they ask for */
if (scanout && renderable && !util_format_is_rgba8_variant(format_desc))
return false;
- switch (format_desc->layout) {
- case UTIL_FORMAT_LAYOUT_PLAIN:
- case UTIL_FORMAT_LAYOUT_OTHER:
- break;
- case UTIL_FORMAT_LAYOUT_ETC:
- case UTIL_FORMAT_LAYOUT_ASTC:
- return true;
- default:
- return false;
- }
+ /* Check we support the format with the given bind */
+
+ unsigned relevant_bind = bind &
+ ( PIPE_BIND_DEPTH_STENCIL | PIPE_BIND_RENDER_TARGET
+ | PIPE_BIND_VERTEX_BUFFER | PIPE_BIND_SAMPLER_VIEW);
+
+ struct panfrost_format fmt = panfrost_pipe_format_table[format];
+
+ /* Also check that compressed texture formats are supported on this
+ * particular chip. They may not be depending on system integration
+ * differences. RGTC can be emulated so is always supported. */
+
+ bool is_rgtc = format_desc->layout == UTIL_FORMAT_LAYOUT_RGTC;
+ bool supported = panfrost_supports_compressed_format(dev, fmt.hw);
+
+ if (!is_rgtc && !supported)
+ return false;
+
+ return fmt.hw && ((relevant_bind & ~fmt.bind) == 0);
+}
+
+/* We always support linear and tiled operations, both external and internal.
+ * We support AFBC for a subset of formats, and colourspace transform for a
+ * subset of those. */
+
+static void
+panfrost_query_dmabuf_modifiers(struct pipe_screen *screen,
+ enum pipe_format format, int max, uint64_t *modifiers, unsigned
+ int *external_only, int *out_count)
+{
+ /* Query AFBC status */
+ bool afbc = panfrost_format_supports_afbc(format);
+ bool ytr = panfrost_afbc_can_ytr(format);
+
+ /* Don't advertise AFBC before T760 */
+ struct panfrost_device *dev = pan_device(screen);
+ afbc &= !(dev->quirks & MIDGARD_NO_AFBC);
+
+ /* XXX: AFBC scanout is broken on mainline RK3399 with older kernels */
+ afbc &= (dev->debug & PAN_DBG_AFBC);
+
+ unsigned count = 0;
+
+ for (unsigned i = 0; i < PAN_MODIFIER_COUNT; ++i) {
+ if (drm_is_afbc(pan_best_modifiers[i]) && !afbc)
+ continue;
+
+ if ((pan_best_modifiers[i] & AFBC_FORMAT_MOD_YTR) && !ytr)
+ continue;
+
+ count++;
+
+ if (max > (int) count) {
+ modifiers[count] = pan_best_modifiers[i];
- /* Internally, formats that are depth/stencil renderable are limited.
- *
- * In particular: Z16, Z24, Z24S8, S8 are all identical from the GPU
- * rendering perspective. That is, we render to Z24S8 (which we can
- * AFBC compress), ignore the different when texturing (who cares?),
- * and then in the off-chance there's a CPU read we blit back to
- * staging.
- *
- * ...alternatively, we can make the state tracker deal with that. */
-
- if (bind & PIPE_BIND_DEPTH_STENCIL) {
- switch (format) {
- case PIPE_FORMAT_Z24_UNORM_S8_UINT:
- case PIPE_FORMAT_Z24X8_UNORM:
- case PIPE_FORMAT_Z32_UNORM:
- case PIPE_FORMAT_Z32_FLOAT:
- case PIPE_FORMAT_Z32_FLOAT_S8X24_UINT:
- return true;
-
- default:
- return false;
+ if (external_only)
+ external_only[count] = false;
}
}
- return true;
+ *out_count = count;
}
static int
panfrost_get_compute_param(struct pipe_screen *pscreen, enum pipe_shader_ir ir_type,
enum pipe_compute_cap param, void *ret)
{
+ struct panfrost_device *dev = pan_device(pscreen);
const char * const ir = "panfrost";
- if (!(pan_debug & PAN_DBG_DEQP))
+ if (!(dev->debug & PAN_DBG_DEQP))
return 0;
#define RET(x) do { \
static void
panfrost_destroy_screen(struct pipe_screen *pscreen)
{
- struct panfrost_screen *screen = pan_screen(pscreen);
- panfrost_bo_cache_evict_all(screen);
- pthread_mutex_destroy(&screen->bo_cache.lock);
- pthread_mutex_destroy(&screen->active_bos_lock);
- drmFreeVersion(screen->kernel_version);
- ralloc_free(screen);
+ panfrost_close_device(pan_device(pscreen));
+ ralloc_free(pscreen);
}
static uint64_t
struct pipe_fence_handle **ptr,
struct pipe_fence_handle *fence)
{
+ struct panfrost_device *dev = pan_device(pscreen);
struct panfrost_fence **p = (struct panfrost_fence **)ptr;
struct panfrost_fence *f = (struct panfrost_fence *)fence;
struct panfrost_fence *old = *p;
if (pipe_reference(&(*p)->reference, &f->reference)) {
- util_dynarray_foreach(&old->syncfds, int, fd)
- close(*fd);
- util_dynarray_fini(&old->syncfds);
+ drmSyncobjDestroy(dev->fd, old->syncobj);
free(old);
}
*p = f;
struct pipe_fence_handle *fence,
uint64_t timeout)
{
- struct panfrost_screen *screen = pan_screen(pscreen);
+ struct panfrost_device *dev = pan_device(pscreen);
struct panfrost_fence *f = (struct panfrost_fence *)fence;
- struct util_dynarray syncobjs;
int ret;
- /* All fences were already signaled */
- if (!util_dynarray_num_elements(&f->syncfds, int))
+ if (f->signaled)
return true;
- util_dynarray_init(&syncobjs, NULL);
- util_dynarray_foreach(&f->syncfds, int, fd) {
- uint32_t syncobj;
-
- ret = drmSyncobjCreate(screen->fd, 0, &syncobj);
- assert(!ret);
-
- ret = drmSyncobjImportSyncFile(screen->fd, syncobj, *fd);
- assert(!ret);
- util_dynarray_append(&syncobjs, uint32_t, syncobj);
- }
-
uint64_t abs_timeout = os_time_get_absolute_timeout(timeout);
if (abs_timeout == OS_TIMEOUT_INFINITE)
abs_timeout = INT64_MAX;
- ret = drmSyncobjWait(screen->fd, util_dynarray_begin(&syncobjs),
- util_dynarray_num_elements(&syncobjs, uint32_t),
+ ret = drmSyncobjWait(dev->fd, &f->syncobj,
+ 1,
abs_timeout, DRM_SYNCOBJ_WAIT_FLAGS_WAIT_ALL,
NULL);
- util_dynarray_foreach(&syncobjs, uint32_t, syncobj)
- drmSyncobjDestroy(screen->fd, *syncobj);
-
- return ret >= 0;
+ f->signaled = (ret >= 0);
+ return f->signaled;
}
struct panfrost_fence *
panfrost_fence_create(struct panfrost_context *ctx,
- struct util_dynarray *fences)
+ uint32_t syncobj)
{
- struct panfrost_screen *screen = pan_screen(ctx->base.screen);
struct panfrost_fence *f = calloc(1, sizeof(*f));
if (!f)
return NULL;
- util_dynarray_init(&f->syncfds, NULL);
-
- /* Export fences from all pending batches. */
- util_dynarray_foreach(fences, struct panfrost_batch_fence *, fence) {
- int fd = -1;
-
- /* The fence is already signaled, no need to export it. */
- if ((*fence)->signaled)
- continue;
-
- drmSyncobjExportSyncFile(screen->fd, (*fence)->syncobj, &fd);
- if (fd == -1)
- fprintf(stderr, "export failed: %m\n");
-
- assert(fd != -1);
- util_dynarray_append(&f->syncfds, int, fd);
- }
-
pipe_reference_init(&f->reference, 1);
+ f->syncobj = syncobj;
return f;
}
enum pipe_shader_ir ir,
enum pipe_shader_type shader)
{
- return &midgard_nir_options;
-}
-
-static uint32_t
-panfrost_active_bos_hash(const void *key)
-{
- const struct panfrost_bo *bo = key;
-
- return _mesa_hash_data(&bo->gem_handle, sizeof(bo->gem_handle));
-}
-
-static bool
-panfrost_active_bos_cmp(const void *keya, const void *keyb)
-{
- const struct panfrost_bo *a = keya, *b = keyb;
-
- return a->gem_handle == b->gem_handle;
+ if (pan_device(pscreen)->quirks & IS_BIFROST)
+ return &bifrost_nir_options;
+ else
+ return &midgard_nir_options;
}
struct pipe_screen *
panfrost_create_screen(int fd, struct renderonly *ro)
{
- pan_debug = debug_get_option_pan_debug();
-
- /* Blacklist apps known to be buggy under Panfrost */
- const char *proc = util_get_process_name();
- const char *blacklist[] = {
- "chromium",
- "chrome",
- };
-
- for (unsigned i = 0; i < ARRAY_SIZE(blacklist); ++i) {
- if ((strcmp(blacklist[i], proc) == 0))
- return NULL;
- }
-
/* Create the screen */
struct panfrost_screen *screen = rzalloc(NULL, struct panfrost_screen);
if (!screen)
return NULL;
+ struct panfrost_device *dev = pan_device(&screen->base);
+ panfrost_open_device(screen, fd, dev);
+
+ dev->debug = debug_get_flags_option("PAN_MESA_DEBUG", debug_options, 0);
+
if (ro) {
- screen->ro = renderonly_dup(ro);
- if (!screen->ro) {
- DBG("Failed to dup renderonly object\n");
+ dev->ro = renderonly_dup(ro);
+ if (!dev->ro) {
+ if (dev->debug & PAN_DBG_MSGS)
+ fprintf(stderr, "Failed to dup renderonly object\n");
+
free(screen);
return NULL;
}
}
- screen->fd = fd;
-
- screen->gpu_id = panfrost_query_gpu_version(screen->fd);
- screen->core_count = panfrost_query_core_count(screen->fd);
- screen->thread_tls_alloc = panfrost_query_thread_tls_alloc(screen->fd);
- screen->quirks = panfrost_get_quirks(screen->gpu_id);
- screen->kernel_version = drmGetVersion(fd);
-
/* Check if we're loading against a supported GPU model. */
- switch (screen->gpu_id) {
+ switch (dev->gpu_id) {
case 0x720: /* T720 */
case 0x750: /* T760 */
case 0x820: /* T820 */
case 0x860: /* T860 */
break;
+ case 0x7093: /* G31 */
+ case 0x7212: /* G52 */
+ if (dev->debug & PAN_DBG_BIFROST)
+ break;
+
+ /* fallthrough */
default:
/* Fail to load against untested models */
- debug_printf("panfrost: Unsupported model %X", screen->gpu_id);
+ debug_printf("panfrost: Unsupported model %X", dev->gpu_id);
+ panfrost_destroy_screen(&(screen->base));
return NULL;
}
- pthread_mutex_init(&screen->active_bos_lock, NULL);
- screen->active_bos = _mesa_set_create(screen, panfrost_active_bos_hash,
- panfrost_active_bos_cmp);
-
- pthread_mutex_init(&screen->bo_cache.lock, NULL);
- list_inithead(&screen->bo_cache.lru);
- for (unsigned i = 0; i < ARRAY_SIZE(screen->bo_cache.buckets); ++i)
- list_inithead(&screen->bo_cache.buckets[i]);
-
- if (pan_debug & (PAN_DBG_TRACE | PAN_DBG_SYNC))
- pandecode_initialize(!(pan_debug & PAN_DBG_TRACE));
+ if (dev->debug & (PAN_DBG_TRACE | PAN_DBG_SYNC))
+ pandecode_initialize(!(dev->debug & PAN_DBG_TRACE));
screen->base.destroy = panfrost_destroy_screen;
screen->base.get_paramf = panfrost_get_paramf;
screen->base.get_timestamp = panfrost_get_timestamp;
screen->base.is_format_supported = panfrost_is_format_supported;
+ screen->base.query_dmabuf_modifiers = panfrost_query_dmabuf_modifiers;
screen->base.context_create = panfrost_create_context;
screen->base.get_compiler_options = panfrost_screen_get_compiler_options;
screen->base.fence_reference = panfrost_fence_reference;
screen->base.fence_finish = panfrost_fence_finish;
screen->base.set_damage_region = panfrost_resource_set_damage_region;
- panfrost_resource_screen_init(screen);
+ panfrost_resource_screen_init(&screen->base);
+
+ if (!(dev->quirks & IS_BIFROST))
+ panfrost_init_blit_shaders(dev);
return &screen->base;
}