uint32_t handle;
uint32_t size;
+ /* This will be read/written by multiple threads without a lock -- you
+ * should take a snapshot and use it to see if you happen to be in the
+ * CL's handles at this position, to make most lookups O(1). It's
+ * volatile to make sure that the compiler doesn't emit multiple loads
+ * from the address, which would make the lookup racy.
+ */
+ volatile uint32_t last_hindex;
+
/** Entry in the linked list of buffers freed, by age. */
struct list_head time_list;
/** Entry in the per-page-count linked list of buffers freed (by age). */
bool vc4_bo_flink(struct vc4_bo *bo, uint32_t *name);
int vc4_bo_get_dmabuf(struct vc4_bo *bo);
-static inline void
-vc4_bo_set_reference(struct vc4_bo **old_bo, struct vc4_bo *new_bo)
-{
- if (pipe_reference(&(*old_bo)->reference, &new_bo->reference))
- vc4_bo_last_unreference(*old_bo);
- *old_bo = new_bo;
-}
-
+void vc4_bo_debug_describe(char* buf, const struct vc4_bo *ptr);
static inline struct vc4_bo *
vc4_bo_reference(struct vc4_bo *bo)
{
- pipe_reference(NULL, &bo->reference);
+ pipe_reference_described(NULL, &bo->reference,
+ (debug_reference_descriptor)
+ vc4_bo_debug_describe);
return bo;
}
if ((*bo)->private) {
/* Avoid the mutex for private BOs */
- if (pipe_reference(&(*bo)->reference, NULL))
+ if (pipe_reference_described(&(*bo)->reference, NULL,
+ (debug_reference_descriptor)
+ vc4_bo_debug_describe)) {
vc4_bo_last_unreference(*bo);
+ }
} else {
screen = (*bo)->screen;
mtx_lock(&screen->bo_handles_mutex);
- if (pipe_reference(&(*bo)->reference, NULL)) {
+ if (pipe_reference_described(&(*bo)->reference, NULL,
+ (debug_reference_descriptor)
+ vc4_bo_debug_describe)) {
util_hash_table_remove(screen->bo_handles,
(void *)(uintptr_t)(*bo)->handle);
vc4_bo_last_unreference(*bo);
if (!*bo)
return;
- if (pipe_reference(&(*bo)->reference, NULL))
+ if (pipe_reference_described(&(*bo)->reference, NULL,
+ (debug_reference_descriptor)
+ vc4_bo_debug_describe)) {
vc4_bo_last_unreference_locked_timed(*bo, time);
+ }
*bo = NULL;
}