#endif
struct gen_device_info;
+struct brw_context;
struct brw_bo {
/**
* The size may be larger than the size originally requested for the
* allocation, such as being aligned to page size.
*/
- unsigned long size;
+ uint64_t size;
/**
* Alignment requirement for object
*
* Used for GTT mapping & pinning the object.
*/
- unsigned long align;
-
- /**
- * Virtual address for accessing the buffer data. Only valid while
- * mapped.
- */
-#ifdef __cplusplus
- void *virt;
-#else
- void *virtual;
-#endif
+ uint64_t align;
/** Buffer manager context associated with this buffer object */
struct brw_bufmgr *bufmgr;
int refcount;
const char *name;
+#ifndef EXEC_OBJECT_CAPTURE
+#define EXEC_OBJECT_CAPTURE (1<<7)
+#endif
+ uint64_t kflags;
+
/**
* Kenel-assigned global name for this object
*
*/
uint32_t tiling_mode;
uint32_t swizzle_mode;
- unsigned long stride;
+ uint32_t stride;
time_t free_time;
/** Mapped address for the buffer, saved across map/unmap cycles */
- void *mem_virtual;
+ void *map_cpu;
/** GTT virtual address for the buffer, saved across map/unmap cycles */
- void *gtt_virtual;
+ void *map_gtt;
/** WC CPU address for the buffer, saved across map/unmap cycles */
- void *wc_virtual;
+ void *map_wc;
int map_count;
- struct list_head vma_list;
/** BO cache list */
struct list_head head;
* Boolean of whether this buffer can be re-used
*/
bool reusable;
+
+ /**
+ * Boolean of whether this buffer is cache coherent
+ */
+ bool cache_coherent;
};
#define BO_ALLOC_FOR_RENDER (1<<0)
*
* Buffer objects are not necessarily initially mapped into CPU virtual
* address space or graphics device aperture. They must be mapped
- * using bo_map() or brw_bo_map_gtt() to be used by the CPU.
+ * using brw_bo_map() to be used by the CPU.
*/
struct brw_bo *brw_bo_alloc(struct brw_bufmgr *bufmgr, const char *name,
- unsigned long size, unsigned int alignment);
-/**
- * Allocate a buffer object, hinting that it will be used as a
- * render target.
- *
- * This is otherwise the same as bo_alloc.
- */
-struct brw_bo *brw_bo_alloc_for_render(struct brw_bufmgr *bufmgr,
- const char *name,
- unsigned long size,
- unsigned int alignment);
+ uint64_t size, uint64_t alignment);
/**
* Allocate a tiled buffer object.
struct brw_bo *brw_bo_alloc_tiled(struct brw_bufmgr *bufmgr,
const char *name,
int x, int y, int cpp,
- uint32_t *tiling_mode,
- unsigned long *pitch,
- unsigned long flags);
+ uint32_t tiling_mode,
+ uint32_t *pitch,
+ unsigned flags);
/** Takes a reference on a buffer object */
void brw_bo_reference(struct brw_bo *bo);
*/
void brw_bo_unreference(struct brw_bo *bo);
+/* Must match MapBufferRange interface (for convenience) */
+#define MAP_READ GL_MAP_READ_BIT
+#define MAP_WRITE GL_MAP_WRITE_BIT
+#define MAP_ASYNC GL_MAP_UNSYNCHRONIZED_BIT
+#define MAP_PERSISTENT GL_MAP_PERSISTENT_BIT
+#define MAP_COHERENT GL_MAP_COHERENT_BIT
+/* internal */
+#define MAP_INTERNAL_MASK (0xff << 24)
+#define MAP_RAW (0x01 << 24)
+
/**
* Maps the buffer into userspace.
*
* This function will block waiting for any existing execution on the
- * buffer to complete, first. The resulting mapping is available at
- * buf->virtual.
+ * buffer to complete, first. The resulting mapping is returned.
*/
-int brw_bo_map(struct brw_bo *bo, int write_enable);
+MUST_CHECK void *brw_bo_map(struct brw_context *brw, struct brw_bo *bo, unsigned flags);
/**
* Reduces the refcount on the userspace mapping of the buffer
int brw_bo_unmap(struct brw_bo *bo);
/** Write data into an object. */
-int brw_bo_subdata(struct brw_bo *bo, unsigned long offset,
- unsigned long size, const void *data);
+int brw_bo_subdata(struct brw_bo *bo, uint64_t offset,
+ uint64_t size, const void *data);
/** Read data from an object. */
-int brw_bo_get_subdata(struct brw_bo *bo, unsigned long offset,
- unsigned long size, void *data);
+int brw_bo_get_subdata(struct brw_bo *bo, uint64_t offset,
+ uint64_t size, void *data);
/**
* Waits for rendering to an object by the GPU to have completed.
*
* bo_subdata, etc. It is merely a way for the driver to implement
* glFinish.
*/
-void brw_bo_wait_rendering(struct brw_bo *bo);
+void brw_bo_wait_rendering(struct brw_context *brw, struct brw_bo *bo);
/**
* Tears down the buffer manager instance.
*/
void brw_bufmgr_destroy(struct brw_bufmgr *bufmgr);
-/**
- * Ask that the buffer be placed in tiling mode
- *
- * \param buf Buffer to set tiling mode for
- * \param tiling_mode desired, and returned tiling mode
- */
-int brw_bo_set_tiling(struct brw_bo *bo, uint32_t *tiling_mode,
- uint32_t stride);
/**
* Get the current tiling (and resulting swizzling) mode for the bo.
*
*/
int brw_bo_madvise(struct brw_bo *bo, int madv);
-/**
- * Disable buffer reuse for buffers which will be shared in some way,
- * as with scanout buffers. When the buffer reference count goes to
- * zero, it will be freed and not placed in the reuse list.
- *
- * \param bo Buffer to disable reuse for
- */
-int brw_bo_disable_reuse(struct brw_bo *bo);
-
-/**
- * Query whether a buffer is reusable.
- *
- * \param bo Buffer to query
- */
-int brw_bo_is_reusable(struct brw_bo *bo);
-
/* drm_bacon_bufmgr_gem.c */
struct brw_bufmgr *brw_bufmgr_init(struct gen_device_info *devinfo,
int fd, int batch_size);
const char *name,
unsigned int handle);
void brw_bufmgr_enable_reuse(struct brw_bufmgr *bufmgr);
-void brw_bufmgr_gem_set_vma_cache_size(struct brw_bufmgr *bufmgr,
- int limit);
-int brw_bo_map_unsynchronized(struct brw_bo *bo);
-int brw_bo_map_gtt(struct brw_bo *bo);
-
-void *brw_bo_map__cpu(struct brw_bo *bo);
-void *brw_bo_map__gtt(struct brw_bo *bo);
-void *brw_bo_map__wc(struct brw_bo *bo);
-
-void brw_bo_start_gtt_access(struct brw_bo *bo, int write_enable);
int brw_bo_wait(struct brw_bo *bo, int64_t timeout_ns);