/**********************************************************
- * Copyright 2009 VMware, Inc. All rights reserved.
+ * Copyright 2009-2015 VMware, Inc. All rights reserved.
*
* Permission is hereby granted, free of charge, to any person
* obtaining a copy of this software and associated documentation
* @author Jose Fonseca
*/
+#include <libsync.h>
#include "svga_cmd.h"
#include "svga3d_caps.h"
#include "vmw_surface.h"
#include "vmw_buffer.h"
#include "vmw_fence.h"
+#include "vmw_shader.h"
+#include "vmw_query.h"
+#include "svga3d_surfacedefs.h"
+/**
+ * Try to get a surface backing buffer from the cache
+ * if it's this size or smaller.
+ */
+#define VMW_TRY_CACHED_SIZE (2*1024*1024)
static struct svga_winsys_buffer *
vmw_svga_winsys_buffer_create(struct svga_winsys_screen *sws,
unsigned size)
{
struct vmw_winsys_screen *vws = vmw_winsys_screen(sws);
- struct pb_desc desc;
+ struct vmw_buffer_desc desc;
struct pb_manager *provider;
struct pb_buffer *buffer;
memset(&desc, 0, sizeof desc);
- desc.alignment = alignment;
- desc.usage = usage;
+ desc.pb_desc.alignment = alignment;
+ desc.pb_desc.usage = usage;
if (usage == SVGA_BUFFER_USAGE_PINNED) {
if (vws->pools.query_fenced == NULL && !vmw_query_pools_init(vws))
return NULL;
provider = vws->pools.query_fenced;
+ } else if (usage == SVGA_BUFFER_USAGE_SHADER) {
+ provider = vws->pools.mob_shader_slab_fenced;
} else
provider = vws->pools.gmr_fenced;
assert(provider);
- buffer = provider->create_buffer(provider, size, &desc);
- if(!buffer)
- return NULL;
+ buffer = provider->create_buffer(provider, size, &desc.pb_desc);
- return vmw_svga_winsys_buffer(buffer);
-}
+ if(!buffer && provider == vws->pools.gmr_fenced) {
+ assert(provider);
+ provider = vws->pools.gmr_slab_fenced;
+ buffer = provider->create_buffer(provider, size, &desc.pb_desc);
+ }
-static void *
-vmw_svga_winsys_buffer_map(struct svga_winsys_screen *sws,
- struct svga_winsys_buffer *buf,
- unsigned flags)
-{
- (void)sws;
- return pb_map(vmw_pb_buffer(buf), flags, NULL);
-}
-
-
-static void
-vmw_svga_winsys_buffer_unmap(struct svga_winsys_screen *sws,
- struct svga_winsys_buffer *buf)
-{
- (void)sws;
- pb_unmap(vmw_pb_buffer(buf));
-}
-
+ if (!buffer)
+ return NULL;
-static void
-vmw_svga_winsys_buffer_destroy(struct svga_winsys_screen *sws,
- struct svga_winsys_buffer *buf)
-{
- struct pb_buffer *pbuf = vmw_pb_buffer(buf);
- (void)sws;
- pb_reference(&pbuf, NULL);
+ return vmw_svga_winsys_buffer_wrap(buffer);
}
static int
vmw_svga_winsys_fence_finish(struct svga_winsys_screen *sws,
struct pipe_fence_handle *fence,
+ uint64_t timeout,
unsigned flag)
{
struct vmw_winsys_screen *vws = vmw_winsys_screen(sws);
- return vmw_fence_finish(vws, fence, flag);
+ return vmw_fence_finish(vws, fence, timeout, flag);
+}
+
+
+static int
+vmw_svga_winsys_fence_get_fd(struct svga_winsys_screen *sws,
+ struct pipe_fence_handle *fence,
+ boolean duplicate)
+{
+ if (duplicate)
+ return dup(vmw_fence_get_fd(fence));
+ else
+ return vmw_fence_get_fd(fence);
}
+static void
+vmw_svga_winsys_fence_create_fd(struct svga_winsys_screen *sws,
+ struct pipe_fence_handle **fence,
+ int32_t fd)
+{
+ *fence = vmw_fence_create(NULL, 0, 0, 0, dup(fd));
+}
+
+static int
+vmw_svga_winsys_fence_server_sync(struct svga_winsys_screen *sws,
+ int32_t *context_fd,
+ struct pipe_fence_handle *fence)
+{
+ int32_t fd = sws->fence_get_fd(sws, fence, FALSE);
+
+ /* If we don't have fd, we don't need to merge fd into the context's fd. */
+ if (fd == -1)
+ return 0;
+
+ return sync_accumulate("vmwgfx", context_fd, fd);
+}
+
static struct svga_winsys_surface *
vmw_svga_winsys_surface_create(struct svga_winsys_screen *sws,
- SVGA3dSurfaceFlags flags,
+ SVGA3dSurfaceAllFlags flags,
SVGA3dSurfaceFormat format,
+ unsigned usage,
SVGA3dSize size,
- uint32 numFaces,
- uint32 numMipLevels)
+ uint32 numLayers,
+ uint32 numMipLevels,
+ unsigned sampleCount)
{
struct vmw_winsys_screen *vws = vmw_winsys_screen(sws);
struct vmw_svga_winsys_surface *surface;
+ struct vmw_buffer_desc desc;
+ struct pb_manager *provider;
+ uint32_t buffer_size;
+ uint32_t num_samples = 1;
+ SVGA3dMSPattern multisample_pattern = SVGA3D_MS_PATTERN_NONE;
+ memset(&desc, 0, sizeof(desc));
surface = CALLOC_STRUCT(vmw_svga_winsys_surface);
if(!surface)
goto no_surface;
pipe_reference_init(&surface->refcnt, 1);
p_atomic_set(&surface->validated, 0);
surface->screen = vws;
- surface->sid = vmw_ioctl_surface_create(vws,
- flags, format, size,
- numFaces, numMipLevels);
- if(surface->sid == SVGA3D_INVALID_ID)
+ (void) mtx_init(&surface->mutex, mtx_plain);
+ surface->shared = !!(usage & SVGA_SURFACE_USAGE_SHARED);
+ provider = (surface->shared) ? vws->pools.gmr : vws->pools.mob_fenced;
+
+ /*
+ * When multisampling is not supported sample count received is 0,
+ * otherwise should have a valid sample count.
+ */
+ if ((flags & SVGA3D_SURFACE_MULTISAMPLE) != 0) {
+ if (sampleCount == 0)
+ goto no_sid;
+ num_samples = sampleCount;
+ multisample_pattern = SVGA3D_MS_PATTERN_STANDARD;
+ }
+
+ /*
+ * Used for the backing buffer GB surfaces, and to approximate
+ * when to flush on non-GB hosts.
+ */
+ buffer_size = svga3dsurface_get_serialized_size_extended(format, size,
+ numMipLevels,
+ numLayers,
+ num_samples);
+ if (flags & SVGA3D_SURFACE_BIND_STREAM_OUTPUT)
+ buffer_size += sizeof(SVGA3dDXSOState);
+
+ if (buffer_size > vws->ioctl.max_texture_size) {
goto no_sid;
+ }
+
+ if (sws->have_gb_objects) {
+ SVGAGuestPtr ptr = {0,0};
+
+ /*
+ * If the backing buffer size is small enough, try to allocate a
+ * buffer out of the buffer cache. Otherwise, let the kernel allocate
+ * a suitable buffer for us.
+ */
+ if (buffer_size < VMW_TRY_CACHED_SIZE && !surface->shared) {
+ struct pb_buffer *pb_buf;
+
+ surface->size = buffer_size;
+ desc.pb_desc.alignment = 4096;
+ desc.pb_desc.usage = 0;
+ pb_buf = provider->create_buffer(provider, buffer_size, &desc.pb_desc);
+ surface->buf = vmw_svga_winsys_buffer_wrap(pb_buf);
+ if (surface->buf && !vmw_gmr_bufmgr_region_ptr(pb_buf, &ptr))
+ assert(0);
+ }
+
+ surface->sid = vmw_ioctl_gb_surface_create(vws, flags, format, usage,
+ size, numLayers,
+ numMipLevels, sampleCount,
+ ptr.gmrId,
+ multisample_pattern,
+ surface->buf ? NULL :
+ &desc.region);
+
+ if (surface->sid == SVGA3D_INVALID_ID) {
+ if (surface->buf == NULL) {
+ goto no_sid;
+ } else {
+ /*
+ * Kernel refused to allocate a surface for us.
+ * Perhaps something was wrong with our buffer?
+ * This is really a guard against future new size requirements
+ * on the backing buffers.
+ */
+ vmw_svga_winsys_buffer_destroy(sws, surface->buf);
+ surface->buf = NULL;
+ surface->sid = vmw_ioctl_gb_surface_create(vws, flags, format, usage,
+ size, numLayers,
+ numMipLevels, sampleCount,
+ 0, multisample_pattern,
+ &desc.region);
+ if (surface->sid == SVGA3D_INVALID_ID)
+ goto no_sid;
+ }
+ }
+
+ /*
+ * If the kernel created the buffer for us, wrap it into a
+ * vmw_svga_winsys_buffer.
+ */
+ if (surface->buf == NULL) {
+ struct pb_buffer *pb_buf;
+
+ surface->size = vmw_region_size(desc.region);
+ desc.pb_desc.alignment = 4096;
+ desc.pb_desc.usage = VMW_BUFFER_USAGE_SHARED;
+ pb_buf = provider->create_buffer(provider, surface->size,
+ &desc.pb_desc);
+ surface->buf = vmw_svga_winsys_buffer_wrap(pb_buf);
+ if (surface->buf == NULL) {
+ vmw_ioctl_region_destroy(desc.region);
+ vmw_ioctl_surface_destroy(vws, surface->sid);
+ goto no_sid;
+ }
+ }
+ } else {
+ /* Legacy surface only support 32-bit svga3d flags */
+ surface->sid = vmw_ioctl_surface_create(vws, (SVGA3dSurface1Flags)flags,
+ format, usage, size, numLayers,
+ numMipLevels, sampleCount);
+ if(surface->sid == SVGA3D_INVALID_ID)
+ goto no_sid;
+
+ /* Best estimate for surface size, used for early flushing. */
+ surface->size = buffer_size;
+ surface->buf = NULL;
+ }
return svga_winsys_surface(surface);
no_sid:
+ if (surface->buf)
+ vmw_svga_winsys_buffer_destroy(sws, surface->buf);
+
FREE(surface);
no_surface:
return NULL;
}
+static boolean
+vmw_svga_winsys_surface_can_create(struct svga_winsys_screen *sws,
+ SVGA3dSurfaceFormat format,
+ SVGA3dSize size,
+ uint32 numLayers,
+ uint32 numMipLevels,
+ uint32 numSamples)
+{
+ struct vmw_winsys_screen *vws = vmw_winsys_screen(sws);
+ uint32_t buffer_size;
+
+ buffer_size = svga3dsurface_get_serialized_size(format, size,
+ numMipLevels,
+ numLayers);
+ if (numSamples > 1)
+ buffer_size *= numSamples;
+
+ if (buffer_size > vws->ioctl.max_texture_size) {
+ return FALSE;
+ }
+ return TRUE;
+}
+
static boolean
vmw_svga_winsys_surface_is_flushed(struct svga_winsys_screen *sws,
{
struct vmw_winsys_screen *vws = vmw_winsys_screen(sws);
+ if (sws->have_gb_objects)
+ return SVGA3D_HWVERSION_WS8_B1;
+
return (SVGA3dHardwareVersion) vws->ioctl.hwversion;
}
vmw_svga_winsys_get_cap(struct svga_winsys_screen *sws,
SVGA3dDevCapIndex index,
SVGA3dDevCapResult *result)
-{
+{
struct vmw_winsys_screen *vws = vmw_winsys_screen(sws);
- const uint32 *capsBlock;
- const SVGA3dCapsRecord *capsRecord = NULL;
- uint32 offset;
- const SVGA3dCapPair *capArray;
- int numCaps, first, last;
- if(vws->ioctl.hwversion < SVGA3D_HWVERSION_WS6_B1)
+ if (index > vws->ioctl.num_cap_3d ||
+ index >= SVGA3D_DEVCAP_MAX ||
+ !vws->ioctl.cap_3d[index].has_cap)
return FALSE;
- /*
- * Search linearly through the caps block records for the specified type.
- */
- capsBlock = (const uint32 *)vws->ioctl.buffer;
- for (offset = 0; capsBlock[offset] != 0; offset += capsBlock[offset]) {
- const SVGA3dCapsRecord *record;
- assert(offset < SVGA_FIFO_3D_CAPS_SIZE);
- record = (const SVGA3dCapsRecord *) (capsBlock + offset);
- if ((record->header.type >= SVGA3DCAPS_RECORD_DEVCAPS_MIN) &&
- (record->header.type <= SVGA3DCAPS_RECORD_DEVCAPS_MAX) &&
- (!capsRecord || (record->header.type > capsRecord->header.type))) {
- capsRecord = record;
- }
+ *result = vws->ioctl.cap_3d[index].result;
+ return TRUE;
+}
+
+struct svga_winsys_gb_shader *
+vmw_svga_winsys_shader_create(struct svga_winsys_screen *sws,
+ SVGA3dShaderType type,
+ const uint32 *bytecode,
+ uint32 bytecodeLen)
+{
+ struct vmw_winsys_screen *vws = vmw_winsys_screen(sws);
+ struct vmw_svga_winsys_shader *shader;
+ void *code;
+
+ shader = CALLOC_STRUCT(vmw_svga_winsys_shader);
+ if(!shader)
+ goto out_no_shader;
+
+ pipe_reference_init(&shader->refcnt, 1);
+ p_atomic_set(&shader->validated, 0);
+ shader->screen = vws;
+ shader->buf = vmw_svga_winsys_buffer_create(sws, 64,
+ SVGA_BUFFER_USAGE_SHADER,
+ bytecodeLen);
+ if (!shader->buf)
+ goto out_no_buf;
+
+ code = vmw_svga_winsys_buffer_map(sws, shader->buf, PIPE_TRANSFER_WRITE);
+ if (!code)
+ goto out_no_buf;
+
+ memcpy(code, bytecode, bytecodeLen);
+ vmw_svga_winsys_buffer_unmap(sws, shader->buf);
+
+ if (!sws->have_vgpu10) {
+ shader->shid = vmw_ioctl_shader_create(vws, type, bytecodeLen);
+ if (shader->shid == SVGA3D_INVALID_ID)
+ goto out_no_shid;
}
- if(!capsRecord)
- return FALSE;
+ return svga_winsys_shader(shader);
- /*
- * Calculate the number of caps from the size of the record.
- */
- capArray = (const SVGA3dCapPair *) capsRecord->data;
- numCaps = (int) ((capsRecord->header.length * sizeof(uint32) -
- sizeof capsRecord->header) / (2 * sizeof(uint32)));
+out_no_shid:
+ vmw_svga_winsys_buffer_destroy(sws, shader->buf);
+out_no_buf:
+ FREE(shader);
+out_no_shader:
+ return NULL;
+}
- /*
- * Binary-search for the cap with the specified index.
- */
- for (first = 0, last = numCaps - 1; first <= last; ) {
- int mid = (first + last) / 2;
-
- if ((SVGA3dDevCapIndex) capArray[mid][0] == index) {
- /*
- * Found it.
- */
- result->u = capArray[mid][1];
- return TRUE;
- }
+void
+vmw_svga_winsys_shader_destroy(struct svga_winsys_screen *sws,
+ struct svga_winsys_gb_shader *shader)
+{
+ struct vmw_svga_winsys_shader *d_shader =
+ vmw_svga_winsys_shader(shader);
- /*
- * Divide and conquer.
- */
- if ((SVGA3dDevCapIndex) capArray[mid][0] > index) {
- last = mid - 1;
- } else {
- first = mid + 1;
- }
- }
+ vmw_svga_winsys_shader_reference(&d_shader, NULL);
+}
+
+static void
+vmw_svga_winsys_stats_inc(enum svga_stats_count index)
+{
+}
- return FALSE;
+static void
+vmw_svga_winsys_stats_time_push(enum svga_stats_time index,
+ struct svga_winsys_stats_timeframe *tf)
+{
}
+static void
+vmw_svga_winsys_stats_time_pop()
+{
+}
boolean
vmw_winsys_screen_init_svga(struct vmw_winsys_screen *vws)
vws->base.surface_create = vmw_svga_winsys_surface_create;
vws->base.surface_is_flushed = vmw_svga_winsys_surface_is_flushed;
vws->base.surface_reference = vmw_svga_winsys_surface_ref;
+ vws->base.surface_can_create = vmw_svga_winsys_surface_can_create;
vws->base.buffer_create = vmw_svga_winsys_buffer_create;
vws->base.buffer_map = vmw_svga_winsys_buffer_map;
vws->base.buffer_unmap = vmw_svga_winsys_buffer_unmap;
vws->base.buffer_destroy = vmw_svga_winsys_buffer_destroy;
vws->base.fence_reference = vmw_svga_winsys_fence_reference;
vws->base.fence_signalled = vmw_svga_winsys_fence_signalled;
+ vws->base.shader_create = vmw_svga_winsys_shader_create;
+ vws->base.shader_destroy = vmw_svga_winsys_shader_destroy;
vws->base.fence_finish = vmw_svga_winsys_fence_finish;
+ vws->base.fence_get_fd = vmw_svga_winsys_fence_get_fd;
+ vws->base.fence_create_fd = vmw_svga_winsys_fence_create_fd;
+ vws->base.fence_server_sync = vmw_svga_winsys_fence_server_sync;
+
+ vws->base.query_create = vmw_svga_winsys_query_create;
+ vws->base.query_init = vmw_svga_winsys_query_init;
+ vws->base.query_destroy = vmw_svga_winsys_query_destroy;
+ vws->base.query_get_result = vmw_svga_winsys_query_get_result;
+
+ vws->base.stats_inc = vmw_svga_winsys_stats_inc;
+ vws->base.stats_time_push = vmw_svga_winsys_stats_time_push;
+ vws->base.stats_time_pop = vmw_svga_winsys_stats_time_pop;
return TRUE;
}