X-Git-Url: https://git.libre-soc.org/?a=blobdiff_plain;f=src%2Fvulkan%2Fwsi%2Fwsi_common_x11.c;h=7e7b3a94e4bb04bac6130fa83a81836150ccf683;hb=8f3b58ebee413fc704fb2fa0fab06dbfe146b08a;hp=548352eec24be3a72ac07f38023e2c4c7e6f1639;hpb=750d8cad72a532d977df10ffbbdd1902bd06f50b;p=mesa.git diff --git a/src/vulkan/wsi/wsi_common_x11.c b/src/vulkan/wsi/wsi_common_x11.c index 548352eec24..7e7b3a94e4b 100644 --- a/src/vulkan/wsi/wsi_common_x11.c +++ b/src/vulkan/wsi/wsi_common_x11.c @@ -33,20 +33,27 @@ #include #include #include - +#include +#include +#include +#include #include "util/hash_table.h" -#include "wsi_common.h" +#include "vk_util.h" +#include "wsi_common_private.h" #include "wsi_common_x11.h" +#include "wsi_common_queue.h" #define typed_memcpy(dest, src, count) ({ \ - static_assert(sizeof(*src) == sizeof(*dest), ""); \ + STATIC_ASSERT(sizeof(*src) == sizeof(*dest)); \ memcpy((dest), (src), (count) * sizeof(*(src))); \ }) struct wsi_x11_connection { bool has_dri3; + bool has_dri3_modifiers; bool has_present; + bool is_proprietary_x11; }; struct wsi_x11 { @@ -57,12 +64,74 @@ struct wsi_x11 { struct hash_table *connections; }; + +/** wsi_dri3_open + * + * Wrapper around xcb_dri3_open + */ +static int +wsi_dri3_open(xcb_connection_t *conn, + xcb_window_t root, + uint32_t provider) +{ + xcb_dri3_open_cookie_t cookie; + xcb_dri3_open_reply_t *reply; + int fd; + + cookie = xcb_dri3_open(conn, + root, + provider); + + reply = xcb_dri3_open_reply(conn, cookie, NULL); + if (!reply) + return -1; + + if (reply->nfd != 1) { + free(reply); + return -1; + } + + fd = xcb_dri3_open_reply_fds(conn, reply)[0]; + free(reply); + fcntl(fd, F_SETFD, fcntl(fd, F_GETFD) | FD_CLOEXEC); + + return fd; +} + +static bool +wsi_x11_check_dri3_compatible(xcb_connection_t *conn, int local_fd) +{ + xcb_screen_iterator_t screen_iter = + xcb_setup_roots_iterator(xcb_get_setup(conn)); + xcb_screen_t *screen = screen_iter.data; + + int dri3_fd = wsi_dri3_open(conn, screen->root, None); + if (dri3_fd != -1) { + char *local_dev = drmGetRenderDeviceNameFromFd(local_fd); + char *dri3_dev = drmGetRenderDeviceNameFromFd(dri3_fd); + int ret; + + close(dri3_fd); + + ret = strcmp(local_dev, dri3_dev); + + free(local_dev); + free(dri3_dev); + + if (ret != 0) + return false; + } + return true; +} + static struct wsi_x11_connection * wsi_x11_connection_create(const VkAllocationCallbacks *alloc, xcb_connection_t *conn) { - xcb_query_extension_cookie_t dri3_cookie, pres_cookie; - xcb_query_extension_reply_t *dri3_reply, *pres_reply; + xcb_query_extension_cookie_t dri3_cookie, pres_cookie, amd_cookie, nv_cookie; + xcb_query_extension_reply_t *dri3_reply, *pres_reply, *amd_reply, *nv_reply; + bool has_dri3_v1_2 = false; + bool has_present_v1_2 = false; struct wsi_x11_connection *wsi_conn = vk_alloc(alloc, sizeof(*wsi_conn), 8, @@ -71,22 +140,72 @@ wsi_x11_connection_create(const VkAllocationCallbacks *alloc, return NULL; dri3_cookie = xcb_query_extension(conn, 4, "DRI3"); - pres_cookie = xcb_query_extension(conn, 7, "PRESENT"); + pres_cookie = xcb_query_extension(conn, 7, "Present"); + + /* We try to be nice to users and emit a warning if they try to use a + * Vulkan application on a system without DRI3 enabled. However, this ends + * up spewing the warning when a user has, for example, both Intel + * integrated graphics and a discrete card with proprietary drivers and are + * running on the discrete card with the proprietary DDX. In this case, we + * really don't want to print the warning because it just confuses users. + * As a heuristic to detect this case, we check for a couple of proprietary + * X11 extensions. + */ + amd_cookie = xcb_query_extension(conn, 11, "ATIFGLRXDRI"); + nv_cookie = xcb_query_extension(conn, 10, "NV-CONTROL"); dri3_reply = xcb_query_extension_reply(conn, dri3_cookie, NULL); pres_reply = xcb_query_extension_reply(conn, pres_cookie, NULL); - if (dri3_reply == NULL || pres_reply == NULL) { + amd_reply = xcb_query_extension_reply(conn, amd_cookie, NULL); + nv_reply = xcb_query_extension_reply(conn, nv_cookie, NULL); + if (!dri3_reply || !pres_reply) { free(dri3_reply); free(pres_reply); + free(amd_reply); + free(nv_reply); vk_free(alloc, wsi_conn); return NULL; } wsi_conn->has_dri3 = dri3_reply->present != 0; +#ifdef HAVE_DRI3_MODIFIERS + if (wsi_conn->has_dri3) { + xcb_dri3_query_version_cookie_t ver_cookie; + xcb_dri3_query_version_reply_t *ver_reply; + + ver_cookie = xcb_dri3_query_version(conn, 1, 2); + ver_reply = xcb_dri3_query_version_reply(conn, ver_cookie, NULL); + has_dri3_v1_2 = + (ver_reply->major_version > 1 || ver_reply->minor_version >= 2); + free(ver_reply); + } +#endif + wsi_conn->has_present = pres_reply->present != 0; +#ifdef HAVE_DRI3_MODIFIERS + if (wsi_conn->has_present) { + xcb_present_query_version_cookie_t ver_cookie; + xcb_present_query_version_reply_t *ver_reply; + + ver_cookie = xcb_present_query_version(conn, 1, 2); + ver_reply = xcb_present_query_version_reply(conn, ver_cookie, NULL); + has_present_v1_2 = + (ver_reply->major_version > 1 || ver_reply->minor_version >= 2); + free(ver_reply); + } +#endif + + wsi_conn->has_dri3_modifiers = has_dri3_v1_2 && has_present_v1_2; + wsi_conn->is_proprietary_x11 = false; + if (amd_reply && amd_reply->present) + wsi_conn->is_proprietary_x11 = true; + if (nv_reply && nv_reply->present) + wsi_conn->is_proprietary_x11 = true; free(dri3_reply); free(pres_reply); + free(amd_reply); + free(nv_reply); return wsi_conn; } @@ -98,6 +217,18 @@ wsi_x11_connection_destroy(const VkAllocationCallbacks *alloc, vk_free(alloc, conn); } +static bool +wsi_x11_check_for_dri3(struct wsi_x11_connection *wsi_conn) +{ + if (wsi_conn->has_dri3) + return true; + if (!wsi_conn->is_proprietary_x11) { + fprintf(stderr, "vulkan: No DRI3 support detected - required for presentation\n" + "Note: you can probably enable DRI3 in your Xorg config\n"); + } + return false; +} + static struct wsi_x11_connection * wsi_x11_get_connection(struct wsi_device *wsi_dev, const VkAllocationCallbacks *alloc, @@ -117,6 +248,8 @@ wsi_x11_get_connection(struct wsi_device *wsi_dev, struct wsi_x11_connection *wsi_conn = wsi_x11_connection_create(alloc, conn); + if (!wsi_conn) + return NULL; pthread_mutex_lock(&wsi->mutex); @@ -134,13 +267,15 @@ wsi_x11_get_connection(struct wsi_device *wsi_dev, return entry->data; } -static const VkSurfaceFormatKHR formats[] = { - { .format = VK_FORMAT_B8G8R8A8_SRGB, }, - { .format = VK_FORMAT_B8G8R8A8_UNORM, }, +static const VkFormat formats[] = { + VK_FORMAT_B8G8R8A8_SRGB, + VK_FORMAT_B8G8R8A8_UNORM, }; static const VkPresentModeKHR present_modes[] = { + VK_PRESENT_MODE_IMMEDIATE_KHR, VK_PRESENT_MODE_MAILBOX_KHR, + VK_PRESENT_MODE_FIFO_KHR, }; static xcb_screen_t * @@ -249,16 +384,23 @@ VkBool32 wsi_get_physical_device_xcb_presentation_support( struct wsi_device *wsi_device, VkAllocationCallbacks *alloc, uint32_t queueFamilyIndex, + int fd, + bool can_handle_different_gpu, xcb_connection_t* connection, xcb_visualid_t visual_id) { struct wsi_x11_connection *wsi_conn = wsi_x11_get_connection(wsi_device, alloc, connection); - if (!wsi_conn->has_dri3) { - fprintf(stderr, "vulkan: No DRI3 support\n"); + if (!wsi_conn) + return false; + + if (!wsi_x11_check_for_dri3(wsi_conn)) return false; - } + + if (!can_handle_different_gpu) + if (!wsi_x11_check_dri3_compatible(connection, fd)) + return false; unsigned visual_depth; if (!connection_get_visualtype(connection, visual_id, &visual_depth)) @@ -293,6 +435,7 @@ x11_surface_get_support(VkIcdSurfaceBase *icd_surface, struct wsi_device *wsi_device, const VkAllocationCallbacks *alloc, uint32_t queueFamilyIndex, + int local_fd, VkBool32* pSupported) { xcb_connection_t *conn = x11_surface_get_connection(icd_surface); @@ -303,8 +446,7 @@ x11_surface_get_support(VkIcdSurfaceBase *icd_surface, if (!wsi_conn) return VK_ERROR_OUT_OF_HOST_MEMORY; - if (!wsi_conn->has_dri3) { - fprintf(stderr, "vulkan: No DRI3 support\n"); + if (!wsi_x11_check_for_dri3(wsi_conn)) { *pSupported = false; return VK_SUCCESS; } @@ -343,6 +485,9 @@ x11_surface_get_capabilities(VkIcdSurfaceBase *icd_surface, xcb_visualtype_t *visual = get_visualtype_for_window(conn, window, &visual_depth); + if (!visual) + return VK_ERROR_SURFACE_LOST_KHR; + geom = xcb_get_geometry_reply(conn, geom_cookie, &err); if (geom) { VkExtent2D extent = { geom->width, geom->height }; @@ -356,7 +501,8 @@ x11_surface_get_capabilities(VkIcdSurfaceBase *icd_surface, */ caps->currentExtent = (VkExtent2D) { -1, -1 }; caps->minImageExtent = (VkExtent2D) { 1, 1 }; - caps->maxImageExtent = (VkExtent2D) { INT16_MAX, INT16_MAX }; + /* This is the maximum supported size on Intel */ + caps->maxImageExtent = (VkExtent2D) { 1 << 14, 1 << 14 }; } free(err); free(geom); @@ -369,8 +515,16 @@ x11_surface_get_capabilities(VkIcdSurfaceBase *icd_surface, VK_COMPOSITE_ALPHA_OPAQUE_BIT_KHR; } + /* For true mailbox mode, we need at least 4 images: + * 1) One to scan out from + * 2) One to have queued for scan-out + * 3) One to be currently held by the X server + * 4) One to render to + */ caps->minImageCount = 2; - caps->maxImageCount = 4; + /* There is no real maximum */ + caps->maxImageCount = 0; + caps->supportedTransforms = VK_SURFACE_TRANSFORM_IDENTITY_BIT_KHR; caps->currentTransform = VK_SURFACE_TRANSFORM_IDENTITY_BIT_KHR; caps->maxImageArrayLayers = 1; @@ -383,27 +537,52 @@ x11_surface_get_capabilities(VkIcdSurfaceBase *icd_surface, return VK_SUCCESS; } +static VkResult +x11_surface_get_capabilities2(VkIcdSurfaceBase *icd_surface, + const void *info_next, + VkSurfaceCapabilities2KHR *caps) +{ + assert(caps->sType == VK_STRUCTURE_TYPE_SURFACE_CAPABILITIES_2_KHR); + + return x11_surface_get_capabilities(icd_surface, &caps->surfaceCapabilities); +} + static VkResult x11_surface_get_formats(VkIcdSurfaceBase *surface, struct wsi_device *wsi_device, uint32_t *pSurfaceFormatCount, VkSurfaceFormatKHR *pSurfaceFormats) { - if (pSurfaceFormats == NULL) { - *pSurfaceFormatCount = ARRAY_SIZE(formats); - return VK_SUCCESS; + VK_OUTARRAY_MAKE(out, pSurfaceFormats, pSurfaceFormatCount); + + for (unsigned i = 0; i < ARRAY_SIZE(formats); i++) { + vk_outarray_append(&out, f) { + f->format = formats[i]; + f->colorSpace = VK_COLORSPACE_SRGB_NONLINEAR_KHR; + } } - VkResult result = VK_SUCCESS; + return vk_outarray_status(&out); +} - if (*pSurfaceFormatCount > ARRAY_SIZE(formats)) - *pSurfaceFormatCount = ARRAY_SIZE(formats); - else if (*pSurfaceFormatCount < ARRAY_SIZE(formats)) - result = VK_INCOMPLETE; +static VkResult +x11_surface_get_formats2(VkIcdSurfaceBase *surface, + struct wsi_device *wsi_device, + const void *info_next, + uint32_t *pSurfaceFormatCount, + VkSurfaceFormat2KHR *pSurfaceFormats) +{ + VK_OUTARRAY_MAKE(out, pSurfaceFormats, pSurfaceFormatCount); - typed_memcpy(pSurfaceFormats, formats, *pSurfaceFormatCount); + for (unsigned i = 0; i < ARRAY_SIZE(formats); i++) { + vk_outarray_append(&out, f) { + assert(f->sType == VK_STRUCTURE_TYPE_SURFACE_FORMAT_2_KHR); + f->surfaceFormat.format = formats[i]; + f->surfaceFormat.colorSpace = VK_COLORSPACE_SRGB_NONLINEAR_KHR; + } + } - return result; + return vk_outarray_status(&out); } static VkResult @@ -416,11 +595,11 @@ x11_surface_get_present_modes(VkIcdSurfaceBase *surface, return VK_SUCCESS; } - assert(*pPresentModeCount >= ARRAY_SIZE(present_modes)); + *pPresentModeCount = MIN2(*pPresentModeCount, ARRAY_SIZE(present_modes)); typed_memcpy(pPresentModes, present_modes, *pPresentModeCount); - *pPresentModeCount = ARRAY_SIZE(present_modes); - return VK_SUCCESS; + return *pPresentModeCount < ARRAY_SIZE(present_modes) ? + VK_INCOMPLETE : VK_SUCCESS; } VkResult wsi_create_xcb_surface(const VkAllocationCallbacks *pAllocator, @@ -438,7 +617,7 @@ VkResult wsi_create_xcb_surface(const VkAllocationCallbacks *pAllocator, surface->connection = pCreateInfo->connection; surface->window = pCreateInfo->window; - *pSurface = _VkIcdSurfaceBase_to_handle(&surface->base); + *pSurface = VkIcdSurfaceBase_to_handle(&surface->base); return VK_SUCCESS; } @@ -457,13 +636,12 @@ VkResult wsi_create_xlib_surface(const VkAllocationCallbacks *pAllocator, surface->dpy = pCreateInfo->dpy; surface->window = pCreateInfo->window; - *pSurface = _VkIcdSurfaceBase_to_handle(&surface->base); + *pSurface = VkIcdSurfaceBase_to_handle(&surface->base); return VK_SUCCESS; } struct x11_image { - VkImage image; - VkDeviceMemory memory; + struct wsi_image base; xcb_pixmap_t pixmap; bool busy; struct xshmfence * shm_fence; @@ -473,40 +651,78 @@ struct x11_image { struct x11_swapchain { struct wsi_swapchain base; + bool has_dri3_modifiers; + xcb_connection_t * conn; xcb_window_t window; xcb_gc_t gc; + uint32_t depth; VkExtent2D extent; - uint32_t image_count; xcb_present_event_t event_id; xcb_special_event_t * special_event; uint64_t send_sbc; + uint64_t last_present_msc; uint32_t stamp; + bool threaded; + VkResult status; + xcb_present_complete_mode_t last_present_mode; + struct wsi_queue present_queue; + struct wsi_queue acquire_queue; + pthread_t queue_manager; + struct x11_image images[0]; }; +/** + * Update the swapchain status with the result of an operation, and return + * the combined status. The chain status will eventually be returned from + * AcquireNextImage and QueuePresent. + * + * We make sure to 'stick' more pessimistic statuses: an out-of-date error + * is permanent once seen, and every subsequent call will return this. If + * this has not been seen, success will be returned. + */ static VkResult -x11_get_images(struct wsi_swapchain *anv_chain, - uint32_t* pCount, VkImage *pSwapchainImages) +x11_swapchain_result(struct x11_swapchain *chain, VkResult result) { - struct x11_swapchain *chain = (struct x11_swapchain *)anv_chain; + /* Prioritise returning existing errors for consistency. */ + if (chain->status < 0) + return chain->status; - if (pSwapchainImages == NULL) { - *pCount = chain->image_count; - return VK_SUCCESS; + /* If we have a new error, mark it as permanent on the chain and return. */ + if (result < 0) { + chain->status = result; + return result; } - assert(chain->image_count <= *pCount); - for (uint32_t i = 0; i < chain->image_count; i++) - pSwapchainImages[i] = chain->images[i].image; + /* Return temporary errors, but don't persist them. */ + if (result == VK_TIMEOUT || result == VK_NOT_READY) + return result; - *pCount = chain->image_count; + /* Suboptimal isn't an error, but is a status which sticks to the swapchain + * and is always returned rather than success. + */ + if (result == VK_SUBOPTIMAL_KHR) { + chain->status = result; + return result; + } - return VK_SUCCESS; + /* No changes, so return the last status. */ + return chain->status; +} + +static struct wsi_image * +x11_get_wsi_image(struct wsi_swapchain *wsi_chain, uint32_t image_index) +{ + struct x11_swapchain *chain = (struct x11_swapchain *)wsi_chain; + return &chain->images[image_index].base; } +/** + * Process an X11 Present event. Does not update chain->status. + */ static VkResult x11_handle_dri3_present_event(struct x11_swapchain *chain, xcb_present_generic_event_t *event) @@ -525,9 +741,11 @@ x11_handle_dri3_present_event(struct x11_swapchain *chain, case XCB_PRESENT_EVENT_IDLE_NOTIFY: { xcb_present_idle_notify_event_t *idle = (void *) event; - for (unsigned i = 0; i < chain->image_count; i++) { + for (unsigned i = 0; i < chain->base.image_count; i++) { if (chain->images[i].pixmap == idle->pixmap) { chain->images[i].busy = false; + if (chain->threaded) + wsi_queue_push(&chain->acquire_queue, i); break; } } @@ -535,7 +753,34 @@ x11_handle_dri3_present_event(struct x11_swapchain *chain, break; } - case XCB_PRESENT_COMPLETE_NOTIFY: + case XCB_PRESENT_EVENT_COMPLETE_NOTIFY: { + xcb_present_complete_notify_event_t *complete = (void *) event; + if (complete->kind == XCB_PRESENT_COMPLETE_KIND_PIXMAP) + chain->last_present_msc = complete->msc; + + VkResult result = VK_SUCCESS; + + /* The winsys is now trying to flip directly and cannot due to our + * configuration. Request the user reallocate. + */ +#ifdef HAVE_DRI3_MODIFIERS + if (complete->mode == XCB_PRESENT_COMPLETE_MODE_SUBOPTIMAL_COPY && + chain->last_present_mode != XCB_PRESENT_COMPLETE_MODE_SUBOPTIMAL_COPY) + result = VK_SUBOPTIMAL_KHR; +#endif + + /* When we go from flipping to copying, the odds are very likely that + * we could reallocate in a more optimal way if we didn't have to care + * about scanout, so we always do this. + */ + if (complete->mode == XCB_PRESENT_COMPLETE_MODE_COPY && + chain->last_present_mode == XCB_PRESENT_COMPLETE_MODE_FLIP) + result = VK_SUBOPTIMAL_KHR; + + chain->last_present_mode = complete->mode; + return result; + } + default: break; } @@ -543,54 +788,136 @@ x11_handle_dri3_present_event(struct x11_swapchain *chain, return VK_SUCCESS; } -static VkResult -x11_acquire_next_image(struct wsi_swapchain *anv_chain, - uint64_t timeout, - VkSemaphore semaphore, - uint32_t *image_index) + +static uint64_t wsi_get_current_time(void) { - struct x11_swapchain *chain = (struct x11_swapchain *)anv_chain; + uint64_t current_time; + struct timespec tv; + + clock_gettime(CLOCK_MONOTONIC, &tv); + current_time = tv.tv_nsec + tv.tv_sec*1000000000ull; + return current_time; +} + +static uint64_t wsi_get_absolute_timeout(uint64_t timeout) +{ + uint64_t current_time = wsi_get_current_time(); + timeout = MIN2(UINT64_MAX - current_time, timeout); + + return current_time + timeout; +} + +static VkResult +x11_acquire_next_image_poll_x11(struct x11_swapchain *chain, + uint32_t *image_index, uint64_t timeout) +{ + xcb_generic_event_t *event; + struct pollfd pfds; + uint64_t atimeout; while (1) { - for (uint32_t i = 0; i < chain->image_count; i++) { + for (uint32_t i = 0; i < chain->base.image_count; i++) { if (!chain->images[i].busy) { /* We found a non-busy image */ xshmfence_await(chain->images[i].shm_fence); *image_index = i; chain->images[i].busy = true; - return VK_SUCCESS; + return x11_swapchain_result(chain, VK_SUCCESS); } } xcb_flush(chain->conn); - xcb_generic_event_t *event = - xcb_wait_for_special_event(chain->conn, chain->special_event); - if (!event) - return VK_ERROR_OUT_OF_DATE_KHR; + if (timeout == UINT64_MAX) { + event = xcb_wait_for_special_event(chain->conn, chain->special_event); + if (!event) + return x11_swapchain_result(chain, VK_ERROR_OUT_OF_DATE_KHR); + } else { + event = xcb_poll_for_special_event(chain->conn, chain->special_event); + if (!event) { + int ret; + if (timeout == 0) + return x11_swapchain_result(chain, VK_NOT_READY); + + atimeout = wsi_get_absolute_timeout(timeout); + + pfds.fd = xcb_get_file_descriptor(chain->conn); + pfds.events = POLLIN; + ret = poll(&pfds, 1, timeout / 1000 / 1000); + if (ret == 0) + return x11_swapchain_result(chain, VK_TIMEOUT); + if (ret == -1) + return x11_swapchain_result(chain, VK_ERROR_OUT_OF_DATE_KHR); + + /* If a non-special event happens, the fd will still + * poll. So recalculate the timeout now just in case. + */ + uint64_t current_time = wsi_get_current_time(); + if (atimeout > current_time) + timeout = atimeout - current_time; + else + timeout = 0; + continue; + } + } + + /* Update the swapchain status here. We may catch non-fatal errors here, + * in which case we need to update the status and continue. + */ VkResult result = x11_handle_dri3_present_event(chain, (void *)event); free(event); - if (result != VK_SUCCESS) - return result; + if (result < 0) + return x11_swapchain_result(chain, result); } } static VkResult -x11_queue_present(struct wsi_swapchain *anv_chain, - uint32_t image_index) +x11_acquire_next_image_from_queue(struct x11_swapchain *chain, + uint32_t *image_index_out, uint64_t timeout) +{ + assert(chain->threaded); + + uint32_t image_index; + VkResult result = wsi_queue_pull(&chain->acquire_queue, + &image_index, timeout); + if (result < 0 || result == VK_TIMEOUT) { + /* On error, the thread has shut down, so safe to update chain->status. + * Calling x11_swapchain_result with VK_TIMEOUT won't modify + * chain->status so that is also safe. + */ + return x11_swapchain_result(chain, result); + } else if (chain->status < 0) { + return chain->status; + } + + assert(image_index < chain->base.image_count); + xshmfence_await(chain->images[image_index].shm_fence); + + *image_index_out = image_index; + + return chain->status; +} + +static VkResult +x11_present_to_x11(struct x11_swapchain *chain, uint32_t image_index, + uint32_t target_msc) { - struct x11_swapchain *chain = (struct x11_swapchain *)anv_chain; struct x11_image *image = &chain->images[image_index]; - assert(image_index < chain->image_count); + assert(image_index < chain->base.image_count); uint32_t options = XCB_PRESENT_OPTION_NONE; - int64_t target_msc = 0; int64_t divisor = 0; int64_t remainder = 0; - options |= XCB_PRESENT_OPTION_ASYNC; + if (chain->base.present_mode == VK_PRESENT_MODE_IMMEDIATE_KHR) + options |= XCB_PRESENT_OPTION_ASYNC; + +#ifdef HAVE_DRI3_MODIFIERS + if (chain->has_dri3_modifiers) + options |= XCB_PRESENT_OPTION_SUBOPTIMAL; +#endif xshmfence_reset(image->shm_fence); @@ -616,49 +943,165 @@ x11_queue_present(struct wsi_swapchain *anv_chain, xcb_flush(chain->conn); - return VK_SUCCESS; + return x11_swapchain_result(chain, VK_SUCCESS); +} + +static VkResult +x11_acquire_next_image(struct wsi_swapchain *anv_chain, + uint64_t timeout, + VkSemaphore semaphore, + uint32_t *image_index) +{ + struct x11_swapchain *chain = (struct x11_swapchain *)anv_chain; + + if (chain->threaded) { + return x11_acquire_next_image_from_queue(chain, image_index, timeout); + } else { + return x11_acquire_next_image_poll_x11(chain, image_index, timeout); + } +} + +static VkResult +x11_queue_present(struct wsi_swapchain *anv_chain, + uint32_t image_index, + const VkPresentRegionKHR *damage) +{ + struct x11_swapchain *chain = (struct x11_swapchain *)anv_chain; + + if (chain->threaded) { + wsi_queue_push(&chain->present_queue, image_index); + return chain->status; + } else { + return x11_present_to_x11(chain, image_index, 0); + } +} + +static void * +x11_manage_fifo_queues(void *state) +{ + struct x11_swapchain *chain = state; + VkResult result; + + assert(chain->base.present_mode == VK_PRESENT_MODE_FIFO_KHR); + + while (chain->status >= 0) { + /* It should be safe to unconditionally block here. Later in the loop + * we blocks until the previous present has landed on-screen. At that + * point, we should have received IDLE_NOTIFY on all images presented + * before that point so the client should be able to acquire any image + * other than the currently presented one. + */ + uint32_t image_index; + result = wsi_queue_pull(&chain->present_queue, &image_index, INT64_MAX); + assert(result != VK_TIMEOUT); + if (result < 0) { + goto fail; + } else if (chain->status < 0) { + /* The status can change underneath us if the swapchain is destroyed + * from another thread. + */ + return NULL; + } + + uint64_t target_msc = chain->last_present_msc + 1; + result = x11_present_to_x11(chain, image_index, target_msc); + if (result < 0) + goto fail; + + while (chain->last_present_msc < target_msc) { + xcb_generic_event_t *event = + xcb_wait_for_special_event(chain->conn, chain->special_event); + if (!event) { + result = VK_ERROR_OUT_OF_DATE_KHR; + goto fail; + } + + result = x11_handle_dri3_present_event(chain, (void *)event); + free(event); + if (result < 0) + goto fail; + } + } + +fail: + result = x11_swapchain_result(chain, result); + wsi_queue_push(&chain->acquire_queue, UINT32_MAX); + + return NULL; } static VkResult x11_image_init(VkDevice device_h, struct x11_swapchain *chain, const VkSwapchainCreateInfoKHR *pCreateInfo, const VkAllocationCallbacks* pAllocator, - struct x11_image *image) + const uint64_t *const *modifiers, + const uint32_t *num_modifiers, + int num_tranches, struct x11_image *image) { xcb_void_cookie_t cookie; VkResult result; - uint32_t row_pitch; - uint32_t offset; uint32_t bpp = 32; - uint32_t depth = 24; - int fd; - uint32_t size; - - result = chain->base.image_fns->create_wsi_image(device_h, - pCreateInfo, - pAllocator, - &image->image, - &image->memory, - &size, - &offset, - &row_pitch, - &fd); - if (result != VK_SUCCESS) + + if (chain->base.use_prime_blit) { + bool use_modifier = num_tranches > 0; + result = wsi_create_prime_image(&chain->base, pCreateInfo, use_modifier, &image->base); + } else { + result = wsi_create_native_image(&chain->base, pCreateInfo, + num_tranches, num_modifiers, modifiers, + &image->base); + } + if (result < 0) return result; image->pixmap = xcb_generate_id(chain->conn); - cookie = - xcb_dri3_pixmap_from_buffer_checked(chain->conn, - image->pixmap, - chain->window, - size, - pCreateInfo->imageExtent.width, - pCreateInfo->imageExtent.height, - row_pitch, - depth, bpp, fd); +#ifdef HAVE_DRI3_MODIFIERS + if (image->base.drm_modifier != DRM_FORMAT_MOD_INVALID) { + /* If the image has a modifier, we must have DRI3 v1.2. */ + assert(chain->has_dri3_modifiers); + + cookie = + xcb_dri3_pixmap_from_buffers_checked(chain->conn, + image->pixmap, + chain->window, + image->base.num_planes, + pCreateInfo->imageExtent.width, + pCreateInfo->imageExtent.height, + image->base.row_pitches[0], + image->base.offsets[0], + image->base.row_pitches[1], + image->base.offsets[1], + image->base.row_pitches[2], + image->base.offsets[2], + image->base.row_pitches[3], + image->base.offsets[3], + chain->depth, bpp, + image->base.drm_modifier, + image->base.fds); + } else +#endif + { + /* Without passing modifiers, we can't have multi-plane RGB images. */ + assert(image->base.num_planes == 1); + + cookie = + xcb_dri3_pixmap_from_buffer_checked(chain->conn, + image->pixmap, + chain->window, + image->base.sizes[0], + pCreateInfo->imageExtent.width, + pCreateInfo->imageExtent.height, + image->base.row_pitches[0], + chain->depth, bpp, + image->base.fds[0]); + } + xcb_discard_reply(chain->conn, cookie.sequence); + /* XCB has now taken ownership of the FDs. */ + for (int i = 0; i < image->base.num_planes; i++) + image->base.fds[i] = -1; + int fence_fd = xshmfence_alloc_shm(); if (fence_fd < 0) goto fail_pixmap; @@ -686,8 +1129,7 @@ fail_pixmap: cookie = xcb_free_pixmap(chain->conn, image->pixmap); xcb_discard_reply(chain->conn, cookie.sequence); - chain->base.image_fns->free_wsi_image(device_h, pAllocator, - image->image, image->memory); + wsi_destroy_image(&chain->base, &image->base); return result; } @@ -706,8 +1148,84 @@ x11_image_finish(struct x11_swapchain *chain, cookie = xcb_free_pixmap(chain->conn, image->pixmap); xcb_discard_reply(chain->conn, cookie.sequence); - chain->base.image_fns->free_wsi_image(chain->base.device, pAllocator, - image->image, image->memory); + wsi_destroy_image(&chain->base, &image->base); +} + +static void +wsi_x11_get_dri3_modifiers(struct wsi_x11_connection *wsi_conn, + xcb_connection_t *conn, xcb_window_t window, + uint8_t depth, uint8_t bpp, + VkCompositeAlphaFlagsKHR vk_alpha, + uint64_t **modifiers_in, uint32_t *num_modifiers_in, + uint32_t *num_tranches_in, + const VkAllocationCallbacks *pAllocator) +{ + if (!wsi_conn->has_dri3_modifiers) + goto out; + +#ifdef HAVE_DRI3_MODIFIERS + xcb_generic_error_t *error = NULL; + xcb_dri3_get_supported_modifiers_cookie_t mod_cookie = + xcb_dri3_get_supported_modifiers(conn, window, depth, bpp); + xcb_dri3_get_supported_modifiers_reply_t *mod_reply = + xcb_dri3_get_supported_modifiers_reply(conn, mod_cookie, &error); + free(error); + + if (!mod_reply || (mod_reply->num_window_modifiers == 0 && + mod_reply->num_screen_modifiers == 0)) { + free(mod_reply); + goto out; + } + + uint32_t n = 0; + uint32_t counts[2]; + uint64_t *modifiers[2]; + + if (mod_reply->num_window_modifiers) { + counts[n] = mod_reply->num_window_modifiers; + modifiers[n] = vk_alloc(pAllocator, + counts[n] * sizeof(uint64_t), + 8, VK_SYSTEM_ALLOCATION_SCOPE_OBJECT); + if (!modifiers[n]) { + free(mod_reply); + goto out; + } + + memcpy(modifiers[n], + xcb_dri3_get_supported_modifiers_window_modifiers(mod_reply), + counts[n] * sizeof(uint64_t)); + n++; + } + + if (mod_reply->num_screen_modifiers) { + counts[n] = mod_reply->num_screen_modifiers; + modifiers[n] = vk_alloc(pAllocator, + counts[n] * sizeof(uint64_t), + 8, VK_SYSTEM_ALLOCATION_SCOPE_OBJECT); + if (!modifiers[n]) { + if (n > 0) + vk_free(pAllocator, modifiers[0]); + free(mod_reply); + goto out; + } + + memcpy(modifiers[n], + xcb_dri3_get_supported_modifiers_screen_modifiers(mod_reply), + counts[n] * sizeof(uint64_t)); + n++; + } + + for (int i = 0; i < n; i++) { + modifiers_in[i] = modifiers[i]; + num_modifiers_in[i] = counts[i]; + } + *num_tranches_in = n; + + free(mod_reply); + return; +#endif +out: + *num_tranches_in = 0; } static VkResult @@ -715,10 +1233,27 @@ x11_swapchain_destroy(struct wsi_swapchain *anv_chain, const VkAllocationCallbacks *pAllocator) { struct x11_swapchain *chain = (struct x11_swapchain *)anv_chain; - for (uint32_t i = 0; i < chain->image_count; i++) + xcb_void_cookie_t cookie; + + if (chain->threaded) { + chain->status = VK_ERROR_OUT_OF_DATE_KHR; + /* Push a UINT32_MAX to wake up the manager */ + wsi_queue_push(&chain->present_queue, UINT32_MAX); + pthread_join(chain->queue_manager, NULL); + wsi_queue_destroy(&chain->acquire_queue); + wsi_queue_destroy(&chain->present_queue); + } + + for (uint32_t i = 0; i < chain->base.image_count; i++) x11_image_finish(chain, pAllocator, &chain->images[i]); xcb_unregister_for_special_event(chain->conn, chain->special_event); + cookie = xcb_present_select_input_checked(chain->conn, chain->event_id, + chain->window, + XCB_PRESENT_EVENT_MASK_NO_EVENT); + xcb_discard_reply(chain->conn, cookie.sequence); + + wsi_swapchain_finish(&chain->base); vk_free(pAllocator, chain); @@ -729,9 +1264,9 @@ static VkResult x11_surface_create_swapchain(VkIcdSurfaceBase *icd_surface, VkDevice device, struct wsi_device *wsi_device, + int local_fd, const VkSwapchainCreateInfoKHR *pCreateInfo, const VkAllocationCallbacks* pAllocator, - const struct wsi_image_fns *image_fns, struct wsi_swapchain **swapchain_out) { struct x11_swapchain *chain; @@ -740,16 +1275,22 @@ x11_surface_create_swapchain(VkIcdSurfaceBase *icd_surface, assert(pCreateInfo->sType == VK_STRUCTURE_TYPE_SWAPCHAIN_CREATE_INFO_KHR); - int num_images = pCreateInfo->minImageCount; + const unsigned num_images = pCreateInfo->minImageCount; - /* For true mailbox mode, we need at least 4 images: - * 1) One to scan out from - * 2) One to have queued for scan-out - * 3) One to be currently held by the Wayland compositor - * 4) One to render to - */ - if (pCreateInfo->presentMode == VK_PRESENT_MODE_MAILBOX_KHR) - num_images = MAX2(num_images, 4); + xcb_connection_t *conn = x11_surface_get_connection(icd_surface); + struct wsi_x11_connection *wsi_conn = + wsi_x11_get_connection(wsi_device, pAllocator, conn); + if (!wsi_conn) + return VK_ERROR_OUT_OF_HOST_MEMORY; + + /* Check for whether or not we have a window up-front */ + xcb_window_t window = x11_surface_get_window(icd_surface); + xcb_get_geometry_reply_t *geometry = + xcb_get_geometry_reply(conn, xcb_get_geometry(conn, window), NULL); + if (geometry == NULL) + return VK_ERROR_SURFACE_LOST_KHR; + const uint32_t bit_depth = geometry->depth; + free(geometry); size_t size = sizeof(*chain) + num_images * sizeof(chain->images[0]); chain = vk_alloc(pAllocator, size, 8, @@ -757,17 +1298,41 @@ x11_surface_create_swapchain(VkIcdSurfaceBase *icd_surface, if (chain == NULL) return VK_ERROR_OUT_OF_HOST_MEMORY; - chain->base.device = device; + result = wsi_swapchain_init(wsi_device, &chain->base, device, + pCreateInfo, pAllocator); + if (result != VK_SUCCESS) + goto fail_alloc; + chain->base.destroy = x11_swapchain_destroy; - chain->base.get_images = x11_get_images; + chain->base.get_wsi_image = x11_get_wsi_image; chain->base.acquire_next_image = x11_acquire_next_image; chain->base.queue_present = x11_queue_present; - chain->base.image_fns = image_fns; - chain->conn = x11_surface_get_connection(icd_surface); - chain->window = x11_surface_get_window(icd_surface); + chain->base.present_mode = pCreateInfo->presentMode; + chain->base.image_count = num_images; + chain->conn = conn; + chain->window = window; + chain->depth = bit_depth; chain->extent = pCreateInfo->imageExtent; - chain->image_count = num_images; chain->send_sbc = 0; + chain->last_present_msc = 0; + chain->threaded = false; + chain->status = VK_SUCCESS; + chain->has_dri3_modifiers = wsi_conn->has_dri3_modifiers; + + /* If we are reallocating from an old swapchain, then we inherit its + * last completion mode, to ensure we don't get into reallocation + * cycles. If we are starting anew, we set 'COPY', as that is the only + * mode which provokes reallocation when anything changes, to make + * sure we have the most optimal allocation. + */ + struct x11_swapchain *old_chain = (void *) pCreateInfo->oldSwapchain; + if (old_chain) + chain->last_present_mode = old_chain->last_present_mode; + else + chain->last_present_mode = XCB_PRESENT_COMPLETE_MODE_COPY; + + if (!wsi_x11_check_dri3_compatible(conn, local_fd)) + chain->base.use_prime_blit = true; chain->event_id = xcb_generate_id(chain->conn); xcb_present_select_input(chain->conn, chain->event_id, chain->window, @@ -796,14 +1361,58 @@ x11_surface_create_swapchain(VkIcdSurfaceBase *icd_surface, (uint32_t []) { 0 }); xcb_discard_reply(chain->conn, cookie.sequence); + uint64_t *modifiers[2] = {NULL, NULL}; + uint32_t num_modifiers[2] = {0, 0}; + uint32_t num_tranches = 0; + if (wsi_device->supports_modifiers) + wsi_x11_get_dri3_modifiers(wsi_conn, conn, window, chain->depth, 32, + pCreateInfo->compositeAlpha, + modifiers, num_modifiers, &num_tranches, + pAllocator); + uint32_t image = 0; - for (; image < chain->image_count; image++) { + for (; image < chain->base.image_count; image++) { result = x11_image_init(device, chain, pCreateInfo, pAllocator, + (const uint64_t *const *)modifiers, + num_modifiers, num_tranches, &chain->images[image]); if (result != VK_SUCCESS) goto fail_init_images; } + if (chain->base.present_mode == VK_PRESENT_MODE_FIFO_KHR) { + chain->threaded = true; + + /* Initialize our queues. We make them base.image_count + 1 because we will + * occasionally use UINT32_MAX to signal the other thread that an error + * has occurred and we don't want an overflow. + */ + int ret; + ret = wsi_queue_init(&chain->acquire_queue, chain->base.image_count + 1); + if (ret) { + goto fail_init_images; + } + + ret = wsi_queue_init(&chain->present_queue, chain->base.image_count + 1); + if (ret) { + wsi_queue_destroy(&chain->acquire_queue); + goto fail_init_images; + } + + for (unsigned i = 0; i < chain->base.image_count; i++) + wsi_queue_push(&chain->acquire_queue, i); + + ret = pthread_create(&chain->queue_manager, NULL, + x11_manage_fifo_queues, chain); + if (ret) { + wsi_queue_destroy(&chain->present_queue); + wsi_queue_destroy(&chain->acquire_queue); + goto fail_init_images; + } + } + + for (int i = 0; i < ARRAY_SIZE(modifiers); i++) + vk_free(pAllocator, modifiers[i]); *swapchain_out = &chain->base; return VK_SUCCESS; @@ -812,9 +1421,15 @@ fail_init_images: for (uint32_t j = 0; j < image; j++) x11_image_finish(chain, pAllocator, &chain->images[j]); + for (int i = 0; i < ARRAY_SIZE(modifiers); i++) + vk_free(pAllocator, modifiers[i]); + fail_register: xcb_unregister_for_special_event(chain->conn, chain->special_event); + wsi_swapchain_finish(&chain->base); + +fail_alloc: vk_free(pAllocator, chain); return result; @@ -854,8 +1469,9 @@ wsi_x11_init_wsi(struct wsi_device *wsi_device, } wsi->base.get_support = x11_surface_get_support; - wsi->base.get_capabilities = x11_surface_get_capabilities; + wsi->base.get_capabilities2 = x11_surface_get_capabilities2; wsi->base.get_formats = x11_surface_get_formats; + wsi->base.get_formats2 = x11_surface_get_formats2; wsi->base.get_present_modes = x11_surface_get_present_modes; wsi->base.create_swapchain = x11_surface_create_swapchain; @@ -883,6 +1499,10 @@ wsi_x11_finish_wsi(struct wsi_device *wsi_device, (struct wsi_x11 *)wsi_device->wsi[VK_ICD_WSI_PLATFORM_XCB]; if (wsi) { + struct hash_entry *entry; + hash_table_foreach(wsi->connections, entry) + wsi_x11_connection_destroy(alloc, entry->data); + _mesa_hash_table_destroy(wsi->connections, NULL); pthread_mutex_destroy(&wsi->mutex);