--- /dev/null
+/*
+ * Copyright © 2017, Google Inc.
+ *
+ * Permission is hereby granted, free of charge, to any person obtaining a
+ * copy of this software and associated documentation files (the "Software"),
+ * to deal in the Software without restriction, including without limitation
+ * the rights to use, copy, modify, merge, publish, distribute, sublicense,
+ * and/or sell copies of the Software, and to permit persons to whom the
+ * Software is furnished to do so, subject to the following conditions:
+ *
+ * The above copyright notice and this permission notice (including the next
+ * paragraph) shall be included in all copies or substantial portions of the
+ * Software.
+ *
+ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
+ * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
+ * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
+ * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
+ * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
+ * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
+ * IN THE SOFTWARE.
+ */
+
+#include <hardware/gralloc.h>
+#include <hardware/hardware.h>
+#include <hardware/hwvulkan.h>
+#include <vulkan/vk_android_native_buffer.h>
+#include <vulkan/vk_icd.h>
+#include <sync/sync.h>
+
+#include "anv_private.h"
+
+static int anv_hal_open(const struct hw_module_t* mod, const char* id, struct hw_device_t** dev);
+static int anv_hal_close(struct hw_device_t *dev);
+
+static void UNUSED
+static_asserts(void)
+{
+ STATIC_ASSERT(HWVULKAN_DISPATCH_MAGIC == ICD_LOADER_MAGIC);
+}
+
+PUBLIC struct hwvulkan_module_t HAL_MODULE_INFO_SYM = {
+ .common = {
+ .tag = HARDWARE_MODULE_TAG,
+ .module_api_version = HWVULKAN_MODULE_API_VERSION_0_1,
+ .hal_api_version = HARDWARE_MAKE_API_VERSION(1, 0),
+ .id = HWVULKAN_HARDWARE_MODULE_ID,
+ .name = "Intel Vulkan HAL",
+ .author = "Intel",
+ .methods = &(hw_module_methods_t) {
+ .open = anv_hal_open,
+ },
+ },
+};
+
+/* If any bits in test_mask are set, then unset them and return true. */
+static inline bool
+unmask32(uint32_t *inout_mask, uint32_t test_mask)
+{
+ uint32_t orig_mask = *inout_mask;
+ *inout_mask &= ~test_mask;
+ return *inout_mask != orig_mask;
+}
+
+static int
+anv_hal_open(const struct hw_module_t* mod, const char* id,
+ struct hw_device_t** dev)
+{
+ assert(mod == &HAL_MODULE_INFO_SYM.common);
+ assert(strcmp(id, HWVULKAN_DEVICE_0) == 0);
+
+ hwvulkan_device_t *hal_dev = malloc(sizeof(*hal_dev));
+ if (!hal_dev)
+ return -1;
+
+ *hal_dev = (hwvulkan_device_t) {
+ .common = {
+ .tag = HARDWARE_DEVICE_TAG,
+ .version = HWVULKAN_DEVICE_API_VERSION_0_1,
+ .module = &HAL_MODULE_INFO_SYM.common,
+ .close = anv_hal_close,
+ },
+ .EnumerateInstanceExtensionProperties = anv_EnumerateInstanceExtensionProperties,
+ .CreateInstance = anv_CreateInstance,
+ .GetInstanceProcAddr = anv_GetInstanceProcAddr,
+ };
+
+ *dev = &hal_dev->common;
+ return 0;
+}
+
+static int
+anv_hal_close(struct hw_device_t *dev)
+{
+ /* hwvulkan.h claims that hw_device_t::close() is never called. */
+ return -1;
+}
+
+VkResult
+anv_image_from_gralloc(VkDevice device_h,
+ const VkImageCreateInfo *base_info,
+ const VkNativeBufferANDROID *gralloc_info,
+ const VkAllocationCallbacks *alloc,
+ VkImage *out_image_h)
+
+{
+ ANV_FROM_HANDLE(anv_device, device, device_h);
+ VkImage image_h = VK_NULL_HANDLE;
+ struct anv_image *image = NULL;
+ struct anv_bo *bo = NULL;
+ VkResult result;
+
+ struct anv_image_create_info anv_info = {
+ .vk_info = base_info,
+ .isl_extra_usage_flags = ISL_SURF_USAGE_DISABLE_AUX_BIT,
+ };
+
+ if (gralloc_info->handle->numFds != 1) {
+ return vk_errorf(device->instance, device,
+ VK_ERROR_INVALID_EXTERNAL_HANDLE_KHR,
+ "VkNativeBufferANDROID::handle::numFds is %d, "
+ "expected 1", gralloc_info->handle->numFds);
+ }
+
+ /* Do not close the gralloc handle's dma_buf. The lifetime of the dma_buf
+ * must exceed that of the gralloc handle, and we do not own the gralloc
+ * handle.
+ */
+ int dma_buf = gralloc_info->handle->data[0];
+
+ result = anv_bo_cache_import(device, &device->bo_cache, dma_buf, &bo);
+ if (result != VK_SUCCESS) {
+ return vk_errorf(device->instance, device, result,
+ "failed to import dma-buf from VkNativeBufferANDROID");
+ }
+
+ int i915_tiling = anv_gem_get_tiling(device, bo->gem_handle);
+ switch (i915_tiling) {
+ case I915_TILING_NONE:
+ anv_info.isl_tiling_flags = ISL_TILING_LINEAR_BIT;
+ break;
+ case I915_TILING_X:
+ anv_info.isl_tiling_flags = ISL_TILING_X_BIT;
+ break;
+ case I915_TILING_Y:
+ anv_info.isl_tiling_flags = ISL_TILING_Y0_BIT;
+ break;
+ case -1:
+ result = vk_errorf(device->instance, device,
+ VK_ERROR_INVALID_EXTERNAL_HANDLE_KHR,
+ "DRM_IOCTL_I915_GEM_GET_TILING failed for "
+ "VkNativeBufferANDROID");
+ goto fail_tiling;
+ default:
+ result = vk_errorf(device->instance, device,
+ VK_ERROR_INVALID_EXTERNAL_HANDLE_KHR,
+ "DRM_IOCTL_I915_GEM_GET_TILING returned unknown "
+ "tiling %d for VkNativeBufferANDROID", i915_tiling);
+ goto fail_tiling;
+ }
+
+ enum isl_format format = anv_get_isl_format(&device->info,
+ base_info->format,
+ VK_IMAGE_ASPECT_COLOR_BIT,
+ base_info->tiling);
+ assert(format != ISL_FORMAT_UNSUPPORTED);
+
+ anv_info.stride = gralloc_info->stride *
+ (isl_format_get_layout(format)->bpb / 8);
+
+ result = anv_image_create(device_h, &anv_info, alloc, &image_h);
+ image = anv_image_from_handle(image_h);
+ if (result != VK_SUCCESS)
+ goto fail_create;
+
+ if (bo->size < image->size) {
+ result = vk_errorf(device, device->instance,
+ VK_ERROR_INVALID_EXTERNAL_HANDLE_KHR,
+ "dma-buf from VkNativeBufferANDROID is too small for "
+ "VkImage: %"PRIu64"B < %"PRIu64"B",
+ bo->size, image->size);
+ goto fail_size;
+ }
+
+ assert(image->n_planes == 1);
+ assert(image->planes[0].bo_offset == 0);
+
+ image->planes[0].bo = bo;
+ image->planes[0].bo_is_owned = true;
+
+ /* We need to set the WRITE flag on window system buffers so that GEM will
+ * know we're writing to them and synchronize uses on other rings (for
+ * example, if the display server uses the blitter ring).
+ *
+ * If this function fails and if the imported bo was resident in the cache,
+ * we should avoid updating the bo's flags. Therefore, we defer updating
+ * the flags until success is certain.
+ *
+ */
+ bo->flags &= ~EXEC_OBJECT_ASYNC;
+ bo->flags |= EXEC_OBJECT_WRITE;
+
+ /* Don't clobber the out-parameter until success is certain. */
+ *out_image_h = image_h;
+
+ return VK_SUCCESS;
+
+ fail_size:
+ anv_DestroyImage(device_h, image_h, alloc);
+ fail_create:
+ fail_tiling:
+ anv_bo_cache_release(device, &device->bo_cache, bo);
+
+ return result;
+}
+
+VkResult anv_GetSwapchainGrallocUsageANDROID(
+ VkDevice device_h,
+ VkFormat format,
+ VkImageUsageFlags imageUsage,
+ int* grallocUsage)
+{
+ ANV_FROM_HANDLE(anv_device, device, device_h);
+ struct anv_physical_device *phys_dev = &device->instance->physicalDevice;
+ VkPhysicalDevice phys_dev_h = anv_physical_device_to_handle(phys_dev);
+ VkResult result;
+
+ *grallocUsage = 0;
+ intel_logd("%s: format=%d, usage=0x%x", __func__, format, imageUsage);
+
+ /* WARNING: Android Nougat's libvulkan.so hardcodes the VkImageUsageFlags
+ * returned to applications via VkSurfaceCapabilitiesKHR::supportedUsageFlags.
+ * The relevant code in libvulkan/swapchain.cpp contains this fun comment:
+ *
+ * TODO(jessehall): I think these are right, but haven't thought hard
+ * about it. Do we need to query the driver for support of any of
+ * these?
+ *
+ * Any disagreement between this function and the hardcoded
+ * VkSurfaceCapabilitiesKHR:supportedUsageFlags causes tests
+ * dEQP-VK.wsi.android.swapchain.*.image_usage to fail.
+ */
+
+ const VkPhysicalDeviceImageFormatInfo2KHR image_format_info = {
+ .sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_IMAGE_FORMAT_INFO_2_KHR,
+ .format = format,
+ .type = VK_IMAGE_TYPE_2D,
+ .tiling = VK_IMAGE_TILING_OPTIMAL,
+ .usage = imageUsage,
+ };
+
+ VkImageFormatProperties2KHR image_format_props = {
+ .sType = VK_STRUCTURE_TYPE_IMAGE_FORMAT_PROPERTIES_2_KHR,
+ };
+
+ /* Check that requested format and usage are supported. */
+ result = anv_GetPhysicalDeviceImageFormatProperties2KHR(phys_dev_h,
+ &image_format_info, &image_format_props);
+ if (result != VK_SUCCESS) {
+ return vk_errorf(device->instance, device, result,
+ "anv_GetPhysicalDeviceImageFormatProperties2KHR failed "
+ "inside %s", __func__);
+ }
+
+ /* Reject STORAGE here to avoid complexity elsewhere. */
+ if (imageUsage & VK_IMAGE_USAGE_STORAGE_BIT) {
+ return vk_errorf(device->instance, device, VK_ERROR_FORMAT_NOT_SUPPORTED,
+ "VK_IMAGE_USAGE_STORAGE_BIT unsupported for gralloc "
+ "swapchain");
+ }
+
+ if (unmask32(&imageUsage, VK_IMAGE_USAGE_TRANSFER_DST_BIT |
+ VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT))
+ *grallocUsage |= GRALLOC_USAGE_HW_RENDER;
+
+ if (unmask32(&imageUsage, VK_IMAGE_USAGE_TRANSFER_SRC_BIT |
+ VK_IMAGE_USAGE_SAMPLED_BIT |
+ VK_IMAGE_USAGE_INPUT_ATTACHMENT_BIT))
+ *grallocUsage |= GRALLOC_USAGE_HW_TEXTURE;
+
+ /* All VkImageUsageFlags not explicitly checked here are unsupported for
+ * gralloc swapchains.
+ */
+ if (imageUsage != 0) {
+ return vk_errorf(device->instance, device, VK_ERROR_FORMAT_NOT_SUPPORTED,
+ "unsupported VkImageUsageFlags(0x%x) for gralloc "
+ "swapchain", imageUsage);
+ }
+
+ /* The below formats support GRALLOC_USAGE_HW_FB (that is, display
+ * scanout). This short list of formats is univserally supported on Intel
+ * but is incomplete. The full set of supported formats is dependent on
+ * kernel and hardware.
+ *
+ * FINISHME: Advertise all display-supported formats.
+ */
+ if (format == VK_FORMAT_B8G8R8A8_UNORM ||
+ format == VK_FORMAT_B5G6R5_UNORM_PACK16) {
+ *grallocUsage |= GRALLOC_USAGE_HW_FB |
+ GRALLOC_USAGE_HW_COMPOSER |
+ GRALLOC_USAGE_EXTERNAL_DISP;
+ }
+
+ if (*grallocUsage == 0)
+ return VK_ERROR_FORMAT_NOT_SUPPORTED;
+
+ return VK_SUCCESS;
+}
+
+VkResult
+anv_AcquireImageANDROID(
+ VkDevice device_h,
+ VkImage image_h,
+ int nativeFenceFd,
+ VkSemaphore semaphore_h,
+ VkFence fence_h)
+{
+ ANV_FROM_HANDLE(anv_device, device, device_h);
+ VkResult result = VK_SUCCESS;
+
+ if (nativeFenceFd != -1) {
+ /* As a simple, firstpass implementation of VK_ANDROID_native_buffer, we
+ * block on the nativeFenceFd. This may introduce latency and is
+ * definitiely inefficient, yet it's correct.
+ *
+ * FINISHME(chadv): Import the nativeFenceFd into the VkSemaphore and
+ * VkFence.
+ */
+ if (sync_wait(nativeFenceFd, /*timeout*/ -1) < 0) {
+ result = vk_errorf(device->instance, device, VK_ERROR_DEVICE_LOST,
+ "%s: failed to wait on nativeFenceFd=%d",
+ __func__, nativeFenceFd);
+ }
+
+ /* From VK_ANDROID_native_buffer's pseudo spec
+ * (https://source.android.com/devices/graphics/implement-vulkan):
+ *
+ * The driver takes ownership of the fence fd and is responsible for
+ * closing it [...] even if vkAcquireImageANDROID fails and returns
+ * an error.
+ */
+ close(nativeFenceFd);
+
+ if (result != VK_SUCCESS)
+ return result;
+ }
+
+ if (semaphore_h || fence_h) {
+ /* Thanks to implicit sync, the image is ready for GPU access. But we
+ * must still put the semaphore into the "submit" state; otherwise the
+ * client may get unexpected behavior if the client later uses it as
+ * a wait semaphore.
+ *
+ * Because we blocked above on the nativeFenceFd, the image is also
+ * ready for foreign-device access (including CPU access). But we must
+ * still signal the fence; otherwise the client may get unexpected
+ * behavior if the client later waits on it.
+ *
+ * For some values of anv_semaphore_type, we must submit the semaphore
+ * to execbuf in order to signal it. Likewise for anv_fence_type.
+ * Instead of open-coding here the signal operation for each
+ * anv_semaphore_type and anv_fence_type, we piggy-back on
+ * vkQueueSubmit.
+ */
+ const VkSubmitInfo submit = {
+ .sType = VK_STRUCTURE_TYPE_SUBMIT_INFO,
+ .waitSemaphoreCount = 0,
+ .commandBufferCount = 0,
+ .signalSemaphoreCount = (semaphore_h ? 1 : 0),
+ .pSignalSemaphores = &semaphore_h,
+ };
+
+ result = anv_QueueSubmit(anv_queue_to_handle(&device->queue), 1,
+ &submit, fence_h);
+ if (result != VK_SUCCESS) {
+ return vk_errorf(device->instance, device, result,
+ "anv_QueueSubmit failed inside %s", __func__);
+ }
+ }
+
+ return VK_SUCCESS;
+}
+
+VkResult
+anv_QueueSignalReleaseImageANDROID(
+ VkQueue queue,
+ uint32_t waitSemaphoreCount,
+ const VkSemaphore* pWaitSemaphores,
+ VkImage image,
+ int* pNativeFenceFd)
+{
+ VkResult result;
+
+ if (waitSemaphoreCount == 0)
+ goto done;
+
+ result = anv_QueueSubmit(queue, 1,
+ &(VkSubmitInfo) {
+ .sType = VK_STRUCTURE_TYPE_SUBMIT_INFO,
+ .waitSemaphoreCount = 1,
+ .pWaitSemaphores = pWaitSemaphores,
+ },
+ (VkFence) VK_NULL_HANDLE);
+ if (result != VK_SUCCESS)
+ return result;
+
+ done:
+ if (pNativeFenceFd) {
+ /* We can rely implicit on sync because above we submitted all
+ * semaphores to the queue.
+ */
+ *pNativeFenceFd = -1;
+ }
+
+ return VK_SUCCESS;
+}