diff options
author | Jason Ekstrand <[email protected]> | 2017-08-03 14:19:44 -0700 |
---|---|---|
committer | Jason Ekstrand <[email protected]> | 2017-08-28 18:35:27 -0700 |
commit | 92286dc08a58226a2155f07d845978633f23590b (patch) | |
tree | 9a5b5eca876894ed9a5ec42543578b1fed1b6da1 /src/intel | |
parent | 738e5e3c1d261d4f262c2ae25ccca8cb67075d72 (diff) |
anv: Pull the guts of anv_fence into anv_fence_impl
This is just a refactor, similar to what we did for semaphores, in
preparation for handling VK_KHR_external_fence.
Reviewed-by: Lionel Landwerlin <[email protected]>
Diffstat (limited to 'src/intel')
-rw-r--r-- | src/intel/vulkan/anv_batch_chain.c | 22 | ||||
-rw-r--r-- | src/intel/vulkan/anv_private.h | 42 | ||||
-rw-r--r-- | src/intel/vulkan/anv_queue.c | 144 |
3 files changed, 159 insertions, 49 deletions
diff --git a/src/intel/vulkan/anv_batch_chain.c b/src/intel/vulkan/anv_batch_chain.c index ef6ada49ff3..775009c913a 100644 --- a/src/intel/vulkan/anv_batch_chain.c +++ b/src/intel/vulkan/anv_batch_chain.c @@ -1549,10 +1549,20 @@ anv_cmd_buffer_execbuf(struct anv_device *device, } if (fence) { - result = anv_execbuf_add_bo(&execbuf, &fence->bo, NULL, - EXEC_OBJECT_WRITE, &device->alloc); - if (result != VK_SUCCESS) - return result; + assert(fence->temporary.type == ANV_FENCE_TYPE_NONE); + struct anv_fence_impl *impl = &fence->permanent; + + switch (impl->type) { + case ANV_FENCE_TYPE_BO: + result = anv_execbuf_add_bo(&execbuf, &impl->bo.bo, NULL, + EXEC_OBJECT_WRITE, &device->alloc); + if (result != VK_SUCCESS) + return result; + break; + + default: + unreachable("Invalid fence type"); + } } if (cmd_buffer) @@ -1598,7 +1608,7 @@ anv_cmd_buffer_execbuf(struct anv_device *device, anv_semaphore_reset_temporary(device, semaphore); } - if (fence) { + if (fence && fence->permanent.type == ANV_FENCE_TYPE_BO) { /* Once the execbuf has returned, we need to set the fence state to * SUBMITTED. We can't do this before calling execbuf because * anv_GetFenceStatus does take the global device lock before checking @@ -1609,7 +1619,7 @@ anv_cmd_buffer_execbuf(struct anv_device *device, * vkGetFenceStatus() return a valid result (VK_ERROR_DEVICE_LOST or * VK_SUCCESS) in a finite amount of time even if execbuf fails. */ - fence->state = ANV_FENCE_STATE_SUBMITTED; + fence->permanent.bo.state = ANV_FENCE_STATE_SUBMITTED; } if (result == VK_SUCCESS && need_out_fence) { diff --git a/src/intel/vulkan/anv_private.h b/src/intel/vulkan/anv_private.h index 715e0adebfb..ab6e5e25658 100644 --- a/src/intel/vulkan/anv_private.h +++ b/src/intel/vulkan/anv_private.h @@ -1707,6 +1707,12 @@ anv_cmd_buffer_alloc_blorp_binding_table(struct anv_cmd_buffer *cmd_buffer, void anv_cmd_buffer_dump(struct anv_cmd_buffer *cmd_buffer); +enum anv_fence_type { + ANV_FENCE_TYPE_NONE = 0, + ANV_FENCE_TYPE_BO, + ANV_FENCE_TYPE_SYNCOBJ, +}; + enum anv_fence_state { /** Indicates that this is a new (or newly reset fence) */ ANV_FENCE_STATE_RESET, @@ -1719,9 +1725,41 @@ enum anv_fence_state { ANV_FENCE_STATE_SIGNALED, }; +struct anv_fence_impl { + enum anv_fence_type type; + + union { + /** Fence implementation for BO fences + * + * These fences use a BO and a set of CPU-tracked state flags. The BO + * is added to the object list of the last execbuf call in a QueueSubmit + * and is marked EXEC_WRITE. The state flags track when the BO has been + * submitted to the kernel. We need to do this because Vulkan lets you + * wait on a fence that has not yet been submitted and I915_GEM_BUSY + * will say it's idle in this case. + */ + struct { + struct anv_bo bo; + enum anv_fence_state state; + } bo; + }; +}; + struct anv_fence { - struct anv_bo bo; - enum anv_fence_state state; + /* Permanent fence state. Every fence has some form of permanent state + * (type != ANV_SEMAPHORE_TYPE_NONE). This may be a BO to fence on (for + * cross-process fences0 or it could just be a dummy for use internally. + */ + struct anv_fence_impl permanent; + + /* Temporary fence state. A fence *may* have temporary state. That state + * is added to the fence by an import operation and is reset back to + * ANV_SEMAPHORE_TYPE_NONE when the fence is reset. A fence with temporary + * state cannot be signaled because the fence must already be signaled + * before the temporary state can be exported from the fence in the other + * process and imported here. + */ + struct anv_fence_impl temporary; }; struct anv_event { diff --git a/src/intel/vulkan/anv_queue.c b/src/intel/vulkan/anv_queue.c index 5023172250a..f8a2f6448ee 100644 --- a/src/intel/vulkan/anv_queue.c +++ b/src/intel/vulkan/anv_queue.c @@ -262,23 +262,26 @@ VkResult anv_CreateFence( VkFence* pFence) { ANV_FROM_HANDLE(anv_device, device, _device); - struct anv_bo fence_bo; struct anv_fence *fence; assert(pCreateInfo->sType == VK_STRUCTURE_TYPE_FENCE_CREATE_INFO); - VkResult result = anv_bo_pool_alloc(&device->batch_bo_pool, &fence_bo, 4096); + fence = vk_zalloc2(&device->alloc, pAllocator, sizeof(*fence), 8, + VK_SYSTEM_ALLOCATION_SCOPE_OBJECT); + if (fence == NULL) + return vk_error(VK_ERROR_OUT_OF_HOST_MEMORY); + + fence->permanent.type = ANV_FENCE_TYPE_BO; + + VkResult result = anv_bo_pool_alloc(&device->batch_bo_pool, + &fence->permanent.bo.bo, 4096); if (result != VK_SUCCESS) return result; - /* Fences are small. Just store the CPU data structure in the BO. */ - fence = fence_bo.map; - fence->bo = fence_bo; - if (pCreateInfo->flags & VK_FENCE_CREATE_SIGNALED_BIT) { - fence->state = ANV_FENCE_STATE_SIGNALED; + fence->permanent.bo.state = ANV_FENCE_STATE_SIGNALED; } else { - fence->state = ANV_FENCE_STATE_RESET; + fence->permanent.bo.state = ANV_FENCE_STATE_RESET; } *pFence = anv_fence_to_handle(fence); @@ -286,6 +289,23 @@ VkResult anv_CreateFence( return VK_SUCCESS; } +static void +anv_fence_impl_cleanup(struct anv_device *device, + struct anv_fence_impl *impl) +{ + switch (impl->type) { + case ANV_FENCE_TYPE_NONE: + /* Dummy. Nothing to do */ + return; + + case ANV_FENCE_TYPE_BO: + anv_bo_pool_free(&device->batch_bo_pool, &impl->bo.bo); + return; + } + + unreachable("Invalid fence type"); +} + void anv_DestroyFence( VkDevice _device, VkFence _fence, @@ -297,8 +317,10 @@ void anv_DestroyFence( if (!fence) return; - assert(fence->bo.map == fence); - anv_bo_pool_free(&device->batch_bo_pool, &fence->bo); + anv_fence_impl_cleanup(device, &fence->temporary); + anv_fence_impl_cleanup(device, &fence->permanent); + + vk_free2(&device->alloc, pAllocator, fence); } VkResult anv_ResetFences( @@ -308,7 +330,18 @@ VkResult anv_ResetFences( { for (uint32_t i = 0; i < fenceCount; i++) { ANV_FROM_HANDLE(anv_fence, fence, pFences[i]); - fence->state = ANV_FENCE_STATE_RESET; + + assert(fence->temporary.type == ANV_FENCE_TYPE_NONE); + struct anv_fence_impl *impl = &fence->permanent; + + switch (impl->type) { + case ANV_FENCE_TYPE_BO: + impl->bo.state = ANV_FENCE_STATE_RESET; + break; + + default: + unreachable("Invalid fence type"); + } } return VK_SUCCESS; @@ -324,45 +357,50 @@ VkResult anv_GetFenceStatus( if (unlikely(device->lost)) return VK_ERROR_DEVICE_LOST; - switch (fence->state) { - case ANV_FENCE_STATE_RESET: - /* If it hasn't even been sent off to the GPU yet, it's not ready */ - return VK_NOT_READY; - - case ANV_FENCE_STATE_SIGNALED: - /* It's been signaled, return success */ - return VK_SUCCESS; + assert(fence->temporary.type == ANV_FENCE_TYPE_NONE); + struct anv_fence_impl *impl = &fence->permanent; - case ANV_FENCE_STATE_SUBMITTED: { - VkResult result = anv_device_bo_busy(device, &fence->bo); - if (result == VK_SUCCESS) { - fence->state = ANV_FENCE_STATE_SIGNALED; + switch (impl->type) { + case ANV_FENCE_TYPE_BO: + switch (impl->bo.state) { + case ANV_FENCE_STATE_RESET: + /* If it hasn't even been sent off to the GPU yet, it's not ready */ + return VK_NOT_READY; + + case ANV_FENCE_STATE_SIGNALED: + /* It's been signaled, return success */ return VK_SUCCESS; - } else { - return result; + + case ANV_FENCE_STATE_SUBMITTED: { + VkResult result = anv_device_bo_busy(device, &impl->bo.bo); + if (result == VK_SUCCESS) { + impl->bo.state = ANV_FENCE_STATE_SIGNALED; + return VK_SUCCESS; + } else { + return result; + } } - } + default: + unreachable("Invalid fence status"); + } + default: - unreachable("Invalid fence status"); + unreachable("Invalid fence type"); } } #define NSEC_PER_SEC 1000000000 #define INT_TYPE_MAX(type) ((1ull << (sizeof(type) * 8 - 1)) - 1) -VkResult anv_WaitForFences( - VkDevice _device, - uint32_t fenceCount, - const VkFence* pFences, - VkBool32 waitAll, - uint64_t _timeout) +static VkResult +anv_wait_for_bo_fences(struct anv_device *device, + uint32_t fenceCount, + const VkFence *pFences, + bool waitAll, + uint64_t _timeout) { - ANV_FROM_HANDLE(anv_device, device, _device); int ret; - if (unlikely(device->lost)) - return VK_ERROR_DEVICE_LOST; - /* DRM_IOCTL_I915_GEM_WAIT uses a signed 64 bit timeout and is supposed * to block indefinitely timeouts <= 0. Unfortunately, this was broken * for a couple of kernel releases. Since there's no way to know @@ -379,7 +417,16 @@ VkResult anv_WaitForFences( bool signaled_fences = false; for (uint32_t i = 0; i < fenceCount; i++) { ANV_FROM_HANDLE(anv_fence, fence, pFences[i]); - switch (fence->state) { + + /* This function assumes that all fences are BO fences and that they + * have no temporary state. Since BO fences will never be exported, + * this should be a safe assumption. + */ + assert(fence->permanent.type == ANV_FENCE_TYPE_BO); + assert(fence->temporary.type == ANV_FENCE_TYPE_NONE); + struct anv_fence_impl *impl = &fence->permanent; + + switch (impl->bo.state) { case ANV_FENCE_STATE_RESET: /* This fence hasn't been submitted yet, we'll catch it the next * time around. Yes, this may mean we dead-loop but, short of @@ -403,10 +450,10 @@ VkResult anv_WaitForFences( /* These are the fences we really care about. Go ahead and wait * on it until we hit a timeout. */ - result = anv_device_wait(device, &fence->bo, timeout); + result = anv_device_wait(device, &impl->bo.bo, timeout); switch (result) { case VK_SUCCESS: - fence->state = ANV_FENCE_STATE_SIGNALED; + impl->bo.state = ANV_FENCE_STATE_SIGNALED; signaled_fences = true; if (!waitAll) goto done; @@ -436,7 +483,7 @@ VkResult anv_WaitForFences( uint32_t now_pending_fences = 0; for (uint32_t i = 0; i < fenceCount; i++) { ANV_FROM_HANDLE(anv_fence, fence, pFences[i]); - if (fence->state == ANV_FENCE_STATE_RESET) + if (fence->permanent.bo.state == ANV_FENCE_STATE_RESET) now_pending_fences++; } assert(now_pending_fences <= pending_fences); @@ -487,6 +534,21 @@ done: return result; } +VkResult anv_WaitForFences( + VkDevice _device, + uint32_t fenceCount, + const VkFence* pFences, + VkBool32 waitAll, + uint64_t timeout) +{ + ANV_FROM_HANDLE(anv_device, device, _device); + + if (unlikely(device->lost)) + return VK_ERROR_DEVICE_LOST; + + return anv_wait_for_bo_fences(device, fenceCount, pFences, waitAll, timeout); +} + // Queue semaphore functions VkResult anv_CreateSemaphore( |