summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorJason Ekstrand <[email protected]>2016-03-07 13:45:25 -0800
committerJason Ekstrand <[email protected]>2016-03-07 22:23:44 -0800
commit2308891edea4d8508d3e95f29c58b4089e96b5e7 (patch)
treed549cc19b88686cd7495003e6bffa7e8f478d0f9
parentf61d40adc2a09221453b7a87880e134a5424773e (diff)
anv: Store CPU-side fence information in the BO
This reduces the number of allocations a bit and cuts back on memory usage. Kind-of a micro-optimization but it also makes the error handling a bit simpler so it seems like a win.
-rw-r--r--src/intel/vulkan/anv_device.c33
1 files changed, 15 insertions, 18 deletions
diff --git a/src/intel/vulkan/anv_device.c b/src/intel/vulkan/anv_device.c
index 44eb0ed2d6c..768e2eb3be1 100644
--- a/src/intel/vulkan/anv_device.c
+++ b/src/intel/vulkan/anv_device.c
@@ -1383,29 +1383,31 @@ VkResult anv_CreateFence(
VkFence* pFence)
{
ANV_FROM_HANDLE(anv_device, device, _device);
+ struct anv_bo fence_bo;
struct anv_fence *fence;
struct anv_batch batch;
VkResult result;
assert(pCreateInfo->sType == VK_STRUCTURE_TYPE_FENCE_CREATE_INFO);
- fence = anv_alloc2(&device->alloc, pAllocator, sizeof(*fence), 8,
- VK_SYSTEM_ALLOCATION_SCOPE_OBJECT);
- if (fence == NULL)
- return vk_error(VK_ERROR_OUT_OF_HOST_MEMORY);
-
- result = anv_bo_pool_alloc(&device->batch_bo_pool, &fence->bo);
+ result = anv_bo_pool_alloc(&device->batch_bo_pool, &fence_bo);
if (result != VK_SUCCESS)
- goto fail;
+ return result;
+
+ /* Fences are small. Just store the CPU data structure in the BO. */
+ fence = fence_bo.map;
+ fence->bo = fence_bo;
- batch.next = batch.start = fence->bo.map;
+ /* Place the batch after the CPU data but on its own cache line. */
+ const uint32_t batch_offset = align_u32(sizeof(*fence), CACHELINE_SIZE);
+ batch.next = batch.start = fence->bo.map + batch_offset;
batch.end = fence->bo.map + fence->bo.size;
anv_batch_emit(&batch, GEN7_MI_BATCH_BUFFER_END);
anv_batch_emit(&batch, GEN7_MI_NOOP);
if (!device->info.has_llc) {
- assert(((uintptr_t) fence->bo.map & CACHELINE_MASK) == 0);
- assert(batch.next - fence->bo.map <= CACHELINE_SIZE);
+ assert(((uintptr_t) batch.start & CACHELINE_MASK) == 0);
+ assert(batch.next - batch.start <= CACHELINE_SIZE);
__builtin_ia32_mfence();
__builtin_ia32_clflush(fence->bo.map);
}
@@ -1421,8 +1423,8 @@ VkResult anv_CreateFence(
fence->execbuf.buffers_ptr = (uintptr_t) fence->exec2_objects;
fence->execbuf.buffer_count = 1;
- fence->execbuf.batch_start_offset = 0;
- fence->execbuf.batch_len = batch.next - fence->bo.map;
+ fence->execbuf.batch_start_offset = batch.start - fence->bo.map;
+ fence->execbuf.batch_len = batch.next - batch.start;
fence->execbuf.cliprects_ptr = 0;
fence->execbuf.num_cliprects = 0;
fence->execbuf.DR1 = 0;
@@ -1438,11 +1440,6 @@ VkResult anv_CreateFence(
*pFence = anv_fence_to_handle(fence);
return VK_SUCCESS;
-
- fail:
- anv_free2(&device->alloc, pAllocator, fence);
-
- return result;
}
void anv_DestroyFence(
@@ -1453,8 +1450,8 @@ void anv_DestroyFence(
ANV_FROM_HANDLE(anv_device, device, _device);
ANV_FROM_HANDLE(anv_fence, fence, _fence);
+ assert(fence->bo.map == fence);
anv_bo_pool_free(&device->batch_bo_pool, &fence->bo);
- anv_free2(&device->alloc, pAllocator, fence);
}
VkResult anv_ResetFences(