summaryrefslogtreecommitdiffstats
path: root/src/gallium/drivers/r600/r600_pipe.c
diff options
context:
space:
mode:
Diffstat (limited to 'src/gallium/drivers/r600/r600_pipe.c')
-rw-r--r--src/gallium/drivers/r600/r600_pipe.c173
1 files changed, 2 insertions, 171 deletions
diff --git a/src/gallium/drivers/r600/r600_pipe.c b/src/gallium/drivers/r600/r600_pipe.c
index e091b084a15..9da42d17d22 100644
--- a/src/gallium/drivers/r600/r600_pipe.c
+++ b/src/gallium/drivers/r600/r600_pipe.c
@@ -68,81 +68,6 @@ static const struct debug_named_value r600_debug_options[] = {
/*
* pipe_context
*/
-static struct r600_fence *r600_create_fence(struct r600_context *rctx)
-{
- struct r600_screen *rscreen = rctx->screen;
- struct r600_fence *fence = NULL;
-
- pipe_mutex_lock(rscreen->fences.mutex);
-
- if (!rscreen->fences.bo) {
- /* Create the shared buffer object */
- rscreen->fences.bo = (struct r600_resource*)
- pipe_buffer_create(&rscreen->b.b, PIPE_BIND_CUSTOM,
- PIPE_USAGE_STAGING, 4096);
- if (!rscreen->fences.bo) {
- R600_ERR("r600: failed to create bo for fence objects\n");
- goto out;
- }
- rscreen->fences.data = r600_buffer_map_sync_with_rings(&rctx->b, rscreen->fences.bo, PIPE_TRANSFER_READ_WRITE);
- }
-
- if (!LIST_IS_EMPTY(&rscreen->fences.pool)) {
- struct r600_fence *entry;
-
- /* Try to find a freed fence that has been signalled */
- LIST_FOR_EACH_ENTRY(entry, &rscreen->fences.pool, head) {
- if (rscreen->fences.data[entry->index] != 0) {
- LIST_DELINIT(&entry->head);
- fence = entry;
- break;
- }
- }
- }
-
- if (!fence) {
- /* Allocate a new fence */
- struct r600_fence_block *block;
- unsigned index;
-
- if ((rscreen->fences.next_index + 1) >= 1024) {
- R600_ERR("r600: too many concurrent fences\n");
- goto out;
- }
-
- index = rscreen->fences.next_index++;
-
- if (!(index % FENCE_BLOCK_SIZE)) {
- /* Allocate a new block */
- block = CALLOC_STRUCT(r600_fence_block);
- if (block == NULL)
- goto out;
-
- LIST_ADD(&block->head, &rscreen->fences.blocks);
- } else {
- block = LIST_ENTRY(struct r600_fence_block, rscreen->fences.blocks.next, head);
- }
-
- fence = &block->fences[index % FENCE_BLOCK_SIZE];
- fence->index = index;
- }
-
- pipe_reference_init(&fence->reference, 1);
-
- rscreen->fences.data[fence->index] = 0;
- r600_context_emit_fence(rctx, rscreen->fences.bo, fence->index, 1);
-
- /* Create a dummy BO so that fence_finish without a timeout can sleep waiting for completion */
- fence->sleep_bo = (struct r600_resource*)
- pipe_buffer_create(&rctx->screen->b.b, PIPE_BIND_CUSTOM,
- PIPE_USAGE_STAGING, 1);
- /* Add the fence as a dummy relocation. */
- r600_context_bo_reloc(&rctx->b, &rctx->b.rings.gfx, fence->sleep_bo, RADEON_USAGE_READWRITE);
-
-out:
- pipe_mutex_unlock(rscreen->fences.mutex);
- return fence;
-}
static void r600_flush(struct pipe_context *ctx, unsigned flags)
{
@@ -180,12 +105,11 @@ static void r600_flush_from_st(struct pipe_context *ctx,
unsigned flags)
{
struct r600_context *rctx = (struct r600_context *)ctx;
- struct r600_fence **rfence = (struct r600_fence**)fence;
unsigned fflags;
fflags = flags & PIPE_FLUSH_END_OF_FRAME ? RADEON_FLUSH_END_OF_FRAME : 0;
- if (rfence) {
- *rfence = r600_create_fence(rctx);
+ if (fence) {
+ *fence = rctx->b.ws->cs_create_fence(rctx->b.rings.gfx.cs);
}
/* flush gfx & dma ring, order does not matter as only one can be live */
if (rctx->b.rings.dma.cs) {
@@ -888,98 +812,15 @@ static void r600_destroy_screen(struct pipe_screen* pscreen)
compute_memory_pool_delete(rscreen->global_pool);
}
- if (rscreen->fences.bo) {
- struct r600_fence_block *entry, *tmp;
-
- LIST_FOR_EACH_ENTRY_SAFE(entry, tmp, &rscreen->fences.blocks, head) {
- LIST_DEL(&entry->head);
- FREE(entry);
- }
-
- rscreen->b.ws->buffer_unmap(rscreen->fences.bo->cs_buf);
- pipe_resource_reference((struct pipe_resource**)&rscreen->fences.bo, NULL);
- }
if (rscreen->trace_bo) {
rscreen->b.ws->buffer_unmap(rscreen->trace_bo->cs_buf);
pipe_resource_reference((struct pipe_resource**)&rscreen->trace_bo, NULL);
}
- pipe_mutex_destroy(rscreen->fences.mutex);
rscreen->b.ws->destroy(rscreen->b.ws);
FREE(rscreen);
}
-static void r600_fence_reference(struct pipe_screen *pscreen,
- struct pipe_fence_handle **ptr,
- struct pipe_fence_handle *fence)
-{
- struct r600_fence **oldf = (struct r600_fence**)ptr;
- struct r600_fence *newf = (struct r600_fence*)fence;
-
- if (pipe_reference(&(*oldf)->reference, &newf->reference)) {
- struct r600_screen *rscreen = (struct r600_screen *)pscreen;
- pipe_mutex_lock(rscreen->fences.mutex);
- pipe_resource_reference((struct pipe_resource**)&(*oldf)->sleep_bo, NULL);
- LIST_ADDTAIL(&(*oldf)->head, &rscreen->fences.pool);
- pipe_mutex_unlock(rscreen->fences.mutex);
- }
-
- *ptr = fence;
-}
-
-static boolean r600_fence_signalled(struct pipe_screen *pscreen,
- struct pipe_fence_handle *fence)
-{
- struct r600_screen *rscreen = (struct r600_screen *)pscreen;
- struct r600_fence *rfence = (struct r600_fence*)fence;
-
- return rscreen->fences.data[rfence->index] != 0;
-}
-
-static boolean r600_fence_finish(struct pipe_screen *pscreen,
- struct pipe_fence_handle *fence,
- uint64_t timeout)
-{
- struct r600_screen *rscreen = (struct r600_screen *)pscreen;
- struct r600_fence *rfence = (struct r600_fence*)fence;
- int64_t start_time = 0;
- unsigned spins = 0;
-
- if (timeout != PIPE_TIMEOUT_INFINITE) {
- start_time = os_time_get();
-
- /* Convert to microseconds. */
- timeout /= 1000;
- }
-
- while (rscreen->fences.data[rfence->index] == 0) {
- /* Special-case infinite timeout - wait for the dummy BO to become idle */
- if (timeout == PIPE_TIMEOUT_INFINITE) {
- rscreen->b.ws->buffer_wait(rfence->sleep_bo->buf, RADEON_USAGE_READWRITE);
- break;
- }
-
- /* The dummy BO will be busy until the CS including the fence has completed, or
- * the GPU is reset. Don't bother continuing to spin when the BO is idle. */
- if (!rscreen->b.ws->buffer_is_busy(rfence->sleep_bo->buf, RADEON_USAGE_READWRITE))
- break;
-
- if (++spins % 256)
- continue;
-#ifdef PIPE_OS_UNIX
- sched_yield();
-#else
- os_time_sleep(10);
-#endif
- if (timeout != PIPE_TIMEOUT_INFINITE &&
- os_time_get() - start_time >= timeout) {
- break;
- }
- }
-
- return rscreen->fences.data[rfence->index] != 0;
-}
-
static uint64_t r600_get_timestamp(struct pipe_screen *screen)
{
struct r600_screen *rscreen = (struct r600_screen*)screen;
@@ -1035,9 +876,6 @@ struct pipe_screen *r600_screen_create(struct radeon_winsys *ws)
} else {
rscreen->b.b.is_format_supported = r600_is_format_supported;
}
- rscreen->b.b.fence_reference = r600_fence_reference;
- rscreen->b.b.fence_signalled = r600_fence_signalled;
- rscreen->b.b.fence_finish = r600_fence_finish;
rscreen->b.b.get_driver_query_info = r600_get_driver_query_info;
if (rscreen->b.info.has_uvd) {
rscreen->b.b.get_video_param = ruvd_get_video_param;
@@ -1113,13 +951,6 @@ struct pipe_screen *r600_screen_create(struct radeon_winsys *ws)
rscreen->has_cp_dma = rscreen->b.info.drm_minor >= 27 &&
!(rscreen->b.debug_flags & DBG_NO_CP_DMA);
- rscreen->fences.bo = NULL;
- rscreen->fences.data = NULL;
- rscreen->fences.next_index = 0;
- LIST_INITHEAD(&rscreen->fences.pool);
- LIST_INITHEAD(&rscreen->fences.blocks);
- pipe_mutex_init(rscreen->fences.mutex);
-
rscreen->global_pool = compute_memory_pool_new(rscreen);
rscreen->cs_count = 0;