diff options
author | Marek Olšák <[email protected]> | 2018-08-31 20:58:10 -0400 |
---|---|---|
committer | Marek Olšák <[email protected]> | 2018-09-07 17:59:02 -0400 |
commit | 21ca322e637291b89a445159fc45b8dbf638e6c9 (patch) | |
tree | c0f2417651f7c6dba6c0ec9da1ee6755132e7181 /src/gallium/auxiliary | |
parent | 9ce2cef68f37a05cd2e4e4128395af38e7be963f (diff) |
st/mesa: throttle texture uploads if their memory usage goes beyond a limit
This prevents radeonsi from running out of memory. It also increases
texture upload performance by being nice to the kernel memory manager.
Diffstat (limited to 'src/gallium/auxiliary')
-rw-r--r-- | src/gallium/auxiliary/util/u_helpers.c | 120 | ||||
-rw-r--r-- | src/gallium/auxiliary/util/u_helpers.h | 17 |
2 files changed, 137 insertions, 0 deletions
diff --git a/src/gallium/auxiliary/util/u_helpers.c b/src/gallium/auxiliary/util/u_helpers.c index 7d45b2f06db..f773360adde 100644 --- a/src/gallium/auxiliary/util/u_helpers.c +++ b/src/gallium/auxiliary/util/u_helpers.c @@ -221,3 +221,123 @@ util_wait_for_idle(struct pipe_context *ctx) ctx->flush(ctx, &fence, 0); ctx->screen->fence_finish(ctx->screen, NULL, fence, PIPE_TIMEOUT_INFINITE); } + +void +util_throttle_init(struct util_throttle *t, uint64_t max_mem_usage) +{ + t->max_mem_usage = max_mem_usage; +} + +void +util_throttle_deinit(struct pipe_screen *screen, struct util_throttle *t) +{ + for (unsigned i = 0; i < ARRAY_SIZE(t->ring); i++) + screen->fence_reference(screen, &t->ring[i].fence, NULL); +} + +static uint64_t +util_get_throttle_total_memory_usage(struct util_throttle *t) +{ + uint64_t total_usage = 0; + + for (unsigned i = 0; i < ARRAY_SIZE(t->ring); i++) + total_usage += t->ring[i].mem_usage; + return total_usage; +} + +static void util_dump_throttle_ring(struct util_throttle *t) +{ + printf("Throttle:\n"); + for (unsigned i = 0; i < ARRAY_SIZE(t->ring); i++) { + printf(" ring[%u]: fence = %s, mem_usage = %"PRIu64"%s%s\n", + i, t->ring[i].fence ? "yes" : " no", + t->ring[i].mem_usage, + t->flush_index == i ? " [flush]" : "", + t->wait_index == i ? " [wait]" : ""); + } +} + +/** + * Notify util_throttle that the next operation allocates memory. + * util_throttle tracks memory usage and waits for fences until its tracked + * memory usage decreases. + * + * Example: + * util_throttle_memory_usage(..., w*h*d*Bpp); + * TexSubImage(..., w, h, d, ...); + * + * This means that TexSubImage can't allocate more memory its maximum limit + * set during initialization. + */ +void +util_throttle_memory_usage(struct pipe_context *pipe, + struct util_throttle *t, uint64_t memory_size) +{ + (void)util_dump_throttle_ring; /* silence warning */ + + if (!t->max_mem_usage) + return; + + struct pipe_screen *screen = pipe->screen; + struct pipe_fence_handle **fence = NULL; + unsigned ring_size = ARRAY_SIZE(t->ring); + uint64_t total = util_get_throttle_total_memory_usage(t); + + /* If there is not enough memory, walk the list of fences and find + * the latest one that we need to wait for. + */ + while (t->wait_index != t->flush_index && + total && total + memory_size > t->max_mem_usage) { + assert(t->ring[t->wait_index].fence); + + /* Release an older fence if we need to wait for a newer one. */ + if (fence) + screen->fence_reference(screen, fence, NULL); + + fence = &t->ring[t->wait_index].fence; + t->ring[t->wait_index].mem_usage = 0; + t->wait_index = (t->wait_index + 1) % ring_size; + + total = util_get_throttle_total_memory_usage(t); + } + + /* Wait for the fence to decrease memory usage. */ + if (fence) { + screen->fence_finish(screen, pipe, *fence, PIPE_TIMEOUT_INFINITE); + screen->fence_reference(screen, fence, NULL); + } + + /* Flush and get a fence if we've exhausted memory usage for the current + * slot. + */ + if (t->ring[t->flush_index].mem_usage && + t->ring[t->flush_index].mem_usage + memory_size > + t->max_mem_usage / (ring_size / 2)) { + struct pipe_fence_handle **fence = + &t->ring[t->flush_index].fence; + + /* Expect that the current flush slot doesn't have a fence yet. */ + assert(!*fence); + + pipe->flush(pipe, fence, PIPE_FLUSH_ASYNC); + t->flush_index = (t->flush_index + 1) % ring_size; + + /* Vacate the next slot if it's occupied. This should be rare. */ + if (t->flush_index == t->wait_index) { + struct pipe_fence_handle **fence = + &t->ring[t->wait_index].fence; + + t->ring[t->wait_index].mem_usage = 0; + t->wait_index = (t->wait_index + 1) % ring_size; + + assert(*fence); + screen->fence_finish(screen, pipe, *fence, PIPE_TIMEOUT_INFINITE); + screen->fence_reference(screen, fence, NULL); + } + + assert(!t->ring[t->flush_index].mem_usage); + assert(!t->ring[t->flush_index].fence); + } + + t->ring[t->flush_index].mem_usage += memory_size; +} diff --git a/src/gallium/auxiliary/util/u_helpers.h b/src/gallium/auxiliary/util/u_helpers.h index f00436e3b11..38c47c1cc98 100644 --- a/src/gallium/auxiliary/util/u_helpers.h +++ b/src/gallium/auxiliary/util/u_helpers.h @@ -64,6 +64,23 @@ util_end_pipestat_query(struct pipe_context *ctx, struct pipe_query *q, void util_wait_for_idle(struct pipe_context *ctx); +/* A utility for throttling execution based on memory usage. */ +struct util_throttle { + struct { + struct pipe_fence_handle *fence; + uint64_t mem_usage; + } ring[10]; + + unsigned flush_index; + unsigned wait_index; + uint64_t max_mem_usage; +}; + +void util_throttle_init(struct util_throttle *t, uint64_t max_mem_usage); +void util_throttle_deinit(struct pipe_screen *screen, struct util_throttle *t); +void util_throttle_memory_usage(struct pipe_context *pipe, + struct util_throttle *t, uint64_t memory_size); + #ifdef __cplusplus } #endif |