diff options
author | Kenneth Graunke <[email protected]> | 2019-04-08 00:45:41 -0700 |
---|---|---|
committer | Kenneth Graunke <[email protected]> | 2019-04-15 14:51:01 -0700 |
commit | 62b2ce0592963b5328d126b05a726913adeb5ba2 (patch) | |
tree | a273ddf3b3a60071378a965f8010a4e2d2a1cbac | |
parent | 9c52dce6a9219f031ffe0a80fd6f389960cb0018 (diff) |
iris: Preserve all PIPE_TRANSFER flags in xfer->usage
We need to preserve PIPE_TRANSFER_FLUSH_EXPLICIT, DISCARD_RANGE, and
so on, but don't want to pass them to iris_bo_map(). So, keep them all,
but mask them off when calling map.
Chris Wilson told me to do this a long time ago and he was right.
-rw-r--r-- | src/gallium/drivers/iris/iris_bufmgr.h | 3 | ||||
-rw-r--r-- | src/gallium/drivers/iris/iris_resource.c | 22 |
2 files changed, 12 insertions, 13 deletions
diff --git a/src/gallium/drivers/iris/iris_bufmgr.h b/src/gallium/drivers/iris/iris_bufmgr.h index a8a65088036..fa74c978191 100644 --- a/src/gallium/drivers/iris/iris_bufmgr.h +++ b/src/gallium/drivers/iris/iris_bufmgr.h @@ -251,6 +251,9 @@ void iris_bo_unreference(struct iris_bo *bo); #define MAP_INTERNAL_MASK (0xff << 24) #define MAP_RAW (0x01 << 24) +#define MAP_FLAGS (MAP_READ | MAP_WRITE | MAP_ASYNC | \ + MAP_PERSISTENT | MAP_COHERENT | MAP_INTERNAL_MASK) + /** * Maps the buffer into userspace. * diff --git a/src/gallium/drivers/iris/iris_resource.c b/src/gallium/drivers/iris/iris_resource.c index 82e1add94af..e05714cdcdc 100644 --- a/src/gallium/drivers/iris/iris_resource.c +++ b/src/gallium/drivers/iris/iris_resource.c @@ -968,7 +968,8 @@ iris_map_copy_region(struct iris_transfer *map) if (iris_batch_references(map->batch, staging_bo)) iris_batch_flush(map->batch); - map->ptr = iris_bo_map(map->dbg, staging_bo, xfer->usage) + extra; + map->ptr = + iris_bo_map(map->dbg, staging_bo, xfer->usage & MAP_FLAGS) + extra; map->unmap = iris_unmap_copy_region; } @@ -1052,7 +1053,7 @@ iris_unmap_s8(struct iris_transfer *map) if (xfer->usage & PIPE_TRANSFER_WRITE) { uint8_t *untiled_s8_map = map->ptr; uint8_t *tiled_s8_map = - iris_bo_map(map->dbg, res->bo, xfer->usage | MAP_RAW); + iris_bo_map(map->dbg, res->bo, (xfer->usage | MAP_RAW) & MAP_FLAGS); for (int s = 0; s < box->depth; s++) { unsigned x0_el, y0_el; @@ -1102,7 +1103,7 @@ iris_map_s8(struct iris_transfer *map) if (!(xfer->usage & PIPE_TRANSFER_DISCARD_RANGE)) { uint8_t *untiled_s8_map = map->ptr; uint8_t *tiled_s8_map = - iris_bo_map(map->dbg, res->bo, xfer->usage | MAP_RAW); + iris_bo_map(map->dbg, res->bo, (xfer->usage | MAP_RAW) & MAP_FLAGS); for (int s = 0; s < box->depth; s++) { unsigned x0_el, y0_el; @@ -1160,7 +1161,8 @@ iris_unmap_tiled_memcpy(struct iris_transfer *map) const bool has_swizzling = false; if (xfer->usage & PIPE_TRANSFER_WRITE) { - char *dst = iris_bo_map(map->dbg, res->bo, xfer->usage | MAP_RAW); + char *dst = + iris_bo_map(map->dbg, res->bo, (xfer->usage | MAP_RAW) & MAP_FLAGS); for (int s = 0; s < box->depth; s++) { unsigned x1, x2, y1, y2; @@ -1204,7 +1206,8 @@ iris_map_tiled_memcpy(struct iris_transfer *map) // XXX: PIPE_TRANSFER_READ? if (!(xfer->usage & PIPE_TRANSFER_DISCARD_RANGE)) { - char *src = iris_bo_map(map->dbg, res->bo, xfer->usage | MAP_RAW); + char *src = + iris_bo_map(map->dbg, res->bo, (xfer->usage | MAP_RAW) & MAP_FLAGS); for (int s = 0; s < box->depth; s++) { unsigned x1, x2, y1, y2; @@ -1229,7 +1232,7 @@ iris_map_direct(struct iris_transfer *map) struct pipe_box *box = &xfer->box; struct iris_resource *res = (struct iris_resource *) xfer->resource; - void *ptr = iris_bo_map(map->dbg, res->bo, xfer->usage); + void *ptr = iris_bo_map(map->dbg, res->bo, xfer->usage & MAP_FLAGS); if (res->base.target == PIPE_BUFFER) { xfer->stride = 0; @@ -1308,13 +1311,6 @@ iris_transfer_map(struct pipe_context *ctx, xfer->box = *box; *ptransfer = xfer; - xfer->usage &= (PIPE_TRANSFER_READ | - PIPE_TRANSFER_WRITE | - PIPE_TRANSFER_UNSYNCHRONIZED | - PIPE_TRANSFER_PERSISTENT | - PIPE_TRANSFER_COHERENT | - PIPE_TRANSFER_DISCARD_RANGE); - /* Avoid using GPU copies for persistent/coherent buffers, as the idea * there is to access them simultaneously on the CPU & GPU. This also * avoids trying to use GPU copies for our u_upload_mgr buffers which |