diff options
author | Samuel Pitoiset <[email protected]> | 2018-11-14 16:24:02 +0100 |
---|---|---|
committer | Samuel Pitoiset <[email protected]> | 2019-06-19 10:06:39 +0200 |
commit | e91c1ea06c56a45b6040b36e6ef32d0b126eebbc (patch) | |
tree | 93a4169491ea7c74e61dd02c3ddddf54910ac236 /src/amd/vulkan/radv_image.c | |
parent | a7f75377aba6853e08561acb9609c5e262e4d9a6 (diff) |
radv: implement compressed FMASK texture reads with RADV_PERFTEST=tccompatcmask
This allows us to disable the FMASK decompress pass when
transitioning from CB writes to shader reads.
This will likely be improved and enabled by default in the future.
No CTS regressions on GFX8 but a few number of multisample CTS
failures on GFX9 (they look related to the small hint).
Signed-off-by: Samuel Pitoiset <[email protected]>
Reviewed-by: Bas Nieuwenhuizen <[email protected]>
Diffstat (limited to 'src/amd/vulkan/radv_image.c')
-rw-r--r-- | src/amd/vulkan/radv_image.c | 41 |
1 files changed, 41 insertions, 0 deletions
diff --git a/src/amd/vulkan/radv_image.c b/src/amd/vulkan/radv_image.c index 98df24d3546..4bea09a8a2b 100644 --- a/src/amd/vulkan/radv_image.c +++ b/src/amd/vulkan/radv_image.c @@ -219,6 +219,29 @@ radv_use_dcc_for_image(struct radv_device *device, return true; } +static bool +radv_use_tc_compat_cmask_for_image(struct radv_device *device, + struct radv_image *image) +{ + if (!(device->instance->perftest_flags & RADV_PERFTEST_TC_COMPAT_CMASK)) + return false; + + /* TC-compat CMASK is only available for GFX8+. */ + if (device->physical_device->rad_info.chip_class < GFX8) + return false; + + if (image->usage & VK_IMAGE_USAGE_STORAGE_BIT) + return false; + + if (radv_image_has_dcc(image)) + return false; + + if (!radv_image_has_cmask(image)) + return false; + + return true; +} + static void radv_prefill_surface_from_metadata(struct radv_device *device, struct radeon_surf *surface, @@ -729,11 +752,26 @@ si_make_texture_descriptor(struct radv_device *device, S_008F20_PITCH(image->planes[0].surface.u.gfx9.fmask.epitch); fmask_state[5] |= S_008F24_META_PIPE_ALIGNED(image->planes[0].surface.u.gfx9.cmask.pipe_aligned) | S_008F24_META_RB_ALIGNED(image->planes[0].surface.u.gfx9.cmask.rb_aligned); + + if (radv_image_is_tc_compat_cmask(image)) { + va = gpu_address + image->offset + image->cmask.offset; + + fmask_state[5] |= S_008F24_META_DATA_ADDRESS(va >> 40); + fmask_state[6] |= S_008F28_COMPRESSION_EN(1); + fmask_state[7] |= va >> 8; + } } else { fmask_state[3] |= S_008F1C_TILING_INDEX(image->fmask.tile_mode_index); fmask_state[4] |= S_008F20_DEPTH(depth - 1) | S_008F20_PITCH(image->fmask.pitch_in_pixels - 1); fmask_state[5] |= S_008F24_LAST_ARRAY(last_layer); + + if (radv_image_is_tc_compat_cmask(image)) { + va = gpu_address + image->offset + image->cmask.offset; + + fmask_state[6] |= S_008F28_COMPRESSION_EN(1); + fmask_state[7] |= va >> 8; + } } } else if (fmask_state) memset(fmask_state, 0, 8 * 4); @@ -1122,6 +1160,9 @@ radv_image_create(VkDevice _device, /* Try to enable FMASK for multisampled images. */ if (radv_image_can_enable_fmask(image)) { radv_image_alloc_fmask(device, image); + + if (radv_use_tc_compat_cmask_for_image(device, image)) + image->tc_compatible_cmask = true; } else { /* Otherwise, try to enable HTILE for depth surfaces. */ if (radv_image_can_enable_htile(image) && |