summaryrefslogtreecommitdiffstats
path: root/src/gallium/drivers/svga
diff options
context:
space:
mode:
authorMarek Olšák <[email protected]>2012-02-23 23:44:36 +0100
committerMarek Olšák <[email protected]>2012-02-27 02:03:23 +0100
commit04e324008759282728a95a1394bac2c4c2a1a3f9 (patch)
tree71691cd97f13431ca8ddb7c902638e5b775eac67 /src/gallium/drivers/svga
parent74d303521e6ba41d1cbeb75edb2f834ebbe8c550 (diff)
gallium: remove PIPE_SHADER_CAP_OUTPUT_READ
r600g is the only driver which has made use of it. The reason the CAP was added was to fix some piglit tests when the GLSL pass lower_output_reads didn't exist. However, not removing output reads breaks the fallback for glClampColorARB, which assumes outputs are not readable. The fix would be non-trivial and my personal preference is to remove the CAP, considering that reading outputs is uncommon and that we can now use lower_output_reads to fix the issue that the CAP was supposed to workaround in the first place.
Diffstat (limited to 'src/gallium/drivers/svga')
-rw-r--r--src/gallium/drivers/svga/svga_screen.c4
1 files changed, 0 insertions, 4 deletions
diff --git a/src/gallium/drivers/svga/svga_screen.c b/src/gallium/drivers/svga/svga_screen.c
index a4de4932652..fe6e6a030b0 100644
--- a/src/gallium/drivers/svga/svga_screen.c
+++ b/src/gallium/drivers/svga/svga_screen.c
@@ -262,8 +262,6 @@ static int svga_get_shader_param(struct pipe_screen *screen, unsigned shader, en
return 0;
case PIPE_SHADER_CAP_MAX_TEXTURE_SAMPLERS:
return 16;
- case PIPE_SHADER_CAP_OUTPUT_READ:
- return 0;
}
break;
case PIPE_SHADER_VERTEX:
@@ -307,8 +305,6 @@ static int svga_get_shader_param(struct pipe_screen *screen, unsigned shader, en
return 0;
case PIPE_SHADER_CAP_INTEGERS:
return 0;
- case PIPE_SHADER_CAP_OUTPUT_READ:
- return 0;
default:
break;
}