/* * Copyright (c) 2016 Etnaviv Project * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sub license, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice (including the * next paragraph) shall be included in all copies or substantial portions * of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER * DEALINGS IN THE SOFTWARE. * * Authors: * Christian Gmeiner */ #include "etnaviv_format.h" #include "hw/state.xml.h" #include "hw/state_3d.xml.h" #include "pipe/p_defines.h" /* Specifies the table of all the formats and their features. Also supplies * the helpers that look up various data in those tables. */ struct etna_format { unsigned vtx; unsigned tex; unsigned rs; boolean present; }; #define RS_FORMAT_NONE ~0 #define RS_FORMAT_MASK 0xf #define RS_FORMAT(x) ((x) & RS_FORMAT_MASK) #define RS_FORMAT_RB_SWAP 0x10 #define RS_FORMAT_X8B8G8R8 (RS_FORMAT_X8R8G8B8 | RS_FORMAT_RB_SWAP) #define RS_FORMAT_A8B8G8R8 (RS_FORMAT_A8R8G8B8 | RS_FORMAT_RB_SWAP) /* vertex + texture */ #define VT(pipe, vtxfmt, texfmt, rsfmt) \ [PIPE_FORMAT_##pipe] = { \ .vtx = VIVS_FE_VERTEX_ELEMENT_CONFIG_TYPE_##vtxfmt, \ .tex = TEXTURE_FORMAT_##texfmt, \ .rs = RS_FORMAT_##rsfmt, \ .present = 1, \ } /* texture-only */ #define _T(pipe, fmt, rsfmt) \ [PIPE_FORMAT_##pipe] = { \ .vtx = ETNA_NO_MATCH, \ .tex = TEXTURE_FORMAT_##fmt, \ .rs = RS_FORMAT_##rsfmt, \ .present = 1, \ } /* vertex-only */ #define V_(pipe, fmt, rsfmt) \ [PIPE_FORMAT_##pipe] = { \ .vtx = VIVS_FE_VERTEX_ELEMENT_CONFIG_TYPE_##fmt, \ .tex = ETNA_NO_MATCH, \ .rs = RS_FORMAT_##rsfmt, \ .present = 1, \ } static struct etna_format formats[PIPE_FORMAT_COUNT] = { /* 8-bit */ V_(R8_UNORM, UNSIGNED_BYTE, NONE), V_(R8_SNORM, BYTE, NONE), V_(R8_UINT, UNSIGNED_BYTE, NONE), V_(R8_SINT, BYTE, NONE), V_(R8_USCALED, UNSIGNED_BYTE, NONE), V_(R8_SSCALED, BYTE, NONE), _T(A8_UNORM, A8, NONE), _T(L8_UNORM, L8, NONE), _T(I8_UNORM, I8, NONE), /* 16-bit */ V_(R16_UNORM, UNSIGNED_SHORT, NONE), V_(R16_SNORM, SHORT, NONE), V_(R16_UINT, UNSIGNED_SHORT, NONE), V_(R16_SINT, SHORT, NONE), V_(R16_USCALED, UNSIGNED_SHORT, NONE), V_(R16_SSCALED, SHORT, NONE), V_(R16_FLOAT, HALF_FLOAT, NONE), _T(B4G4R4A4_UNORM, A4R4G4B4, A4R4G4B4), _T(B4G4R4X4_UNORM, X4R4G4B4, X4R4G4B4), _T(Z16_UNORM, D16, A4R4G4B4), _T(B5G6R5_UNORM, R5G6B5, R5G6B5), _T(B5G5R5A1_UNORM, A1R5G5B5, A1R5G5B5), _T(B5G5R5X1_UNORM, X1R5G5B5, X1R5G5B5), V_(R8G8_UNORM, UNSIGNED_BYTE, NONE), V_(R8G8_SNORM, BYTE, NONE), V_(R8G8_UINT, UNSIGNED_BYTE, NONE), V_(R8G8_SINT, BYTE, NONE), V_(R8G8_USCALED, UNSIGNED_BYTE, NONE), V_(R8G8_SSCALED, BYTE, NONE), /* 24-bit */ V_(R8G8B8_UNORM, UNSIGNED_BYTE, NONE), V_(R8G8B8_SNORM, BYTE, NONE), V_(R8G8B8_UINT, UNSIGNED_BYTE, NONE), V_(R8G8B8_SINT, BYTE, NONE), V_(R8G8B8_USCALED, UNSIGNED_BYTE, NONE), V_(R8G8B8_SSCALED, BYTE, NONE), /* 32-bit */ V_(R32_UNORM, UNSIGNED_INT, NONE), V_(R32_SNORM, INT, NONE), V_(R32_SINT, INT, NONE), V_(R32_UINT, UNSIGNED_INT, NONE), V_(R32_USCALED, UNSIGNED_INT, NONE), V_(R32_SSCALED, INT, NONE), V_(R32_FLOAT, FLOAT, NONE), V_(R32_FIXED, FIXED, NONE), V_(R16G16_UNORM, UNSIGNED_SHORT, NONE), V_(R16G16_SNORM, SHORT, NONE), V_(R16G16_UINT, UNSIGNED_SHORT, NONE), V_(R16G16_SINT, SHORT, NONE), V_(R16G16_USCALED, UNSIGNED_SHORT, NONE), V_(R16G16_SSCALED, SHORT, NONE), V_(R16G16_FLOAT, HALF_FLOAT, NONE), V_(A8B8G8R8_UNORM, UNSIGNED_BYTE, NONE), V_(R8G8B8A8_UNORM, UNSIGNED_BYTE, A8B8G8R8), V_(R8G8B8A8_SNORM, BYTE, A8B8G8R8), _T(R8G8B8X8_UNORM, X8B8G8R8, X8B8G8R8), V_(R8G8B8A8_UINT, UNSIGNED_BYTE, A8B8G8R8), V_(R8G8B8A8_SINT, BYTE, A8B8G8R8), V_(R8G8B8A8_USCALED, UNSIGNED_BYTE, A8B8G8R8), V_(R8G8B8A8_SSCALED, BYTE, A8B8G8R8), _T(R8G8B8A8_UNORM, A8B8G8R8, A8B8G8R8), _T(R8G8B8X8_UNORM, X8B8G8R8, X8B8G8R8), _T(B8G8R8A8_UNORM, A8R8G8B8, A8R8G8B8), _T(B8G8R8X8_UNORM, X8R8G8B8, X8R8G8B8), V_(R10G10B10A2_UNORM, UNSIGNED_INT_10_10_10_2, NONE), V_(R10G10B10A2_SNORM, INT_10_10_10_2, NONE), V_(R10G10B10A2_USCALED, UNSIGNED_INT_10_10_10_2, NONE), V_(R10G10B10A2_SSCALED, INT_10_10_10_2, NONE), _T(X8Z24_UNORM, D24S8, A8R8G8B8), _T(S8_UINT_Z24_UNORM, D24S8, A8R8G8B8), /* 48-bit */ V_(R16G16B16_UNORM, UNSIGNED_SHORT, NONE), V_(R16G16B16_SNORM, SHORT, NONE), V_(R16G16B16_UINT, UNSIGNED_SHORT, NONE), V_(R16G16B16_SINT, SHORT, NONE), V_(R16G16B16_USCALED, UNSIGNED_SHORT, NONE), V_(R16G16B16_SSCALED, SHORT, NONE), V_(R16G16B16_FLOAT, HALF_FLOAT, NONE), /* 64-bit */ V_(R16G16B16A16_UNORM, UNSIGNED_SHORT, NONE), V_(R16G16B16A16_SNORM, SHORT, NONE), V_(R16G16B16A16_UINT, UNSIGNED_SHORT, NONE), V_(R16G16B16A16_SINT, SHORT, NONE), V_(R16G16B16A16_USCALED, UNSIGNED_SHORT, NONE), V_(R16G16B16A16_SSCALED, SHORT, NONE), V_(R16G16B16A16_FLOAT, HALF_FLOAT, NONE), V_(R32G32_UNORM, UNSIGNED_INT, NONE), V_(R32G32_SNORM, INT, NONE), V_(R32G32_UINT, UNSIGNED_INT, NONE), V_(R32G32_SINT, INT, NONE), V_(R32G32_USCALED, UNSIGNED_INT, NONE), V_(R32G32_SSCALED, INT, NONE), V_(R32G32_FLOAT, FLOAT, NONE), V_(R32G32_FIXED, FIXED, NONE), /* 96-bit */ V_(R32G32B32_UNORM, UNSIGNED_INT, NONE), V_(R32G32B32_SNORM, INT, NONE), V_(R32G32B32_UINT, UNSIGNED_INT, NONE), V_(R32G32B32_SINT, INT, NONE), V_(R32G32B32_USCALED, UNSIGNED_INT, NONE), V_(R32G32B32_SSCALED, INT, NONE), V_(R32G32B32_FLOAT, FLOAT, NONE), V_(R32G32B32_FIXED, FIXED, NONE), /* 128-bit */ V_(R32G32B32A32_UNORM, UNSIGNED_INT, NONE), V_(R32G32B32A32_SNORM, INT, NONE), V_(R32G32B32A32_UINT, UNSIGNED_INT, NONE), V_(R32G32B32A32_SINT, INT, NONE), V_(R32G32B32A32_USCALED, UNSIGNED_INT, NONE), V_(R32G32B32A32_SSCALED, INT, NONE), V_(R32G32B32A32_FLOAT, FLOAT, NONE), V_(R32G32B32A32_FIXED, FIXED, NONE), /* compressed */ _T(ETC1_RGB8, ETC1, NONE), _T(DXT1_RGB, DXT1, NONE), _T(DXT1_RGBA, DXT1, NONE), _T(DXT3_RGBA, DXT2_DXT3, NONE), _T(DXT3_RGBA, DXT2_DXT3, NONE), _T(DXT5_RGBA, DXT4_DXT5, NONE), /* YUV */ _T(YUYV, YUY2, YUY2), _T(UYVY, UYVY, NONE), }; uint32_t translate_texture_format(enum pipe_format fmt) { /* XXX with TEXTURE_FORMAT_EXT and swizzle on newer chips we can * support much more */ if (!formats[fmt].present) return ETNA_NO_MATCH; return formats[fmt].tex; } uint32_t translate_rs_format(enum pipe_format fmt) { if (!formats[fmt].present) return ETNA_NO_MATCH; if (formats[fmt].rs == ETNA_NO_MATCH) return ETNA_NO_MATCH; return RS_FORMAT(formats[fmt].rs); } int translate_rs_format_rb_swap(enum pipe_format fmt) { assert(formats[fmt].present); return formats[fmt].rs & RS_FORMAT_RB_SWAP; } /* Return type flags for vertex element format */ uint32_t translate_vertex_format_type(enum pipe_format fmt) { if (!formats[fmt].present) return ETNA_NO_MATCH; return formats[fmt].vtx; }