1/* 2 * Copyright (c) 2016 Etnaviv Project 3 * 4 * Permission is hereby granted, free of charge, to any person obtaining a 5 * copy of this software and associated documentation files (the "Software"), 6 * to deal in the Software without restriction, including without limitation 7 * the rights to use, copy, modify, merge, publish, distribute, sub license, 8 * and/or sell copies of the Software, and to permit persons to whom the 9 * Software is furnished to do so, subject to the following conditions: 10 * 11 * The above copyright notice and this permission notice (including the 12 * next paragraph) shall be included in all copies or substantial portions 13 * of the Software. 14 * 15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 17 * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL 18 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER 19 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING 20 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER 21 * DEALINGS IN THE SOFTWARE. 22 * 23 * Authors: 24 * Christian Gmeiner <christian.gmeiner@gmail.com> 25 */ 26 27#include "etnaviv_format.h" 28 29#include "hw/common_3d.xml.h" 30#include "hw/state.xml.h" 31#include "hw/state_3d.xml.h" 32 33#include "pipe/p_defines.h" 34#include "util/compiler.h" 35 36/* Specifies the table of all the formats and their features. Also supplies 37 * the helpers that look up various data in those tables. 38 */ 39 40struct etna_format { 41 unsigned vtx; 42 unsigned tex; 43 unsigned pe; 44 bool present; 45}; 46 47#define PE_FORMAT_NONE ~0 48 49#define PE_FORMAT_MASK 0x7f 50#define PE_FORMAT(x) ((x) & PE_FORMAT_MASK) 51#define PE_FORMAT_RB_SWAP 0x80 52 53#define PE_FORMAT_X8B8G8R8 (PE_FORMAT_X8R8G8B8 | PE_FORMAT_RB_SWAP) 54#define PE_FORMAT_A8B8G8R8 (PE_FORMAT_A8R8G8B8 | PE_FORMAT_RB_SWAP) 55 56#define TS_SAMPLER_FORMAT_NONE ETNA_NO_MATCH 57 58/* vertex + texture */ 59#define VT(pipe, vtxfmt, texfmt, rsfmt) \ 60 [PIPE_FORMAT_##pipe] = { \ 61 .vtx = FE_DATA_TYPE_##vtxfmt, \ 62 .tex = TEXTURE_FORMAT_##texfmt, \ 63 .pe = PE_FORMAT_##rsfmt, \ 64 .present = 1, \ 65 } 66 67/* texture-only */ 68#define _T(pipe, fmt, rsfmt) \ 69 [PIPE_FORMAT_##pipe] = { \ 70 .vtx = ETNA_NO_MATCH, \ 71 .tex = TEXTURE_FORMAT_##fmt, \ 72 .pe = PE_FORMAT_##rsfmt, \ 73 .present = 1, \ 74 } 75 76/* vertex-only */ 77#define V_(pipe, fmt, rsfmt) \ 78 [PIPE_FORMAT_##pipe] = { \ 79 .vtx = FE_DATA_TYPE_##fmt, \ 80 .tex = ETNA_NO_MATCH, \ 81 .pe = PE_FORMAT_##rsfmt, \ 82 .present = 1, \ 83 } 84 85static struct etna_format formats[PIPE_FORMAT_COUNT] = { 86 /* 8-bit */ 87 VT(R8_UNORM, UNSIGNED_BYTE, L8, R8), 88 VT(R8_SNORM, BYTE, EXT_R8_SNORM | EXT_FORMAT, NONE), 89 VT(R8_UINT, BYTE_I, EXT_R8I | EXT_FORMAT, R8I), 90 VT(R8_SINT, BYTE_I, EXT_R8I | EXT_FORMAT, R8I), 91 V_(R8_USCALED, UNSIGNED_BYTE, NONE), 92 V_(R8_SSCALED, BYTE, NONE), 93 94 _T(A8_UNORM, A8, NONE), 95 _T(L8_UNORM, L8, NONE), 96 _T(I8_UNORM, I8, NONE), 97 98 /* 16-bit */ 99 V_(R16_UNORM, UNSIGNED_SHORT, NONE), 100 V_(R16_SNORM, SHORT, NONE), 101 VT(R16_UINT, SHORT_I, EXT_R16I | EXT_FORMAT, R16I), 102 VT(R16_SINT, SHORT_I, EXT_R16I | EXT_FORMAT, R16I), 103 V_(R16_USCALED, UNSIGNED_SHORT, NONE), 104 V_(R16_SSCALED, SHORT, NONE), 105 VT(R16_FLOAT, HALF_FLOAT, EXT_R16F | EXT_FORMAT, R16F), 106 107 _T(B4G4R4A4_UNORM, A4R4G4B4, A4R4G4B4), 108 _T(B4G4R4X4_UNORM, X4R4G4B4, X4R4G4B4), 109 110 _T(L8A8_UNORM, A8L8, NONE), 111 112 _T(Z16_UNORM, D16, NONE), 113 _T(B5G6R5_UNORM, R5G6B5, R5G6B5), 114 _T(B5G5R5A1_UNORM, A1R5G5B5, A1R5G5B5), 115 _T(B5G5R5X1_UNORM, X1R5G5B5, X1R5G5B5), 116 117 VT(R8G8_UNORM, UNSIGNED_BYTE, EXT_G8R8 | EXT_FORMAT, G8R8), 118 VT(R8G8_SNORM, BYTE, EXT_G8R8_SNORM | EXT_FORMAT, NONE), 119 VT(R8G8_UINT, BYTE_I, EXT_G8R8I | EXT_FORMAT, G8R8I), 120 VT(R8G8_SINT, BYTE_I, EXT_G8R8I | EXT_FORMAT, G8R8I), 121 V_(R8G8_USCALED, UNSIGNED_BYTE, NONE), 122 V_(R8G8_SSCALED, BYTE, NONE), 123 124 /* 24-bit */ 125 V_(R8G8B8_UNORM, UNSIGNED_BYTE, NONE), 126 V_(R8G8B8_SNORM, BYTE, NONE), 127 V_(R8G8B8_UINT, BYTE_I, NONE), 128 V_(R8G8B8_SINT, BYTE_I, NONE), 129 V_(R8G8B8_USCALED, UNSIGNED_BYTE, NONE), 130 V_(R8G8B8_SSCALED, BYTE, NONE), 131 132 /* 32-bit */ 133 V_(R32_UNORM, UNSIGNED_INT, NONE), 134 V_(R32_SNORM, INT, NONE), 135 VT(R32_SINT, FLOAT, EXT_R32F | EXT_FORMAT, R32F), 136 VT(R32_UINT, FLOAT, EXT_R32F | EXT_FORMAT, R32F), 137 V_(R32_USCALED, UNSIGNED_INT, NONE), 138 V_(R32_SSCALED, INT, NONE), 139 VT(R32_FLOAT, FLOAT, EXT_R32F | EXT_FORMAT, R32F), 140 V_(R32_FIXED, FIXED, NONE), 141 142 V_(R16G16_UNORM, UNSIGNED_SHORT, NONE), 143 V_(R16G16_SNORM, SHORT, NONE), 144 VT(R16G16_UINT, SHORT_I, EXT_G16R16I | EXT_FORMAT, G16R16I), 145 VT(R16G16_SINT, SHORT_I, EXT_G16R16I | EXT_FORMAT, G16R16I), 146 V_(R16G16_USCALED, UNSIGNED_SHORT, NONE), 147 V_(R16G16_SSCALED, SHORT, NONE), 148 VT(R16G16_FLOAT, HALF_FLOAT, EXT_G16R16F | EXT_FORMAT, G16R16F), 149 150 V_(A8B8G8R8_UNORM, UNSIGNED_BYTE, NONE), 151 152 VT(R8G8B8A8_UNORM, UNSIGNED_BYTE, A8B8G8R8, A8B8G8R8), 153 VT(R8G8B8A8_SNORM, BYTE, EXT_A8B8G8R8_SNORM | EXT_FORMAT, NONE), 154 _T(R8G8B8X8_UNORM, X8B8G8R8, X8B8G8R8), 155 _T(R8G8B8X8_SNORM, EXT_X8B8G8R8_SNORM | EXT_FORMAT, NONE), 156 VT(R8G8B8A8_UINT, BYTE_I, EXT_A8B8G8R8I | EXT_FORMAT, A8B8G8R8I), 157 VT(R8G8B8A8_SINT, BYTE_I, EXT_A8B8G8R8I | EXT_FORMAT, A8B8G8R8I), 158 V_(R8G8B8A8_USCALED, UNSIGNED_BYTE, A8B8G8R8), 159 V_(R8G8B8A8_SSCALED, BYTE, A8B8G8R8), 160 161 _T(B8G8R8A8_UNORM, A8R8G8B8, A8R8G8B8), 162 _T(B8G8R8X8_UNORM, X8R8G8B8, X8R8G8B8), 163 164 VT(R10G10B10A2_UNORM, UNSIGNED_INT_2_10_10_10_REV, EXT_A2B10G10R10 | EXT_FORMAT, A2B10G10R10), 165 _T(R10G10B10X2_UNORM, EXT_A2B10G10R10 | EXT_FORMAT, A2B10G10R10), 166 V_(R10G10B10A2_SNORM, INT_2_10_10_10_REV, NONE), 167 _T(R10G10B10A2_UINT, EXT_A2B10G10R10UI | EXT_FORMAT, A2B10G10R10UI), 168 V_(R10G10B10A2_USCALED, UNSIGNED_INT_2_10_10_10_REV, NONE), 169 V_(R10G10B10A2_SSCALED, INT_2_10_10_10_REV, NONE), 170 171 _T(X8Z24_UNORM, D24X8, NONE), 172 _T(S8_UINT_Z24_UNORM, D24X8, NONE), 173 174 _T(R9G9B9E5_FLOAT, E5B9G9R9, NONE), 175 _T(R11G11B10_FLOAT, EXT_B10G11R11F | EXT_FORMAT, B10G11R11F), 176 177 /* 48-bit */ 178 V_(R16G16B16_UNORM, UNSIGNED_SHORT, NONE), 179 V_(R16G16B16_SNORM, SHORT, NONE), 180 V_(R16G16B16_UINT, SHORT_I, NONE), 181 V_(R16G16B16_SINT, SHORT_I, NONE), 182 V_(R16G16B16_USCALED, UNSIGNED_SHORT, NONE), 183 V_(R16G16B16_SSCALED, SHORT, NONE), 184 V_(R16G16B16_FLOAT, HALF_FLOAT, NONE), 185 186 /* 64-bit */ 187 V_(R16G16B16A16_UNORM, UNSIGNED_SHORT, NONE), 188 V_(R16G16B16A16_SNORM, SHORT, NONE), 189 VT(R16G16B16A16_UINT, SHORT_I, EXT_A16B16G16R16I | EXT_FORMAT, A16B16G16R16I), 190 VT(R16G16B16A16_SINT, SHORT_I, EXT_A16B16G16R16I | EXT_FORMAT, A16B16G16R16I), 191 V_(R16G16B16A16_USCALED, UNSIGNED_SHORT, NONE), 192 V_(R16G16B16A16_SSCALED, SHORT, NONE), 193 VT(R16G16B16A16_FLOAT, HALF_FLOAT, EXT_A16B16G16R16F | EXT_FORMAT, A16B16G16R16F), 194 195 V_(R32G32_UNORM, UNSIGNED_INT, NONE), 196 V_(R32G32_SNORM, INT, NONE), 197 VT(R32G32_UINT, FLOAT, EXT_G32R32F | EXT_FORMAT, G32R32F), 198 VT(R32G32_SINT, FLOAT, EXT_G32R32F | EXT_FORMAT, G32R32F), 199 V_(R32G32_USCALED, UNSIGNED_INT, NONE), 200 V_(R32G32_SSCALED, INT, NONE), 201 VT(R32G32_FLOAT, FLOAT, EXT_G32R32F | EXT_FORMAT, G32R32F), 202 V_(R32G32_FIXED, FIXED, NONE), 203 204 /* 96-bit */ 205 V_(R32G32B32_UNORM, UNSIGNED_INT, NONE), 206 V_(R32G32B32_SNORM, INT, NONE), 207 V_(R32G32B32_UINT, FLOAT, NONE), 208 V_(R32G32B32_SINT, FLOAT, NONE), 209 V_(R32G32B32_USCALED, UNSIGNED_INT, NONE), 210 V_(R32G32B32_SSCALED, INT, NONE), 211 V_(R32G32B32_FLOAT, FLOAT, NONE), 212 V_(R32G32B32_FIXED, FIXED, NONE), 213 214 /* 128-bit */ 215 V_(R32G32B32A32_UNORM, UNSIGNED_INT, NONE), 216 V_(R32G32B32A32_SNORM, INT, NONE), 217 V_(R32G32B32A32_UINT, FLOAT, NONE), 218 V_(R32G32B32A32_SINT, FLOAT, NONE), 219 V_(R32G32B32A32_USCALED, UNSIGNED_INT, NONE), 220 V_(R32G32B32A32_SSCALED, INT, NONE), 221 V_(R32G32B32A32_FLOAT, FLOAT, NONE), 222 V_(R32G32B32A32_FIXED, FIXED, NONE), 223 224 /* compressed */ 225 _T(ETC1_RGB8, ETC1, NONE), 226 227 _T(DXT1_RGB, DXT1, NONE), 228 _T(DXT1_RGBA, DXT1, NONE), 229 _T(DXT3_RGBA, DXT2_DXT3, NONE), 230 _T(DXT5_RGBA, DXT4_DXT5, NONE), 231 232 _T(ETC2_RGB8, EXT_NONE | EXT_FORMAT, NONE), /* Extd. format NONE doubles as ETC2_RGB8 */ 233 _T(ETC2_RGB8A1, EXT_RGB8_PUNCHTHROUGH_ALPHA1_ETC2 | EXT_FORMAT, NONE), 234 _T(ETC2_RGBA8, EXT_RGBA8_ETC2_EAC | EXT_FORMAT, NONE), 235 _T(ETC2_R11_UNORM, EXT_R11_EAC | EXT_FORMAT, NONE), 236 _T(ETC2_R11_SNORM, EXT_SIGNED_R11_EAC | EXT_FORMAT, NONE), 237 _T(ETC2_RG11_UNORM, EXT_RG11_EAC | EXT_FORMAT, NONE), 238 _T(ETC2_RG11_SNORM, EXT_SIGNED_RG11_EAC | EXT_FORMAT, NONE), 239 240 _T(ASTC_4x4, ASTC_RGBA_4x4 | ASTC_FORMAT, NONE), 241 _T(ASTC_5x4, ASTC_RGBA_5x4 | ASTC_FORMAT, NONE), 242 _T(ASTC_5x5, ASTC_RGBA_5x5 | ASTC_FORMAT, NONE), 243 _T(ASTC_6x5, ASTC_RGBA_6x5 | ASTC_FORMAT, NONE), 244 _T(ASTC_6x6, ASTC_RGBA_6x6 | ASTC_FORMAT, NONE), 245 _T(ASTC_8x5, ASTC_RGBA_8x5 | ASTC_FORMAT, NONE), 246 _T(ASTC_8x6, ASTC_RGBA_8x6 | ASTC_FORMAT, NONE), 247 _T(ASTC_8x8, ASTC_RGBA_8x8 | ASTC_FORMAT, NONE), 248 _T(ASTC_10x5, ASTC_RGBA_10x5 | ASTC_FORMAT, NONE), 249 _T(ASTC_10x6, ASTC_RGBA_10x6 | ASTC_FORMAT, NONE), 250 _T(ASTC_10x8, ASTC_RGBA_10x8 | ASTC_FORMAT, NONE), 251 _T(ASTC_10x10, ASTC_RGBA_10x10 | ASTC_FORMAT, NONE), 252 _T(ASTC_12x10, ASTC_RGBA_12x10 | ASTC_FORMAT, NONE), 253 _T(ASTC_12x12, ASTC_RGBA_12x12 | ASTC_FORMAT, NONE), 254 255 /* YUV */ 256 _T(YUYV, YUY2, YUY2), 257 _T(UYVY, UYVY, NONE), 258}; 259 260uint32_t 261translate_texture_format(enum pipe_format fmt) 262{ 263 fmt = util_format_linear(fmt); 264 265 if (!formats[fmt].present) 266 return ETNA_NO_MATCH; 267 268 return formats[fmt].tex; 269} 270 271bool 272texture_use_int_filter(const struct pipe_sampler_view *sv, 273 const struct pipe_sampler_state *ss, 274 bool tex_desc) 275{ 276 switch (sv->target) { 277 case PIPE_TEXTURE_1D_ARRAY: 278 case PIPE_TEXTURE_2D_ARRAY: 279 if (tex_desc) 280 break; 281 FALLTHROUGH; 282 case PIPE_TEXTURE_3D: 283 return false; 284 default: 285 break; 286 } 287 288 /* only unorm formats can use int filter */ 289 if (!util_format_is_unorm(sv->format)) 290 return false; 291 292 if (util_format_is_srgb(sv->format)) 293 return false; 294 295 if (util_format_description(sv->format)->layout == UTIL_FORMAT_LAYOUT_ASTC) 296 return false; 297 298 if (ss->max_anisotropy > 1) 299 return false; 300 301 switch (sv->format) { 302 /* apparently D16 can't use int filter but D24 can */ 303 case PIPE_FORMAT_Z16_UNORM: 304 case PIPE_FORMAT_R10G10B10A2_UNORM: 305 case PIPE_FORMAT_R10G10B10X2_UNORM: 306 case PIPE_FORMAT_ETC2_R11_UNORM: 307 case PIPE_FORMAT_ETC2_RG11_UNORM: 308 return false; 309 default: 310 return true; 311 } 312} 313 314bool 315texture_format_needs_swiz(enum pipe_format fmt) 316{ 317 return util_format_linear(fmt) == PIPE_FORMAT_R8_UNORM; 318} 319 320uint32_t 321get_texture_swiz(enum pipe_format fmt, unsigned swizzle_r, 322 unsigned swizzle_g, unsigned swizzle_b, unsigned swizzle_a) 323{ 324 unsigned char swiz[4] = { 325 swizzle_r, swizzle_g, swizzle_b, swizzle_a, 326 }; 327 328 if (util_format_linear(fmt) == PIPE_FORMAT_R8_UNORM) { 329 /* R8 is emulated with L8, needs yz channels set to zero */ 330 for (unsigned i = 0; i < 4; i++) { 331 if (swiz[i] == PIPE_SWIZZLE_Y || swiz[i] == PIPE_SWIZZLE_Z) 332 swiz[i] = PIPE_SWIZZLE_0; 333 } 334 } 335 336 /* PIPE_SWIZZLE_ maps 1:1 to TEXTURE_SWIZZLE_ */ 337 STATIC_ASSERT(PIPE_SWIZZLE_X == TEXTURE_SWIZZLE_RED); 338 STATIC_ASSERT(PIPE_SWIZZLE_Y == TEXTURE_SWIZZLE_GREEN); 339 STATIC_ASSERT(PIPE_SWIZZLE_Z == TEXTURE_SWIZZLE_BLUE); 340 STATIC_ASSERT(PIPE_SWIZZLE_W == TEXTURE_SWIZZLE_ALPHA); 341 STATIC_ASSERT(PIPE_SWIZZLE_0 == TEXTURE_SWIZZLE_ZERO); 342 STATIC_ASSERT(PIPE_SWIZZLE_1 == TEXTURE_SWIZZLE_ONE); 343 344 return VIVS_TE_SAMPLER_CONFIG1_SWIZZLE_R(swiz[0]) | 345 VIVS_TE_SAMPLER_CONFIG1_SWIZZLE_G(swiz[1]) | 346 VIVS_TE_SAMPLER_CONFIG1_SWIZZLE_B(swiz[2]) | 347 VIVS_TE_SAMPLER_CONFIG1_SWIZZLE_A(swiz[3]); 348} 349 350uint32_t 351translate_pe_format(enum pipe_format fmt) 352{ 353 fmt = util_format_linear(fmt); 354 355 if (!formats[fmt].present) 356 return ETNA_NO_MATCH; 357 358 if (formats[fmt].pe == ETNA_NO_MATCH) 359 return ETNA_NO_MATCH; 360 361 return PE_FORMAT(formats[fmt].pe); 362} 363 364int 365translate_pe_format_rb_swap(enum pipe_format fmt) 366{ 367 fmt = util_format_linear(fmt); 368 assert(formats[fmt].present); 369 370 return formats[fmt].pe & PE_FORMAT_RB_SWAP; 371} 372 373/* Return type flags for vertex element format */ 374uint32_t 375translate_vertex_format_type(enum pipe_format fmt) 376{ 377 if (!formats[fmt].present) 378 return ETNA_NO_MATCH; 379 380 return formats[fmt].vtx; 381} 382