/* * Copyright (C) 2021 Alyssa Rosenzweig * Copyright (c) 2019 Collabora, Ltd. * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice (including the next * paragraph) shall be included in all copies or substantial portions of the * Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE * SOFTWARE. */ #include #include #include #include #include #include "util/macros.h" #include "tiling.h" /* Z-order with square tiles, at most 64x64: * * [y5][x5][y4][x4][y3][x3][y2][x2][y1][x1][y0][x0] * * Efficient tiling algorithm described in * https://fgiesen.wordpress.com/2011/01/17/texture-tiling-and-swizzling/ but * for posterity, we split into X and Y parts, and are faced with the problem * of incrementing: * * 0 [x5] 0 [x4] 0 [x3] 0 [x2] 0 [x1] 0 [x0] * * To do so, we fill in the "holes" with 1's by adding the bitwise inverse of * the mask of bits we care about * * 0 [x5] 0 [x4] 0 [x3] 0 [x2] 0 [x1] 0 [x0] * + 1 0 1 0 1 0 1 0 1 0 1 0 * ------------------------------------------ * 1 [x5] 1 [x4] 1 [x3] 1 [x2] 1 [x1] 1 [x0] * * Then when we add one, the holes are passed over by forcing carry bits high. * Finally, we need to zero out the holes, by ANDing with the mask of bits we * care about. In total, we get the expression (X + ~mask + 1) & mask, and * applying the two's complement identity, we are left with (X - mask) & mask */ /* mask of bits used for X coordinate in a tile */ #define SPACE_MASK 0x555 // 0b010101010101 typedef struct { uint16_t lo; uint8_t hi; } __attribute__((packed)) agx_uint24_t; typedef struct { uint32_t lo; uint16_t hi; } __attribute__((packed)) agx_uint48_t; typedef struct { uint64_t lo; uint32_t hi; } __attribute__((packed)) agx_uint96_t; typedef struct { uint64_t lo; uint64_t hi; } __attribute__((packed)) agx_uint128_t; static uint32_t agx_space_bits(unsigned x) { assert(x < 64); return ((x & 1) << 0) | ((x & 2) << 1) | ((x & 4) << 2) | ((x & 8) << 3) | ((x & 16) << 4) | ((x & 32) << 5); } #define TILED_UNALIGNED_TYPE(pixel_t, is_store, tile_shift) { \ unsigned tile_size = (1 << tile_shift);\ unsigned pixels_per_tile = tile_size * tile_size;\ unsigned tiles_per_row = (width + tile_size - 1) >> tile_shift;\ unsigned y_offs = agx_space_bits(sy & (tile_size - 1)) << 1;\ unsigned x_offs_start = agx_space_bits(sx & (tile_size - 1));\ unsigned space_mask = SPACE_MASK & (pixels_per_tile - 1);\ \ pixel_t *linear = _linear; \ pixel_t *tiled = _tiled; \ \ for (unsigned y = sy; y < smaxy; ++y) {\ unsigned tile_y = (y >> tile_shift);\ unsigned tile_row = tile_y * tiles_per_row;\ unsigned x_offs = x_offs_start;\ \ pixel_t *linear_row = linear;\ \ for (unsigned x = sx; x < smaxx; ++x) {\ unsigned tile_x = (x >> tile_shift);\ unsigned tile_idx = (tile_row + tile_x);\ unsigned tile_base = tile_idx * pixels_per_tile;\ \ pixel_t *ptiled = &tiled[tile_base + y_offs + x_offs];\ pixel_t *plinear = (linear_row++);\ pixel_t *outp = (pixel_t *) (is_store ? ptiled : plinear); \ pixel_t *inp = (pixel_t *) (is_store ? plinear : ptiled); \ *outp = *inp;\ x_offs = (x_offs - space_mask) & space_mask;\ }\ \ y_offs = (((y_offs >> 1) - space_mask) & space_mask) << 1;\ linear += linear_pitch;\ }\ } #define TILED_UNALIGNED_TYPES(bpp, store, tile_shift) { \ if (bpp == 8) \ TILED_UNALIGNED_TYPE(uint8_t, store, tile_shift) \ else if (bpp == 16) \ TILED_UNALIGNED_TYPE(uint16_t, store, tile_shift) \ else if (bpp == 24) \ TILED_UNALIGNED_TYPE(agx_uint24_t, store, tile_shift) \ else if (bpp == 32) \ TILED_UNALIGNED_TYPE(uint32_t, store, tile_shift) \ else if (bpp == 48) \ TILED_UNALIGNED_TYPE(agx_uint48_t, store, tile_shift) \ else if (bpp == 64) \ TILED_UNALIGNED_TYPE(uint64_t, store, tile_shift) \ else if (bpp == 96) \ TILED_UNALIGNED_TYPE(agx_uint96_t, store, tile_shift) \ else if (bpp == 128) \ TILED_UNALIGNED_TYPE(agx_uint128_t, store, tile_shift) \ else \ unreachable("Can't tile this bpp\n"); \ } void agx_detile(void *_tiled, void *_linear, unsigned width, unsigned bpp, unsigned linear_pitch, unsigned sx, unsigned sy, unsigned smaxx, unsigned smaxy, unsigned tile_shift) { TILED_UNALIGNED_TYPES(bpp, false, tile_shift); } void agx_tile(void *_tiled, void *_linear, unsigned width, unsigned bpp, unsigned linear_pitch, unsigned sx, unsigned sy, unsigned smaxx, unsigned smaxy, unsigned tile_shift) { TILED_UNALIGNED_TYPES(bpp, true, tile_shift); }