1/* 2 * VP9 compatible video decoder 3 * 4 * Copyright (C) 2013 Ronald S. Bultje <rsbultje gmail com> 5 * Copyright (C) 2013 Clément Bœsch <u pkh me> 6 * 7 * This file is part of FFmpeg. 8 * 9 * FFmpeg is free software; you can redistribute it and/or 10 * modify it under the terms of the GNU Lesser General Public 11 * License as published by the Free Software Foundation; either 12 * version 2.1 of the License, or (at your option) any later version. 13 * 14 * FFmpeg is distributed in the hope that it will be useful, 15 * but WITHOUT ANY WARRANTY; without even the implied warranty of 16 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU 17 * Lesser General Public License for more details. 18 * 19 * You should have received a copy of the GNU Lesser General Public 20 * License along with FFmpeg; if not, write to the Free Software 21 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA 22 */ 23 24#include "libavutil/avassert.h" 25#include "libavutil/mem_internal.h" 26 27#include "threadframe.h" 28#include "videodsp.h" 29#include "vp9data.h" 30#include "vp9dec.h" 31 32static av_always_inline int check_intra_mode(VP9TileData *td, int mode, uint8_t **a, 33 uint8_t *dst_edge, ptrdiff_t stride_edge, 34 uint8_t *dst_inner, ptrdiff_t stride_inner, 35 uint8_t *l, int col, int x, int w, 36 int row, int y, enum TxfmMode tx, 37 int p, int ss_h, int ss_v, int bytesperpixel) 38{ 39 VP9Context *s = td->s; 40 int have_top = row > 0 || y > 0; 41 int have_left = col > td->tile_col_start || x > 0; 42 int have_right = x < w - 1; 43 int bpp = s->s.h.bpp; 44 static const uint8_t mode_conv[10][2 /* have_left */][2 /* have_top */] = { 45 [VERT_PRED] = { { DC_127_PRED, VERT_PRED }, 46 { DC_127_PRED, VERT_PRED } }, 47 [HOR_PRED] = { { DC_129_PRED, DC_129_PRED }, 48 { HOR_PRED, HOR_PRED } }, 49 [DC_PRED] = { { DC_128_PRED, TOP_DC_PRED }, 50 { LEFT_DC_PRED, DC_PRED } }, 51 [DIAG_DOWN_LEFT_PRED] = { { DC_127_PRED, DIAG_DOWN_LEFT_PRED }, 52 { DC_127_PRED, DIAG_DOWN_LEFT_PRED } }, 53 [DIAG_DOWN_RIGHT_PRED] = { { DIAG_DOWN_RIGHT_PRED, DIAG_DOWN_RIGHT_PRED }, 54 { DIAG_DOWN_RIGHT_PRED, DIAG_DOWN_RIGHT_PRED } }, 55 [VERT_RIGHT_PRED] = { { VERT_RIGHT_PRED, VERT_RIGHT_PRED }, 56 { VERT_RIGHT_PRED, VERT_RIGHT_PRED } }, 57 [HOR_DOWN_PRED] = { { HOR_DOWN_PRED, HOR_DOWN_PRED }, 58 { HOR_DOWN_PRED, HOR_DOWN_PRED } }, 59 [VERT_LEFT_PRED] = { { DC_127_PRED, VERT_LEFT_PRED }, 60 { DC_127_PRED, VERT_LEFT_PRED } }, 61 [HOR_UP_PRED] = { { DC_129_PRED, DC_129_PRED }, 62 { HOR_UP_PRED, HOR_UP_PRED } }, 63 [TM_VP8_PRED] = { { DC_129_PRED, VERT_PRED }, 64 { HOR_PRED, TM_VP8_PRED } }, 65 }; 66 static const struct { 67 uint8_t needs_left:1; 68 uint8_t needs_top:1; 69 uint8_t needs_topleft:1; 70 uint8_t needs_topright:1; 71 uint8_t invert_left:1; 72 } edges[N_INTRA_PRED_MODES] = { 73 [VERT_PRED] = { .needs_top = 1 }, 74 [HOR_PRED] = { .needs_left = 1 }, 75 [DC_PRED] = { .needs_top = 1, .needs_left = 1 }, 76 [DIAG_DOWN_LEFT_PRED] = { .needs_top = 1, .needs_topright = 1 }, 77 [DIAG_DOWN_RIGHT_PRED] = { .needs_left = 1, .needs_top = 1, 78 .needs_topleft = 1 }, 79 [VERT_RIGHT_PRED] = { .needs_left = 1, .needs_top = 1, 80 .needs_topleft = 1 }, 81 [HOR_DOWN_PRED] = { .needs_left = 1, .needs_top = 1, 82 .needs_topleft = 1 }, 83 [VERT_LEFT_PRED] = { .needs_top = 1, .needs_topright = 1 }, 84 [HOR_UP_PRED] = { .needs_left = 1, .invert_left = 1 }, 85 [TM_VP8_PRED] = { .needs_left = 1, .needs_top = 1, 86 .needs_topleft = 1 }, 87 [LEFT_DC_PRED] = { .needs_left = 1 }, 88 [TOP_DC_PRED] = { .needs_top = 1 }, 89 [DC_128_PRED] = { 0 }, 90 [DC_127_PRED] = { 0 }, 91 [DC_129_PRED] = { 0 } 92 }; 93 94 av_assert2(mode >= 0 && mode < 10); 95 mode = mode_conv[mode][have_left][have_top]; 96 if (edges[mode].needs_top) { 97 uint8_t *top, *topleft; 98 int n_px_need = 4 << tx, n_px_have = (((s->cols - col) << !ss_h) - x) * 4; 99 int n_px_need_tr = 0; 100 101 if (tx == TX_4X4 && edges[mode].needs_topright && have_right) 102 n_px_need_tr = 4; 103 104 // if top of sb64-row, use s->intra_pred_data[] instead of 105 // dst[-stride] for intra prediction (it contains pre- instead of 106 // post-loopfilter data) 107 if (have_top) { 108 top = !(row & 7) && !y ? 109 s->intra_pred_data[p] + (col * (8 >> ss_h) + x * 4) * bytesperpixel : 110 y == 0 ? &dst_edge[-stride_edge] : &dst_inner[-stride_inner]; 111 if (have_left) 112 topleft = !(row & 7) && !y ? 113 s->intra_pred_data[p] + (col * (8 >> ss_h) + x * 4) * bytesperpixel : 114 y == 0 || x == 0 ? &dst_edge[-stride_edge] : 115 &dst_inner[-stride_inner]; 116 } 117 118 if (have_top && 119 (!edges[mode].needs_topleft || (have_left && top == topleft)) && 120 (tx != TX_4X4 || !edges[mode].needs_topright || have_right) && 121 n_px_need + n_px_need_tr <= n_px_have) { 122 *a = top; 123 } else { 124 if (have_top) { 125 if (n_px_need <= n_px_have) { 126 memcpy(*a, top, n_px_need * bytesperpixel); 127 } else { 128#define memset_bpp(c, i1, v, i2, num) do { \ 129 if (bytesperpixel == 1) { \ 130 memset(&(c)[(i1)], (v)[(i2)], (num)); \ 131 } else { \ 132 int n, val = AV_RN16A(&(v)[(i2) * 2]); \ 133 for (n = 0; n < (num); n++) { \ 134 AV_WN16A(&(c)[((i1) + n) * 2], val); \ 135 } \ 136 } \ 137} while (0) 138 memcpy(*a, top, n_px_have * bytesperpixel); 139 memset_bpp(*a, n_px_have, (*a), n_px_have - 1, n_px_need - n_px_have); 140 } 141 } else { 142#define memset_val(c, val, num) do { \ 143 if (bytesperpixel == 1) { \ 144 memset((c), (val), (num)); \ 145 } else { \ 146 int n; \ 147 for (n = 0; n < (num); n++) { \ 148 AV_WN16A(&(c)[n * 2], (val)); \ 149 } \ 150 } \ 151} while (0) 152 memset_val(*a, (128 << (bpp - 8)) - 1, n_px_need); 153 } 154 if (edges[mode].needs_topleft) { 155 if (have_left && have_top) { 156#define assign_bpp(c, i1, v, i2) do { \ 157 if (bytesperpixel == 1) { \ 158 (c)[(i1)] = (v)[(i2)]; \ 159 } else { \ 160 AV_COPY16(&(c)[(i1) * 2], &(v)[(i2) * 2]); \ 161 } \ 162} while (0) 163 assign_bpp(*a, -1, topleft, -1); 164 } else { 165#define assign_val(c, i, v) do { \ 166 if (bytesperpixel == 1) { \ 167 (c)[(i)] = (v); \ 168 } else { \ 169 AV_WN16A(&(c)[(i) * 2], (v)); \ 170 } \ 171} while (0) 172 assign_val((*a), -1, (128 << (bpp - 8)) + (have_top ? +1 : -1)); 173 } 174 } 175 if (tx == TX_4X4 && edges[mode].needs_topright) { 176 if (have_top && have_right && 177 n_px_need + n_px_need_tr <= n_px_have) { 178 memcpy(&(*a)[4 * bytesperpixel], &top[4 * bytesperpixel], 4 * bytesperpixel); 179 } else { 180 memset_bpp(*a, 4, *a, 3, 4); 181 } 182 } 183 } 184 } 185 if (edges[mode].needs_left) { 186 if (have_left) { 187 int n_px_need = 4 << tx, i, n_px_have = (((s->rows - row) << !ss_v) - y) * 4; 188 uint8_t *dst = x == 0 ? dst_edge : dst_inner; 189 ptrdiff_t stride = x == 0 ? stride_edge : stride_inner; 190 191 if (edges[mode].invert_left) { 192 if (n_px_need <= n_px_have) { 193 for (i = 0; i < n_px_need; i++) 194 assign_bpp(l, i, &dst[i * stride], -1); 195 } else { 196 for (i = 0; i < n_px_have; i++) 197 assign_bpp(l, i, &dst[i * stride], -1); 198 memset_bpp(l, n_px_have, l, n_px_have - 1, n_px_need - n_px_have); 199 } 200 } else { 201 if (n_px_need <= n_px_have) { 202 for (i = 0; i < n_px_need; i++) 203 assign_bpp(l, n_px_need - 1 - i, &dst[i * stride], -1); 204 } else { 205 for (i = 0; i < n_px_have; i++) 206 assign_bpp(l, n_px_need - 1 - i, &dst[i * stride], -1); 207 memset_bpp(l, 0, l, n_px_need - n_px_have, n_px_need - n_px_have); 208 } 209 } 210 } else { 211 memset_val(l, (128 << (bpp - 8)) + 1, 4 << tx); 212 } 213 } 214 215 return mode; 216} 217 218static av_always_inline void intra_recon(VP9TileData *td, ptrdiff_t y_off, 219 ptrdiff_t uv_off, int bytesperpixel) 220{ 221 VP9Context *s = td->s; 222 VP9Block *b = td->b; 223 int row = td->row, col = td->col; 224 int w4 = ff_vp9_bwh_tab[1][b->bs][0] << 1, step1d = 1 << b->tx, n; 225 int h4 = ff_vp9_bwh_tab[1][b->bs][1] << 1, x, y, step = 1 << (b->tx * 2); 226 int end_x = FFMIN(2 * (s->cols - col), w4); 227 int end_y = FFMIN(2 * (s->rows - row), h4); 228 int tx = 4 * s->s.h.lossless + b->tx, uvtx = b->uvtx + 4 * s->s.h.lossless; 229 int uvstep1d = 1 << b->uvtx, p; 230 uint8_t *dst = td->dst[0], *dst_r = s->s.frames[CUR_FRAME].tf.f->data[0] + y_off; 231 LOCAL_ALIGNED_32(uint8_t, a_buf, [96]); 232 LOCAL_ALIGNED_32(uint8_t, l, [64]); 233 234 for (n = 0, y = 0; y < end_y; y += step1d) { 235 uint8_t *ptr = dst, *ptr_r = dst_r; 236 for (x = 0; x < end_x; x += step1d, ptr += 4 * step1d * bytesperpixel, 237 ptr_r += 4 * step1d * bytesperpixel, n += step) { 238 int mode = b->mode[b->bs > BS_8x8 && b->tx == TX_4X4 ? 239 y * 2 + x : 0]; 240 uint8_t *a = &a_buf[32]; 241 enum TxfmType txtp = ff_vp9_intra_txfm_type[mode]; 242 int eob = b->skip ? 0 : b->tx > TX_8X8 ? AV_RN16A(&td->eob[n]) : td->eob[n]; 243 244 mode = check_intra_mode(td, mode, &a, ptr_r, 245 s->s.frames[CUR_FRAME].tf.f->linesize[0], 246 ptr, td->y_stride, l, 247 col, x, w4, row, y, b->tx, 0, 0, 0, bytesperpixel); 248 s->dsp.intra_pred[b->tx][mode](ptr, td->y_stride, l, a); 249 if (eob) 250 s->dsp.itxfm_add[tx][txtp](ptr, td->y_stride, 251 td->block + 16 * n * bytesperpixel, eob); 252 } 253 dst_r += 4 * step1d * s->s.frames[CUR_FRAME].tf.f->linesize[0]; 254 dst += 4 * step1d * td->y_stride; 255 } 256 257 // U/V 258 w4 >>= s->ss_h; 259 end_x >>= s->ss_h; 260 end_y >>= s->ss_v; 261 step = 1 << (b->uvtx * 2); 262 for (p = 0; p < 2; p++) { 263 dst = td->dst[1 + p]; 264 dst_r = s->s.frames[CUR_FRAME].tf.f->data[1 + p] + uv_off; 265 for (n = 0, y = 0; y < end_y; y += uvstep1d) { 266 uint8_t *ptr = dst, *ptr_r = dst_r; 267 for (x = 0; x < end_x; x += uvstep1d, ptr += 4 * uvstep1d * bytesperpixel, 268 ptr_r += 4 * uvstep1d * bytesperpixel, n += step) { 269 int mode = b->uvmode; 270 uint8_t *a = &a_buf[32]; 271 int eob = b->skip ? 0 : b->uvtx > TX_8X8 ? AV_RN16A(&td->uveob[p][n]) : td->uveob[p][n]; 272 273 mode = check_intra_mode(td, mode, &a, ptr_r, 274 s->s.frames[CUR_FRAME].tf.f->linesize[1], 275 ptr, td->uv_stride, l, col, x, w4, row, y, 276 b->uvtx, p + 1, s->ss_h, s->ss_v, bytesperpixel); 277 s->dsp.intra_pred[b->uvtx][mode](ptr, td->uv_stride, l, a); 278 if (eob) 279 s->dsp.itxfm_add[uvtx][DCT_DCT](ptr, td->uv_stride, 280 td->uvblock[p] + 16 * n * bytesperpixel, eob); 281 } 282 dst_r += 4 * uvstep1d * s->s.frames[CUR_FRAME].tf.f->linesize[1]; 283 dst += 4 * uvstep1d * td->uv_stride; 284 } 285 } 286} 287 288void ff_vp9_intra_recon_8bpp(VP9TileData *td, ptrdiff_t y_off, ptrdiff_t uv_off) 289{ 290 intra_recon(td, y_off, uv_off, 1); 291} 292 293void ff_vp9_intra_recon_16bpp(VP9TileData *td, ptrdiff_t y_off, ptrdiff_t uv_off) 294{ 295 intra_recon(td, y_off, uv_off, 2); 296} 297 298static av_always_inline void mc_luma_unscaled(VP9TileData *td, vp9_mc_func (*mc)[2], 299 uint8_t *dst, ptrdiff_t dst_stride, 300 const uint8_t *ref, ptrdiff_t ref_stride, 301 ThreadFrame *ref_frame, 302 ptrdiff_t y, ptrdiff_t x, const VP56mv *mv, 303 int bw, int bh, int w, int h, int bytesperpixel) 304{ 305 VP9Context *s = td->s; 306 int mx = mv->x, my = mv->y, th; 307 308 y += my >> 3; 309 x += mx >> 3; 310 ref += y * ref_stride + x * bytesperpixel; 311 mx &= 7; 312 my &= 7; 313 // FIXME bilinear filter only needs 0/1 pixels, not 3/4 314 // we use +7 because the last 7 pixels of each sbrow can be changed in 315 // the longest loopfilter of the next sbrow 316 th = (y + bh + 4 * !!my + 7) >> 6; 317 ff_thread_await_progress(ref_frame, FFMAX(th, 0), 0); 318 // The arm/aarch64 _hv filters read one more row than what actually is 319 // needed, so switch to emulated edge one pixel sooner vertically 320 // (!!my * 5) than horizontally (!!mx * 4). 321 if (x < !!mx * 3 || y < !!my * 3 || 322 x + !!mx * 4 > w - bw || y + !!my * 5 > h - bh) { 323 s->vdsp.emulated_edge_mc(td->edge_emu_buffer, 324 ref - !!my * 3 * ref_stride - !!mx * 3 * bytesperpixel, 325 160, ref_stride, 326 bw + !!mx * 7, bh + !!my * 7, 327 x - !!mx * 3, y - !!my * 3, w, h); 328 ref = td->edge_emu_buffer + !!my * 3 * 160 + !!mx * 3 * bytesperpixel; 329 ref_stride = 160; 330 } 331 mc[!!mx][!!my](dst, dst_stride, ref, ref_stride, bh, mx << 1, my << 1); 332} 333 334static av_always_inline void mc_chroma_unscaled(VP9TileData *td, vp9_mc_func (*mc)[2], 335 uint8_t *dst_u, uint8_t *dst_v, 336 ptrdiff_t dst_stride, 337 const uint8_t *ref_u, ptrdiff_t src_stride_u, 338 const uint8_t *ref_v, ptrdiff_t src_stride_v, 339 ThreadFrame *ref_frame, 340 ptrdiff_t y, ptrdiff_t x, const VP56mv *mv, 341 int bw, int bh, int w, int h, int bytesperpixel) 342{ 343 VP9Context *s = td->s; 344 int mx = mv->x * (1 << !s->ss_h), my = mv->y * (1 << !s->ss_v), th; 345 346 y += my >> 4; 347 x += mx >> 4; 348 ref_u += y * src_stride_u + x * bytesperpixel; 349 ref_v += y * src_stride_v + x * bytesperpixel; 350 mx &= 15; 351 my &= 15; 352 // FIXME bilinear filter only needs 0/1 pixels, not 3/4 353 // we use +7 because the last 7 pixels of each sbrow can be changed in 354 // the longest loopfilter of the next sbrow 355 th = (y + bh + 4 * !!my + 7) >> (6 - s->ss_v); 356 ff_thread_await_progress(ref_frame, FFMAX(th, 0), 0); 357 // The arm/aarch64 _hv filters read one more row than what actually is 358 // needed, so switch to emulated edge one pixel sooner vertically 359 // (!!my * 5) than horizontally (!!mx * 4). 360 if (x < !!mx * 3 || y < !!my * 3 || 361 x + !!mx * 4 > w - bw || y + !!my * 5 > h - bh) { 362 s->vdsp.emulated_edge_mc(td->edge_emu_buffer, 363 ref_u - !!my * 3 * src_stride_u - !!mx * 3 * bytesperpixel, 364 160, src_stride_u, 365 bw + !!mx * 7, bh + !!my * 7, 366 x - !!mx * 3, y - !!my * 3, w, h); 367 ref_u = td->edge_emu_buffer + !!my * 3 * 160 + !!mx * 3 * bytesperpixel; 368 mc[!!mx][!!my](dst_u, dst_stride, ref_u, 160, bh, mx, my); 369 370 s->vdsp.emulated_edge_mc(td->edge_emu_buffer, 371 ref_v - !!my * 3 * src_stride_v - !!mx * 3 * bytesperpixel, 372 160, src_stride_v, 373 bw + !!mx * 7, bh + !!my * 7, 374 x - !!mx * 3, y - !!my * 3, w, h); 375 ref_v = td->edge_emu_buffer + !!my * 3 * 160 + !!mx * 3 * bytesperpixel; 376 mc[!!mx][!!my](dst_v, dst_stride, ref_v, 160, bh, mx, my); 377 } else { 378 mc[!!mx][!!my](dst_u, dst_stride, ref_u, src_stride_u, bh, mx, my); 379 mc[!!mx][!!my](dst_v, dst_stride, ref_v, src_stride_v, bh, mx, my); 380 } 381} 382 383#define mc_luma_dir(td, mc, dst, dst_ls, src, src_ls, tref, row, col, mv, \ 384 px, py, pw, ph, bw, bh, w, h, i) \ 385 mc_luma_unscaled(td, s->dsp.mc, dst, dst_ls, src, src_ls, tref, row, col, \ 386 mv, bw, bh, w, h, bytesperpixel) 387#define mc_chroma_dir(td, mc, dstu, dstv, dst_ls, srcu, srcu_ls, srcv, srcv_ls, tref, \ 388 row, col, mv, px, py, pw, ph, bw, bh, w, h, i) \ 389 mc_chroma_unscaled(td, s->dsp.mc, dstu, dstv, dst_ls, srcu, srcu_ls, srcv, srcv_ls, tref, \ 390 row, col, mv, bw, bh, w, h, bytesperpixel) 391#define SCALED 0 392#define FN(x) x##_8bpp 393#define BYTES_PER_PIXEL 1 394#include "vp9_mc_template.c" 395#undef FN 396#undef BYTES_PER_PIXEL 397#define FN(x) x##_16bpp 398#define BYTES_PER_PIXEL 2 399#include "vp9_mc_template.c" 400#undef mc_luma_dir 401#undef mc_chroma_dir 402#undef FN 403#undef BYTES_PER_PIXEL 404#undef SCALED 405 406static av_always_inline void mc_luma_scaled(VP9TileData *td, vp9_scaled_mc_func smc, 407 vp9_mc_func (*mc)[2], 408 uint8_t *dst, ptrdiff_t dst_stride, 409 const uint8_t *ref, ptrdiff_t ref_stride, 410 ThreadFrame *ref_frame, 411 ptrdiff_t y, ptrdiff_t x, const VP56mv *in_mv, 412 int px, int py, int pw, int ph, 413 int bw, int bh, int w, int h, int bytesperpixel, 414 const uint16_t *scale, const uint8_t *step) 415{ 416 VP9Context *s = td->s; 417 if (s->s.frames[CUR_FRAME].tf.f->width == ref_frame->f->width && 418 s->s.frames[CUR_FRAME].tf.f->height == ref_frame->f->height) { 419 mc_luma_unscaled(td, mc, dst, dst_stride, ref, ref_stride, ref_frame, 420 y, x, in_mv, bw, bh, w, h, bytesperpixel); 421 } else { 422#define scale_mv(n, dim) (((int64_t)(n) * scale[dim]) >> 14) 423 int mx, my; 424 int refbw_m1, refbh_m1; 425 int th; 426 VP56mv mv; 427 428 mv.x = av_clip(in_mv->x, -(x + pw - px + 4) * 8, (s->cols * 8 - x + px + 3) * 8); 429 mv.y = av_clip(in_mv->y, -(y + ph - py + 4) * 8, (s->rows * 8 - y + py + 3) * 8); 430 // BUG libvpx seems to scale the two components separately. This introduces 431 // rounding errors but we have to reproduce them to be exactly compatible 432 // with the output from libvpx... 433 mx = scale_mv(mv.x * 2, 0) + scale_mv(x * 16, 0); 434 my = scale_mv(mv.y * 2, 1) + scale_mv(y * 16, 1); 435 436 y = my >> 4; 437 x = mx >> 4; 438 ref += y * ref_stride + x * bytesperpixel; 439 mx &= 15; 440 my &= 15; 441 refbw_m1 = ((bw - 1) * step[0] + mx) >> 4; 442 refbh_m1 = ((bh - 1) * step[1] + my) >> 4; 443 // FIXME bilinear filter only needs 0/1 pixels, not 3/4 444 // we use +7 because the last 7 pixels of each sbrow can be changed in 445 // the longest loopfilter of the next sbrow 446 th = (y + refbh_m1 + 4 + 7) >> 6; 447 ff_thread_await_progress(ref_frame, FFMAX(th, 0), 0); 448 // The arm/aarch64 _hv filters read one more row than what actually is 449 // needed, so switch to emulated edge one pixel sooner vertically 450 // (y + 5 >= h - refbh_m1) than horizontally (x + 4 >= w - refbw_m1). 451 if (x < 3 || y < 3 || x + 4 >= w - refbw_m1 || y + 5 >= h - refbh_m1) { 452 s->vdsp.emulated_edge_mc(td->edge_emu_buffer, 453 ref - 3 * ref_stride - 3 * bytesperpixel, 454 288, ref_stride, 455 refbw_m1 + 8, refbh_m1 + 8, 456 x - 3, y - 3, w, h); 457 ref = td->edge_emu_buffer + 3 * 288 + 3 * bytesperpixel; 458 ref_stride = 288; 459 } 460 smc(dst, dst_stride, ref, ref_stride, bh, mx, my, step[0], step[1]); 461 } 462} 463 464static av_always_inline void mc_chroma_scaled(VP9TileData *td, vp9_scaled_mc_func smc, 465 vp9_mc_func (*mc)[2], 466 uint8_t *dst_u, uint8_t *dst_v, 467 ptrdiff_t dst_stride, 468 const uint8_t *ref_u, ptrdiff_t src_stride_u, 469 const uint8_t *ref_v, ptrdiff_t src_stride_v, 470 ThreadFrame *ref_frame, 471 ptrdiff_t y, ptrdiff_t x, const VP56mv *in_mv, 472 int px, int py, int pw, int ph, 473 int bw, int bh, int w, int h, int bytesperpixel, 474 const uint16_t *scale, const uint8_t *step) 475{ 476 VP9Context *s = td->s; 477 if (s->s.frames[CUR_FRAME].tf.f->width == ref_frame->f->width && 478 s->s.frames[CUR_FRAME].tf.f->height == ref_frame->f->height) { 479 mc_chroma_unscaled(td, mc, dst_u, dst_v, dst_stride, ref_u, src_stride_u, 480 ref_v, src_stride_v, ref_frame, 481 y, x, in_mv, bw, bh, w, h, bytesperpixel); 482 } else { 483 int mx, my; 484 int refbw_m1, refbh_m1; 485 int th; 486 VP56mv mv; 487 488 if (s->ss_h) { 489 // BUG https://code.google.com/p/webm/issues/detail?id=820 490 mv.x = av_clip(in_mv->x, -(x + pw - px + 4) * 16, (s->cols * 4 - x + px + 3) * 16); 491 mx = scale_mv(mv.x, 0) + (scale_mv(x * 16, 0) & ~15) + (scale_mv(x * 32, 0) & 15); 492 } else { 493 mv.x = av_clip(in_mv->x, -(x + pw - px + 4) * 8, (s->cols * 8 - x + px + 3) * 8); 494 mx = scale_mv(mv.x * 2, 0) + scale_mv(x * 16, 0); 495 } 496 if (s->ss_v) { 497 // BUG https://code.google.com/p/webm/issues/detail?id=820 498 mv.y = av_clip(in_mv->y, -(y + ph - py + 4) * 16, (s->rows * 4 - y + py + 3) * 16); 499 my = scale_mv(mv.y, 1) + (scale_mv(y * 16, 1) & ~15) + (scale_mv(y * 32, 1) & 15); 500 } else { 501 mv.y = av_clip(in_mv->y, -(y + ph - py + 4) * 8, (s->rows * 8 - y + py + 3) * 8); 502 my = scale_mv(mv.y * 2, 1) + scale_mv(y * 16, 1); 503 } 504#undef scale_mv 505 y = my >> 4; 506 x = mx >> 4; 507 ref_u += y * src_stride_u + x * bytesperpixel; 508 ref_v += y * src_stride_v + x * bytesperpixel; 509 mx &= 15; 510 my &= 15; 511 refbw_m1 = ((bw - 1) * step[0] + mx) >> 4; 512 refbh_m1 = ((bh - 1) * step[1] + my) >> 4; 513 // FIXME bilinear filter only needs 0/1 pixels, not 3/4 514 // we use +7 because the last 7 pixels of each sbrow can be changed in 515 // the longest loopfilter of the next sbrow 516 th = (y + refbh_m1 + 4 + 7) >> (6 - s->ss_v); 517 ff_thread_await_progress(ref_frame, FFMAX(th, 0), 0); 518 // The arm/aarch64 _hv filters read one more row than what actually is 519 // needed, so switch to emulated edge one pixel sooner vertically 520 // (y + 5 >= h - refbh_m1) than horizontally (x + 4 >= w - refbw_m1). 521 if (x < 3 || y < 3 || x + 4 >= w - refbw_m1 || y + 5 >= h - refbh_m1) { 522 s->vdsp.emulated_edge_mc(td->edge_emu_buffer, 523 ref_u - 3 * src_stride_u - 3 * bytesperpixel, 524 288, src_stride_u, 525 refbw_m1 + 8, refbh_m1 + 8, 526 x - 3, y - 3, w, h); 527 ref_u = td->edge_emu_buffer + 3 * 288 + 3 * bytesperpixel; 528 smc(dst_u, dst_stride, ref_u, 288, bh, mx, my, step[0], step[1]); 529 530 s->vdsp.emulated_edge_mc(td->edge_emu_buffer, 531 ref_v - 3 * src_stride_v - 3 * bytesperpixel, 532 288, src_stride_v, 533 refbw_m1 + 8, refbh_m1 + 8, 534 x - 3, y - 3, w, h); 535 ref_v = td->edge_emu_buffer + 3 * 288 + 3 * bytesperpixel; 536 smc(dst_v, dst_stride, ref_v, 288, bh, mx, my, step[0], step[1]); 537 } else { 538 smc(dst_u, dst_stride, ref_u, src_stride_u, bh, mx, my, step[0], step[1]); 539 smc(dst_v, dst_stride, ref_v, src_stride_v, bh, mx, my, step[0], step[1]); 540 } 541 } 542} 543 544#define mc_luma_dir(td, mc, dst, dst_ls, src, src_ls, tref, row, col, mv, \ 545 px, py, pw, ph, bw, bh, w, h, i) \ 546 mc_luma_scaled(td, s->dsp.s##mc, s->dsp.mc, dst, dst_ls, src, src_ls, tref, row, col, \ 547 mv, px, py, pw, ph, bw, bh, w, h, bytesperpixel, \ 548 s->mvscale[b->ref[i]], s->mvstep[b->ref[i]]) 549#define mc_chroma_dir(td, mc, dstu, dstv, dst_ls, srcu, srcu_ls, srcv, srcv_ls, tref, \ 550 row, col, mv, px, py, pw, ph, bw, bh, w, h, i) \ 551 mc_chroma_scaled(td, s->dsp.s##mc, s->dsp.mc, dstu, dstv, dst_ls, srcu, srcu_ls, srcv, srcv_ls, tref, \ 552 row, col, mv, px, py, pw, ph, bw, bh, w, h, bytesperpixel, \ 553 s->mvscale[b->ref[i]], s->mvstep[b->ref[i]]) 554#define SCALED 1 555#define FN(x) x##_scaled_8bpp 556#define BYTES_PER_PIXEL 1 557#include "vp9_mc_template.c" 558#undef FN 559#undef BYTES_PER_PIXEL 560#define FN(x) x##_scaled_16bpp 561#define BYTES_PER_PIXEL 2 562#include "vp9_mc_template.c" 563#undef mc_luma_dir 564#undef mc_chroma_dir 565#undef FN 566#undef BYTES_PER_PIXEL 567#undef SCALED 568 569static av_always_inline void inter_recon(VP9TileData *td, int bytesperpixel) 570{ 571 VP9Context *s = td->s; 572 VP9Block *b = td->b; 573 int row = td->row, col = td->col; 574 575 if (s->mvscale[b->ref[0]][0] == REF_INVALID_SCALE || 576 (b->comp && s->mvscale[b->ref[1]][0] == REF_INVALID_SCALE)) { 577 if (!s->td->error_info) { 578 s->td->error_info = AVERROR_INVALIDDATA; 579 av_log(NULL, AV_LOG_ERROR, "Bitstream not supported, " 580 "reference frame has invalid dimensions\n"); 581 } 582 return; 583 } 584 585 if (s->mvscale[b->ref[0]][0] || (b->comp && s->mvscale[b->ref[1]][0])) { 586 if (bytesperpixel == 1) { 587 inter_pred_scaled_8bpp(td); 588 } else { 589 inter_pred_scaled_16bpp(td); 590 } 591 } else { 592 if (bytesperpixel == 1) { 593 inter_pred_8bpp(td); 594 } else { 595 inter_pred_16bpp(td); 596 } 597 } 598 599 if (!b->skip) { 600 /* mostly copied intra_recon() */ 601 602 int w4 = ff_vp9_bwh_tab[1][b->bs][0] << 1, step1d = 1 << b->tx, n; 603 int h4 = ff_vp9_bwh_tab[1][b->bs][1] << 1, x, y, step = 1 << (b->tx * 2); 604 int end_x = FFMIN(2 * (s->cols - col), w4); 605 int end_y = FFMIN(2 * (s->rows - row), h4); 606 int tx = 4 * s->s.h.lossless + b->tx, uvtx = b->uvtx + 4 * s->s.h.lossless; 607 int uvstep1d = 1 << b->uvtx, p; 608 uint8_t *dst = td->dst[0]; 609 610 // y itxfm add 611 for (n = 0, y = 0; y < end_y; y += step1d) { 612 uint8_t *ptr = dst; 613 for (x = 0; x < end_x; x += step1d, 614 ptr += 4 * step1d * bytesperpixel, n += step) { 615 int eob = b->tx > TX_8X8 ? AV_RN16A(&td->eob[n]) : td->eob[n]; 616 617 if (eob) 618 s->dsp.itxfm_add[tx][DCT_DCT](ptr, td->y_stride, 619 td->block + 16 * n * bytesperpixel, eob); 620 } 621 dst += 4 * td->y_stride * step1d; 622 } 623 624 // uv itxfm add 625 end_x >>= s->ss_h; 626 end_y >>= s->ss_v; 627 step = 1 << (b->uvtx * 2); 628 for (p = 0; p < 2; p++) { 629 dst = td->dst[p + 1]; 630 for (n = 0, y = 0; y < end_y; y += uvstep1d) { 631 uint8_t *ptr = dst; 632 for (x = 0; x < end_x; x += uvstep1d, 633 ptr += 4 * uvstep1d * bytesperpixel, n += step) { 634 int eob = b->uvtx > TX_8X8 ? AV_RN16A(&td->uveob[p][n]) : td->uveob[p][n]; 635 636 if (eob) 637 s->dsp.itxfm_add[uvtx][DCT_DCT](ptr, td->uv_stride, 638 td->uvblock[p] + 16 * n * bytesperpixel, eob); 639 } 640 dst += 4 * uvstep1d * td->uv_stride; 641 } 642 } 643 } 644} 645 646void ff_vp9_inter_recon_8bpp(VP9TileData *td) 647{ 648 inter_recon(td, 1); 649} 650 651void ff_vp9_inter_recon_16bpp(VP9TileData *td) 652{ 653 inter_recon(td, 2); 654} 655