xref: /third_party/ffmpeg/libavcodec/vp9recon.c (revision cabdff1a)
1/*
2 * VP9 compatible video decoder
3 *
4 * Copyright (C) 2013 Ronald S. Bultje <rsbultje gmail com>
5 * Copyright (C) 2013 Clément Bœsch <u pkh me>
6 *
7 * This file is part of FFmpeg.
8 *
9 * FFmpeg is free software; you can redistribute it and/or
10 * modify it under the terms of the GNU Lesser General Public
11 * License as published by the Free Software Foundation; either
12 * version 2.1 of the License, or (at your option) any later version.
13 *
14 * FFmpeg is distributed in the hope that it will be useful,
15 * but WITHOUT ANY WARRANTY; without even the implied warranty of
16 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
17 * Lesser General Public License for more details.
18 *
19 * You should have received a copy of the GNU Lesser General Public
20 * License along with FFmpeg; if not, write to the Free Software
21 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
22 */
23
24#include "libavutil/avassert.h"
25#include "libavutil/mem_internal.h"
26
27#include "threadframe.h"
28#include "videodsp.h"
29#include "vp9data.h"
30#include "vp9dec.h"
31
32static av_always_inline int check_intra_mode(VP9TileData *td, int mode, uint8_t **a,
33                                             uint8_t *dst_edge, ptrdiff_t stride_edge,
34                                             uint8_t *dst_inner, ptrdiff_t stride_inner,
35                                             uint8_t *l, int col, int x, int w,
36                                             int row, int y, enum TxfmMode tx,
37                                             int p, int ss_h, int ss_v, int bytesperpixel)
38{
39    VP9Context *s = td->s;
40    int have_top = row > 0 || y > 0;
41    int have_left = col > td->tile_col_start || x > 0;
42    int have_right = x < w - 1;
43    int bpp = s->s.h.bpp;
44    static const uint8_t mode_conv[10][2 /* have_left */][2 /* have_top */] = {
45        [VERT_PRED]            = { { DC_127_PRED,          VERT_PRED            },
46                                   { DC_127_PRED,          VERT_PRED            } },
47        [HOR_PRED]             = { { DC_129_PRED,          DC_129_PRED          },
48                                   { HOR_PRED,             HOR_PRED             } },
49        [DC_PRED]              = { { DC_128_PRED,          TOP_DC_PRED          },
50                                   { LEFT_DC_PRED,         DC_PRED              } },
51        [DIAG_DOWN_LEFT_PRED]  = { { DC_127_PRED,          DIAG_DOWN_LEFT_PRED  },
52                                   { DC_127_PRED,          DIAG_DOWN_LEFT_PRED  } },
53        [DIAG_DOWN_RIGHT_PRED] = { { DIAG_DOWN_RIGHT_PRED, DIAG_DOWN_RIGHT_PRED },
54                                   { DIAG_DOWN_RIGHT_PRED, DIAG_DOWN_RIGHT_PRED } },
55        [VERT_RIGHT_PRED]      = { { VERT_RIGHT_PRED,      VERT_RIGHT_PRED      },
56                                   { VERT_RIGHT_PRED,      VERT_RIGHT_PRED      } },
57        [HOR_DOWN_PRED]        = { { HOR_DOWN_PRED,        HOR_DOWN_PRED        },
58                                   { HOR_DOWN_PRED,        HOR_DOWN_PRED        } },
59        [VERT_LEFT_PRED]       = { { DC_127_PRED,          VERT_LEFT_PRED       },
60                                   { DC_127_PRED,          VERT_LEFT_PRED       } },
61        [HOR_UP_PRED]          = { { DC_129_PRED,          DC_129_PRED          },
62                                   { HOR_UP_PRED,          HOR_UP_PRED          } },
63        [TM_VP8_PRED]          = { { DC_129_PRED,          VERT_PRED            },
64                                   { HOR_PRED,             TM_VP8_PRED          } },
65    };
66    static const struct {
67        uint8_t needs_left:1;
68        uint8_t needs_top:1;
69        uint8_t needs_topleft:1;
70        uint8_t needs_topright:1;
71        uint8_t invert_left:1;
72    } edges[N_INTRA_PRED_MODES] = {
73        [VERT_PRED]            = { .needs_top  = 1 },
74        [HOR_PRED]             = { .needs_left = 1 },
75        [DC_PRED]              = { .needs_top  = 1, .needs_left = 1 },
76        [DIAG_DOWN_LEFT_PRED]  = { .needs_top  = 1, .needs_topright = 1 },
77        [DIAG_DOWN_RIGHT_PRED] = { .needs_left = 1, .needs_top = 1,
78                                   .needs_topleft = 1 },
79        [VERT_RIGHT_PRED]      = { .needs_left = 1, .needs_top = 1,
80                                   .needs_topleft = 1 },
81        [HOR_DOWN_PRED]        = { .needs_left = 1, .needs_top = 1,
82                                   .needs_topleft = 1 },
83        [VERT_LEFT_PRED]       = { .needs_top  = 1, .needs_topright = 1 },
84        [HOR_UP_PRED]          = { .needs_left = 1, .invert_left = 1 },
85        [TM_VP8_PRED]          = { .needs_left = 1, .needs_top = 1,
86                                   .needs_topleft = 1 },
87        [LEFT_DC_PRED]         = { .needs_left = 1 },
88        [TOP_DC_PRED]          = { .needs_top  = 1 },
89        [DC_128_PRED]          = { 0 },
90        [DC_127_PRED]          = { 0 },
91        [DC_129_PRED]          = { 0 }
92    };
93
94    av_assert2(mode >= 0 && mode < 10);
95    mode = mode_conv[mode][have_left][have_top];
96    if (edges[mode].needs_top) {
97        uint8_t *top, *topleft;
98        int n_px_need = 4 << tx, n_px_have = (((s->cols - col) << !ss_h) - x) * 4;
99        int n_px_need_tr = 0;
100
101        if (tx == TX_4X4 && edges[mode].needs_topright && have_right)
102            n_px_need_tr = 4;
103
104        // if top of sb64-row, use s->intra_pred_data[] instead of
105        // dst[-stride] for intra prediction (it contains pre- instead of
106        // post-loopfilter data)
107        if (have_top) {
108            top = !(row & 7) && !y ?
109                s->intra_pred_data[p] + (col * (8 >> ss_h) + x * 4) * bytesperpixel :
110                y == 0 ? &dst_edge[-stride_edge] : &dst_inner[-stride_inner];
111            if (have_left)
112                topleft = !(row & 7) && !y ?
113                    s->intra_pred_data[p] + (col * (8 >> ss_h) + x * 4) * bytesperpixel :
114                    y == 0 || x == 0 ? &dst_edge[-stride_edge] :
115                    &dst_inner[-stride_inner];
116        }
117
118        if (have_top &&
119            (!edges[mode].needs_topleft || (have_left && top == topleft)) &&
120            (tx != TX_4X4 || !edges[mode].needs_topright || have_right) &&
121            n_px_need + n_px_need_tr <= n_px_have) {
122            *a = top;
123        } else {
124            if (have_top) {
125                if (n_px_need <= n_px_have) {
126                    memcpy(*a, top, n_px_need * bytesperpixel);
127                } else {
128#define memset_bpp(c, i1, v, i2, num) do { \
129    if (bytesperpixel == 1) { \
130        memset(&(c)[(i1)], (v)[(i2)], (num)); \
131    } else { \
132        int n, val = AV_RN16A(&(v)[(i2) * 2]); \
133        for (n = 0; n < (num); n++) { \
134            AV_WN16A(&(c)[((i1) + n) * 2], val); \
135        } \
136    } \
137} while (0)
138                    memcpy(*a, top, n_px_have * bytesperpixel);
139                    memset_bpp(*a, n_px_have, (*a), n_px_have - 1, n_px_need - n_px_have);
140                }
141            } else {
142#define memset_val(c, val, num) do { \
143    if (bytesperpixel == 1) { \
144        memset((c), (val), (num)); \
145    } else { \
146        int n; \
147        for (n = 0; n < (num); n++) { \
148            AV_WN16A(&(c)[n * 2], (val)); \
149        } \
150    } \
151} while (0)
152                memset_val(*a, (128 << (bpp - 8)) - 1, n_px_need);
153            }
154            if (edges[mode].needs_topleft) {
155                if (have_left && have_top) {
156#define assign_bpp(c, i1, v, i2) do { \
157    if (bytesperpixel == 1) { \
158        (c)[(i1)] = (v)[(i2)]; \
159    } else { \
160        AV_COPY16(&(c)[(i1) * 2], &(v)[(i2) * 2]); \
161    } \
162} while (0)
163                    assign_bpp(*a, -1, topleft, -1);
164                } else {
165#define assign_val(c, i, v) do { \
166    if (bytesperpixel == 1) { \
167        (c)[(i)] = (v); \
168    } else { \
169        AV_WN16A(&(c)[(i) * 2], (v)); \
170    } \
171} while (0)
172                    assign_val((*a), -1, (128 << (bpp - 8)) + (have_top ? +1 : -1));
173                }
174            }
175            if (tx == TX_4X4 && edges[mode].needs_topright) {
176                if (have_top && have_right &&
177                    n_px_need + n_px_need_tr <= n_px_have) {
178                    memcpy(&(*a)[4 * bytesperpixel], &top[4 * bytesperpixel], 4 * bytesperpixel);
179                } else {
180                    memset_bpp(*a, 4, *a, 3, 4);
181                }
182            }
183        }
184    }
185    if (edges[mode].needs_left) {
186        if (have_left) {
187            int n_px_need = 4 << tx, i, n_px_have = (((s->rows - row) << !ss_v) - y) * 4;
188            uint8_t *dst = x == 0 ? dst_edge : dst_inner;
189            ptrdiff_t stride = x == 0 ? stride_edge : stride_inner;
190
191            if (edges[mode].invert_left) {
192                if (n_px_need <= n_px_have) {
193                    for (i = 0; i < n_px_need; i++)
194                        assign_bpp(l, i, &dst[i * stride], -1);
195                } else {
196                    for (i = 0; i < n_px_have; i++)
197                        assign_bpp(l, i, &dst[i * stride], -1);
198                    memset_bpp(l, n_px_have, l, n_px_have - 1, n_px_need - n_px_have);
199                }
200            } else {
201                if (n_px_need <= n_px_have) {
202                    for (i = 0; i < n_px_need; i++)
203                        assign_bpp(l, n_px_need - 1 - i, &dst[i * stride], -1);
204                } else {
205                    for (i = 0; i < n_px_have; i++)
206                        assign_bpp(l, n_px_need - 1 - i, &dst[i * stride], -1);
207                    memset_bpp(l, 0, l, n_px_need - n_px_have, n_px_need - n_px_have);
208                }
209            }
210        } else {
211            memset_val(l, (128 << (bpp - 8)) + 1, 4 << tx);
212        }
213    }
214
215    return mode;
216}
217
218static av_always_inline void intra_recon(VP9TileData *td, ptrdiff_t y_off,
219                                         ptrdiff_t uv_off, int bytesperpixel)
220{
221    VP9Context *s = td->s;
222    VP9Block *b = td->b;
223    int row = td->row, col = td->col;
224    int w4 = ff_vp9_bwh_tab[1][b->bs][0] << 1, step1d = 1 << b->tx, n;
225    int h4 = ff_vp9_bwh_tab[1][b->bs][1] << 1, x, y, step = 1 << (b->tx * 2);
226    int end_x = FFMIN(2 * (s->cols - col), w4);
227    int end_y = FFMIN(2 * (s->rows - row), h4);
228    int tx = 4 * s->s.h.lossless + b->tx, uvtx = b->uvtx + 4 * s->s.h.lossless;
229    int uvstep1d = 1 << b->uvtx, p;
230    uint8_t *dst = td->dst[0], *dst_r = s->s.frames[CUR_FRAME].tf.f->data[0] + y_off;
231    LOCAL_ALIGNED_32(uint8_t, a_buf, [96]);
232    LOCAL_ALIGNED_32(uint8_t, l, [64]);
233
234    for (n = 0, y = 0; y < end_y; y += step1d) {
235        uint8_t *ptr = dst, *ptr_r = dst_r;
236        for (x = 0; x < end_x; x += step1d, ptr += 4 * step1d * bytesperpixel,
237                               ptr_r += 4 * step1d * bytesperpixel, n += step) {
238            int mode = b->mode[b->bs > BS_8x8 && b->tx == TX_4X4 ?
239                               y * 2 + x : 0];
240            uint8_t *a = &a_buf[32];
241            enum TxfmType txtp = ff_vp9_intra_txfm_type[mode];
242            int eob = b->skip ? 0 : b->tx > TX_8X8 ? AV_RN16A(&td->eob[n]) : td->eob[n];
243
244            mode = check_intra_mode(td, mode, &a, ptr_r,
245                                    s->s.frames[CUR_FRAME].tf.f->linesize[0],
246                                    ptr, td->y_stride, l,
247                                    col, x, w4, row, y, b->tx, 0, 0, 0, bytesperpixel);
248            s->dsp.intra_pred[b->tx][mode](ptr, td->y_stride, l, a);
249            if (eob)
250                s->dsp.itxfm_add[tx][txtp](ptr, td->y_stride,
251                                           td->block + 16 * n * bytesperpixel, eob);
252        }
253        dst_r += 4 * step1d * s->s.frames[CUR_FRAME].tf.f->linesize[0];
254        dst   += 4 * step1d * td->y_stride;
255    }
256
257    // U/V
258    w4    >>= s->ss_h;
259    end_x >>= s->ss_h;
260    end_y >>= s->ss_v;
261    step = 1 << (b->uvtx * 2);
262    for (p = 0; p < 2; p++) {
263        dst   = td->dst[1 + p];
264        dst_r = s->s.frames[CUR_FRAME].tf.f->data[1 + p] + uv_off;
265        for (n = 0, y = 0; y < end_y; y += uvstep1d) {
266            uint8_t *ptr = dst, *ptr_r = dst_r;
267            for (x = 0; x < end_x; x += uvstep1d, ptr += 4 * uvstep1d * bytesperpixel,
268                                   ptr_r += 4 * uvstep1d * bytesperpixel, n += step) {
269                int mode = b->uvmode;
270                uint8_t *a = &a_buf[32];
271                int eob = b->skip ? 0 : b->uvtx > TX_8X8 ? AV_RN16A(&td->uveob[p][n]) : td->uveob[p][n];
272
273                mode = check_intra_mode(td, mode, &a, ptr_r,
274                                        s->s.frames[CUR_FRAME].tf.f->linesize[1],
275                                        ptr, td->uv_stride, l, col, x, w4, row, y,
276                                        b->uvtx, p + 1, s->ss_h, s->ss_v, bytesperpixel);
277                s->dsp.intra_pred[b->uvtx][mode](ptr, td->uv_stride, l, a);
278                if (eob)
279                    s->dsp.itxfm_add[uvtx][DCT_DCT](ptr, td->uv_stride,
280                                                    td->uvblock[p] + 16 * n * bytesperpixel, eob);
281            }
282            dst_r += 4 * uvstep1d * s->s.frames[CUR_FRAME].tf.f->linesize[1];
283            dst   += 4 * uvstep1d * td->uv_stride;
284        }
285    }
286}
287
288void ff_vp9_intra_recon_8bpp(VP9TileData *td, ptrdiff_t y_off, ptrdiff_t uv_off)
289{
290    intra_recon(td, y_off, uv_off, 1);
291}
292
293void ff_vp9_intra_recon_16bpp(VP9TileData *td, ptrdiff_t y_off, ptrdiff_t uv_off)
294{
295    intra_recon(td, y_off, uv_off, 2);
296}
297
298static av_always_inline void mc_luma_unscaled(VP9TileData *td, vp9_mc_func (*mc)[2],
299                                              uint8_t *dst, ptrdiff_t dst_stride,
300                                              const uint8_t *ref, ptrdiff_t ref_stride,
301                                              ThreadFrame *ref_frame,
302                                              ptrdiff_t y, ptrdiff_t x, const VP56mv *mv,
303                                              int bw, int bh, int w, int h, int bytesperpixel)
304{
305    VP9Context *s = td->s;
306    int mx = mv->x, my = mv->y, th;
307
308    y += my >> 3;
309    x += mx >> 3;
310    ref += y * ref_stride + x * bytesperpixel;
311    mx &= 7;
312    my &= 7;
313    // FIXME bilinear filter only needs 0/1 pixels, not 3/4
314    // we use +7 because the last 7 pixels of each sbrow can be changed in
315    // the longest loopfilter of the next sbrow
316    th = (y + bh + 4 * !!my + 7) >> 6;
317    ff_thread_await_progress(ref_frame, FFMAX(th, 0), 0);
318    // The arm/aarch64 _hv filters read one more row than what actually is
319    // needed, so switch to emulated edge one pixel sooner vertically
320    // (!!my * 5) than horizontally (!!mx * 4).
321    if (x < !!mx * 3 || y < !!my * 3 ||
322        x + !!mx * 4 > w - bw || y + !!my * 5 > h - bh) {
323        s->vdsp.emulated_edge_mc(td->edge_emu_buffer,
324                                 ref - !!my * 3 * ref_stride - !!mx * 3 * bytesperpixel,
325                                 160, ref_stride,
326                                 bw + !!mx * 7, bh + !!my * 7,
327                                 x - !!mx * 3, y - !!my * 3, w, h);
328        ref = td->edge_emu_buffer + !!my * 3 * 160 + !!mx * 3 * bytesperpixel;
329        ref_stride = 160;
330    }
331    mc[!!mx][!!my](dst, dst_stride, ref, ref_stride, bh, mx << 1, my << 1);
332}
333
334static av_always_inline void mc_chroma_unscaled(VP9TileData *td, vp9_mc_func (*mc)[2],
335                                                uint8_t *dst_u, uint8_t *dst_v,
336                                                ptrdiff_t dst_stride,
337                                                const uint8_t *ref_u, ptrdiff_t src_stride_u,
338                                                const uint8_t *ref_v, ptrdiff_t src_stride_v,
339                                                ThreadFrame *ref_frame,
340                                                ptrdiff_t y, ptrdiff_t x, const VP56mv *mv,
341                                                int bw, int bh, int w, int h, int bytesperpixel)
342{
343    VP9Context *s = td->s;
344    int mx = mv->x * (1 << !s->ss_h), my = mv->y * (1 << !s->ss_v), th;
345
346    y += my >> 4;
347    x += mx >> 4;
348    ref_u += y * src_stride_u + x * bytesperpixel;
349    ref_v += y * src_stride_v + x * bytesperpixel;
350    mx &= 15;
351    my &= 15;
352    // FIXME bilinear filter only needs 0/1 pixels, not 3/4
353    // we use +7 because the last 7 pixels of each sbrow can be changed in
354    // the longest loopfilter of the next sbrow
355    th = (y + bh + 4 * !!my + 7) >> (6 - s->ss_v);
356    ff_thread_await_progress(ref_frame, FFMAX(th, 0), 0);
357    // The arm/aarch64 _hv filters read one more row than what actually is
358    // needed, so switch to emulated edge one pixel sooner vertically
359    // (!!my * 5) than horizontally (!!mx * 4).
360    if (x < !!mx * 3 || y < !!my * 3 ||
361        x + !!mx * 4 > w - bw || y + !!my * 5 > h - bh) {
362        s->vdsp.emulated_edge_mc(td->edge_emu_buffer,
363                                 ref_u - !!my * 3 * src_stride_u - !!mx * 3 * bytesperpixel,
364                                 160, src_stride_u,
365                                 bw + !!mx * 7, bh + !!my * 7,
366                                 x - !!mx * 3, y - !!my * 3, w, h);
367        ref_u = td->edge_emu_buffer + !!my * 3 * 160 + !!mx * 3 * bytesperpixel;
368        mc[!!mx][!!my](dst_u, dst_stride, ref_u, 160, bh, mx, my);
369
370        s->vdsp.emulated_edge_mc(td->edge_emu_buffer,
371                                 ref_v - !!my * 3 * src_stride_v - !!mx * 3 * bytesperpixel,
372                                 160, src_stride_v,
373                                 bw + !!mx * 7, bh + !!my * 7,
374                                 x - !!mx * 3, y - !!my * 3, w, h);
375        ref_v = td->edge_emu_buffer + !!my * 3 * 160 + !!mx * 3 * bytesperpixel;
376        mc[!!mx][!!my](dst_v, dst_stride, ref_v, 160, bh, mx, my);
377    } else {
378        mc[!!mx][!!my](dst_u, dst_stride, ref_u, src_stride_u, bh, mx, my);
379        mc[!!mx][!!my](dst_v, dst_stride, ref_v, src_stride_v, bh, mx, my);
380    }
381}
382
383#define mc_luma_dir(td, mc, dst, dst_ls, src, src_ls, tref, row, col, mv, \
384                    px, py, pw, ph, bw, bh, w, h, i) \
385    mc_luma_unscaled(td, s->dsp.mc, dst, dst_ls, src, src_ls, tref, row, col, \
386                     mv, bw, bh, w, h, bytesperpixel)
387#define mc_chroma_dir(td, mc, dstu, dstv, dst_ls, srcu, srcu_ls, srcv, srcv_ls, tref, \
388                      row, col, mv, px, py, pw, ph, bw, bh, w, h, i) \
389    mc_chroma_unscaled(td, s->dsp.mc, dstu, dstv, dst_ls, srcu, srcu_ls, srcv, srcv_ls, tref, \
390                       row, col, mv, bw, bh, w, h, bytesperpixel)
391#define SCALED 0
392#define FN(x) x##_8bpp
393#define BYTES_PER_PIXEL 1
394#include "vp9_mc_template.c"
395#undef FN
396#undef BYTES_PER_PIXEL
397#define FN(x) x##_16bpp
398#define BYTES_PER_PIXEL 2
399#include "vp9_mc_template.c"
400#undef mc_luma_dir
401#undef mc_chroma_dir
402#undef FN
403#undef BYTES_PER_PIXEL
404#undef SCALED
405
406static av_always_inline void mc_luma_scaled(VP9TileData *td, vp9_scaled_mc_func smc,
407                                            vp9_mc_func (*mc)[2],
408                                            uint8_t *dst, ptrdiff_t dst_stride,
409                                            const uint8_t *ref, ptrdiff_t ref_stride,
410                                            ThreadFrame *ref_frame,
411                                            ptrdiff_t y, ptrdiff_t x, const VP56mv *in_mv,
412                                            int px, int py, int pw, int ph,
413                                            int bw, int bh, int w, int h, int bytesperpixel,
414                                            const uint16_t *scale, const uint8_t *step)
415{
416    VP9Context *s = td->s;
417    if (s->s.frames[CUR_FRAME].tf.f->width == ref_frame->f->width &&
418        s->s.frames[CUR_FRAME].tf.f->height == ref_frame->f->height) {
419        mc_luma_unscaled(td, mc, dst, dst_stride, ref, ref_stride, ref_frame,
420                         y, x, in_mv, bw, bh, w, h, bytesperpixel);
421    } else {
422#define scale_mv(n, dim) (((int64_t)(n) * scale[dim]) >> 14)
423    int mx, my;
424    int refbw_m1, refbh_m1;
425    int th;
426    VP56mv mv;
427
428    mv.x = av_clip(in_mv->x, -(x + pw - px + 4) * 8, (s->cols * 8 - x + px + 3) * 8);
429    mv.y = av_clip(in_mv->y, -(y + ph - py + 4) * 8, (s->rows * 8 - y + py + 3) * 8);
430    // BUG libvpx seems to scale the two components separately. This introduces
431    // rounding errors but we have to reproduce them to be exactly compatible
432    // with the output from libvpx...
433    mx = scale_mv(mv.x * 2, 0) + scale_mv(x * 16, 0);
434    my = scale_mv(mv.y * 2, 1) + scale_mv(y * 16, 1);
435
436    y = my >> 4;
437    x = mx >> 4;
438    ref += y * ref_stride + x * bytesperpixel;
439    mx &= 15;
440    my &= 15;
441    refbw_m1 = ((bw - 1) * step[0] + mx) >> 4;
442    refbh_m1 = ((bh - 1) * step[1] + my) >> 4;
443    // FIXME bilinear filter only needs 0/1 pixels, not 3/4
444    // we use +7 because the last 7 pixels of each sbrow can be changed in
445    // the longest loopfilter of the next sbrow
446    th = (y + refbh_m1 + 4 + 7) >> 6;
447    ff_thread_await_progress(ref_frame, FFMAX(th, 0), 0);
448    // The arm/aarch64 _hv filters read one more row than what actually is
449    // needed, so switch to emulated edge one pixel sooner vertically
450    // (y + 5 >= h - refbh_m1) than horizontally (x + 4 >= w - refbw_m1).
451    if (x < 3 || y < 3 || x + 4 >= w - refbw_m1 || y + 5 >= h - refbh_m1) {
452        s->vdsp.emulated_edge_mc(td->edge_emu_buffer,
453                                 ref - 3 * ref_stride - 3 * bytesperpixel,
454                                 288, ref_stride,
455                                 refbw_m1 + 8, refbh_m1 + 8,
456                                 x - 3, y - 3, w, h);
457        ref = td->edge_emu_buffer + 3 * 288 + 3 * bytesperpixel;
458        ref_stride = 288;
459    }
460    smc(dst, dst_stride, ref, ref_stride, bh, mx, my, step[0], step[1]);
461    }
462}
463
464static av_always_inline void mc_chroma_scaled(VP9TileData *td, vp9_scaled_mc_func smc,
465                                              vp9_mc_func (*mc)[2],
466                                              uint8_t *dst_u, uint8_t *dst_v,
467                                              ptrdiff_t dst_stride,
468                                              const uint8_t *ref_u, ptrdiff_t src_stride_u,
469                                              const uint8_t *ref_v, ptrdiff_t src_stride_v,
470                                              ThreadFrame *ref_frame,
471                                              ptrdiff_t y, ptrdiff_t x, const VP56mv *in_mv,
472                                              int px, int py, int pw, int ph,
473                                              int bw, int bh, int w, int h, int bytesperpixel,
474                                              const uint16_t *scale, const uint8_t *step)
475{
476    VP9Context *s = td->s;
477    if (s->s.frames[CUR_FRAME].tf.f->width == ref_frame->f->width &&
478        s->s.frames[CUR_FRAME].tf.f->height == ref_frame->f->height) {
479        mc_chroma_unscaled(td, mc, dst_u, dst_v, dst_stride, ref_u, src_stride_u,
480                           ref_v, src_stride_v, ref_frame,
481                           y, x, in_mv, bw, bh, w, h, bytesperpixel);
482    } else {
483    int mx, my;
484    int refbw_m1, refbh_m1;
485    int th;
486    VP56mv mv;
487
488    if (s->ss_h) {
489        // BUG https://code.google.com/p/webm/issues/detail?id=820
490        mv.x = av_clip(in_mv->x, -(x + pw - px + 4) * 16, (s->cols * 4 - x + px + 3) * 16);
491        mx = scale_mv(mv.x, 0) + (scale_mv(x * 16, 0) & ~15) + (scale_mv(x * 32, 0) & 15);
492    } else {
493        mv.x = av_clip(in_mv->x, -(x + pw - px + 4) * 8, (s->cols * 8 - x + px + 3) * 8);
494        mx = scale_mv(mv.x * 2, 0) + scale_mv(x * 16, 0);
495    }
496    if (s->ss_v) {
497        // BUG https://code.google.com/p/webm/issues/detail?id=820
498        mv.y = av_clip(in_mv->y, -(y + ph - py + 4) * 16, (s->rows * 4 - y + py + 3) * 16);
499        my = scale_mv(mv.y, 1) + (scale_mv(y * 16, 1) & ~15) + (scale_mv(y * 32, 1) & 15);
500    } else {
501        mv.y = av_clip(in_mv->y, -(y + ph - py + 4) * 8, (s->rows * 8 - y + py + 3) * 8);
502        my = scale_mv(mv.y * 2, 1) + scale_mv(y * 16, 1);
503    }
504#undef scale_mv
505    y = my >> 4;
506    x = mx >> 4;
507    ref_u += y * src_stride_u + x * bytesperpixel;
508    ref_v += y * src_stride_v + x * bytesperpixel;
509    mx &= 15;
510    my &= 15;
511    refbw_m1 = ((bw - 1) * step[0] + mx) >> 4;
512    refbh_m1 = ((bh - 1) * step[1] + my) >> 4;
513    // FIXME bilinear filter only needs 0/1 pixels, not 3/4
514    // we use +7 because the last 7 pixels of each sbrow can be changed in
515    // the longest loopfilter of the next sbrow
516    th = (y + refbh_m1 + 4 + 7) >> (6 - s->ss_v);
517    ff_thread_await_progress(ref_frame, FFMAX(th, 0), 0);
518    // The arm/aarch64 _hv filters read one more row than what actually is
519    // needed, so switch to emulated edge one pixel sooner vertically
520    // (y + 5 >= h - refbh_m1) than horizontally (x + 4 >= w - refbw_m1).
521    if (x < 3 || y < 3 || x + 4 >= w - refbw_m1 || y + 5 >= h - refbh_m1) {
522        s->vdsp.emulated_edge_mc(td->edge_emu_buffer,
523                                 ref_u - 3 * src_stride_u - 3 * bytesperpixel,
524                                 288, src_stride_u,
525                                 refbw_m1 + 8, refbh_m1 + 8,
526                                 x - 3, y - 3, w, h);
527        ref_u = td->edge_emu_buffer + 3 * 288 + 3 * bytesperpixel;
528        smc(dst_u, dst_stride, ref_u, 288, bh, mx, my, step[0], step[1]);
529
530        s->vdsp.emulated_edge_mc(td->edge_emu_buffer,
531                                 ref_v - 3 * src_stride_v - 3 * bytesperpixel,
532                                 288, src_stride_v,
533                                 refbw_m1 + 8, refbh_m1 + 8,
534                                 x - 3, y - 3, w, h);
535        ref_v = td->edge_emu_buffer + 3 * 288 + 3 * bytesperpixel;
536        smc(dst_v, dst_stride, ref_v, 288, bh, mx, my, step[0], step[1]);
537    } else {
538        smc(dst_u, dst_stride, ref_u, src_stride_u, bh, mx, my, step[0], step[1]);
539        smc(dst_v, dst_stride, ref_v, src_stride_v, bh, mx, my, step[0], step[1]);
540    }
541    }
542}
543
544#define mc_luma_dir(td, mc, dst, dst_ls, src, src_ls, tref, row, col, mv, \
545                    px, py, pw, ph, bw, bh, w, h, i) \
546    mc_luma_scaled(td, s->dsp.s##mc, s->dsp.mc, dst, dst_ls, src, src_ls, tref, row, col, \
547                   mv, px, py, pw, ph, bw, bh, w, h, bytesperpixel, \
548                   s->mvscale[b->ref[i]], s->mvstep[b->ref[i]])
549#define mc_chroma_dir(td, mc, dstu, dstv, dst_ls, srcu, srcu_ls, srcv, srcv_ls, tref, \
550                      row, col, mv, px, py, pw, ph, bw, bh, w, h, i) \
551    mc_chroma_scaled(td, s->dsp.s##mc, s->dsp.mc, dstu, dstv, dst_ls, srcu, srcu_ls, srcv, srcv_ls, tref, \
552                     row, col, mv, px, py, pw, ph, bw, bh, w, h, bytesperpixel, \
553                     s->mvscale[b->ref[i]], s->mvstep[b->ref[i]])
554#define SCALED 1
555#define FN(x) x##_scaled_8bpp
556#define BYTES_PER_PIXEL 1
557#include "vp9_mc_template.c"
558#undef FN
559#undef BYTES_PER_PIXEL
560#define FN(x) x##_scaled_16bpp
561#define BYTES_PER_PIXEL 2
562#include "vp9_mc_template.c"
563#undef mc_luma_dir
564#undef mc_chroma_dir
565#undef FN
566#undef BYTES_PER_PIXEL
567#undef SCALED
568
569static av_always_inline void inter_recon(VP9TileData *td, int bytesperpixel)
570{
571    VP9Context *s = td->s;
572    VP9Block *b = td->b;
573    int row = td->row, col = td->col;
574
575    if (s->mvscale[b->ref[0]][0] == REF_INVALID_SCALE ||
576        (b->comp && s->mvscale[b->ref[1]][0] == REF_INVALID_SCALE)) {
577        if (!s->td->error_info) {
578            s->td->error_info = AVERROR_INVALIDDATA;
579            av_log(NULL, AV_LOG_ERROR, "Bitstream not supported, "
580                                       "reference frame has invalid dimensions\n");
581        }
582        return;
583    }
584
585    if (s->mvscale[b->ref[0]][0] || (b->comp && s->mvscale[b->ref[1]][0])) {
586        if (bytesperpixel == 1) {
587            inter_pred_scaled_8bpp(td);
588        } else {
589            inter_pred_scaled_16bpp(td);
590        }
591    } else {
592        if (bytesperpixel == 1) {
593            inter_pred_8bpp(td);
594        } else {
595            inter_pred_16bpp(td);
596        }
597    }
598
599    if (!b->skip) {
600        /* mostly copied intra_recon() */
601
602        int w4 = ff_vp9_bwh_tab[1][b->bs][0] << 1, step1d = 1 << b->tx, n;
603        int h4 = ff_vp9_bwh_tab[1][b->bs][1] << 1, x, y, step = 1 << (b->tx * 2);
604        int end_x = FFMIN(2 * (s->cols - col), w4);
605        int end_y = FFMIN(2 * (s->rows - row), h4);
606        int tx = 4 * s->s.h.lossless + b->tx, uvtx = b->uvtx + 4 * s->s.h.lossless;
607        int uvstep1d = 1 << b->uvtx, p;
608        uint8_t *dst = td->dst[0];
609
610        // y itxfm add
611        for (n = 0, y = 0; y < end_y; y += step1d) {
612            uint8_t *ptr = dst;
613            for (x = 0; x < end_x; x += step1d,
614                 ptr += 4 * step1d * bytesperpixel, n += step) {
615                int eob = b->tx > TX_8X8 ? AV_RN16A(&td->eob[n]) : td->eob[n];
616
617                if (eob)
618                    s->dsp.itxfm_add[tx][DCT_DCT](ptr, td->y_stride,
619                                                  td->block + 16 * n * bytesperpixel, eob);
620            }
621            dst += 4 * td->y_stride * step1d;
622        }
623
624        // uv itxfm add
625        end_x >>= s->ss_h;
626        end_y >>= s->ss_v;
627        step = 1 << (b->uvtx * 2);
628        for (p = 0; p < 2; p++) {
629            dst = td->dst[p + 1];
630            for (n = 0, y = 0; y < end_y; y += uvstep1d) {
631                uint8_t *ptr = dst;
632                for (x = 0; x < end_x; x += uvstep1d,
633                     ptr += 4 * uvstep1d * bytesperpixel, n += step) {
634                    int eob = b->uvtx > TX_8X8 ? AV_RN16A(&td->uveob[p][n]) : td->uveob[p][n];
635
636                    if (eob)
637                        s->dsp.itxfm_add[uvtx][DCT_DCT](ptr, td->uv_stride,
638                                                        td->uvblock[p] + 16 * n * bytesperpixel, eob);
639                }
640                dst += 4 * uvstep1d * td->uv_stride;
641            }
642        }
643    }
644}
645
646void ff_vp9_inter_recon_8bpp(VP9TileData *td)
647{
648    inter_recon(td, 1);
649}
650
651void ff_vp9_inter_recon_16bpp(VP9TileData *td)
652{
653    inter_recon(td, 2);
654}
655