1/* 2 * DSP functions for Indeo Video Interactive codecs (Indeo4 and Indeo5) 3 * 4 * Copyright (c) 2009-2011 Maxim Poliakovski 5 * 6 * This file is part of FFmpeg. 7 * 8 * FFmpeg is free software; you can redistribute it and/or 9 * modify it under the terms of the GNU Lesser General Public 10 * License as published by the Free Software Foundation; either 11 * version 2.1 of the License, or (at your option) any later version. 12 * 13 * FFmpeg is distributed in the hope that it will be useful, 14 * but WITHOUT ANY WARRANTY; without even the implied warranty of 15 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU 16 * Lesser General Public License for more details. 17 * 18 * You should have received a copy of the GNU Lesser General Public 19 * License along with FFmpeg; if not, write to the Free Software 20 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA 21 */ 22 23/** 24 * @file 25 * DSP functions (inverse transforms, motion compensation, wavelet recompositions) 26 * for Indeo Video Interactive codecs. 27 */ 28 29#include <string.h> 30#include "libavutil/common.h" 31#include "ivi.h" 32#include "ivi_dsp.h" 33 34void ff_ivi_recompose53(const IVIPlaneDesc *plane, uint8_t *dst, 35 const ptrdiff_t dst_pitch) 36{ 37 int x, y, indx; 38 int32_t p0, p1, p2, p3, tmp0, tmp1, tmp2; 39 int32_t b0_1, b0_2, b1_1, b1_2, b1_3, b2_1, b2_2, b2_3, b2_4, b2_5, b2_6; 40 int32_t b3_1, b3_2, b3_3, b3_4, b3_5, b3_6, b3_7, b3_8, b3_9; 41 ptrdiff_t pitch, back_pitch; 42 const short *b0_ptr, *b1_ptr, *b2_ptr, *b3_ptr; 43 const int num_bands = 4; 44 45 /* all bands should have the same pitch */ 46 pitch = plane->bands[0].pitch; 47 48 /* pixels at the position "y-1" will be set to pixels at the "y" for the 1st iteration */ 49 back_pitch = 0; 50 51 /* get pointers to the wavelet bands */ 52 b0_ptr = plane->bands[0].buf; 53 b1_ptr = plane->bands[1].buf; 54 b2_ptr = plane->bands[2].buf; 55 b3_ptr = plane->bands[3].buf; 56 57 for (y = 0; y < plane->height; y += 2) { 58 59 if (y+2 >= plane->height) 60 pitch= 0; 61 /* load storage variables with values */ 62 if (num_bands > 0) { 63 b0_1 = b0_ptr[0]; 64 b0_2 = b0_ptr[pitch]; 65 } 66 67 if (num_bands > 1) { 68 b1_1 = b1_ptr[back_pitch]; 69 b1_2 = b1_ptr[0]; 70 b1_3 = b1_1 - b1_2*6 + b1_ptr[pitch]; 71 } 72 73 if (num_bands > 2) { 74 b2_2 = b2_ptr[0]; // b2[x, y ] 75 b2_3 = b2_2; // b2[x+1,y ] = b2[x,y] 76 b2_5 = b2_ptr[pitch]; // b2[x ,y+1] 77 b2_6 = b2_5; // b2[x+1,y+1] = b2[x,y+1] 78 } 79 80 if (num_bands > 3) { 81 b3_2 = b3_ptr[back_pitch]; // b3[x ,y-1] 82 b3_3 = b3_2; // b3[x+1,y-1] = b3[x ,y-1] 83 b3_5 = b3_ptr[0]; // b3[x ,y ] 84 b3_6 = b3_5; // b3[x+1,y ] = b3[x ,y ] 85 b3_8 = b3_2 - b3_5*6 + b3_ptr[pitch]; 86 b3_9 = b3_8; 87 } 88 89 for (x = 0, indx = 0; x < plane->width; x+=2, indx++) { 90 if (x+2 >= plane->width) { 91 b0_ptr --; 92 b1_ptr --; 93 b2_ptr --; 94 b3_ptr --; 95 } 96 97 /* some values calculated in the previous iterations can */ 98 /* be reused in the next ones, so do appropriate copying */ 99 b2_1 = b2_2; // b2[x-1,y ] = b2[x, y ] 100 b2_2 = b2_3; // b2[x ,y ] = b2[x+1,y ] 101 b2_4 = b2_5; // b2[x-1,y+1] = b2[x ,y+1] 102 b2_5 = b2_6; // b2[x ,y+1] = b2[x+1,y+1] 103 b3_1 = b3_2; // b3[x-1,y-1] = b3[x ,y-1] 104 b3_2 = b3_3; // b3[x ,y-1] = b3[x+1,y-1] 105 b3_4 = b3_5; // b3[x-1,y ] = b3[x ,y ] 106 b3_5 = b3_6; // b3[x ,y ] = b3[x+1,y ] 107 b3_7 = b3_8; // vert_HPF(x-1) 108 b3_8 = b3_9; // vert_HPF(x ) 109 110 p0 = p1 = p2 = p3 = 0; 111 112 /* process the LL-band by applying LPF both vertically and horizontally */ 113 if (num_bands > 0) { 114 tmp0 = b0_1; 115 tmp2 = b0_2; 116 b0_1 = b0_ptr[indx+1]; 117 b0_2 = b0_ptr[pitch+indx+1]; 118 tmp1 = tmp0 + b0_1; 119 120 p0 = tmp0 * 16; 121 p1 = tmp1 * 8; 122 p2 = (tmp0 + tmp2) * 8; 123 p3 = (tmp1 + tmp2 + b0_2) * 4; 124 } 125 126 /* process the HL-band by applying HPF vertically and LPF horizontally */ 127 if (num_bands > 1) { 128 tmp0 = b1_2; 129 tmp1 = b1_1; 130 b1_2 = b1_ptr[indx+1]; 131 b1_1 = b1_ptr[back_pitch+indx+1]; 132 133 tmp2 = tmp1 - tmp0*6 + b1_3; 134 b1_3 = b1_1 - b1_2*6 + b1_ptr[pitch+indx+1]; 135 136 p0 += (tmp0 + tmp1) * 8; 137 p1 += (tmp0 + tmp1 + b1_1 + b1_2) * 4; 138 p2 += tmp2 * 4; 139 p3 += (tmp2 + b1_3) * 2; 140 } 141 142 /* process the LH-band by applying LPF vertically and HPF horizontally */ 143 if (num_bands > 2) { 144 b2_3 = b2_ptr[indx+1]; 145 b2_6 = b2_ptr[pitch+indx+1]; 146 147 tmp0 = b2_1 + b2_2; 148 tmp1 = b2_1 - b2_2*6 + b2_3; 149 150 p0 += tmp0 * 8; 151 p1 += tmp1 * 4; 152 p2 += (tmp0 + b2_4 + b2_5) * 4; 153 p3 += (tmp1 + b2_4 - b2_5*6 + b2_6) * 2; 154 } 155 156 /* process the HH-band by applying HPF both vertically and horizontally */ 157 if (num_bands > 3) { 158 b3_6 = b3_ptr[indx+1]; // b3[x+1,y ] 159 b3_3 = b3_ptr[back_pitch+indx+1]; // b3[x+1,y-1] 160 161 tmp0 = b3_1 + b3_4; 162 tmp1 = b3_2 + b3_5; 163 tmp2 = b3_3 + b3_6; 164 165 b3_9 = b3_3 - b3_6*6 + b3_ptr[pitch+indx+1]; 166 167 p0 += (tmp0 + tmp1) * 4; 168 p1 += (tmp0 - tmp1*6 + tmp2) * 2; 169 p2 += (b3_7 + b3_8) * 2; 170 p3 += b3_7 - b3_8*6 + b3_9; 171 } 172 173 /* output four pixels */ 174 dst[x] = av_clip_uint8((p0 >> 6) + 128); 175 dst[x+1] = av_clip_uint8((p1 >> 6) + 128); 176 dst[dst_pitch+x] = av_clip_uint8((p2 >> 6) + 128); 177 dst[dst_pitch+x+1] = av_clip_uint8((p3 >> 6) + 128); 178 }// for x 179 180 dst += dst_pitch << 1; 181 182 back_pitch = -pitch; 183 184 b0_ptr += pitch + 1; 185 b1_ptr += pitch + 1; 186 b2_ptr += pitch + 1; 187 b3_ptr += pitch + 1; 188 } 189} 190 191void ff_ivi_recompose_haar(const IVIPlaneDesc *plane, uint8_t *dst, 192 const ptrdiff_t dst_pitch) 193{ 194 int x, y, indx, b0, b1, b2, b3, p0, p1, p2, p3; 195 const short *b0_ptr, *b1_ptr, *b2_ptr, *b3_ptr; 196 ptrdiff_t pitch; 197 198 /* all bands should have the same pitch */ 199 pitch = plane->bands[0].pitch; 200 201 /* get pointers to the wavelet bands */ 202 b0_ptr = plane->bands[0].buf; 203 b1_ptr = plane->bands[1].buf; 204 b2_ptr = plane->bands[2].buf; 205 b3_ptr = plane->bands[3].buf; 206 207 for (y = 0; y < plane->height; y += 2) { 208 for (x = 0, indx = 0; x < plane->width; x += 2, indx++) { 209 /* load coefficients */ 210 b0 = b0_ptr[indx]; //should be: b0 = (num_bands > 0) ? b0_ptr[indx] : 0; 211 b1 = b1_ptr[indx]; //should be: b1 = (num_bands > 1) ? b1_ptr[indx] : 0; 212 b2 = b2_ptr[indx]; //should be: b2 = (num_bands > 2) ? b2_ptr[indx] : 0; 213 b3 = b3_ptr[indx]; //should be: b3 = (num_bands > 3) ? b3_ptr[indx] : 0; 214 215 /* haar wavelet recomposition */ 216 p0 = (b0 + b1 + b2 + b3 + 2) >> 2; 217 p1 = (b0 + b1 - b2 - b3 + 2) >> 2; 218 p2 = (b0 - b1 + b2 - b3 + 2) >> 2; 219 p3 = (b0 - b1 - b2 + b3 + 2) >> 2; 220 221 /* bias, convert and output four pixels */ 222 dst[x] = av_clip_uint8(p0 + 128); 223 dst[x + 1] = av_clip_uint8(p1 + 128); 224 dst[dst_pitch + x] = av_clip_uint8(p2 + 128); 225 dst[dst_pitch + x + 1] = av_clip_uint8(p3 + 128); 226 }// for x 227 228 dst += dst_pitch << 1; 229 230 b0_ptr += pitch; 231 b1_ptr += pitch; 232 b2_ptr += pitch; 233 b3_ptr += pitch; 234 }// for y 235} 236 237/** butterfly operation for the inverse Haar transform */ 238#define IVI_HAAR_BFLY(s1, s2, o1, o2, t) \ 239 t = ((s1) - (s2)) >> 1;\ 240 o1 = ((s1) + (s2)) >> 1;\ 241 o2 = (t);\ 242 243/** inverse 8-point Haar transform */ 244#define INV_HAAR8(s1, s5, s3, s7, s2, s4, s6, s8,\ 245 d1, d2, d3, d4, d5, d6, d7, d8,\ 246 t0, t1, t2, t3, t4, t5, t6, t7, t8) {\ 247 t1 = (s1) * 2; t5 = (s5) * 2;\ 248 IVI_HAAR_BFLY(t1, t5, t1, t5, t0); IVI_HAAR_BFLY(t1, s3, t1, t3, t0);\ 249 IVI_HAAR_BFLY(t5, s7, t5, t7, t0); IVI_HAAR_BFLY(t1, s2, t1, t2, t0);\ 250 IVI_HAAR_BFLY(t3, s4, t3, t4, t0); IVI_HAAR_BFLY(t5, s6, t5, t6, t0);\ 251 IVI_HAAR_BFLY(t7, s8, t7, t8, t0);\ 252 d1 = COMPENSATE(t1);\ 253 d2 = COMPENSATE(t2);\ 254 d3 = COMPENSATE(t3);\ 255 d4 = COMPENSATE(t4);\ 256 d5 = COMPENSATE(t5);\ 257 d6 = COMPENSATE(t6);\ 258 d7 = COMPENSATE(t7);\ 259 d8 = COMPENSATE(t8); } 260 261/** inverse 4-point Haar transform */ 262#define INV_HAAR4(s1, s3, s5, s7, d1, d2, d3, d4, t0, t1, t2, t3, t4) {\ 263 IVI_HAAR_BFLY(s1, s3, t0, t1, t4);\ 264 IVI_HAAR_BFLY(t0, s5, t2, t3, t4);\ 265 d1 = COMPENSATE(t2);\ 266 d2 = COMPENSATE(t3);\ 267 IVI_HAAR_BFLY(t1, s7, t2, t3, t4);\ 268 d3 = COMPENSATE(t2);\ 269 d4 = COMPENSATE(t3); } 270 271void ff_ivi_inverse_haar_8x8(const int32_t *in, int16_t *out, ptrdiff_t pitch, 272 const uint8_t *flags) 273{ 274 int i, shift, sp1, sp2, sp3, sp4; 275 const int32_t *src; 276 int32_t *dst; 277 int tmp[64]; 278 int t0, t1, t2, t3, t4, t5, t6, t7, t8; 279 280 /* apply the InvHaar8 to all columns */ 281#define COMPENSATE(x) (x) 282 src = in; 283 dst = tmp; 284 for (i = 0; i < 8; i++) { 285 if (flags[i]) { 286 /* pre-scaling */ 287 shift = !(i & 4); 288 sp1 = src[ 0] * (1 << shift); 289 sp2 = src[ 8] * (1 << shift); 290 sp3 = src[16] * (1 << shift); 291 sp4 = src[24] * (1 << shift); 292 INV_HAAR8( sp1, sp2, sp3, sp4, 293 src[32], src[40], src[48], src[56], 294 dst[ 0], dst[ 8], dst[16], dst[24], 295 dst[32], dst[40], dst[48], dst[56], 296 t0, t1, t2, t3, t4, t5, t6, t7, t8); 297 } else 298 dst[ 0] = dst[ 8] = dst[16] = dst[24] = 299 dst[32] = dst[40] = dst[48] = dst[56] = 0; 300 301 src++; 302 dst++; 303 } 304#undef COMPENSATE 305 306 /* apply the InvHaar8 to all rows */ 307#define COMPENSATE(x) (x) 308 src = tmp; 309 for (i = 0; i < 8; i++) { 310 if ( !src[0] && !src[1] && !src[2] && !src[3] 311 && !src[4] && !src[5] && !src[6] && !src[7]) { 312 memset(out, 0, 8 * sizeof(out[0])); 313 } else { 314 INV_HAAR8(src[0], src[1], src[2], src[3], 315 src[4], src[5], src[6], src[7], 316 out[0], out[1], out[2], out[3], 317 out[4], out[5], out[6], out[7], 318 t0, t1, t2, t3, t4, t5, t6, t7, t8); 319 } 320 src += 8; 321 out += pitch; 322 } 323#undef COMPENSATE 324} 325 326void ff_ivi_row_haar8(const int32_t *in, int16_t *out, ptrdiff_t pitch, 327 const uint8_t *flags) 328{ 329 int i; 330 int t0, t1, t2, t3, t4, t5, t6, t7, t8; 331 332 /* apply the InvHaar8 to all rows */ 333#define COMPENSATE(x) (x) 334 for (i = 0; i < 8; i++) { 335 if ( !in[0] && !in[1] && !in[2] && !in[3] 336 && !in[4] && !in[5] && !in[6] && !in[7]) { 337 memset(out, 0, 8 * sizeof(out[0])); 338 } else { 339 INV_HAAR8(in[0], in[1], in[2], in[3], 340 in[4], in[5], in[6], in[7], 341 out[0], out[1], out[2], out[3], 342 out[4], out[5], out[6], out[7], 343 t0, t1, t2, t3, t4, t5, t6, t7, t8); 344 } 345 in += 8; 346 out += pitch; 347 } 348#undef COMPENSATE 349} 350 351void ff_ivi_col_haar8(const int32_t *in, int16_t *out, ptrdiff_t pitch, 352 const uint8_t *flags) 353{ 354 int i; 355 int t0, t1, t2, t3, t4, t5, t6, t7, t8; 356 357 /* apply the InvHaar8 to all columns */ 358#define COMPENSATE(x) (x) 359 for (i = 0; i < 8; i++) { 360 if (flags[i]) { 361 INV_HAAR8(in[ 0], in[ 8], in[16], in[24], 362 in[32], in[40], in[48], in[56], 363 out[0 * pitch], out[1 * pitch], 364 out[2 * pitch], out[3 * pitch], 365 out[4 * pitch], out[5 * pitch], 366 out[6 * pitch], out[7 * pitch], 367 t0, t1, t2, t3, t4, t5, t6, t7, t8); 368 } else 369 out[0 * pitch] = out[1 * pitch] = 370 out[2 * pitch] = out[3 * pitch] = 371 out[4 * pitch] = out[5 * pitch] = 372 out[6 * pitch] = out[7 * pitch] = 0; 373 374 in++; 375 out++; 376 } 377#undef COMPENSATE 378} 379 380void ff_ivi_inverse_haar_4x4(const int32_t *in, int16_t *out, ptrdiff_t pitch, 381 const uint8_t *flags) 382{ 383 int i, shift, sp1, sp2; 384 const int32_t *src; 385 int32_t *dst; 386 int tmp[16]; 387 int t0, t1, t2, t3, t4; 388 389 /* apply the InvHaar4 to all columns */ 390#define COMPENSATE(x) (x) 391 src = in; 392 dst = tmp; 393 for (i = 0; i < 4; i++) { 394 if (flags[i]) { 395 /* pre-scaling */ 396 shift = !(i & 2); 397 sp1 = src[0] * (1 << shift); 398 sp2 = src[4] * (1 << shift); 399 INV_HAAR4( sp1, sp2, src[8], src[12], 400 dst[0], dst[4], dst[8], dst[12], 401 t0, t1, t2, t3, t4); 402 } else 403 dst[0] = dst[4] = dst[8] = dst[12] = 0; 404 405 src++; 406 dst++; 407 } 408#undef COMPENSATE 409 410 /* apply the InvHaar8 to all rows */ 411#define COMPENSATE(x) (x) 412 src = tmp; 413 for (i = 0; i < 4; i++) { 414 if (!src[0] && !src[1] && !src[2] && !src[3]) { 415 memset(out, 0, 4 * sizeof(out[0])); 416 } else { 417 INV_HAAR4(src[0], src[1], src[2], src[3], 418 out[0], out[1], out[2], out[3], 419 t0, t1, t2, t3, t4); 420 } 421 src += 4; 422 out += pitch; 423 } 424#undef COMPENSATE 425} 426 427void ff_ivi_row_haar4(const int32_t *in, int16_t *out, ptrdiff_t pitch, 428 const uint8_t *flags) 429{ 430 int i; 431 int t0, t1, t2, t3, t4; 432 433 /* apply the InvHaar4 to all rows */ 434#define COMPENSATE(x) (x) 435 for (i = 0; i < 4; i++) { 436 if (!in[0] && !in[1] && !in[2] && !in[3]) { 437 memset(out, 0, 4 * sizeof(out[0])); 438 } else { 439 INV_HAAR4(in[0], in[1], in[2], in[3], 440 out[0], out[1], out[2], out[3], 441 t0, t1, t2, t3, t4); 442 } 443 in += 4; 444 out += pitch; 445 } 446#undef COMPENSATE 447} 448 449void ff_ivi_col_haar4(const int32_t *in, int16_t *out, ptrdiff_t pitch, 450 const uint8_t *flags) 451{ 452 int i; 453 int t0, t1, t2, t3, t4; 454 455 /* apply the InvHaar8 to all columns */ 456#define COMPENSATE(x) (x) 457 for (i = 0; i < 4; i++) { 458 if (flags[i]) { 459 INV_HAAR4(in[0], in[4], in[8], in[12], 460 out[0 * pitch], out[1 * pitch], 461 out[2 * pitch], out[3 * pitch], 462 t0, t1, t2, t3, t4); 463 } else 464 out[0 * pitch] = out[1 * pitch] = 465 out[2 * pitch] = out[3 * pitch] = 0; 466 467 in++; 468 out++; 469 } 470#undef COMPENSATE 471} 472 473void ff_ivi_dc_haar_2d(const int32_t *in, int16_t *out, ptrdiff_t pitch, 474 int blk_size) 475{ 476 int x, y; 477 int16_t dc_coeff; 478 479 dc_coeff = (*in + 0) >> 3; 480 481 for (y = 0; y < blk_size; out += pitch, y++) { 482 for (x = 0; x < blk_size; x++) 483 out[x] = dc_coeff; 484 } 485} 486 487/** butterfly operation for the inverse slant transform */ 488#define IVI_SLANT_BFLY(s1, s2, o1, o2, t) \ 489 t = (s1) - (s2);\ 490 o1 = (s1) + (s2);\ 491 o2 = (t);\ 492 493/** This is a reflection a,b = 1/2, 5/4 for the inverse slant transform */ 494#define IVI_IREFLECT(s1, s2, o1, o2, t) \ 495 t = (((s1) + (s2)*2 + 2) >> 2) + (s1);\ 496 o2 = (((s1)*2 - (s2) + 2) >> 2) - (s2);\ 497 o1 = (t);\ 498 499/** This is a reflection a,b = 1/2, 7/8 for the inverse slant transform */ 500#define IVI_SLANT_PART4(s1, s2, o1, o2, t) \ 501 t = (s2) + (((s1)*4 - (s2) + 4) >> 3);\ 502 o2 = (s1) + ((-(s1) - (s2)*4 + 4) >> 3);\ 503 o1 = (t);\ 504 505/** inverse slant8 transform */ 506#define IVI_INV_SLANT8(s1, s4, s8, s5, s2, s6, s3, s7,\ 507 d1, d2, d3, d4, d5, d6, d7, d8,\ 508 t0, t1, t2, t3, t4, t5, t6, t7, t8) {\ 509 IVI_SLANT_PART4(s4, s5, t4, t5, t0);\ 510\ 511 IVI_SLANT_BFLY(s1, t5, t1, t5, t0); IVI_SLANT_BFLY(s2, s6, t2, t6, t0);\ 512 IVI_SLANT_BFLY(s7, s3, t7, t3, t0); IVI_SLANT_BFLY(t4, s8, t4, t8, t0);\ 513\ 514 IVI_SLANT_BFLY(t1, t2, t1, t2, t0); IVI_IREFLECT (t4, t3, t4, t3, t0);\ 515 IVI_SLANT_BFLY(t5, t6, t5, t6, t0); IVI_IREFLECT (t8, t7, t8, t7, t0);\ 516 IVI_SLANT_BFLY(t1, t4, t1, t4, t0); IVI_SLANT_BFLY(t2, t3, t2, t3, t0);\ 517 IVI_SLANT_BFLY(t5, t8, t5, t8, t0); IVI_SLANT_BFLY(t6, t7, t6, t7, t0);\ 518 d1 = COMPENSATE(t1);\ 519 d2 = COMPENSATE(t2);\ 520 d3 = COMPENSATE(t3);\ 521 d4 = COMPENSATE(t4);\ 522 d5 = COMPENSATE(t5);\ 523 d6 = COMPENSATE(t6);\ 524 d7 = COMPENSATE(t7);\ 525 d8 = COMPENSATE(t8);} 526 527/** inverse slant4 transform */ 528#define IVI_INV_SLANT4(s1, s4, s2, s3, d1, d2, d3, d4, t0, t1, t2, t3, t4) {\ 529 IVI_SLANT_BFLY(s1, s2, t1, t2, t0); IVI_IREFLECT (s4, s3, t4, t3, t0);\ 530\ 531 IVI_SLANT_BFLY(t1, t4, t1, t4, t0); IVI_SLANT_BFLY(t2, t3, t2, t3, t0);\ 532 d1 = COMPENSATE(t1);\ 533 d2 = COMPENSATE(t2);\ 534 d3 = COMPENSATE(t3);\ 535 d4 = COMPENSATE(t4);} 536 537void ff_ivi_inverse_slant_8x8(const int32_t *in, int16_t *out, ptrdiff_t pitch, const uint8_t *flags) 538{ 539 int i; 540 const int32_t *src; 541 int32_t *dst; 542 int tmp[64]; 543 int t0, t1, t2, t3, t4, t5, t6, t7, t8; 544 545#define COMPENSATE(x) (x) 546 src = in; 547 dst = tmp; 548 for (i = 0; i < 8; i++) { 549 if (flags[i]) { 550 IVI_INV_SLANT8(src[0], src[8], src[16], src[24], src[32], src[40], src[48], src[56], 551 dst[0], dst[8], dst[16], dst[24], dst[32], dst[40], dst[48], dst[56], 552 t0, t1, t2, t3, t4, t5, t6, t7, t8); 553 } else 554 dst[0] = dst[8] = dst[16] = dst[24] = dst[32] = dst[40] = dst[48] = dst[56] = 0; 555 556 src++; 557 dst++; 558 } 559#undef COMPENSATE 560 561#define COMPENSATE(x) (((x) + 1)>>1) 562 src = tmp; 563 for (i = 0; i < 8; i++) { 564 if (!src[0] && !src[1] && !src[2] && !src[3] && !src[4] && !src[5] && !src[6] && !src[7]) { 565 memset(out, 0, 8*sizeof(out[0])); 566 } else { 567 IVI_INV_SLANT8(src[0], src[1], src[2], src[3], src[4], src[5], src[6], src[7], 568 out[0], out[1], out[2], out[3], out[4], out[5], out[6], out[7], 569 t0, t1, t2, t3, t4, t5, t6, t7, t8); 570 } 571 src += 8; 572 out += pitch; 573 } 574#undef COMPENSATE 575} 576 577void ff_ivi_inverse_slant_4x4(const int32_t *in, int16_t *out, ptrdiff_t pitch, const uint8_t *flags) 578{ 579 int i; 580 const int32_t *src; 581 int32_t *dst; 582 int tmp[16]; 583 int t0, t1, t2, t3, t4; 584 585#define COMPENSATE(x) (x) 586 src = in; 587 dst = tmp; 588 for (i = 0; i < 4; i++) { 589 if (flags[i]) { 590 IVI_INV_SLANT4(src[0], src[4], src[8], src[12], 591 dst[0], dst[4], dst[8], dst[12], 592 t0, t1, t2, t3, t4); 593 } else 594 dst[0] = dst[4] = dst[8] = dst[12] = 0; 595 596 src++; 597 dst++; 598 } 599#undef COMPENSATE 600 601#define COMPENSATE(x) (((x) + 1)>>1) 602 src = tmp; 603 for (i = 0; i < 4; i++) { 604 if (!src[0] && !src[1] && !src[2] && !src[3]) { 605 out[0] = out[1] = out[2] = out[3] = 0; 606 } else { 607 IVI_INV_SLANT4(src[0], src[1], src[2], src[3], 608 out[0], out[1], out[2], out[3], 609 t0, t1, t2, t3, t4); 610 } 611 src += 4; 612 out += pitch; 613 } 614#undef COMPENSATE 615} 616 617void ff_ivi_dc_slant_2d(const int32_t *in, int16_t *out, ptrdiff_t pitch, int blk_size) 618{ 619 int x, y; 620 int16_t dc_coeff; 621 622 dc_coeff = (*in + 1) >> 1; 623 624 for (y = 0; y < blk_size; out += pitch, y++) { 625 for (x = 0; x < blk_size; x++) 626 out[x] = dc_coeff; 627 } 628} 629 630void ff_ivi_row_slant8(const int32_t *in, int16_t *out, ptrdiff_t pitch, const uint8_t *flags) 631{ 632 int i; 633 int t0, t1, t2, t3, t4, t5, t6, t7, t8; 634 635#define COMPENSATE(x) (((x) + 1)>>1) 636 for (i = 0; i < 8; i++) { 637 if (!in[0] && !in[1] && !in[2] && !in[3] && !in[4] && !in[5] && !in[6] && !in[7]) { 638 memset(out, 0, 8*sizeof(out[0])); 639 } else { 640 IVI_INV_SLANT8( in[0], in[1], in[2], in[3], in[4], in[5], in[6], in[7], 641 out[0], out[1], out[2], out[3], out[4], out[5], out[6], out[7], 642 t0, t1, t2, t3, t4, t5, t6, t7, t8); 643 } 644 in += 8; 645 out += pitch; 646 } 647#undef COMPENSATE 648} 649 650void ff_ivi_dc_row_slant(const int32_t *in, int16_t *out, ptrdiff_t pitch, int blk_size) 651{ 652 int x, y; 653 int16_t dc_coeff; 654 655 dc_coeff = (*in + 1) >> 1; 656 657 for (x = 0; x < blk_size; x++) 658 out[x] = dc_coeff; 659 660 out += pitch; 661 662 for (y = 1; y < blk_size; out += pitch, y++) { 663 for (x = 0; x < blk_size; x++) 664 out[x] = 0; 665 } 666} 667 668void ff_ivi_col_slant8(const int32_t *in, int16_t *out, ptrdiff_t pitch, const uint8_t *flags) 669{ 670 int i, row2, row4, row8; 671 int t0, t1, t2, t3, t4, t5, t6, t7, t8; 672 673 row2 = pitch << 1; 674 row4 = pitch << 2; 675 row8 = pitch << 3; 676 677#define COMPENSATE(x) (((x) + 1)>>1) 678 for (i = 0; i < 8; i++) { 679 if (flags[i]) { 680 IVI_INV_SLANT8(in[0], in[8], in[16], in[24], in[32], in[40], in[48], in[56], 681 out[0], out[pitch], out[row2], out[row2 + pitch], out[row4], 682 out[row4 + pitch], out[row4 + row2], out[row8 - pitch], 683 t0, t1, t2, t3, t4, t5, t6, t7, t8); 684 } else { 685 out[0] = out[pitch] = out[row2] = out[row2 + pitch] = out[row4] = 686 out[row4 + pitch] = out[row4 + row2] = out[row8 - pitch] = 0; 687 } 688 689 in++; 690 out++; 691 } 692#undef COMPENSATE 693} 694 695void ff_ivi_dc_col_slant(const int32_t *in, int16_t *out, ptrdiff_t pitch, int blk_size) 696{ 697 int x, y; 698 int16_t dc_coeff; 699 700 dc_coeff = (*in + 1) >> 1; 701 702 for (y = 0; y < blk_size; out += pitch, y++) { 703 out[0] = dc_coeff; 704 for (x = 1; x < blk_size; x++) 705 out[x] = 0; 706 } 707} 708 709void ff_ivi_row_slant4(const int32_t *in, int16_t *out, ptrdiff_t pitch, const uint8_t *flags) 710{ 711 int i; 712 int t0, t1, t2, t3, t4; 713 714#define COMPENSATE(x) (((x) + 1)>>1) 715 for (i = 0; i < 4; i++) { 716 if (!in[0] && !in[1] && !in[2] && !in[3]) { 717 memset(out, 0, 4*sizeof(out[0])); 718 } else { 719 IVI_INV_SLANT4( in[0], in[1], in[2], in[3], 720 out[0], out[1], out[2], out[3], 721 t0, t1, t2, t3, t4); 722 } 723 in += 4; 724 out += pitch; 725 } 726#undef COMPENSATE 727} 728 729void ff_ivi_col_slant4(const int32_t *in, int16_t *out, ptrdiff_t pitch, const uint8_t *flags) 730{ 731 int i, row2; 732 int t0, t1, t2, t3, t4; 733 734 row2 = pitch << 1; 735 736#define COMPENSATE(x) (((x) + 1)>>1) 737 for (i = 0; i < 4; i++) { 738 if (flags[i]) { 739 IVI_INV_SLANT4(in[0], in[4], in[8], in[12], 740 out[0], out[pitch], out[row2], out[row2 + pitch], 741 t0, t1, t2, t3, t4); 742 } else { 743 out[0] = out[pitch] = out[row2] = out[row2 + pitch] = 0; 744 } 745 746 in++; 747 out++; 748 } 749#undef COMPENSATE 750} 751 752void ff_ivi_put_pixels_8x8(const int32_t *in, int16_t *out, ptrdiff_t pitch, 753 const uint8_t *flags) 754{ 755 int x, y; 756 757 for (y = 0; y < 8; out += pitch, in += 8, y++) 758 for (x = 0; x < 8; x++) 759 out[x] = in[x]; 760} 761 762void ff_ivi_put_dc_pixel_8x8(const int32_t *in, int16_t *out, ptrdiff_t pitch, 763 int blk_size) 764{ 765 int y; 766 767 out[0] = in[0]; 768 memset(out + 1, 0, 7*sizeof(out[0])); 769 out += pitch; 770 771 for (y = 1; y < 8; out += pitch, y++) 772 memset(out, 0, 8*sizeof(out[0])); 773} 774 775#define IVI_MC_TEMPLATE(size, suffix, OP) \ 776static void ivi_mc_ ## size ##x## size ## suffix(int16_t *buf, \ 777 ptrdiff_t dpitch, \ 778 const int16_t *ref_buf, \ 779 ptrdiff_t pitch, int mc_type) \ 780{ \ 781 int i, j; \ 782 const int16_t *wptr; \ 783\ 784 switch (mc_type) { \ 785 case 0: /* fullpel (no interpolation) */ \ 786 for (i = 0; i < size; i++, buf += dpitch, ref_buf += pitch) { \ 787 for (j = 0; j < size; j++) {\ 788 OP(buf[j], ref_buf[j]); \ 789 } \ 790 } \ 791 break; \ 792 case 1: /* horizontal halfpel interpolation */ \ 793 for (i = 0; i < size; i++, buf += dpitch, ref_buf += pitch) \ 794 for (j = 0; j < size; j++) \ 795 OP(buf[j], (ref_buf[j] + ref_buf[j+1]) >> 1); \ 796 break; \ 797 case 2: /* vertical halfpel interpolation */ \ 798 wptr = ref_buf + pitch; \ 799 for (i = 0; i < size; i++, buf += dpitch, wptr += pitch, ref_buf += pitch) \ 800 for (j = 0; j < size; j++) \ 801 OP(buf[j], (ref_buf[j] + wptr[j]) >> 1); \ 802 break; \ 803 case 3: /* vertical and horizontal halfpel interpolation */ \ 804 wptr = ref_buf + pitch; \ 805 for (i = 0; i < size; i++, buf += dpitch, wptr += pitch, ref_buf += pitch) \ 806 for (j = 0; j < size; j++) \ 807 OP(buf[j], (ref_buf[j] + ref_buf[j+1] + wptr[j] + wptr[j+1]) >> 2); \ 808 break; \ 809 } \ 810} \ 811\ 812void ff_ivi_mc_ ## size ##x## size ## suffix(int16_t *buf, const int16_t *ref_buf, \ 813 ptrdiff_t pitch, int mc_type) \ 814{ \ 815 ivi_mc_ ## size ##x## size ## suffix(buf, pitch, ref_buf, pitch, mc_type); \ 816} \ 817 818#define IVI_MC_AVG_TEMPLATE(size, suffix, OP) \ 819void ff_ivi_mc_avg_ ## size ##x## size ## suffix(int16_t *buf, \ 820 const int16_t *ref_buf, \ 821 const int16_t *ref_buf2, \ 822 ptrdiff_t pitch, \ 823 int mc_type, int mc_type2) \ 824{ \ 825 int16_t tmp[size * size]; \ 826 int i, j; \ 827\ 828 ivi_mc_ ## size ##x## size ## _no_delta(tmp, size, ref_buf, pitch, mc_type); \ 829 ivi_mc_ ## size ##x## size ## _delta(tmp, size, ref_buf2, pitch, mc_type2); \ 830 for (i = 0; i < size; i++, buf += pitch) { \ 831 for (j = 0; j < size; j++) {\ 832 OP(buf[j], tmp[i * size + j] >> 1); \ 833 } \ 834 } \ 835} \ 836 837#define OP_PUT(a, b) (a) = (b) 838#define OP_ADD(a, b) (a) += (b) 839 840IVI_MC_TEMPLATE(8, _no_delta, OP_PUT) 841IVI_MC_TEMPLATE(8, _delta, OP_ADD) 842IVI_MC_TEMPLATE(4, _no_delta, OP_PUT) 843IVI_MC_TEMPLATE(4, _delta, OP_ADD) 844IVI_MC_AVG_TEMPLATE(8, _no_delta, OP_PUT) 845IVI_MC_AVG_TEMPLATE(8, _delta, OP_ADD) 846IVI_MC_AVG_TEMPLATE(4, _no_delta, OP_PUT) 847IVI_MC_AVG_TEMPLATE(4, _delta, OP_ADD) 848