1/* 2 * Loongson SIMD optimized vc1dsp 3 * 4 * Copyright (c) 2019 Loongson Technology Corporation Limited 5 * gxw <guxiwei-hf@loongson.cn> 6 * 7 * This file is part of FFmpeg. 8 * 9 * FFmpeg is free software; you can redistribute it and/or 10 * modify it under the terms of the GNU Lesser General Public 11 * License as published by the Free Software Foundation; either 12 * version 2.1 of the License, or (at your option) any later version. 13 * 14 * FFmpeg is distributed in the hope that it will be useful, 15 * but WITHOUT ANY WARRANTY; without even the implied warranty of 16 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU 17 * Lesser General Public License for more details. 18 * 19 * You should have received a copy of the GNU Lesser General Public 20 * License along with FFmpeg; if not, write to the Free Software 21 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA 22 */ 23 24#include "vc1dsp_mips.h" 25#include "constants.h" 26#include "libavutil/mips/generic_macros_msa.h" 27 28void ff_vc1_inv_trans_8x8_msa(int16_t block[64]) 29{ 30 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; 31 v4i32 in_r0, in_r1, in_r2, in_r3, in_r4, in_r5, in_r6, in_r7; 32 v4i32 in_l0, in_l1, in_l2, in_l3, in_l4, in_l5, in_l6, in_l7; 33 v4i32 t_r1, t_r2, t_r3, t_r4, t_r5, t_r6, t_r7, t_r8; 34 v4i32 t_l1, t_l2, t_l3, t_l4, t_l5, t_l6, t_l7, t_l8; 35 v4i32 cnst_12 = {12, 12, 12, 12}; 36 v4i32 cnst_4 = {4, 4, 4, 4}; 37 v4i32 cnst_16 = {16, 16, 16, 16}; 38 v4i32 cnst_6 = {6, 6, 6, 6}; 39 v4i32 cnst_15 = {15, 15, 15, 15}; 40 v4i32 cnst_9 = {9, 9, 9, 9}; 41 v4i32 cnst_1 = {1, 1, 1, 1}; 42 v4i32 cnst_64 = {64, 64, 64, 64}; 43 44 LD_SH8(block, 8, in0, in1, in2, in3, in4, in5, in6, in7); 45 UNPCK_SH_SW(in0, in_r0, in_l0); 46 UNPCK_SH_SW(in1, in_r1, in_l1); 47 UNPCK_SH_SW(in2, in_r2, in_l2); 48 UNPCK_SH_SW(in3, in_r3, in_l3); 49 UNPCK_SH_SW(in4, in_r4, in_l4); 50 UNPCK_SH_SW(in5, in_r5, in_l5); 51 UNPCK_SH_SW(in6, in_r6, in_l6); 52 UNPCK_SH_SW(in7, in_r7, in_l7); 53 // First loop 54 t_r1 = cnst_12 * (in_r0 + in_r4) + cnst_4; 55 t_l1 = cnst_12 * (in_l0 + in_l4) + cnst_4; 56 t_r2 = cnst_12 * (in_r0 - in_r4) + cnst_4; 57 t_l2 = cnst_12 * (in_l0 - in_l4) + cnst_4; 58 t_r3 = cnst_16 * in_r2 + cnst_6 * in_r6; 59 t_l3 = cnst_16 * in_l2 + cnst_6 * in_l6; 60 t_r4 = cnst_6 * in_r2 - cnst_16 * in_r6; 61 t_l4 = cnst_6 * in_l2 - cnst_16 * in_l6; 62 63 ADD4(t_r1, t_r3, t_l1, t_l3, t_r2, t_r4, t_l2, t_l4, t_r5, t_l5, t_r6, t_l6); 64 SUB4(t_r2, t_r4, t_l2, t_l4, t_r1, t_r3, t_l1, t_l3, t_r7, t_l7, t_r8, t_l8); 65 t_r1 = cnst_16 * in_r1 + cnst_15 * in_r3 + cnst_9 * in_r5 + cnst_4 * in_r7; 66 t_l1 = cnst_16 * in_l1 + cnst_15 * in_l3 + cnst_9 * in_l5 + cnst_4 * in_l7; 67 t_r2 = cnst_15 * in_r1 - cnst_4 * in_r3 - cnst_16 * in_r5 - cnst_9 * in_r7; 68 t_l2 = cnst_15 * in_l1 - cnst_4 * in_l3 - cnst_16 * in_l5 - cnst_9 * in_l7; 69 t_r3 = cnst_9 * in_r1 - cnst_16 * in_r3 + cnst_4 * in_r5 + cnst_15 * in_r7; 70 t_l3 = cnst_9 * in_l1 - cnst_16 * in_l3 + cnst_4 * in_l5 + cnst_15 * in_l7; 71 t_r4 = cnst_4 * in_r1 - cnst_9 * in_r3 + cnst_15 * in_r5 - cnst_16 * in_r7; 72 t_l4 = cnst_4 * in_l1 - cnst_9 * in_l3 + cnst_15 * in_l5 - cnst_16 * in_l7; 73 74 in_r0 = (t_r5 + t_r1) >> 3; 75 in_l0 = (t_l5 + t_l1) >> 3; 76 in_r1 = (t_r6 + t_r2) >> 3; 77 in_l1 = (t_l6 + t_l2) >> 3; 78 in_r2 = (t_r7 + t_r3) >> 3; 79 in_l2 = (t_l7 + t_l3) >> 3; 80 in_r3 = (t_r8 + t_r4) >> 3; 81 in_l3 = (t_l8 + t_l4) >> 3; 82 83 in_r4 = (t_r8 - t_r4) >> 3; 84 in_l4 = (t_l8 - t_l4) >> 3; 85 in_r5 = (t_r7 - t_r3) >> 3; 86 in_l5 = (t_l7 - t_l3) >> 3; 87 in_r6 = (t_r6 - t_r2) >> 3; 88 in_l6 = (t_l6 - t_l2) >> 3; 89 in_r7 = (t_r5 - t_r1) >> 3; 90 in_l7 = (t_l5 - t_l1) >> 3; 91 TRANSPOSE4x4_SW_SW(in_r0, in_r1, in_r2, in_r3, in_r0, in_r1, in_r2, in_r3); 92 TRANSPOSE4x4_SW_SW(in_l0, in_l1, in_l2, in_l3, in_l0, in_l1, in_l2, in_l3); 93 TRANSPOSE4x4_SW_SW(in_r4, in_r5, in_r6, in_r7, in_r4, in_r5, in_r6, in_r7); 94 TRANSPOSE4x4_SW_SW(in_l4, in_l5, in_l6, in_l7, in_l4, in_l5, in_l6, in_l7); 95 // Second loop 96 t_r1 = cnst_12 * (in_r0 + in_l0) + cnst_64; 97 t_l1 = cnst_12 * (in_r4 + in_l4) + cnst_64; 98 t_r2 = cnst_12 * (in_r0 - in_l0) + cnst_64; 99 t_l2 = cnst_12 * (in_r4 - in_l4) + cnst_64; 100 t_r3 = cnst_16 * in_r2 + cnst_6 * in_l2; 101 t_l3 = cnst_16 * in_r6 + cnst_6 * in_l6; 102 t_r4 = cnst_6 * in_r2 - cnst_16 * in_l2; 103 t_l4 = cnst_6 * in_r6 - cnst_16 * in_l6; 104 105 ADD4(t_r1, t_r3, t_l1, t_l3, t_r2, t_r4, t_l2, t_l4, t_r5, t_l5, t_r6, t_l6); 106 SUB4(t_r2, t_r4, t_l2, t_l4, t_r1, t_r3, t_l1, t_l3, t_r7, t_l7, t_r8, t_l8); 107 t_r1 = cnst_16 * in_r1 + cnst_15 * in_r3 + cnst_9 * in_l1 + cnst_4 * in_l3; 108 t_l1 = cnst_16 * in_r5 + cnst_15 * in_r7 + cnst_9 * in_l5 + cnst_4 * in_l7; 109 t_r2 = cnst_15 * in_r1 - cnst_4 * in_r3 - cnst_16 * in_l1 - cnst_9 * in_l3; 110 t_l2 = cnst_15 * in_r5 - cnst_4 * in_r7 - cnst_16 * in_l5 - cnst_9 * in_l7; 111 t_r3 = cnst_9 * in_r1 - cnst_16 * in_r3 + cnst_4 * in_l1 + cnst_15 * in_l3; 112 t_l3 = cnst_9 * in_r5 - cnst_16 * in_r7 + cnst_4 * in_l5 + cnst_15 * in_l7; 113 t_r4 = cnst_4 * in_r1 - cnst_9 * in_r3 + cnst_15 * in_l1 - cnst_16 * in_l3; 114 t_l4 = cnst_4 * in_r5 - cnst_9 * in_r7 + cnst_15 * in_l5 - cnst_16 * in_l7; 115 116 in_r0 = (t_r5 + t_r1) >> 7; 117 in_l0 = (t_l5 + t_l1) >> 7; 118 in_r1 = (t_r6 + t_r2) >> 7; 119 in_l1 = (t_l6 + t_l2) >> 7; 120 in_r2 = (t_r7 + t_r3) >> 7; 121 in_l2 = (t_l7 + t_l3) >> 7; 122 in_r3 = (t_r8 + t_r4) >> 7; 123 in_l3 = (t_l8 + t_l4) >> 7; 124 125 in_r4 = (t_r8 - t_r4 + cnst_1) >> 7; 126 in_l4 = (t_l8 - t_l4 + cnst_1) >> 7; 127 in_r5 = (t_r7 - t_r3 + cnst_1) >> 7; 128 in_l5 = (t_l7 - t_l3 + cnst_1) >> 7; 129 in_r6 = (t_r6 - t_r2 + cnst_1) >> 7; 130 in_l6 = (t_l6 - t_l2 + cnst_1) >> 7; 131 in_r7 = (t_r5 - t_r1 + cnst_1) >> 7; 132 in_l7 = (t_l5 - t_l1 + cnst_1) >> 7; 133 PCKEV_H4_SH(in_l0, in_r0, in_l1, in_r1, in_l2, in_r2, in_l3, in_r3, 134 in0, in1, in2, in3); 135 PCKEV_H4_SH(in_l4, in_r4, in_l5, in_r5, in_l6, in_r6, in_l7, in_r7, 136 in4, in5, in6, in7); 137 ST_SH8(in0, in1, in2, in3, in4, in5, in6, in7, block, 8); 138} 139 140void ff_vc1_inv_trans_4x8_msa(uint8_t *dest, ptrdiff_t linesize, int16_t *block) 141{ 142 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; 143 v4i32 in_r0, in_r1, in_r2, in_r3, in_r4, in_r5, in_r6, in_r7; 144 v4i32 t1, t2, t3, t4, t5, t6, t7, t8; 145 v4i32 dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7; 146 v16i8 zero_m = { 0 }; 147 v4i32 cnst_17 = {17, 17, 17, 17}; 148 v4i32 cnst_22 = {22, 22, 22, 22}; 149 v4i32 cnst_10 = {10, 10, 10, 10}; 150 v4i32 cnst_12 = {12, 12, 12, 12}; 151 v4i32 cnst_64 = {64, 64, 64, 64}; 152 v4i32 cnst_16 = {16, 16, 16, 16}; 153 v4i32 cnst_15 = {15, 15, 15, 15}; 154 v4i32 cnst_4 = {4, 4, 4, 4}; 155 v4i32 cnst_6 = {6, 6, 6, 6}; 156 v4i32 cnst_9 = {9, 9, 9, 9}; 157 v4i32 cnst_1 = {1, 1, 1, 1}; 158 159 LD_SH8(block, 8, in0, in1, in2, in3, in4, in5, in6, in7); 160 UNPCK_R_SH_SW(in0, in_r0); 161 UNPCK_R_SH_SW(in1, in_r1); 162 UNPCK_R_SH_SW(in2, in_r2); 163 UNPCK_R_SH_SW(in3, in_r3); 164 UNPCK_R_SH_SW(in4, in_r4); 165 UNPCK_R_SH_SW(in5, in_r5); 166 UNPCK_R_SH_SW(in6, in_r6); 167 UNPCK_R_SH_SW(in7, in_r7); 168 // First loop 169 TRANSPOSE4x4_SW_SW(in_r0, in_r1, in_r2, in_r3, in_r0, in_r1, in_r2, in_r3); 170 TRANSPOSE4x4_SW_SW(in_r4, in_r5, in_r6, in_r7, in_r4, in_r5, in_r6, in_r7); 171 t1 = cnst_17 * (in_r0 + in_r2) + cnst_4; 172 t5 = cnst_17 * (in_r4 + in_r6) + cnst_4; 173 t2 = cnst_17 * (in_r0 - in_r2) + cnst_4; 174 t6 = cnst_17 * (in_r4 - in_r6) + cnst_4; 175 t3 = cnst_22 * in_r1 + cnst_10 * in_r3; 176 t7 = cnst_22 * in_r5 + cnst_10 * in_r7; 177 t4 = cnst_22 * in_r3 - cnst_10 * in_r1; 178 t8 = cnst_22 * in_r7 - cnst_10 * in_r5; 179 180 in_r0 = (t1 + t3) >> 3; 181 in_r4 = (t5 + t7) >> 3; 182 in_r1 = (t2 - t4) >> 3; 183 in_r5 = (t6 - t8) >> 3; 184 in_r2 = (t2 + t4) >> 3; 185 in_r6 = (t6 + t8) >> 3; 186 in_r3 = (t1 - t3) >> 3; 187 in_r7 = (t5 - t7) >> 3; 188 TRANSPOSE4x4_SW_SW(in_r0, in_r1, in_r2, in_r3, in_r0, in_r1, in_r2, in_r3); 189 TRANSPOSE4x4_SW_SW(in_r4, in_r5, in_r6, in_r7, in_r4, in_r5, in_r6, in_r7); 190 PCKEV_H4_SH(in_r1, in_r0, in_r3, in_r2, in_r5, in_r4, in_r7, in_r6, 191 in0, in1, in2, in3); 192 ST_D8(in0, in1, in2, in3, 0, 1, 0, 1, 0, 1, 0, 1, block, 8); 193 // Second loop 194 t1 = cnst_12 * (in_r0 + in_r4) + cnst_64; 195 t2 = cnst_12 * (in_r0 - in_r4) + cnst_64; 196 t3 = cnst_16 * in_r2 + cnst_6 * in_r6; 197 t4 = cnst_6 * in_r2 - cnst_16 * in_r6; 198 t5 = t1 + t3, t6 = t2 + t4; 199 t7 = t2 - t4, t8 = t1 - t3; 200 t1 = cnst_16 * in_r1 + cnst_15 * in_r3 + cnst_9 * in_r5 + cnst_4 * in_r7; 201 t2 = cnst_15 * in_r1 - cnst_4 * in_r3 - cnst_16 * in_r5 - cnst_9 * in_r7; 202 t3 = cnst_9 * in_r1 - cnst_16 * in_r3 + cnst_4 * in_r5 + cnst_15 * in_r7; 203 t4 = cnst_4 * in_r1 - cnst_9 * in_r3 + cnst_15 * in_r5 - cnst_16 * in_r7; 204 LD_SW8(dest, linesize, dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7); 205 ILVR_B8_SW(zero_m, dst0, zero_m, dst1, zero_m, dst2, zero_m, dst3, 206 zero_m, dst4, zero_m, dst5, zero_m, dst6, zero_m, dst7, 207 dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7); 208 ILVR_H4_SW(zero_m, dst0, zero_m, dst1, zero_m, dst2, zero_m, dst3, 209 dst0, dst1, dst2, dst3); 210 ILVR_H4_SW(zero_m, dst4, zero_m, dst5, zero_m, dst6, zero_m, dst7, 211 dst4, dst5, dst6, dst7); 212 in_r0 = (t5 + t1) >> 7; 213 in_r1 = (t6 + t2) >> 7; 214 in_r2 = (t7 + t3) >> 7; 215 in_r3 = (t8 + t4) >> 7; 216 in_r4 = (t8 - t4 + cnst_1) >> 7; 217 in_r5 = (t7 - t3 + cnst_1) >> 7; 218 in_r6 = (t6 - t2 + cnst_1) >> 7; 219 in_r7 = (t5 - t1 + cnst_1) >> 7; 220 ADD4(in_r0, dst0, in_r1, dst1, in_r2, dst2, in_r3, dst3, 221 in_r0, in_r1, in_r2, in_r3); 222 ADD4(in_r4, dst4, in_r5, dst5, in_r6, dst6, in_r7, dst7, 223 in_r4, in_r5, in_r6, in_r7); 224 CLIP_SW8_0_255(in_r0, in_r1, in_r2, in_r3, in_r4, in_r5, in_r6, in_r7); 225 PCKEV_H4_SH(in_r1, in_r0, in_r3, in_r2, in_r5, in_r4, in_r7, in_r6, 226 in0, in1, in2, in3); 227 PCKEV_B2_SH(in1, in0, in3, in2, in0, in1); 228 ST_W8(in0, in1, 0, 1, 2, 3, 0, 1, 2, 3, dest, linesize); 229} 230 231void ff_vc1_inv_trans_8x4_msa(uint8_t *dest, ptrdiff_t linesize, int16_t *block) 232{ 233 v4i32 in0, in1, in2, in3, in4, in5, in6, in7; 234 v4i32 t1, t2, t3, t4, t5, t6, t7, t8; 235 v4i32 dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7; 236 v16i8 zero_m = { 0 }; 237 v4i32 cnst_17 = {17, 17, 17, 17}; 238 v4i32 cnst_22 = {22, 22, 22, 22}; 239 v4i32 cnst_10 = {10, 10, 10, 10}; 240 v4i32 cnst_12 = {12, 12, 12, 12}; 241 v4i32 cnst_64 = {64, 64, 64, 64}; 242 v4i32 cnst_16 = {16, 16, 16, 16}; 243 v4i32 cnst_15 = {15, 15, 15, 15}; 244 v4i32 cnst_4 = {4, 4, 4, 4}; 245 v4i32 cnst_6 = {6, 6, 6, 6}; 246 v4i32 cnst_9 = {9, 9, 9, 9}; 247 248 LD_SW4(block, 8, t1, t2, t3, t4); 249 UNPCK_SH_SW(t1, in0, in4); 250 UNPCK_SH_SW(t2, in1, in5); 251 UNPCK_SH_SW(t3, in2, in6); 252 UNPCK_SH_SW(t4, in3, in7); 253 TRANSPOSE4x4_SW_SW(in0, in1, in2, in3, in0, in1, in2, in3); 254 TRANSPOSE4x4_SW_SW(in4, in5, in6, in7, in4, in5, in6, in7); 255 // First loop 256 t1 = cnst_12 * (in0 + in4) + cnst_4; 257 t2 = cnst_12 * (in0 - in4) + cnst_4; 258 t3 = cnst_16 * in2 + cnst_6 * in6; 259 t4 = cnst_6 * in2 - cnst_16 * in6; 260 t5 = t1 + t3, t6 = t2 + t4; 261 t7 = t2 - t4, t8 = t1 - t3; 262 t1 = cnst_16 * in1 + cnst_15 * in3 + cnst_9 * in5 + cnst_4 * in7; 263 t2 = cnst_15 * in1 - cnst_4 * in3 - cnst_16 * in5 - cnst_9 * in7; 264 t3 = cnst_9 * in1 - cnst_16 * in3 + cnst_4 * in5 + cnst_15 * in7; 265 t4 = cnst_4 * in1 - cnst_9 * in3 + cnst_15 * in5 - cnst_16 * in7; 266 in0 = (t5 + t1) >> 3; 267 in1 = (t6 + t2) >> 3; 268 in2 = (t7 + t3) >> 3; 269 in3 = (t8 + t4) >> 3; 270 in4 = (t8 - t4) >> 3; 271 in5 = (t7 - t3) >> 3; 272 in6 = (t6 - t2) >> 3; 273 in7 = (t5 - t1) >> 3; 274 TRANSPOSE4x4_SW_SW(in0, in1, in2, in3, in0, in1, in2, in3); 275 TRANSPOSE4x4_SW_SW(in4, in5, in6, in7, in4, in5, in6, in7); 276 PCKEV_H4_SW(in4, in0, in5, in1, in6, in2, in7, in3, t1, t2, t3, t4); 277 ST_SW4(t1, t2, t3, t4, block, 8); 278 // Second loop 279 LD_SW4(dest, linesize, dst0, dst1, dst2, dst3); 280 ILVR_B4_SW(zero_m, dst0, zero_m, dst1, zero_m, dst2, zero_m, dst3, 281 dst0, dst1, dst2, dst3); 282 ILVL_H4_SW(zero_m, dst0, zero_m, dst1, zero_m, dst2, zero_m, dst3, 283 dst4, dst5, dst6, dst7); 284 ILVR_H4_SW(zero_m, dst0, zero_m, dst1, zero_m, dst2, zero_m, dst3, 285 dst0, dst1, dst2, dst3); 286 // Right part 287 t1 = cnst_17 * (in0 + in2) + cnst_64; 288 t2 = cnst_17 * (in0 - in2) + cnst_64; 289 t3 = cnst_22 * in1 + cnst_10 * in3; 290 t4 = cnst_22 * in3 - cnst_10 * in1; 291 in0 = (t1 + t3) >> 7; 292 in1 = (t2 - t4) >> 7; 293 in2 = (t2 + t4) >> 7; 294 in3 = (t1 - t3) >> 7; 295 ADD4(in0, dst0, in1, dst1, in2, dst2, in3, dst3, in0, in1, in2, in3); 296 CLIP_SW4_0_255(in0, in1, in2, in3); 297 // Left part 298 t5 = cnst_17 * (in4 + in6) + cnst_64; 299 t6 = cnst_17 * (in4 - in6) + cnst_64; 300 t7 = cnst_22 * in5 + cnst_10 * in7; 301 t8 = cnst_22 * in7 - cnst_10 * in5; 302 in4 = (t5 + t7) >> 7; 303 in5 = (t6 - t8) >> 7; 304 in6 = (t6 + t8) >> 7; 305 in7 = (t5 - t7) >> 7; 306 ADD4(in4, dst4, in5, dst5, in6, dst6, in7, dst7, in4, in5, in6, in7); 307 CLIP_SW4_0_255(in4, in5, in6, in7); 308 PCKEV_H4_SW(in4, in0, in5, in1, in6, in2, in7, in3, in0, in1, in2, in3); 309 PCKEV_B2_SW(in1, in0, in3, in2, in0, in1); 310 ST_D4(in0, in1, 0, 1, 0, 1, dest, linesize); 311} 312 313static void put_vc1_mspel_mc_h_v_msa(uint8_t *dst, const uint8_t *src, 314 ptrdiff_t stride, int hmode, int vmode, 315 int rnd) 316{ 317 v8i16 in_r0, in_r1, in_r2, in_r3, in_l0, in_l1, in_l2, in_l3; 318 v8i16 t0, t1, t2, t3, t4, t5, t6, t7; 319 v8i16 t8, t9, t10, t11, t12, t13, t14, t15; 320 v8i16 cnst_para0, cnst_para1, cnst_para2, cnst_para3, cnst_r; 321 static const int para_value[][4] = {{4, 53, 18, 3}, 322 {1, 9, 9, 1}, 323 {3, 18, 53, 4}}; 324 static const int shift_value[] = {0, 5, 1, 5}; 325 int shift = (shift_value[hmode] + shift_value[vmode]) >> 1; 326 int r = (1 << (shift - 1)) + rnd - 1; 327 cnst_r = __msa_fill_h(r); 328 src -= 1, src -= stride; 329 cnst_para0 = __msa_fill_h(para_value[vmode - 1][0]); 330 cnst_para1 = __msa_fill_h(para_value[vmode - 1][1]); 331 cnst_para2 = __msa_fill_h(para_value[vmode - 1][2]); 332 cnst_para3 = __msa_fill_h(para_value[vmode - 1][3]); 333 LD_SH4(src, stride, in_l0, in_l1, in_l2, in_l3); 334 UNPCK_UB_SH(in_l0, in_r0, in_l0); 335 UNPCK_UB_SH(in_l1, in_r1, in_l1); 336 UNPCK_UB_SH(in_l2, in_r2, in_l2); 337 UNPCK_UB_SH(in_l3, in_r3, in_l3); 338 // row 0 339 t0 = cnst_para1 * in_r1 + cnst_para2 * in_r2 340 - cnst_para0 * in_r0 - cnst_para3 * in_r3; 341 t8 = cnst_para1 * in_l1 + cnst_para2 * in_l2 342 - cnst_para0 * in_l0 - cnst_para3 * in_l3; 343 in_l0 = LD_SH(src + 4 * stride); 344 UNPCK_UB_SH(in_l0, in_r0, in_l0); 345 // row 1 346 t1 = cnst_para1 * in_r2 + cnst_para2 * in_r3 347 - cnst_para0 * in_r1 - cnst_para3 * in_r0; 348 t9 = cnst_para1 * in_l2 + cnst_para2 * in_l3 349 - cnst_para0 * in_l1 - cnst_para3 * in_l0; 350 in_l1 = LD_SH(src + 5 * stride); 351 UNPCK_UB_SH(in_l1, in_r1, in_l1); 352 // row 2 353 t2 = cnst_para1 * in_r3 + cnst_para2 * in_r0 354 - cnst_para0 * in_r2 - cnst_para3 * in_r1; 355 t10 = cnst_para1 * in_l3 + cnst_para2 * in_l0 356 - cnst_para0 * in_l2 - cnst_para3 * in_l1; 357 in_l2 = LD_SH(src + 6 * stride); 358 UNPCK_UB_SH(in_l2, in_r2, in_l2); 359 // row 3 360 t3 = cnst_para1 * in_r0 + cnst_para2 * in_r1 361 - cnst_para0 * in_r3 - cnst_para3 * in_r2; 362 t11 = cnst_para1 * in_l0 + cnst_para2 * in_l1 363 - cnst_para0 * in_l3 - cnst_para3 * in_l2; 364 in_l3 = LD_SH(src + 7 * stride); 365 UNPCK_UB_SH(in_l3, in_r3, in_l3); 366 // row 4 367 t4 = cnst_para1 * in_r1 + cnst_para2 * in_r2 368 - cnst_para0 * in_r0 - cnst_para3 * in_r3; 369 t12 = cnst_para1 * in_l1 + cnst_para2 * in_l2 370 - cnst_para0 * in_l0 - cnst_para3 * in_l3; 371 in_l0 = LD_SH(src + 8 * stride); 372 UNPCK_UB_SH(in_l0, in_r0, in_l0); 373 // row 5 374 t5 = cnst_para1 * in_r2 + cnst_para2 * in_r3 375 - cnst_para0 * in_r1 - cnst_para3 * in_r0; 376 t13 = cnst_para1 * in_l2 + cnst_para2 * in_l3 377 - cnst_para0 * in_l1 - cnst_para3 * in_l0; 378 in_l1 = LD_SH(src + 9 * stride); 379 UNPCK_UB_SH(in_l1, in_r1, in_l1); 380 // row 6 381 t6 = cnst_para1 * in_r3 + cnst_para2 * in_r0 382 - cnst_para0 * in_r2 - cnst_para3 * in_r1; 383 t14 = cnst_para1 * in_l3 + cnst_para2 * in_l0 384 - cnst_para0 * in_l2 - cnst_para3 * in_l1; 385 in_l2 = LD_SH(src + 10 * stride); 386 UNPCK_UB_SH(in_l2, in_r2, in_l2); 387 // row 7 388 t7 = cnst_para1 * in_r0 + cnst_para2 * in_r1 389 - cnst_para0 * in_r3 - cnst_para3 * in_r2; 390 t15 = cnst_para1 * in_l0 + cnst_para2 * in_l1 391 - cnst_para0 * in_l3 - cnst_para3 * in_l2; 392 393 ADD4(t0, cnst_r, t1, cnst_r, t2, cnst_r, t3, cnst_r, t0, t1, t2, t3); 394 ADD4(t4, cnst_r, t5, cnst_r, t6, cnst_r, t7, cnst_r, t4, t5, t6, t7); 395 ADD4(t8, cnst_r, t9, cnst_r, t10, cnst_r, t11, cnst_r, 396 t8, t9, t10, t11); 397 ADD4(t12, cnst_r, t13, cnst_r, t14, cnst_r, t15, cnst_r, 398 t12, t13, t14, t15); 399 t0 >>= shift, t1 >>= shift, t2 >>= shift, t3 >>= shift; 400 t4 >>= shift, t5 >>= shift, t6 >>= shift, t7 >>= shift; 401 t8 >>= shift, t9 >>= shift, t10 >>= shift, t11 >>= shift; 402 t12 >>= shift, t13 >>= shift, t14 >>= shift, t15 >>= shift; 403 TRANSPOSE8x8_SH_SH(t0, t1, t2, t3, t4, t5, t6, t7, 404 t0, t1, t2, t3, t4, t5, t6, t7); 405 TRANSPOSE8x8_SH_SH(t8, t9, t10, t11, t12, t13, t14, t15, 406 t8, t9, t10, t11, t12, t13, t14, t15); 407 cnst_para0 = __msa_fill_h(para_value[hmode - 1][0]); 408 cnst_para1 = __msa_fill_h(para_value[hmode - 1][1]); 409 cnst_para2 = __msa_fill_h(para_value[hmode - 1][2]); 410 cnst_para3 = __msa_fill_h(para_value[hmode - 1][3]); 411 r = 64 - rnd; 412 cnst_r = __msa_fill_h(r); 413 // col 0 ~ 7 414 t0 = cnst_para1 * t1 + cnst_para2 * t2 - cnst_para0 * t0 - cnst_para3 * t3; 415 t1 = cnst_para1 * t2 + cnst_para2 * t3 - cnst_para0 * t1 - cnst_para3 * t4; 416 t2 = cnst_para1 * t3 + cnst_para2 * t4 - cnst_para0 * t2 - cnst_para3 * t5; 417 t3 = cnst_para1 * t4 + cnst_para2 * t5 - cnst_para0 * t3 - cnst_para3 * t6; 418 t4 = cnst_para1 * t5 + cnst_para2 * t6 - cnst_para0 * t4 - cnst_para3 * t7; 419 t5 = cnst_para1 * t6 + cnst_para2 * t7 - cnst_para0 * t5 - cnst_para3 * t8; 420 t6 = cnst_para1 * t7 + cnst_para2 * t8 - cnst_para0 * t6 - cnst_para3 * t9; 421 t7 = cnst_para1 * t8 + cnst_para2 * t9 - cnst_para0 * t7 - cnst_para3 * t10; 422 ADD4(t0, cnst_r, t1, cnst_r, t2, cnst_r, t3, cnst_r, t0, t1, t2, t3); 423 ADD4(t4, cnst_r, t5, cnst_r, t6, cnst_r, t7, cnst_r, t4, t5, t6, t7); 424 t0 >>= 7, t1 >>= 7, t2 >>= 7, t3 >>= 7; 425 t4 >>= 7, t5 >>= 7, t6 >>= 7, t7 >>= 7; 426 TRANSPOSE8x8_SH_SH(t0, t1, t2, t3, t4, t5, t6, t7, 427 t0, t1, t2, t3, t4, t5, t6, t7); 428 CLIP_SH8_0_255(t0, t1, t2, t3, t4, t5, t6, t7); 429 PCKEV_B4_SH(t1, t0, t3, t2, t5, t4, t7, t6, t0, t1, t2, t3); 430 ST_D8(t0, t1, t2, t3, 0, 1, 0, 1, 0, 1, 0, 1, dst, stride); 431} 432 433#define PUT_VC1_MSPEL_MC_MSA(hmode, vmode) \ 434void ff_put_vc1_mspel_mc ## hmode ## vmode ## _msa(uint8_t *dst, \ 435 const uint8_t *src, \ 436 ptrdiff_t stride, int rnd) \ 437{ \ 438 put_vc1_mspel_mc_h_v_msa(dst, src, stride, hmode, vmode, rnd); \ 439} \ 440void ff_put_vc1_mspel_mc ## hmode ## vmode ## _16_msa(uint8_t *dst, \ 441 const uint8_t *src, \ 442 ptrdiff_t stride, int rnd) \ 443{ \ 444 put_vc1_mspel_mc_h_v_msa(dst, src, stride, hmode, vmode, rnd); \ 445 put_vc1_mspel_mc_h_v_msa(dst + 8, src + 8, stride, hmode, vmode, rnd); \ 446 dst += 8 * stride, src += 8 * stride; \ 447 put_vc1_mspel_mc_h_v_msa(dst, src, stride, hmode, vmode, rnd); \ 448 put_vc1_mspel_mc_h_v_msa(dst + 8, src + 8, stride, hmode, vmode, rnd); \ 449} 450 451PUT_VC1_MSPEL_MC_MSA(1, 1); 452PUT_VC1_MSPEL_MC_MSA(1, 2); 453PUT_VC1_MSPEL_MC_MSA(1, 3); 454 455PUT_VC1_MSPEL_MC_MSA(2, 1); 456PUT_VC1_MSPEL_MC_MSA(2, 2); 457PUT_VC1_MSPEL_MC_MSA(2, 3); 458 459PUT_VC1_MSPEL_MC_MSA(3, 1); 460PUT_VC1_MSPEL_MC_MSA(3, 2); 461PUT_VC1_MSPEL_MC_MSA(3, 3); 462