Home
last modified time | relevance | path

Searched refs:in7 (Results 1 - 25 of 31) sorted by relevance

12

/third_party/ffmpeg/libavcodec/mips/
H A Didctdsp_msa.c28 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in put_pixels_clamped_msa() local
30 LD_SH8(block, 8, in0, in1, in2, in3, in4, in5, in6, in7); in put_pixels_clamped_msa()
31 CLIP_SH8_0_255(in0, in1, in2, in3, in4, in5, in6, in7); in put_pixels_clamped_msa()
33 PCKEV_B4_SH(in4, in4, in5, in5, in6, in6, in7, in7, in4, in5, in6, in7); in put_pixels_clamped_msa()
42 in7_d = __msa_copy_u_d((v2i64) in7, 0); in put_pixels_clamped_msa()
52 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in put_signed_pixels_clamped_msa() local
54 LD_SH8(block, 8, in0, in1, in2, in3, in4, in5, in6, in7); in put_signed_pixels_clamped_msa()
63 in7 in put_signed_pixels_clamped_msa()
86 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; add_pixels_clamped_msa() local
[all...]
H A Dhevc_idct_msa.c94 #define HEVC_IDCT8x8_COL(in0, in1, in2, in3, in4, in5, in6, in7, shift) \
104 ILVR_H4_SH(in4, in0, in6, in2, in5, in1, in3, in7, \
106 ILVL_H4_SH(in4, in0, in6, in2, in5, in1, in3, in7, \
133 PCKEV_H2_SH(sum0_l, sum0_r, sum3_l, sum3_r, in0, in7); \
334 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in hevc_idct_8x8_msa() local
336 LD_SH8(coeffs, 8, in0, in1, in2, in3, in4, in5, in6, in7); in hevc_idct_8x8_msa()
337 HEVC_IDCT8x8_COL(in0, in1, in2, in3, in4, in5, in6, in7, 7); in hevc_idct_8x8_msa()
338 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in hevc_idct_8x8_msa()
339 in0, in1, in2, in3, in4, in5, in6, in7); in hevc_idct_8x8_msa()
340 HEVC_IDCT8x8_COL(in0, in1, in2, in3, in4, in5, in6, in7, 1 in hevc_idct_8x8_msa()
353 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; hevc_idct_16x16_msa() local
443 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; hevc_idct_8x32_column_msa() local
604 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; hevc_idct_transpose_32x8_to_8x32() local
617 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; hevc_idct_transpose_8x32_to_32x8() local
740 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; hevc_addblk_8x8_msa() local
775 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; hevc_addblk_16x16_msa() local
842 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; hevc_addblk_32x32_msa() local
[all...]
H A Dmpegvideoencdsp_msa.c27 v16u8 in0, in1, in2, in3, in4, in5, in6, in7; in sum_u8src_16width_msa() local
30 LD_UB8(src, stride, in0, in1, in2, in3, in4, in5, in6, in7); in sum_u8src_16width_msa()
35 HADD_UB4_UB(in4, in5, in6, in7, in4, in5, in6, in7); in sum_u8src_16width_msa()
46 sum += HADD_UH_U32(in7); in sum_u8src_16width_msa()
H A Dsimple_idct_msa.c28 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in simple_idct_msa() local
41 LD_SH8(block, 8, in0, in1, in2, in3, in4, in5, in6, in7); in simple_idct_msa()
46 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in simple_idct_msa()
47 in0, in1, in2, in3, in4, in5, in6, in7); in simple_idct_msa()
48 select_vec = in1 | in2 | in3 | in4 | in5 | in6 | in7; in simple_idct_msa()
81 ILVRL_H2_SW(in5, in7, temp0_r, temp0_l); in simple_idct_msa()
120 in7 = (v8i16) __msa_bmnz_v((v16u8) a0_r, (v16u8) temp, (v16u8) select_vec); in simple_idct_msa()
121 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in simple_idct_msa()
122 in0, in1, in2, in3, in4, in5, in6, in7); in simple_idct_msa()
160 ILVRL_H2_SW(in5, in7, temp0_ in simple_idct_msa()
191 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; simple_idct_put_msa() local
367 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; simple_idct_add_msa() local
[all...]
H A Dvc1dsp_msa.c30 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in ff_vc1_inv_trans_8x8_msa() local
44 LD_SH8(block, 8, in0, in1, in2, in3, in4, in5, in6, in7); in ff_vc1_inv_trans_8x8_msa()
52 UNPCK_SH_SW(in7, in_r7, in_l7); in ff_vc1_inv_trans_8x8_msa()
136 in4, in5, in6, in7); in ff_vc1_inv_trans_8x8_msa()
137 ST_SH8(in0, in1, in2, in3, in4, in5, in6, in7, block, 8); in ff_vc1_inv_trans_8x8_msa()
142 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in ff_vc1_inv_trans_4x8_msa() local
159 LD_SH8(block, 8, in0, in1, in2, in3, in4, in5, in6, in7); in ff_vc1_inv_trans_4x8_msa()
167 UNPCK_R_SH_SW(in7, in_r7); in ff_vc1_inv_trans_4x8_msa()
233 v4i32 in0, in1, in2, in3, in4, in5, in6, in7; in ff_vc1_inv_trans_8x4_msa() local
252 UNPCK_SH_SW(t4, in3, in7); in ff_vc1_inv_trans_8x4_msa()
[all...]
H A Dcompute_antialias_float.h68 float in1, in2, in3, in4, in5, in6, in7, in8; in compute_antialias_mips_float() local
96 "lwc1 %[in7], 5*4(%[csa]) \t\n" in compute_antialias_mips_float()
101 "mul.s %[out4], %[in5], %[in7] \t\n" in compute_antialias_mips_float()
105 "nmsub.s %[out3], %[out3], %[in7], %[in8] \t\n" in compute_antialias_mips_float()
116 "lwc1 %[in7], 13*4(%[csa]) \t\n" in compute_antialias_mips_float()
121 "mul.s %[out4], %[in5], %[in7] \t\n" in compute_antialias_mips_float()
124 "nmsub.s %[out3], %[out3], %[in7], %[in8] \t\n" in compute_antialias_mips_float()
136 "lwc1 %[in7], 21*4(%[csa]) \t\n" in compute_antialias_mips_float()
141 "mul.s %[out4], %[in5], %[in7] \t\n" in compute_antialias_mips_float()
145 "nmsub.s %[out3], %[out3], %[in7], in compute_antialias_mips_float()
[all...]
H A Dvp9_idct_msa.c86 #define VP9_DOT_ADD_SUB_SRARI_PCK(in0, in1, in2, in3, in4, in5, in6, in7, \
94 DOTP_SH4_SW(in2, in3, in2, in3, in6, in6, in7, in7, \
116 #define VP9_ADST8(in0, in1, in2, in3, in4, in5, in6, in7, \
133 ILVRL_H2_SH(in0, in7, vec1_m, vec0_m); \
136 cnst1_m, cnst2_m, cnst3_m, in7, in0, \
152 BUTTERFLY_4(in7, in0, in2, in5, s1_m, s0_m, in2, in5); \
323 #define TRANSPOSE4X8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, \
330 ILVR_H4_SH(in1, in0, in3, in2, in5, in4, in7, in6, \
479 #define VP9_IDCT8x8_1D(in0, in1, in2, in3, in4, in5, in6, in7, \
620 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; vp9_idct8x8_12_colcol_addblk_msa() local
684 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; vp9_idct8x8_colcol_addblk_msa() local
711 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; vp9_iadst8x8_colcol_addblk_msa() local
837 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; vp9_iadst_idct_8x8_add_msa() local
864 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; vp9_idct_iadst_8x8_add_msa() local
[all...]
H A Dh263dsp_msa.c32 v16u8 in0, in1, in2, in3, in4, in5, in6, in7; in h263_h_loop_filter_msa() local
38 LD_UB8(src, stride, in0, in1, in2, in3, in4, in5, in6, in7); in h263_h_loop_filter_msa()
39 TRANSPOSE8x4_UB_UB(in0, in1, in2, in3, in4, in5, in6, in7, in h263_h_loop_filter_msa()
H A Dhevc_mc_bi_msa.c142 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in hevc_bi_copy_6w_msa() local
154 LD_SH8(src1_ptr, src2_stride, in0, in1, in2, in3, in4, in5, in6, in7); in hevc_bi_copy_6w_msa()
164 HEVC_BI_RND_CLIP4_MAX_SATU(in4, in5, in6, in7, dst4, dst5, dst6, dst7, in hevc_bi_copy_6w_msa()
193 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in hevc_bi_copy_8w_msa() local
253 in7); in hevc_bi_copy_8w_msa()
259 HEVC_BI_RND_CLIP4_MAX_SATU(in4, in5, in6, in7, dst4, dst5, dst6, in hevc_bi_copy_8w_msa()
281 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in hevc_bi_copy_12w_msa() local
289 LD_SH4(src1_ptr + 8, src2_stride, in4, in5, in6, in7); in hevc_bi_copy_12w_msa()
291 ILVR_D2_SH(in5, in4, in7, in6, in4, in5); in hevc_bi_copy_12w_msa()
319 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in hevc_bi_copy_16w_msa() local
358 v8i16 in0, in1, in2, in3, in4, in5, in6, in7, in8, in9, in10, in11, dst11; hevc_bi_copy_24w_msa() local
405 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; hevc_bi_copy_32w_msa() local
449 v8i16 in0, in1, in2, in3, in4, in5, in6, in7, in8, in9, in10, in11, dst11; hevc_bi_copy_48w_msa() local
503 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; hevc_bi_copy_64w_msa() local
544 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; hevc_hz_bi_8t_4w_msa() local
1187 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; hevc_vt_bi_8t_4w_msa() local
1350 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; hevc_vt_bi_8t_12w_msa() local
2283 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; hevc_hz_bi_4t_4x8multiple_msa() local
2615 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; hevc_hz_bi_4t_12w_msa() local
2750 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; hevc_hz_bi_4t_24w_msa() local
3012 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; hevc_vt_bi_4t_4x8multiple_msa() local
3393 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; hevc_vt_bi_4t_12w_msa() local
3686 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; hevc_vt_bi_4t_32w_msa() local
[all...]
H A Dmpegaudiodsp_mips_float.c75 float in1, in2, in3, in4, in5, in6, in7, in8; in ff_mpadsp_apply_window_mips_float() local
101 "lwc1 %[in7], 192*4(%[window]) \t\n" in ff_mpadsp_apply_window_mips_float()
110 "madd.s %[sum], %[sum], %[in7], %[in8] \t\n" in ff_mpadsp_apply_window_mips_float()
112 "lwc1 %[in7], 448*4(%[window]) \t\n" in ff_mpadsp_apply_window_mips_float()
123 "madd.s %[sum], %[sum], %[in7], %[in8] \t\n" in ff_mpadsp_apply_window_mips_float()
124 "lwc1 %[in7], 224*4(%[window]) \t\n" in ff_mpadsp_apply_window_mips_float()
135 "nmsub.s %[sum], %[sum], %[in7], %[in8] \t\n" in ff_mpadsp_apply_window_mips_float()
136 "lwc1 %[in7], 480*4(%[window]) \t\n" in ff_mpadsp_apply_window_mips_float()
145 "nmsub.s %[sum], %[sum], %[in7], %[in8] \t\n" in ff_mpadsp_apply_window_mips_float()
251 "lwc1 %[in7], 24 in ff_mpadsp_apply_window_mips_float()
[all...]
H A Dhevc_mc_biw_msa.c331 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in hevc_biwgt_copy_12w_msa() local
347 LD_SH4(src1_ptr + 8, src2_stride, in4, in5, in6, in7); in hevc_biwgt_copy_12w_msa()
350 ILVR_D2_SH(in5, in4, in7, in6, in4, in5); in hevc_biwgt_copy_12w_msa()
390 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in hevc_biwgt_copy_16w_msa() local
406 LD_SH4(src1_ptr + 8, src2_stride, in4, in5, in6, in7); in hevc_biwgt_copy_16w_msa()
417 HEVC_BIW_RND_CLIP4_MAX_SATU(tmp2, tmp3, tmp6, tmp7, in2, in3, in6, in7, in hevc_biwgt_copy_16w_msa()
445 v8i16 in0, in1, in2, in3, in4, in5, in6, in7, in8, in9, in10, in11, dst11; in hevc_biwgt_copy_24w_msa() local
461 LD_SH4(src1_ptr + 8, src2_stride, in4, in5, in6, in7); in hevc_biwgt_copy_24w_msa()
480 HEVC_BIW_RND_CLIP4_MAX_SATU(dst8, dst9, dst10, dst11, in3, in7, in10, in hevc_biwgt_copy_24w_msa()
509 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in hevc_biwgt_copy_32w_msa() local
622 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; hevc_biwgt_copy_64w_msa() local
1419 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; hevc_vt_biwgt_8t_4w_msa() local
2774 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; hevc_hz_biwgt_4t_4x8multiple_msa() local
3181 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; hevc_hz_biwgt_4t_12w_msa() local
3266 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; hevc_hz_biwgt_4t_16w_msa() local
3647 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; hevc_vt_biwgt_4t_4x8multiple_msa() local
4074 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; hevc_vt_biwgt_4t_12w_msa() local
4398 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; hevc_vt_biwgt_4t_32w_msa() local
[all...]
/third_party/ffmpeg/libavcodec/loongarch/
H A Dhevc_idct_lsx.c101 #define HEVC_IDCT8x8_COL(in0, in1, in2, in3, in4, in5, in6, in7, shift) \
111 DUP4_ARG2(__lsx_vilvl_h, in4, in0, in6, in2, in5, in1, in3, in7, \
113 DUP4_ARG2(__lsx_vilvh_h, in4, in0, in6, in2, in5, in1, in3, in7, \
140 in7 = __lsx_vssrarni_h_w(sum3_l, sum3_r, shift); \
363 __m128i in0, in1, in2, in3, in4, in5, in6, in7; in ff_hevc_idct_8x8_lsx() local
368 coeffs, 112, in4, in5, in6, in7); in ff_hevc_idct_8x8_lsx()
369 HEVC_IDCT8x8_COL(in0, in1, in2, in3, in4, in5, in6, in7, 7); in ff_hevc_idct_8x8_lsx()
370 LSX_TRANSPOSE8x8_H(in0, in1, in2, in3, in4, in5, in6, in7, in ff_hevc_idct_8x8_lsx()
371 in0, in1, in2, in3, in4, in5, in6, in7); in ff_hevc_idct_8x8_lsx()
372 HEVC_IDCT8x8_COL(in0, in1, in2, in3, in4, in5, in6, in7, 1 in ff_hevc_idct_8x8_lsx()
393 __m128i in0, in1, in2, in3, in4, in5, in6, in7; ff_hevc_idct_16x16_lsx() local
544 __m128i in0, in1, in2, in3, in4, in5, in6, in7; hevc_idct_8x32_column_lsx() local
764 __m128i in0, in1, in2, in3, in4, in5, in6, in7; hevc_idct_transpose_32x8_to_8x32() local
789 __m128i in0, in1, in2, in3, in4, in5, in6, in7; hevc_idct_transpose_8x32_to_32x8() local
[all...]
H A Dvp9_idct_lsx.c132 #define VP9_ILVLTRANS4x8_H(in0, in1, in2, in3, in4, in5, in6, in7, \
139 DUP4_ARG2(__lsx_vilvl_h, in1, in0, in3, in2, in5, in4, in7, in6, \
194 #define VP9_IDCT8x8_1D(in0, in1, in2, in3, in4, in5, in6, in7, \
207 VP9_MADD(in1, in7, in3, in5, k0_m, k1_m, k2_m, k3_m, in1, in7, in3, in5); \
208 DUP2_ARG2(__lsx_vsub_h, in1, in3, in7, in5, res0_m, res1_m); \
221 tp7_m = __lsx_vadd_h(in7, in5); \
253 __m128i in0, in1, in2, in3, in4, in5, in6, in7; in vp9_idct8x8_12_colcol_addblk_lsx() local
262 in4, in5, in6, in7); in vp9_idct8x8_12_colcol_addblk_lsx()
271 DUP4_ARG2(__lsx_vilvl_d,in1, in0, in3, in2, in5, in4, in7, in vp9_idct8x8_12_colcol_addblk_lsx()
336 __m128i in0, in1, in2, in3, in4, in5, in6, in7; vp9_idct8x8_colcol_addblk_lsx() local
[all...]
H A Dvc1dsp_lasx.c142 __m256i in0, in1, in2, in3, in4, in5, in6, in7; in ff_vc1_inv_trans_8x8_dc_lasx() local
153 0, dst + stride3, 0, in4, in5, in6, in7); in ff_vc1_inv_trans_8x8_dc_lasx()
155 DUP4_ARG2(__lasx_xvilvl_d, in1, in0, in3, in2, in5, in4, in7, in6, in ff_vc1_inv_trans_8x8_dc_lasx()
297 __m256i in0, in1, in2, in3, in4, in5, in6, in7; in ff_vc1_inv_trans_4x8_dc_lasx() local
307 0, dst + stride3, 0, in4, in5, in6, in7); in ff_vc1_inv_trans_4x8_dc_lasx()
309 DUP4_ARG2(__lasx_xvilvl_w, in1, in0, in3, in2, in5, in4, in7, in6, in ff_vc1_inv_trans_4x8_dc_lasx()
808 __m256i in0, in1, in2, in3, in4, in5, in6, in7, in put_vc1_mspel_mc_h_lasx() local
839 in7 = __lasx_xvldx(_src, stride3); in put_vc1_mspel_mc_h_lasx()
848 DUP4_ARG2(__lasx_xvilvl_b, in2, in0, in3, in1, in6, in4, in7, in5, in put_vc1_mspel_mc_h_lasx()
865 DUP4_ARG2(__lasx_xvilvh_b, in2, in0, in3, in1, in6, in4, in7, in in put_vc1_mspel_mc_h_lasx()
[all...]
H A Dhevc_mc_bi_lsx.c148 __m128i in0, in1, in2, in3, in4, in5, in6, in7; in hevc_bi_copy_6w_lsx() local
173 in7 = __lsx_vldx(src1_ptr, src2_stride_3x); in hevc_bi_copy_6w_lsx()
184 out3 = hevc_bi_rnd_clip(in6, dst6, in7, dst7); in hevc_bi_copy_6w_lsx()
244 __m128i in0, in1, in2, in3, in4, in5, in6, in7; in hevc_bi_copy_8w_lsx() local
275 in7 = __lsx_vldx(src1_ptr, src2_stride_3x); in hevc_bi_copy_8w_lsx()
280 out3 = hevc_bi_rnd_clip(in6, dst6, in7, dst7); in hevc_bi_copy_8w_lsx()
328 __m128i in0, in1, in2, in3, in4, in5, in6, in7; in hevc_bi_copy_12w_lsx() local
345 in7 = __lsx_vldx(_src1, src2_stride_3x); in hevc_bi_copy_12w_lsx()
348 DUP2_ARG2(__lsx_vilvl_d, in5, in4, in7, in6, in4, in5); in hevc_bi_copy_12w_lsx()
386 __m128i in0, in1, in2, in3, in4, in5, in6, in7; in hevc_bi_copy_16w_lsx() local
1073 __m128i in0, in1, in2, in3, in4, in5, in6, in7; hevc_hz_4t_24w_lsx() local
1221 __m128i in0, in1, in2, in3, in4, in5, in6, in7; hevc_vt_4t_12w_lsx() local
[all...]
/third_party/ffmpeg/libavutil/mips/
H A Dgeneric_macros_msa.h386 #define ST_V8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, pdst, stride) \
389 ST_V4(RTYPE, in4, in5, in6, in7, (pdst) + 4 * stride, stride); \
520 Arguments : Inputs - in0, in1, in2, in3, in4, in5, in6, in7, pdst, stride
527 #define ST12x8_UB(in0, in1, in2, in3, in4, in5, in6, in7, pdst, stride) \
542 out7_m = __msa_copy_u_d((v2i64) in7, 0); \
551 out15_m = __msa_copy_u_w((v4i32) in7, 2); \
597 #define AVER_UB4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \
601 AVER_UB2(RTYPE, in4, in5, in6, in7, out2, out3) \
954 in4, in5, in6, in7) \
957 CLIP_SH4_0_255(in4, in5, in6, in7); \
[all...]
/third_party/skia/third_party/externals/libwebp/src/dsp/
H A Dmsa_macro.h310 #define ST_B8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \
313 ST_B4(RTYPE, in4, in5, in6, in7, pdst + 4 * stride, stride); \
873 #define ILVR_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \
876 ILVR_B2(RTYPE, in4, in5, in6, in7, out2, out3); \
899 #define ILVR_H4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \
902 ILVR_H2(RTYPE, in4, in5, in6, in7, out2, out3); \
923 #define ILVR_D4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \
926 ILVR_D2(RTYPE, in4, in5, in6, in7, out2, out3); \
984 #define PCKEV_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \
987 PCKEV_B2(RTYPE, in4, in5, in6, in7, out
[all...]
H A Dlossless_sse2.c502 __m128i in7 = _mm_loadu_si128(in + 7); in ConvertBGRAToRGB_SSE2() local
504 VP8L32bToPlanar_SSE2(&in4, &in5, &in6, &in7); in ConvertBGRAToRGB_SSE2()
507 VP8PlanarTo24b_SSE2(&in1, &in5, &in2, &in6, &in3, &in7); in ConvertBGRAToRGB_SSE2()
513 _mm_storeu_si128(out + 5, in7); in ConvertBGRAToRGB_SSE2()
/third_party/ffmpeg/libavcodec/aarch64/
H A Dhevcdsp_idct_neon.S280 .macro tr_8x4 shift, in0,in0t, in1,in1t, in2,in2t, in3,in3t, in4,in4t, in5,in5t, in6,in6t, in7,in7t, p1, p2
294 sum_sub v30.4s, \in7\in7t, v0.h[5], +, \p2
295 sum_sub v28.4s, \in7\in7t, v0.h[7], +, \p2
296 sum_sub v29.4s, \in7\in7t, v0.h[6], -, \p2
306 sum_sub v31.4s, \in7\in7t, v0.h[4], -, \p2
318 fixsqrshrn \in7,\in7t, v24, \shift
411 .macro butterfly16 in0, in1, in2, in3, in4, in5, in6, in7
418 add \in5, \in6, \in7
419 sub \in6, \in6, \in7
433 .macro scale out0, out1, out2, out3, in0, in1, in2, in3, in4, in5, in6, in7, shif
[all...]
/third_party/node/deps/openssl/openssl/crypto/aes/asm/
H A Daesp8-ppc.pl678 my ($in0, $in1, $in2, $in3, $in4, $in5, $in6, $in7 )=map("v$_",(0..3,10..13));
795 lvx_u $in7,$x70,$inp
801 le?vperm $in7,$in7,$in7,$inpperm
804 vxor $out7,$in7,$rndkey0
864 # loop inX-in7 are loaded
933 vmr $ivec,$in7
935 lvx_u $in7,$x70,$inp
945 le?vperm $in7,
[all...]
/third_party/openssl/crypto/aes/asm/
H A Daesp8-ppc.pl678 my ($in0, $in1, $in2, $in3, $in4, $in5, $in6, $in7 )=map("v$_",(0..3,10..13));
795 lvx_u $in7,$x70,$inp
801 le?vperm $in7,$in7,$in7,$inpperm
804 vxor $out7,$in7,$rndkey0
864 # loop inX-in7 are loaded
933 vmr $ivec,$in7
935 lvx_u $in7,$x70,$inp
945 le?vperm $in7,
[all...]
/third_party/skia/third_party/externals/libjpeg-turbo/simd/x86_64/
H A Djidctflt-sse2.asm213 psrad xmm1, (DWORD_BIT-WORD_BIT) ; xmm1=in7=(70 71 72 73)
215 cvtdq2ps xmm1, xmm1 ; xmm1=in7=(70 71 72 73)
/third_party/ffmpeg/libavcodec/x86/
H A Dflacdsp.asm185 cglobal flac_decorrelate_indep%2_%1, 2, %2+2, %3+1, out, in0, in1, len, in2, in3, in4, in5, in6, in7
/third_party/skia/third_party/externals/libjpeg-turbo/simd/i386/
H A Djidctflt-sse2.asm220 psrad xmm1, (DWORD_BIT-WORD_BIT) ; xmm1=in7=(70 71 72 73)
222 cvtdq2ps xmm1, xmm1 ; xmm1=in7=(70 71 72 73)
/third_party/skia/third_party/externals/libpng/mips/
H A Dfilter_msa_intrinsics.c292 #define ADD4(in0, in1, in2, in3, in4, in5, in6, in7, \
296 ADD2(in4, in5, in6, in7, out2, out3); \

Completed in 48 milliseconds

12