/third_party/python/Modules/_blake2/impl/ |
H A D | blake2s-round.h | 44 #define G1(row1,row2,row3,row4,buf) \ 46 row4 = _mm_xor_si128( row4, row1 ); \ 47 row4 = _mm_roti_epi32(row4, -16); \ 48 row3 = _mm_add_epi32( row3, row4 ); \ 52 #define G2(row1,row2,row3,row4,buf) \ 54 row4 = _mm_xor_si128( row4, row1 ); \ 55 row4 [all...] |
H A D | blake2s.c | 268 __m128i row1, row2, row3, row4; in blake2s_compress() local 307 row4 = _mm_xor_si128( _mm_setr_epi32( 0x510E527F, 0x9B05688C, 0x1F83D9AB, 0x5BE0CD19 ), LOADU( &S->t[0] ) ); in blake2s_compress() 319 STOREU( &S->h[4], _mm_xor_si128( ff1, _mm_xor_si128( row2, row4 ) ) ); in blake2s_compress()
|
/third_party/skia/third_party/externals/libjpeg-turbo/simd/arm/aarch32/ |
H A D | jchuff-neon.c | 138 int16x8_t row4 = vld1q_dup_s16(block + 35); in jsimd_huff_encode_one_block_neon() local 139 row4 = vld1q_lane_s16(block + 42, row4, 1); in jsimd_huff_encode_one_block_neon() 140 row4 = vld1q_lane_s16(block + 49, row4, 2); in jsimd_huff_encode_one_block_neon() 141 row4 = vld1q_lane_s16(block + 56, row4, 3); in jsimd_huff_encode_one_block_neon() 142 row4 = vld1q_lane_s16(block + 57, row4, 4); in jsimd_huff_encode_one_block_neon() 143 row4 in jsimd_huff_encode_one_block_neon() [all...] |
/third_party/skia/third_party/externals/libjpeg-turbo/simd/arm/ |
H A D | jfdctfst-neon.c | 156 int16x8_t row4 = vreinterpretq_s16_s32(rows_04.val[1]); in jsimd_fdct_ifast_neon() local 169 tmp3 = vaddq_s16(row3, row4); in jsimd_fdct_ifast_neon() 170 tmp4 = vsubq_s16(row3, row4); in jsimd_fdct_ifast_neon() 179 row4 = vsubq_s16(tmp10, tmp11); in jsimd_fdct_ifast_neon() 210 vst1q_s16(data + 4 * DCTSIZE, row4); in jsimd_fdct_ifast_neon()
|
H A D | jidctfst-neon.c | 73 int16x8_t row4 = vld1q_s16(coef_block + 4 * DCTSIZE); in jsimd_idct_ifast_neon() local 86 bitmap = vorrq_s16(bitmap, row4); in jsimd_idct_ifast_neon() 105 row4 = dcval; in jsimd_idct_ifast_neon() 129 int16x4_t tmp2 = vmul_s16(vget_high_s16(row4), quant_row4); in jsimd_idct_ifast_neon() 182 row4 = vcombine_s16(dcval, vadd_s16(tmp3, tmp4)); in jsimd_idct_ifast_neon() 204 int16x4_t tmp2 = vmul_s16(vget_low_s16(row4), quant_row4); in jsimd_idct_ifast_neon() 257 row4 = vcombine_s16(vadd_s16(tmp3, tmp4), dcval); in jsimd_idct_ifast_neon() 274 int16x8_t tmp2 = vmulq_s16(row4, quant_row4); in jsimd_idct_ifast_neon() 327 row4 = vaddq_s16(tmp3, tmp4); in jsimd_idct_ifast_neon() 334 int16x8x2_t rows_45 = vtrnq_s16(row4, row in jsimd_idct_ifast_neon() [all...] |
H A D | jidctint-neon.c | 101 int16x4_t row4, 205 int16x4_t row4 = vld1_s16(coef_block + 4 * DCTSIZE); in jsimd_idct_islow_neon() local 223 bitmap = vorr_s16(bitmap, row4); in jsimd_idct_islow_neon() 244 jsimd_idct_islow_pass1_regular(row0, row1, row2, row3, row4, row5, in jsimd_idct_islow_neon() 258 row4 = vld1_s16(coef_block + 4 * DCTSIZE + 4); in jsimd_idct_islow_neon() 276 bitmap = vorr_s16(bitmap, row4); in jsimd_idct_islow_neon() 304 jsimd_idct_islow_pass1_regular(row0, row1, row2, row3, row4, row5, in jsimd_idct_islow_neon() 342 int16x4_t row4, in jsimd_idct_islow_pass1_regular() 377 z3_s16 = vmul_s16(row4, quant_row4); in jsimd_idct_islow_pass1_regular() 338 jsimd_idct_islow_pass1_regular(int16x4_t row0, int16x4_t row1, int16x4_t row2, int16x4_t row3, int16x4_t row4, int16x4_t row5, int16x4_t row6, int16x4_t row7, int16x4_t quant_row0, int16x4_t quant_row1, int16x4_t quant_row2, int16x4_t quant_row3, int16x4_t quant_row4, int16x4_t quant_row5, int16x4_t quant_row6, int16x4_t quant_row7, int16_t *workspace_1, int16_t *workspace_2) jsimd_idct_islow_pass1_regular() argument
|
H A D | jfdctint-neon.c | 255 int16x8_t row4 = vreinterpretq_s16_s32(rows_04.val[1]); in jsimd_fdct_islow_neon() local 268 tmp3 = vaddq_s16(row3, row4); in jsimd_fdct_islow_neon() 269 tmp4 = vsubq_s16(row3, row4); in jsimd_fdct_islow_neon() 278 row4 = vrshrq_n_s16(vsubq_s16(tmp10, tmp11), PASS1_BITS); in jsimd_fdct_islow_neon() 372 vst1q_s16(data + 4 * DCTSIZE, row4); in jsimd_fdct_islow_neon()
|
H A D | jquanti-neon.c | 65 int16x8_t row4 = in jsimd_convsamp_neon() local 78 vst1q_s16(workspace + 4 * DCTSIZE, row4); in jsimd_convsamp_neon()
|
H A D | jcphuff-neon.c | 202 int16x8_t row4 = vld1q_s16(values + 4 * DCTSIZE); in jsimd_encode_mcu_AC_first_prepare_neon() local 211 uint8x8_t row4_eq0 = vmovn_u16(vceqq_s16(row4, vdupq_n_s16(0))); in jsimd_encode_mcu_AC_first_prepare_neon()
|
/third_party/ffmpeg/libavcodec/mips/ |
H A D | vp8_lpf_msa.c | 380 v16u8 row0, row1, row2, row3, row4, row5, row6, row7, row8; in ff_vp8_h_loop_filter16_msa() local 388 LD_UB8(temp_src, pitch, row0, row1, row2, row3, row4, row5, row6, row7); in ff_vp8_h_loop_filter16_msa() 392 TRANSPOSE16x8_UB_UB(row0, row1, row2, row3, row4, row5, row6, row7, in ff_vp8_h_loop_filter16_msa() 445 v16u8 row0, row1, row2, row3, row4, row5, row6, row7, row8; in ff_vp8_h_loop_filter8uv_msa() local 453 LD_UB8(src_u - 4, pitch, row0, row1, row2, row3, row4, row5, row6, row7); in ff_vp8_h_loop_filter8uv_msa() 456 TRANSPOSE16x8_UB_UB(row0, row1, row2, row3, row4, row5, row6, row7, in ff_vp8_h_loop_filter8uv_msa() 525 v16u8 row0, row1, row2, row3, row4, row5, row6, row7, row8; in ff_vp8_h_loop_filter_simple_msa() local 531 LD_UB8(temp_src, pitch, row0, row1, row2, row3, row4, row5, row6, row7); in ff_vp8_h_loop_filter_simple_msa() 535 TRANSPOSE16x4_UB_UB(row0, row1, row2, row3, row4, row5, row6, row7, in ff_vp8_h_loop_filter_simple_msa() 595 v16u8 row0, row1, row2, row3, row4, row in ff_vp8_h_loop_filter8uv_inner_msa() local 649 v16u8 row0, row1, row2, row3, row4, row5, row6, row7; ff_vp8_h_loop_filter16_inner_msa() local [all...] |
H A D | vp9_lpf_msa.c | 1233 v16u8 row0, row1, row2, row3, row4, row5, row6, row7; in ff_loop_filter_h_44_16_msa() local 1237 LD_UB8(src - 4, pitch, row0, row1, row2, row3, row4, row5, row6, row7); in ff_loop_filter_h_44_16_msa() 1241 TRANSPOSE16x8_UB_UB(row0, row1, row2, row3, row4, row5, row6, row7, in ff_loop_filter_h_44_16_msa() 1358 v16u8 row4, row5, row6, row7, row12, row13, row14, row15; in ff_loop_filter_h_88_16_msa() local 1370 LD_UB8(temp_src, pitch, p0, p1, p2, p3, row4, row5, row6, row7); in ff_loop_filter_h_88_16_msa() 1375 TRANSPOSE16x8_UB_UB(p0, p1, p2, p3, row4, row5, row6, row7, in ff_loop_filter_h_88_16_msa() 1471 v16u8 row4, row5, row6, row7, row12, row13, row14, row15; in ff_loop_filter_h_84_16_msa() local 1480 LD_UB8(temp_src, pitch, p0, p1, p2, p3, row4, row5, row6, row7); in ff_loop_filter_h_84_16_msa() 1485 TRANSPOSE16x8_UB_UB(p0, p1, p2, p3, row4, row5, row6, row7, in ff_loop_filter_h_84_16_msa() 1574 v16u8 row4, row in ff_loop_filter_h_48_16_msa() local 1713 v16u8 row0, row1, row2, row3, row4, row5, row6, row7; vp9_transpose_16x16() local [all...] |
H A D | h264dsp_msa.c | 829 v16u8 row0, row1, row2, row3, row4, row5, row6, row7; in avc_loopfilter_luma_intra_edge_ver_msa() local 832 LD_UB8(src, img_width, row0, row1, row2, row3, row4, row5, row6, row7); in avc_loopfilter_luma_intra_edge_ver_msa() 837 row4, row5, row6, row7, in avc_loopfilter_luma_intra_edge_ver_msa() 1240 v16u8 row0, row1, row2, row3, row4, row5, row6, row7; in avc_loopfilter_cb_or_cr_intra_edge_ver_msa() local 1243 row0, row1, row2, row3, row4, row5, row6, row7); in avc_loopfilter_cb_or_cr_intra_edge_ver_msa() 1245 TRANSPOSE8x4_UB_UB(row0, row1, row2, row3, row4, row5, row6, row7, in avc_loopfilter_cb_or_cr_intra_edge_ver_msa() 1926 v16u8 row0, row1, row2, row3, row4, row5, row6, row7; in avc_loopfilter_cb_or_cr_inter_edge_ver_msa() local 1952 row0, row1, row2, row3, row4, row5, row6, row7); in avc_loopfilter_cb_or_cr_inter_edge_ver_msa() 1955 row4, row5, row6, row7, in avc_loopfilter_cb_or_cr_inter_edge_ver_msa()
|
/third_party/ffmpeg/libavcodec/loongarch/ |
H A D | h264dsp_lasx.c | 96 __m256i row0, row1, row2, row3, row4, row5, row6, row7; in ff_h264_h_lpf_luma_8_lasx() local 103 src, img_width_3x, row4, row5, row6, row7); in ff_h264_h_lpf_luma_8_lasx() 114 LASX_TRANSPOSE16x8_B(row0, row1, row2, row3, row4, row5, row6, in ff_h264_h_lpf_luma_8_lasx() 201 __m256i row0, row1, row2, row3, row4, row5, row6, row7; in ff_h264_h_lpf_luma_8_lasx() local 212 DUP2_ARG2(__lasx_xvilvl_b, row2, row0, row3, row1, row4, row6); in ff_h264_h_lpf_luma_8_lasx() 214 DUP4_ARG2(__lasx_xvperm_w, row4, control, row5, control, row6, in ff_h264_h_lpf_luma_8_lasx() 215 control, row7, control, row4, row5, row6, row7); in ff_h264_h_lpf_luma_8_lasx() 216 __lasx_xvstelm_d(row4, src, 0, 0); in ff_h264_h_lpf_luma_8_lasx() 217 __lasx_xvstelm_d(row4, src + img_width, 0, 1); in ff_h264_h_lpf_luma_8_lasx() 219 __lasx_xvstelm_d(row4, sr in ff_h264_h_lpf_luma_8_lasx() 393 __m256i row0, row1, row2, row3, row4, row5, row6, row7; ff_h264_h_lpf_chroma_8_lasx() local 597 __m256i row0, row1, row2, row3, row4, row5, row6, row7; ff_h264_h_lpf_luma_intra_8_lasx() local 717 __m256i row0, row1, row2, row3, row4, row5, row6, row7; ff_h264_h_lpf_luma_intra_8_lasx() local 892 __m256i row0, row1, row2, row3, row4, row5, row6, row7; ff_h264_h_lpf_chroma_intra_8_lasx() local [all...] |
H A D | vp8_lpf_lsx.c | 335 __m128i row0, row1, row2, row3, row4, row5, row6, row7, row8; in ff_vp8_h_loop_filter16_lsx() local 352 temp_src + stride3, 0, row4, row5, row6, row7); in ff_vp8_h_loop_filter16_lsx() 360 LSX_TRANSPOSE16x8_B(row0, row1, row2, row3, row4, row5, row6, row7, row8, row9, row10, in ff_vp8_h_loop_filter16_lsx() 422 __m128i row0, row1, row2, row3, row4, row5, row6, row7, row8; in ff_vp8_h_loop_filter8uv_lsx() local 439 temp_src + stride3, 0, row4, row5, row6, row7); in ff_vp8_h_loop_filter8uv_lsx() 448 LSX_TRANSPOSE16x8_B(row0, row1, row2, row3, row4, row5, row6, row7, in ff_vp8_h_loop_filter8uv_lsx()
|
H A D | vp9_lpf_lsx.c | 1408 __m128i row0, row1, row2, row3, row4, row5, row6, row7; in ff_loop_filter_h_44_16_lsx() local 1416 row4 = __lsx_vld(dst_tmp, 0); in ff_loop_filter_h_44_16_lsx() 1428 LSX_TRANSPOSE16x8_B(row0, row1, row2, row3, row4, row5, row6, row7, in ff_loop_filter_h_44_16_lsx() 1606 __m128i row4, row5, row6, row7, row12, row13, row14, row15; in ff_loop_filter_h_88_16_lsx() local 1620 row4 = __lsx_vld(dst_tmp, 0); in ff_loop_filter_h_88_16_lsx() 1633 LSX_TRANSPOSE16x8_B(p0, p1, p2, p3, row4, row5, row6, row7, in ff_loop_filter_h_88_16_lsx() 1791 __m128i row4, row5, row6, row7, row12, row13, row14, row15; in ff_loop_filter_h_84_16_lsx() local 1802 row4 = __lsx_vld(dst_tmp, 0); in ff_loop_filter_h_84_16_lsx() 1815 LSX_TRANSPOSE16x8_B(p0, p1, p2, p3, row4, row5, row6, row7, in ff_loop_filter_h_84_16_lsx() 1966 __m128i row4, row in ff_loop_filter_h_48_16_lsx() local 2198 __m128i row0, row1, row2, row3, row4, row5, row6, row7; vp9_transpose_16x16() local [all...] |
/third_party/skia/gm/ |
H A D | hardstop_gradients.cpp | 118 SkScalar row4[] = {0.00f, 0.25f, 0.50f, 0.50f, 1.00f}; variable 128 row4,
|
/third_party/skia/third_party/externals/libwebp/src/dsp/ |
H A D | dec_msa.c | 364 v16u8 row0, row1, row2, row3, row4, row5, row6, row7, row8; in HFilter16() local 371 LD_UB8(ptmp, stride, row0, row1, row2, row3, row4, row5, row6, row7); in HFilter16() 374 TRANSPOSE16x8_UB_UB(row0, row1, row2, row3, row4, row5, row6, row7, in HFilter16() 446 v16u8 row0, row1, row2, row3, row4, row5, row6, row7; in HFilterVertEdge16i() local 453 LD_UB8(src - 4, stride, row0, row1, row2, row3, row4, row5, row6, row7); in HFilterVertEdge16i() 456 TRANSPOSE16x8_UB_UB(row0, row1, row2, row3, row4, row5, row6, row7, in HFilterVertEdge16i() 530 v16u8 row0, row1, row2, row3, row4, row5, row6, row7, row8; in HFilter8() local 537 LD_UB8(ptmp_src_u, stride, row0, row1, row2, row3, row4, row5, row6, row7); in HFilter8() 540 TRANSPOSE16x8_UB_UB(row0, row1, row2, row3, row4, row5, row6, row7, in HFilter8() 595 v16u8 row0, row1, row2, row3, row4, row in HFilter8i() local 636 v16u8 p1, p0, q1, q0, mask, row0, row1, row2, row3, row4, row5, row6, row7; SimpleHFilter16() local [all...] |
H A D | dec_neon.c | 175 const uint8x16_t row4 = LOAD_UV_8(4); in Load8x8x2T_NEON() local 192 const uint8x16x2_t row45 = vtrnq_u8(row4, row5); // ... in Load8x8x2T_NEON()
|
/third_party/skia/third_party/externals/d3d12allocator/src/ |
H A D | Common.h | 187 const vec4& row4) :
in mat4() 191 _41(row4.x), _42(row4.y), _43(row4.z), _44(row4.w)
in mat4()
|
/third_party/ffmpeg/libavcodec/ |
H A D | ivi_dsp.c | 670 int i, row2, row4, row8; in ff_ivi_col_slant8() local 674 row4 = pitch << 2; in ff_ivi_col_slant8() 681 out[0], out[pitch], out[row2], out[row2 + pitch], out[row4], in ff_ivi_col_slant8() 682 out[row4 + pitch], out[row4 + row2], out[row8 - pitch], in ff_ivi_col_slant8() 685 out[0] = out[pitch] = out[row2] = out[row2 + pitch] = out[row4] = in ff_ivi_col_slant8() 686 out[row4 + pitch] = out[row4 + row2] = out[row8 - pitch] = 0; in ff_ivi_col_slant8()
|
/third_party/skia/third_party/externals/libjpeg-turbo/simd/arm/aarch64/ |
H A D | jchuff-neon.c | 129 int16x8_t row4 = in jsimd_huff_encode_one_block_neon() local 165 int16x8_t abs_row4 = vabsq_s16(row4); in jsimd_huff_encode_one_block_neon() 180 vreinterpretq_u16_s16(veorq_s16(abs_row4, vshrq_n_s16(row4, 15))); in jsimd_huff_encode_one_block_neon()
|
/third_party/astc-encoder/Source/ |
H A D | stb_image.h | 2531 __m128i row0, row1, row2, row3, row4, row5, row6, row7; in stbi__idct_simd() local 2589 __m128i sum04 = _mm_add_epi16(row0, row4); \ in stbi__idct_simd() 2590 __m128i dif04 = _mm_sub_epi16(row0, row4); \ in stbi__idct_simd() 2610 dct_bfly32o(row3,row4, x3,x4,bias,shift); \ in stbi__idct_simd() 2631 row4 = _mm_load_si128((const __m128i *) (data + 4*8)); in stbi__idct_simd() 2641 dct_interleave16(row0, row4); in stbi__idct_simd() 2649 dct_interleave16(row4, row6); in stbi__idct_simd() 2655 dct_interleave16(row4, row5); in stbi__idct_simd() 2666 __m128i p2 = _mm_packus_epi16(row4, row5); in stbi__idct_simd() 2711 int16x8_t row0, row1, row2, row3, row4, row in stbi__idct_simd() local [all...] |