/third_party/ffmpeg/tests/checkasm/ |
H A D | fixed_dsp.c | 35 src1[i] = sign_extend(rnd(), 24); \ 40 static void check_vector_fmul(const int *src0, const int *src1) in check_vector_fmul() argument 45 declare_func(void, int *dst, const int *src0, const int *src1, int len); in check_vector_fmul() 47 call_ref(ref, src0, src1, BUF_SIZE); in check_vector_fmul() 48 call_new(new, src0, src1, BUF_SIZE); in check_vector_fmul() 51 bench_new(new, src0, src1, BUF_SIZE); in check_vector_fmul() 54 static void check_vector_fmul_add(const int *src0, const int *src1, const int *src2) in check_vector_fmul_add() argument 59 declare_func(void, int *dst, const int *src0, const int *src1, const int *src2, int len); in check_vector_fmul_add() 61 call_ref(ref, src0, src1, src2, BUF_SIZE); in check_vector_fmul_add() 62 call_new(new, src0, src1, src in check_vector_fmul_add() 68 check_vector_fmul_window(const int32_t *src0, const int32_t *src1, const int32_t *win) check_vector_fmul_window() argument 82 check_vector_fmul_window_scaled(const int32_t *src0, const int32_t *src1, const int32_t *win) check_vector_fmul_window_scaled() argument 96 check_butterflies(const int *src0, const int *src1) check_butterflies() argument 120 check_scalarproduct_fixed(const int *src0, const int *src1) check_scalarproduct_fixed() argument [all...] |
H A D | float_dsp.c | 44 static void test_vector_fmul(const float *src0, const float *src1) in test_vector_fmul() argument 50 declare_func(void, float *dst, const float *src0, const float *src1, in test_vector_fmul() 53 call_ref(cdst, src0, src1, LEN); in test_vector_fmul() 54 call_new(odst, src0, src1, LEN); in test_vector_fmul() 56 double t = fabs(src0[i]) + fabs(src1[i]) + fabs(src0[i] * src1[i]) + 1.0; in test_vector_fmul() 64 bench_new(odst, src0, src1, LEN); in test_vector_fmul() 67 static void test_vector_dmul(const double *src0, const double *src1) in test_vector_dmul() argument 73 declare_func(void, double *dst, const double *src0, const double *src1, in test_vector_dmul() 76 call_ref(cdst, src0, src1, LE in test_vector_dmul() 91 test_vector_fmul_add(const float *src0, const float *src1, const float *src2) test_vector_fmul_add() argument 113 test_vector_fmul_scalar(const float *src0, const float *src1) test_vector_fmul_scalar() argument 136 test_vector_fmul_window(const float *src0, const float *src1, const float *win) test_vector_fmul_window() argument 159 test_vector_fmac_scalar(const float *src0, const float *src1, const float *src2) test_vector_fmac_scalar() argument 184 test_vector_dmul_scalar(const double *src0, const double *src1) test_vector_dmul_scalar() argument 207 test_vector_dmac_scalar(const double *src0, const double *src1, const double *src2) test_vector_dmac_scalar() argument 231 test_butterflies_float(const float *src0, const float *src1) test_butterflies_float() argument 266 test_scalarproduct_float(const float *src0, const float *src1) test_scalarproduct_float() argument [all...] |
H A D | llviddsp.c | 50 uint8_t *src1 = av_calloc(width, sizeof(*src1)); in check_add_bytes() local 53 init_buffer(src0, src1, uint8_t, width); in check_add_bytes() 61 call_new(dst1, src1, width); in check_add_bytes() 64 bench_new(dst1, src1, width); in check_add_bytes() 68 av_free(src1); in check_add_bytes() 78 uint8_t *src1 = av_calloc(width, sizeof(*src1)); in check_add_median_pred() local 81 declare_func_emms(AV_CPU_FLAG_MMX, void, uint8_t *dst, const uint8_t *src1, in check_add_median_pred() 85 init_buffer(src0, src1, uint8_ in check_add_median_pred() 116 uint8_t *src1 = av_calloc(width, sizeof(*src1)); check_add_left_pred() local 145 uint16_t *src1 = av_calloc(width, sizeof(*src1)); check_add_left_pred_16() local 170 uint8_t *src0, *src1; check_add_gradient_pred() local [all...] |
/third_party/ltp/tools/sparse/sparse-src/ |
H A D | simplify.c | 352 kill_use(&insn->src1); in kill_insn() 498 replace_pseudo(insn, &insn->src1, src); in replace_with_unop() 546 pseudo_t old_a = in->src1; in replace_insn_pair() 548 pseudo_t old_1 = out->src1; in replace_insn_pair() 551 use_pseudo(in, a, &in->src1); in replace_insn_pair() 553 use_pseudo(out, in->target, &out->src1); in replace_insn_pair() 556 remove_usage(old_a, &in->src1); in replace_insn_pair() 558 remove_usage(old_1, &out->src1); in replace_insn_pair() 629 static pseudo_t eval_op(int op, unsigned size, pseudo_t src1, pseudo_t src2) in eval_op() argument 632 long long left = src1 in eval_op() 818 pseudo_t src1 = or->src1; simplify_mask_or() local 1175 pseudo_t src1, src2; simplify_compare_constant() local 1593 pseudo_t src1 = insn->src1; simplify_const_leftsub() local 1831 pseudo_t src1 = *p1; simplify_add_one_side() local 1874 pseudo_t src1 = insn->src1; simplify_sub() local 1904 pseudo_t src1 = insn->src1; simplify_compare() local 1929 pseudo_t src1 = *p1; simplify_and_one_side() local 1994 pseudo_t src1 = *p1; simplify_ior_one_side() local 2048 pseudo_t src1 = *p1; simplify_xor_one_side() local 2428 pseudo_t cond, src1, src2; simplify_select() local 2575 pseudo_t src1, src2, src3; simplify_range() local [all...] |
/third_party/mesa3d/src/imagination/vulkan/pds/ |
H A D | pvr_pds_printer.c | 70 char src1[32]; in pvr_pds_disassemble_instruction_add64() local 73 pvr_pds_disassemble_operand(add->src1, src1, sizeof(src1)); in pvr_pds_disassemble_instruction_add64() 84 src1, in pvr_pds_disassemble_instruction_add64() 94 char src1[32]; in pvr_pds_disassemble_instruction_add32() local 97 pvr_pds_disassemble_operand(add->src1, src1, sizeof(src1)); in pvr_pds_disassemble_instruction_add32() 108 src1, in pvr_pds_disassemble_instruction_add32() 119 char src1[32]; pvr_pds_disassemble_instruction_sftlp32() local 171 char src1[32]; pvr_pds_disassemble_instruction_stm() local 226 char src1[32]; pvr_pds_disassemble_instruction_sftlp64() local 278 char src1[32]; pvr_pds_disassemble_instruction_cmp() local 386 char src1[PVR_PDS_MAX_INST_STR_LEN]; pvr_pds_disassemble_instruction_ddmad() local 487 char src1[PVR_PDS_MAX_INST_STR_LEN]; pvr_pds_disassemble_instruction_mad() local 514 char src1[PVR_PDS_MAX_INST_STR_LEN]; pvr_pds_disassemble_instruction_dout() local [all...] |
/kernel/linux/linux-5.10/drivers/crypto/caam/ |
H A D | desc_constr.h | 415 #define append_math_add(desc, dest, src0, src1, len) \ 416 APPEND_MATH(ADD, desc, dest, src0, src1, len) 417 #define append_math_sub(desc, dest, src0, src1, len) \ 418 APPEND_MATH(SUB, desc, dest, src0, src1, len) 419 #define append_math_add_c(desc, dest, src0, src1, len) \ 420 APPEND_MATH(ADDC, desc, dest, src0, src1, len) 421 #define append_math_sub_b(desc, dest, src0, src1, len) \ 422 APPEND_MATH(SUBB, desc, dest, src0, src1, len) 423 #define append_math_and(desc, dest, src0, src1, len) \ 424 APPEND_MATH(AND, desc, dest, src0, src1, le [all...] |
/kernel/linux/linux-6.6/drivers/crypto/caam/ |
H A D | desc_constr.h | 416 #define append_math_add(desc, dest, src0, src1, len) \ 417 APPEND_MATH(ADD, desc, dest, src0, src1, len) 418 #define append_math_sub(desc, dest, src0, src1, len) \ 419 APPEND_MATH(SUB, desc, dest, src0, src1, len) 420 #define append_math_add_c(desc, dest, src0, src1, len) \ 421 APPEND_MATH(ADDC, desc, dest, src0, src1, len) 422 #define append_math_sub_b(desc, dest, src0, src1, len) \ 423 APPEND_MATH(SUBB, desc, dest, src0, src1, len) 424 #define append_math_and(desc, dest, src0, src1, len) \ 425 APPEND_MATH(AND, desc, dest, src0, src1, le [all...] |
/third_party/pcre2/pcre2/src/sljit/ |
H A D | sljitNativePPC_64.c | 132 FAIL_IF(push_inst(compiler, EXTSW | S(src1) | A(TMP_REG1))); \ 133 src1 = TMP_REG1; \ 143 FAIL_IF(push_inst(compiler, EXTSW | S(src1) | A(TMP_REG1))); \ 144 src1 = TMP_REG1; \ 148 sljit_s32 dst, sljit_s32 src1, sljit_s32 src2) in emit_single_op() 155 SLJIT_ASSERT(src1 == TMP_REG1); in emit_single_op() 162 SLJIT_ASSERT(src1 == TMP_REG1); in emit_single_op() 175 SLJIT_ASSERT(src1 == TMP_REG1); in emit_single_op() 190 SLJIT_ASSERT(src1 == TMP_REG1); in emit_single_op() 202 SLJIT_ASSERT(src1 in emit_single_op() 147 emit_single_op(struct sljit_compiler *compiler, sljit_s32 op, sljit_s32 flags, sljit_s32 dst, sljit_s32 src1, sljit_s32 src2) emit_single_op() argument [all...] |
H A D | sljitNativePPC_32.c | 46 sljit_s32 dst, sljit_s32 src1, sljit_s32 src2) in emit_single_op() 55 SLJIT_ASSERT(src1 == TMP_REG1); in emit_single_op() 62 SLJIT_ASSERT(src1 == TMP_REG1); in emit_single_op() 77 SLJIT_ASSERT(src1 == TMP_REG1); in emit_single_op() 89 SLJIT_ASSERT(src1 == TMP_REG1); in emit_single_op() 93 SLJIT_ASSERT(src1 == TMP_REG1); in emit_single_op() 97 SLJIT_ASSERT(src1 == TMP_REG1); in emit_single_op() 109 return push_inst(compiler, ADD | OE(ALT_SET_FLAGS) | RC(ALT_SET_FLAGS) | D(dst) | A(src1) | B(src2)); in emit_single_op() 117 return push_inst(compiler, ADDIS | D(dst) | A(src1) | compiler->imm); in emit_single_op() 122 FAIL_IF(push_inst(compiler, ADDIS | D(dst) | A(src1) | (((im in emit_single_op() 45 emit_single_op(struct sljit_compiler *compiler, sljit_s32 op, sljit_s32 flags, sljit_s32 dst, sljit_s32 src1, sljit_s32 src2) emit_single_op() argument [all...] |
/third_party/skia/third_party/externals/swiftshader/src/Shader/ |
H A D | ShaderCore.hpp | 247 void add(Vector4f &dst, const Vector4f &src0, const Vector4f &src1); 248 void iadd(Vector4f &dst, const Vector4f &src0, const Vector4f &src1); 249 void sub(Vector4f &dst, const Vector4f &src0, const Vector4f &src1); 250 void isub(Vector4f &dst, const Vector4f &src0, const Vector4f &src1); 251 void mad(Vector4f &dst, const Vector4f &src0, const Vector4f &src1, const Vector4f &src2); 252 void imad(Vector4f &dst, const Vector4f &src0, const Vector4f &src1, const Vector4f &src2); 253 void mul(Vector4f &dst, const Vector4f &src0, const Vector4f &src1); 254 void imul(Vector4f &dst, const Vector4f &src0, const Vector4f &src1); 256 void div(Vector4f &dst, const Vector4f &src0, const Vector4f &src1); 257 void idiv(Vector4f &dst, const Vector4f &src0, const Vector4f &src1); [all...] |
H A D | ShaderCore.cpp | 766 void ShaderCore::add(Vector4f &dst, const Vector4f &src0, const Vector4f &src1) in add() argument 768 dst.x = src0.x + src1.x; in add() 769 dst.y = src0.y + src1.y; in add() 770 dst.z = src0.z + src1.z; in add() 771 dst.w = src0.w + src1.w; in add() 774 void ShaderCore::iadd(Vector4f &dst, const Vector4f &src0, const Vector4f &src1) in iadd() argument 776 dst.x = As<Float4>(As<Int4>(src0.x) + As<Int4>(src1.x)); in iadd() 777 dst.y = As<Float4>(As<Int4>(src0.y) + As<Int4>(src1.y)); in iadd() 778 dst.z = As<Float4>(As<Int4>(src0.z) + As<Int4>(src1.z)); in iadd() 779 dst.w = As<Float4>(As<Int4>(src0.w) + As<Int4>(src1 in iadd() 782 sub(Vector4f &dst, const Vector4f &src0, const Vector4f &src1) sub() argument 790 isub(Vector4f &dst, const Vector4f &src0, const Vector4f &src1) isub() argument 798 mad(Vector4f &dst, const Vector4f &src0, const Vector4f &src1, const Vector4f &src2) mad() argument 806 imad(Vector4f &dst, const Vector4f &src0, const Vector4f &src1, const Vector4f &src2) imad() argument 814 mul(Vector4f &dst, const Vector4f &src0, const Vector4f &src1) mul() argument 822 imul(Vector4f &dst, const Vector4f &src0, const Vector4f &src1) imul() argument 840 div(Vector4f &dst, const Vector4f &src0, const Vector4f &src1) div() argument 848 idiv(Vector4f &dst, const Vector4f &src0, const Vector4f &src1) idiv() argument 861 udiv(Vector4f &dst, const Vector4f &src0, const Vector4f &src1) udiv() argument 874 mod(Vector4f &dst, const Vector4f &src0, const Vector4f &src1) mod() argument 882 imod(Vector4f &dst, const Vector4f &src0, const Vector4f &src1) imod() argument 895 umod(Vector4f &dst, const Vector4f &src0, const Vector4f &src1) umod() argument 908 shl(Vector4f &dst, const Vector4f &src0, const Vector4f &src1) shl() argument 916 ishr(Vector4f &dst, const Vector4f &src0, const Vector4f &src1) ishr() argument 924 ushr(Vector4f &dst, const Vector4f &src0, const Vector4f &src1) ushr() argument 973 dist1(Float4 &dst, const Vector4f &src0, const Vector4f &src1, bool pp) dist1() argument 978 dist2(Float4 &dst, const Vector4f &src0, const Vector4f &src1, bool pp) dist2() argument 986 dist3(Float4 &dst, const Vector4f &src0, const Vector4f &src1, bool pp) dist3() argument 995 dist4(Float4 &dst, const Vector4f &src0, const Vector4f &src1, bool pp) dist4() argument 1005 dp1(Vector4f &dst, const Vector4f &src0, const Vector4f &src1) dp1() argument 1015 dp2(Vector4f &dst, const Vector4f &src0, const Vector4f &src1) dp2() argument 1025 dp2add(Vector4f &dst, const Vector4f &src0, const Vector4f &src1, const Vector4f &src2) dp2add() argument 1035 dp3(Vector4f &dst, const Vector4f &src0, const Vector4f &src1) dp3() argument 1045 dp4(Vector4f &dst, const Vector4f &src0, const Vector4f &src1) dp4() argument 1055 min(Vector4f &dst, const Vector4f &src0, const Vector4f &src1) min() argument 1063 imin(Vector4f &dst, const Vector4f &src0, const Vector4f &src1) imin() argument 1071 umin(Vector4f &dst, const Vector4f &src0, const Vector4f &src1) umin() argument 1079 max(Vector4f &dst, const Vector4f &src0, const Vector4f &src1) max() argument 1087 imax(Vector4f &dst, const Vector4f &src0, const Vector4f &src1) imax() argument 1095 umax(Vector4f &dst, const Vector4f &src0, const Vector4f &src1) umax() argument 1103 slt(Vector4f &dst, const Vector4f &src0, const Vector4f &src1) slt() argument 1189 att(Vector4f &dst, const Vector4f &src0, const Vector4f &src1) att() argument 1198 lrp(Vector4f &dst, const Vector4f &src0, const Vector4f &src1, const Vector4f &src2) lrp() argument 1309 det2(Vector4f &dst, const Vector4f &src0, const Vector4f &src1) det2() argument 1315 det3(Vector4f &dst, const Vector4f &src0, const Vector4f &src1, const Vector4f &src2) det3() argument 1321 det4(Vector4f &dst, const Vector4f &src0, const Vector4f &src1, const Vector4f &src2, const Vector4f &src3) det4() argument 1396 powx(Vector4f &dst, const Vector4f &src0, const Vector4f &src1, bool pp) powx() argument 1406 pow(Vector4f &dst, const Vector4f &src0, const Vector4f &src1, bool pp) pow() argument 1414 crs(Vector4f &dst, const Vector4f &src0, const Vector4f &src1) crs() argument 1654 atan2(Vector4f &dst, const Vector4f &src0, const Vector4f &src1, bool pp) atan2() argument 1759 cmp0(Vector4f &dst, const Vector4f &src0, const Vector4f &src1, const Vector4f &src2) cmp0() argument 1767 select(Vector4f &dst, const Vector4f &src0, const Vector4f &src1, const Vector4f &src2) select() argument 1775 extract(Float4 &dst, const Vector4f &src0, const Float4 &src1) extract() argument 1804 cmp0(Float4 &dst, const Float4 &src0, const Float4 &src1, const Float4 &src2) cmp0() argument 1810 cmp0i(Float4 &dst, const Float4 &src0, const Float4 &src1, const Float4 &src2) cmp0i() argument 1816 select(Float4 &dst, RValue<Int4> src0, const Float4 &src1, const Float4 &src2) select() argument 1822 cmp(Vector4f &dst, const Vector4f &src0, const Vector4f &src1, Control control) cmp() argument 1867 icmp(Vector4f &dst, const Vector4f &src0, const Vector4f &src1, Control control) icmp() argument 1912 ucmp(Vector4f &dst, const Vector4f &src0, const Vector4f &src1, Control control) ucmp() argument 1975 bitwise_or(Vector4f &dst, const Vector4f &src0, const Vector4f &src1) bitwise_or() argument 1983 bitwise_xor(Vector4f &dst, const Vector4f &src0, const Vector4f &src1) bitwise_xor() argument 1991 bitwise_and(Vector4f &dst, const Vector4f &src0, const Vector4f &src1) bitwise_and() argument 1999 equal(Vector4f &dst, const Vector4f &src0, const Vector4f &src1) equal() argument 2010 notEqual(Vector4f &dst, const Vector4f &src0, const Vector4f &src1) notEqual() argument [all...] |
/third_party/ffmpeg/libavcodec/loongarch/ |
H A D | hpeldsp_lasx.c | 26 put_pixels8_l2_8_lsx(uint8_t *dst, const uint8_t *src1, const uint8_t *src2, in put_pixels8_l2_8_lsx() argument 39 "vld $vr0, %[src1], 0 \n\t" in put_pixels8_l2_8_lsx() 40 "vldx $vr1, %[src1], %[srcStride1] \n\t" in put_pixels8_l2_8_lsx() 41 "vldx $vr2, %[src1], %[stride1_2] \n\t" in put_pixels8_l2_8_lsx() 42 "vldx $vr3, %[src1], %[stride1_3] \n\t" in put_pixels8_l2_8_lsx() 43 "add.d %[src1], %[src1], %[stride1_4] \n\t" in put_pixels8_l2_8_lsx() 67 : [dst]"+&r"(dst), [src2]"+&r"(src2), [src1]"+&r"(src1), in put_pixels8_l2_8_lsx() 79 put_pixels16_l2_8_lsx(uint8_t *dst, const uint8_t *src1, cons argument 284 __m256i src0, src1, src2, src3, src4, src5, src6, src7; common_hz_bil_no_rnd_16x16_lasx() local 393 __m256i src0, src1, src2, src3, src4, src5, src6, src7; common_hz_bil_no_rnd_8x16_lasx() local 462 __m256i src0, src1, src2, src3, src4, src5, src6, src7, src8; common_vt_bil_no_rnd_16x16_lasx() local 554 __m256i src0, src1, src2, src3, src4, src5, src6, src7, src8; common_vt_bil_no_rnd_8x16_lasx() local 616 __m256i src0, src1, src2, src3, src4, src5, src6, src7, src8, src9; common_hv_bil_no_rnd_16x16_lasx() local 776 __m256i src0, src1, src2, src3, src4, src5, src6, src7, src8, src9; common_hv_bil_no_rnd_8x16_lasx() local 873 __m256i src0, src1, src2, src3, src4, src5, src6, src7; common_hz_bil_no_rnd_8x8_lasx() local 923 __m256i src0, src1, src2, src3, src4, src5, src6, src7; common_hz_bil_no_rnd_4x8_lasx() local 960 __m256i src0, src1, src2, src3, src4, src5, src6, src7, src8; common_vt_bil_no_rnd_8x8_lasx() local 1001 __m256i src0, src1, src2, src3, src4; common_vt_bil_no_rnd_4x8_lasx() local 1035 __m256i src0, src1, src2, src3, src4, src5, src6, src7; common_hv_bil_no_rnd_8x8_lasx() local 1098 __m256i src0, src1, src2, src3, src4, src5, src6, src7; common_hv_bil_no_rnd_4x8_lasx() local 1148 __m256i src0, src1, src2, src3, src4, src5, src6, src7, src8, src9; common_hv_bil_16w_lasx() local 1241 __m256i src0, src1, src2, src3, src4, src5, src6, src7; common_hv_bil_8w_lasx() local [all...] |
/kernel/linux/linux-5.10/arch/powerpc/include/asm/ |
H A D | kvm_fpu.h | 14 extern void fps_fres(u64 *fpscr, u32 *dst, u32 *src1); 15 extern void fps_frsqrte(u64 *fpscr, u32 *dst, u32 *src1); 16 extern void fps_fsqrts(u64 *fpscr, u32 *dst, u32 *src1); 18 extern void fps_fadds(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2); 19 extern void fps_fdivs(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2); 20 extern void fps_fmuls(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2); 21 extern void fps_fsubs(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2); 23 extern void fps_fmadds(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2, 25 extern void fps_fmsubs(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2, 27 extern void fps_fnmadds(u64 *fpscr, u32 *dst, u32 *src1, u3 [all...] |
/kernel/linux/linux-6.6/arch/powerpc/include/asm/ |
H A D | kvm_fpu.h | 14 extern void fps_fres(u64 *fpscr, u32 *dst, u32 *src1); 15 extern void fps_frsqrte(u64 *fpscr, u32 *dst, u32 *src1); 16 extern void fps_fsqrts(u64 *fpscr, u32 *dst, u32 *src1); 18 extern void fps_fadds(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2); 19 extern void fps_fdivs(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2); 20 extern void fps_fmuls(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2); 21 extern void fps_fsubs(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2); 23 extern void fps_fmadds(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2, 25 extern void fps_fmsubs(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2, 27 extern void fps_fnmadds(u64 *fpscr, u32 *dst, u32 *src1, u3 [all...] |
/third_party/ffmpeg/libavcodec/mips/ |
H A D | simple_idct_mmi.c | 64 #define IDCT_ROW_COND_DC(src1, src2) \ in ff_simple_idct_8_mmi() 65 "dmfc1 $11, "#src1" \n\t" \ in ff_simple_idct_8_mmi() 71 "punpcklhw $f30, "#src1", "#src2" \n\t" \ in ff_simple_idct_8_mmi() 72 "punpckhhw $f31, "#src1", "#src2" \n\t" \ in ff_simple_idct_8_mmi() 74 "punpcklhw "#src1", $f30, $f31 \n\t" \ in ff_simple_idct_8_mmi() 78 "pmaddhw $f30, "#src1", $f18 \n\t" \ in ff_simple_idct_8_mmi() 89 "pmaddhw $f30, "#src1", $f20 \n\t" \ in ff_simple_idct_8_mmi() 100 "pmaddhw $f30, "#src1", $f22 \n\t" \ in ff_simple_idct_8_mmi() 111 "pmaddhw $f30, "#src1", $f24 \n\t" \ in ff_simple_idct_8_mmi() 113 "paddw "#src1", in ff_simple_idct_8_mmi() [all...] |
H A D | hpeldsp_msa.c | 61 v16u8 src0, src1, src0_sld1, src1_sld1, res0, res1; in common_hz_bil_4w_msa() local 65 LD_UB2(src, src_stride, src0, src1); in common_hz_bil_4w_msa() 68 SLDI_B2_UB(zeros, src0, zeros, src1, 1, src0_sld1, src1_sld1); in common_hz_bil_4w_msa() 69 AVER_UB2_UB(src0_sld1, src0, src1_sld1, src1, res0, res1); in common_hz_bil_4w_msa() 85 v16i8 src0, src1, src2, src3, src0_sld1, src1_sld1, src2_sld1, src3_sld1; in common_hz_bil_8w_msa() local 89 LD_SB4(src, src_stride, src0, src1, src2, src3); in common_hz_bil_8w_msa() 92 SLDI_B4_SB(zeros, src0, zeros, src1, zeros, src2, zeros, src3, 1, in common_hz_bil_8w_msa() 94 AVER_ST8x4_UB(src0, src0_sld1, src1, src1_sld1, in common_hz_bil_8w_msa() 105 v16u8 src0, src1, src2, src3, src4, src5, src6, src7; in common_hz_bil_16w_msa() local 109 LD_UB8(src, src_stride, src0, src1, src in common_hz_bil_16w_msa() 127 v16i8 src0, src1, src2, src3, src4, src5, src6, src7; common_hz_bil_no_rnd_8x8_msa() local 150 v16i8 src0, src1, src2, src3, src0_sld1, src1_sld1, src2_sld1, src3_sld1; common_hz_bil_no_rnd_4x8_msa() local 164 v16u8 src0, src1, src2, src3, src4, src5, src6, src7, src8; common_hz_bil_no_rnd_16x16_msa() local 199 v16u8 src0, src1, src2, src3, src4, src5, src6, src7, src8; common_hz_bil_no_rnd_8x16_msa() local 220 v16u8 src0, src1, src0_sld1, src1_sld1, res0, res1; common_hz_bil_and_aver_dst_4w_msa() local 254 v16i8 src0, src1, src2, src3, src0_sld1, src1_sld1, src2_sld1, src3_sld1; common_hz_bil_and_aver_dst_8w_msa() local 276 v16u8 src0, src1, src2, src3, src4, src5, src6, src7, src8; common_hz_bil_and_aver_dst_16w_msa() local 300 v16u8 src0, src1, src2, res0, res1; common_vt_bil_4w_msa() local 327 v16u8 src0, src1, src2, src3, src4; common_vt_bil_8w_msa() local 349 v16u8 src0, src1, src2, src3, src4, src5, src6, src7, src8; common_vt_bil_16w_msa() local 372 v16u8 src0, src1, src2, src3, src4, src5, src6, src7, src8; common_vt_bil_no_rnd_8x8_msa() local 389 v16u8 src0, src1, src2, src3, src4; common_vt_bil_no_rnd_4x8_msa() local 400 v16u8 src0, src1, src2, src3, src4, src5, src6, src7, src8; common_vt_bil_no_rnd_16x16_msa() local 427 v16u8 src0, src1, src2, src3, src4, src5, src6, src7, src8; common_vt_bil_no_rnd_8x16_msa() local 447 v16u8 src0, src1, src2; common_vt_bil_and_aver_dst_4w_msa() local 480 v16u8 src0, src1, src2, src3, src4; common_vt_bil_and_aver_dst_8w_msa() local 502 v16u8 src0, src1, src2, src3, src4, src5, src6, src7, src8; common_vt_bil_and_aver_dst_16w_msa() local 535 v16i8 src0, src1, src2, src0_sld1, src1_sld1, src2_sld1; common_hv_bil_4w_msa() local 571 v16i8 src0, src1, src2, src3, src4; common_hv_bil_8w_msa() local 608 v16u8 src0, src1, src2, src3, src4, src5, src6, src7, src8, src9; common_hv_bil_16w_msa() local 664 v16u8 src0, src1, src2, src3, src4, src5, src6, src7, src8; common_hv_bil_no_rnd_8x8_msa() local 712 v16i8 src0, src1, src2, src3, src4; common_hv_bil_no_rnd_4x8_msa() local 747 v16u8 src0, src1, src2, src3, src4, src5, src6, src7, src8, src9; common_hv_bil_no_rnd_16x16_msa() local 864 v16u8 src0, src1, src2, src3, src4, src5, src6, src7, src8, src9; common_hv_bil_no_rnd_8x16_msa() local 931 v16i8 src0, src1, src2, src0_sld1, src1_sld1, src2_sld1; common_hv_bil_and_aver_dst_4w_msa() local 972 v16i8 src0, src1, src2, src3, src4; common_hv_bil_and_aver_dst_8w_msa() local 1012 v16u8 src0, src1, src2, src3, src4, src5, src6, src7, src8, src9, src10; common_hv_bil_and_aver_dst_16w_msa() local 1084 v16u8 src0, src1, src2, src3, src4, src5, src6, src7; copy_width8_msa() local 1171 v16u8 src0, src1, src2, src3, src4, src5, src6, src7; copy_16multx8mult_msa() local 1197 v16u8 src0, src1, src2, src3, src4, src5, src6, src7; copy_width16_msa() local 1232 v16u8 src0, src1, src2, src3; avg_width4_msa() local 1277 v16u8 src0, src1, src2, src3; avg_width8_msa() local 1302 v16u8 src0, src1, src2, src3, src4, src5, src6, src7; avg_width16_msa() local [all...] |
/third_party/ffmpeg/libavcodec/ |
H A D | lossless_videoencdsp.c | 28 static void diff_bytes_c(uint8_t *dst, const uint8_t *src1, const uint8_t *src2, intptr_t w) in diff_bytes_c() argument 33 if (((long)src1 | (long)src2) & (sizeof(long) - 1)) { in diff_bytes_c() 35 dst[i + 0] = src1[i + 0] - src2[i + 0]; in diff_bytes_c() 36 dst[i + 1] = src1[i + 1] - src2[i + 1]; in diff_bytes_c() 37 dst[i + 2] = src1[i + 2] - src2[i + 2]; in diff_bytes_c() 38 dst[i + 3] = src1[i + 3] - src2[i + 3]; in diff_bytes_c() 39 dst[i + 4] = src1[i + 4] - src2[i + 4]; in diff_bytes_c() 40 dst[i + 5] = src1[i + 5] - src2[i + 5]; in diff_bytes_c() 41 dst[i + 6] = src1[i + 6] - src2[i + 6]; in diff_bytes_c() 42 dst[i + 7] = src1[ in diff_bytes_c() 56 sub_median_pred_c(uint8_t *dst, const uint8_t *src1, const uint8_t *src2, intptr_t w, int *left, int *left_top) sub_median_pred_c() argument [all...] |
H A D | jpeg2000dsp.c | 45 float *src0 = _src0, *src1 = _src1, *src2 = _src2; in ict_float() local 51 i1f = *src0 - (f_ict_params[1] * *src1) in ict_float() 53 i2f = *src0 + (f_ict_params[3] * *src1); in ict_float() 55 *src1++ = i1f; in ict_float() 62 int32_t *src0 = _src0, *src1 = _src1, *src2 = _src2; in ict_int() local 68 i1 = *src0 - ((int)(((unsigned)i_ict_params[1] * *src1) + (1 << 15)) >> 16) in ict_int() 70 i2 = *src0 + (2 * *src1) + ((int)((-14942U * *src1) + (1 << 15)) >> 16); in ict_int() 72 *src1++ = i1; in ict_int() 79 int32_t *src0 = _src0, *src1 in rct_int() local [all...] |
/third_party/skia/third_party/externals/angle2/src/image_util/ |
H A D | imageformats.cpp | 31 void L8::average(L8 *dst, const L8 *src1, const L8 *src2) in average() argument 33 dst->L = gl::average(src1->L, src2->L); in average() 62 void R8::average(R8 *dst, const R8 *src1, const R8 *src2) in average() argument 64 dst->R = gl::average(src1->R, src2->R); in average() 80 void A8::average(A8 *dst, const A8 *src1, const A8 *src2) in average() argument 82 dst->A = gl::average(src1->A, src2->A); in average() 100 void L8A8::average(L8A8 *dst, const L8A8 *src1, const L8A8 *src2) in average() argument 102 *(uint16_t *)dst = (((*(uint16_t *)src1 ^ *(uint16_t *)src2) & 0xFEFE) >> 1) + in average() 103 (*(uint16_t *)src1 & *(uint16_t *)src2); in average() 121 void A8L8::average(A8L8 *dst, const A8L8 *src1, cons argument 155 average(R8G8 *dst, const R8G8 *src1, const R8G8 *src2) average() argument 191 average(R8G8B8 *dst, const R8G8B8 *src1, const R8G8B8 *src2) average() argument 228 average(B8G8R8 *dst, const B8G8R8 *src1, const B8G8R8 *src2) average() argument 250 average(R5G6B5 *dst, const R5G6B5 *src1, const R5G6B5 *src2) average() argument 275 average(B5G6R5 *dst, const B5G6R5 *src1, const B5G6R5 *src2) average() argument 317 average(A8R8G8B8 *dst, const A8R8G8B8 *src1, const A8R8G8B8 *src2) average() argument 355 average(R8G8B8A8 *dst, const R8G8B8A8 *src1, const R8G8B8A8 *src2) average() argument 377 average(R8G8B8A8SRGB *dst, const R8G8B8A8SRGB *src1, const R8G8B8A8SRGB *src2) average() argument 427 average(B8G8R8A8 *dst, const B8G8R8A8 *src1, const B8G8R8A8 *src2) average() argument 465 average(B8G8R8X8 *dst, const B8G8R8X8 *src1, const B8G8R8X8 *src2) average() argument 504 average(R8G8B8X8 *dst, const R8G8B8X8 *src1, const R8G8B8X8 *src2) average() argument 527 average(A1R5G5B5 *dst, const A1R5G5B5 *src1, const A1R5G5B5 *src2) average() argument 555 average(R5G5B5A1 *dst, const R5G5B5A1 *src1, const R5G5B5A1 *src2) average() argument 583 average(R4G4B4A4 *dst, const R4G4B4A4 *src1, const R4G4B4A4 *src2) average() argument 611 average(A4R4G4B4 *dst, const A4R4G4B4 *src1, const A4R4G4B4 *src2) average() argument 649 average(R16 *dst, const R16 *src1, const R16 *src2) average() argument 682 average(R16G16 *dst, const R16G16 *src1, const R16G16 *src2) average() argument 718 average(R16G16B16 *dst, const R16G16B16 *src1, const R16G16B16 *src2) average() argument 757 average(R16G16B16A16 *dst, const R16G16B16A16 *src1, const R16G16B16A16 *src2) average() argument 791 average(R32 *dst, const R32 *src1, const R32 *src2) average() argument 824 average(R32G32 *dst, const R32G32 *src1, const R32G32 *src2) average() argument 860 average(R32G32B32 *dst, const R32G32B32 *src1, const R32G32B32 *src2) average() argument 899 average(R32G32B32A32 *dst, const R32G32B32A32 *src1, const R32G32B32A32 *src2) average() argument 933 average(R8S *dst, const R8S *src1, const R8S *src2) average() argument 966 average(R8G8S *dst, const R8G8S *src1, const R8G8S *src2) average() argument 1002 average(R8G8B8S *dst, const R8G8B8S *src1, const R8G8B8S *src2) average() argument 1041 average(R8G8B8A8S *dst, const R8G8B8A8S *src1, const R8G8B8A8S *src2) average() argument 1075 average(R16S *dst, const R16S *src1, const R16S *src2) average() argument 1108 average(R16G16S *dst, const R16G16S *src1, const R16G16S *src2) average() argument 1144 average(R16G16B16S *dst, const R16G16B16S *src1, const R16G16B16S *src2) average() argument 1183 average(R16G16B16A16S *dst, const R16G16B16A16S *src1, const R16G16B16A16S *src2) average() argument 1219 average(R32S *dst, const R32S *src1, const R32S *src2) average() argument 1252 average(R32G32S *dst, const R32G32S *src1, const R32G32S *src2) average() argument 1288 average(R32G32B32S *dst, const R32G32B32S *src1, const R32G32B32S *src2) average() argument 1327 average(R32G32B32A32S *dst, const R32G32B32A32S *src1, const R32G32B32A32S *src2) average() argument 1353 average(A16B16G16R16F *dst, const A16B16G16R16F *src1, const A16B16G16R16F *src2) average() argument 1379 average(R16G16B16A16F *dst, const R16G16B16A16F *src1, const R16G16B16A16F *src2) average() argument 1402 average(R16F *dst, const R16F *src1, const R16F *src2) average() argument 1420 average(A16F *dst, const A16F *src1, const A16F *src2) average() argument 1439 average(L16F *dst, const L16F *src1, const L16F *src2) average() argument 1459 average(L16A16F *dst, const L16A16F *src1, const L16A16F *src2) average() argument 1479 average(R16G16F *dst, const R16G16F *src1, const R16G16F *src2) average() argument 1500 average(R16G16B16F *dst, const R16G16B16F *src1, const R16G16B16F *src2) average() argument 1523 average(A32B32G32R32F *dst, const A32B32G32R32F *src1, const A32B32G32R32F *src2) average() argument 1549 average(R32G32B32A32F *dst, const R32G32B32A32F *src1, const R32G32B32A32F *src2) average() argument 1572 average(R32F *dst, const R32F *src1, const R32F *src2) average() argument 1590 average(A32F *dst, const A32F *src1, const A32F *src2) average() argument 1608 average(L32F *dst, const L32F *src1, const L32F *src2) average() argument 1627 average(L32A32F *dst, const L32A32F *src1, const L32A32F *src2) average() argument 1647 average(R32G32F *dst, const R32G32F *src1, const R32G32F *src2) average() argument 1668 average(R32G32B32F *dst, const R32G32B32F *src1, const R32G32B32F *src2) average() argument 1707 average(R10G10B10A2 *dst, const R10G10B10A2 *src1, const R10G10B10A2 *src2) average() argument 1747 average(R10G10B10A2S *dst, const R10G10B10A2S *src1, const R10G10B10A2S *src2) average() argument 1785 average(R10G10B10X2 *dst, const R10G10B10X2 *src1, const R10G10B10X2 *src2) average() argument 1824 average(B10G10R10A2 *dst, const B10G10R10A2 *src1, const B10G10R10A2 *src2) average() argument 1844 average(R9G9B9E5 *dst, const R9G9B9E5 *src1, const R9G9B9E5 *src2) average() argument 1871 average(R11G11B10F *dst, const R11G11B10F *src1, const R11G11B10F *src2) average() argument [all...] |
/kernel/linux/linux-6.6/lib/crypto/ |
H A D | utils.c | 13 * XOR @len bytes from @src1 and @src2 together, writing the result to @dst 17 void __crypto_xor(u8 *dst, const u8 *src1, const u8 *src2, unsigned int len) in __crypto_xor() argument 23 int d = (((unsigned long)dst ^ (unsigned long)src1) | in __crypto_xor() 36 *dst++ = *src1++ ^ *src2++; in __crypto_xor() 43 u64 l = get_unaligned((u64 *)src1) ^ in __crypto_xor() 47 *(u64 *)dst = *(u64 *)src1 ^ *(u64 *)src2; in __crypto_xor() 50 src1 += 8; in __crypto_xor() 57 u32 l = get_unaligned((u32 *)src1) ^ in __crypto_xor() 61 *(u32 *)dst = *(u32 *)src1 ^ *(u32 *)src2; in __crypto_xor() 64 src1 in __crypto_xor() [all...] |
/third_party/node/deps/v8/src/codegen/shared-ia32-x64/ |
H A D | macro-assembler-shared-ia32-x64.cc | 80 void SharedTurboAssembler::Movhps(XMMRegister dst, XMMRegister src1, in Movhps() argument 84 vmovhps(dst, src1, src2); in Movhps() 86 if (dst != src1) { in Movhps() 87 movaps(dst, src1); in Movhps() 93 void SharedTurboAssembler::Movlps(XMMRegister dst, XMMRegister src1, in Movlps() argument 97 vmovlps(dst, src1, src2); in Movlps() 99 if (dst != src1) { in Movlps() 100 movaps(dst, src1); in Movlps() 106 void SharedTurboAssembler::Pblendvb(XMMRegister dst, XMMRegister src1, in Pblendvb() argument 110 vpblendvb(dst, src1, src in Pblendvb() 119 Shufps(XMMRegister dst, XMMRegister src1, XMMRegister src2, uint8_t imm8) Shufps() argument 400 I8x16Shl(XMMRegister dst, XMMRegister src1, uint8_t src2, Register tmp1, XMMRegister tmp2) I8x16Shl() argument 422 I8x16Shl(XMMRegister dst, XMMRegister src1, Register src2, Register tmp1, XMMRegister tmp2, XMMRegister tmp3) I8x16Shl() argument 449 I8x16ShrS(XMMRegister dst, XMMRegister src1, uint8_t src2, XMMRegister tmp) I8x16ShrS() argument 463 I8x16ShrS(XMMRegister dst, XMMRegister src1, Register src2, Register tmp1, XMMRegister tmp2, XMMRegister tmp3) I8x16ShrS() argument 484 I8x16ShrU(XMMRegister dst, XMMRegister src1, uint8_t src2, Register tmp1, XMMRegister tmp2) I8x16ShrU() argument 506 I8x16ShrU(XMMRegister dst, XMMRegister src1, Register src2, Register tmp1, XMMRegister tmp2, XMMRegister tmp3) I8x16ShrU() argument 557 I16x8ExtMulLow(XMMRegister dst, XMMRegister src1, XMMRegister src2, XMMRegister scratch, bool is_signed) I16x8ExtMulLow() argument 566 I16x8ExtMulHighS(XMMRegister dst, XMMRegister src1, XMMRegister src2, XMMRegister scratch) I16x8ExtMulHighS() argument 590 I16x8ExtMulHighU(XMMRegister dst, XMMRegister src1, XMMRegister src2, XMMRegister scratch) I16x8ExtMulHighU() argument 689 I16x8Q15MulRSatS(XMMRegister dst, XMMRegister src1, XMMRegister src2, XMMRegister scratch) I16x8Q15MulRSatS() argument 751 I32x4ExtMul(XMMRegister dst, XMMRegister src1, XMMRegister src2, XMMRegister scratch, bool low, bool is_signed) I32x4ExtMul() argument 856 I64x2GtS(XMMRegister dst, XMMRegister src0, XMMRegister src1, XMMRegister scratch) I64x2GtS() argument 890 I64x2GeS(XMMRegister dst, XMMRegister src0, XMMRegister src1, XMMRegister scratch) I64x2GeS() argument 1038 I64x2ExtMul(XMMRegister dst, XMMRegister src1, XMMRegister src2, XMMRegister scratch, bool low, bool is_signed) I64x2ExtMul() argument 1121 S128Select(XMMRegister dst, XMMRegister mask, XMMRegister src1, XMMRegister src2, XMMRegister scratch) S128Select() argument 1281 F32x4Qfma(XMMRegister dst, XMMRegister src1, XMMRegister src2, XMMRegister src3, XMMRegister tmp) F32x4Qfma() argument 1287 F32x4Qfms(XMMRegister dst, XMMRegister src1, XMMRegister src2, XMMRegister src3, XMMRegister tmp) F32x4Qfms() argument 1293 F64x2Qfma(XMMRegister dst, XMMRegister src1, XMMRegister src2, XMMRegister src3, XMMRegister tmp) F64x2Qfma() argument 1299 F64x2Qfms(XMMRegister dst, XMMRegister src1, XMMRegister src2, XMMRegister src3, XMMRegister tmp) F64x2Qfms() argument [all...] |
/third_party/mesa3d/src/compiler/nir/tests/ |
H A D | ssa_def_bits_used_tests.cpp | 59 nir_ssa_def *src0, nir_ssa_def *src1) in build_alu_instr() 61 nir_ssa_def *def = nir_build_alu(&bld, op, src0, src1, NULL, NULL); in build_alu_instr() 84 nir_ssa_def *src1 = nir_imm_int(&bld, 0xffffffff); in TEST_F() local 86 nir_alu_instr *alu = build_alu_instr(nir_op_iand, src0, src1); in TEST_F() 91 /* If the test is changed, and somehow src1 is used multiple times, in TEST_F() 95 ASSERT_TRUE(is_used_once(src1)); in TEST_F() 113 nir_ssa_def *src1 = nir_imm_int(&bld, 0xffffffff); in TEST_F() local 115 nir_alu_instr *alu = build_alu_instr(nir_op_ior, src0, src1); in TEST_F() 120 /* If the test is changed, and somehow src1 is used multiple times, in TEST_F() 124 ASSERT_TRUE(is_used_once(src1)); in TEST_F() 58 build_alu_instr(nir_op op, nir_ssa_def *src0, nir_ssa_def *src1) build_alu_instr() argument 141 nir_ssa_def *src1 = nir_imm_ivec4(&bld, TEST_F() local 172 nir_ssa_def *src1 = nir_imm_ivec4(&bld, TEST_F() local 203 nir_ssa_def *src1 = nir_imm_ivec4(&bld, TEST_F() local 234 nir_ssa_def *src1 = nir_imm_ivec4(&bld, TEST_F() local [all...] |
/kernel/linux/linux-5.10/arch/ia64/lib/ |
H A D | copy_user.S | 66 #define src1 r24 define 103 mov src1=src // copy because of rotation 115 EX(.failure_in_pipe1,(p16) ld1 val1[0]=[src1],1) 130 and src2=0x7,src1 // src offset 134 // that we can reach 8-byte alignment for both src1 and dst1. 155 // We know src1 is not 8-byte aligned in this case. 197 EX(.failure_in_pipe2,(p16) ld1 val1[0]=[src1],1) 212 (p14) sub src1=src1,t2 213 (p15) sub src1 [all...] |
/kernel/linux/linux-6.6/arch/ia64/lib/ |
H A D | copy_user.S | 66 #define src1 r24 define 103 mov src1=src // copy because of rotation 115 EX(.failure_in_pipe1,(p16) ld1 val1[0]=[src1],1) 130 and src2=0x7,src1 // src offset 134 // that we can reach 8-byte alignment for both src1 and dst1. 155 // We know src1 is not 8-byte aligned in this case. 197 EX(.failure_in_pipe2,(p16) ld1 val1[0]=[src1],1) 212 (p14) sub src1=src1,t2 213 (p15) sub src1 [all...] |
/third_party/node/deps/v8/src/codegen/arm/ |
H A D | assembler-arm.h | 423 void and_(Register dst, Register src1, const Operand& src2, SBit s = LeaveCC, 425 void and_(Register dst, Register src1, Register src2, SBit s = LeaveCC, 428 void eor(Register dst, Register src1, const Operand& src2, SBit s = LeaveCC, 430 void eor(Register dst, Register src1, Register src2, SBit s = LeaveCC, 433 void sub(Register dst, Register src1, const Operand& src2, SBit s = LeaveCC, 435 void sub(Register dst, Register src1, Register src2, SBit s = LeaveCC, 438 void rsb(Register dst, Register src1, const Operand& src2, SBit s = LeaveCC, 441 void add(Register dst, Register src1, const Operand& src2, SBit s = LeaveCC, 443 void add(Register dst, Register src1, Register src2, SBit s = LeaveCC, 446 void adc(Register dst, Register src1, cons [all...] |