/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/Mips/ |
H A D | MipsLegalizerInfo.cpp | 57 const LLT s32 = LLT::scalar(32); in MipsLegalizerInfo() local 67 if (CheckTyN(0, Query, {s32})) in MipsLegalizerInfo() 73 .clampScalar(0, s32, s32); in MipsLegalizerInfo() 76 .lowerFor({{s32, s1}}); in MipsLegalizerInfo() 79 .legalFor({s32}) in MipsLegalizerInfo() 80 .maxScalar(0, s32); in MipsLegalizerInfo() 84 if (CheckTy0Ty1MemSizeAlign(Query, {{s32, p0, 8, ST.hasMips32r6()}, in MipsLegalizerInfo() 85 {s32, p0, 16, ST.hasMips32r6()}, in MipsLegalizerInfo() 86 {s32, p in MipsLegalizerInfo() 269 const LLT s32 = LLT::scalar(32); legalizeCustom() local [all...] |
/third_party/ffmpeg/libavcodec/arm/ |
H A D | fmtconvert_vfp.S | 55 vcvt.f32.s32 s16, s16 56 vcvt.f32.s32 s17, s17 57 vcvt.f32.s32 s18, s18 58 vcvt.f32.s32 s19, s19 59 vcvt.f32.s32 s20, s20 60 vcvt.f32.s32 s21, s21 61 vcvt.f32.s32 s22, s22 62 vcvt.f32.s32 s23, s23 68 vcvt.f32.s32 s24, s24 69 vcvt.f32.s32 s2 [all...] |
H A D | rv34dsp_neon.S | 39 vadd.s32 q13, q13, q9 @ z3 = 17*block[i+4*1] + 7*block[i+4*3] 40 vsub.s32 q12, q12, q1 @ z2 = 7*block[i+4*1] - 17*block[i+4*3] 41 vadd.s32 q1, q10, q13 @ z0 + z3 42 vadd.s32 q2, q11, q12 @ z1 + z2 43 vsub.s32 q8, q10, q13 @ z0 - z3 44 vsub.s32 q3, q11, q12 @ z1 - z2 49 vmov.s32 d0, #13 50 vadd.s32 q10, q1, q3 51 vsub.s32 q11, q1, q3 52 vshl.s32 q1 [all...] |
H A D | sbcdsp_neon.S | 67 vpadd.s32 d0, d0, d1 68 vpadd.s32 d1, d2, d3 70 vrshrn.s32 d0, q0, SBC_PROTO_FIXED_SCALE 82 vpadd.s32 d0, d20, d21 /* TODO: can be eliminated */ 83 vpadd.s32 d1, d22, d23 /* TODO: can be eliminated */ 141 vpadd.s32 d0, d24, d25 142 vpadd.s32 d1, d26, d27 143 vpadd.s32 d2, d28, d29 144 vpadd.s32 d3, d30, d31 146 vrshr.s32 q [all...] |
H A D | vp3dsp_neon.S | 133 vshrn.s32 d4, q2, #16 134 vshrn.s32 d5, q3, #16 135 vshrn.s32 d6, q4, #16 136 vshrn.s32 d7, q5, #16 137 vshrn.s32 d8, q6, #16 138 vshrn.s32 d9, q7, #16 149 vshrn.s32 d4, q2, #16 150 vshrn.s32 d5, q3, #16 151 vshrn.s32 d6, q4, #16 // ip[7] * C7 152 vshrn.s32 d [all...] |
H A D | int_neon.S | 40 vpadd.s32 d16, d0, d1 41 vpadd.s32 d17, d2, d3 42 vpadd.s32 d18, d4, d5 43 vpadd.s32 d19, d6, d7 44 vpadd.s32 d0, d16, d17 45 vpadd.s32 d1, d18, d19 46 vpadd.s32 d2, d0, d1 47 vpaddl.s32 d3, d2
|
H A D | fmtconvert_neon.S | 33 vcvt.f32.s32 q3, q1 35 vcvt.f32.s32 q8, q2 42 vcvt.f32.s32 q3, q1 44 vcvt.f32.s32 q8, q2 64 vcvt.f32.s32 q0, q0 65 vcvt.f32.s32 q1, q1 67 vcvt.f32.s32 q2, q2 68 vcvt.f32.s32 q3, q3 82 vcvt.f32.s32 q0, q0 83 vcvt.f32.s32 q [all...] |
H A D | lossless_audiodsp_neon.S | 52 vpadd.s32 d16, d0, d1 53 vpadd.s32 d17, d2, d3 54 vpadd.s32 d18, d4, d5 55 vpadd.s32 d19, d6, d7 56 vpadd.s32 d0, d16, d17 57 vpadd.s32 d1, d18, d19 58 vpadd.s32 d2, d0, d1 59 vpaddl.s32 d3, d2
|
H A D | hevcdsp_qpel_neon.S | 90 vsub.s32 q9, q11 // 58 * d0 - 10 * c0 91 vsub.s32 q10, q12 // 58 * d1 - 10 * c1 94 vadd.s32 q9, q13 // 58 * d0 - 10 * c0 + 17 * e0 95 vadd.s32 q10, q14 // 58 * d1 - 10 * c1 + 17 * e1 98 vadd.s32 q9, q11 // 58 * d0 - 10 * c0 + 17 * e0 + 4 * b0 99 vadd.s32 q10, q12 // 58 * d1 - 10 * c1 + 17 * e1 + 4 * b1 100 vsub.s32 q13, q15 // g0 - a0 - 5 * f0 101 vsub.s32 q14, q8 // g1 - a1 - 5 * f1 102 vadd.s32 q9, q13 // 58 * d0 - 10 * c0 + 17 * e0 + 4 * b0 + g0 - a0 - 5 * f0 103 vadd.s32 q1 [all...] |
H A D | ac3dsp_neon.S | 76 vshl.s32 q1, q1, q0 85 vcvt.s32.f32 q0, q0, #24 87 vcvt.s32.f32 q1, q1, #24 88 vcvt.s32.f32 q2, q2, #24 90 vcvt.s32.f32 q3, q3, #24 101 vabs.s32 q1, q0 143 vadd.s32 d18, d16, d17 144 vsub.s32 d19, d16, d17 145 vmlal.s32 q0, d16, d16 146 vmlal.s32 q [all...] |
H A D | vp9itxfm_16bpp_neon.S | 74 vadd.s32 \tmpd1, \in1, \in2 75 vsub.s32 \tmpd2, \in1, \in2 77 vneg.s32 \tmpd1, \tmpd1 79 vmull.s32 \tmpq3, \tmpd1, d0[0] 80 vmull.s32 \tmpq4, \tmpd2, d0[0] 88 vmull.s32 \tmpq3, \in1, d0[0] 99 vadd.s32 \tmpq1, \in1, \in2 100 vsub.s32 \tmpq2, \in1, \in2 101 vmull.s32 \tmpq3, \tmpd11, d0[0] 102 vmull.s32 \tmpq [all...] |
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/ARM/ |
H A D | ARMLegalizerInfo.cpp | 74 const LLT s32 = LLT::scalar(32); in ARMLegalizerInfo() local 85 .legalForCartesianProduct({s8, s16, s32}, {s1, s8, s16}); in ARMLegalizerInfo() 90 .legalFor({s32}) in ARMLegalizerInfo() 91 .minScalar(0, s32); in ARMLegalizerInfo() 95 .legalFor({s32, s64}) in ARMLegalizerInfo() 96 .minScalar(0, s32); in ARMLegalizerInfo() 99 .legalFor({s32}) in ARMLegalizerInfo() 100 .minScalar(0, s32); in ARMLegalizerInfo() 103 .legalFor({{s32, s32}}) in ARMLegalizerInfo() [all...] |
/third_party/ffmpeg/libavresample/arm/ |
H A D | audio_convert_neon.S | 27 vcvt.s32.f32 q8, q0, #31 29 vcvt.s32.f32 q9, q1, #31 34 vqrshrn.s32 d4, q8, #16 36 vcvt.s32.f32 q0, q0, #31 37 vqrshrn.s32 d5, q9, #16 39 vcvt.s32.f32 q1, q1, #31 40 vqrshrn.s32 d6, q0, #16 42 vqrshrn.s32 d7, q1, #16 44 vcvt.s32.f32 q8, q8, #31 46 vcvt.s32 [all...] |
/third_party/ffmpeg/libswresample/arm/ |
H A D | audio_convert_neon.S | 28 vcvt.s32.f32 q8, q0, #31 30 vcvt.s32.f32 q9, q1, #31 35 vqrshrn.s32 d4, q8, #16 37 vcvt.s32.f32 q0, q0, #31 38 vqrshrn.s32 d5, q9, #16 40 vcvt.s32.f32 q1, q1, #31 41 vqrshrn.s32 d6, q0, #16 43 vqrshrn.s32 d7, q1, #16 45 vcvt.s32.f32 q8, q8, #31 47 vcvt.s32 [all...] |
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/AArch64/ |
H A D | AArch64LegalizerInfo.cpp | 39 const LLT s32 = LLT::scalar(32); in AArch64LegalizerInfo() local 62 .legalFor({p0, s1, s8, s16, s32, s64, v2s32, v4s32, v2s64}) in AArch64LegalizerInfo() 79 .legalFor({p0, s16, s32, s64, v2s32, v4s32, v2s64}) in AArch64LegalizerInfo() 84 .legalFor({s32, s64, v4s32, v2s32, v2s64}) in AArch64LegalizerInfo() 85 .clampScalar(0, s32, s64) in AArch64LegalizerInfo() 89 .legalFor({s32, s64, v2s32, v4s32, v2s64, v8s16, v16s8}) in AArch64LegalizerInfo() 90 .clampScalar(0, s32, s64) in AArch64LegalizerInfo() 97 .legalFor({{s32, s32}, {s64, s64}, in AArch64LegalizerInfo() 99 .clampScalar(1, s32, s6 in AArch64LegalizerInfo() [all...] |
/third_party/ltp/tools/sparse/sparse-src/validation/linear/ |
H A D | shift-assign1.c | 2 typedef __INT32_TYPE__ s32; typedef 9 s16 s16s32(s16 a, s32 b) { a >>= b; return a; } in s16s32() 14 s32 s32s16(s32 a, s16 b) { a >>= b; return a; } in s32s16() 15 s32 s32s32(s32 a, s32 b) { a >>= b; return a; } in s32s32() 16 s32 s32s64(s32 a, s64 b) { a >>= b; return a; } in s32s64() 17 s32 s32u1 in s32s64() [all...] |
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/X86/ |
H A D | X86LegalizerInfo.cpp | 112 const LLT s32 = LLT::scalar(32); in setLegalizerInfo32bit() local 116 for (auto Ty : {p0, s1, s8, s16, s32}) in setLegalizerInfo32bit() 119 for (auto Ty : {s8, s16, s32, p0}) in setLegalizerInfo32bit() 123 for (auto Ty : {s8, s16, s32}) in setLegalizerInfo32bit() 127 setAction({Op, s32}, Legal); in setLegalizerInfo32bit() 132 for (auto Ty : {s8, s16, s32, p0}) in setLegalizerInfo32bit() 144 setAction({G_PTR_ADD, 1, s32}, Legal); in setLegalizerInfo32bit() 148 .legalForCartesianProduct({s1, s8, s16, s32}, {p0}) in setLegalizerInfo32bit() 149 .maxScalar(0, s32) in setLegalizerInfo32bit() 151 getActionDefinitionsBuilder(G_INTTOPTR).legalFor({{p0, s32}}); in setLegalizerInfo32bit() 208 const LLT s32 = LLT::scalar(32); setLegalizerInfo64bit() local 288 const LLT s32 = LLT::scalar(32); setLegalizerInfoSSE1() local 317 const LLT s32 = LLT::scalar(32); setLegalizerInfoSSE2() local [all...] |
/third_party/ltp/tools/sparse/sparse-src/validation/optim/ |
H A D | shift-big.c | 2 typedef int s32; typedef 4 s32 asr31(s32 a) { return a >> 31; } in asr31() 5 s32 asr32(s32 a) { return a >> 32; } in asr32() 6 s32 asr33(s32 a) { return a >> 33; } in asr33()
|
/third_party/ffmpeg/libswscale/arm/ |
H A D | hscale.S | 33 vmov.s32 q4, #0 @ val accumulator 34 vmov.s32 q5, #0 @ val accumulator 49 vpadd.s32 d16, d16, d17 @ horizontal pair adding of the 8x32-bit multiplied values into 4x32-bit (part 1) 50 vpadd.s32 d17, d18, d19 @ horizontal pair adding of the 8x32-bit multiplied values into 4x32-bit (part 2) 51 vpadd.s32 d20, d20, d21 @ horizontal pair adding of the 8x32-bit multiplied values into 4x32-bit (part 1) 52 vpadd.s32 d21, d22, d23 @ horizontal pair adding of the 8x32-bit multiplied values into 4x32-bit (part 2) 53 vadd.s32 q4, q8 @ update val accumulator 54 vadd.s32 q5, q10 @ update val accumulator 60 vpadd.s32 d8, d8, d9 @ horizontal pair adding of the 8x32-bit sums into 4x32-bit (part 1) 61 vpadd.s32 d [all...] |
H A D | output.S | 58 vpadd.s32 d10, d18, d19 @ A*X+I*Y,B*X+J*Y 59 vpadd.s32 d11, d20, d21 @ C*X+K*Y,D*X+L*Y 60 vpadd.s32 d12, d22, d23 @ E*X+M*Y,F*X+N*Y 61 vpadd.s32 d13, d24, d25 @ G*X+O*Y,H*X+P*Y 62 vadd.s32 q3, q5 @ update val accumulator (part 1) 63 vadd.s32 q4, q6 @ update val accumulator (part 2) 66 vshr.s32 q3, q3, #19 @ val>>19 (part 1) 67 vshr.s32 q4, q4, #19 @ val>>19 (part 2) 68 vqmovun.s32 d6, q3 @ clip16(val>>19) (part 1) 69 vqmovun.s32 d [all...] |
/third_party/ntfs-3g/include/ntfs-3g/ |
H A D | endians.h | 204 #define sle32_to_cpu(x) (s32)__le32_to_cpu((s32)(x)) 208 #define sle32_to_cpup(x) (s32)__le32_to_cpu(*(s32*)(x)) 224 #define cpu_to_sle32(x) (s32)__cpu_to_le32((s32)(x)) 228 #define cpu_to_sle32p(x) (s32)__cpu_to_le32(*(s32*)(x)) 244 #define sbe32_to_cpu(x) (s32)__be32_to_cpu((s32)( [all...] |
/third_party/alsa-lib/test/ |
H A D | mixtest.c | 15 typedef int s32; typedef 90 volatile s32 *sum, in mix_areas_srv() 103 s16 *dst, const s32 *sum, in saturate() 109 s32 sample = *sum; in saturate() 123 volatile s32 *sum, in mix_areas0() 132 s32 sample = *dst + *src; in mix_areas0() 159 volatile s32 *sum, in mix_areas2() 166 s32 sample = *src; in mix_areas2() 167 s32 old_sample = *sum; in mix_areas2() 205 void init(s16 *dst, s32 *su [all...] |
/third_party/skia/third_party/externals/spirv-tools/test/opt/ |
H A D | types_test.cpp | 133 auto* s32 = types.back().get(); in GenerateAllTypes() local 145 types.emplace_back(new Vector(s32, 2)); in GenerateAllTypes() 146 types.emplace_back(new Vector(s32, 3)); in GenerateAllTypes() 158 types.emplace_back(new Image(s32, SpvDim2D, 0, 0, 0, 0, SpvImageFormatRg8, in GenerateAllTypes() 161 types.emplace_back(new Image(s32, SpvDim2D, 0, 1, 0, 0, SpvImageFormatRg8, in GenerateAllTypes() 163 types.emplace_back(new Image(s32, SpvDim3D, 0, 1, 0, 0, SpvImageFormatRg8, in GenerateAllTypes() 209 types.emplace_back(new Struct(std::vector<const Type*>{s32})); in GenerateAllTypes() 210 types.emplace_back(new Struct(std::vector<const Type*>{s32, f32})); in GenerateAllTypes() 229 types.emplace_back(new Function(voidt, {boolt, s32})); in GenerateAllTypes() 230 types.emplace_back(new Function(s32, {bool in GenerateAllTypes() 303 auto s32 = MakeUnique<Integer>(32, true); TEST() local 311 auto s32 = MakeUnique<Integer>(32, true); TEST() local [all...] |
/third_party/skia/third_party/externals/swiftshader/third_party/SPIRV-Tools/test/opt/ |
H A D | types_test.cpp | 133 auto* s32 = types.back().get(); in GenerateAllTypes() local 145 types.emplace_back(new Vector(s32, 2)); in GenerateAllTypes() 146 types.emplace_back(new Vector(s32, 3)); in GenerateAllTypes() 158 types.emplace_back(new Image(s32, SpvDim2D, 0, 0, 0, 0, SpvImageFormatRg8, in GenerateAllTypes() 161 types.emplace_back(new Image(s32, SpvDim2D, 0, 1, 0, 0, SpvImageFormatRg8, in GenerateAllTypes() 163 types.emplace_back(new Image(s32, SpvDim3D, 0, 1, 0, 0, SpvImageFormatRg8, in GenerateAllTypes() 209 types.emplace_back(new Struct(std::vector<const Type*>{s32})); in GenerateAllTypes() 210 types.emplace_back(new Struct(std::vector<const Type*>{s32, f32})); in GenerateAllTypes() 229 types.emplace_back(new Function(voidt, {boolt, s32})); in GenerateAllTypes() 230 types.emplace_back(new Function(s32, {bool in GenerateAllTypes() 303 auto s32 = MakeUnique<Integer>(32, true); TEST() local 311 auto s32 = MakeUnique<Integer>(32, true); TEST() local [all...] |
/third_party/mesa3d/src/nouveau/codegen/lib/ |
H A D | gf100.asm | 48 set $p2 0x1 lt s32 $r0 0x0 49 set $p3 0x1 lt s32 $r1 0x0 xor $p2 50 cvt s32 $r0 abs s32 $r0 51 cvt s32 $r1 abs s32 $r1 77 $p3 cvt s32 $r0 neg s32 $r0 78 $p2 cvt s32 $r1 neg s32 [all...] |