/third_party/skia/third_party/externals/harfbuzz/src/ |
H A D | hb-cff2-interp-cs.hh | 92 scalars.init (); in init() 99 scalars.fini (); in fini() 136 if (unlikely (!scalars.resize (region_count))) in process_blend() 140 &scalars[0], region_count); in process_blend() local 171 if (likely (scalars.length == arg.deltas.length)) in blend_arg() 174 for (unsigned int i = 0; i < scalars.length; i++) in blend_arg() 176 v += (double)scalars[i] * arg.deltas[i].to_real (); in blend_arg() 190 hb_vector_t<float> scalars; member
|
/third_party/skia/modules/svg/src/ |
H A D | SkSVGAttributeParser.cpp | 418 SkScalar scalars[6]; 420 if (!(this->parseScalarToken(scalars + i) && 426 m->setAll(scalars[0], scalars[2], scalars[4], scalars[1], scalars[3], scalars[5], 0, 0, 1);
|
/third_party/vk-gl-cts/modules/gles2/scripts/ |
H A D | genutil.py | 596 def __init__ (self, numCols, numRows, scalars): 597 assert len(scalars) == numRows*numCols 600 self.scalars = scalars 604 scalars = [] 607 scalars.append(1.0 if col == row else 0.0) 608 return Mat(numCols, numRows, scalars) 613 return self.scalars[colNdx*self.numRows + rowNdx] 618 self.scalars[colNdx*self.numRows + rowNdx] = scalar 644 return "%s(%s)" % (self.typeString(), ", ".join([str(s) for s in self.scalars])) [all...] |
H A D | gen-conversions.py | 35 # - scalars types 82 scalars = reduce(operator.add, [x[ndx].toFloat().getScalars() for x in comps]) 83 # print "->", scalars 84 res.append(Vec.fromScalarList(scalars)) 102 scalars = reduce(operator.add, [x[ndx].toFloat().getScalars() for x in comps]) 103 res.append(Mat(numCols, numRows, scalars))
|
/third_party/openssl/test/ |
H A D | ectest.c | 38 /* test multiplication with group order, long and negative scalars */ 78 const BIGNUM *scalars[6]; in group_order_tests() local 117 scalars[0] = scalars[1] = BN_value_one(); in group_order_tests() 120 if (!TEST_true(EC_POINTs_mul(group, R, NULL, 2, points, scalars, ctx)) in group_order_tests() 125 scalars[0] = n1; in group_order_tests() 127 scalars[1] = n2; in group_order_tests() 129 scalars[2] = n1; in group_order_tests() 131 scalars[3] = n2; in group_order_tests() 133 scalars[ in group_order_tests() 168 const BIGNUM *scalars[4]; prime_field_tests() local 789 const BIGNUM *scalars[3]; char2_curve_test() local [all...] |
/third_party/node/deps/openssl/openssl/crypto/ec/ |
H A D | ecp_s390x_nistp.c | 50 const BIGNUM *scalars[], in ec_GFp_s390x_nistp_mul() 81 || (scalar == NULL && num == 1 && BN_is_negative(scalars[0]) == 0)) { in ec_GFp_s390x_nistp_mul() 88 scalar_ptr = scalars[0]; in ec_GFp_s390x_nistp_mul() 118 rc = ossl_ec_wNAF_mul(group, r, scalar, num, points, scalars, ctx); in ec_GFp_s390x_nistp_mul() 291 const BIGNUM *scalars[], \ 295 scalars, ctx, \ 47 ec_GFp_s390x_nistp_mul(const EC_GROUP *group, EC_POINT *r, const BIGNUM *scalar, size_t num, const EC_POINT *points[], const BIGNUM *scalars[], BN_CTX *ctx, unsigned int fc, int len) ec_GFp_s390x_nistp_mul() argument
|
H A D | ec_local.h | 122 * + points[0] * scalars[0] 124 * + points[num-1] * scalars[num-1]. 129 * should provide an input scalar (either scalar or scalars[0]) 136 size_t num, const EC_POINT *points[], const BIGNUM *scalars[], 354 const BIGNUM *scalars[], BN_CTX *); 526 const EC_POINT *points[], const BIGNUM *scalars[], 531 const BIGNUM *scalars[], BN_CTX *ctx); 546 const EC_POINT *points[], const BIGNUM *scalars[], 551 const BIGNUM *scalars[], BN_CTX *ctx); 566 const EC_POINT *points[], const BIGNUM *scalars[], [all...] |
H A D | ecp_nistz256.c | 630 const BIGNUM **scalars = NULL; in ecp_nistz256_windowed_mul() local 639 || (scalars = OPENSSL_malloc(num * sizeof(BIGNUM *))) == NULL) { in ecp_nistz256_windowed_mul() 660 scalars[i] = mod; in ecp_nistz256_windowed_mul() 662 scalars[i] = scalar[i]; in ecp_nistz256_windowed_mul() 664 for (j = 0; j < bn_get_top(scalars[i]) * BN_BYTES; j += BN_BYTES) { in ecp_nistz256_windowed_mul() 665 BN_ULONG d = bn_get_words(scalars[i])[j / BN_BYTES]; in ecp_nistz256_windowed_mul() 785 OPENSSL_free(scalars); in ecp_nistz256_windowed_mul() 951 /* r = scalar*G + sum(scalars[i]*points[i]) */ 957 const BIGNUM *scalars[], BN_CTX *ctx) in ecp_nistz256_points_mul() 1137 memcpy(new_scalars, scalars, nu in ecp_nistz256_points_mul() 952 ecp_nistz256_points_mul(const EC_GROUP *group, EC_POINT *r, const BIGNUM *scalar, size_t num, const EC_POINT *points[], const BIGNUM *scalars[], BN_CTX *ctx) ecp_nistz256_points_mul() argument [all...] |
H A D | ec_mult.c | 405 * \sum scalars[i]*points[i], 412 const BIGNUM *scalars[], BN_CTX *ctx) in ossl_ec_wNAF_mul() 434 * treated like other scalars, i.e. in ossl_ec_wNAF_mul() 455 if ((scalar == NULL) && (num == 1) && (scalars[0] != group->order)) { in ossl_ec_wNAF_mul() 463 return ossl_ec_scalar_mul_ladder(group, r, scalars[0], points[0], in ossl_ec_wNAF_mul() 507 * 'scalars' */ in ossl_ec_wNAF_mul() 536 bits = i < num ? BN_num_bits(scalars[i]) : BN_num_bits(scalar); in ossl_ec_wNAF_mul() 541 bn_compute_wNAF((i < num ? scalars[i] : scalar), wsize[i], in ossl_ec_wNAF_mul() 410 ossl_ec_wNAF_mul(const EC_GROUP *group, EC_POINT *r, const BIGNUM *scalar, size_t num, const EC_POINT *points[], const BIGNUM *scalars[], BN_CTX *ctx) ossl_ec_wNAF_mul() argument
|
H A D | ecp_nistp256.c | 1667 * small point multiples 0*P, 1*P, ..., 17*P are in pre_comp[], the scalars 1668 * in scalars[]. If g_scalar is non-NULL, we also add this multiple of the 1673 const felem_bytearray scalars[], in batch_mul() 1689 * Loop over all scalars msb-to-lsb, interleaving additions of multiples in batch_mul() 1735 /* loop over all scalars */ in batch_mul() 1737 bits = get_bit(scalars[num], i + 4) << 5; in batch_mul() 1738 bits |= get_bit(scalars[num], i + 3) << 4; in batch_mul() 1739 bits |= get_bit(scalars[num], i + 2) << 3; in batch_mul() 1740 bits |= get_bit(scalars[num], i + 1) << 2; in batch_mul() 1741 bits |= get_bit(scalars[nu in batch_mul() 1672 batch_mul(felem x_out, felem y_out, felem z_out, const felem_bytearray scalars[], const unsigned num_points, const u8 *g_scalar, const int mixed, const smallfelem pre_comp[][17][3], const smallfelem g_pre_comp[2][16][3]) batch_mul() argument 2023 ossl_ec_GFp_nistp256_points_mul(const EC_GROUP *group, EC_POINT *r, const BIGNUM *scalar, size_t num, const EC_POINT *points[], const BIGNUM *scalars[], BN_CTX *ctx) ossl_ec_GFp_nistp256_points_mul() argument [all...] |
H A D | ecp_nistp521.c | 1577 * small point multiples 0*P, 1*P, ..., 16*P are in pre_comp[], the scalars 1578 * in scalars[]. If g_scalar is non-NULL, we also add this multiple of the 1583 const felem_bytearray scalars[], in batch_mul() 1598 * Loop over all scalars msb-to-lsb, interleaving additions of multiples in batch_mul() 1631 /* loop over all scalars */ in batch_mul() 1633 bits = get_bit(scalars[num], i + 4) << 5; in batch_mul() 1634 bits |= get_bit(scalars[num], i + 3) << 4; in batch_mul() 1635 bits |= get_bit(scalars[num], i + 2) << 3; in batch_mul() 1636 bits |= get_bit(scalars[num], i + 1) << 2; in batch_mul() 1637 bits |= get_bit(scalars[nu in batch_mul() 1582 batch_mul(felem x_out, felem y_out, felem z_out, const felem_bytearray scalars[], const unsigned num_points, const u8 *g_scalar, const int mixed, const felem pre_comp[][17][3], const felem g_pre_comp[16][3]) batch_mul() argument 1916 ossl_ec_GFp_nistp521_points_mul(const EC_GROUP *group, EC_POINT *r, const BIGNUM *scalar, size_t num, const EC_POINT *points[], const BIGNUM *scalars[], BN_CTX *ctx) ossl_ec_GFp_nistp521_points_mul() argument [all...] |
H A D | ecp_nistp224.c | 85 * scalars for point multiplication. 466 * Multiply a field element by a scalar: out = out * scalar The scalars we 479 * scalars we actually use are small, so results fit without overflow 1134 * small point multiples 0*P, 1*P, ..., 16*P are in pre_comp[], the scalars 1135 * in scalars[]. If g_scalar is non-NULL, we also add this multiple of the 1140 const felem_bytearray scalars[], in batch_mul() 1156 * Loop over all scalars msb-to-lsb, interleaving additions of multiples in batch_mul() 1200 /* loop over all scalars */ in batch_mul() 1202 bits = get_bit(scalars[num], i + 4) << 5; in batch_mul() 1203 bits |= get_bit(scalars[nu in batch_mul() 1139 batch_mul(felem x_out, felem y_out, felem z_out, const felem_bytearray scalars[], const unsigned num_points, const u8 *g_scalar, const int mixed, const felem pre_comp[][17][3], const felem g_pre_comp[2][16][3]) batch_mul() argument 1411 ossl_ec_GFp_nistp224_points_mul(const EC_GROUP *group, EC_POINT *r, const BIGNUM *scalar, size_t num, const EC_POINT *points[], const BIGNUM *scalars[], BN_CTX *ctx) ossl_ec_GFp_nistp224_points_mul() argument [all...] |
/third_party/openssl/crypto/ec/ |
H A D | ecp_s390x_nistp.c | 50 const BIGNUM *scalars[], in ec_GFp_s390x_nistp_mul() 81 || (scalar == NULL && num == 1 && BN_is_negative(scalars[0]) == 0)) { in ec_GFp_s390x_nistp_mul() 88 scalar_ptr = scalars[0]; in ec_GFp_s390x_nistp_mul() 118 rc = ossl_ec_wNAF_mul(group, r, scalar, num, points, scalars, ctx); in ec_GFp_s390x_nistp_mul() 291 const BIGNUM *scalars[], \ 295 scalars, ctx, \ 47 ec_GFp_s390x_nistp_mul(const EC_GROUP *group, EC_POINT *r, const BIGNUM *scalar, size_t num, const EC_POINT *points[], const BIGNUM *scalars[], BN_CTX *ctx, unsigned int fc, int len) ec_GFp_s390x_nistp_mul() argument
|
H A D | ec_local.h | 122 * + points[0] * scalars[0] 124 * + points[num-1] * scalars[num-1]. 129 * should provide an input scalar (either scalar or scalars[0]) 136 size_t num, const EC_POINT *points[], const BIGNUM *scalars[], 354 const BIGNUM *scalars[], BN_CTX *); 526 const EC_POINT *points[], const BIGNUM *scalars[], 531 const BIGNUM *scalars[], BN_CTX *ctx); 546 const EC_POINT *points[], const BIGNUM *scalars[], 551 const BIGNUM *scalars[], BN_CTX *ctx); 566 const EC_POINT *points[], const BIGNUM *scalars[], [all...] |
H A D | ecp_nistz256.c | 630 const BIGNUM **scalars = NULL; in ecp_nistz256_windowed_mul() local 639 || (scalars = OPENSSL_malloc(num * sizeof(BIGNUM *))) == NULL) { in ecp_nistz256_windowed_mul() 660 scalars[i] = mod; in ecp_nistz256_windowed_mul() 662 scalars[i] = scalar[i]; in ecp_nistz256_windowed_mul() 664 for (j = 0; j < bn_get_top(scalars[i]) * BN_BYTES; j += BN_BYTES) { in ecp_nistz256_windowed_mul() 665 BN_ULONG d = bn_get_words(scalars[i])[j / BN_BYTES]; in ecp_nistz256_windowed_mul() 785 OPENSSL_free(scalars); in ecp_nistz256_windowed_mul() 951 /* r = scalar*G + sum(scalars[i]*points[i]) */ 957 const BIGNUM *scalars[], BN_CTX *ctx) in ecp_nistz256_points_mul() 1137 memcpy(new_scalars, scalars, nu in ecp_nistz256_points_mul() 952 ecp_nistz256_points_mul(const EC_GROUP *group, EC_POINT *r, const BIGNUM *scalar, size_t num, const EC_POINT *points[], const BIGNUM *scalars[], BN_CTX *ctx) ecp_nistz256_points_mul() argument [all...] |
H A D | ec_mult.c | 405 * \sum scalars[i]*points[i], 412 const BIGNUM *scalars[], BN_CTX *ctx) in ossl_ec_wNAF_mul() 434 * treated like other scalars, i.e. in ossl_ec_wNAF_mul() 455 if ((scalar == NULL) && (num == 1) && (scalars[0] != group->order)) { in ossl_ec_wNAF_mul() 463 return ossl_ec_scalar_mul_ladder(group, r, scalars[0], points[0], in ossl_ec_wNAF_mul() 507 * 'scalars' */ in ossl_ec_wNAF_mul() 536 bits = i < num ? BN_num_bits(scalars[i]) : BN_num_bits(scalar); in ossl_ec_wNAF_mul() 541 bn_compute_wNAF((i < num ? scalars[i] : scalar), wsize[i], in ossl_ec_wNAF_mul() 410 ossl_ec_wNAF_mul(const EC_GROUP *group, EC_POINT *r, const BIGNUM *scalar, size_t num, const EC_POINT *points[], const BIGNUM *scalars[], BN_CTX *ctx) ossl_ec_wNAF_mul() argument
|
H A D | ecp_nistp256.c | 1667 * small point multiples 0*P, 1*P, ..., 17*P are in pre_comp[], the scalars 1668 * in scalars[]. If g_scalar is non-NULL, we also add this multiple of the 1673 const felem_bytearray scalars[], in batch_mul() 1689 * Loop over all scalars msb-to-lsb, interleaving additions of multiples in batch_mul() 1735 /* loop over all scalars */ in batch_mul() 1737 bits = get_bit(scalars[num], i + 4) << 5; in batch_mul() 1738 bits |= get_bit(scalars[num], i + 3) << 4; in batch_mul() 1739 bits |= get_bit(scalars[num], i + 2) << 3; in batch_mul() 1740 bits |= get_bit(scalars[num], i + 1) << 2; in batch_mul() 1741 bits |= get_bit(scalars[nu in batch_mul() 1672 batch_mul(felem x_out, felem y_out, felem z_out, const felem_bytearray scalars[], const unsigned num_points, const u8 *g_scalar, const int mixed, const smallfelem pre_comp[][17][3], const smallfelem g_pre_comp[2][16][3]) batch_mul() argument 2023 ossl_ec_GFp_nistp256_points_mul(const EC_GROUP *group, EC_POINT *r, const BIGNUM *scalar, size_t num, const EC_POINT *points[], const BIGNUM *scalars[], BN_CTX *ctx) ossl_ec_GFp_nistp256_points_mul() argument [all...] |
H A D | ecp_nistp521.c | 1577 * small point multiples 0*P, 1*P, ..., 16*P are in pre_comp[], the scalars 1578 * in scalars[]. If g_scalar is non-NULL, we also add this multiple of the 1583 const felem_bytearray scalars[], in batch_mul() 1598 * Loop over all scalars msb-to-lsb, interleaving additions of multiples in batch_mul() 1631 /* loop over all scalars */ in batch_mul() 1633 bits = get_bit(scalars[num], i + 4) << 5; in batch_mul() 1634 bits |= get_bit(scalars[num], i + 3) << 4; in batch_mul() 1635 bits |= get_bit(scalars[num], i + 2) << 3; in batch_mul() 1636 bits |= get_bit(scalars[num], i + 1) << 2; in batch_mul() 1637 bits |= get_bit(scalars[nu in batch_mul() 1582 batch_mul(felem x_out, felem y_out, felem z_out, const felem_bytearray scalars[], const unsigned num_points, const u8 *g_scalar, const int mixed, const felem pre_comp[][17][3], const felem g_pre_comp[16][3]) batch_mul() argument 1916 ossl_ec_GFp_nistp521_points_mul(const EC_GROUP *group, EC_POINT *r, const BIGNUM *scalar, size_t num, const EC_POINT *points[], const BIGNUM *scalars[], BN_CTX *ctx) ossl_ec_GFp_nistp521_points_mul() argument [all...] |
H A D | ecp_nistp224.c | 85 * scalars for point multiplication. 466 * Multiply a field element by a scalar: out = out * scalar The scalars we 479 * scalars we actually use are small, so results fit without overflow 1134 * small point multiples 0*P, 1*P, ..., 16*P are in pre_comp[], the scalars 1135 * in scalars[]. If g_scalar is non-NULL, we also add this multiple of the 1140 const felem_bytearray scalars[], in batch_mul() 1156 * Loop over all scalars msb-to-lsb, interleaving additions of multiples in batch_mul() 1200 /* loop over all scalars */ in batch_mul() 1202 bits = get_bit(scalars[num], i + 4) << 5; in batch_mul() 1203 bits |= get_bit(scalars[nu in batch_mul() 1139 batch_mul(felem x_out, felem y_out, felem z_out, const felem_bytearray scalars[], const unsigned num_points, const u8 *g_scalar, const int mixed, const felem pre_comp[][17][3], const felem g_pre_comp[2][16][3]) batch_mul() argument 1411 ossl_ec_GFp_nistp224_points_mul(const EC_GROUP *group, EC_POINT *r, const BIGNUM *scalar, size_t num, const EC_POINT *points[], const BIGNUM *scalars[], BN_CTX *ctx) ossl_ec_GFp_nistp224_points_mul() argument [all...] |
/third_party/openssl/ohos_lite/crypto/ec/ |
H A D | ec_local.h | 126 * + points[0] * scalars[0] 128 * + points[num-1] * scalars[num-1]. 133 * should provide an input scalar (either scalar or scalars[0]) 140 size_t num, const EC_POINT *points[], const BIGNUM *scalars[], 342 size_t num, const EC_POINT *points[], const BIGNUM *scalars[], 505 const EC_POINT *points[], const BIGNUM *scalars[], 510 const BIGNUM *scalars[], BN_CTX *ctx); 525 const EC_POINT *points[], const BIGNUM *scalars[], 530 const BIGNUM *scalars[], BN_CTX *ctx); 545 const EC_POINT *points[], const BIGNUM *scalars[], [all...] |
/third_party/vk-gl-cts/modules/gles3/scripts/ |
H A D | genutil.py | 739 def __init__ (self, numCols, numRows, scalars): 740 assert len(scalars) == numRows*numCols 743 self.scalars = scalars 747 scalars = [] 750 scalars.append(scalar if col == row else 0.0) 751 return Mat(numCols, numRows, scalars) 760 return self.scalars[colNdx*self.numRows + rowNdx] 765 self.scalars[colNdx*self.numRows + rowNdx] = scalar 791 return "%s(%s)" % (self.typeString(), ", ".join(["%s" % s for s in self.scalars])) [all...] |
H A D | gen-conversions.py | 35 # - scalars types 82 scalars = reduce(operator.add, [x[ndx].toFloat().getScalars() for x in comps]) 83 # print "->", scalars 84 res.append(Vec.fromScalarList(scalars)) 105 scalars = reduce(operator.add, [x[ndx].toFloat().getScalars() for x in comps]) 106 res.append(Mat(numCols, numRows, scalars))
|
/third_party/vk-gl-cts/modules/gles31/scripts/ |
H A D | genutil.py | 741 def __init__ (self, numCols, numRows, scalars): 742 assert len(scalars) == numRows*numCols 745 self.scalars = scalars 749 scalars = [] 752 scalars.append(scalar if col == row else 0.0) 753 return Mat(numCols, numRows, scalars) 762 return self.scalars[colNdx*self.numRows + rowNdx] 767 self.scalars[colNdx*self.numRows + rowNdx] = scalar 793 return "%s(%s)" % (self.typeString(), ", ".join(["%s" % s for s in self.scalars])) [all...] |
/third_party/mesa3d/src/asahi/lib/ |
H A D | agx_device.c | 405 uint64_t scalars[2] = { in agx_create_command_queue() local 412 scalars, 2, NULL, NULL); in agx_create_command_queue() 418 uint64_t scalars[2] = { in agx_create_command_queue() local 425 scalars, 2, NULL, NULL); in agx_create_command_queue()
|
/third_party/vk-gl-cts/external/openglcts/modules/common/ |
H A D | glcShaderLibraryCase.cpp | 388 std::vector<float>& scalars = attribValues[attribValueNdx++]; in execute() local 389 scalars.resize(numVerticesPerDraw * scalarSize); in execute() 394 scalars[repNdx * scalarSize + ndx] = val.elements[arrayNdx * scalarSize + ndx].float32; in execute() 405 scalars[repNdx * scalarSize + ndx] = v; in execute() 435 &scalars[i * numRows])); in execute() 441 vertexArrays.push_back(va::Float(attribLoc, scalarSize, numVerticesPerDraw, 0, &scalars[0])); in execute()
|