Home
last modified time | relevance | path

Searched refs:scalars (Results 1 - 25 of 39) sorted by relevance

12

/third_party/skia/third_party/externals/harfbuzz/src/
H A Dhb-cff2-interp-cs.hh92 scalars.init (); in init()
99 scalars.fini (); in fini()
136 if (unlikely (!scalars.resize (region_count))) in process_blend()
140 &scalars[0], region_count); in process_blend() local
171 if (likely (scalars.length == arg.deltas.length)) in blend_arg()
174 for (unsigned int i = 0; i < scalars.length; i++) in blend_arg()
176 v += (double)scalars[i] * arg.deltas[i].to_real (); in blend_arg()
190 hb_vector_t<float> scalars; member
/third_party/skia/modules/svg/src/
H A DSkSVGAttributeParser.cpp418 SkScalar scalars[6];
420 if (!(this->parseScalarToken(scalars + i) &&
426 m->setAll(scalars[0], scalars[2], scalars[4], scalars[1], scalars[3], scalars[5], 0, 0, 1);
/third_party/vk-gl-cts/modules/gles2/scripts/
H A Dgenutil.py596 def __init__ (self, numCols, numRows, scalars):
597 assert len(scalars) == numRows*numCols
600 self.scalars = scalars
604 scalars = []
607 scalars.append(1.0 if col == row else 0.0)
608 return Mat(numCols, numRows, scalars)
613 return self.scalars[colNdx*self.numRows + rowNdx]
618 self.scalars[colNdx*self.numRows + rowNdx] = scalar
644 return "%s(%s)" % (self.typeString(), ", ".join([str(s) for s in self.scalars]))
[all...]
H A Dgen-conversions.py35 # - scalars types
82 scalars = reduce(operator.add, [x[ndx].toFloat().getScalars() for x in comps])
83 # print "->", scalars
84 res.append(Vec.fromScalarList(scalars))
102 scalars = reduce(operator.add, [x[ndx].toFloat().getScalars() for x in comps])
103 res.append(Mat(numCols, numRows, scalars))
/third_party/openssl/test/
H A Dectest.c38 /* test multiplication with group order, long and negative scalars */
78 const BIGNUM *scalars[6]; in group_order_tests() local
117 scalars[0] = scalars[1] = BN_value_one(); in group_order_tests()
120 if (!TEST_true(EC_POINTs_mul(group, R, NULL, 2, points, scalars, ctx)) in group_order_tests()
125 scalars[0] = n1; in group_order_tests()
127 scalars[1] = n2; in group_order_tests()
129 scalars[2] = n1; in group_order_tests()
131 scalars[3] = n2; in group_order_tests()
133 scalars[ in group_order_tests()
168 const BIGNUM *scalars[4]; prime_field_tests() local
789 const BIGNUM *scalars[3]; char2_curve_test() local
[all...]
/third_party/node/deps/openssl/openssl/crypto/ec/
H A Decp_s390x_nistp.c50 const BIGNUM *scalars[], in ec_GFp_s390x_nistp_mul()
81 || (scalar == NULL && num == 1 && BN_is_negative(scalars[0]) == 0)) { in ec_GFp_s390x_nistp_mul()
88 scalar_ptr = scalars[0]; in ec_GFp_s390x_nistp_mul()
118 rc = ossl_ec_wNAF_mul(group, r, scalar, num, points, scalars, ctx); in ec_GFp_s390x_nistp_mul()
291 const BIGNUM *scalars[], \
295 scalars, ctx, \
47 ec_GFp_s390x_nistp_mul(const EC_GROUP *group, EC_POINT *r, const BIGNUM *scalar, size_t num, const EC_POINT *points[], const BIGNUM *scalars[], BN_CTX *ctx, unsigned int fc, int len) ec_GFp_s390x_nistp_mul() argument
H A Dec_local.h122 * + points[0] * scalars[0]
124 * + points[num-1] * scalars[num-1].
129 * should provide an input scalar (either scalar or scalars[0])
136 size_t num, const EC_POINT *points[], const BIGNUM *scalars[],
354 const BIGNUM *scalars[], BN_CTX *);
526 const EC_POINT *points[], const BIGNUM *scalars[],
531 const BIGNUM *scalars[], BN_CTX *ctx);
546 const EC_POINT *points[], const BIGNUM *scalars[],
551 const BIGNUM *scalars[], BN_CTX *ctx);
566 const EC_POINT *points[], const BIGNUM *scalars[],
[all...]
H A Decp_nistz256.c630 const BIGNUM **scalars = NULL; in ecp_nistz256_windowed_mul() local
639 || (scalars = OPENSSL_malloc(num * sizeof(BIGNUM *))) == NULL) { in ecp_nistz256_windowed_mul()
660 scalars[i] = mod; in ecp_nistz256_windowed_mul()
662 scalars[i] = scalar[i]; in ecp_nistz256_windowed_mul()
664 for (j = 0; j < bn_get_top(scalars[i]) * BN_BYTES; j += BN_BYTES) { in ecp_nistz256_windowed_mul()
665 BN_ULONG d = bn_get_words(scalars[i])[j / BN_BYTES]; in ecp_nistz256_windowed_mul()
785 OPENSSL_free(scalars); in ecp_nistz256_windowed_mul()
951 /* r = scalar*G + sum(scalars[i]*points[i]) */
957 const BIGNUM *scalars[], BN_CTX *ctx) in ecp_nistz256_points_mul()
1137 memcpy(new_scalars, scalars, nu in ecp_nistz256_points_mul()
952 ecp_nistz256_points_mul(const EC_GROUP *group, EC_POINT *r, const BIGNUM *scalar, size_t num, const EC_POINT *points[], const BIGNUM *scalars[], BN_CTX *ctx) ecp_nistz256_points_mul() argument
[all...]
H A Dec_mult.c405 * \sum scalars[i]*points[i],
412 const BIGNUM *scalars[], BN_CTX *ctx) in ossl_ec_wNAF_mul()
434 * treated like other scalars, i.e. in ossl_ec_wNAF_mul()
455 if ((scalar == NULL) && (num == 1) && (scalars[0] != group->order)) { in ossl_ec_wNAF_mul()
463 return ossl_ec_scalar_mul_ladder(group, r, scalars[0], points[0], in ossl_ec_wNAF_mul()
507 * 'scalars' */ in ossl_ec_wNAF_mul()
536 bits = i < num ? BN_num_bits(scalars[i]) : BN_num_bits(scalar); in ossl_ec_wNAF_mul()
541 bn_compute_wNAF((i < num ? scalars[i] : scalar), wsize[i], in ossl_ec_wNAF_mul()
410 ossl_ec_wNAF_mul(const EC_GROUP *group, EC_POINT *r, const BIGNUM *scalar, size_t num, const EC_POINT *points[], const BIGNUM *scalars[], BN_CTX *ctx) ossl_ec_wNAF_mul() argument
H A Decp_nistp256.c1667 * small point multiples 0*P, 1*P, ..., 17*P are in pre_comp[], the scalars
1668 * in scalars[]. If g_scalar is non-NULL, we also add this multiple of the
1673 const felem_bytearray scalars[], in batch_mul()
1689 * Loop over all scalars msb-to-lsb, interleaving additions of multiples in batch_mul()
1735 /* loop over all scalars */ in batch_mul()
1737 bits = get_bit(scalars[num], i + 4) << 5; in batch_mul()
1738 bits |= get_bit(scalars[num], i + 3) << 4; in batch_mul()
1739 bits |= get_bit(scalars[num], i + 2) << 3; in batch_mul()
1740 bits |= get_bit(scalars[num], i + 1) << 2; in batch_mul()
1741 bits |= get_bit(scalars[nu in batch_mul()
1672 batch_mul(felem x_out, felem y_out, felem z_out, const felem_bytearray scalars[], const unsigned num_points, const u8 *g_scalar, const int mixed, const smallfelem pre_comp[][17][3], const smallfelem g_pre_comp[2][16][3]) batch_mul() argument
2023 ossl_ec_GFp_nistp256_points_mul(const EC_GROUP *group, EC_POINT *r, const BIGNUM *scalar, size_t num, const EC_POINT *points[], const BIGNUM *scalars[], BN_CTX *ctx) ossl_ec_GFp_nistp256_points_mul() argument
[all...]
H A Decp_nistp521.c1577 * small point multiples 0*P, 1*P, ..., 16*P are in pre_comp[], the scalars
1578 * in scalars[]. If g_scalar is non-NULL, we also add this multiple of the
1583 const felem_bytearray scalars[], in batch_mul()
1598 * Loop over all scalars msb-to-lsb, interleaving additions of multiples in batch_mul()
1631 /* loop over all scalars */ in batch_mul()
1633 bits = get_bit(scalars[num], i + 4) << 5; in batch_mul()
1634 bits |= get_bit(scalars[num], i + 3) << 4; in batch_mul()
1635 bits |= get_bit(scalars[num], i + 2) << 3; in batch_mul()
1636 bits |= get_bit(scalars[num], i + 1) << 2; in batch_mul()
1637 bits |= get_bit(scalars[nu in batch_mul()
1582 batch_mul(felem x_out, felem y_out, felem z_out, const felem_bytearray scalars[], const unsigned num_points, const u8 *g_scalar, const int mixed, const felem pre_comp[][17][3], const felem g_pre_comp[16][3]) batch_mul() argument
1916 ossl_ec_GFp_nistp521_points_mul(const EC_GROUP *group, EC_POINT *r, const BIGNUM *scalar, size_t num, const EC_POINT *points[], const BIGNUM *scalars[], BN_CTX *ctx) ossl_ec_GFp_nistp521_points_mul() argument
[all...]
H A Decp_nistp224.c85 * scalars for point multiplication.
466 * Multiply a field element by a scalar: out = out * scalar The scalars we
479 * scalars we actually use are small, so results fit without overflow
1134 * small point multiples 0*P, 1*P, ..., 16*P are in pre_comp[], the scalars
1135 * in scalars[]. If g_scalar is non-NULL, we also add this multiple of the
1140 const felem_bytearray scalars[], in batch_mul()
1156 * Loop over all scalars msb-to-lsb, interleaving additions of multiples in batch_mul()
1200 /* loop over all scalars */ in batch_mul()
1202 bits = get_bit(scalars[num], i + 4) << 5; in batch_mul()
1203 bits |= get_bit(scalars[nu in batch_mul()
1139 batch_mul(felem x_out, felem y_out, felem z_out, const felem_bytearray scalars[], const unsigned num_points, const u8 *g_scalar, const int mixed, const felem pre_comp[][17][3], const felem g_pre_comp[2][16][3]) batch_mul() argument
1411 ossl_ec_GFp_nistp224_points_mul(const EC_GROUP *group, EC_POINT *r, const BIGNUM *scalar, size_t num, const EC_POINT *points[], const BIGNUM *scalars[], BN_CTX *ctx) ossl_ec_GFp_nistp224_points_mul() argument
[all...]
/third_party/openssl/crypto/ec/
H A Decp_s390x_nistp.c50 const BIGNUM *scalars[], in ec_GFp_s390x_nistp_mul()
81 || (scalar == NULL && num == 1 && BN_is_negative(scalars[0]) == 0)) { in ec_GFp_s390x_nistp_mul()
88 scalar_ptr = scalars[0]; in ec_GFp_s390x_nistp_mul()
118 rc = ossl_ec_wNAF_mul(group, r, scalar, num, points, scalars, ctx); in ec_GFp_s390x_nistp_mul()
291 const BIGNUM *scalars[], \
295 scalars, ctx, \
47 ec_GFp_s390x_nistp_mul(const EC_GROUP *group, EC_POINT *r, const BIGNUM *scalar, size_t num, const EC_POINT *points[], const BIGNUM *scalars[], BN_CTX *ctx, unsigned int fc, int len) ec_GFp_s390x_nistp_mul() argument
H A Dec_local.h122 * + points[0] * scalars[0]
124 * + points[num-1] * scalars[num-1].
129 * should provide an input scalar (either scalar or scalars[0])
136 size_t num, const EC_POINT *points[], const BIGNUM *scalars[],
354 const BIGNUM *scalars[], BN_CTX *);
526 const EC_POINT *points[], const BIGNUM *scalars[],
531 const BIGNUM *scalars[], BN_CTX *ctx);
546 const EC_POINT *points[], const BIGNUM *scalars[],
551 const BIGNUM *scalars[], BN_CTX *ctx);
566 const EC_POINT *points[], const BIGNUM *scalars[],
[all...]
H A Decp_nistz256.c630 const BIGNUM **scalars = NULL; in ecp_nistz256_windowed_mul() local
639 || (scalars = OPENSSL_malloc(num * sizeof(BIGNUM *))) == NULL) { in ecp_nistz256_windowed_mul()
660 scalars[i] = mod; in ecp_nistz256_windowed_mul()
662 scalars[i] = scalar[i]; in ecp_nistz256_windowed_mul()
664 for (j = 0; j < bn_get_top(scalars[i]) * BN_BYTES; j += BN_BYTES) { in ecp_nistz256_windowed_mul()
665 BN_ULONG d = bn_get_words(scalars[i])[j / BN_BYTES]; in ecp_nistz256_windowed_mul()
785 OPENSSL_free(scalars); in ecp_nistz256_windowed_mul()
951 /* r = scalar*G + sum(scalars[i]*points[i]) */
957 const BIGNUM *scalars[], BN_CTX *ctx) in ecp_nistz256_points_mul()
1137 memcpy(new_scalars, scalars, nu in ecp_nistz256_points_mul()
952 ecp_nistz256_points_mul(const EC_GROUP *group, EC_POINT *r, const BIGNUM *scalar, size_t num, const EC_POINT *points[], const BIGNUM *scalars[], BN_CTX *ctx) ecp_nistz256_points_mul() argument
[all...]
H A Dec_mult.c405 * \sum scalars[i]*points[i],
412 const BIGNUM *scalars[], BN_CTX *ctx) in ossl_ec_wNAF_mul()
434 * treated like other scalars, i.e. in ossl_ec_wNAF_mul()
455 if ((scalar == NULL) && (num == 1) && (scalars[0] != group->order)) { in ossl_ec_wNAF_mul()
463 return ossl_ec_scalar_mul_ladder(group, r, scalars[0], points[0], in ossl_ec_wNAF_mul()
507 * 'scalars' */ in ossl_ec_wNAF_mul()
536 bits = i < num ? BN_num_bits(scalars[i]) : BN_num_bits(scalar); in ossl_ec_wNAF_mul()
541 bn_compute_wNAF((i < num ? scalars[i] : scalar), wsize[i], in ossl_ec_wNAF_mul()
410 ossl_ec_wNAF_mul(const EC_GROUP *group, EC_POINT *r, const BIGNUM *scalar, size_t num, const EC_POINT *points[], const BIGNUM *scalars[], BN_CTX *ctx) ossl_ec_wNAF_mul() argument
H A Decp_nistp256.c1667 * small point multiples 0*P, 1*P, ..., 17*P are in pre_comp[], the scalars
1668 * in scalars[]. If g_scalar is non-NULL, we also add this multiple of the
1673 const felem_bytearray scalars[], in batch_mul()
1689 * Loop over all scalars msb-to-lsb, interleaving additions of multiples in batch_mul()
1735 /* loop over all scalars */ in batch_mul()
1737 bits = get_bit(scalars[num], i + 4) << 5; in batch_mul()
1738 bits |= get_bit(scalars[num], i + 3) << 4; in batch_mul()
1739 bits |= get_bit(scalars[num], i + 2) << 3; in batch_mul()
1740 bits |= get_bit(scalars[num], i + 1) << 2; in batch_mul()
1741 bits |= get_bit(scalars[nu in batch_mul()
1672 batch_mul(felem x_out, felem y_out, felem z_out, const felem_bytearray scalars[], const unsigned num_points, const u8 *g_scalar, const int mixed, const smallfelem pre_comp[][17][3], const smallfelem g_pre_comp[2][16][3]) batch_mul() argument
2023 ossl_ec_GFp_nistp256_points_mul(const EC_GROUP *group, EC_POINT *r, const BIGNUM *scalar, size_t num, const EC_POINT *points[], const BIGNUM *scalars[], BN_CTX *ctx) ossl_ec_GFp_nistp256_points_mul() argument
[all...]
H A Decp_nistp521.c1577 * small point multiples 0*P, 1*P, ..., 16*P are in pre_comp[], the scalars
1578 * in scalars[]. If g_scalar is non-NULL, we also add this multiple of the
1583 const felem_bytearray scalars[], in batch_mul()
1598 * Loop over all scalars msb-to-lsb, interleaving additions of multiples in batch_mul()
1631 /* loop over all scalars */ in batch_mul()
1633 bits = get_bit(scalars[num], i + 4) << 5; in batch_mul()
1634 bits |= get_bit(scalars[num], i + 3) << 4; in batch_mul()
1635 bits |= get_bit(scalars[num], i + 2) << 3; in batch_mul()
1636 bits |= get_bit(scalars[num], i + 1) << 2; in batch_mul()
1637 bits |= get_bit(scalars[nu in batch_mul()
1582 batch_mul(felem x_out, felem y_out, felem z_out, const felem_bytearray scalars[], const unsigned num_points, const u8 *g_scalar, const int mixed, const felem pre_comp[][17][3], const felem g_pre_comp[16][3]) batch_mul() argument
1916 ossl_ec_GFp_nistp521_points_mul(const EC_GROUP *group, EC_POINT *r, const BIGNUM *scalar, size_t num, const EC_POINT *points[], const BIGNUM *scalars[], BN_CTX *ctx) ossl_ec_GFp_nistp521_points_mul() argument
[all...]
H A Decp_nistp224.c85 * scalars for point multiplication.
466 * Multiply a field element by a scalar: out = out * scalar The scalars we
479 * scalars we actually use are small, so results fit without overflow
1134 * small point multiples 0*P, 1*P, ..., 16*P are in pre_comp[], the scalars
1135 * in scalars[]. If g_scalar is non-NULL, we also add this multiple of the
1140 const felem_bytearray scalars[], in batch_mul()
1156 * Loop over all scalars msb-to-lsb, interleaving additions of multiples in batch_mul()
1200 /* loop over all scalars */ in batch_mul()
1202 bits = get_bit(scalars[num], i + 4) << 5; in batch_mul()
1203 bits |= get_bit(scalars[nu in batch_mul()
1139 batch_mul(felem x_out, felem y_out, felem z_out, const felem_bytearray scalars[], const unsigned num_points, const u8 *g_scalar, const int mixed, const felem pre_comp[][17][3], const felem g_pre_comp[2][16][3]) batch_mul() argument
1411 ossl_ec_GFp_nistp224_points_mul(const EC_GROUP *group, EC_POINT *r, const BIGNUM *scalar, size_t num, const EC_POINT *points[], const BIGNUM *scalars[], BN_CTX *ctx) ossl_ec_GFp_nistp224_points_mul() argument
[all...]
/third_party/openssl/ohos_lite/crypto/ec/
H A Dec_local.h126 * + points[0] * scalars[0]
128 * + points[num-1] * scalars[num-1].
133 * should provide an input scalar (either scalar or scalars[0])
140 size_t num, const EC_POINT *points[], const BIGNUM *scalars[],
342 size_t num, const EC_POINT *points[], const BIGNUM *scalars[],
505 const EC_POINT *points[], const BIGNUM *scalars[],
510 const BIGNUM *scalars[], BN_CTX *ctx);
525 const EC_POINT *points[], const BIGNUM *scalars[],
530 const BIGNUM *scalars[], BN_CTX *ctx);
545 const EC_POINT *points[], const BIGNUM *scalars[],
[all...]
/third_party/vk-gl-cts/modules/gles3/scripts/
H A Dgenutil.py739 def __init__ (self, numCols, numRows, scalars):
740 assert len(scalars) == numRows*numCols
743 self.scalars = scalars
747 scalars = []
750 scalars.append(scalar if col == row else 0.0)
751 return Mat(numCols, numRows, scalars)
760 return self.scalars[colNdx*self.numRows + rowNdx]
765 self.scalars[colNdx*self.numRows + rowNdx] = scalar
791 return "%s(%s)" % (self.typeString(), ", ".join(["%s" % s for s in self.scalars]))
[all...]
H A Dgen-conversions.py35 # - scalars types
82 scalars = reduce(operator.add, [x[ndx].toFloat().getScalars() for x in comps])
83 # print "->", scalars
84 res.append(Vec.fromScalarList(scalars))
105 scalars = reduce(operator.add, [x[ndx].toFloat().getScalars() for x in comps])
106 res.append(Mat(numCols, numRows, scalars))
/third_party/vk-gl-cts/modules/gles31/scripts/
H A Dgenutil.py741 def __init__ (self, numCols, numRows, scalars):
742 assert len(scalars) == numRows*numCols
745 self.scalars = scalars
749 scalars = []
752 scalars.append(scalar if col == row else 0.0)
753 return Mat(numCols, numRows, scalars)
762 return self.scalars[colNdx*self.numRows + rowNdx]
767 self.scalars[colNdx*self.numRows + rowNdx] = scalar
793 return "%s(%s)" % (self.typeString(), ", ".join(["%s" % s for s in self.scalars]))
[all...]
/third_party/mesa3d/src/asahi/lib/
H A Dagx_device.c405 uint64_t scalars[2] = { in agx_create_command_queue() local
412 scalars, 2, NULL, NULL); in agx_create_command_queue()
418 uint64_t scalars[2] = { in agx_create_command_queue() local
425 scalars, 2, NULL, NULL); in agx_create_command_queue()
/third_party/vk-gl-cts/external/openglcts/modules/common/
H A DglcShaderLibraryCase.cpp388 std::vector<float>& scalars = attribValues[attribValueNdx++]; in execute() local
389 scalars.resize(numVerticesPerDraw * scalarSize); in execute()
394 scalars[repNdx * scalarSize + ndx] = val.elements[arrayNdx * scalarSize + ndx].float32; in execute()
405 scalars[repNdx * scalarSize + ndx] = v; in execute()
435 &scalars[i * numRows])); in execute()
441 vertexArrays.push_back(va::Float(attribLoc, scalarSize, numVerticesPerDraw, 0, &scalars[0])); in execute()

Completed in 36 milliseconds

12