/external/eigen/test/ |
D | evaluators.cpp | 8 prod(const Lhs& lhs, const Rhs& rhs) in prod() function 158 VERIFY_IS_APPROX_EVALUATOR2(d, prod(a,b), a*b); in test_evaluators() 159 VERIFY_IS_APPROX_EVALUATOR2(d.noalias(), prod(a,b), a*b); in test_evaluators() 160 VERIFY_IS_APPROX_EVALUATOR2(d, prod(a,b) + c, a*b + c); in test_evaluators() 161 VERIFY_IS_APPROX_EVALUATOR2(d, s * prod(a,b), s * a*b); in test_evaluators() 162 VERIFY_IS_APPROX_EVALUATOR2(d, prod(a,b).transpose(), (a*b).transpose()); in test_evaluators() 163 VERIFY_IS_APPROX_EVALUATOR2(d, prod(a,b) + prod(b,c), a*b + b*c); in test_evaluators() 167 copy_using_evaluator(a, prod(a,a)); in test_evaluators() 172 add_assign_using_evaluator(c.noalias(), prod(a,b)); in test_evaluators() 177 subtract_assign_using_evaluator(c.noalias(), prod(a,b)); in test_evaluators() [all …]
|
D | zerosized.cpp | 17 VERIFY(m.prod()==1); in zeroReduction() 59 …Matrix<Scalar,Dynamic,Dynamic> prod = m.template block<MatrixType::RowsAtCompileTime,0>(0,0,rows,0… in zeroSizedMatrix() local 60 VERIFY(prod.rows()==rows && prod.cols()==cols); in zeroSizedMatrix() 61 VERIFY(prod.isZero()); in zeroSizedMatrix() 62 prod = m.template block<1,0>(0,0) * m.template block<0,1>(0,0); in zeroSizedMatrix() 63 VERIFY(prod.size()==1); in zeroSizedMatrix() 64 VERIFY(prod.isZero()); in zeroSizedMatrix()
|
/external/llvm/test/CodeGen/ARM/ |
D | neon-v8.1a.ll | 23 %prod = call <4 x i16> @llvm.arm.neon.vqrdmulh.v4i16(<4 x i16> %mhs, <4 x i16> %rhs) 24 %retval = call <4 x i16> @llvm.arm.neon.vqadds.v4i16(<4 x i16> %acc, <4 x i16> %prod) 31 %prod = call <8 x i16> @llvm.arm.neon.vqrdmulh.v8i16(<8 x i16> %mhs, <8 x i16> %rhs) 32 %retval = call <8 x i16> @llvm.arm.neon.vqadds.v8i16(<8 x i16> %acc, <8 x i16> %prod) 39 %prod = call <2 x i32> @llvm.arm.neon.vqrdmulh.v2i32(<2 x i32> %mhs, <2 x i32> %rhs) 40 %retval = call <2 x i32> @llvm.arm.neon.vqadds.v2i32(<2 x i32> %acc, <2 x i32> %prod) 47 %prod = call <4 x i32> @llvm.arm.neon.vqrdmulh.v4i32(<4 x i32> %mhs, <4 x i32> %rhs) 48 %retval = call <4 x i32> @llvm.arm.neon.vqadds.v4i32(<4 x i32> %acc, <4 x i32> %prod) 55 %prod = call <4 x i16> @llvm.arm.neon.vqrdmulh.v4i16(<4 x i16> %mhs, <4 x i16> %rhs) 56 %retval = call <4 x i16> @llvm.arm.neon.vqsubs.v4i16(<4 x i16> %acc, <4 x i16> %prod) [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/ARM/ |
D | neon-v8.1a.ll | 23 %prod = call <4 x i16> @llvm.arm.neon.vqrdmulh.v4i16(<4 x i16> %mhs, <4 x i16> %rhs) 24 %retval = call <4 x i16> @llvm.arm.neon.vqadds.v4i16(<4 x i16> %acc, <4 x i16> %prod) 31 %prod = call <8 x i16> @llvm.arm.neon.vqrdmulh.v8i16(<8 x i16> %mhs, <8 x i16> %rhs) 32 %retval = call <8 x i16> @llvm.arm.neon.vqadds.v8i16(<8 x i16> %acc, <8 x i16> %prod) 39 %prod = call <2 x i32> @llvm.arm.neon.vqrdmulh.v2i32(<2 x i32> %mhs, <2 x i32> %rhs) 40 %retval = call <2 x i32> @llvm.arm.neon.vqadds.v2i32(<2 x i32> %acc, <2 x i32> %prod) 47 %prod = call <4 x i32> @llvm.arm.neon.vqrdmulh.v4i32(<4 x i32> %mhs, <4 x i32> %rhs) 48 %retval = call <4 x i32> @llvm.arm.neon.vqadds.v4i32(<4 x i32> %acc, <4 x i32> %prod) 55 %prod = call <4 x i16> @llvm.arm.neon.vqrdmulh.v4i16(<4 x i16> %mhs, <4 x i16> %rhs) 56 %retval = call <4 x i16> @llvm.arm.neon.vqsubs.v4i16(<4 x i16> %acc, <4 x i16> %prod) [all …]
|
/external/skqp/src/core/ |
D | SkMathPriv.h | 103 unsigned prod = a*b + 1; in SkMulDiv255Trunc() local 104 return (prod + (prod >> 8)) >> 8; in SkMulDiv255Trunc() 113 unsigned prod = a*b + 255; in SkMulDiv255Ceiling() local 114 return (prod + (prod >> 8)) >> 8; in SkMulDiv255Ceiling() 119 static inline unsigned SkDiv255Round(unsigned prod) { in SkDiv255Round() argument 120 prod += 128; in SkDiv255Round() 121 return (prod + (prod >> 8)) >> 8; in SkDiv255Round()
|
/external/skia/src/core/ |
D | SkMathPriv.h | 103 unsigned prod = a*b + 1; in SkMulDiv255Trunc() local 104 return (prod + (prod >> 8)) >> 8; in SkMulDiv255Trunc() 113 unsigned prod = a*b + 255; in SkMulDiv255Ceiling() local 114 return (prod + (prod >> 8)) >> 8; in SkMulDiv255Ceiling() 119 static inline unsigned SkDiv255Round(unsigned prod) { in SkDiv255Round() argument 120 prod += 128; in SkDiv255Round() 121 return (prod + (prod >> 8)) >> 8; in SkDiv255Round()
|
/external/llvm/test/CodeGen/AArch64/ |
D | arm64-neon-v8.1a.ll | 32 %prod = call <4 x i16> @llvm.aarch64.neon.sqrdmulh.v4i16(<4 x i16> %mhs, <4 x i16> %rhs) 33 %retval = call <4 x i16> @llvm.aarch64.neon.sqadd.v4i16(<4 x i16> %acc, <4 x i16> %prod) 42 %prod = call <8 x i16> @llvm.aarch64.neon.sqrdmulh.v8i16(<8 x i16> %mhs, <8 x i16> %rhs) 43 %retval = call <8 x i16> @llvm.aarch64.neon.sqadd.v8i16(<8 x i16> %acc, <8 x i16> %prod) 52 %prod = call <2 x i32> @llvm.aarch64.neon.sqrdmulh.v2i32(<2 x i32> %mhs, <2 x i32> %rhs) 53 %retval = call <2 x i32> @llvm.aarch64.neon.sqadd.v2i32(<2 x i32> %acc, <2 x i32> %prod) 62 %prod = call <4 x i32> @llvm.aarch64.neon.sqrdmulh.v4i32(<4 x i32> %mhs, <4 x i32> %rhs) 63 %retval = call <4 x i32> @llvm.aarch64.neon.sqadd.v4i32(<4 x i32> %acc, <4 x i32> %prod) 72 %prod = call <4 x i16> @llvm.aarch64.neon.sqrdmulh.v4i16(<4 x i16> %mhs, <4 x i16> %rhs) 73 %retval = call <4 x i16> @llvm.aarch64.neon.sqsub.v4i16(<4 x i16> %acc, <4 x i16> %prod) [all …]
|
D | dp-3source.ll | 39 %prod = mul i64 %ext1, %ext2 40 %res = add i64 %acc, %prod 49 %prod = mul i64 %ext1, %ext2 50 %res = sub i64 %acc, %prod 59 %prod = mul i64 %ext1, %ext2 60 %res = add i64 %acc, %prod 69 %prod = mul i64 %ext1, %ext2 70 %res = sub i64 %acc, %prod 113 %prod = mul i32 %lhs, %rhs 114 %res = sub i32 0, %prod [all …]
|
D | fp-dp3.ll | 90 %prod = fmul float %b, %c 91 %sum = fadd float %a, %prod 102 %prod = fmul float %b, %c 103 %diff = fsub float %a, %prod 115 %prod = fmul float %b, %c 116 %diff = fsub float %nega, %prod 129 %prod = fmul float %b, %c 130 %sum = fadd float %nega, %prod 142 %prod = fmul float %b, %c 143 %sum = fadd float %a, %prod [all …]
|
D | mul-lohi.ll | 17 %prod = mul i128 %lhs, %rhs 18 ret i128 %prod 32 %prod = mul i128 %lhs, %rhs 33 ret i128 %prod 44 %prod = mul i128 %lhs, %rhs 45 ret i128 %prod
|
/external/skia/include/core/ |
D | SkMath.h | 60 unsigned prod = a*b + (1 << (shift - 1)); in SkMul16ShiftRound() local 61 return (prod + (prod >> shift)) >> shift; in SkMul16ShiftRound() 71 unsigned prod = a*b + 128; in SkMulDiv255Round() local 72 return (prod + (prod >> 8)) >> 8; in SkMulDiv255Round()
|
/external/skqp/include/core/ |
D | SkMath.h | 60 unsigned prod = a*b + (1 << (shift - 1)); in SkMul16ShiftRound() local 61 return (prod + (prod >> shift)) >> shift; in SkMul16ShiftRound() 71 unsigned prod = a*b + 128; in SkMulDiv255Round() local 72 return (prod + (prod >> 8)) >> 8; in SkMulDiv255Round()
|
/external/webrtc/webrtc/modules/audio_coding/codecs/isac/fix/source/ |
D | filters_neon.c | 26 int64_t prod = 0; in WebRtcIsacfix_AutocorrNeon() local 48 prod = vaddvq_s64(tmpb_v); in WebRtcIsacfix_AutocorrNeon() 50 prod = vget_lane_s64(vadd_s64(vget_low_s64(tmpb_v), vget_high_s64(tmpb_v)), in WebRtcIsacfix_AutocorrNeon() 54 temp = (uint32_t)(prod >> 31); in WebRtcIsacfix_AutocorrNeon() 57 r[0] = (int32_t)(prod >> scaling); in WebRtcIsacfix_AutocorrNeon() 94 prod = vaddvq_s64(tmpb_v); in WebRtcIsacfix_AutocorrNeon() 96 prod = vget_lane_s64(vadd_s64(vget_low_s64(tmpb_v), vget_high_s64(tmpb_v)), in WebRtcIsacfix_AutocorrNeon() 107 r[i] = (int32_t)((prod + prod_tail) >> scaling); in WebRtcIsacfix_AutocorrNeon()
|
D | filters.c | 27 int64_t prod = 0; in WebRtcIsacfix_AutocorrC() local 35 prod += x[i] * x[i]; in WebRtcIsacfix_AutocorrC() 39 temp = (uint32_t)(prod >> 31); in WebRtcIsacfix_AutocorrC() 45 r[0] = (int32_t)(prod >> scaling); in WebRtcIsacfix_AutocorrC() 49 prod = 0; in WebRtcIsacfix_AutocorrC() 51 prod += x[j] * x[i + j]; in WebRtcIsacfix_AutocorrC() 53 sum = (int32_t)(prod >> scaling); in WebRtcIsacfix_AutocorrC()
|
/external/toolchain-utils/deprecated/ |
D | compare_benchmarks.py | 147 prod = 1.0 155 prod = prod * (stats[key] / basestats[key]) 164 prod = math.exp(1.0 / count * math.log(prod)) 165 prod = (1.0 - prod) * 100 167 print('%s,,,%f' % ('Geomean', prod)) 169 print('%15.15s%15.15s%15.15s%14.2f%%' % ('Geomean', '', '', prod))
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/AArch64/ |
D | dp-3source.ll | 39 %prod = mul i64 %ext1, %ext2 40 %res = add i64 %acc, %prod 49 %prod = mul i64 %ext1, %ext2 50 %res = sub i64 %acc, %prod 59 %prod = mul i64 %ext1, %ext2 60 %res = add i64 %acc, %prod 69 %prod = mul i64 %ext1, %ext2 70 %res = sub i64 %acc, %prod 113 %prod = mul i32 %lhs, %rhs 114 %res = sub i32 0, %prod [all …]
|
D | arm64-neon-v8.1a.ll | 35 %prod = call <4 x i16> @llvm.aarch64.neon.sqrdmulh.v4i16(<4 x i16> %mhs, <4 x i16> %rhs) 36 %retval = call <4 x i16> @llvm.aarch64.neon.sqadd.v4i16(<4 x i16> %acc, <4 x i16> %prod) 45 %prod = call <8 x i16> @llvm.aarch64.neon.sqrdmulh.v8i16(<8 x i16> %mhs, <8 x i16> %rhs) 46 %retval = call <8 x i16> @llvm.aarch64.neon.sqadd.v8i16(<8 x i16> %acc, <8 x i16> %prod) 55 %prod = call <2 x i32> @llvm.aarch64.neon.sqrdmulh.v2i32(<2 x i32> %mhs, <2 x i32> %rhs) 56 %retval = call <2 x i32> @llvm.aarch64.neon.sqadd.v2i32(<2 x i32> %acc, <2 x i32> %prod) 65 %prod = call <4 x i32> @llvm.aarch64.neon.sqrdmulh.v4i32(<4 x i32> %mhs, <4 x i32> %rhs) 66 %retval = call <4 x i32> @llvm.aarch64.neon.sqadd.v4i32(<4 x i32> %acc, <4 x i32> %prod) 75 %prod = call <4 x i16> @llvm.aarch64.neon.sqrdmulh.v4i16(<4 x i16> %mhs, <4 x i16> %rhs) 76 %retval = call <4 x i16> @llvm.aarch64.neon.sqsub.v4i16(<4 x i16> %acc, <4 x i16> %prod) [all …]
|
D | fp-dp3.ll | 90 %prod = fmul float %b, %c 91 %sum = fadd float %a, %prod 102 %prod = fmul float %b, %c 103 %diff = fsub float %a, %prod 115 %prod = fmul float %b, %c 116 %diff = fsub float %nega, %prod 129 %prod = fmul float %b, %c 130 %sum = fadd float %nega, %prod 142 %prod = fmul float %b, %c 143 %sum = fadd float %a, %prod [all …]
|
D | mul-lohi.ll | 19 %prod = mul i128 %lhs, %rhs 20 ret i128 %prod 34 %prod = mul i128 %lhs, %rhs 35 ret i128 %prod 46 %prod = mul i128 %lhs, %rhs 47 ret i128 %prod
|
/external/python/cpython3/Parser/ |
D | asdl_c.py | 277 def visitProduct(self, prod, name): argument 279 self.get_args(prod.fields), 280 self.get_args(prod.attributes), 362 def visitProduct(self, prod, name): argument 451 def visitProduct(self, prod, name): argument 457 for f in prod.fields: 458 self.visitFieldDeclaration(f, name, prod=prod, depth=1) 459 for a in prod.attributes: 460 self.visitFieldDeclaration(a, name, prod=prod, depth=1) 462 for f in prod.fields: [all …]
|
/external/tensorflow/tensorflow/python/ops/ragged/ |
D | ragged_segment_op_test.py | 35 def prod(values): function 88 (ragged_math_ops.segment_prod, prod, [0, 0, 1, 1, 2, 2]), 89 (ragged_math_ops.segment_prod, prod, [0, 0, 0, 1, 1, 1]), 90 (ragged_math_ops.segment_prod, prod, [5, 4, 3, 2, 1, 0]), 91 (ragged_math_ops.segment_prod, prod, [0, 0, 0, 10, 10, 10]), 120 (ragged_math_ops.segment_prod, prod, [0, 0, 1, 1, 2, 2]), 121 (ragged_math_ops.segment_prod, prod, [0, 0, 0, 1, 1, 1]), 122 (ragged_math_ops.segment_prod, prod, [5, 4, 3, 2, 1, 0]), 123 (ragged_math_ops.segment_prod, prod, [0, 0, 0, 10, 10, 10]),
|
/external/eigen/bench/btl/libs/ublas/ |
D | ublas_interface.hh | 92 X = prod(A,B); in matrix_vector_product_slow() 96 X = prod(A,B); in matrix_matrix_product_slow() 106 X.assign(prod(A,B)); in matrix_vector_product() 110 X.assign(prod(trans(A),B)); in atv_product() 114 X.assign(prod(A,B)); in matrix_matrix_product() 127 X.assign(prod(trans(A),A)); in ata_product() 132 X.assign(prod(A,trans(A))); in aat_product()
|
/external/epid-sdk/ext/ipp/sources/ippcp/ |
D | gsmodmethodstuff.h | 198 static BNU_CHUNK_T* gs_mont_red(BNU_CHUNK_T* pr, BNU_CHUNK_T* prod, gsModEngine* pME) in gs_mont_red() argument 208 for(i=0; i<mLen; i++, prod++) { in gs_mont_red() 213 BNU_CHUNK_T u = prod[0] * k0; in gs_mont_red() 217 ADD_AB(ex_mu, t, prod[0], muL); in gs_mont_red() 223 ADD_AB(ex_mu, t, prod[j], ex_mu); /* carry in ex_mu,t */ in gs_mont_red() 224 ADD_AB(c, prod[j], t, muL); /* carry in c */ in gs_mont_red() 227 ADD_ABC(carry, prod[mLen], prod[mLen], ex_mu, carry); in gs_mont_red() 231 carry -= cpSub_BNU(pr, prod, pm, mLen); in gs_mont_red() 232 cpMaskMove_gs(pr, prod, mLen, cpIsNonZero(carry)); in gs_mont_red() 238 static BNU_CHUNK_T* gs_mont_red(BNU_CHUNK_T* pr, BNU_CHUNK_T* prod, gsModEngine* pME) in gs_mont_red() argument [all …]
|
/external/libchrome/third_party/ply/ |
D | yacc.py | 287 … prod = self.productions # Local reference to production list (to avoid lookup on self.) 378 p = prod[-t] 602 … prod = self.productions # Local reference to production list (to avoid lookup on self.) 675 p = prod[-t] 874 … prod = self.productions # Local reference to production list (to avoid lookup on self.) 947 p = prod[-t] 1154 def __init__(self,number,name,prod,precedence=('right',0),func=None,file='',line=0): argument 1156 self.prod = tuple(prod) 1166 self.len = len(self.prod) # Length of the production 1170 for s in self.prod: [all …]
|
/external/eigen/Eigen/src/Core/products/ |
D | GeneralMatrixMatrixTriangular.h | 202 …static void run(MatrixType& mat, const ProductType& prod, const typename MatrixType::Scalar& alpha… 210 … internal::add_const_on_value_type<ActualLhs>::type actualLhs = LhsBlasTraits::extract(prod.lhs()); 216 … internal::add_const_on_value_type<ActualRhs>::type actualRhs = RhsBlasTraits::extract(prod.rhs()); 218 …= alpha * LhsBlasTraits::extractScalarFactor(prod.lhs().derived()) * RhsBlasTraits::extractScalarF… 250 …static void run(MatrixType& mat, const ProductType& prod, const typename MatrixType::Scalar& alpha… 256 … internal::add_const_on_value_type<ActualLhs>::type actualLhs = LhsBlasTraits::extract(prod.lhs()); 262 … internal::add_const_on_value_type<ActualRhs>::type actualRhs = RhsBlasTraits::extract(prod.rhs()); 264 …= alpha * LhsBlasTraits::extractScalarFactor(prod.lhs().derived()) * RhsBlasTraits::extractScalarF… 299 …angularViewImpl<MatrixType,UpLo,Dense>::_assignProduct(const ProductType& prod, const Scalar& alph… 302 …eigen_assert(derived().nestedExpression().rows() == prod.rows() && derived().cols() == prod.cols()… [all …]
|