/external/rust/crates/quiche/deps/boringssl/src/crypto/poly1305/ |
D | poly1305_vec.c | 54 static inline uint128_t add128(uint128_t a, uint128_t b) { return a + b; } in add128() function 176 d[0] = add128(mul64x64_128(r20, r20), mul64x64_128(r21 * 2, s22)); in poly1305_first_block() 177 d[1] = add128(mul64x64_128(r22, s22), mul64x64_128(r20 * 2, r21)); in poly1305_first_block() 178 d[2] = add128(mul64x64_128(r21, r21), mul64x64_128(r22 * 2, r20)); in poly1305_first_block() 773 d[0] = add128(add128(mul64x64_128(h0, r0), mul64x64_128(h1, s2)), in CRYPTO_poly1305_finish() 775 d[1] = add128(add128(mul64x64_128(h0, r1), mul64x64_128(h1, r0)), in CRYPTO_poly1305_finish() 777 d[2] = add128(add128(mul64x64_128(h0, r2), mul64x64_128(h1, r1)), in CRYPTO_poly1305_finish()
|
/external/rust/crates/ring/crypto/poly1305/ |
D | poly1305_vec.c | 58 static inline uint128_t add128(uint128_t a, uint128_t b) { return a + b; } in add128() function 184 d[0] = add128(mul64x64_128(r20, r20), mul64x64_128(r21 * 2, s22)); in poly1305_first_block() 185 d[1] = add128(mul64x64_128(r22, s22), mul64x64_128(r20 * 2, r21)); in poly1305_first_block() 186 d[2] = add128(mul64x64_128(r21, r21), mul64x64_128(r22 * 2, r20)); in poly1305_first_block() 781 d[0] = add128(add128(mul64x64_128(h0, r0), mul64x64_128(h1, s2)), in GFp_poly1305_finish() 783 d[1] = add128(add128(mul64x64_128(h0, r1), mul64x64_128(h1, r0)), in GFp_poly1305_finish() 785 d[2] = add128(add128(mul64x64_128(h0, r2), mul64x64_128(h1, r1)), in GFp_poly1305_finish()
|
/external/boringssl/src/crypto/poly1305/ |
D | poly1305_vec.c | 54 static inline uint128_t add128(uint128_t a, uint128_t b) { return a + b; } in add128() function 180 d[0] = add128(mul64x64_128(r20, r20), mul64x64_128(r21 * 2, s22)); in poly1305_first_block() 181 d[1] = add128(mul64x64_128(r22, s22), mul64x64_128(r20 * 2, r21)); in poly1305_first_block() 182 d[2] = add128(mul64x64_128(r21, r21), mul64x64_128(r22 * 2, r20)); in poly1305_first_block() 777 d[0] = add128(add128(mul64x64_128(h0, r0), mul64x64_128(h1, s2)), in CRYPTO_poly1305_finish() 779 d[1] = add128(add128(mul64x64_128(h0, r1), mul64x64_128(h1, r0)), in CRYPTO_poly1305_finish() 781 d[2] = add128(add128(mul64x64_128(h0, r2), mul64x64_128(h1, r1)), in CRYPTO_poly1305_finish()
|
/external/llvm-project/llvm/test/Analysis/CostModel/AArch64/ |
D | sve-fixed-length.ll | 30 ; CHECK: cost of [[#div(127,VBITS)+1]] for instruction: %add128 = add <4 x i32> undef, undef 35 %add128 = add <4 x i32> undef, undef
|
/external/llvm-project/llvm/test/CodeGen/WebAssembly/ |
D | i128.ll | 12 ; CHECK-LABEL: add128: 13 ; CHECK-NEXT: .functype add128 (i32, i64, i64, i64, i64) -> (){{$}} 20 define i128 @add128(i128 %x, i128 %y) {
|
/external/llvm-project/llvm/test/CodeGen/AArch64/ |
D | shrink-wrap.ll | 134 %add128 = add nsw i32 %shr127, 0 135 %add134 = add nsw i32 0, %add128
|
/external/llvm/test/CodeGen/WebAssembly/ |
D | i128.ll | 12 ; CHECK-LABEL: add128: 20 define i128 @add128(i128 %x, i128 %y) {
|
/external/llvm/test/CodeGen/AArch64/ |
D | shrink-wrap.ll | 134 %add128 = add nsw i32 %shr127, 0 135 %add134 = add nsw i32 0, %add128
|
/external/llvm-project/llvm/test/CodeGen/X86/ |
D | misched-matmul.ll | 102 %add128.i = fadd double %mul122.i, %mul127.i 104 %add134.i = fadd double %add128.i, %mul133.i
|
D | addcarry.ll | 7 define i128 @add128(i128 %a, i128 %b) nounwind { 8 ; CHECK-LABEL: add128:
|
/external/llvm/test/CodeGen/X86/ |
D | misched-matmul.ll | 102 %add128.i = fadd double %mul122.i, %mul127.i 104 %add134.i = fadd double %add128.i, %mul133.i
|
/external/OpenCL-CTS/test_conformance/math_brute_force/ |
D | reference_math.cpp | 2767 static inline void add128(cl_ulong *hi, cl_ulong *lo, cl_ulong chi, in add128() function 2939 add128(&hi, &lo, chi, clo, &exponent); in reference_fmal() 2950 add128(&chi, &clo, hi, lo, &expC); in reference_fmal()
|