/external/llvm-project/mlir/test/Dialect/Affine/ |
D | dma.mlir | 11 %c64 = constant 64 : index 14 affine.dma_start %0[%i0, %i1], %1[%i0, %i1], %2[%c0], %c64 16 affine.dma_wait %2[%c0], %c64 : memref<1xi32> 32 %c64 = constant 64 : index 37 affine.dma_start %0[%i0, %i1], %1[%i0, %i1], %2[%c0], %c64, %c128, %c256 39 affine.dma_wait %2[%c0], %c64 : memref<1xi32> 55 %c64 = constant 64 : index 59 %2[%c0], %c64 61 affine.dma_wait %2[%c0], %c64 : memref<1xi32> 77 %c64 = constant 64 : index [all …]
|
D | load-store-invalid.mlir | 113 %c64 = constant 64 : index 117 affine.dma_start %0[%3], %1[%i0], %2[%c0], %c64 130 %c64 = constant 64 : index 134 affine.dma_start %0[%i0], %1[%3], %2[%c0], %c64 147 %c64 = constant 64 : index 151 affine.dma_start %0[%i0], %1[%arg0], %2[%3], %c64 164 %c64 = constant 64 : index 168 affine.dma_wait %2[%3], %c64 : memref<1xi32, 4>
|
/external/tensorflow/tensorflow/compiler/xla/service/gpu/tests/ |
D | gemm_rewrite_test.cc | 157 x = c64[2,2] parameter(0) 158 y = c64[2,2] parameter(1) 159 k = c64[] constant((3.0, 3.0)) 160 k_broadcast = c64[2, 2] broadcast(k), dimensions={} 161 dot_a = c64[2,2] dot(x, y), lhs_contracting_dims={1}, rhs_contracting_dims={0} 162 ROOT dot_a_multiplied = c64[2, 2] multiply(dot_a, k_broadcast) 171 ; CHECK-LABEL: ENTRY %AddDotsFunc (x: c64[2,2], y: c64[2,2]) -> c64[2,2] { 172 ; CHECK-NEXT: %x = c64[2,2]{1,0} parameter(0) 173 ; CHECK-NEXT: %y = c64[2,2]{1,0} parameter(1) 174 ; CHECK-NEXT: ROOT %custom-call = c64[2,2]{1,0} custom-call(%x, %y), custom_call_target="__cubla…
|
/external/libwebsockets/lib/tls/mbedtls/ |
D | lws-genaes.c | 168 int n, m, ret = -1, c64 = cek_bits / 64; in lws_genaes_rfc3394_wrap() local 200 memcpy(out + 8, in, 8 * c64); in lws_genaes_rfc3394_wrap() 217 memcpy(out, in + 8, 8 * c64); in lws_genaes_rfc3394_wrap() 229 for (m = 1; m <= c64; m++) { in lws_genaes_rfc3394_wrap() 236 out[7] ^= c64 * n + m; in lws_genaes_rfc3394_wrap() 247 uint8_t *r = out + (c64 - 1) * 8; in lws_genaes_rfc3394_wrap() 248 for (m = c64; m >= 1; m--) { in lws_genaes_rfc3394_wrap() 250 b[7] ^= c64 * n + m; in lws_genaes_rfc3394_wrap()
|
/external/llvm-project/mlir/integration_test/Dialect/Vector/CPU/ |
D | test-vector-distribute.mlir | 39 %c64 = constant 64 : index 40 %out = alloc(%c64) : memref<?xf32> 41 %in1 = call @alloc_1d_filled_inc_f32(%c64, %cf1) : (index, f32) -> memref<?xf32> 42 %in2 = call @alloc_1d_filled_inc_f32(%c64, %cf2) : (index, f32) -> memref<?xf32>
|
/external/llvm/test/CodeGen/X86/ |
D | 2007-05-15-maskmovq.ll | 6 define void @test(<1 x i64> %c64, <1 x i64> %mask1, i8* %P) { 9 %tmp6 = bitcast <1 x i64> %c64 to x86_mmx ; <x86_mmx> [#uses=1]
|
/external/tensorflow/tensorflow/compiler/xla/service/ |
D | hlo_parser_test.cc | 805 ENTRY %Fft (input: c64[8,32]) -> c64[8,32] { in CreateTestCases() 806 %input = c64[8,32]{1,0} parameter(0) in CreateTestCases() 807 ROOT %fft = c64[8,32]{1,0} fft(c64[8,32]{1,0} %input), fft_type=FFT, fft_length={32} in CreateTestCases() 817 ENTRY %Ifft2d (input: c64[5,8,32]) -> c64[5,8,32] { in CreateTestCases() 818 %input = c64[5,8,32]{2,1,0} parameter(0) in CreateTestCases() 819 ROOT %fft = c64[5,8,32]{2,1,0} fft(c64[5,8,32]{2,1,0} %input), fft_type=IFFT, fft_length={8,32} in CreateTestCases() 829 ENTRY %Rfft2d (input: f32[5,64,32]) -> c64[5,64,17] { in CreateTestCases() 831 ROOT %fft = c64[5,64,17]{2,1,0} fft(f32[5,64,32]{2,1,0} %input), fft_type=RFFT, fft_length={64,32} in CreateTestCases() 841 ENTRY %Irfft3d (input: c64[5,64,128,33]) -> f32[5,64,128,64] { in CreateTestCases() 842 %input = c64[5,64,128,33]{3,2,1,0} parameter(0) in CreateTestCases() [all …]
|
/external/llvm-project/llvm/test/CodeGen/X86/ |
D | 2007-05-15-maskmovq.ll | 4 define void @test(<1 x i64> %c64, <1 x i64> %mask1, i8* %P) { 29 %tmp6 = bitcast <1 x i64> %c64 to x86_mmx ; <x86_mmx> [#uses=1]
|
/external/libtextclassifier/abseil-cpp/absl/random/internal/ |
D | wide_multiply.h | 62 const uint64_t c64 = a32 * b32; in MultiplyU64ToU128() 69 return absl::MakeUint128(c64 + (c32a >> 32) + (c32b >> 32) + carry, in MultiplyU64ToU128()
|
/external/abseil-cpp/absl/random/internal/ |
D | wide_multiply.h | 62 const uint64_t c64 = a32 * b32; in MultiplyU64ToU128() 69 return absl::MakeUint128(c64 + (c32a >> 32) + (c32b >> 32) + carry, in MultiplyU64ToU128()
|
/external/webrtc/third_party/abseil-cpp/absl/random/internal/ |
D | wide_multiply.h | 62 const uint64_t c64 = a32 * b32; in MultiplyU64ToU128() 69 return absl::MakeUint128(c64 + (c32a >> 32) + (c32b >> 32) + carry, in MultiplyU64ToU128()
|
/external/rust/crates/grpcio-sys/grpc/third_party/abseil-cpp/absl/random/internal/ |
D | wide_multiply.h | 62 const uint64_t c64 = a32 * b32; in MultiplyU64ToU128() 69 return absl::MakeUint128(c64 + (c32a >> 32) + (c32b >> 32) + carry, in MultiplyU64ToU128()
|
/external/openscreen/third_party/abseil/src/absl/random/internal/ |
D | wide_multiply.h | 62 const uint64_t c64 = a32 * b32; in MultiplyU64ToU128() 69 return absl::MakeUint128(c64 + (c32a >> 32) + (c32b >> 32) + carry, in MultiplyU64ToU128()
|
/external/tensorflow/tensorflow/lite/kernels/ |
D | cast.cc | 96 copyCast(in, reinterpret_cast<std::complex<float>*>(out->data.c64), in copyToTensor() 130 context, reinterpret_cast<std::complex<float>*>(input->data.c64), in Eval()
|
/external/tensorflow/tensorflow/core/kernels/mlir_generated/ |
D | BUILD | 522 "c64", 577 "c64", 628 "c64", 662 "c64", 672 # TODO(b/162577610): Add bf16, c64 and c128. 719 "c64", 739 "c64",
|
/external/llvm-project/llvm/test/Transforms/SLPVectorizer/X86/ |
D | arith-mul.ll | 14 @c64 = common global [8 x i64] zeroinitializer, align 64 51 ; SSE-NEXT: store i64 [[R0]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i6… 52 ; SSE-NEXT: store i64 [[R1]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i6… 53 ; SSE-NEXT: store i64 [[R2]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i6… 54 ; SSE-NEXT: store i64 [[R3]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i6… 55 ; SSE-NEXT: store i64 [[R4]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i6… 56 ; SSE-NEXT: store i64 [[R5]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i6… 57 ; SSE-NEXT: store i64 [[R6]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i6… 58 ; SSE-NEXT: store i64 [[R7]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i6… 86 ; SLM-NEXT: store i64 [[R0]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i6… [all …]
|
D | arith-sub-ssat.ll | 12 @c64 = common global [8 x i64] zeroinitializer, align 64 54 ; SSE-NEXT: store i64 [[R0]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i6… 55 ; SSE-NEXT: store i64 [[R1]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i6… 56 ; SSE-NEXT: store i64 [[R2]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i6… 57 ; SSE-NEXT: store i64 [[R3]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i6… 58 ; SSE-NEXT: store i64 [[R4]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i6… 59 ; SSE-NEXT: store i64 [[R5]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i6… 60 ; SSE-NEXT: store i64 [[R6]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i6… 61 ; SSE-NEXT: store i64 [[R7]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i6… 89 ; SLM-NEXT: store i64 [[R0]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i6… [all …]
|
D | arith-add-ssat.ll | 12 @c64 = common global [8 x i64] zeroinitializer, align 64 54 ; SSE-NEXT: store i64 [[R0]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i6… 55 ; SSE-NEXT: store i64 [[R1]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i6… 56 ; SSE-NEXT: store i64 [[R2]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i6… 57 ; SSE-NEXT: store i64 [[R3]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i6… 58 ; SSE-NEXT: store i64 [[R4]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i6… 59 ; SSE-NEXT: store i64 [[R5]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i6… 60 ; SSE-NEXT: store i64 [[R6]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i6… 61 ; SSE-NEXT: store i64 [[R7]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i6… 89 ; SLM-NEXT: store i64 [[R0]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i6… [all …]
|
D | arith-abs.ll | 12 @c64 = common global [8 x i64] zeroinitializer, align 64 38 ; SSE-NEXT: store <2 x i64> [[TMP5]], <2 x i64>* bitcast ([8 x i64]* @c64 to <2 x i64>*), align 8 39 …], <2 x i64>* bitcast (i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i64 2) to <… 40 …], <2 x i64>* bitcast (i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i64 4) to <… 41 …], <2 x i64>* bitcast (i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i64 6) to <… 53 ; SLM-NEXT: store <2 x i64> [[TMP5]], <2 x i64>* bitcast ([8 x i64]* @c64 to <2 x i64>*), align 8 54 …], <2 x i64>* bitcast (i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i64 2) to <… 55 …], <2 x i64>* bitcast (i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i64 4) to <… 56 …], <2 x i64>* bitcast (i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i64 6) to <… 64 ; AVX-NEXT: store <4 x i64> [[TMP3]], <4 x i64>* bitcast ([8 x i64]* @c64 to <4 x i64>*), align 8 [all …]
|
D | shift-ashr.ll | 12 @c64 = common global [8 x i64] zeroinitializer, align 64 49 ; SSE-NEXT: store i64 [[R0]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i6… 50 ; SSE-NEXT: store i64 [[R1]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i6… 51 ; SSE-NEXT: store i64 [[R2]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i6… 52 ; SSE-NEXT: store i64 [[R3]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i6… 53 ; SSE-NEXT: store i64 [[R4]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i6… 54 ; SSE-NEXT: store i64 [[R5]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i6… 55 ; SSE-NEXT: store i64 [[R6]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i6… 56 ; SSE-NEXT: store i64 [[R7]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i6… 84 ; AVX1-NEXT: store i64 [[R0]], i64* getelementptr inbounds ([8 x i64], [8 x i64]* @c64, i32 0, i… [all …]
|
/external/rust/crates/crossbeam-channel/tests/ |
D | golang.rs | 439 let c64 = make::<i64>(buffer); in main() localVariable 449 recv(c64.rx()) -> _ => panic!("blocked i64sender"), in main() 506 go!(c64, sync, i64receiver(c64, sync)); in main() 510 send(c64.tx(), 123456) -> _ => break, in main() 522 go!(c64, sync, i64sender(c64, sync)); in main() 529 recv(c64.rx()) -> v => { in main()
|
/external/bouncycastle/bcprov/src/main/java/org/bouncycastle/math/ |
D | Primes.java | 500 long c64 = c & 0xFFFFFFFFL; in implSTRandomPrime() local 501 if (isPrime32(c64)) in implSTRandomPrime() 503 return new STOutput(BigInteger.valueOf(c64), primeSeed, primeGenCounter); in implSTRandomPrime()
|
/external/bouncycastle/repackaged/bcprov/src/main/java/com/android/org/bouncycastle/math/ |
D | Primes.java | 504 long c64 = c & 0xFFFFFFFFL; in implSTRandomPrime() local 505 if (isPrime32(c64)) in implSTRandomPrime() 507 return new STOutput(BigInteger.valueOf(c64), primeSeed, primeGenCounter); in implSTRandomPrime()
|
/external/bouncycastle/repackaged_platform/bcprov/src/main/java/com/android/internal/org/bouncycastle/math/ |
D | Primes.java | 504 long c64 = c & 0xFFFFFFFFL; in implSTRandomPrime() local 505 if (isPrime32(c64)) in implSTRandomPrime() 507 return new STOutput(BigInteger.valueOf(c64), primeSeed, primeGenCounter); in implSTRandomPrime()
|
/external/tensorflow/tensorflow/compiler/mlir/xla/tests/translate/ |
D | import.hlotxt | 160 %test_complex (Arg_0.1: f32[4], Arg_1.2: f32[4]) -> c64[4] { 165 ROOT %complex.3 = c64[4] complex(f32[4] %Arg_0.1, f32[4] %Arg_1.2) 196 %constant.4 = c64[] constant((1, 0)) 398 …ROOT %fft.2 = c64[3,5]{1,0} fft(%arg0.1), fft_type=RFFT, fft_length={9}, metadata={op_type="RFFT" … 441 %test_imag (Arg_0.1: c64[4]) -> f32[4] { 442 %Arg_0.1 = c64[4] parameter(0) 445 ROOT %imag.3 = f32[4] imag(c64[4] %Arg_0.1) 635 %test_real (Arg_0.1: c64[4]) -> f32[4] { 636 %Arg_0.1 = c64[4] parameter(0) 639 ROOT %real.3 = f32[4] real(c64[4] %Arg_0.1) [all …]
|