/external/libvpx/libvpx/vp9/common/mips/msa/ |
D | vp9_idct16x16_msa.c | 21 int16_t *out_ptr = &out[0]; in vp9_iht16x16_256_add_msa() local 28 vpx_idct16_1d_rows_msa((input + (i << 7)), (out_ptr + (i << 7))); in vp9_iht16x16_256_add_msa() 34 vpx_idct16_1d_columns_addblk_msa((out_ptr + (i << 3)), (dst + (i << 3)), in vp9_iht16x16_256_add_msa() 42 vpx_idct16_1d_rows_msa((input + (i << 7)), (out_ptr + (i << 7))); in vp9_iht16x16_256_add_msa() 47 vpx_iadst16_1d_columns_addblk_msa((out_ptr + (i << 3)), in vp9_iht16x16_256_add_msa() 55 vpx_iadst16_1d_rows_msa((input + (i << 7)), (out_ptr + (i << 7))); in vp9_iht16x16_256_add_msa() 61 vpx_idct16_1d_columns_addblk_msa((out_ptr + (i << 3)), (dst + (i << 3)), in vp9_iht16x16_256_add_msa() 69 vpx_iadst16_1d_rows_msa((input + (i << 7)), (out_ptr + (i << 7))); in vp9_iht16x16_256_add_msa() 74 vpx_iadst16_1d_columns_addblk_msa((out_ptr + (i << 3)), in vp9_iht16x16_256_add_msa()
|
/external/llvm-project/llvm/test/CodeGen/AMDGPU/ |
D | schedule-xdl-resource.ll | 28 %out_ptr.1 = getelementptr <32 x float>, <32 x float> addrspace(1)* %out, i32 %stride 29 %out_ptr.2 = getelementptr <32 x float>, <32 x float> addrspace(1)* %out_ptr.1, i32 %stride 30 %out_ptr.3 = getelementptr <32 x float>, <32 x float> addrspace(1)* %out_ptr.2, i32 %stride 31 %out_ptr.4 = getelementptr <32 x float>, <32 x float> addrspace(1)* %out_ptr.3, i32 %stride 32 %out_ptr.5 = getelementptr <32 x float>, <32 x float> addrspace(1)* %out_ptr.4, i32 %stride 33 %out_ptr.6 = getelementptr <32 x float>, <32 x float> addrspace(1)* %out_ptr.5, i32 %stride 34 store <32 x float> %mai.1, <32 x float> addrspace(1)* %out_ptr.1 35 store <32 x float> %mai.2, <32 x float> addrspace(1)* %out_ptr.2 36 store <32 x float> %mai.3, <32 x float> addrspace(1)* %out_ptr.3 37 store <32 x float> %mai.4, <32 x float> addrspace(1)* %out_ptr.4 [all …]
|
D | load-constant-f32.ll | 9 …ddrspace(4)* noalias nocapture readonly %weights, float addrspace(1)* noalias nocapture %out_ptr) { 11 %out_ptr.promoted = load float, float addrspace(1)* %out_ptr, align 4 13 %add = fadd float %tmp, %out_ptr.promoted 35 store float %add.7, float addrspace(1)* %out_ptr, align 4
|
D | load-constant-f64.ll | 20 …drspace(4)* noalias nocapture readonly %weights, double addrspace(1)* noalias nocapture %out_ptr) { 22 %out_ptr.promoted = load double, double addrspace(1)* %out_ptr, align 4 24 %add = fadd double %tmp, %out_ptr.promoted 46 store double %add.7, double addrspace(1)* %out_ptr, align 4
|
/external/libxaac/decoder/ |
D | ixheaacd_aac_imdct.h | 28 VOID ixheaacd_post_twiddle_dec(WORD32 out_ptr[], WORD32 spec_data[], 33 WORD32 out_ptr[], WORD32 spec_data[], 37 WORD32 out_ptr[], WORD32 spec_data[], 59 WORD32 *spec_data1, WORD32 *spec_data2, WORD32 *out_ptr, 64 WORD32 *spec_data1, WORD32 *spec_data2, WORD32 *out_ptr, 69 WORD32 *spec_data1, WORD32 *spec_data2, WORD32 *out_ptr,
|
D | ixheaacd_acelp_mdct.c | 138 WORD32 *out_ptr = &out[2 * nlength - 1]; in ixheaacd_post_twid() local 142 out_ptr[0] = -ixheaacd_mul_add64_sat_32(data_re[i], data_im[i], sin_ptr[i], in ixheaacd_post_twid() 147 out_ptr[-2] = -ixheaacd_mul_add64_sat_32(data_re[i + 1], data_im[i + 1], in ixheaacd_post_twid() 152 out_ptr[-4] = -ixheaacd_mul_add64_sat_32(data_re[i + 2], data_im[i + 2], in ixheaacd_post_twid() 157 out_ptr[-6] = -ixheaacd_mul_add64_sat_32(data_re[i + 3], data_im[i + 3], in ixheaacd_post_twid() 160 out_ptr -= 8; in ixheaacd_post_twid()
|
D | ixheaacd_aac_imdct.c | 129 WORD32 *spec_data1, WORD32 *spec_data2, WORD32 *out_ptr, in ixheaacd_pretwiddle_compute_dec() argument 136 WORD32 *out_ptr1 = out_ptr + (npoints2 << 1) - 1; in ixheaacd_pretwiddle_compute_dec() 153 *out_ptr = in ixheaacd_pretwiddle_compute_dec() 156 *out_ptr = ixheaacd_shl32(*out_ptr, neg_expo); in ixheaacd_pretwiddle_compute_dec() 157 out_ptr++; in ixheaacd_pretwiddle_compute_dec() 159 *out_ptr = ixheaacd_sub32(ixheaacd_mult32x16in32(tempi, cos), in ixheaacd_pretwiddle_compute_dec() 162 *out_ptr = ixheaacd_shl32(*out_ptr, neg_expo); in ixheaacd_pretwiddle_compute_dec() 163 out_ptr++; in ixheaacd_pretwiddle_compute_dec() 191 *out_ptr = in ixheaacd_pretwiddle_compute_dec() 193 *out_ptr = ixheaacd_shl32(*out_ptr, neg_expo); in ixheaacd_pretwiddle_compute_dec() [all …]
|
D | ixheaacd_hbe_trans.c | 745 FLOAT32 *out_ptr = &ptr_hbe_txposer->qmf_out_buf[1][2 * qmf_band_idx]; in ixheaacd_hbe_post_anal_prod2() local 764 *out_ptr++ += in ixheaacd_hbe_post_anal_prod2() 767 *out_ptr++ += in ixheaacd_hbe_post_anal_prod2() 771 out_ptr += 126; in ixheaacd_hbe_post_anal_prod2() 775 out_ptr -= 128 * 8; in ixheaacd_hbe_post_anal_prod2() 778 out_ptr -= (128 * 2 * qmf_voc_columns) - 2; in ixheaacd_hbe_post_anal_prod2() 1020 FLOAT32 *out_ptr = &ptr_hbe_txposer->qmf_out_buf[3][2 * qmf_band_idx]; in ixheaacd_hbe_post_anal_prod4() local 1056 *out_ptr++ += (temp_r * 0.6666667f); in ixheaacd_hbe_post_anal_prod4() 1057 *out_ptr++ += (temp_i * 0.6666667f); in ixheaacd_hbe_post_anal_prod4() 1060 out_ptr += 126; in ixheaacd_hbe_post_anal_prod4() [all …]
|
/external/webrtc/modules/rtp_rtcp/source/ |
D | rtp_format_video_generic.cc | 59 uint8_t* out_ptr = in NextPacket() local 61 RTC_CHECK(out_ptr); in NextPacket() 64 memcpy(out_ptr, header_, header_size_); in NextPacket() 69 memcpy(out_ptr + header_size_, remaining_payload_.data(), in NextPacket()
|
/external/tensorflow/tensorflow/lite/kernels/internal/optimized/ |
D | batch_matmul.h | 104 float* out_ptr = output_data + ((b0 * batch_dim1 * batch_dim2) + in BatchMatMul() local 109 dst_params, out_ptr, gemm_params, context); in BatchMatMul() 223 float* out_ptr = output_data + ((b0 * batch_dim1 * batch_dim2) + in BatchMatMul() local 252 const float32x4_t result0 = vmlaq_f32(vld1q_f32(out_ptr + idx), in BatchMatMul() 254 const float32x4_t result1 = vmlaq_f32(vld1q_f32(out_ptr + idx + 4), in BatchMatMul() 256 vst1q_f32(out_ptr + idx, result0); in BatchMatMul() 257 vst1q_f32(out_ptr + idx + 4, result1); in BatchMatMul() 263 out_ptr[idx] += batch_scaling_factor * accum_scratch[idx]; in BatchMatMul() 361 int8_t* out_ptr = output_data + ((b0 * batch_dim1 * batch_dim2) + in BatchMatMul() local 371 dst_params, out_ptr, gemm_params, context); in BatchMatMul()
|
/external/rust/crates/quiche/deps/boringssl/src/crypto/bytestring/ |
D | cbs.c | 55 int CBS_stow(const CBS *cbs, uint8_t **out_ptr, size_t *out_len) { in CBS_stow() argument 56 OPENSSL_free(*out_ptr); in CBS_stow() 57 *out_ptr = NULL; in CBS_stow() 63 *out_ptr = OPENSSL_memdup(cbs->data, cbs->len); in CBS_stow() 64 if (*out_ptr == NULL) { in CBS_stow() 71 int CBS_strdup(const CBS *cbs, char **out_ptr) { in CBS_strdup() argument 72 if (*out_ptr != NULL) { in CBS_strdup() 73 OPENSSL_free(*out_ptr); in CBS_strdup() 75 *out_ptr = OPENSSL_strndup((const char*)cbs->data, cbs->len); in CBS_strdup() 76 return (*out_ptr != NULL); in CBS_strdup()
|
/external/libgav1/libgav1/src/ |
D | decoder.cc | 66 Libgav1Decoder* decoder, const Libgav1DecoderBuffer** out_ptr) { in Libgav1DecoderDequeueFrame() argument 68 return cxx_decoder->DequeueFrame(out_ptr); in Libgav1DecoderDequeueFrame() 102 StatusCode Decoder::DequeueFrame(const DecoderBuffer** out_ptr) { in DequeueFrame() argument 104 return impl_->DequeueFrame(out_ptr); in DequeueFrame()
|
/external/libvpx/libvpx/vp9/encoder/mips/msa/ |
D | vp9_fdct16x16_msa.c | 88 int16_t *out_ptr = out + 128; in fadst16_cols_step2_msa() local 110 ST_SH(out5, (out_ptr + 4 * 16)); in fadst16_cols_step2_msa() 117 ST_SH(out13, (out_ptr + 5 * 16)); in fadst16_cols_step2_msa() 127 ST_SH(out1, (out_ptr + 7 * 16)); in fadst16_cols_step2_msa() 135 ST_SH(out9, (out_ptr + 6 * 16)); in fadst16_cols_step2_msa() 142 ST_SH(out3, (out_ptr)); in fadst16_cols_step2_msa() 146 ST_SH(out7, (out_ptr + 3 * 16)); in fadst16_cols_step2_msa() 150 ST_SH(out11, (out_ptr + 16)); in fadst16_cols_step2_msa() 154 ST_SH(out15, (out_ptr + 2 * 16)); in fadst16_cols_step2_msa() 266 int16_t *out_ptr = out + 8; in fadst16_rows_step2_msa() local [all …]
|
/external/boringssl/src/crypto/bytestring/ |
D | cbs.c | 55 int CBS_stow(const CBS *cbs, uint8_t **out_ptr, size_t *out_len) { in CBS_stow() argument 56 OPENSSL_free(*out_ptr); in CBS_stow() 57 *out_ptr = NULL; in CBS_stow() 63 *out_ptr = OPENSSL_memdup(cbs->data, cbs->len); in CBS_stow() 64 if (*out_ptr == NULL) { in CBS_stow() 71 int CBS_strdup(const CBS *cbs, char **out_ptr) { in CBS_strdup() argument 72 if (*out_ptr != NULL) { in CBS_strdup() 73 OPENSSL_free(*out_ptr); in CBS_strdup() 75 *out_ptr = OPENSSL_strndup((const char*)cbs->data, cbs->len); in CBS_strdup() 76 return (*out_ptr != NULL); in CBS_strdup()
|
/external/libjpeg-turbo/simd/arm/ |
D | jquanti-neon.c | 97 JCOEFPTR out_ptr = coef_block; in jsimd_quantize_neon() local 185 vst1q_s16(out_ptr + (i + 0) * DCTSIZE, row0); in jsimd_quantize_neon() 186 vst1q_s16(out_ptr + (i + 1) * DCTSIZE, row1); in jsimd_quantize_neon() 187 vst1q_s16(out_ptr + (i + 2) * DCTSIZE, row2); in jsimd_quantize_neon() 188 vst1q_s16(out_ptr + (i + 3) * DCTSIZE, row3); in jsimd_quantize_neon()
|
/external/tensorflow/tensorflow/core/kernels/ |
D | roll_op.cc | 208 T* out_ptr = &output[0]; in DoRollWithMemcpy() local 210 out_ptr += start; in DoRollWithMemcpy() 233 out_ptr += (out_indx - indx) * stride; in DoRollWithMemcpy() 255 memcpy(out_ptr, in_ptr, group_size * sizeof(T)); in DoRollWithMemcpy() 259 out_ptr += group_size; in DoRollWithMemcpy() 277 out_ptr -= dim_range[j]; // now wraps around in DoRollWithMemcpy() 281 out_ptr += dim_range[j]; // indx became 0 so reverse wrap around in DoRollWithMemcpy()
|
D | reverse_op.cc | 55 T* out_ptr = result->bit_casted_tensor<T, 3>().data(); in ReverseRows() local 58 out_ptr += start * row_size; in ReverseRows() 61 out_ptr += row_size; in ReverseRows() 64 out_ptr -= inner_size; in ReverseRows() 65 memcpy(out_ptr, in_ptr, inner_size * sizeof(T)); in ReverseRows() 70 out_ptr += row_size; in ReverseRows()
|
/external/webrtc/common_audio/signal_processing/ |
D | filter_ma_fast_q12.c | 23 int16_t* out_ptr, in WebRtcSpl_FilterMAFastQ12() argument 52 *out_ptr++ = (int16_t)((o + (int32_t)2048) >> 12); in WebRtcSpl_FilterMAFastQ12()
|
/external/tensorflow/tensorflow/lite/kernels/internal/reference/ |
D | batch_matmul.h | 89 float* out_ptr = output_data + ((b0 * batch_dim1 * batch_dim2) + in BatchMatMul() local 100 out_ptr[idx] = total; in BatchMatMul() 175 float* out_ptr = output_data + ((b0 * batch_dim1 * batch_dim2) + in BatchMatMul() local 190 out_ptr[idx] += batch_scaling_factor * total; in BatchMatMul() 245 T* out_ptr = output_data + in BatchMatMul() local 263 out_ptr[idx] = static_cast<T>(total_scaled); in BatchMatMul()
|
D | pad.h | 87 T* out_ptr = output_data; in PadImpl() local 100 *out_ptr++ = pad_value; in PadImpl() 102 *out_ptr++ = *in_ptr++; in PadImpl()
|
/external/libtextclassifier/native/tensorflow_models/seq_flow_lite/tflite_ops/ |
D | layer_norm.cc | 208 uint8_t* out_ptr = output->data.uint8; in FlexibleLayerNorm() local 225 out_ptr[input_offset] = in FlexibleLayerNorm() 238 float* out_ptr = output->data.f; in DefaultLayerNormFloat() local 254 out_ptr[index] = input->data.f[index] * multiplier + bias; in DefaultLayerNormFloat() 270 uint8_t* out_ptr = output->data.uint8; in DefaultLayerNorm() local 287 out_ptr[index] = PodQuantize(value, out_zero_point, out_inverse_scale); in DefaultLayerNorm()
|
/external/tensorflow/tensorflow/compiler/xla/python/ |
D | custom_call_for_test.pyx | 9 cdef void test_subtract_f32(void* out_ptr, void** data_ptr) nogil: argument 12 cdef float* out = <float*>(out_ptr)
|
/external/libgav1/libgav1/src/gav1/ |
D | decoder.h | 55 Libgav1Decoder* decoder, const Libgav1DecoderBuffer** out_ptr); 122 StatusCode DequeueFrame(const DecoderBuffer** out_ptr);
|
/external/llvm/test/Transforms/InstCombine/ |
D | obfuscated_splat.ll | 3 define void @test(<4 x float> *%in_ptr, <4 x float> *%out_ptr) { 9 store <4 x float> %D, <4 x float> *%out_ptr
|
/external/llvm-project/llvm/test/Transforms/InstCombine/ |
D | obfuscated_splat.ll | 3 define void @test(<4 x float> *%in_ptr, <4 x float> *%out_ptr) { 9 store <4 x float> %D, <4 x float> *%out_ptr
|