/external/libvpx/libvpx/vp8/encoder/mips/msa/ |
D | encodeopt_msa.c | 18 v8i16 coeff, dq_coeff, coeff0, coeff1; in vp8_block_error_msa() local 26 ILVRL_H2_SH(coeff, dq_coeff, coeff0, coeff1); in vp8_block_error_msa() 27 HSUB_UH2_SW(coeff0, coeff1, diff0, diff1); in vp8_block_error_msa() 47 v8i16 coeff, coeff0, coeff1, coeff2, coeff3, coeff4; in vp8_mbblock_error_msa() local 79 ILVRL_H2_SH(coeff, dq_coeff, coeff0, coeff1); in vp8_mbblock_error_msa() 80 HSUB_UH2_SW(coeff0, coeff1, diff0, diff1); in vp8_mbblock_error_msa() 83 ILVRL_H2_SH(coeff2, dq_coeff2, coeff0, coeff1); in vp8_mbblock_error_msa() 84 HSUB_UH2_SW(coeff0, coeff1, diff0, diff1); in vp8_mbblock_error_msa() 91 ILVRL_H2_SH(coeff3, dq_coeff3, coeff0, coeff1); in vp8_mbblock_error_msa() 92 HSUB_UH2_SW(coeff0, coeff1, diff0, diff1); in vp8_mbblock_error_msa() [all …]
|
D | quantize_msa.c | 24 v8i16 coeff0, coeff1, z0, z1; in fast_quantize_b_msa() local 35 LD_SH2(coeff_ptr, 8, coeff0, coeff1); in fast_quantize_b_msa() 36 VSHF_H2_SH(coeff0, coeff1, coeff0, coeff1, zigzag_mask0, zigzag_mask1, z0, in fast_quantize_b_msa() 38 LD_SH2(round, 8, coeff0, coeff1); in fast_quantize_b_msa() 39 VSHF_H2_SH(coeff0, coeff1, coeff0, coeff1, zigzag_mask0, zigzag_mask1, round0, in fast_quantize_b_msa() 41 LD_SH2(quant, 8, coeff0, coeff1); in fast_quantize_b_msa() 42 VSHF_H2_SH(coeff0, coeff1, coeff0, coeff1, zigzag_mask0, zigzag_mask1, quant0, in fast_quantize_b_msa() 93 v8i16 coeff0, coeff1, z0, z1; in exact_regular_quantize_b_msa() local 105 LD_SH2(coeff_ptr, 8, coeff0, coeff1); in exact_regular_quantize_b_msa() 106 VSHF_H2_SH(coeff0, coeff1, coeff0, coeff1, zigzag_mask0, zigzag_mask1, z0, in exact_regular_quantize_b_msa() [all …]
|
/external/libaom/libaom/aom_dsp/x86/ |
D | quantize_sse2.c | 32 __m128i coeff0, coeff1, coeff0_sign, coeff1_sign; in aom_quantize_b_sse2() local 45 coeff1 = load_coefficients(coeff_ptr + 8); in aom_quantize_b_sse2() 49 coeff1_sign = _mm_srai_epi16(coeff1, 15); in aom_quantize_b_sse2() 51 qcoeff1 = invert_sign_sse2(coeff1, coeff1_sign); in aom_quantize_b_sse2() 78 coeff1 = calculate_dqcoeff(qcoeff1, dequant); in aom_quantize_b_sse2() 81 store_coefficients(coeff1, dqcoeff_ptr + 8); in aom_quantize_b_sse2() 84 scan_for_eob(&coeff0, &coeff1, cmp_mask0, cmp_mask1, iscan_ptr, 0, zero); in aom_quantize_b_sse2() 89 coeff1 = load_coefficients(coeff_ptr + index + 8); in aom_quantize_b_sse2() 92 coeff1_sign = _mm_srai_epi16(coeff1, 15); in aom_quantize_b_sse2() 94 qcoeff1 = invert_sign_sse2(coeff1, coeff1_sign); in aom_quantize_b_sse2() [all …]
|
D | highbd_adaptive_quantize_avx2.c | 100 __m256i coeff1, in highbd_store_coefficients_avx2() argument 103 _mm256_store_si256((__m256i *)(coeff_ptr + 8), coeff1); in highbd_store_coefficients_avx2() 119 __m256i coeff0, qcoeff0, coeff1, qcoeff1; in aom_highbd_quantize_b_adaptive_avx2() local 147 coeff1 = _mm256_load_si256((__m256i *)(coeff_ptr + 8)); in aom_highbd_quantize_b_adaptive_avx2() 148 qcoeff1 = _mm256_abs_epi32(coeff1); in aom_highbd_quantize_b_adaptive_avx2() 174 qcoeff1 = _mm256_sign_epi32(qcoeff1, coeff1); in aom_highbd_quantize_b_adaptive_avx2() 181 coeff1 = highbd_calculate_dqcoeff_avx2(qcoeff1, dequant); in aom_highbd_quantize_b_adaptive_avx2() 182 highbd_store_coefficients_avx2(coeff0, coeff1, dqcoeff_ptr); in aom_highbd_quantize_b_adaptive_avx2() 189 coeff1 = _mm256_load_si256((__m256i *)(coeff_ptr + index + 8)); in aom_highbd_quantize_b_adaptive_avx2() 190 qcoeff1 = _mm256_abs_epi32(coeff1); in aom_highbd_quantize_b_adaptive_avx2() [all …]
|
D | highbd_adaptive_quantize_sse2.c | 109 __m128i coeff0, coeff1, coeff0_sign, coeff1_sign; in aom_highbd_quantize_b_adaptive_sse2() local 152 coeff1 = _mm_load_si128((__m128i *)(coeff_ptr + 4)); in aom_highbd_quantize_b_adaptive_sse2() 155 coeff1_sign = _mm_srai_epi32(coeff1, 31); in aom_highbd_quantize_b_adaptive_sse2() 157 qcoeff1 = invert_sign_32_sse2(coeff1, coeff1_sign); in aom_highbd_quantize_b_adaptive_sse2() 200 coeff1 = highbd_calculate_dqcoeff(qcoeff1, dequant, log_scale); in aom_highbd_quantize_b_adaptive_sse2() 202 _mm_store_si128((__m128i *)(dqcoeff_ptr + 4), coeff1); in aom_highbd_quantize_b_adaptive_sse2() 208 coeff1 = _mm_load_si128((__m128i *)(coeff_ptr + index + 4)); in aom_highbd_quantize_b_adaptive_sse2() 211 coeff1_sign = _mm_srai_epi32(coeff1, 31); in aom_highbd_quantize_b_adaptive_sse2() 213 qcoeff1 = invert_sign_32_sse2(coeff1, coeff1_sign); in aom_highbd_quantize_b_adaptive_sse2() 245 coeff1 = highbd_calculate_dqcoeff(qcoeff1, dequant, log_scale); in aom_highbd_quantize_b_adaptive_sse2() [all …]
|
D | adaptive_quantize_sse2.c | 32 __m128i coeff0, coeff1, coeff0_sign, coeff1_sign; in aom_quantize_b_adaptive_sse2() local 59 coeff1 = load_coefficients(coeff_ptr + 8); in aom_quantize_b_adaptive_sse2() 63 coeff1_sign = _mm_srai_epi16(coeff1, 15); in aom_quantize_b_adaptive_sse2() 65 qcoeff1 = invert_sign_sse2(coeff1, coeff1_sign); in aom_quantize_b_adaptive_sse2() 112 coeff1 = calculate_dqcoeff(qcoeff1, dequant); in aom_quantize_b_adaptive_sse2() 115 store_coefficients(coeff1, dqcoeff_ptr + 8); in aom_quantize_b_adaptive_sse2() 121 coeff1 = load_coefficients(coeff_ptr + index + 8); in aom_quantize_b_adaptive_sse2() 124 coeff1_sign = _mm_srai_epi16(coeff1, 15); in aom_quantize_b_adaptive_sse2() 126 qcoeff1 = invert_sign_sse2(coeff1, coeff1_sign); in aom_quantize_b_adaptive_sse2() 162 coeff1 = calculate_dqcoeff(qcoeff1, dequant); in aom_quantize_b_adaptive_sse2() [all …]
|
D | avg_intrin_avx2.c | 149 const __m256i coeff1 = _mm256_loadu_si256((const __m256i *)(t_coeff + 64)); in hadamard_16x16_avx2() local 153 __m256i b0 = _mm256_add_epi16(coeff0, coeff1); in hadamard_16x16_avx2() 154 __m256i b1 = _mm256_sub_epi16(coeff0, coeff1); in hadamard_16x16_avx2() 194 const __m256i coeff1 = _mm256_loadu_si256((const __m256i *)(t_coeff + 64)); in aom_hadamard_lp_16x16_avx2() local 198 __m256i b0 = _mm256_add_epi16(coeff0, coeff1); in aom_hadamard_lp_16x16_avx2() 199 __m256i b1 = _mm256_sub_epi16(coeff0, coeff1); in aom_hadamard_lp_16x16_avx2() 235 const __m256i coeff1 = _mm256_loadu_si256((const __m256i *)(t_coeff + 256)); in aom_hadamard_32x32_avx2() local 239 __m256i b0 = _mm256_add_epi16(coeff0, coeff1); in aom_hadamard_32x32_avx2() 240 __m256i b1 = _mm256_sub_epi16(coeff0, coeff1); in aom_hadamard_32x32_avx2() 391 __m256i coeff1 = _mm256_loadu_si256((const __m256i *)(t_coeff + 64)); in aom_highbd_hadamard_16x16_avx2() local [all …]
|
D | quantize_ssse3.c | 76 __m128i coeff0, coeff1, qcoeff0, qcoeff1; in aom_quantize_b_64x64_ssse3() local 98 coeff1 = load_coefficients(coeff_ptr + 8); in aom_quantize_b_64x64_ssse3() 101 qcoeff1 = _mm_abs_epi16(coeff1); in aom_quantize_b_64x64_ssse3() 129 qcoeff1 = _mm_sign_epi16(qcoeff1, coeff1); in aom_quantize_b_64x64_ssse3() 149 coeff1 = load_coefficients(coeff_ptr + index + 8); in aom_quantize_b_64x64_ssse3() 152 qcoeff1 = _mm_abs_epi16(coeff1); in aom_quantize_b_64x64_ssse3() 173 qcoeff1 = _mm_sign_epi16(qcoeff1, coeff1); in aom_quantize_b_64x64_ssse3()
|
D | avg_intrin_sse2.c | 327 __m128i coeff1 = _mm_load_si128((const __m128i *)(t_coeff + 64)); in hadamard_16x16_sse2() local 331 __m128i b0 = _mm_add_epi16(coeff0, coeff1); in hadamard_16x16_sse2() 332 __m128i b1 = _mm_sub_epi16(coeff0, coeff1); in hadamard_16x16_sse2() 342 coeff1 = _mm_add_epi16(b1, b3); in hadamard_16x16_sse2() 348 store_tran_low(coeff1, coeff + 64); in hadamard_16x16_sse2() 354 _mm_store_si128((__m128i *)(coeff16 + 64), coeff1); in hadamard_16x16_sse2() 387 __m128i coeff1 = _mm_load_si128((const __m128i *)(t_coeff + 256)); in aom_hadamard_32x32_sse2() local 391 __m128i b0 = _mm_add_epi16(coeff0, coeff1); in aom_hadamard_32x32_sse2() 392 __m128i b1 = _mm_sub_epi16(coeff0, coeff1); in aom_hadamard_32x32_sse2() 402 coeff1 = _mm_add_epi16(b1, b3); in aom_hadamard_32x32_sse2() [all …]
|
D | quantize_x86.h | 98 static INLINE __m128i scan_for_eob(__m128i *coeff0, __m128i *coeff1, in scan_for_eob() argument 104 const __m128i zero_coeff1 = _mm_cmpeq_epi16(*coeff1, zero); in scan_for_eob() 129 const __m128i coeff1 = _mm_load_si128((__m128i *)(coeff_ptr)); in load_coefficients() local 131 return _mm_packs_epi32(coeff1, coeff2); in load_coefficients()
|
/external/libvpx/libvpx/vp9/encoder/x86/ |
D | vp9_quantize_sse2.c | 44 __m128i coeff0, coeff1; in vp9_quantize_fp_sse2() local 59 coeff1 = load_tran_low(coeff_ptr + n_coeffs + 8); in vp9_quantize_fp_sse2() 63 coeff1_sign = _mm_srai_epi16(coeff1, 15); in vp9_quantize_fp_sse2() 65 qcoeff1 = _mm_xor_si128(coeff1, coeff1_sign); in vp9_quantize_fp_sse2() 87 coeff1 = _mm_mullo_epi16(qcoeff1, dequant); in vp9_quantize_fp_sse2() 90 store_tran_low(coeff1, dqcoeff_ptr + n_coeffs + 8); in vp9_quantize_fp_sse2() 100 zero_coeff1 = _mm_cmpeq_epi16(coeff1, zero); in vp9_quantize_fp_sse2() 119 __m128i coeff0, coeff1; in vp9_quantize_fp_sse2() local 126 coeff1 = load_tran_low(coeff_ptr + n_coeffs + 8); in vp9_quantize_fp_sse2() 130 coeff1_sign = _mm_srai_epi16(coeff1, 15); in vp9_quantize_fp_sse2() [all …]
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
D | quantize_ssse3.c | 31 __m128i coeff0, coeff1; in vpx_quantize_b_ssse3() local 45 coeff1 = load_tran_low(coeff_ptr + 8); in vpx_quantize_b_ssse3() 48 qcoeff1 = _mm_abs_epi16(coeff1); in vpx_quantize_b_ssse3() 62 qcoeff1 = _mm_sign_epi16(qcoeff1, coeff1); in vpx_quantize_b_ssse3() 80 coeff1 = load_tran_low(coeff_ptr + index + 8); in vpx_quantize_b_ssse3() 83 qcoeff1 = _mm_abs_epi16(coeff1); in vpx_quantize_b_ssse3() 92 qcoeff1 = _mm_sign_epi16(qcoeff1, coeff1); in vpx_quantize_b_ssse3() 126 __m128i coeff0, coeff1; in vpx_quantize_b_32x32_ssse3() local 161 coeff1 = load_tran_low(coeff_ptr + 8); in vpx_quantize_b_32x32_ssse3() 164 qcoeff1 = _mm_abs_epi16(coeff1); in vpx_quantize_b_32x32_ssse3() [all …]
|
D | quantize_avx.c | 35 __m128i coeff0, coeff1; in vpx_quantize_b_avx() local 52 coeff1 = load_tran_low(coeff_ptr + 8); in vpx_quantize_b_avx() 55 qcoeff1 = _mm_abs_epi16(coeff1); in vpx_quantize_b_avx() 85 qcoeff1 = _mm_sign_epi16(qcoeff1, coeff1); in vpx_quantize_b_avx() 105 coeff1 = load_tran_low(coeff_ptr + index + 8); in vpx_quantize_b_avx() 108 qcoeff1 = _mm_abs_epi16(coeff1); in vpx_quantize_b_avx() 128 qcoeff1 = _mm_sign_epi16(qcoeff1, coeff1); in vpx_quantize_b_avx() 161 __m128i coeff0, coeff1; in vpx_quantize_b_32x32_avx() local 193 coeff1 = load_tran_low(coeff_ptr + 8); in vpx_quantize_b_32x32_avx() 196 qcoeff1 = _mm_abs_epi16(coeff1); in vpx_quantize_b_32x32_avx() [all …]
|
D | quantize_sse2.c | 31 __m128i coeff0, coeff1, coeff0_sign, coeff1_sign; in vpx_quantize_b_sse2() local 46 coeff1 = load_tran_low(coeff_ptr + 8); in vpx_quantize_b_sse2() 50 coeff1_sign = _mm_srai_epi16(coeff1, 15); in vpx_quantize_b_sse2() 52 qcoeff1 = invert_sign_sse2(coeff1, coeff1_sign); in vpx_quantize_b_sse2() 86 coeff1 = load_tran_low(coeff_ptr + index + 8); in vpx_quantize_b_sse2() 89 coeff1_sign = _mm_srai_epi16(coeff1, 15); in vpx_quantize_b_sse2() 91 qcoeff1 = invert_sign_sse2(coeff1, coeff1_sign); in vpx_quantize_b_sse2()
|
D | avg_intrin_avx2.c | 150 __m256i coeff1 = _mm256_loadu_si256((const __m256i *)(t_coeff + 64)); in vpx_highbd_hadamard_16x16_avx2() local 154 __m256i b0 = _mm256_add_epi32(coeff0, coeff1); in vpx_highbd_hadamard_16x16_avx2() 155 __m256i b1 = _mm256_sub_epi32(coeff0, coeff1); in vpx_highbd_hadamard_16x16_avx2() 165 coeff1 = _mm256_add_epi32(b1, b3); in vpx_highbd_hadamard_16x16_avx2() 170 _mm256_storeu_si256((__m256i *)(coeff + 64), coeff1); in vpx_highbd_hadamard_16x16_avx2() 191 __m256i coeff1 = _mm256_loadu_si256((const __m256i *)(t_coeff + 256)); in vpx_highbd_hadamard_32x32_avx2() local 195 __m256i b0 = _mm256_add_epi32(coeff0, coeff1); in vpx_highbd_hadamard_32x32_avx2() 196 __m256i b1 = _mm256_sub_epi32(coeff0, coeff1); in vpx_highbd_hadamard_32x32_avx2() 206 coeff1 = _mm256_add_epi32(b1, b3); in vpx_highbd_hadamard_32x32_avx2() 211 _mm256_storeu_si256((__m256i *)(coeff + 256), coeff1); in vpx_highbd_hadamard_32x32_avx2() [all …]
|
D | avg_intrin_sse2.c | 348 __m128i coeff1 = _mm_load_si128((const __m128i *)(t_coeff + 64)); in hadamard_16x16_sse2() local 352 __m128i b0 = _mm_add_epi16(coeff0, coeff1); in hadamard_16x16_sse2() 353 __m128i b1 = _mm_sub_epi16(coeff0, coeff1); in hadamard_16x16_sse2() 363 coeff1 = _mm_add_epi16(b1, b3); in hadamard_16x16_sse2() 369 store_tran_low(coeff1, coeff + 64); in hadamard_16x16_sse2() 375 _mm_store_si128((__m128i *)(coeff16 + 64), coeff1); in hadamard_16x16_sse2() 412 __m128i coeff1 = _mm_load_si128((const __m128i *)(t_coeff + 256)); in vpx_hadamard_32x32_sse2() local 416 __m128i b0 = _mm_add_epi16(coeff0, coeff1); in vpx_hadamard_32x32_sse2() 417 __m128i b1 = _mm_sub_epi16(coeff0, coeff1); in vpx_hadamard_32x32_sse2() 427 coeff1 = _mm_add_epi16(b1, b3); in vpx_hadamard_32x32_sse2() [all …]
|
/external/libvpx/libvpx/vp9/encoder/ppc/ |
D | vp9_quantize_vsx.c | 54 int16x8_t coeff1 = vec_vsx_ld(16, coeff_ptr); in vp9_quantize_fp_vsx() local 79 qcoeff1 = vec_mulhi(vec_vaddshs(vec_abs(coeff1), round), quant); in vp9_quantize_fp_vsx() 81 qcoeff1 = vec_sign(qcoeff1, coeff1); in vp9_quantize_fp_vsx() 104 coeff1 = vec_vsx_ld(off1, coeff_ptr); in vp9_quantize_fp_vsx() 117 qcoeff1 = vec_mulhi(vec_vaddshs(vec_abs(coeff1), round), quant); in vp9_quantize_fp_vsx() 119 qcoeff1 = vec_sign(qcoeff1, coeff1); in vp9_quantize_fp_vsx() 189 int16x8_t coeff1 = vec_vsx_ld(16, coeff_ptr); in vp9_quantize_fp_32x32_vsx() local 194 int16x8_t abs_coeff1 = vec_abs(coeff1); in vp9_quantize_fp_32x32_vsx() 223 vec_madds(vec_vaddshs(vec_abs(coeff1), round), quant, vec_zeros_s16); in vp9_quantize_fp_32x32_vsx() 226 qcoeff1 = vec_sign(qcoeff1, coeff1); in vp9_quantize_fp_32x32_vsx() [all …]
|
/external/libvpx/libvpx/vpx_dsp/ppc/ |
D | quantize_vsx.c | 115 int16x8_t coeff1 = vec_vsx_ld(16, coeff_ptr); in vpx_quantize_b_vsx() local 118 int16x8_t coeff1_abs = vec_abs(coeff1); in vpx_quantize_b_vsx() 135 quantize_coeff(coeff1, coeff1_abs, round, quant, quant_shift, zero_mask1); in vpx_quantize_b_vsx() 156 coeff1 = vec_vsx_ld(off1, coeff_ptr); in vpx_quantize_b_vsx() 159 coeff1_abs = vec_abs(coeff1); in vpx_quantize_b_vsx() 166 qcoeff1 = quantize_coeff(coeff1, coeff1_abs, round, quant, quant_shift, in vpx_quantize_b_vsx() 224 int16x8_t coeff1 = vec_vsx_ld(16, coeff_ptr); in vpx_quantize_b_32x32_vsx() local 227 int16x8_t coeff1_abs = vec_abs(coeff1); in vpx_quantize_b_32x32_vsx() 247 qcoeff1 = quantize_coeff_32(coeff1, coeff1_abs, round, quant, quant_shift, in vpx_quantize_b_32x32_vsx() 265 coeff1 = vec_vsx_ld(off1, coeff_ptr); in vpx_quantize_b_32x32_vsx() [all …]
|
/external/libaom/libaom/av1/encoder/x86/ |
D | av1_quantize_sse2.c | 79 __m128i coeff0, coeff1; in quantize() local 81 read_coeff(coeff_ptr, n_coeffs, &coeff0, &coeff1); in quantize() 85 const __m128i coeff1_sign = _mm_srai_epi16(coeff1, 15); in quantize() 87 __m128i qcoeff1 = _mm_xor_si128(coeff1, coeff1_sign); in quantize() 111 coeff1 = _mm_mullo_epi16(qcoeff1, *dequant1); in quantize() 113 write_qcoeff(&coeff0, &coeff1, dqcoeff_ptr, n_coeffs); in quantize() 118 const __m128i zero_coeff1 = _mm_cmpeq_epi16(coeff1, zero); in quantize()
|
/external/apache-commons-math/src/main/java/org/apache/commons/math/ode/nonstiff/ |
D | MidpointStepInterpolator.java | 86 final double coeff1 = oneMinusThetaH * theta; in computeInterpolatedStateAndDerivatives() local 94 interpolatedState[i] = currentState[i] + coeff1 * yDot1 - coeff2 * yDot2; in computeInterpolatedStateAndDerivatives()
|
D | ClassicalRungeKuttaStepInterpolator.java | 92 final double coeff1 = s * ((-fourTheta + 5) * theta - 1); in computeInterpolatedStateAndDerivatives() local 103 currentState[i] + coeff1 * yDot1 + coeff23 * yDot23 + coeff4 * yDot4; in computeInterpolatedStateAndDerivatives()
|
D | ThreeEighthesStepInterpolator.java | 93 final double coeff1 = s * (1 - 7 * theta + 2 * fourTheta2); in computeInterpolatedStateAndDerivatives() local 108 currentState[i] - coeff1 * yDot1 - coeff2 * yDot2 - coeff3 * yDot3 - coeff4 * yDot4; in computeInterpolatedStateAndDerivatives()
|
D | GillStepInterpolator.java | 102 final double coeff1 = soMt * (1 - fourTheta); in computeInterpolatedStateAndDerivatives() local 118 currentState[i] - coeff1 * yDot1 - coeff2 * yDot2 - coeff3 * yDot3 - coeff4 * yDot4; in computeInterpolatedStateAndDerivatives()
|
/external/webp/src/dsp/ |
D | upsampling_neon.c | 145 const int16x8_t Y1 = vqdmulhq_lane_s16(Y0, coeff1, 0); \ 146 const int16x8_t R0 = vqdmulhq_lane_s16(V0, coeff1, 1); \ 147 const int16x8_t G0 = vqdmulhq_lane_s16(U0, coeff1, 2); \ 148 const int16x8_t G1 = vqdmulhq_lane_s16(V0, coeff1, 3); \ 210 const int16x4_t coeff1 = vld1_s16(kCoeffs1); \
|
/external/apache-commons-math/src/main/java/org/apache/commons/math/estimation/ |
D | LevenbergMarquardtEstimator.java | 394 double coeff1 = 0; in estimate() local 396 coeff1 += work1[j] * work1[j]; in estimate() 399 coeff1 = coeff1 / pc2; in estimate() 401 double preRed = coeff1 + 2 * coeff2; in estimate() 402 double dirDer = -(coeff1 + coeff2); in estimate()
|