Home
last modified time | relevance | path

Searched refs:coeff1 (Results 1 – 25 of 34) sorted by relevance

12

/external/libvpx/libvpx/vp8/encoder/mips/msa/
Dencodeopt_msa.c18 v8i16 coeff, dq_coeff, coeff0, coeff1; in vp8_block_error_msa() local
26 ILVRL_H2_SH(coeff, dq_coeff, coeff0, coeff1); in vp8_block_error_msa()
27 HSUB_UH2_SW(coeff0, coeff1, diff0, diff1); in vp8_block_error_msa()
47 v8i16 coeff, coeff0, coeff1, coeff2, coeff3, coeff4; in vp8_mbblock_error_msa() local
79 ILVRL_H2_SH(coeff, dq_coeff, coeff0, coeff1); in vp8_mbblock_error_msa()
80 HSUB_UH2_SW(coeff0, coeff1, diff0, diff1); in vp8_mbblock_error_msa()
83 ILVRL_H2_SH(coeff2, dq_coeff2, coeff0, coeff1); in vp8_mbblock_error_msa()
84 HSUB_UH2_SW(coeff0, coeff1, diff0, diff1); in vp8_mbblock_error_msa()
91 ILVRL_H2_SH(coeff3, dq_coeff3, coeff0, coeff1); in vp8_mbblock_error_msa()
92 HSUB_UH2_SW(coeff0, coeff1, diff0, diff1); in vp8_mbblock_error_msa()
[all …]
Dquantize_msa.c24 v8i16 coeff0, coeff1, z0, z1; in fast_quantize_b_msa() local
35 LD_SH2(coeff_ptr, 8, coeff0, coeff1); in fast_quantize_b_msa()
36 VSHF_H2_SH(coeff0, coeff1, coeff0, coeff1, zigzag_mask0, zigzag_mask1, z0, in fast_quantize_b_msa()
38 LD_SH2(round, 8, coeff0, coeff1); in fast_quantize_b_msa()
39 VSHF_H2_SH(coeff0, coeff1, coeff0, coeff1, zigzag_mask0, zigzag_mask1, round0, in fast_quantize_b_msa()
41 LD_SH2(quant, 8, coeff0, coeff1); in fast_quantize_b_msa()
42 VSHF_H2_SH(coeff0, coeff1, coeff0, coeff1, zigzag_mask0, zigzag_mask1, quant0, in fast_quantize_b_msa()
93 v8i16 coeff0, coeff1, z0, z1; in exact_regular_quantize_b_msa() local
105 LD_SH2(coeff_ptr, 8, coeff0, coeff1); in exact_regular_quantize_b_msa()
106 VSHF_H2_SH(coeff0, coeff1, coeff0, coeff1, zigzag_mask0, zigzag_mask1, z0, in exact_regular_quantize_b_msa()
[all …]
/external/libaom/libaom/aom_dsp/x86/
Dquantize_sse2.c32 __m128i coeff0, coeff1, coeff0_sign, coeff1_sign; in aom_quantize_b_sse2() local
45 coeff1 = load_coefficients(coeff_ptr + 8); in aom_quantize_b_sse2()
49 coeff1_sign = _mm_srai_epi16(coeff1, 15); in aom_quantize_b_sse2()
51 qcoeff1 = invert_sign_sse2(coeff1, coeff1_sign); in aom_quantize_b_sse2()
78 coeff1 = calculate_dqcoeff(qcoeff1, dequant); in aom_quantize_b_sse2()
81 store_coefficients(coeff1, dqcoeff_ptr + 8); in aom_quantize_b_sse2()
84 scan_for_eob(&coeff0, &coeff1, cmp_mask0, cmp_mask1, iscan_ptr, 0, zero); in aom_quantize_b_sse2()
89 coeff1 = load_coefficients(coeff_ptr + index + 8); in aom_quantize_b_sse2()
92 coeff1_sign = _mm_srai_epi16(coeff1, 15); in aom_quantize_b_sse2()
94 qcoeff1 = invert_sign_sse2(coeff1, coeff1_sign); in aom_quantize_b_sse2()
[all …]
Dhighbd_adaptive_quantize_avx2.c100 __m256i coeff1, in highbd_store_coefficients_avx2() argument
103 _mm256_store_si256((__m256i *)(coeff_ptr + 8), coeff1); in highbd_store_coefficients_avx2()
119 __m256i coeff0, qcoeff0, coeff1, qcoeff1; in aom_highbd_quantize_b_adaptive_avx2() local
147 coeff1 = _mm256_load_si256((__m256i *)(coeff_ptr + 8)); in aom_highbd_quantize_b_adaptive_avx2()
148 qcoeff1 = _mm256_abs_epi32(coeff1); in aom_highbd_quantize_b_adaptive_avx2()
174 qcoeff1 = _mm256_sign_epi32(qcoeff1, coeff1); in aom_highbd_quantize_b_adaptive_avx2()
181 coeff1 = highbd_calculate_dqcoeff_avx2(qcoeff1, dequant); in aom_highbd_quantize_b_adaptive_avx2()
182 highbd_store_coefficients_avx2(coeff0, coeff1, dqcoeff_ptr); in aom_highbd_quantize_b_adaptive_avx2()
189 coeff1 = _mm256_load_si256((__m256i *)(coeff_ptr + index + 8)); in aom_highbd_quantize_b_adaptive_avx2()
190 qcoeff1 = _mm256_abs_epi32(coeff1); in aom_highbd_quantize_b_adaptive_avx2()
[all …]
Dhighbd_adaptive_quantize_sse2.c109 __m128i coeff0, coeff1, coeff0_sign, coeff1_sign; in aom_highbd_quantize_b_adaptive_sse2() local
152 coeff1 = _mm_load_si128((__m128i *)(coeff_ptr + 4)); in aom_highbd_quantize_b_adaptive_sse2()
155 coeff1_sign = _mm_srai_epi32(coeff1, 31); in aom_highbd_quantize_b_adaptive_sse2()
157 qcoeff1 = invert_sign_32_sse2(coeff1, coeff1_sign); in aom_highbd_quantize_b_adaptive_sse2()
200 coeff1 = highbd_calculate_dqcoeff(qcoeff1, dequant, log_scale); in aom_highbd_quantize_b_adaptive_sse2()
202 _mm_store_si128((__m128i *)(dqcoeff_ptr + 4), coeff1); in aom_highbd_quantize_b_adaptive_sse2()
208 coeff1 = _mm_load_si128((__m128i *)(coeff_ptr + index + 4)); in aom_highbd_quantize_b_adaptive_sse2()
211 coeff1_sign = _mm_srai_epi32(coeff1, 31); in aom_highbd_quantize_b_adaptive_sse2()
213 qcoeff1 = invert_sign_32_sse2(coeff1, coeff1_sign); in aom_highbd_quantize_b_adaptive_sse2()
245 coeff1 = highbd_calculate_dqcoeff(qcoeff1, dequant, log_scale); in aom_highbd_quantize_b_adaptive_sse2()
[all …]
Dadaptive_quantize_sse2.c32 __m128i coeff0, coeff1, coeff0_sign, coeff1_sign; in aom_quantize_b_adaptive_sse2() local
59 coeff1 = load_coefficients(coeff_ptr + 8); in aom_quantize_b_adaptive_sse2()
63 coeff1_sign = _mm_srai_epi16(coeff1, 15); in aom_quantize_b_adaptive_sse2()
65 qcoeff1 = invert_sign_sse2(coeff1, coeff1_sign); in aom_quantize_b_adaptive_sse2()
112 coeff1 = calculate_dqcoeff(qcoeff1, dequant); in aom_quantize_b_adaptive_sse2()
115 store_coefficients(coeff1, dqcoeff_ptr + 8); in aom_quantize_b_adaptive_sse2()
121 coeff1 = load_coefficients(coeff_ptr + index + 8); in aom_quantize_b_adaptive_sse2()
124 coeff1_sign = _mm_srai_epi16(coeff1, 15); in aom_quantize_b_adaptive_sse2()
126 qcoeff1 = invert_sign_sse2(coeff1, coeff1_sign); in aom_quantize_b_adaptive_sse2()
162 coeff1 = calculate_dqcoeff(qcoeff1, dequant); in aom_quantize_b_adaptive_sse2()
[all …]
Davg_intrin_avx2.c149 const __m256i coeff1 = _mm256_loadu_si256((const __m256i *)(t_coeff + 64)); in hadamard_16x16_avx2() local
153 __m256i b0 = _mm256_add_epi16(coeff0, coeff1); in hadamard_16x16_avx2()
154 __m256i b1 = _mm256_sub_epi16(coeff0, coeff1); in hadamard_16x16_avx2()
194 const __m256i coeff1 = _mm256_loadu_si256((const __m256i *)(t_coeff + 64)); in aom_hadamard_lp_16x16_avx2() local
198 __m256i b0 = _mm256_add_epi16(coeff0, coeff1); in aom_hadamard_lp_16x16_avx2()
199 __m256i b1 = _mm256_sub_epi16(coeff0, coeff1); in aom_hadamard_lp_16x16_avx2()
235 const __m256i coeff1 = _mm256_loadu_si256((const __m256i *)(t_coeff + 256)); in aom_hadamard_32x32_avx2() local
239 __m256i b0 = _mm256_add_epi16(coeff0, coeff1); in aom_hadamard_32x32_avx2()
240 __m256i b1 = _mm256_sub_epi16(coeff0, coeff1); in aom_hadamard_32x32_avx2()
391 __m256i coeff1 = _mm256_loadu_si256((const __m256i *)(t_coeff + 64)); in aom_highbd_hadamard_16x16_avx2() local
[all …]
Dquantize_ssse3.c76 __m128i coeff0, coeff1, qcoeff0, qcoeff1; in aom_quantize_b_64x64_ssse3() local
98 coeff1 = load_coefficients(coeff_ptr + 8); in aom_quantize_b_64x64_ssse3()
101 qcoeff1 = _mm_abs_epi16(coeff1); in aom_quantize_b_64x64_ssse3()
129 qcoeff1 = _mm_sign_epi16(qcoeff1, coeff1); in aom_quantize_b_64x64_ssse3()
149 coeff1 = load_coefficients(coeff_ptr + index + 8); in aom_quantize_b_64x64_ssse3()
152 qcoeff1 = _mm_abs_epi16(coeff1); in aom_quantize_b_64x64_ssse3()
173 qcoeff1 = _mm_sign_epi16(qcoeff1, coeff1); in aom_quantize_b_64x64_ssse3()
Davg_intrin_sse2.c327 __m128i coeff1 = _mm_load_si128((const __m128i *)(t_coeff + 64)); in hadamard_16x16_sse2() local
331 __m128i b0 = _mm_add_epi16(coeff0, coeff1); in hadamard_16x16_sse2()
332 __m128i b1 = _mm_sub_epi16(coeff0, coeff1); in hadamard_16x16_sse2()
342 coeff1 = _mm_add_epi16(b1, b3); in hadamard_16x16_sse2()
348 store_tran_low(coeff1, coeff + 64); in hadamard_16x16_sse2()
354 _mm_store_si128((__m128i *)(coeff16 + 64), coeff1); in hadamard_16x16_sse2()
387 __m128i coeff1 = _mm_load_si128((const __m128i *)(t_coeff + 256)); in aom_hadamard_32x32_sse2() local
391 __m128i b0 = _mm_add_epi16(coeff0, coeff1); in aom_hadamard_32x32_sse2()
392 __m128i b1 = _mm_sub_epi16(coeff0, coeff1); in aom_hadamard_32x32_sse2()
402 coeff1 = _mm_add_epi16(b1, b3); in aom_hadamard_32x32_sse2()
[all …]
Dquantize_x86.h98 static INLINE __m128i scan_for_eob(__m128i *coeff0, __m128i *coeff1, in scan_for_eob() argument
104 const __m128i zero_coeff1 = _mm_cmpeq_epi16(*coeff1, zero); in scan_for_eob()
129 const __m128i coeff1 = _mm_load_si128((__m128i *)(coeff_ptr)); in load_coefficients() local
131 return _mm_packs_epi32(coeff1, coeff2); in load_coefficients()
/external/libvpx/libvpx/vp9/encoder/x86/
Dvp9_quantize_sse2.c44 __m128i coeff0, coeff1; in vp9_quantize_fp_sse2() local
59 coeff1 = load_tran_low(coeff_ptr + n_coeffs + 8); in vp9_quantize_fp_sse2()
63 coeff1_sign = _mm_srai_epi16(coeff1, 15); in vp9_quantize_fp_sse2()
65 qcoeff1 = _mm_xor_si128(coeff1, coeff1_sign); in vp9_quantize_fp_sse2()
87 coeff1 = _mm_mullo_epi16(qcoeff1, dequant); in vp9_quantize_fp_sse2()
90 store_tran_low(coeff1, dqcoeff_ptr + n_coeffs + 8); in vp9_quantize_fp_sse2()
100 zero_coeff1 = _mm_cmpeq_epi16(coeff1, zero); in vp9_quantize_fp_sse2()
119 __m128i coeff0, coeff1; in vp9_quantize_fp_sse2() local
126 coeff1 = load_tran_low(coeff_ptr + n_coeffs + 8); in vp9_quantize_fp_sse2()
130 coeff1_sign = _mm_srai_epi16(coeff1, 15); in vp9_quantize_fp_sse2()
[all …]
/external/libvpx/libvpx/vpx_dsp/x86/
Dquantize_ssse3.c31 __m128i coeff0, coeff1; in vpx_quantize_b_ssse3() local
45 coeff1 = load_tran_low(coeff_ptr + 8); in vpx_quantize_b_ssse3()
48 qcoeff1 = _mm_abs_epi16(coeff1); in vpx_quantize_b_ssse3()
62 qcoeff1 = _mm_sign_epi16(qcoeff1, coeff1); in vpx_quantize_b_ssse3()
80 coeff1 = load_tran_low(coeff_ptr + index + 8); in vpx_quantize_b_ssse3()
83 qcoeff1 = _mm_abs_epi16(coeff1); in vpx_quantize_b_ssse3()
92 qcoeff1 = _mm_sign_epi16(qcoeff1, coeff1); in vpx_quantize_b_ssse3()
126 __m128i coeff0, coeff1; in vpx_quantize_b_32x32_ssse3() local
161 coeff1 = load_tran_low(coeff_ptr + 8); in vpx_quantize_b_32x32_ssse3()
164 qcoeff1 = _mm_abs_epi16(coeff1); in vpx_quantize_b_32x32_ssse3()
[all …]
Dquantize_avx.c35 __m128i coeff0, coeff1; in vpx_quantize_b_avx() local
52 coeff1 = load_tran_low(coeff_ptr + 8); in vpx_quantize_b_avx()
55 qcoeff1 = _mm_abs_epi16(coeff1); in vpx_quantize_b_avx()
85 qcoeff1 = _mm_sign_epi16(qcoeff1, coeff1); in vpx_quantize_b_avx()
105 coeff1 = load_tran_low(coeff_ptr + index + 8); in vpx_quantize_b_avx()
108 qcoeff1 = _mm_abs_epi16(coeff1); in vpx_quantize_b_avx()
128 qcoeff1 = _mm_sign_epi16(qcoeff1, coeff1); in vpx_quantize_b_avx()
161 __m128i coeff0, coeff1; in vpx_quantize_b_32x32_avx() local
193 coeff1 = load_tran_low(coeff_ptr + 8); in vpx_quantize_b_32x32_avx()
196 qcoeff1 = _mm_abs_epi16(coeff1); in vpx_quantize_b_32x32_avx()
[all …]
Dquantize_sse2.c31 __m128i coeff0, coeff1, coeff0_sign, coeff1_sign; in vpx_quantize_b_sse2() local
46 coeff1 = load_tran_low(coeff_ptr + 8); in vpx_quantize_b_sse2()
50 coeff1_sign = _mm_srai_epi16(coeff1, 15); in vpx_quantize_b_sse2()
52 qcoeff1 = invert_sign_sse2(coeff1, coeff1_sign); in vpx_quantize_b_sse2()
86 coeff1 = load_tran_low(coeff_ptr + index + 8); in vpx_quantize_b_sse2()
89 coeff1_sign = _mm_srai_epi16(coeff1, 15); in vpx_quantize_b_sse2()
91 qcoeff1 = invert_sign_sse2(coeff1, coeff1_sign); in vpx_quantize_b_sse2()
Davg_intrin_avx2.c150 __m256i coeff1 = _mm256_loadu_si256((const __m256i *)(t_coeff + 64)); in vpx_highbd_hadamard_16x16_avx2() local
154 __m256i b0 = _mm256_add_epi32(coeff0, coeff1); in vpx_highbd_hadamard_16x16_avx2()
155 __m256i b1 = _mm256_sub_epi32(coeff0, coeff1); in vpx_highbd_hadamard_16x16_avx2()
165 coeff1 = _mm256_add_epi32(b1, b3); in vpx_highbd_hadamard_16x16_avx2()
170 _mm256_storeu_si256((__m256i *)(coeff + 64), coeff1); in vpx_highbd_hadamard_16x16_avx2()
191 __m256i coeff1 = _mm256_loadu_si256((const __m256i *)(t_coeff + 256)); in vpx_highbd_hadamard_32x32_avx2() local
195 __m256i b0 = _mm256_add_epi32(coeff0, coeff1); in vpx_highbd_hadamard_32x32_avx2()
196 __m256i b1 = _mm256_sub_epi32(coeff0, coeff1); in vpx_highbd_hadamard_32x32_avx2()
206 coeff1 = _mm256_add_epi32(b1, b3); in vpx_highbd_hadamard_32x32_avx2()
211 _mm256_storeu_si256((__m256i *)(coeff + 256), coeff1); in vpx_highbd_hadamard_32x32_avx2()
[all …]
Davg_intrin_sse2.c348 __m128i coeff1 = _mm_load_si128((const __m128i *)(t_coeff + 64)); in hadamard_16x16_sse2() local
352 __m128i b0 = _mm_add_epi16(coeff0, coeff1); in hadamard_16x16_sse2()
353 __m128i b1 = _mm_sub_epi16(coeff0, coeff1); in hadamard_16x16_sse2()
363 coeff1 = _mm_add_epi16(b1, b3); in hadamard_16x16_sse2()
369 store_tran_low(coeff1, coeff + 64); in hadamard_16x16_sse2()
375 _mm_store_si128((__m128i *)(coeff16 + 64), coeff1); in hadamard_16x16_sse2()
412 __m128i coeff1 = _mm_load_si128((const __m128i *)(t_coeff + 256)); in vpx_hadamard_32x32_sse2() local
416 __m128i b0 = _mm_add_epi16(coeff0, coeff1); in vpx_hadamard_32x32_sse2()
417 __m128i b1 = _mm_sub_epi16(coeff0, coeff1); in vpx_hadamard_32x32_sse2()
427 coeff1 = _mm_add_epi16(b1, b3); in vpx_hadamard_32x32_sse2()
[all …]
/external/libvpx/libvpx/vp9/encoder/ppc/
Dvp9_quantize_vsx.c54 int16x8_t coeff1 = vec_vsx_ld(16, coeff_ptr); in vp9_quantize_fp_vsx() local
79 qcoeff1 = vec_mulhi(vec_vaddshs(vec_abs(coeff1), round), quant); in vp9_quantize_fp_vsx()
81 qcoeff1 = vec_sign(qcoeff1, coeff1); in vp9_quantize_fp_vsx()
104 coeff1 = vec_vsx_ld(off1, coeff_ptr); in vp9_quantize_fp_vsx()
117 qcoeff1 = vec_mulhi(vec_vaddshs(vec_abs(coeff1), round), quant); in vp9_quantize_fp_vsx()
119 qcoeff1 = vec_sign(qcoeff1, coeff1); in vp9_quantize_fp_vsx()
189 int16x8_t coeff1 = vec_vsx_ld(16, coeff_ptr); in vp9_quantize_fp_32x32_vsx() local
194 int16x8_t abs_coeff1 = vec_abs(coeff1); in vp9_quantize_fp_32x32_vsx()
223 vec_madds(vec_vaddshs(vec_abs(coeff1), round), quant, vec_zeros_s16); in vp9_quantize_fp_32x32_vsx()
226 qcoeff1 = vec_sign(qcoeff1, coeff1); in vp9_quantize_fp_32x32_vsx()
[all …]
/external/libvpx/libvpx/vpx_dsp/ppc/
Dquantize_vsx.c115 int16x8_t coeff1 = vec_vsx_ld(16, coeff_ptr); in vpx_quantize_b_vsx() local
118 int16x8_t coeff1_abs = vec_abs(coeff1); in vpx_quantize_b_vsx()
135 quantize_coeff(coeff1, coeff1_abs, round, quant, quant_shift, zero_mask1); in vpx_quantize_b_vsx()
156 coeff1 = vec_vsx_ld(off1, coeff_ptr); in vpx_quantize_b_vsx()
159 coeff1_abs = vec_abs(coeff1); in vpx_quantize_b_vsx()
166 qcoeff1 = quantize_coeff(coeff1, coeff1_abs, round, quant, quant_shift, in vpx_quantize_b_vsx()
224 int16x8_t coeff1 = vec_vsx_ld(16, coeff_ptr); in vpx_quantize_b_32x32_vsx() local
227 int16x8_t coeff1_abs = vec_abs(coeff1); in vpx_quantize_b_32x32_vsx()
247 qcoeff1 = quantize_coeff_32(coeff1, coeff1_abs, round, quant, quant_shift, in vpx_quantize_b_32x32_vsx()
265 coeff1 = vec_vsx_ld(off1, coeff_ptr); in vpx_quantize_b_32x32_vsx()
[all …]
/external/libaom/libaom/av1/encoder/x86/
Dav1_quantize_sse2.c79 __m128i coeff0, coeff1; in quantize() local
81 read_coeff(coeff_ptr, n_coeffs, &coeff0, &coeff1); in quantize()
85 const __m128i coeff1_sign = _mm_srai_epi16(coeff1, 15); in quantize()
87 __m128i qcoeff1 = _mm_xor_si128(coeff1, coeff1_sign); in quantize()
111 coeff1 = _mm_mullo_epi16(qcoeff1, *dequant1); in quantize()
113 write_qcoeff(&coeff0, &coeff1, dqcoeff_ptr, n_coeffs); in quantize()
118 const __m128i zero_coeff1 = _mm_cmpeq_epi16(coeff1, zero); in quantize()
/external/apache-commons-math/src/main/java/org/apache/commons/math/ode/nonstiff/
DMidpointStepInterpolator.java86 final double coeff1 = oneMinusThetaH * theta; in computeInterpolatedStateAndDerivatives() local
94 interpolatedState[i] = currentState[i] + coeff1 * yDot1 - coeff2 * yDot2; in computeInterpolatedStateAndDerivatives()
DClassicalRungeKuttaStepInterpolator.java92 final double coeff1 = s * ((-fourTheta + 5) * theta - 1); in computeInterpolatedStateAndDerivatives() local
103 currentState[i] + coeff1 * yDot1 + coeff23 * yDot23 + coeff4 * yDot4; in computeInterpolatedStateAndDerivatives()
DThreeEighthesStepInterpolator.java93 final double coeff1 = s * (1 - 7 * theta + 2 * fourTheta2); in computeInterpolatedStateAndDerivatives() local
108 currentState[i] - coeff1 * yDot1 - coeff2 * yDot2 - coeff3 * yDot3 - coeff4 * yDot4; in computeInterpolatedStateAndDerivatives()
DGillStepInterpolator.java102 final double coeff1 = soMt * (1 - fourTheta); in computeInterpolatedStateAndDerivatives() local
118 currentState[i] - coeff1 * yDot1 - coeff2 * yDot2 - coeff3 * yDot3 - coeff4 * yDot4; in computeInterpolatedStateAndDerivatives()
/external/webp/src/dsp/
Dupsampling_neon.c145 const int16x8_t Y1 = vqdmulhq_lane_s16(Y0, coeff1, 0); \
146 const int16x8_t R0 = vqdmulhq_lane_s16(V0, coeff1, 1); \
147 const int16x8_t G0 = vqdmulhq_lane_s16(U0, coeff1, 2); \
148 const int16x8_t G1 = vqdmulhq_lane_s16(V0, coeff1, 3); \
210 const int16x4_t coeff1 = vld1_s16(kCoeffs1); \
/external/apache-commons-math/src/main/java/org/apache/commons/math/estimation/
DLevenbergMarquardtEstimator.java394 double coeff1 = 0; in estimate() local
396 coeff1 += work1[j] * work1[j]; in estimate()
399 coeff1 = coeff1 / pc2; in estimate()
401 double preRed = coeff1 + 2 * coeff2; in estimate()
402 double dirDer = -(coeff1 + coeff2); in estimate()

12