/external/libgav1/libgav1/src/dsp/x86/ |
D | average_blend_sse4.cc | 154 const __m128i& round_offset, const __m128i& max, in AverageBlendRow() argument 170 const __m128i round_0 = _mm_add_epi32(compound_offset_0, round_offset); in AverageBlendRow() 171 const __m128i round_1 = _mm_add_epi32(compound_offset_1, round_offset); in AverageBlendRow() 195 const __m128i round_offset = in AverageBlend10bpp_SSE4_1() local 206 AverageBlendRow<4, 0>(pred_0, pred_1, compound_offset, round_offset, max, in AverageBlend10bpp_SSE4_1() 220 AverageBlendRow<8, 0>(pred_0, pred_1, compound_offset, round_offset, max, in AverageBlend10bpp_SSE4_1() 224 round_offset, max, zero, dst + dest_stride, in AverageBlend10bpp_SSE4_1() 238 AverageBlendRow<8, 0>(pred_0, pred_1, compound_offset, round_offset, max, in AverageBlend10bpp_SSE4_1() 240 AverageBlendRow<8, 8>(pred_0, pred_1, compound_offset, round_offset, max, in AverageBlend10bpp_SSE4_1() 244 round_offset, max, zero, dst + dest_stride, in AverageBlend10bpp_SSE4_1() [all …]
|
/external/libaom/libaom/aom_dsp/x86/ |
D | blend_mask_sse4.h | 68 const __m128i *round_offset, int shift) { in aom_lowbd_blend_a64_d16_mask_subw0_subh0_w4_sse4_1() argument 74 blend_a64_d16_mask_w4_sse41(dst, src0, src1, &m, round_offset, &v_maxval, in aom_lowbd_blend_a64_d16_mask_subw0_subh0_w4_sse4_1() 87 const __m128i *round_offset, int shift) { in aom_lowbd_blend_a64_d16_mask_subw0_subh0_w8_sse4_1() argument 92 blend_a64_d16_mask_w8_sse41(dst, src0, src1, &m, round_offset, &v_maxval, in aom_lowbd_blend_a64_d16_mask_subw0_subh0_w8_sse4_1() 105 const __m128i *round_offset, int shift) { in aom_lowbd_blend_a64_d16_mask_subw1_subh1_w4_sse4_1() argument 117 blend_a64_d16_mask_w4_sse41(dst, src0, src1, &m, round_offset, &v_maxval, in aom_lowbd_blend_a64_d16_mask_subw1_subh1_w4_sse4_1() 130 const __m128i *round_offset, int shift) { in aom_lowbd_blend_a64_d16_mask_subw1_subh1_w8_sse4_1() argument 142 blend_a64_d16_mask_w8_sse41(dst, src0, src1, &m, round_offset, &v_maxval, in aom_lowbd_blend_a64_d16_mask_subw1_subh1_w8_sse4_1() 155 const __m128i *round_offset, int shift) { in aom_lowbd_blend_a64_d16_mask_subw1_subh0_w4_sse4_1() argument 164 blend_a64_d16_mask_w4_sse41(dst, src0, src1, &m, round_offset, &v_maxval, in aom_lowbd_blend_a64_d16_mask_subw1_subh0_w4_sse4_1() [all …]
|
D | masked_variance_intrin_ssse3.h | 28 const __m128i round_offset = in comp_mask_pred_16_ssse3() local 45 const __m128i roundAL = _mm_mulhrs_epi16(blendAL, round_offset); in comp_mask_pred_16_ssse3() 46 const __m128i roundAH = _mm_mulhrs_epi16(blendAH, round_offset); in comp_mask_pred_16_ssse3() 57 const __m128i round_offset = in comp_mask_pred_8_ssse3() local 79 const __m128i roundA = _mm_mulhrs_epi16(blendA, round_offset); in comp_mask_pred_8_ssse3() 80 const __m128i roundB = _mm_mulhrs_epi16(blendB, round_offset); in comp_mask_pred_8_ssse3()
|
D | blend_a64_mask_avx2.c | 86 const __m256i *round_offset, int shift) { in lowbd_blend_a64_d16_mask_subw0_subh0_w16_avx2() argument 92 blend_a64_d16_mask_w16_avx2(dst, src0, src1, &m0, round_offset, &v_maxval, in lowbd_blend_a64_d16_mask_subw0_subh0_w16_avx2() 105 const __m256i *round_offset, int shift) { in lowbd_blend_a64_d16_mask_subw0_subh0_w32_avx2() argument 114 round_offset, &v_maxval, shift); in lowbd_blend_a64_d16_mask_subw0_subh0_w32_avx2() 127 const __m256i *round_offset, int shift) { in lowbd_blend_a64_d16_mask_subw1_subh1_w16_avx2() argument 139 blend_a64_d16_mask_w16_avx2(dst, src0, src1, &m0, round_offset, &v_maxval, in lowbd_blend_a64_d16_mask_subw1_subh1_w16_avx2() 152 const __m256i *round_offset, int shift) { in lowbd_blend_a64_d16_mask_subw1_subh1_w32_avx2() argument 171 round_offset, &v_maxval, shift); in lowbd_blend_a64_d16_mask_subw1_subh1_w32_avx2() 184 const __m256i *round_offset, int shift) { in lowbd_blend_a64_d16_mask_subw1_subh0_w16_avx2() argument 195 round_offset, &v_maxval, shift); in lowbd_blend_a64_d16_mask_subw1_subh0_w16_avx2() [all …]
|
D | blend_a64_mask_sse4.c | 908 const __m128i *round_offset, int shift) { in lowbd_blend_a64_d16_mask_subw0_subh0_w16_sse4_1() argument 917 round_offset, &v_maxval, shift); in lowbd_blend_a64_d16_mask_subw0_subh0_w16_sse4_1() 930 const __m128i *round_offset, int shift) { in lowbd_blend_a64_d16_mask_subw1_subh1_w16_sse4_1() argument 949 round_offset, &v_maxval, shift); in lowbd_blend_a64_d16_mask_subw1_subh1_w16_sse4_1() 962 const __m128i *round_offset, int shift) { in lowbd_blend_a64_d16_mask_subw1_subh0_w16_sse4_1() argument 976 round_offset, &v_maxval, shift); in lowbd_blend_a64_d16_mask_subw1_subh0_w16_sse4_1() 989 const __m128i *round_offset, int shift) { in lowbd_blend_a64_d16_mask_subw0_subh1_w16_sse4_1() argument 1002 round_offset, &v_maxval, shift); in lowbd_blend_a64_d16_mask_subw0_subh1_w16_sse4_1() 1020 const int round_offset = in aom_lowbd_blend_a64_d16_mask_sse4_1() local 1034 const __m128i v_round_offset = _mm_set1_epi32(round_offset); in aom_lowbd_blend_a64_d16_mask_sse4_1() [all …]
|
D | variance_avx2.c | 347 const __m256i round_offset = _mm256_set1_epi16(1 << (round_bits)); in comp_mask_pred_line_avx2() local 358 const __m256i roundAL = _mm256_mulhrs_epi16(blendAL, round_offset); in comp_mask_pred_line_avx2() 359 const __m256i roundAH = _mm256_mulhrs_epi16(blendAH, round_offset); in comp_mask_pred_line_avx2()
|
/external/libaom/libaom/aom_dsp/ |
D | blend_a64_mask.c | 44 const int round_offset = (1 << (offset_bits - conv_params->round_1)) + in aom_lowbd_blend_a64_d16_mask_c() local 66 res -= round_offset; in aom_lowbd_blend_a64_d16_mask_c() 85 res -= round_offset; in aom_lowbd_blend_a64_d16_mask_c() 100 res -= round_offset; in aom_lowbd_blend_a64_d16_mask_c() 115 res -= round_offset; in aom_lowbd_blend_a64_d16_mask_c() 130 const int round_offset = (1 << (offset_bits - conv_params->round_1)) + in aom_highbd_blend_a64_d16_mask_c() local 161 res -= round_offset; in aom_highbd_blend_a64_d16_mask_c() 180 res -= round_offset; in aom_highbd_blend_a64_d16_mask_c() 196 res -= round_offset; in aom_highbd_blend_a64_d16_mask_c() 212 res -= round_offset; in aom_highbd_blend_a64_d16_mask_c()
|
/external/libaom/libaom/av1/common/ |
D | convolve.c | 324 const int round_offset = (1 << (offset_bits - conv_params->round_1)) + in av1_dist_wtd_convolve_y_c() local 341 res = ROUND_POWER_OF_TWO(res, conv_params->round_1) + round_offset; in av1_dist_wtd_convolve_y_c() 352 tmp -= round_offset; in av1_dist_wtd_convolve_y_c() 374 const int round_offset = (1 << (offset_bits - conv_params->round_1)) + in av1_dist_wtd_convolve_x_c() local 391 res += round_offset; in av1_dist_wtd_convolve_x_c() 402 tmp -= round_offset; in av1_dist_wtd_convolve_x_c() 425 const int round_offset = (1 << (offset_bits - conv_params->round_1)) + in av1_dist_wtd_convolve_2d_copy_c() local 435 res += round_offset; in av1_dist_wtd_convolve_2d_copy_c() 446 tmp -= round_offset; in av1_dist_wtd_convolve_2d_copy_c() 813 const int round_offset = (1 << (offset_bits - conv_params->round_1)) + in av1_highbd_dist_wtd_convolve_x_c() local [all …]
|
D | cfl.h | 166 #define CFL_SUB_AVG_X(arch, width, height, round_offset, num_pel_log2) \ argument 169 subtract_average_##arch(src, dst, width, height, round_offset, \
|
D | cfl.c | 118 int height, int round_offset, int num_pel_log2) { in subtract_average_c() argument 119 int sum = round_offset; in subtract_average_c()
|
/external/libaom/libaom/av1/common/x86/ |
D | cfl_sse2.c | 24 int height, int round_offset, in subtract_average_sse2() argument 27 const __m128i round_offset_epi32 = _mm_set1_epi32(round_offset); in subtract_average_sse2()
|
D | cfl_avx2.c | 404 int height, int round_offset, in subtract_average_avx2() argument 441 _mm256_add_epi32(fill, _mm256_set1_epi32(round_offset)), num_pel_log2); in subtract_average_avx2()
|
/external/libaom/libaom/av1/common/ppc/ |
D | cfl_ppc.c | 36 int width, int height, int round_offset, in subtract_average_vsx() argument 48 int32x4_t sum_32x4_0 = { 0, 0, 0, round_offset }; in subtract_average_vsx()
|
/external/libaom/libaom/av1/common/arm/ |
D | jnt_convolve_neon.c | 772 const int round_offset = (1 << (offset_bits - conv_params->round_1)) + in av1_dist_wtd_convolve_2d_copy_neon() local 774 const int16x4_t sub_const_vec = vdup_n_s16((int16_t)round_offset); in av1_dist_wtd_convolve_2d_copy_neon() 775 const uint16x8_t dup_round_offset16x8 = vdupq_n_u16((uint16_t)round_offset); in av1_dist_wtd_convolve_2d_copy_neon() 894 const int round_offset = (1 << (offset_bits - conv_params->round_1)) + in av1_dist_wtd_convolve_x_neon() local 948 const int16x4_t round_offset_vec = vdup_n_s16(round_offset); in av1_dist_wtd_convolve_x_neon() 1126 const int16x8_t round_offset128 = vdupq_n_s16(round_offset); in av1_dist_wtd_convolve_x_neon() 1127 const int16x4_t round_offset64 = vdup_n_s16(round_offset); in av1_dist_wtd_convolve_x_neon() 1357 const int round_offset = (1 << (offset_bits - conv_params->round_1)) + in av1_dist_wtd_convolve_y_neon() local 1412 const int16x4_t round_offset64 = vdup_n_s16(round_offset); in av1_dist_wtd_convolve_y_neon() 1567 const int16x8_t round_offset128 = vdupq_n_s16(round_offset); in av1_dist_wtd_convolve_y_neon() [all …]
|
D | cfl_neon.c | 257 int round_offset, in CFL_GET_SUBSAMPLE_FUNCTION() 262 (void)round_offset; in CFL_GET_SUBSAMPLE_FUNCTION()
|
/external/libaom/libaom/aom_dsp/arm/ |
D | blend_a64_mask_neon.c | 144 const int round_offset = (1 << (offset_bits - conv_params->round_1)) + in aom_lowbd_blend_a64_d16_mask_neon() local 169 const uint16_t offset = round_offset - (1 << (round_bits - 1)); in aom_lowbd_blend_a64_d16_mask_neon()
|
/external/tensorflow/tensorflow/core/data/service/ |
D | dispatcher_impl.cc | 738 int64 round_offset = 2; in ClientHeartbeat() local 740 round_offset *= 2; in ClientHeartbeat() 743 round_robin_rounds_[request->job_client_id()] + round_offset); in ClientHeartbeat()
|