Home
last modified time | relevance | path

Searched refs:int16x8_t (Results 1 – 25 of 179) sorted by relevance

12345678

/external/libvpx/libvpx/vpx_dsp/ppc/
Dtxfm_common_vsx.h22 static const int16x8_t cospi1_v = { 16364, 16364, 16364, 16364,
24 static const int16x8_t cospi2_v = { 16305, 16305, 16305, 16305,
26 static const int16x8_t cospi3_v = { 16207, 16207, 16207, 16207,
28 static const int16x8_t cospi4_v = { 16069, 16069, 16069, 16069,
30 static const int16x8_t cospi4m_v = { -16069, -16069, -16069, -16069,
32 static const int16x8_t cospi5_v = { 15893, 15893, 15893, 15893,
34 static const int16x8_t cospi6_v = { 15679, 15679, 15679, 15679,
36 static const int16x8_t cospi7_v = { 15426, 15426, 15426, 15426,
38 static const int16x8_t cospi8_v = { 15137, 15137, 15137, 15137,
40 static const int16x8_t cospi8m_v = { -15137, -15137, -15137, -15137,
[all …]
Dhadamard_vsx.c16 static void vpx_hadamard_s16_8x8_one_pass(int16x8_t v[8]) { in vpx_hadamard_s16_8x8_one_pass()
17 const int16x8_t b0 = vec_add(v[0], v[1]); in vpx_hadamard_s16_8x8_one_pass()
18 const int16x8_t b1 = vec_sub(v[0], v[1]); in vpx_hadamard_s16_8x8_one_pass()
19 const int16x8_t b2 = vec_add(v[2], v[3]); in vpx_hadamard_s16_8x8_one_pass()
20 const int16x8_t b3 = vec_sub(v[2], v[3]); in vpx_hadamard_s16_8x8_one_pass()
21 const int16x8_t b4 = vec_add(v[4], v[5]); in vpx_hadamard_s16_8x8_one_pass()
22 const int16x8_t b5 = vec_sub(v[4], v[5]); in vpx_hadamard_s16_8x8_one_pass()
23 const int16x8_t b6 = vec_add(v[6], v[7]); in vpx_hadamard_s16_8x8_one_pass()
24 const int16x8_t b7 = vec_sub(v[6], v[7]); in vpx_hadamard_s16_8x8_one_pass()
26 const int16x8_t c0 = vec_add(b0, b2); in vpx_hadamard_s16_8x8_one_pass()
[all …]
Dquantize_vsx.c18 static INLINE int16x8_t vec_sign(int16x8_t a, int16x8_t b) { in vec_sign()
19 const int16x8_t mask = vec_sra(b, vec_shift_sign_s16); in vec_sign()
32 static INLINE int16x8_t vec_mulhi(int16x8_t a, int16x8_t b) { in vec_mulhi()
39 static INLINE int16x8_t quantize_coeff(int16x8_t coeff, int16x8_t coeff_abs, in quantize_coeff()
40 int16x8_t round, int16x8_t quant, in quantize_coeff()
41 int16x8_t quant_shift, bool16x8_t mask) { in quantize_coeff()
42 const int16x8_t rounded = vec_vaddshs(coeff_abs, round); in quantize_coeff()
43 int16x8_t qcoeff = vec_mulhi(rounded, quant); in quantize_coeff()
51 static INLINE int16x8_t quantize_coeff_32(int16x8_t coeff, int16x8_t coeff_abs, in quantize_coeff_32()
52 int16x8_t round, int16x8_t quant, in quantize_coeff_32()
[all …]
Dtranspose_vsx.h17 static INLINE void vpx_transpose_s16_8x8(int16x8_t v[8]) { in vpx_transpose_s16_8x8()
40 int16x8_t b0, b1, b2, b3, b4, b5, b6, b7; in vpx_transpose_s16_8x8()
41 int16x8_t c0, c1, c2, c3, c4, c5, c6, c7; in vpx_transpose_s16_8x8()
101 static INLINE void transpose_8x8(const int16x8_t *a, int16x8_t *b) { in transpose_8x8()
103 const int16x8_t s1_0 = vec_mergeh(a[0], a[4]); in transpose_8x8()
104 const int16x8_t s1_1 = vec_mergel(a[0], a[4]); in transpose_8x8()
105 const int16x8_t s1_2 = vec_mergeh(a[1], a[5]); in transpose_8x8()
106 const int16x8_t s1_3 = vec_mergel(a[1], a[5]); in transpose_8x8()
107 const int16x8_t s1_4 = vec_mergeh(a[2], a[6]); in transpose_8x8()
108 const int16x8_t s1_5 = vec_mergel(a[2], a[6]); in transpose_8x8()
[all …]
Dinv_txfm_vsx.h16 void vpx_round_store4x4_vsx(int16x8_t *in, int16x8_t *out, uint8_t *dest,
18 void vpx_idct4_vsx(int16x8_t *in, int16x8_t *out);
19 void vp9_iadst4_vsx(int16x8_t *in, int16x8_t *out);
21 void vpx_round_store8x8_vsx(int16x8_t *in, uint8_t *dest, int stride);
22 void vpx_idct8_vsx(int16x8_t *in, int16x8_t *out);
23 void vp9_iadst8_vsx(int16x8_t *in, int16x8_t *out);
43 void vpx_round_store16x16_vsx(int16x8_t *src0, int16x8_t *src1, uint8_t *dest,
45 void vpx_idct16_vsx(int16x8_t *src0, int16x8_t *src1);
46 void vpx_iadst16_vsx(int16x8_t *src0, int16x8_t *src1);
Dfdct32x32_vsx.c19 static INLINE void single_butterfly(int16x8_t a, int16x8_t b, int16x8_t *add, in single_butterfly()
20 int16x8_t *sub) { in single_butterfly()
46 *add = (int16x8_t)vec_perm(ssum_e, ssum_o, vec_perm_odd_even_pack); in single_butterfly()
47 *sub = (int16x8_t)vec_perm(sdiff_e, sdiff_o, vec_perm_odd_even_pack); in single_butterfly()
51 static INLINE void double_butterfly(int16x8_t a, int16x8_t c1, int16x8_t b, in double_butterfly()
52 int16x8_t c2, int16x8_t *add, in double_butterfly()
53 int16x8_t *sub) { in double_butterfly()
81 *add = (int16x8_t)vec_perm(ssum_e, ssum_o, vec_perm_odd_even_pack); in double_butterfly()
82 *sub = (int16x8_t)vec_perm(sdiff_e, sdiff_o, vec_perm_odd_even_pack); in double_butterfly()
87 static INLINE void load(const int16_t *a, int stride, int16x8_t *b) { in load()
[all …]
Dvariance_vsx.c21 const int16x8_t a0 = unpack_to_s16_h(read4x2(src_ptr, src_stride)); in vpx_get4x4sse_cs_vsx()
22 const int16x8_t a1 = in vpx_get4x4sse_cs_vsx()
24 const int16x8_t b0 = unpack_to_s16_h(read4x2(ref_ptr, ref_stride)); in vpx_get4x4sse_cs_vsx()
25 const int16x8_t b1 = in vpx_get4x4sse_cs_vsx()
27 const int16x8_t d0 = vec_sub(a0, b0); in vpx_get4x4sse_cs_vsx()
28 const int16x8_t d1 = vec_sub(a1, b1); in vpx_get4x4sse_cs_vsx()
43 const int16x8_t v = vec_vsx_ld(0, src_ptr + i); in vpx_get_mb_ss_vsx()
112 const int16x8_t a0 = unpack_to_s16_h(va0); in variance_inner_32()
113 const int16x8_t b0 = unpack_to_s16_h(vb0); in variance_inner_32()
114 const int16x8_t a1 = unpack_to_s16_l(va0); in variance_inner_32()
[all …]
Dinv_txfm_vsx.c22 static const int16x8_t cospi1_v = { 16364, 16364, 16364, 16364,
24 static const int16x8_t cospi1m_v = { -16364, -16364, -16364, -16364,
26 static const int16x8_t cospi2_v = { 16305, 16305, 16305, 16305,
28 static const int16x8_t cospi2m_v = { -16305, -16305, -16305, -16305,
30 static const int16x8_t cospi3_v = { 16207, 16207, 16207, 16207,
32 static const int16x8_t cospi4_v = { 16069, 16069, 16069, 16069,
34 static const int16x8_t cospi4m_v = { -16069, -16069, -16069, -16069,
36 static const int16x8_t cospi5_v = { 15893, 15893, 15893, 15893,
38 static const int16x8_t cospi5m_v = { -15893, -15893, -15893, -15893,
40 static const int16x8_t cospi6_v = { 15679, 15679, 15679, 15679,
[all …]
/external/gemmlowp/fixedpoint/
Dfixedpoint_neon.h32 struct FixedPointRawTypeTraits<int16x8_t> {
43 inline int16x8_t BitAnd(int16x8_t a, int16x8_t b) {
53 inline int16x8_t BitOr(int16x8_t a, int16x8_t b) {
63 inline int16x8_t BitXor(int16x8_t a, int16x8_t b) {
73 inline int16x8_t BitNot(int16x8_t a) {
83 inline int16x8_t Add(int16x8_t a, int16x8_t b) {
93 inline int16x8_t Sub(int16x8_t a, int16x8_t b) {
103 inline int16x8_t Neg(int16x8_t a) {
113 inline int16x8_t ShiftLeft(int16x8_t a, int offset) {
123 inline int16x8_t ShiftRight(int16x8_t a, int offset) {
[all …]
/external/libvpx/libvpx/vpx_dsp/arm/
Dhadamard_neon.c19 static void hadamard8x8_one_pass(int16x8_t *a0, int16x8_t *a1, int16x8_t *a2, in hadamard8x8_one_pass()
20 int16x8_t *a3, int16x8_t *a4, int16x8_t *a5, in hadamard8x8_one_pass()
21 int16x8_t *a6, int16x8_t *a7) { in hadamard8x8_one_pass()
22 const int16x8_t b0 = vaddq_s16(*a0, *a1); in hadamard8x8_one_pass()
23 const int16x8_t b1 = vsubq_s16(*a0, *a1); in hadamard8x8_one_pass()
24 const int16x8_t b2 = vaddq_s16(*a2, *a3); in hadamard8x8_one_pass()
25 const int16x8_t b3 = vsubq_s16(*a2, *a3); in hadamard8x8_one_pass()
26 const int16x8_t b4 = vaddq_s16(*a4, *a5); in hadamard8x8_one_pass()
27 const int16x8_t b5 = vsubq_s16(*a4, *a5); in hadamard8x8_one_pass()
28 const int16x8_t b6 = vaddq_s16(*a6, *a7); in hadamard8x8_one_pass()
[all …]
Dquantize_neon.c18 static INLINE void calculate_dqcoeff_and_store(const int16x8_t qcoeff, in calculate_dqcoeff_and_store()
19 const int16x8_t dequant, in calculate_dqcoeff_and_store()
41 const int16x8_t one = vdupq_n_s16(1); in vpx_quantize_b_neon()
42 const int16x8_t neg_one = vdupq_n_s16(-1); in vpx_quantize_b_neon()
51 const int16x8_t zbin = vld1q_s16(zbin_ptr); in vpx_quantize_b_neon()
52 const int16x8_t round = vld1q_s16(round_ptr); in vpx_quantize_b_neon()
53 const int16x8_t quant = vld1q_s16(quant_ptr); in vpx_quantize_b_neon()
54 const int16x8_t quant_shift = vld1q_s16(quant_shift_ptr); in vpx_quantize_b_neon()
55 const int16x8_t dequant = vld1q_s16(dequant_ptr); in vpx_quantize_b_neon()
60 const int16x8_t coeff = load_tran_low_to_s16q(coeff_ptr); in vpx_quantize_b_neon()
[all …]
Dhighbd_idct32x32_add_neon.c19 const int16x8_t res, in highbd_idct32x32_1_add_pos_kernel()
20 const int16x8_t max) { in highbd_idct32x32_1_add_pos_kernel()
25 const int16x8_t b0 = vaddq_s16(res, vreinterpretq_s16_u16(a0)); in highbd_idct32x32_1_add_pos_kernel()
26 const int16x8_t b1 = vaddq_s16(res, vreinterpretq_s16_u16(a1)); in highbd_idct32x32_1_add_pos_kernel()
27 const int16x8_t b2 = vaddq_s16(res, vreinterpretq_s16_u16(a2)); in highbd_idct32x32_1_add_pos_kernel()
28 const int16x8_t b3 = vaddq_s16(res, vreinterpretq_s16_u16(a3)); in highbd_idct32x32_1_add_pos_kernel()
29 const int16x8_t c0 = vminq_s16(b0, max); in highbd_idct32x32_1_add_pos_kernel()
30 const int16x8_t c1 = vminq_s16(b1, max); in highbd_idct32x32_1_add_pos_kernel()
31 const int16x8_t c2 = vminq_s16(b2, max); in highbd_idct32x32_1_add_pos_kernel()
32 const int16x8_t c3 = vminq_s16(b3, max); in highbd_idct32x32_1_add_pos_kernel()
[all …]
Dfwd_txfm_neon.c24 int16x8_t input_0 = vshlq_n_s16(vld1q_s16(&input[0 * stride]), 2); in vpx_fdct8x8_neon()
25 int16x8_t input_1 = vshlq_n_s16(vld1q_s16(&input[1 * stride]), 2); in vpx_fdct8x8_neon()
26 int16x8_t input_2 = vshlq_n_s16(vld1q_s16(&input[2 * stride]), 2); in vpx_fdct8x8_neon()
27 int16x8_t input_3 = vshlq_n_s16(vld1q_s16(&input[3 * stride]), 2); in vpx_fdct8x8_neon()
28 int16x8_t input_4 = vshlq_n_s16(vld1q_s16(&input[4 * stride]), 2); in vpx_fdct8x8_neon()
29 int16x8_t input_5 = vshlq_n_s16(vld1q_s16(&input[5 * stride]), 2); in vpx_fdct8x8_neon()
30 int16x8_t input_6 = vshlq_n_s16(vld1q_s16(&input[6 * stride]), 2); in vpx_fdct8x8_neon()
31 int16x8_t input_7 = vshlq_n_s16(vld1q_s16(&input[7 * stride]), 2); in vpx_fdct8x8_neon()
33 int16x8_t out_0, out_1, out_2, out_3, out_4, out_5, out_6, out_7; in vpx_fdct8x8_neon()
34 const int16x8_t v_s0 = vaddq_s16(input_0, input_7); in vpx_fdct8x8_neon()
[all …]
Dfdct_partial_neon.c28 int16x8_t b0, b1; in vpx_fdct4x4_1_neon()
29 int16x8_t c; in vpx_fdct4x4_1_neon()
53 int16x8_t sum = vld1q_s16(&input[0]); in vpx_fdct8x8_1_neon()
56 const int16x8_t input_00 = vld1q_s16(&input[r * stride]); in vpx_fdct8x8_1_neon()
67 int16x8_t left = vld1q_s16(input); in vpx_fdct16x16_1_neon()
68 int16x8_t right = vld1q_s16(input + 8); in vpx_fdct16x16_1_neon()
73 const int16x8_t a = vld1q_s16(input); in vpx_fdct16x16_1_neon()
74 const int16x8_t b = vld1q_s16(input + 8); in vpx_fdct16x16_1_neon()
89 int16x8_t a0 = vld1q_s16(input); in vpx_fdct32x32_1_neon()
90 int16x8_t a1 = vld1q_s16(input + 8); in vpx_fdct32x32_1_neon()
[all …]
Didct_neon.h45 static INLINE int16x8_t final_add(const int16x8_t a, const int16x8_t b) { in final_add()
53 static INLINE int16x8_t final_sub(const int16x8_t a, const int16x8_t b) { in final_sub()
81 static INLINE int16x8_t dct_const_round_shift_low_8(const int32x4_t *const in) { in dct_const_round_shift_low_8()
87 int16x8_t *const d0, in dct_const_round_shift_low_8_dual()
88 int16x8_t *const d1) { in dct_const_round_shift_low_8_dual()
104 static INLINE int16x8_t multiply_shift_and_narrow_s16(const int16x8_t a, in multiply_shift_and_narrow_s16()
118 static INLINE int16x8_t add_multiply_shift_and_narrow_s16( in add_multiply_shift_and_narrow_s16()
119 const int16x8_t a, const int16x8_t b, const int16_t ab_const) { in add_multiply_shift_and_narrow_s16()
137 static INLINE int16x8_t sub_multiply_shift_and_narrow_s16( in sub_multiply_shift_and_narrow_s16()
138 const int16x8_t a, const int16x8_t b, const int16_t ab_const) { in sub_multiply_shift_and_narrow_s16()
[all …]
/external/libvpx/libvpx/vp9/encoder/arm/neon/
Dvp9_quantize_neon.c38 const int16x8_t v_zero = vdupq_n_s16(0); in vp9_quantize_fp_neon()
39 const int16x8_t v_one = vdupq_n_s16(1); in vp9_quantize_fp_neon()
40 int16x8_t v_eobmax_76543210 = vdupq_n_s16(-1); in vp9_quantize_fp_neon()
41 int16x8_t v_round = vmovq_n_s16(round_ptr[1]); in vp9_quantize_fp_neon()
42 int16x8_t v_quant = vmovq_n_s16(quant_ptr[1]); in vp9_quantize_fp_neon()
43 int16x8_t v_dequant = vmovq_n_s16(dequant_ptr[1]); in vp9_quantize_fp_neon()
55 const int16x8_t v_iscan = vld1q_s16(&iscan[0]); in vp9_quantize_fp_neon()
56 const int16x8_t v_coeff = load_tran_low_to_s16q(coeff_ptr); in vp9_quantize_fp_neon()
57 const int16x8_t v_coeff_sign = vshrq_n_s16(v_coeff, 15); in vp9_quantize_fp_neon()
58 const int16x8_t v_tmp = vabaq_s16(v_round, v_coeff, v_zero); in vp9_quantize_fp_neon()
[all …]
/external/libhevc/encoder/arm/
Dihevce_common_utils_neon.c113 int16x8_t a2, a3, a4, a5, a8; in ihevce_wt_avg_2d_16x1_neon()
179 int16x8_t a0, a1, a6; in ihevce_wt_avg_2d_8x1_neon()
234 int16x8_t a8, a9, a6; in ihevce_wt_avg_2d_4xn_neon()
448 static INLINE WORD32 sad_cal(int16x8_t temp_reg) in sad_cal()
473 int16x8_t temp_reg0, temp_reg1, temp_reg2, temp_reg3, temp_reg4; in ihevce_get_luma_eo_sao_params_neon()
474 int16x8_t edgeidx_reg0, edgeidx_reg1, edgeidx_reg2, edgeidx_reg3, edgeidx_reg4; in ihevce_get_luma_eo_sao_params_neon()
475 int16x8_t edgeidx_reg5, edgeidx_reg6, edgeidx_reg7; in ihevce_get_luma_eo_sao_params_neon()
476 int16x8_t pel_error, pel_error1; in ihevce_get_luma_eo_sao_params_neon()
477 int16x8_t sign_reg0, sign_reg1, sign_reg, sign_reg2, sign_reg3; in ihevce_get_luma_eo_sao_params_neon()
478 int16x8_t edgeidx, edgeidx1; in ihevce_get_luma_eo_sao_params_neon()
[all …]
Dihevce_had_compute_neon.c85 hadamard4x4_2_one_pass(int16x8_t *r0, int16x8_t *r1, int16x8_t *r2, int16x8_t *r3) in hadamard4x4_2_one_pass()
87 const int16x8_t a0 = vaddq_s16(*r0, *r2); in hadamard4x4_2_one_pass()
88 const int16x8_t a1 = vaddq_s16(*r1, *r3); in hadamard4x4_2_one_pass()
89 const int16x8_t a2 = vsubq_s16(*r0, *r2); in hadamard4x4_2_one_pass()
90 const int16x8_t a3 = vsubq_s16(*r1, *r3); in hadamard4x4_2_one_pass()
103 int16x8_t *r0, in hadamard4x4_2()
104 int16x8_t *r1, in hadamard4x4_2()
105 int16x8_t *r2, in hadamard4x4_2()
106 int16x8_t *r3) in hadamard4x4_2()
129 int16x8_t *r0, in hadamard4x4_4()
[all …]
/external/libaom/libaom/av1/encoder/arm/neon/
Dquantize_neon.c42 const int16x8_t v_zero = vdupq_n_s16(0); in av1_quantize_fp_neon()
43 const int16x8_t v_one = vdupq_n_s16(1); in av1_quantize_fp_neon()
44 int16x8_t v_eobmax_76543210 = vdupq_n_s16(-1); in av1_quantize_fp_neon()
45 int16x8_t v_round = vmovq_n_s16(round_ptr[1]); in av1_quantize_fp_neon()
46 int16x8_t v_quant = vmovq_n_s16(quant_ptr[1]); in av1_quantize_fp_neon()
47 int16x8_t v_dequant = vmovq_n_s16(dequant_ptr[1]); in av1_quantize_fp_neon()
54 const int16x8_t v_iscan = vld1q_s16(&iscan[0]); in av1_quantize_fp_neon()
55 const int16x8_t v_coeff = vld1q_s16(&coeff_ptr[0]); in av1_quantize_fp_neon()
56 const int16x8_t v_coeff_sign = vshrq_n_s16(v_coeff, 15); in av1_quantize_fp_neon()
57 const int16x8_t v_tmp = vabaq_s16(v_round, v_coeff, v_zero); in av1_quantize_fp_neon()
[all …]
/external/webp/src/dsp/
Dyuv_neon.c105 const int16x8_t tmp3 = vcombine_s16(vshrn_n_s32(tmp2_lo, 16), \
122 int16x8_t U, V; in ConvertRGBA32ToUV_NEON()
142 int16x8_t U_tmp, V_tmp; in ConvertARGBToUV_NEON()
186 const int16x8_t zero = vdupq_n_s16(0); in SharpYUVUpdateY_NEON()
187 const int16x8_t max = vdupq_n_s16(MAX_Y); in SharpYUVUpdateY_NEON()
192 const int16x8_t A = vreinterpretq_s16_u16(vld1q_u16(ref + i)); in SharpYUVUpdateY_NEON()
193 const int16x8_t B = vreinterpretq_s16_u16(vld1q_u16(src + i)); in SharpYUVUpdateY_NEON()
194 const int16x8_t C = vreinterpretq_s16_u16(vld1q_u16(dst + i)); in SharpYUVUpdateY_NEON()
195 const int16x8_t D = vsubq_s16(A, B); // diff_y in SharpYUVUpdateY_NEON()
196 const int16x8_t F = vaddq_s16(C, D); // new_y in SharpYUVUpdateY_NEON()
[all …]
/external/clang/test/CodeGen/
Daarch64-neon-2velem.c21 int16x8_t test_vmlaq_lane_s16(int16x8_t a, int16x8_t b, int16x4_t v) { in test_vmlaq_lane_s16()
48 int16x4_t test_vmla_laneq_s16(int16x4_t a, int16x4_t b, int16x8_t v) { in test_vmla_laneq_s16()
57 int16x8_t test_vmlaq_laneq_s16(int16x8_t a, int16x8_t b, int16x8_t v) { in test_vmlaq_laneq_s16()
93 int16x8_t test_vmlsq_lane_s16(int16x8_t a, int16x8_t b, int16x4_t v) { in test_vmlsq_lane_s16()
120 int16x4_t test_vmls_laneq_s16(int16x4_t a, int16x4_t b, int16x8_t v) { in test_vmls_laneq_s16()
129 int16x8_t test_vmlsq_laneq_s16(int16x8_t a, int16x8_t b, int16x8_t v) { in test_vmlsq_laneq_s16()
163 int16x8_t test_vmulq_lane_s16(int16x8_t a, int16x4_t v) { in test_vmulq_lane_s16()
219 int16x4_t test_vmul_laneq_s16(int16x4_t a, int16x8_t v) { in test_vmul_laneq_s16()
227 int16x8_t test_vmulq_laneq_s16(int16x8_t a, int16x8_t v) { in test_vmulq_laneq_s16()
531 int32x4_t test_vmlal_laneq_s16(int32x4_t a, int16x4_t b, int16x8_t v) { in test_vmlal_laneq_s16()
[all …]
Darm-v8.1a-neon-intrinsics.c34 int16x8_t test_vqrdmlahq_s16(int16x8_t a, int16x8_t b, int16x8_t c) { in test_vqrdmlahq_s16()
78 int16x8_t test_vqrdmlahq_lane_s16(int16x8_t a, int16x8_t b, int16x4_t c) { in test_vqrdmlahq_lane_s16()
122 int16x8_t test_vqrdmlshq_s16(int16x8_t a, int16x8_t b, int16x8_t c) { in test_vqrdmlshq_s16()
166 int16x8_t test_vqrdmlshq_lane_s16(int16x8_t a, int16x8_t b, int16x4_t c) { in test_vqrdmlshq_lane_s16()
/external/libvpx/libvpx/vp9/encoder/ppc/
Dvp9_quantize_vsx.c20 static INLINE int16x8_t vec_mulhi(int16x8_t a, int16x8_t b) { in vec_mulhi()
28 static INLINE int16x8_t vec_sign(int16x8_t a, int16x8_t b) { in vec_sign()
29 const int16x8_t mask = vec_sra(b, vec_shift_sign_s16); in vec_sign()
35 static INLINE int16x8_t vec_max_across(int16x8_t a) { in vec_max_across()
47 int16x8_t qcoeff0, qcoeff1, dqcoeff0, dqcoeff1, eob; in vp9_quantize_fp_vsx()
50 int16x8_t round = vec_vsx_ld(0, round_ptr); in vp9_quantize_fp_vsx()
51 int16x8_t quant = vec_vsx_ld(0, quant_ptr); in vp9_quantize_fp_vsx()
52 int16x8_t dequant = vec_vsx_ld(0, dequant_ptr); in vp9_quantize_fp_vsx()
53 int16x8_t coeff0 = vec_vsx_ld(0, coeff_ptr); in vp9_quantize_fp_vsx()
54 int16x8_t coeff1 = vec_vsx_ld(16, coeff_ptr); in vp9_quantize_fp_vsx()
[all …]
/external/libaom/libaom/aom_dsp/arm/
Dfwd_txfm_neon.c21 int16x8_t input_0 = vshlq_n_s16(vld1q_s16(&input[0 * stride]), 2); in aom_fdct8x8_neon()
22 int16x8_t input_1 = vshlq_n_s16(vld1q_s16(&input[1 * stride]), 2); in aom_fdct8x8_neon()
23 int16x8_t input_2 = vshlq_n_s16(vld1q_s16(&input[2 * stride]), 2); in aom_fdct8x8_neon()
24 int16x8_t input_3 = vshlq_n_s16(vld1q_s16(&input[3 * stride]), 2); in aom_fdct8x8_neon()
25 int16x8_t input_4 = vshlq_n_s16(vld1q_s16(&input[4 * stride]), 2); in aom_fdct8x8_neon()
26 int16x8_t input_5 = vshlq_n_s16(vld1q_s16(&input[5 * stride]), 2); in aom_fdct8x8_neon()
27 int16x8_t input_6 = vshlq_n_s16(vld1q_s16(&input[6 * stride]), 2); in aom_fdct8x8_neon()
28 int16x8_t input_7 = vshlq_n_s16(vld1q_s16(&input[7 * stride]), 2); in aom_fdct8x8_neon()
30 int16x8_t out_0, out_1, out_2, out_3, out_4, out_5, out_6, out_7; in aom_fdct8x8_neon()
31 const int16x8_t v_s0 = vaddq_s16(input_0, input_7); in aom_fdct8x8_neon()
[all …]
/external/libaom/libaom/av1/common/arm/
Dconvolve_neon.h19 const int16x8_t s0, const int16x8_t s1, const int16x8_t s2, in wiener_convolve8_vert_4x8()
20 const int16x8_t s3, const int16x8_t s4, const int16x8_t s5, in wiener_convolve8_vert_4x8()
21 const int16x8_t s6, int16_t *filter_y, const int bd, in wiener_convolve8_vert_4x8()
23 int16x8_t ss0, ss1, ss2; in wiener_convolve8_vert_4x8()
68 const int16x8_t s0, const int16x8_t s1, const int16x8_t s2, in wiener_convolve8_horiz_8x8()
69 const int16x8_t s3, int16_t *filter_x, const int bd, in wiener_convolve8_horiz_8x8()
71 int16x8_t sum; in wiener_convolve8_horiz_8x8()
158 static INLINE int16x8_t
159 convolve8_8x8_s16(const int16x8_t s0, const int16x8_t s1, const int16x8_t s2, in convolve8_8x8_s16()
160 const int16x8_t s3, const int16x8_t s4, const int16x8_t s5, in convolve8_8x8_s16()
[all …]

12345678