Home
last modified time | relevance | path

Searched refs:vreinterpret_u16_u8 (Results 1 – 25 of 29) sorted by relevance

12

/external/XNNPACK/src/x8-zip/
Dxm-neon.c44 …const uint16x4x2_t vxyzw_lo = vzip_u16(vreinterpret_u16_u8(vxy.val[0]), vreinterpret_u16_u8(vzw.va… in xnn_x8_zip_xm_ukernel__neon()
45 …const uint16x4x2_t vxyzw_hi = vzip_u16(vreinterpret_u16_u8(vxy.val[1]), vreinterpret_u16_u8(vzw.va… in xnn_x8_zip_xm_ukernel__neon()
87 …const uint16x4x2_t vxyzw_lo = vzip_u16(vreinterpret_u16_u8(vxy.val[0]), vreinterpret_u16_u8(vzw.va… in xnn_x8_zip_xm_ukernel__neon()
88 …const uint16x4x2_t vxyzw_hi = vzip_u16(vreinterpret_u16_u8(vxy.val[1]), vreinterpret_u16_u8(vzw.va… in xnn_x8_zip_xm_ukernel__neon()
/external/libaom/libaom/av1/common/arm/
Dblend_a64_vmask_neon.c117 vtrn_u16(vreinterpret_u16_u8(m1), vreinterpret_u16_u8(m2)); in aom_blend_a64_vmask_neon()
122 vreinterpret_u16_u8(max_minus_m1), vreinterpret_u16_u8(max_minus_m2)); in aom_blend_a64_vmask_neon()
132 vreinterpret_u16_u8(vrshrn_n_u16(res, AOM_BLEND_A64_ROUND_BITS)), 0); in aom_blend_a64_vmask_neon()
135 vreinterpret_u16_u8(vrshrn_n_u16(res, AOM_BLEND_A64_ROUND_BITS)), 1); in aom_blend_a64_vmask_neon()
Dconvolve_neon.c289 vreinterpret_u16_u8(d01), 0); in av1_convolve_x_sr_neon()
291 vreinterpret_u16_u8(d23), 0); in av1_convolve_x_sr_neon()
293 vreinterpret_u16_u8(d01), 2); in av1_convolve_x_sr_neon()
295 vreinterpret_u16_u8(d23), 2); in av1_convolve_x_sr_neon()
404 vst1_lane_u16((uint16_t *)dst, vreinterpret_u16_u8(t0), 0); // 00 01 in av1_convolve_x_sr_neon()
406 vst1_lane_u16((uint16_t *)dst, vreinterpret_u16_u8(t1), 0); // 10 11 in av1_convolve_x_sr_neon()
408 vst1_lane_u16((uint16_t *)dst, vreinterpret_u16_u8(t2), 0); // 20 21 in av1_convolve_x_sr_neon()
410 vst1_lane_u16((uint16_t *)dst, vreinterpret_u16_u8(t3), 0); // 30 31 in av1_convolve_x_sr_neon()
412 vst1_lane_u16((uint16_t *)dst, vreinterpret_u16_u8(t0), 2); // 40 41 in av1_convolve_x_sr_neon()
414 vst1_lane_u16((uint16_t *)dst, vreinterpret_u16_u8(t1), 2); // 50 51 in av1_convolve_x_sr_neon()
[all …]
Dtranspose_neon.h93 vtrn_u16(vreinterpret_u16_u8(b0.val[0]), vreinterpret_u16_u8(b1.val[0])); in transpose_u8_8x4()
95 vtrn_u16(vreinterpret_u16_u8(b0.val[1]), vreinterpret_u16_u8(b1.val[1])); in transpose_u8_8x4()
112 vtrn_u16(vreinterpret_u16_u8(*a0), vreinterpret_u16_u8(*a1)); in transpose_u8_4x4()
Dblend_a64_hmask_neon.c125 vreinterpret_u16_u8(vrshrn_n_u16(res, AOM_BLEND_A64_ROUND_BITS)), 0); in aom_blend_a64_hmask_neon()
128 vreinterpret_u16_u8(vrshrn_n_u16(res, AOM_BLEND_A64_ROUND_BITS)), 1); in aom_blend_a64_hmask_neon()
/external/exoplayer/tree/extensions/vp9/src/main/jni/
Dvpx_jni.cc127 uint16x4_t added1 = vqadd_u16(values, vreinterpret_u16_u8(randvec)); in convert_16_to_8_neon()
137 uint16x4_t added2 = vqadd_u16(values, vreinterpret_u16_u8(randvec)); in convert_16_to_8_neon()
152 added1 = vqadd_u16(values, vreinterpret_u16_u8(randvec)); in convert_16_to_8_neon()
156 added2 = vqadd_u16(values, vreinterpret_u16_u8(randvec)); in convert_16_to_8_neon()
202 vreinterpret_u16_u8(randvec)); in convert_16_to_8_neon()
206 vreinterpret_u16_u8(randvec)); in convert_16_to_8_neon()
210 vreinterpret_u16_u8(randvec)); in convert_16_to_8_neon()
214 vreinterpret_u16_u8(randvec)); in convert_16_to_8_neon()
/external/libgav1/libgav1/src/dsp/arm/
Dcommon_neon.h210 vld1_lane_u16(&temp, vreinterpret_u16_u8(val), lane)); in Load2()
262 ValueToMem<uint16_t>(buf, vget_lane_u16(vreinterpret_u16_u8(val), lane)); in Store2()
506 vtrn_u16(vreinterpret_u16_u8(*a), vreinterpret_u16_u8(*b)); in Transpose4x4()
532 vtrn_u16(vreinterpret_u16_u8(b0.val[0]), vreinterpret_u16_u8(b1.val[0])); in Transpose8x4()
534 vtrn_u16(vreinterpret_u16_u8(b0.val[1]), vreinterpret_u16_u8(b1.val[1])); in Transpose8x4()
Dintrapred_directional_neon.cc856 const uint16x4x2_t c0 = vtrn_u16(vreinterpret_u16_u8(b0.val[0]), in DirectionalIntraPredictorZone3_NEON()
857 vreinterpret_u16_u8(b1.val[0])); in DirectionalIntraPredictorZone3_NEON()
858 const uint16x4x2_t c1 = vtrn_u16(vreinterpret_u16_u8(b0.val[1]), in DirectionalIntraPredictorZone3_NEON()
859 vreinterpret_u16_u8(b1.val[1])); in DirectionalIntraPredictorZone3_NEON()
/external/libjpeg-turbo/simd/arm/
Djidctint-neon.c684 vreinterpret_u16_u8(cols_01_23.val[0]), in jsimd_idct_islow_pass2_regular()
685 vreinterpret_u16_u8(cols_01_23.val[1]), in jsimd_idct_islow_pass2_regular()
686 vreinterpret_u16_u8(cols_45_67.val[0]), in jsimd_idct_islow_pass2_regular()
687 vreinterpret_u16_u8(cols_45_67.val[1]) in jsimd_idct_islow_pass2_regular()
787 vreinterpret_u16_u8(cols_01_23.val[0]), in jsimd_idct_islow_pass2_sparse()
788 vreinterpret_u16_u8(cols_01_23.val[1]), in jsimd_idct_islow_pass2_sparse()
789 vreinterpret_u16_u8(cols_45_67.val[0]), in jsimd_idct_islow_pass2_sparse()
790 vreinterpret_u16_u8(cols_45_67.val[1]) in jsimd_idct_islow_pass2_sparse()
Djidctred-neon.c473 vreinterpret_u16_u8(output_0123.val[0]), in jsimd_idct_4x4_neon()
474 vreinterpret_u16_u8(output_0123.val[1]) in jsimd_idct_4x4_neon()
/external/libvpx/libvpx/vp9/encoder/arm/neon/
Dvp9_frame_scale_neon.c340 vreinterpret_u16_u8(dd.val[0]), 0); in scale_plane_4_to_1_general()
342 vreinterpret_u16_u8(dd.val[1]), 0); in scale_plane_4_to_1_general()
344 vreinterpret_u16_u8(dd.val[0]), 1); in scale_plane_4_to_1_general()
346 vreinterpret_u16_u8(dd.val[1]), 1); in scale_plane_4_to_1_general()
348 vreinterpret_u16_u8(dd.val[0]), 2); in scale_plane_4_to_1_general()
350 vreinterpret_u16_u8(dd.val[1]), 2); in scale_plane_4_to_1_general()
352 vreinterpret_u16_u8(dd.val[0]), 3); in scale_plane_4_to_1_general()
354 vreinterpret_u16_u8(dd.val[1]), 3); in scale_plane_4_to_1_general()
/external/exoplayer/tree/extensions/av1/src/main/jni/
Dgav1_jni.cc449 uint16x4_t added_1 = vqadd_u16(values, vreinterpret_u16_u8(randvec)); in Convert10BitFrameTo8BitDataBufferNeon()
460 uint16x4_t added_2 = vqadd_u16(values, vreinterpret_u16_u8(randvec)); in Convert10BitFrameTo8BitDataBufferNeon()
477 added_1 = vqadd_u16(values, vreinterpret_u16_u8(randvec)); in Convert10BitFrameTo8BitDataBufferNeon()
483 added_2 = vqadd_u16(values, vreinterpret_u16_u8(randvec)); in Convert10BitFrameTo8BitDataBufferNeon()
/external/gemmlowp/internal/
Doutput_neon.h666 b[0] = vtrn_u16(vreinterpret_u16_u8(a[0].val[0]),
667 vreinterpret_u16_u8(a[1].val[0]));
668 b[1] = vtrn_u16(vreinterpret_u16_u8(a[0].val[1]),
669 vreinterpret_u16_u8(a[1].val[1]));
670 b[2] = vtrn_u16(vreinterpret_u16_u8(a[2].val[0]),
671 vreinterpret_u16_u8(a[3].val[0]));
672 b[3] = vtrn_u16(vreinterpret_u16_u8(a[2].val[1]),
673 vreinterpret_u16_u8(a[3].val[1]));
/external/XNNPACK/src/u8-clamp/
Dneon-x64.c59 vst1_lane_u16(__builtin_assume_aligned(y, 1), vreinterpret_u16_u8(vout), 0); y += 2; in xnn_u8_clamp_ukernel__neon_x64()
/external/libvpx/libvpx/vp8/common/arm/neon/
Dvp8_loopfilter_neon.c271 const uint16x4x2_t r02_u16 = vtrn_u16(vreinterpret_u16_u8(result.val[0]), in write_4x8()
272 vreinterpret_u16_u8(result.val[2])); in write_4x8()
273 const uint16x4x2_t r13_u16 = vtrn_u16(vreinterpret_u16_u8(result.val[1]), in write_4x8()
274 vreinterpret_u16_u8(result.val[3])); in write_4x8()
Dloopfiltersimpleverticaledge_neon.c30 const uint16x4_t x_0_4 = vreinterpret_u16_u8(r01_u8.val[0]); in write_2x4()
31 const uint16x4_t x_1_5 = vreinterpret_u16_u8(r01_u8.val[1]); in write_2x4()
/external/XNNPACK/src/u8-maxpool/
D9p8x-minmax-neon-c16.c134 vst1_lane_u16(__builtin_assume_aligned(o, 1), vreinterpret_u16_u8(vout_lo), 0); o += 2; in xnn_u8_maxpool_minmax_ukernel_9p8x__neon_c16()
238 vst1_lane_u16(__builtin_assume_aligned(o, 1), vreinterpret_u16_u8(vout_lo), 0); o += 2; in xnn_u8_maxpool_minmax_ukernel_9p8x__neon_c16()
/external/libvpx/libvpx/vpx_dsp/arm/
Dtranspose_neon.h76 vtrn_u16(vreinterpret_u16_u8(*a0), vreinterpret_u16_u8(*a1)); in transpose_u8_4x4()
419 vtrn_u16(vreinterpret_u16_u8(b0.val[0]), vreinterpret_u16_u8(b1.val[0])); in transpose_u8_8x4()
421 vtrn_u16(vreinterpret_u16_u8(b0.val[1]), vreinterpret_u16_u8(b1.val[1])); in transpose_u8_8x4()
/external/XNNPACK/src/qu8-gavgpool/
D7x-minmax-neon-c8.c207 vst1_lane_u16(__builtin_assume_aligned(output, 1), vreinterpret_u16_u8(vout), 0); output += 2; in xnn_qu8_gavgpool_minmax_ukernel_7x__neon_c8()
D7p7x-minmax-neon-c8.c285 vst1_lane_u16(__builtin_assume_aligned(output, 1), vreinterpret_u16_u8(vout), 0); output += 2; in xnn_qu8_gavgpool_minmax_ukernel_7p7x__neon_c8()
/external/XNNPACK/src/qu8-avgpool/
D9x-minmax-neon-c8.c265 … vst1_lane_u16(__builtin_assume_aligned(output, 1), vreinterpret_u16_u8(vout), 0); output += 2; in xnn_qu8_avgpool_minmax_ukernel_9x__neon_c8()
D9p8x-minmax-neon-c8.c410 … vst1_lane_u16(__builtin_assume_aligned(output, 1), vreinterpret_u16_u8(vout), 0); output += 2; in xnn_qu8_avgpool_minmax_ukernel_9p8x__neon_c8()
/external/XNNPACK/src/qu8-vadd/
Dminmax-neon.c239 vst1_lane_u16(__builtin_assume_aligned(y, 1), vreinterpret_u16_u8(vy), 0); y += 2; in xnn_qu8_vadd_minmax_ukernel__neon()
/external/XNNPACK/src/qu8-dwconv/
Dup8x9-minmax-neon.c261 … vst1_lane_u16(__builtin_assume_aligned(output, 1), vreinterpret_u16_u8(vout), 0); output += 2; in xnn_qu8_dwconv_minmax_ukernel_up8x9__neon()
/external/llvm-project/clang/test/CodeGen/
Daarch64-neon-intrinsics.c15579 return vreinterpret_u16_u8(a); in test_vreinterpret_u16_u8()

12