/external/XNNPACK/src/x8-zip/ |
D | xm-neon.c | 44 …const uint16x4x2_t vxyzw_lo = vzip_u16(vreinterpret_u16_u8(vxy.val[0]), vreinterpret_u16_u8(vzw.va… in xnn_x8_zip_xm_ukernel__neon() 45 …const uint16x4x2_t vxyzw_hi = vzip_u16(vreinterpret_u16_u8(vxy.val[1]), vreinterpret_u16_u8(vzw.va… in xnn_x8_zip_xm_ukernel__neon() 87 …const uint16x4x2_t vxyzw_lo = vzip_u16(vreinterpret_u16_u8(vxy.val[0]), vreinterpret_u16_u8(vzw.va… in xnn_x8_zip_xm_ukernel__neon() 88 …const uint16x4x2_t vxyzw_hi = vzip_u16(vreinterpret_u16_u8(vxy.val[1]), vreinterpret_u16_u8(vzw.va… in xnn_x8_zip_xm_ukernel__neon()
|
/external/libaom/libaom/av1/common/arm/ |
D | blend_a64_vmask_neon.c | 117 vtrn_u16(vreinterpret_u16_u8(m1), vreinterpret_u16_u8(m2)); in aom_blend_a64_vmask_neon() 122 vreinterpret_u16_u8(max_minus_m1), vreinterpret_u16_u8(max_minus_m2)); in aom_blend_a64_vmask_neon() 132 vreinterpret_u16_u8(vrshrn_n_u16(res, AOM_BLEND_A64_ROUND_BITS)), 0); in aom_blend_a64_vmask_neon() 135 vreinterpret_u16_u8(vrshrn_n_u16(res, AOM_BLEND_A64_ROUND_BITS)), 1); in aom_blend_a64_vmask_neon()
|
D | convolve_neon.c | 289 vreinterpret_u16_u8(d01), 0); in av1_convolve_x_sr_neon() 291 vreinterpret_u16_u8(d23), 0); in av1_convolve_x_sr_neon() 293 vreinterpret_u16_u8(d01), 2); in av1_convolve_x_sr_neon() 295 vreinterpret_u16_u8(d23), 2); in av1_convolve_x_sr_neon() 404 vst1_lane_u16((uint16_t *)dst, vreinterpret_u16_u8(t0), 0); // 00 01 in av1_convolve_x_sr_neon() 406 vst1_lane_u16((uint16_t *)dst, vreinterpret_u16_u8(t1), 0); // 10 11 in av1_convolve_x_sr_neon() 408 vst1_lane_u16((uint16_t *)dst, vreinterpret_u16_u8(t2), 0); // 20 21 in av1_convolve_x_sr_neon() 410 vst1_lane_u16((uint16_t *)dst, vreinterpret_u16_u8(t3), 0); // 30 31 in av1_convolve_x_sr_neon() 412 vst1_lane_u16((uint16_t *)dst, vreinterpret_u16_u8(t0), 2); // 40 41 in av1_convolve_x_sr_neon() 414 vst1_lane_u16((uint16_t *)dst, vreinterpret_u16_u8(t1), 2); // 50 51 in av1_convolve_x_sr_neon() [all …]
|
D | transpose_neon.h | 93 vtrn_u16(vreinterpret_u16_u8(b0.val[0]), vreinterpret_u16_u8(b1.val[0])); in transpose_u8_8x4() 95 vtrn_u16(vreinterpret_u16_u8(b0.val[1]), vreinterpret_u16_u8(b1.val[1])); in transpose_u8_8x4() 112 vtrn_u16(vreinterpret_u16_u8(*a0), vreinterpret_u16_u8(*a1)); in transpose_u8_4x4()
|
D | blend_a64_hmask_neon.c | 125 vreinterpret_u16_u8(vrshrn_n_u16(res, AOM_BLEND_A64_ROUND_BITS)), 0); in aom_blend_a64_hmask_neon() 128 vreinterpret_u16_u8(vrshrn_n_u16(res, AOM_BLEND_A64_ROUND_BITS)), 1); in aom_blend_a64_hmask_neon()
|
/external/exoplayer/tree/extensions/vp9/src/main/jni/ |
D | vpx_jni.cc | 127 uint16x4_t added1 = vqadd_u16(values, vreinterpret_u16_u8(randvec)); in convert_16_to_8_neon() 137 uint16x4_t added2 = vqadd_u16(values, vreinterpret_u16_u8(randvec)); in convert_16_to_8_neon() 152 added1 = vqadd_u16(values, vreinterpret_u16_u8(randvec)); in convert_16_to_8_neon() 156 added2 = vqadd_u16(values, vreinterpret_u16_u8(randvec)); in convert_16_to_8_neon() 202 vreinterpret_u16_u8(randvec)); in convert_16_to_8_neon() 206 vreinterpret_u16_u8(randvec)); in convert_16_to_8_neon() 210 vreinterpret_u16_u8(randvec)); in convert_16_to_8_neon() 214 vreinterpret_u16_u8(randvec)); in convert_16_to_8_neon()
|
/external/libgav1/libgav1/src/dsp/arm/ |
D | common_neon.h | 210 vld1_lane_u16(&temp, vreinterpret_u16_u8(val), lane)); in Load2() 262 ValueToMem<uint16_t>(buf, vget_lane_u16(vreinterpret_u16_u8(val), lane)); in Store2() 506 vtrn_u16(vreinterpret_u16_u8(*a), vreinterpret_u16_u8(*b)); in Transpose4x4() 532 vtrn_u16(vreinterpret_u16_u8(b0.val[0]), vreinterpret_u16_u8(b1.val[0])); in Transpose8x4() 534 vtrn_u16(vreinterpret_u16_u8(b0.val[1]), vreinterpret_u16_u8(b1.val[1])); in Transpose8x4()
|
D | intrapred_directional_neon.cc | 856 const uint16x4x2_t c0 = vtrn_u16(vreinterpret_u16_u8(b0.val[0]), in DirectionalIntraPredictorZone3_NEON() 857 vreinterpret_u16_u8(b1.val[0])); in DirectionalIntraPredictorZone3_NEON() 858 const uint16x4x2_t c1 = vtrn_u16(vreinterpret_u16_u8(b0.val[1]), in DirectionalIntraPredictorZone3_NEON() 859 vreinterpret_u16_u8(b1.val[1])); in DirectionalIntraPredictorZone3_NEON()
|
/external/libjpeg-turbo/simd/arm/ |
D | jidctint-neon.c | 684 vreinterpret_u16_u8(cols_01_23.val[0]), in jsimd_idct_islow_pass2_regular() 685 vreinterpret_u16_u8(cols_01_23.val[1]), in jsimd_idct_islow_pass2_regular() 686 vreinterpret_u16_u8(cols_45_67.val[0]), in jsimd_idct_islow_pass2_regular() 687 vreinterpret_u16_u8(cols_45_67.val[1]) in jsimd_idct_islow_pass2_regular() 787 vreinterpret_u16_u8(cols_01_23.val[0]), in jsimd_idct_islow_pass2_sparse() 788 vreinterpret_u16_u8(cols_01_23.val[1]), in jsimd_idct_islow_pass2_sparse() 789 vreinterpret_u16_u8(cols_45_67.val[0]), in jsimd_idct_islow_pass2_sparse() 790 vreinterpret_u16_u8(cols_45_67.val[1]) in jsimd_idct_islow_pass2_sparse()
|
D | jidctred-neon.c | 473 vreinterpret_u16_u8(output_0123.val[0]), in jsimd_idct_4x4_neon() 474 vreinterpret_u16_u8(output_0123.val[1]) in jsimd_idct_4x4_neon()
|
/external/libvpx/libvpx/vp9/encoder/arm/neon/ |
D | vp9_frame_scale_neon.c | 340 vreinterpret_u16_u8(dd.val[0]), 0); in scale_plane_4_to_1_general() 342 vreinterpret_u16_u8(dd.val[1]), 0); in scale_plane_4_to_1_general() 344 vreinterpret_u16_u8(dd.val[0]), 1); in scale_plane_4_to_1_general() 346 vreinterpret_u16_u8(dd.val[1]), 1); in scale_plane_4_to_1_general() 348 vreinterpret_u16_u8(dd.val[0]), 2); in scale_plane_4_to_1_general() 350 vreinterpret_u16_u8(dd.val[1]), 2); in scale_plane_4_to_1_general() 352 vreinterpret_u16_u8(dd.val[0]), 3); in scale_plane_4_to_1_general() 354 vreinterpret_u16_u8(dd.val[1]), 3); in scale_plane_4_to_1_general()
|
/external/exoplayer/tree/extensions/av1/src/main/jni/ |
D | gav1_jni.cc | 449 uint16x4_t added_1 = vqadd_u16(values, vreinterpret_u16_u8(randvec)); in Convert10BitFrameTo8BitDataBufferNeon() 460 uint16x4_t added_2 = vqadd_u16(values, vreinterpret_u16_u8(randvec)); in Convert10BitFrameTo8BitDataBufferNeon() 477 added_1 = vqadd_u16(values, vreinterpret_u16_u8(randvec)); in Convert10BitFrameTo8BitDataBufferNeon() 483 added_2 = vqadd_u16(values, vreinterpret_u16_u8(randvec)); in Convert10BitFrameTo8BitDataBufferNeon()
|
/external/gemmlowp/internal/ |
D | output_neon.h | 666 b[0] = vtrn_u16(vreinterpret_u16_u8(a[0].val[0]), 667 vreinterpret_u16_u8(a[1].val[0])); 668 b[1] = vtrn_u16(vreinterpret_u16_u8(a[0].val[1]), 669 vreinterpret_u16_u8(a[1].val[1])); 670 b[2] = vtrn_u16(vreinterpret_u16_u8(a[2].val[0]), 671 vreinterpret_u16_u8(a[3].val[0])); 672 b[3] = vtrn_u16(vreinterpret_u16_u8(a[2].val[1]), 673 vreinterpret_u16_u8(a[3].val[1]));
|
/external/XNNPACK/src/u8-clamp/ |
D | neon-x64.c | 59 vst1_lane_u16(__builtin_assume_aligned(y, 1), vreinterpret_u16_u8(vout), 0); y += 2; in xnn_u8_clamp_ukernel__neon_x64()
|
/external/libvpx/libvpx/vp8/common/arm/neon/ |
D | vp8_loopfilter_neon.c | 271 const uint16x4x2_t r02_u16 = vtrn_u16(vreinterpret_u16_u8(result.val[0]), in write_4x8() 272 vreinterpret_u16_u8(result.val[2])); in write_4x8() 273 const uint16x4x2_t r13_u16 = vtrn_u16(vreinterpret_u16_u8(result.val[1]), in write_4x8() 274 vreinterpret_u16_u8(result.val[3])); in write_4x8()
|
D | loopfiltersimpleverticaledge_neon.c | 30 const uint16x4_t x_0_4 = vreinterpret_u16_u8(r01_u8.val[0]); in write_2x4() 31 const uint16x4_t x_1_5 = vreinterpret_u16_u8(r01_u8.val[1]); in write_2x4()
|
/external/XNNPACK/src/u8-maxpool/ |
D | 9p8x-minmax-neon-c16.c | 134 vst1_lane_u16(__builtin_assume_aligned(o, 1), vreinterpret_u16_u8(vout_lo), 0); o += 2; in xnn_u8_maxpool_minmax_ukernel_9p8x__neon_c16() 238 vst1_lane_u16(__builtin_assume_aligned(o, 1), vreinterpret_u16_u8(vout_lo), 0); o += 2; in xnn_u8_maxpool_minmax_ukernel_9p8x__neon_c16()
|
/external/libvpx/libvpx/vpx_dsp/arm/ |
D | transpose_neon.h | 76 vtrn_u16(vreinterpret_u16_u8(*a0), vreinterpret_u16_u8(*a1)); in transpose_u8_4x4() 419 vtrn_u16(vreinterpret_u16_u8(b0.val[0]), vreinterpret_u16_u8(b1.val[0])); in transpose_u8_8x4() 421 vtrn_u16(vreinterpret_u16_u8(b0.val[1]), vreinterpret_u16_u8(b1.val[1])); in transpose_u8_8x4()
|
/external/XNNPACK/src/qu8-gavgpool/ |
D | 7x-minmax-neon-c8.c | 207 vst1_lane_u16(__builtin_assume_aligned(output, 1), vreinterpret_u16_u8(vout), 0); output += 2; in xnn_qu8_gavgpool_minmax_ukernel_7x__neon_c8()
|
D | 7p7x-minmax-neon-c8.c | 285 vst1_lane_u16(__builtin_assume_aligned(output, 1), vreinterpret_u16_u8(vout), 0); output += 2; in xnn_qu8_gavgpool_minmax_ukernel_7p7x__neon_c8()
|
/external/XNNPACK/src/qu8-avgpool/ |
D | 9x-minmax-neon-c8.c | 265 … vst1_lane_u16(__builtin_assume_aligned(output, 1), vreinterpret_u16_u8(vout), 0); output += 2; in xnn_qu8_avgpool_minmax_ukernel_9x__neon_c8()
|
D | 9p8x-minmax-neon-c8.c | 410 … vst1_lane_u16(__builtin_assume_aligned(output, 1), vreinterpret_u16_u8(vout), 0); output += 2; in xnn_qu8_avgpool_minmax_ukernel_9p8x__neon_c8()
|
/external/XNNPACK/src/qu8-vadd/ |
D | minmax-neon.c | 239 vst1_lane_u16(__builtin_assume_aligned(y, 1), vreinterpret_u16_u8(vy), 0); y += 2; in xnn_qu8_vadd_minmax_ukernel__neon()
|
/external/XNNPACK/src/qu8-dwconv/ |
D | up8x9-minmax-neon.c | 261 … vst1_lane_u16(__builtin_assume_aligned(output, 1), vreinterpret_u16_u8(vout), 0); output += 2; in xnn_qu8_dwconv_minmax_ukernel_up8x9__neon()
|
/external/llvm-project/clang/test/CodeGen/ |
D | aarch64-neon-intrinsics.c | 15579 return vreinterpret_u16_u8(a); in test_vreinterpret_u16_u8()
|