/external/libaom/libaom/av1/encoder/ |
D | av1_fwd_txfm2d.c | 126 DECLARE_ALIGNED(32, int32_t, txfm_buf[4 * 8]); in av1_fwd_txfm2d_4x8_c() 129 fwd_txfm2d_c(input, output, stride, &cfg, txfm_buf, bd); in av1_fwd_txfm2d_4x8_c() 134 int32_t txfm_buf[8 * 4]; in av1_fwd_txfm2d_8x4_c() local 137 fwd_txfm2d_c(input, output, stride, &cfg, txfm_buf, bd); in av1_fwd_txfm2d_8x4_c() 142 DECLARE_ALIGNED(32, int32_t, txfm_buf[8 * 16]); in av1_fwd_txfm2d_8x16_c() 145 fwd_txfm2d_c(input, output, stride, &cfg, txfm_buf, bd); in av1_fwd_txfm2d_8x16_c() 150 int32_t txfm_buf[16 * 8]; in av1_fwd_txfm2d_16x8_c() local 153 fwd_txfm2d_c(input, output, stride, &cfg, txfm_buf, bd); in av1_fwd_txfm2d_16x8_c() 158 DECLARE_ALIGNED(32, int32_t, txfm_buf[16 * 32]); in av1_fwd_txfm2d_16x32_c() 161 fwd_txfm2d_c(input, output, stride, &cfg, txfm_buf, bd); in av1_fwd_txfm2d_16x32_c() [all …]
|
/external/libaom/libaom/av1/common/ |
D | av1_inv_txfm2d.c | 256 int32_t *txfm_buf, TX_SIZE tx_size, in inv_txfm2d_add_c() argument 284 int32_t *temp_in = txfm_buf; in inv_txfm2d_add_c() 339 int stride, int32_t *txfm_buf, in inv_txfm2d_add_facade() argument 346 inv_txfm2d_add_c(input, output, stride, &cfg, txfm_buf, tx_size, bd); in inv_txfm2d_add_facade() 351 DECLARE_ALIGNED(32, int, txfm_buf[4 * 8 + 8 + 8]); in av1_inv_txfm2d_add_4x8_c() 352 inv_txfm2d_add_facade(input, output, stride, txfm_buf, tx_type, TX_4X8, bd); in av1_inv_txfm2d_add_4x8_c() 357 DECLARE_ALIGNED(32, int, txfm_buf[8 * 4 + 8 + 8]); in av1_inv_txfm2d_add_8x4_c() 358 inv_txfm2d_add_facade(input, output, stride, txfm_buf, tx_type, TX_8X4, bd); in av1_inv_txfm2d_add_8x4_c() 363 DECLARE_ALIGNED(32, int, txfm_buf[8 * 16 + 16 + 16]); in av1_inv_txfm2d_add_8x16_c() 364 inv_txfm2d_add_facade(input, output, stride, txfm_buf, tx_type, TX_8X16, bd); in av1_inv_txfm2d_add_8x16_c() [all …]
|
/external/libaom/libaom/av1/encoder/x86/ |
D | av1_fwd_txfm2d_sse4.c | 80 int32_t *txfm_buf) { in fwd_txfm2d_sse4_1() argument 96 __m128i *buf_128 = (__m128i *)txfm_buf; in fwd_txfm2d_sse4_1() 101 int16_array_with_stride_to_int32_array_without_stride(input, stride, txfm_buf, in fwd_txfm2d_sse4_1() 115 int32_t *txfm_buf) { in fwd_txfm2d_64x64_sse4_1() argument 123 __m128i *buf_128 = (__m128i *)txfm_buf; in fwd_txfm2d_64x64_sse4_1() 150 DECLARE_ALIGNED(16, int32_t, txfm_buf[1024]); in av1_fwd_txfm2d_32x32_sse4_1() 154 fwd_txfm2d_sse4_1(input, output, stride, &cfg, txfm_buf); in av1_fwd_txfm2d_32x32_sse4_1() 159 DECLARE_ALIGNED(16, int32_t, txfm_buf[4096]); in av1_fwd_txfm2d_64x64_sse4_1() 163 fwd_txfm2d_64x64_sse4_1(input, output, stride, &cfg, txfm_buf); in av1_fwd_txfm2d_64x64_sse4_1()
|
/external/libaom/libaom/av1/common/arm/ |
D | av1_inv_txfm_neon.c | 3911 DECLARE_ALIGNED(32, int, txfm_buf[4 * 4 + 8 + 8]); in lowbd_inv_txfm2d_add_4x4_neon() 3912 int32_t *temp_in = txfm_buf; in lowbd_inv_txfm2d_add_4x4_neon() 3973 DECLARE_ALIGNED(32, int, txfm_buf[4 * 8 + 8 + 8]); in lowbd_inv_txfm2d_add_4x8_neon() 3974 int32_t *temp_in = txfm_buf; in lowbd_inv_txfm2d_add_4x8_neon() 4037 DECLARE_ALIGNED(32, int, txfm_buf[8 * 4 + 8 + 8]); in lowbd_inv_txfm2d_add_8x4_neon() 4038 int32_t *temp_in = txfm_buf; in lowbd_inv_txfm2d_add_8x4_neon() 4101 DECLARE_ALIGNED(32, int, txfm_buf[4 * 16 + 16 + 16]); in lowbd_inv_txfm2d_add_4x16_neon() 4102 int32_t *temp_in = txfm_buf; in lowbd_inv_txfm2d_add_4x16_neon() 4163 DECLARE_ALIGNED(32, int, txfm_buf[16 * 4 + 16 + 16]); in lowbd_inv_txfm2d_add_16x4_neon() 4164 int32_t *temp_in = txfm_buf; in lowbd_inv_txfm2d_add_16x4_neon()
|