Lines Matching refs:uint64x2_t
53 static inline uint64x2_t shift_swap(const uint64x2_t &x, const uint64x2_t &mask) { in shift_swap()
54 uint64x2_t t = vandq_u64(mask, veorq_u64(x, vshrq_n_u64(x, shift))); in shift_swap()
58 static inline uint64x2_t pack_indices(const uint64x2_t &x) { in pack_indices()
61 static const uint64x2_t kMask1 = { 0x3FC0003FC00000ULL, 0x3FC0003FC00000ULL }; in pack_indices()
62 uint64x2_t ret = shift_swap<10>(x, kMask1); in pack_indices()
65 static const uint64x2_t kMask2 = { (0x3FULL << 52), (0x3FULL << 52) }; in pack_indices()
66 static const uint64x2_t kMask3 = { (0x3FULL << 28), (0x3FULL << 28) }; in pack_indices()
67 const uint64x2_t x1 = vandq_u64(vshlq_n_u64(ret, 52), kMask2); in pack_indices()
68 const uint64x2_t x2 = vandq_u64(vshlq_n_u64(ret, 20), kMask3); in pack_indices()
73 static const uint64x2_t kMask4 = { 0xFC0000ULL, 0xFC0000ULL }; in pack_indices()
79 static const uint64x2_t kMask5 = { 0x3FULL, 0x3FULL }; in pack_indices()
84 static const uint64x2_t kMask6 = { 0xFFF000000ULL, 0xFFF000000ULL }; in pack_indices()
89 static const uint64x2_t kMask5 = { 0xFC0ULL, 0xFC0ULL }; in pack_indices()
94 static const uint64x2_t kMask6 = { (0xFFFULL << 36), (0xFFFULL << 36) }; in pack_indices()
95 static const uint64x2_t kMask7 = { 0xFFFFFFULL, 0xFFFFFFULL }; in pack_indices()
96 static const uint64x2_t kMask8 = { 0xFFFULL, 0xFFFULL }; in pack_indices()
97 const uint64x2_t y1 = vandq_u64(ret, kMask6); in pack_indices()
98 const uint64x2_t y2 = vshlq_n_u64(vandq_u64(ret, kMask7), 12); in pack_indices()
99 const uint64x2_t y3 = vandq_u64(vshrq_n_u64(ret, 24), kMask8); in pack_indices()
106 static const uint64x2_t kHeader = { 0x8490000000000000ULL, 0x8490000000000000ULL }; in pack_indices()
137 const uint64x2_t cvt = vreinterpretq_u64_u8(x); in is_zero()
144 static inline uint64x2_t fix_endianness(uint64x2_t x) { in fix_endianness()
148 static inline uint64x2_t fix_endianness(uint64x2_t x) { in fix_endianness()
176 static const uint64x2_t kTransparent = { 0x0020000000002000ULL, in compress_r11eac_blocks()
193 const uint64x2_t indexRow12 = vreinterpretq_u64_u8( in compress_r11eac_blocks()
195 const uint64x2_t indexRow34 = vreinterpretq_u64_u8( in compress_r11eac_blocks()
200 const uint64x2_t blockIndicesLeft = vreinterpretq_u64_u32(vrev64q_u32(blockIndices.val[0])); in compress_r11eac_blocks()
201 … const uint64x2_t blockIndicesRight = vreinterpretq_u64_u32(vrev64q_u32(blockIndices.val[1])); in compress_r11eac_blocks()
203 const uint64x2_t indicesLeft = fix_endianness(pack_indices(blockIndicesLeft)); in compress_r11eac_blocks()
204 const uint64x2_t indicesRight = fix_endianness(pack_indices(blockIndicesRight)); in compress_r11eac_blocks()
206 const uint64x2_t d1 = vcombine_u64(vget_low_u64(indicesLeft), vget_low_u64(indicesRight)); in compress_r11eac_blocks()
207 const uint64x2_t d2 = vcombine_u64(vget_high_u64(indicesLeft), vget_high_u64(indicesRight)); in compress_r11eac_blocks()