/external/valgrind/none/tests/arm/ |
D | neon128.c | 357 TESTINSN_imm("vmov.i8 q2", q2, 0x7); in main() 371 TESTINSN_imm("vmvn.i8 q2", q2, 0x7); in main() 382 TESTINSN_imm("vorr.i16 q2", q2, 0x7); in main() 412 TESTINSN_bin("vadd.i32 q0, q1, q2", q0, q1, i32, 24, q2, i32, 120); in main() 413 TESTINSN_bin("vadd.i64 q0, q1, q2", q0, q1, i32, 140, q2, i32, 120); in main() 414 TESTINSN_bin("vadd.i32 q0, q1, q2", q0, q1, i32, 140, q2, i32, 120); in main() 415 TESTINSN_bin("vadd.i16 q0, q1, q2", q0, q1, i32, 140, q2, i32, 120); in main() 416 TESTINSN_bin("vadd.i8 q0, q1, q2", q0, q1, i32, 140, q2, i32, 120); in main() 417 TESTINSN_bin("vadd.i8 q0, q1, q2", q0, q1, i32, (1 << 31) + 1, q2, i32, (1 << 31) + 2); in main() 418 TESTINSN_bin("vadd.i16 q0, q1, q2", q0, q1, i32, (1 << 31) + 1, q2, i32, (1 << 31) + 2); in main() [all …]
|
D | neon128.stdout.exp | 95 vadd.i32 q0, q1, q2 :: Qd 0x00000090 0x00000090 0x00000090 0x00000090 Qm (i32)0x00000018 Qn (i32)… 96 vadd.i64 q0, q1, q2 :: Qd 0x00000104 0x00000104 0x00000104 0x00000104 Qm (i32)0x0000008c Qn (i32)… 97 vadd.i32 q0, q1, q2 :: Qd 0x00000104 0x00000104 0x00000104 0x00000104 Qm (i32)0x0000008c Qn (i32)… 98 vadd.i16 q0, q1, q2 :: Qd 0x00000104 0x00000104 0x00000104 0x00000104 Qm (i32)0x0000008c Qn (i32)… 99 vadd.i8 q0, q1, q2 :: Qd 0x00000004 0x00000004 0x00000004 0x00000004 Qm (i32)0x0000008c Qn (i32)0… 100 vadd.i8 q0, q1, q2 :: Qd 0x00000003 0x00000003 0x00000003 0x00000003 Qm (i32)0x80000001 Qn (i32)0… 101 vadd.i16 q0, q1, q2 :: Qd 0x00000003 0x00000003 0x00000003 0x00000003 Qm (i32)0x80000001 Qn (i32)… 102 vadd.i32 q0, q1, q2 :: Qd 0x00000003 0x00000003 0x00000003 0x00000003 Qm (i32)0x80000001 Qn (i32)… 103 vadd.i64 q0, q1, q2 :: Qd 0x00000004 0x00000003 0x00000004 0x00000003 Qm (i32)0x80000001 Qn (i32)… 107 vsub.i32 q0, q1, q2 :: Qd 0xffffffa0 0xffffffa0 0xffffffa0 0xffffffa0 Qm (i32)0x00000018 Qn (i32)… [all …]
|
/external/llvm/test/MC/ARM/ |
D | vorr-vbic-illegal-cases.s | 5 vorr.i32 q2, #0xffffffff 7 vorr.i32 q2, #0xabababab 8 vorr.i16 q2, #0xabab 9 vorr.i16 q2, #0xabab 14 @ CHECK: vorr.i32 q2, #0xffffffff 18 @ CHECK: vorr.i32 q2, #0xabababab 20 @ CHECK: vorr.i16 q2, #0xabab 22 @ CHECK: vorr.i16 q2, #0xabab 25 vbic.i32 q2, #0xffffffff 27 vbic.i32 q2, #0xabababab [all …]
|
D | vmov-vmvn-byte-replicate.s | 6 @ CHECK: vmov.i8 q2, #0xff @ encoding: [0x5f,0x4e,0x87,0xf3] 8 @ CHECK: vmov.i8 q2, #0xab @ encoding: [0x5b,0x4e,0x82,0xf3] 9 @ CHECK: vmov.i8 q2, #0xab @ encoding: [0x5b,0x4e,0x82,0xf3] 10 @ CHECK: vmov.i8 q2, #0xab @ encoding: [0x5b,0x4e,0x82,0xf3] 13 @ CHECK: vmov.i8 q2, #0x0 @ encoding: [0x50,0x4e,0x80,0xf2] 15 @ CHECK: vmov.i8 q2, #0x54 @ encoding: [0x54,0x4e,0x85,0xf2] 17 @ CHECK: vmov.i8 q2, #0x54 @ encoding: [0x54,0x4e,0x85,0xf2] 20 vmov.i32 q2, #0xffffffff 22 vmov.i32 q2, #0xabababab 23 vmov.i16 q2, #0xabab [all …]
|
D | thumb-neon-crypto.s | 19 sha1c.32 q0, q1, q2 20 @ CHECK: sha1c.32 q0, q1, q2 @ encoding: [0x02,0xef,0x44,0x0c] 21 sha1m.32 q0, q1, q2 22 @ CHECK: sha1m.32 q0, q1, q2 @ encoding: [0x22,0xef,0x44,0x0c] 23 sha1p.32 q0, q1, q2 24 @ CHECK: sha1p.32 q0, q1, q2 @ encoding: [0x12,0xef,0x44,0x0c] 25 sha1su0.32 q0, q1, q2 26 @ CHECK: sha1su0.32 q0, q1, q2 @ encoding: [0x32,0xef,0x44,0x0c] 27 sha256h.32 q0, q1, q2 28 @ CHECK: sha256h.32 q0, q1, q2 @ encoding: [0x02,0xff,0x44,0x0c] [all …]
|
D | vmov-vmvn-illegal-cases.s | 7 @ CHECK: vmov.i32 q2, #0xffffffab 9 @ CHECK: vmov.i16 q2, #0xffab 11 @ CHECK: vmov.i16 q2, #0xffab 16 @ CHECK: vmvn.i32 q2, #0xffffffab 18 @ CHECK: vmvn.i16 q2, #0xffab 20 @ CHECK: vmvn.i16 q2, #0xffab 23 vmov.i32 q2, #0xffffffab 24 vmov.i16 q2, #0xffab 25 vmov.i16 q2, #0xffab 28 vmvn.i32 q2, #0xffffffab [all …]
|
D | neon-crypto.s | 27 sha1c.32 q0, q1, q2 28 sha1m.32 q0, q1, q2 29 sha1p.32 q0, q1, q2 30 sha1su0.32 q0, q1, q2 31 sha256h.32 q0, q1, q2 32 sha256h2.32 q0, q1, q2 33 sha256su1.32 q0, q1, q2 34 @ CHECK: sha1c.32 q0, q1, q2 @ encoding: [0x44,0x0c,0x02,0xf2] 35 @ CHECK: sha1m.32 q0, q1, q2 @ encoding: [0x44,0x0c,0x22,0xf2] 36 @ CHECK: sha1p.32 q0, q1, q2 @ encoding: [0x44,0x0c,0x12,0xf2] [all …]
|
D | neont2-minmax-encoding.s | 21 vmax.s8 q1, q2, q3 29 vmax.s8 q2, q3 32 vmax.u8 q11, q2 35 vmax.f32 q2, q1 51 @ CHECK: vmax.s8 q1, q2, q3 @ encoding: [0x04,0xef,0x46,0x26] 58 @ CHECK: vmax.s8 q2, q2, q3 @ encoding: [0x04,0xef,0x46,0x46] 61 @ CHECK: vmax.u8 q11, q11, q2 @ encoding: [0x46,0xff,0xc4,0x66] 64 @ CHECK: vmax.f32 q2, q2, q1 @ encoding: [0x04,0xef,0x42,0x4f] 83 vmin.s8 q1, q2, q3 91 vmin.s8 q2, q3 [all …]
|
D | neon-minmax-encoding.s | 19 vmax.s8 q1, q2, q3 27 vmax.s8 q2, q3 30 vmax.u8 q11, q2 33 vmax.f32 q2, q1 49 @ CHECK: vmax.s8 q1, q2, q3 @ encoding: [0x46,0x26,0x04,0xf2] 56 @ CHECK: vmax.s8 q2, q2, q3 @ encoding: [0x46,0x46,0x04,0xf2] 59 @ CHECK: vmax.u8 q11, q11, q2 @ encoding: [0xc4,0x66,0x46,0xf3] 62 @ CHECK: vmax.f32 q2, q2, q1 @ encoding: [0x42,0x4f,0x04,0xf2] 81 vmin.s8 q1, q2, q3 89 vmin.s8 q2, q3 [all …]
|
D | neon-shiftaccum-encoding.s | 7 vsra.s8 q7, q2, #8 16 vsra.u16 q2, q7, #6 25 vsra.s8 q2, #8 42 @ CHECK: vsra.s8 q7, q2, #8 @ encoding: [0x54,0xe1,0x88,0xf2] 51 @ CHECK: vsra.u16 q2, q7, #6 @ encoding: [0x5e,0x41,0x9a,0xf3] 59 @ CHECK: vsra.s8 q2, q2, #8 @ encoding: [0x54,0x41,0x88,0xf2] 80 vrsra.s8 q1, q2, #8 81 vrsra.s16 q2, q3, #16 98 vrsra.s8 q2, #8 115 @ CHECK: vrsra.s8 q1, q2, #8 @ encoding: [0x54,0x23,0x88,0xf2] [all …]
|
D | neont2-shiftaccum-encoding.s | 9 vsra.s8 q7, q2, #8 18 vsra.u16 q2, q7, #6 27 vsra.s8 q2, #8 44 @ CHECK: vsra.s8 q7, q2, #8 @ encoding: [0x88,0xef,0x54,0xe1] 53 @ CHECK: vsra.u16 q2, q7, #6 @ encoding: [0x9a,0xff,0x5e,0x41] 61 @ CHECK: vsra.s8 q2, q2, #8 @ encoding: [0x88,0xef,0x54,0x41] 83 vrsra.s8 q1, q2, #8 84 vrsra.s16 q2, q3, #16 101 vrsra.s8 q2, #8 118 @ CHECK: vrsra.s8 q1, q2, #8 @ encoding: [0x88,0xef,0x54,0x23] [all …]
|
/external/eigen/test/ |
D | geo_quaternion.cpp | 71 Quaternionx q1, q2; in quaternion() local 72 q2.setIdentity(); in quaternion() 73 VERIFY_IS_APPROX(Quaternionx(Quaternionx::Identity()).coeffs(), q2.coeffs()); in quaternion() 75 VERIFY_IS_APPROX(q1.coeffs(), (q1*q2).coeffs()); in quaternion() 78 q1 *= q2; in quaternion() 81 q2 = AngleAxisx(a, v1.normalized()); in quaternion() 84 Scalar refangle = abs(AngleAxisx(q1.inverse()*q2).angle()); in quaternion() 88 if((q1.coeffs()-q2.coeffs()).norm() > 10*largeEps) in quaternion() 90 VERIFY_IS_MUCH_SMALLER_THAN(abs(q1.angularDistance(q2) - refangle), Scalar(1)); in quaternion() 95 VERIFY_IS_APPROX(q1 * q2 * v2, in quaternion() [all …]
|
/external/v8/src/base/ |
D | division-by-constant.cc | 35 T q2 = min / ad; // Init. q2 = 2**p/|d|. in SignedDivisionByConstant() local 36 T r2 = min - q2 * ad; // Init. r2 = rem(2**p, |d|). in SignedDivisionByConstant() 46 q2 = 2 * q2; // Update q2 = 2**p/|d|. in SignedDivisionByConstant() 49 q2 = q2 + 1; in SignedDivisionByConstant() 54 T mul = q2 + 1; in SignedDivisionByConstant() 73 T q2 = max / d; // Init. q2 = (2**p - 1)/d. in UnsignedDivisionByConstant() local 74 T r2 = max - q2 * d; // Init. r2 = rem(2**p - 1, d). in UnsignedDivisionByConstant() 86 if (q2 >= max) a = true; in UnsignedDivisionByConstant() 87 q2 = 2 * q2 + 1; in UnsignedDivisionByConstant() 90 if (q2 >= min) a = true; in UnsignedDivisionByConstant() [all …]
|
/external/libvpx/libvpx/vp8/common/arm/neon/ |
D | dequant_idct_neon.c | 24 int16x8_t q1, q2, q3, q4, q5, q6; in vp8_dequant_idct_add_neon() local 55 q2 = vreinterpretq_s16_u16(vmulq_u16(vreinterpretq_u16_s16(q4), in vp8_dequant_idct_add_neon() 58 d12 = vqadd_s16(vget_low_s16(q1), vget_low_s16(q2)); in vp8_dequant_idct_add_neon() 59 d13 = vqsub_s16(vget_low_s16(q1), vget_low_s16(q2)); in vp8_dequant_idct_add_neon() 61 q2 = vcombine_s16(vget_high_s16(q1), vget_high_s16(q2)); in vp8_dequant_idct_add_neon() 63 q3 = vqdmulhq_n_s16(q2, sinpi8sqrt2); in vp8_dequant_idct_add_neon() 64 q4 = vqdmulhq_n_s16(q2, cospi8sqrt2minus1); in vp8_dequant_idct_add_neon() 69 q3 = vqaddq_s16(q3, q2); in vp8_dequant_idct_add_neon() 70 q4 = vqaddq_s16(q4, q2); in vp8_dequant_idct_add_neon() 88 q2 = vcombine_s16(d2tmp2.val[1], d2tmp3.val[1]); in vp8_dequant_idct_add_neon() [all …]
|
/external/freetype/src/base/ |
D | ftbbox.c | 251 FT_Pos q2, in cubic_peak() argument 269 FT_ABS( q2 ) | in cubic_peak() 280 q2 <<= shift; in cubic_peak() 287 q2 >>= -shift; in cubic_peak() 294 while ( q2 > 0 || q3 > 0 ) in cubic_peak() 297 if ( q1 + q2 > q3 + q4 ) /* first half */ in cubic_peak() 300 q3 = q3 + q2; in cubic_peak() 301 q2 = q2 + q1; in cubic_peak() 303 q3 = q3 + q2; in cubic_peak() 306 q2 = q2 / 2; in cubic_peak() [all …]
|
/external/pdfium/third_party/freetype/src/base/ |
D | ftbbox.c | 251 FT_Pos q2, in cubic_peak() argument 268 FT_MSB( FT_ABS( q1 ) | FT_ABS( q2 ) | FT_ABS( q3 ) | FT_ABS( q4 ) ); in cubic_peak() 277 q2 <<= shift; in cubic_peak() 284 q2 >>= -shift; in cubic_peak() 291 while ( q2 > 0 || q3 > 0 ) in cubic_peak() 294 if ( q1 + q2 > q3 + q4 ) /* first half */ in cubic_peak() 297 q3 = q3 + q2; in cubic_peak() 298 q2 = q2 + q1; in cubic_peak() 300 q3 = q3 + q2; in cubic_peak() 303 q2 = q2 / 2; in cubic_peak() [all …]
|
/external/apache-commons-math/src/main/java/org/apache/commons/math/geometry/ |
D | Rotation.java | 110 private final double q2; field in Rotation 133 public Rotation(double q0, double q1, double q2, double q3, in Rotation() argument 138 double inv = 1.0 / FastMath.sqrt(q0 * q0 + q1 * q1 + q2 * q2 + q3 * q3); in Rotation() 141 q2 *= inv; in Rotation() 147 this.q2 = q2; in Rotation() 185 q2 = coeff * axis.getY(); in Rotation() 261 q2 = inv * (ort[2][0] - ort[0][2]); in Rotation() 270 q2 = inv * (ort[0][1] + ort[1][0]); in Rotation() 276 q2 = 0.5 * FastMath.sqrt(s + 1.0); in Rotation() 277 double inv = 0.25 / q2; in Rotation() [all …]
|
/external/libvpx/libvpx/vp9/common/ |
D | vp9_loopfilter_filters.c | 25 uint8_t q2, uint8_t q3) { in filter_mask() argument 31 mask |= (abs(q2 - q1) > limit) * -1; in filter_mask() 32 mask |= (abs(q3 - q2) > limit) * -1; in filter_mask() 41 uint8_t q2, uint8_t q3) { in flat_mask4() argument 46 mask |= (abs(q2 - q0) > thresh) * -1; in flat_mask4() 56 uint8_t q1, uint8_t q2, in flat_mask5() argument 58 int8_t mask = ~flat_mask4(thresh, p3, p2, p1, p0, q0, q1, q2, q3); in flat_mask5() 114 const uint8_t q0 = s[0 * p], q1 = s[1 * p], q2 = s[2 * p], q3 = s[3 * p]; in vp9_lpf_horizontal_4_c() local 116 p3, p2, p1, p0, q0, q1, q2, q3); in vp9_lpf_horizontal_4_c() 139 const uint8_t q0 = s[0], q1 = s[1], q2 = s[2], q3 = s[3]; in vp9_lpf_vertical_4_c() local [all …]
|
/external/jmonkeyengine/engine/src/core/com/jme3/math/ |
D | Quaternion.java | 181 public Quaternion(Quaternion q1, Quaternion q2, float interp) { in Quaternion() argument 182 slerp(q1, q2, interp); in Quaternion() 629 public Quaternion slerp(Quaternion q1, Quaternion q2, float t) { in slerp() argument 631 if (q1.x == q2.x && q1.y == q2.y && q1.z == q2.z && q1.w == q2.w) { in slerp() 636 float result = (q1.x * q2.x) + (q1.y * q2.y) + (q1.z * q2.z) in slerp() 637 + (q1.w * q2.w); in slerp() 641 q2.x = -q2.x; in slerp() 642 q2.y = -q2.y; in slerp() 643 q2.z = -q2.z; in slerp() 644 q2.w = -q2.w; in slerp() [all …]
|
/external/libvpx/libvpx/vp9/common/arm/neon/ |
D | vp9_loopfilter_16_neon.asm | 63 vld1.u8 {q9}, [r2@64] ; q2 89 ; q2 thresh 96 ; q9 q2 111 vabd.u8 q3, q9, q8 ; m5 = abs(q2 - q1) 112 vabd.u8 q4, q10, q9 ; m6 = abs(q3 - q2) 126 vcgt.u8 q13, q13, q2 ; (abs(p1 - p0) > thresh)*-1 127 vcgt.u8 q14, q14, q2 ; (abs(q1 - q0) > thresh)*-1 130 vabd.u8 q2, q5, q8 ; a = abs(p1 - q1) 137 vshr.u8 q2, q2, #1 ; a = a / 2 141 vqadd.u8 q9, q9, q2 ; a = b + a [all …]
|
/external/boringssl/linux-arm/crypto/aes/ |
D | aesv8-armx.S | 38 vld1.32 {q1,q2},[r3]! 104 vsub.i8 q2,q2,q10 @ adjust the mask 227 vld1.8 {q2},[r0] 232 .byte 0x00,0x43,0xb0,0xf3 @ aese q2,q0 233 .byte 0x84,0x43,0xb0,0xf3 @ aesmc q2,q2 236 .byte 0x02,0x43,0xb0,0xf3 @ aese q2,q1 237 .byte 0x84,0x43,0xb0,0xf3 @ aesmc q2,q2 241 .byte 0x00,0x43,0xb0,0xf3 @ aese q2,q0 242 .byte 0x84,0x43,0xb0,0xf3 @ aesmc q2,q2 244 .byte 0x02,0x43,0xb0,0xf3 @ aese q2,q1 [all …]
|
/external/libavc/common/arm/ |
D | ih264_deblk_luma_a9.s | 112 vld1.8 {q2}, [r0], r1 @q2 values are loaded into q2 119 vabd.u8 q15, q2, q0 @Q15 = Aq = ABS(q2 - q0) 152 vaddl.u8 q2, d16, d4 @Q15,Q2 = q2 + (p0+q0+1)>>1 161 vsub.i16 q2, q2, q8 @ 162 vsub.i16 q15, q15, q13 @Q15,Q2 = [q2 + (p0+q0+1)>>1] - (q1<<1) 167 vqshrn.s16 d30, q2, #1 @Q15 = i_macro_q1 244 vabd.u8 q6, q2, q3 @ABS(p0 - q0) 245 vabd.u8 q7, q4, q2 @ABS(q1 - q0) 252 vld1.8 {d14, d15}, [r0], r1 @load q2 to Q7, q0 = q0 + src_strd 255 vabd.u8 q11, q7, q2 @Aq = ABS(q2 - q0) [all …]
|
D | ih264_padding_neon.s | 184 vdup.u8 q2, r10 186 vst1.8 {q2}, [r4], r1 @ 16 bytes store 196 vdup.u8 q2, r10 199 vst1.8 {q2}, [r4], r1 @ 16 bytes store 213 vdup.u8 q2, r10 216 vst1.8 {q2}, [r4]! @ 16 bytes store 218 vst1.8 {q2}, [r4], r6 @ 16 bytes store 230 vdup.u8 q2, r10 232 vst1.8 {q2}, [r4]! @ 16 bytes store 234 vst1.8 {q2}, [r4], r6 @ 16 bytes store [all …]
|
/external/libhevc/common/arm/ |
D | ihevc_deblk_chroma_horz.s | 85 vmovl.u8 q2,d4 106 vadd.i16 q2,q3,q2 108 vsub.i16 q3,q2,q8 123 vdup.16 q2,r1 128 vzip.16 q2,q8 133 vmin.s16 q8,q3,q2 134 vmax.s16 q2,q9,q8 135 vadd.i16 q1,q1,q2 136 vsub.i16 q0,q0,q2
|
/external/libopus/celt/ |
D | quant_bands.c | 373 int q2; in quant_fine_energy() local 377 q2 = (error[i+c*m->nbEBands]+QCONST16(.5f,DB_SHIFT))>>(DB_SHIFT-fine_quant[i]); in quant_fine_energy() 379 q2 = (int)floor((error[i+c*m->nbEBands]+.5f)*frac); in quant_fine_energy() 381 if (q2 > frac-1) in quant_fine_energy() 382 q2 = frac-1; in quant_fine_energy() 383 if (q2<0) in quant_fine_energy() 384 q2 = 0; in quant_fine_energy() 385 ec_enc_bits(enc, q2, fine_quant[i]); in quant_fine_energy() 387 …offset = SUB16(SHR32(SHL32(EXTEND32(q2),DB_SHIFT)+QCONST16(.5f,DB_SHIFT),fine_quant[i]),QCONST16(.… in quant_fine_energy() 389 offset = (q2+.5f)*(1<<(14-fine_quant[i]))*(1.f/16384) - .5f; in quant_fine_energy() [all …]
|