/external/boringssl/linux-arm/crypto/aes/ |
D | aesv8-armx32.S | 37 veor q0,q0,q0 50 vext.8 q9,q0,q3,#12 52 .byte 0x00,0x43,0xf0,0xf3 @ aese q10,q0 56 vext.8 q9,q0,q9,#12 58 vext.8 q9,q0,q9,#12 69 vext.8 q9,q0,q3,#12 71 .byte 0x00,0x43,0xf0,0xf3 @ aese q10,q0 74 vext.8 q9,q0,q9,#12 76 vext.8 q9,q0,q9,#12 84 vext.8 q9,q0,q3,#12 [all …]
|
D | bsaes-armv7.S | 1140 vld1.8 {q0,q1}, [r0]! @ load input 1157 veor q0, q0, q14 @ ^= IV 1166 vst1.8 {q0,q1}, [r1]! @ write output 1182 vld1.8 {q0}, [r0]! @ load input 1211 veor q0, q0, q14 @ ^= IV 1220 vst1.8 {q0,q1}, [r1]! @ write output 1234 veor q0, q0, q14 @ ^= IV 1243 vst1.8 {q0,q1}, [r1]! @ write output 1255 veor q0, q0, q14 @ ^= IV 1261 vst1.8 {q0,q1}, [r1]! @ write output [all …]
|
/external/valgrind/none/tests/arm/ |
D | neon128.c | 355 TESTINSN_imm("vmov.i32 q0", q0, 0x7); in main() 365 TESTINSN_imm("vmov.f32 q0", q0, 0.328125); in main() 366 TESTINSN_imm("vmov.f32 q0", q0, -0.328125); in main() 369 TESTINSN_imm("vmvn.i32 q0", q0, 0x7); in main() 381 TESTINSN_imm("vorr.i32 q0", q0, 0x7); in main() 389 TESTINSN_imm("vbic.i32 q0", q0, 0x7); in main() 397 TESTINSN_un("vmvn q0, q1", q0, q1, i32, 24); in main() 399 TESTINSN_un("vmvn q0, q14", q0, q14, i32, 24); in main() 402 TESTINSN_un("vmov q0, q1", q0, q1, i32, 24); in main() 404 TESTINSN_un("vmov q0, q14", q0, q14, i32, 24); in main() [all …]
|
/external/llvm/test/MC/ARM/ |
D | thumb-neon-crypto.s | 3 aesd.8 q0, q1 4 @ CHECK: aesd.8 q0, q1 @ encoding: [0xb0,0xff,0x42,0x03] 5 aese.8 q0, q1 6 @ CHECK: aese.8 q0, q1 @ encoding: [0xb0,0xff,0x02,0x03] 7 aesimc.8 q0, q1 8 @ CHECK: aesimc.8 q0, q1 @ encoding: [0xb0,0xff,0xc2,0x03] 9 aesmc.8 q0, q1 10 @ CHECK: aesmc.8 q0, q1 @ encoding: [0xb0,0xff,0x82,0x03] 12 sha1h.32 q0, q1 13 @ CHECK: sha1h.32 q0, q1 @ encoding: [0xb9,0xff,0xc2,0x02] [all …]
|
D | fullfp16-neon.s | 5 vadd.f16 q0, q1, q2 7 @ ARM: vadd.f16 q0, q1, q2 @ encoding: [0x44,0x0d,0x12,0xf2] 9 @ THUMB: vadd.f16 q0, q1, q2 @ encoding: [0x12,0xef,0x44,0x0d] 12 vsub.f16 q0, q1, q2 14 @ ARM: vsub.f16 q0, q1, q2 @ encoding: [0x44,0x0d,0x32,0xf2] 16 @ THUMB: vsub.f16 q0, q1, q2 @ encoding: [0x32,0xef,0x44,0x0d] 19 vmul.f16 q0, q1, q2 21 @ ARM: vmul.f16 q0, q1, q2 @ encoding: [0x54,0x0d,0x12,0xf3] 23 @ THUMB: vmul.f16 q0, q1, q2 @ encoding: [0x12,0xff,0x54,0x0d] 33 vmla.f16 q0, q1, q2 [all …]
|
D | neon-crypto.s | 4 aesd.8 q0, q1 5 aese.8 q0, q1 6 aesimc.8 q0, q1 7 aesmc.8 q0, q1 8 @ CHECK: aesd.8 q0, q1 @ encoding: [0x42,0x03,0xb0,0xf3] 9 @ CHECK: aese.8 q0, q1 @ encoding: [0x02,0x03,0xb0,0xf3] 10 @ CHECK: aesimc.8 q0, q1 @ encoding: [0xc2,0x03,0xb0,0xf3] 11 @ CHECK: aesmc.8 q0, q1 @ encoding: [0x82,0x03,0xb0,0xf3] 17 sha1h.32 q0, q1 18 sha1su1.32 q0, q1 [all …]
|
D | directive-arch_extension-crypto.s | 19 vmull.p64 q0, d0, d1 22 aesd.8 q0, q1 24 aese.8 q0, q1 26 aesimc.8 q0, q1 28 aesmc.8 q0, q1 31 sha1h.32 q0, q1 33 sha1su1.32 q0, q1 35 sha256su0.32 q0, q1 38 sha1c.32 q0, q1, q2 40 sha1m.32 q0, q1, q2 [all …]
|
D | fullfp16-neon-neg.s | 7 vadd.f16 q0, q1, q2 12 vsub.f16 q0, q1, q2 17 vmul.f16 q0, q1, q2 27 vmla.f16 q0, q1, q2 37 vmls.f16 q0, q1, q2 47 vfma.f16 q0, q1, q2 52 vfms.f16 q0, q1, q2 107 vacge.f16 q0, q1, q2 112 vacgt.f16 q0, q1, q2 117 vacle.f16 q0, q1, q2 [all …]
|
/external/libvpx/libvpx/vpx_dsp/arm/ |
D | intrapred_neon_asm.asm | 76 vld1.8 {q0}, [r2] 77 vst1.8 {q0}, [r0], r1 78 vst1.8 {q0}, [r0], r1 79 vst1.8 {q0}, [r0], r1 80 vst1.8 {q0}, [r0], r1 81 vst1.8 {q0}, [r0], r1 82 vst1.8 {q0}, [r0], r1 83 vst1.8 {q0}, [r0], r1 84 vst1.8 {q0}, [r0], r1 85 vst1.8 {q0}, [r0], r1 [all …]
|
/external/libavc/common/arm/ |
D | ih264_deblk_luma_a9.s | 62 @* Pointer to the src sample q0 106 vld1.8 {q0}, [r0], r1 @q0 values are loaded into q0 109 vabd.u8 q11, q3, q0 @Q11 = ABS(p0 - q0) 111 vabd.u8 q12, q1, q0 @Q12 = ABS(q1 - q0) 119 vabd.u8 q15, q2, q0 @Q15 = Aq = ABS(q2 - q0) 122 vcge.u8 q9, q11, q10 @Q9 = ( ABS(p0 - q0) >= Alpha ) 123 vcge.u8 q12, q12, q8 @Q12=( ABS(q1 - q0) >= Beta ) 128 vorr q9, q9, q12 @Q9 = ( ABS(p0 - q0) >= Alpha ) | ( ABS(q1 - q0) >= Beta ) 130 vsubl.u8 q12, d0, d6 @Q15,Q12 = (q0 - p0) 131 …vorr q9, q9, q13 @Q9 = ( ABS(p0 - q0) >= Alpha ) | ( ABS(q1 - q0) >= Beta ) | (… [all …]
|
D | ih264_intra_pred_luma_16x16_a9q.s | 113 vld1.8 {q0}, [r0] 115 vst1.8 {q0}, [r1], r3 116 vst1.8 {q0}, [r1], r3 117 vst1.8 {q0}, [r1], r3 118 vst1.8 {q0}, [r1], r3 119 vst1.8 {q0}, [r1], r3 120 vst1.8 {q0}, [r1], r3 121 vst1.8 {q0}, [r1], r3 122 vst1.8 {q0}, [r1], r3 123 vst1.8 {q0}, [r1], r3 [all …]
|
D | ih264_deblk_chroma_a9.s | 69 @* Pointer to the src sample q0 102 vaddl.u8 q5, d7, d1 @Q4,Q5 = q0 + p1 108 vabd.u8 q11, q2, q0 @Q11 = ABS(p0 - q0) 109 vabd.u8 q12, q1, q0 @Q12 = ABS(q1 - q0) 115 vcge.u8 q9, q11, q10 @Q9 = ( ABS(p0 - q0) >= Alpha ) 116 vcge.u8 q12, q12, q8 @Q12= ( ABS(q1 - q0) >= Beta ) 120 vorr q9, q9, q12 @Q9 = ( ABS(p0 - q0) >= Alpha ) | ( ABS(q1 - q0) >= Beta ) 123 …vorr q9, q9, q13 @Q9 = ( ABS(p0 - q0) >= Alpha ) | ( ABS(q1 - q0) >= Beta ) | (… 125 vbit q4, q0, q9 @ 145 @* Pointer to the src sample q0 [all …]
|
D | ih264_intra_pred_chroma_a9q.s | 117 vld1.u8 {q0}, [r0] @BOTH LEFT AND TOP AVAILABLE 123 vmovl.u8 q0, d0 145 vld1.u8 {q0}, [r0] 150 vpaddl.u32 q0, q0 151 vqrshrun.s16 d0, q0, #2 158 vld1.u8 {q0}, [r0] 163 vpaddl.u32 q0, q0 164 vqrshrun.s16 d0, q0, #2 243 vld1.u8 {q0}, [r0] 251 vext.8 q0, q0, q0, #12 [all …]
|
/external/boringssl/linux-arm/crypto/modes/ |
D | ghashv8-armx32.S | 31 .byte 0xa8,0x0e,0xa8,0xf2 @ pmull q0,q12,q12 36 vext.8 q9,q0,q2,#8 @ Karatsuba post-processing 37 veor q10,q0,q2 40 .byte 0x26,0x4e,0xe0,0xf2 @ pmull q10,q0,q11 @ 1st phase 44 veor q0,q1,q10 46 vext.8 q10,q0,q0,#8 @ 2nd phase 47 .byte 0x26,0x0e,0xa0,0xf2 @ pmull q0,q0,q11 49 veor q14,q0,q10 72 .byte 0x86,0x0e,0xa8,0xf2 @ pmull q0,q12,q3 @ H.lo·Xi.lo 77 vext.8 q9,q0,q2,#8 @ Karatsuba post-processing [all …]
|
/external/libvpx/libvpx/vpx_dsp/ |
D | loopfilter.c | 35 uint8_t p2, uint8_t p1, uint8_t p0, uint8_t q0, in filter_mask() argument 41 mask |= (abs(q1 - q0) > limit) * -1; in filter_mask() 44 mask |= (abs(p0 - q0) * 2 + abs(p1 - q1) / 2 > blimit) * -1; in filter_mask() 49 uint8_t p1, uint8_t p0, uint8_t q0, uint8_t q1, in flat_mask4() argument 53 mask |= (abs(q1 - q0) > thresh) * -1; in flat_mask4() 55 mask |= (abs(q2 - q0) > thresh) * -1; in flat_mask4() 57 mask |= (abs(q3 - q0) > thresh) * -1; in flat_mask4() 62 uint8_t p2, uint8_t p1, uint8_t p0, uint8_t q0, in flat_mask5() argument 65 int8_t mask = ~flat_mask4(thresh, p3, p2, p1, p0, q0, q1, q2, q3); in flat_mask5() 67 mask |= (abs(q4 - q0) > thresh) * -1; in flat_mask5() [all …]
|
/external/llvm/test/MC/Disassembler/ARM/ |
D | neon-crypto.txt | 4 # CHECK: aesd.8 q0, q1 6 # CHECK: aese.8 q0, q1 8 # CHECK: aesimc.8 q0, q1 10 # CHECK: aesmc.8 q0, q1 13 # CHECK: sha1h.32 q0, q1 15 # CHECK: sha1su1.32 q0, q1 17 # CHECK: sha256su0.32 q0, q1 20 # CHECK: sha1c.32 q0, q1, q2 22 # CHECK: sha1m.32 q0, q1, q2 24 # CHECK: sha1p.32 q0, q1, q2 [all …]
|
D | thumb-neon-crypto.txt | 4 # CHECK: aesd.8 q0, q1 6 # CHECK: aese.8 q0, q1 8 # CHECK: aesimc.8 q0, q1 10 # CHECK: aesmc.8 q0, q1 13 # CHECK: sha1h.32 q0, q1 15 # CHECK: sha1su1.32 q0, q1 17 # CHECK: sha256su0.32 q0, q1 20 # CHECK: sha1c.32 q0, q1, q2 22 # CHECK: sha1m.32 q0, q1, q2 24 # CHECK: sha1p.32 q0, q1, q2 [all …]
|
D | fullfp16-neon-thumb.txt | 5 # CHECK: vadd.f16 q0, q1, q2 10 # CHECK: vsub.f16 q0, q1, q2 15 # CHECK: vmul.f16 q0, q1, q2 25 # CHECK: vmla.f16 q0, q1, q2 35 # CHECK: vmls.f16 q0, q1, q2 45 # CHECK: vfma.f16 q0, q1, q2 50 # CHECK: vfms.f16 q0, q1, q2 95 # CHECK: vacge.f16 q0, q1, q2 100 # CHECK: vacgt.f16 q0, q1, q2 105 # CHECK: vabd.f16 q0, q1, q2 [all …]
|
D | fullfp16-neon-arm.txt | 5 # CHECK: vadd.f16 q0, q1, q2 10 # CHECK: vsub.f16 q0, q1, q2 15 # CHECK: vmul.f16 q0, q1, q2 25 # CHECK: vmla.f16 q0, q1, q2 35 # CHECK: vmls.f16 q0, q1, q2 45 # CHECK: vfma.f16 q0, q1, q2 50 # CHECK: vfms.f16 q0, q1, q2 95 # CHECK: vacge.f16 q0, q1, q2 100 # CHECK: vacgt.f16 q0, q1, q2 105 # CHECK: vabd.f16 q0, q1, q2 [all …]
|
/external/llvm/test/CodeGen/ARM/ |
D | thumb-big-stack.ll | 145 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 147 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 149 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 151 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 153 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 155 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 157 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 159 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 161 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 163 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… [all …]
|
/external/libvpx/libvpx/vpx_dsp/mips/ |
D | loopfilter_4_msa.c | 19 v16u8 p3, p2, p1, p0, q3, q2, q1, q0, p1_out, p0_out, q0_out, q1_out; in vpx_lpf_horizontal_4_msa() local 22 LD_UB8((src - 4 * pitch), pitch, p3, p2, p1, p0, q0, q1, q2, q3); in vpx_lpf_horizontal_4_msa() 28 LPF_MASK_HEV(p3, p2, p1, p0, q0, q1, q2, q3, limit, b_limit, thresh, hev, in vpx_lpf_horizontal_4_msa() 30 VP9_LPF_FILTER4_4W(p1, p0, q0, q1, mask, hev, p1_out, p0_out, q0_out, q1_out); in vpx_lpf_horizontal_4_msa() 47 v16u8 p3, p2, p1, p0, q3, q2, q1, q0; in vpx_lpf_horizontal_4_dual_msa() local 50 LD_UB8((src - 4 * pitch), pitch, p3, p2, p1, p0, q0, q1, q2, q3); in vpx_lpf_horizontal_4_dual_msa() 64 LPF_MASK_HEV(p3, p2, p1, p0, q0, q1, q2, q3, limit0, b_limit0, thresh0, hev, in vpx_lpf_horizontal_4_dual_msa() 66 VP9_LPF_FILTER4_4W(p1, p0, q0, q1, mask, hev, p1, p0, q0, q1); in vpx_lpf_horizontal_4_dual_msa() 68 ST_UB4(p1, p0, q0, q1, (src - 2 * pitch), pitch); in vpx_lpf_horizontal_4_dual_msa() 76 v16u8 p3, p2, p1, p0, q3, q2, q1, q0; in vpx_lpf_vertical_4_msa() local [all …]
|
D | loopfilter_masks_dspr2.h | 29 uint32_t p2, uint32_t q0, uint32_t q1, in filter_hev_mask_dspr2() argument 92 [p0] "r"(p0), [q1] "r"(q1), [q0] "r"(q0), [q2] "r"(q2), [q3] "r"(q3), in filter_hev_mask_dspr2() 121 : [p0] "r"(p0), [q0] "r"(q0), [p1] "r"(p1), [r3] "r"(r3), [q1] "r"(q1), in filter_hev_mask_dspr2() 130 uint32_t p3, uint32_t p2, uint32_t q0, uint32_t q1, uint32_t q2, in filter_hev_mask_flatmask4_dspr2() argument 232 [p0] "r"(p0), [q1] "r"(q1), [q0] "r"(q0), [q2] "r"(q2), [q3] "r"(q3), in filter_hev_mask_flatmask4_dspr2() 261 : [p0] "r"(p0), [q0] "r"(q0), [p1] "r"(p1), [r3] "r"(r3), [q1] "r"(q1), in filter_hev_mask_flatmask4_dspr2() 270 uint32_t p0, uint32_t q0, uint32_t q1, uint32_t q2, in flatmask5() argument 345 [q0] "r"(q0), [q1] "r"(q1), [q2] "r"(q2), [q3] "r"(q3), [q4] "r"(q4), in flatmask5()
|
/external/apache-commons-math/src/main/java/org/apache/commons/math/geometry/ |
D | Rotation.java | 104 private final double q0; field in Rotation 133 public Rotation(double q0, double q1, double q2, double q3, in Rotation() argument 138 double inv = 1.0 / FastMath.sqrt(q0 * q0 + q1 * q1 + q2 * q2 + q3 * q3); in Rotation() 139 q0 *= inv; in Rotation() 145 this.q0 = q0; in Rotation() 183 q0 = FastMath.cos (halfAngle); in Rotation() 258 q0 = 0.5 * FastMath.sqrt(s + 1.0); in Rotation() 259 double inv = 0.25 / q0; in Rotation() 269 q0 = inv * (ort[1][2] - ort[2][1]); in Rotation() 278 q0 = inv * (ort[2][0] - ort[0][2]); in Rotation() [all …]
|
/external/llvm/test/MC/Hexagon/ |
D | v60-vcmp.s | 14 #CHECK: 1c80e720 { q0 &= vcmp.gt(v7.ub{{ *}},{{ *}}v0.ub) } 15 q0&=vcmp.gt(v7.ub,v0.ub) 32 #CHECK: 1c85e900 { q0 &= vcmp.eq(v9.b{{ *}},{{ *}}v5.b) } 33 q0&=vcmp.eq(v9.b,v5.b) 35 #CHECK: 1c9cfca8 { q0 ^= vcmp.gt(v28.uw{{ *}},{{ *}}v28.uw) } 36 q0^=vcmp.gt(v28.uw,v28.uw) 38 #CHECK: 1c81faa0 { q0 ^= vcmp.gt(v26.ub{{ *}},{{ *}}v1.ub) } 39 q0^=vcmp.gt(v26.ub,v1.ub) 41 #CHECK: 1c96f0a4 { q0 ^= vcmp.gt(v16.uh{{ *}},{{ *}}v22.uh) } 42 q0^=vcmp.gt(v16.uh,v22.uh) [all …]
|
/external/fdlibm/ |
D | k_rem_pio2.c | 170 int jz,jx,jv,jp,jk,carry,n,iq[20],i,j,k,m,q0,ih; local 180 q0 = e0-24*(jv+1); 201 z = ieee_scalbn(z,q0); /* actual value of z */ 206 if(q0>0) { /* need iq[jz-1] to determine n */ 207 i = (iq[jz-1]>>(24-q0)); n += i; 208 iq[jz-1] -= i<<(24-q0); 209 ih = iq[jz-1]>>(23-q0); 211 else if(q0==0) ih = iq[jz-1]>>23; 224 if(q0>0) { /* rare case: chance is 1 in 12 */ 225 switch(q0) { [all …]
|