/external/llvm/test/MC/ARM/ |
D | neon-bitwise-encoding.s | 110 veor q4, q7, q3 111 veor.8 q4, q7, q3 112 veor.16 q4, q7, q3 113 veor.32 q4, q7, q3 114 veor.64 q4, q7, q3 116 veor.i8 q4, q7, q3 117 veor.i16 q4, q7, q3 118 veor.i32 q4, q7, q3 119 veor.i64 q4, q7, q3 121 veor.s8 q4, q7, q3 [all …]
|
/external/llvm/test/CodeGen/ARM/ |
D | thumb-big-stack.ll | 145 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 147 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 149 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 151 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 153 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 155 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 157 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 159 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 161 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 163 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… [all …]
|
/external/boringssl/linux-arm/crypto/aes/ |
D | aesv8-armx32.S | 38 vld1.8 {q3},[r0]! 48 vtbl.8 d20,{q3},d4 49 vtbl.8 d21,{q3},d5 50 vext.8 q9,q0,q3,#12 51 vst1.32 {q3},[r2]! 55 veor q3,q3,q9 57 veor q3,q3,q9 60 veor q3,q3,q9 62 veor q3,q3,q10 67 vtbl.8 d20,{q3},d4 [all …]
|
/external/pdfium/third_party/freetype/src/base/ |
D | ftbbox.c | 252 FT_Pos q3, in cubic_peak() argument 270 FT_ABS( q3 ) | in cubic_peak() 281 q3 <<= shift; in cubic_peak() 288 q3 >>= -shift; in cubic_peak() 294 while ( q2 > 0 || q3 > 0 ) in cubic_peak() 297 if ( q1 + q2 > q3 + q4 ) /* first half */ in cubic_peak() 299 q4 = q4 + q3; in cubic_peak() 300 q3 = q3 + q2; in cubic_peak() 302 q4 = q4 + q3; in cubic_peak() 303 q3 = q3 + q2; in cubic_peak() [all …]
|
/external/freetype/src/base/ |
D | ftbbox.c | 252 FT_Pos q3, in cubic_peak() argument 270 FT_ABS( q3 ) | in cubic_peak() 281 q3 <<= shift; in cubic_peak() 288 q3 >>= -shift; in cubic_peak() 294 while ( q2 > 0 || q3 > 0 ) in cubic_peak() 297 if ( q1 + q2 > q3 + q4 ) /* first half */ in cubic_peak() 299 q4 = q4 + q3; in cubic_peak() 300 q3 = q3 + q2; in cubic_peak() 302 q4 = q4 + q3; in cubic_peak() 303 q3 = q3 + q2; in cubic_peak() [all …]
|
/external/apache-commons-math/src/main/java/org/apache/commons/math/geometry/ |
D | Rotation.java | 113 private final double q3; field in Rotation 133 public Rotation(double q0, double q1, double q2, double q3, in Rotation() argument 138 double inv = 1.0 / FastMath.sqrt(q0 * q0 + q1 * q1 + q2 * q2 + q3 * q3); in Rotation() 142 q3 *= inv; in Rotation() 148 this.q3 = q3; in Rotation() 186 q3 = coeff * axis.getZ(); in Rotation() 262 q3 = inv * (ort[0][1] - ort[1][0]); in Rotation() 271 q3 = inv * (ort[0][2] + ort[2][0]); in Rotation() 280 q3 = inv * (ort[2][1] + ort[1][2]); in Rotation() 284 q3 = 0.5 * FastMath.sqrt(s + 1.0); in Rotation() [all …]
|
/external/boringssl/linux-arm/crypto/modes/ |
D | ghashv8-armx32.S | 15 vext.8 q3,q9,q9,#8 19 vshr.u64 q10,q3,#63 22 vshl.i64 q3,q3,#1 25 vorr q3,q3,q10 @ H<<<=1 26 veor q12,q3,q8 @ twisted H 70 vext.8 q3,q9,q9,#8 72 .byte 0x86,0x0e,0xa8,0xf2 @ pmull q0,q12,q3 @ H.lo·Xi.lo 73 veor q9,q9,q3 @ Karatsuba pre-processing 74 .byte 0x87,0x4e,0xa9,0xf2 @ pmull2 q2,q12,q3 @ H.hi·Xi.hi 133 vext.8 q3,q8,q8,#8 @ rotate I[0] [all …]
|
/external/libvpx/libvpx/vpx_dsp/ |
D | loopfilter.c | 36 uint8_t q1, uint8_t q2, uint8_t q3) { in filter_mask() argument 43 mask |= (abs(q3 - q2) > limit) * -1; in filter_mask() 50 uint8_t q2, uint8_t q3) { in flat_mask4() argument 57 mask |= (abs(q3 - q0) > thresh) * -1; in flat_mask4() 63 uint8_t q1, uint8_t q2, uint8_t q3, in flat_mask5() argument 65 int8_t mask = ~flat_mask4(thresh, p3, p2, p1, p0, q0, q1, q2, q3); in flat_mask5() 121 const uint8_t q0 = s[0 * p], q1 = s[1 * p], q2 = s[2 * p], q3 = s[3 * p]; in vpx_lpf_horizontal_4_c() local 123 filter_mask(*limit, *blimit, p3, p2, p1, p0, q0, q1, q2, q3); in vpx_lpf_horizontal_4_c() 145 const uint8_t q0 = s[0], q1 = s[1], q2 = s[2], q3 = s[3]; in vpx_lpf_vertical_4_c() local 147 filter_mask(*limit, *blimit, p3, p2, p1, p0, q0, q1, q2, q3); in vpx_lpf_vertical_4_c() [all …]
|
/external/libvpx/libvpx/vp8/common/arm/neon/ |
D | idct_dequant_full_2x_neon.c | 21 int16x8_t q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10, q11; in idct_dequant_full_2x_neon() local 38 q3 = vld1q_s16(q); in idct_dequant_full_2x_neon() 67 q3 = vmulq_s16(q3, q1); in idct_dequant_full_2x_neon() 79 dLow0 = vget_low_s16(q3); in idct_dequant_full_2x_neon() 80 dHigh0 = vget_high_s16(q3); in idct_dequant_full_2x_neon() 83 q3 = vcombine_s16(dLow0, dLow1); in idct_dequant_full_2x_neon() 91 q10 = vqaddq_s16(q2, q3); in idct_dequant_full_2x_neon() 92 q11 = vqsubq_s16(q2, q3); in idct_dequant_full_2x_neon() 101 q3 = vqaddq_s16(q7, q4); in idct_dequant_full_2x_neon() 103 q4 = vqaddq_s16(q10, q3); in idct_dequant_full_2x_neon() [all …]
|
D | dequant_idct_neon.c | 27 int16x8_t q1, q2, q3, q4, q5, q6; in vp8_dequant_idct_add_neon() local 35 q3 = vld1q_s16(input); in vp8_dequant_idct_add_neon() 57 vmulq_u16(vreinterpretq_u16_s16(q3), vreinterpretq_u16_s16(q5))); in vp8_dequant_idct_add_neon() 66 q3 = vqdmulhq_n_s16(q2, sinpi8sqrt2); in vp8_dequant_idct_add_neon() 73 d10 = vqsub_s16(vget_low_s16(q3), vget_high_s16(q4)); in vp8_dequant_idct_add_neon() 74 d11 = vqadd_s16(vget_high_s16(q3), vget_low_s16(q4)); in vp8_dequant_idct_add_neon() 91 q3 = vqdmulhq_n_s16(q2, sinpi8sqrt2); in vp8_dequant_idct_add_neon() 101 d10 = vqsub_s16(vget_low_s16(q3), vget_high_s16(q4)); in vp8_dequant_idct_add_neon() 102 d11 = vqadd_s16(vget_high_s16(q3), vget_low_s16(q4)); in vp8_dequant_idct_add_neon()
|
/external/libavc/common/arm/ |
D | ih264_padding_neon.s | 185 vdup.u8 q3, r11 188 vst1.8 {q3}, [r4], r1 @ 16 bytes store 197 vdup.u8 q3, r11 200 vst1.8 {q3}, [r4], r1 @ 16 bytes store 217 vdup.u8 q3, r11 220 vst1.8 {q3}, [r4]! @ 16 bytes store 223 vst1.8 {q3}, [r4], r6 @ 16 bytes store 233 vdup.u8 q3, r11 236 vst1.8 {q3}, [r4]! @ 16 bytes store 237 vst1.8 {q3}, [r4], r6 @ 16 bytes store [all …]
|
D | ih264_inter_pred_luma_horz_hpel_vert_qpel_a9q.s | 153 vaddl.u8 q3, d0, d5 159 vmla.u16 q3, q4, q11 163 vmls.u16 q3, q4, q12 170 vst1.32 {q3}, [r9], r6 @ store temp buffer 0 249 vaddl.s16 q3, d7, d17 255 vmlal.s16 q3, d31, d22 256 vmlsl.s16 q3, d29, d24 262 vqrshrun.s32 d19, q3, #10 279 vaddl.s16 q3, d9, d21 291 vmlal.s16 q3, d31, d22 [all …]
|
D | ih264_weighted_pred_a9q.s | 142 vmovl.u8 q3, d6 @converting rows 3,4 to 16-bit 145 vmul.s16 q3, q3, d2[0] @weight mult. for rows 3,4 149 vrshl.s16 q3, q3, q0 @rounds off the weighted samples from rows 3,4 152 vaddw.s8 q3, q3, d3 @adding offset for rows 3,4 155 vqmovun.s16 d6, q3 @saturating rows 3,4 to unsigned 8-bit 173 vmovl.u8 q3, d6 @converting row 2 to 16-bit 178 vmul.s16 q3, q3, d2[0] @weight mult. for row 2 183 vrshl.s16 q3, q3, q0 @rounds off the weighted samples from row 2 187 vaddw.s8 q3, q3, d3 @adding offset for row 2 192 vqmovun.s16 d6, q3 @saturating row 2 to unsigned 8-bit [all …]
|
/external/libjpeg-turbo/simd/ |
D | jsimd_arm_neon.S | 111 JLONG q1, q2, q3, q4, q5, q6, q7; \ 133 q3 = ((JLONG) row0 - (JLONG) row4) << 13; \ 137 q1 = q3 + q2; \ 155 q3 = q3 - q2; \ 171 tmp12 = q3; \ 255 vmul.s16 q11, q11, q3 263 vmul.s16 q15, q15, q3 274 vsubl.s16 q3, ROW0L, ROW4L 283 vshl.s32 q3, q3, #13 287 vadd.s32 q1, q3, q2 [all …]
|
/external/libvpx/libvpx/vpx_dsp/mips/ |
D | loopfilter_4_msa.c | 19 v16u8 p3, p2, p1, p0, q3, q2, q1, q0, p1_out, p0_out, q0_out, q1_out; in vpx_lpf_horizontal_4_msa() local 22 LD_UB8((src - 4 * pitch), pitch, p3, p2, p1, p0, q0, q1, q2, q3); in vpx_lpf_horizontal_4_msa() 28 LPF_MASK_HEV(p3, p2, p1, p0, q0, q1, q2, q3, limit, b_limit, thresh, hev, in vpx_lpf_horizontal_4_msa() 47 v16u8 p3, p2, p1, p0, q3, q2, q1, q0; in vpx_lpf_horizontal_4_dual_msa() local 50 LD_UB8((src - 4 * pitch), pitch, p3, p2, p1, p0, q0, q1, q2, q3); in vpx_lpf_horizontal_4_dual_msa() 64 LPF_MASK_HEV(p3, p2, p1, p0, q0, q1, q2, q3, limit0, b_limit0, thresh0, hev, in vpx_lpf_horizontal_4_dual_msa() 76 v16u8 p3, p2, p1, p0, q3, q2, q1, q0; in vpx_lpf_vertical_4_msa() local 79 LD_UB8((src - 4), pitch, p3, p2, p1, p0, q0, q1, q2, q3); in vpx_lpf_vertical_4_msa() 85 TRANSPOSE8x8_UB_UB(p3, p2, p1, p0, q0, q1, q2, q3, p3, p2, p1, p0, q0, q1, q2, in vpx_lpf_vertical_4_msa() 86 q3); in vpx_lpf_vertical_4_msa() [all …]
|
D | loopfilter_masks_dspr2.h | 30 uint32_t q2, uint32_t q3, in filter_hev_mask_dspr2() argument 92 [p0] "r"(p0), [q1] "r"(q1), [q0] "r"(q0), [q2] "r"(q2), [q3] "r"(q3), in filter_hev_mask_dspr2() 131 uint32_t q3, uint32_t *hev, uint32_t *mask, uint32_t *flat) { in filter_hev_mask_flatmask4_dspr2() argument 232 [p0] "r"(p0), [q1] "r"(q1), [q0] "r"(q0), [q2] "r"(q2), [q3] "r"(q3), in filter_hev_mask_flatmask4_dspr2() 271 uint32_t q3, uint32_t q4, uint32_t *flat2) { in flatmask5() argument 345 [q0] "r"(q0), [q1] "r"(q1), [q2] "r"(q2), [q3] "r"(q3), [q4] "r"(q4), in flatmask5()
|
D | loopfilter_8_msa.c | 19 v16u8 p3, p2, p1, p0, q3, q2, q1, q0; in vpx_lpf_horizontal_8_msa() local 26 LD_UB8((src - 4 * pitch), pitch, p3, p2, p1, p0, q0, q1, q2, q3); in vpx_lpf_horizontal_8_msa() 32 LPF_MASK_HEV(p3, p2, p1, p0, q0, q1, q2, q3, limit, b_limit, thresh, hev, in vpx_lpf_horizontal_8_msa() 34 VP9_FLAT4(p3, p2, p0, q0, q2, q3, flat); in vpx_lpf_horizontal_8_msa() 47 q2, zero, q3, p3_r, p2_r, p1_r, p0_r, q0_r, q1_r, q2_r, q3_r); in vpx_lpf_horizontal_8_msa() 85 v16u8 p3, p2, p1, p0, q3, q2, q1, q0; in vpx_lpf_horizontal_8_dual_msa() local 95 LD_UB8(src - (4 * pitch), pitch, p3, p2, p1, p0, q0, q1, q2, q3); in vpx_lpf_horizontal_8_dual_msa() 110 LPF_MASK_HEV(p3, p2, p1, p0, q0, q1, q2, q3, limit, b_limit, thresh, hev, in vpx_lpf_horizontal_8_dual_msa() 112 VP9_FLAT4(p3, p2, p0, q0, q2, q3, flat); in vpx_lpf_horizontal_8_dual_msa() 119 q2, zero, q3, p3_r, p2_r, p1_r, p0_r, q0_r, q1_r, q2_r, q3_r); in vpx_lpf_horizontal_8_dual_msa() [all …]
|
/external/libhevc/common/arm/ |
D | ihevc_intra_pred_luma_horz.s | 124 vdup.8 q3,d1[5] @duplicate the iii value. 133 vst1.8 {q3},[r2],r3 134 vst1.8 {q3},[r9],r3 140 vdup.8 q3,d1[1] 149 vst1.8 {q3},[r2],r3 150 vst1.8 {q3},[r9],r3 156 vdup.8 q3,d0[5] 165 vst1.8 {q3},[r2],r3 166 vst1.8 {q3},[r9],r3 172 vdup.8 q3,d0[1] [all …]
|
D | ihevc_intra_pred_chroma_horz.s | 126 vdup.16 q3,d1[1] @duplicate the iii value. 135 vst1.16 {q3},[r2],r3 136 vst1.16 {q3},[r9],r3 142 vdup.16 q3,d0[1] 151 vst1.16 {q3},[r2],r3 152 vst1.16 {q3},[r9],r3 158 vdup.16 q3,d11[1] 167 vst1.16 {q3},[r2],r3 168 vst1.16 {q3},[r9],r3 174 vdup.16 q3,d10[1] [all …]
|
D | ihevc_deblk_chroma_horz.s | 92 vsub.i16 q3,q0,q1 96 vshl.i16 q3,q3,#2 106 vadd.i16 q2,q3,q2 108 vsub.i16 q3,q2,q8 125 vrshr.s16 q3,q3,#3 133 vmin.s16 q8,q3,q2
|
/external/libvpx/libvpx/vpx_dsp/arm/ |
D | highbd_loopfilter_neon.c | 47 const uint16x8_t q3, uint16x8_t *hev, uint16x8_t *mask) { in filter_hev_mask4() argument 56 *mask = vmaxq_u16(*mask, vabdq_u16(q3, q2)); in filter_hev_mask4() 73 const uint16x8_t q2, const uint16x8_t q3, uint16x8_t *flat, in filter_flat_hev_mask() argument 77 q0, q1, q2, q3, hev, &mask); in filter_flat_hev_mask() 81 *flat = vmaxq_u16(*flat, vabdq_u16(q3, q0)); in filter_flat_hev_mask() 93 const uint16x8_t q3, const uint16x8_t q4, in flat_mask5() argument 102 flat2 = vmaxq_u16(flat2, vabdq_u16(q3, q0)); in flat_mask5() 151 const uint16x8_t q2, const uint16x8_t q3, in calc_7_tap_filter() argument 166 *oq0 = calc_7_tap_filter_kernel(p3, p0, q0, q3, &sum); in calc_7_tap_filter() 167 *oq1 = calc_7_tap_filter_kernel(p2, q0, q1, q3, &sum); in calc_7_tap_filter() [all …]
|
D | vpx_convolve_avg_neon_asm.asm | 37 vld1.8 {q2-q3}, [r0], lr 44 vrhadd.u8 q3, q3, q11 46 vst1.8 {q2-q3}, [r2@128], r4 53 vld1.8 {q2-q3}, [r0], r1 63 vrhadd.u8 q3, q3, q11 65 vst1.8 {q2-q3}, [r2@128], r3 74 vld1.8 {q3}, [r6@128], r3 80 vrhadd.u8 q1, q1, q3
|
D | loopfilter_neon.c | 74 const uint8x##w##_t q3, uint8x##w##_t *hev, uint8x##w##_t *mask) { \ 83 *mask = vmax##r##u8(*mask, vabd##r##u8(q3, q2)); \ 106 const uint8x##w##_t q3, uint8x##w##_t *flat, uint32_t *flat_status, \ 111 q2, q3, hev, &mask); \ 115 *flat = vmax##r##u8(*flat, vabd##r##u8(q3, q0)); \ 131 const uint8x##w##_t q1, const uint8x##w##_t q2, const uint8x##w##_t q3, \ 140 flat2 = vmax##r##u8(flat2, vabd##r##u8(q3, q0)); \ 235 const uint8x8_t q2, const uint8x8_t q3, in calc_7_tap_filter_8() argument 250 *oq0 = calc_7_tap_filter_8_kernel(p3, p0, q0, q3, &sum); in calc_7_tap_filter_8() 251 *oq1 = calc_7_tap_filter_8_kernel(p2, q0, q1, q3, &sum); in calc_7_tap_filter_8() [all …]
|
/external/boringssl/src/crypto/curve25519/asm/ |
D | x25519-asm-arm.S | 83 vshr.u64 q3,q2,#7 113 vand q5,q5,q3 114 vand q7,q7,q3 115 vand q9,q9,q3 116 vand q11,q11,q3 117 vand q3,q13,q3 148 vadd.i64 q3,q3,q13 150 vadd.i64 q14,q3,q0 179 vsub.i64 q3,q3,q10 190 vsub.i64 q3,q6,q7 [all …]
|
/external/icu/icu4c/source/test/perf/howExpensiveIs/ |
D | sieve.cpp | 108 double qs(double *times, int n, double *q1, double *q2, double *q3) { in qs() argument 111 *q3 = medianof(times,n,3); in qs() 112 return *q3-*q1; in qs() 116 double q1,q2,q3; in uprv_getMeanTime() local 121 double iqr = qs(times,n,&q1,&q2,&q3); in uprv_getMeanTime() 123 double rangeMax = (q3+(1.5*iqr)); in uprv_getMeanTime() 128 printf("iqr: %.9f, q1=%.9f, q2=%.9f, q3=%.9f, max=%.9f, n=%d\n", iqr,q1,q2,q3,(double)-1, n); in uprv_getMeanTime() 152 double iqr = qs(times,n,&q1,&q2,&q3); in uprv_getMeanTime() 154 rangeMax = (q3+(1.5*iqr)); in uprv_getMeanTime() 182 printf("min: %.9f, q1=%.9f, q2=%.9f, q3=%.9f, max=%.9f, n=%d\n", minTime,q1,q2,q3,maxTime, n); in uprv_getMeanTime()
|