/external/llvm/test/MC/ARM/ |
D | neon-mul-encoding.s | 3 vmul.i8 d16, d16, d17 4 vmul.i16 d16, d16, d17 5 vmul.i32 d16, d16, d17 6 vmul.f32 d16, d16, d17 7 vmul.i8 q8, q8, q9 8 vmul.i16 q8, q8, q9 9 vmul.i32 q8, q8, q9 10 vmul.f32 q8, q8, q9 11 vmul.p8 d16, d16, d17 12 vmul.p8 q8, q8, q9 [all …]
|
D | neont2-mul-encoding.s | 5 vmul.i8 d16, d16, d17 6 vmul.i16 d16, d16, d17 7 vmul.i32 d16, d16, d17 8 vmul.f32 d16, d16, d17 9 vmul.i8 q8, q8, q9 10 vmul.i16 q8, q8, q9 11 vmul.i32 q8, q8, q9 12 vmul.f32 q8, q8, q9 13 vmul.p8 d16, d16, d17 14 vmul.p8 q8, q8, q9 [all …]
|
D | simple-fp-encoding.s | 24 vmul.f64 d16, d17, d16 25 vmul.f64 d20, d17 26 vmul.f32 s0, s1, s0 27 vmul.f32 s11, s21 30 @ CHECK: vmul.f64 d16, d17, d16 @ encoding: [0xa0,0x0b,0x61,0xee] 31 @ CHECK: vmul.f64 d20, d20, d17 @ encoding: [0xa1,0x4b,0x64,0xee] 32 @ CHECK: vmul.f32 s0, s1, s0 @ encoding: [0x80,0x0a,0x20,0xee] 33 @ CHECK: vmul.f32 s11, s11, s21 @ encoding: [0xaa,0x5a,0x65,0xee]
|
/external/llvm/test/CodeGen/ARM/ |
D | neon-spfp.ll | 47 ; CHECK-LINUXA5: vmul.f32 s{{[0-9]*}} 48 ; CHECK-LINUXA8: vmul.f32 s{{[0-9]*}} 49 ; CHECK-LINUXA9: vmul.f32 s{{[0-9]*}} 50 ; CHECK-LINUXA15: vmul.f32 s{{[0-9]*}} 52 ; CHECK-LINUXSWIFT: vmul.f32 d{{[0-9]*}} 54 ; CHECK-UNSAFEA5: vmul.f32 d{{[0-9]*}} 55 ; CHECK-UNSAFEA8: vmul.f32 d{{[0-9]*}} 57 ; CHECK-UNSAFEA9: vmul.f32 s{{[0-9]*}} 58 ; CHECK-UNSAFEA15: vmul.f32 s{{[0-9]*}} 59 ; CHECK-UNSAFESWIFT: vmul.f32 d{{[0-9]*}} [all …]
|
D | fmacs.ll | 16 ; A8: vmul.f32 32 ; A8: vmul.f64 48 ; A8: vmul.f32 60 ; A8: vmul.f32 61 ; A8: vmul.f32 85 ; A8: vmul.f32 86 ; A8: vmul.f32 92 ; A9: vmul.f32 97 ; HARD: vmul.f32 s0, s2, s3
|
D | fmuls.ll | 26 ; VFP2: vmul.f32 s 29 ; NFP1: vmul.f32 d 31 ; NFP0: vmul.f32 s 34 ; CORTEXA8: vmul.f32 s 36 ; CORTEXA8U: vmul.f32 d 38 ; CORTEXA9: vmul.f32 s
|
D | fp-fast.ll | 7 ; CHECK: vmul.f32 16 ; CHECK-NOT: vmul 18 ; CHECK-NOT: vmul 45 ; CHECK: vmul.f32 54 ; CHECK: vmul.f32
|
D | saxpy10-a9.ll | 18 ; CHECK-NEXT: vmul 25 ; CHECK-NEXT: vmul 30 ; CHECK-NEXT: vmul 36 ; CHECK-NEXT: vmul 41 ; CHECK-NEXT: vmul 46 ; CHECK-NEXT: vmul 51 ; CHECK-NEXT: vmul 54 ; CHECK-NEXT: vmul
|
D | vcvt_combine.ll | 7 ; CHECK-NOT: vmul 23 ; CHECK-NOT: vmul 39 ; CHECK: vmul 53 ; CHECK: vmul 67 ; CHECK-NOT: vmul 81 ; CHECK-NOT: vmul
|
D | no-fpu.ll | 26 ; NONEON-NOVFP-NOT: vmul.f64 28 ; NOVFP-NOT: vmul.f64 30 ; NONEON-VFP: vmul.f64
|
D | vdiv_combine.ll | 11 ; CHECK-NOT: {{vdiv|vmul}} 27 ; CHECK-NOT: {{vdiv|vmul}} 41 ; CHECK: {{vdiv|vmul}} 55 ; CHECK: {{vdiv|vmul}} 69 ; CHECK-NOT: {{vdiv|vmul}} 83 ; CHECK-NOT: {{vdiv|vmul}}
|
D | vmul.ll | 5 ;CHECK: vmul.i8 14 ;CHECK: vmul.i16 23 ;CHECK: vmul.i32 32 ;CHECK: vmul.f32 41 ;CHECK: vmul.p8 50 ;CHECK: vmul.i8 59 ;CHECK: vmul.i16 68 ;CHECK: vmul.i32 77 ;CHECK: vmul.f32 86 ;CHECK: vmul.p8 [all …]
|
D | fnmacs.ll | 14 ; A8: vmul.f32 30 ; A8: vmul.f64
|
D | fmscs.ll | 14 ; A8: vmul.f32 30 ; A8: vmul.f64
|
D | arm-modifier.ll | 11 …call void asm sideeffect "vmul.f32 q0, q0, ${0:y} \0A\09vmul.f32 q1, q1, ${0:y} \0A\09vmul.f…
|
/external/llvm/test/CodeGen/Thumb2/ |
D | cortex-fp.ll | 11 ; CORTEXM4: vmul.f32 s 12 ; CORTEXM7: vmul.f32 s 13 ; CORTEXA8: vmul.f32 d 24 ; CORTEXM7: vmul.f64 d 25 ; CORTEXA8: vmul.f64 d
|
/external/libavc/common/arm/ |
D | ih264_intra_pred_chroma_a9q.s | 406 vmul.s16 q7, q5, q4 407 vmul.s16 q8, q6, q4 463 vmul.s16 q6, q2, q4 464 vmul.s16 q8, q2, q5 470 vmul.s16 q2, q3, q10 472 vmul.s16 q9, q3, q10 473 vmul.s16 q7, q3, q15 474 vmul.s16 q4, q3, q15 485 vmul.s16 q2, q3, q10 486 vmul.s16 q9, q3, q10 [all …]
|
D | ih264_weighted_pred_a9q.s | 144 vmul.s16 q2, q2, d2[0] @weight mult. for rows 1,2 145 vmul.s16 q3, q3, d2[0] @weight mult. for rows 3,4 176 vmul.s16 q2, q2, d2[0] @weight mult. for row 1 178 vmul.s16 q3, q3, d2[0] @weight mult. for row 2 179 vmul.s16 q4, q4, d2[0] @weight mult. for row 3 180 vmul.s16 q5, q5, d2[0] @weight mult. for row 4 216 vmul.s16 q6, q6, d2[0] @weight mult. for row 1L 218 vmul.s16 q7, q7, d2[0] @weight mult. for row 1H 220 vmul.s16 q8, q8, d2[0] @weight mult. for row 2L 222 vmul.s16 q9, q9, d2[0] @weight mult. for row 2H [all …]
|
D | ih264_weighted_bi_pred_a9q.s | 184 vmul.s16 q2, q2, d2[0] @weight 1 mult. for rows 1,2 186 vmul.s16 q4, q4, d2[0] @weight 1 mult. for rows 3,4 224 vmul.s16 q2, q2, d2[0] @weight 1 mult. for row 1 228 vmul.s16 q4, q4, d2[0] @weight 1 mult. for row 2 233 vmul.s16 q6, q6, d2[0] @weight 1 mult. for row 3 235 vmul.s16 q8, q8, d2[0] @weight 1 mult. for row 4 279 vmul.s16 q10, q10, d2[0] @weight 1 mult. for row 1L 284 vmul.s16 q2, q2, d2[0] @weight 1 mult. for row 1H 289 vmul.s16 q12, q12, d2[0] @weight 1 mult. for row 2L 294 vmul.s16 q4, q4, d2[0] @weight 1 mult. for row 2H [all …]
|
D | ih264_ihadamard_scaling_a9.s | 141 vmul.s32 q0, q0, q9 @ Q0 = p[i] = (x[i] * trns_coeff[i]) where i = 0..3 142 vmul.s32 q1, q1, q9 @ Q1 = p[i] = (x[i] * trns_coeff[i]) where i = 4..7 143 vmul.s32 q2, q2, q9 @ Q2 = p[i] = (x[i] * trns_coeff[i]) where i = 8..11 144 vmul.s32 q3, q3, q9 @ Q3 = p[i] = (x[i] * trns_coeff[i]) where i = 12..15 236 vmul.s32 q5, q3, q15 237 vmul.s32 q6, q1, q15
|
/external/jpeg/ |
D | jsimd_arm_neon.S | 198 vmul.s16 q2, q2, q10 200 vmul.s16 q3, q3, q11 201 vmul.s16 q4, q4, q12 203 vmul.s16 q5, q5, q13 204 vmul.s16 q6, q6, q14 206 vmul.s16 q7, q7, q15 207 vmul.s16 q8, q8, q10 208 vmul.s16 q9, q9, q11 420 vmul.s16 q2, q2, q9 422 vmul.s16 q3, q3, q10 [all …]
|
/external/freetype/src/smooth/ |
D | ftsmooth.c | 118 FT_Int vmul = mode == FT_RENDER_MODE_LCD_V; in ft_smooth_render_generic() local 176 if ( vmul ) in ft_smooth_render_generic() 194 if ( vmul ) in ft_smooth_render_generic() 275 if ( vmul ) in ft_smooth_render_generic() 294 if ( vmul ) in ft_smooth_render_generic() 339 if ( vmul ) in ft_smooth_render_generic()
|
/external/pdfium/third_party/freetype/src/smooth/ |
D | ftsmooth.c | 118 FT_Int vmul = mode == FT_RENDER_MODE_LCD_V; in ft_smooth_render_generic() local 176 if ( vmul ) in ft_smooth_render_generic() 194 if ( vmul ) in ft_smooth_render_generic() 275 if ( vmul ) in ft_smooth_render_generic() 294 if ( vmul ) in ft_smooth_render_generic() 339 if ( vmul ) in ft_smooth_render_generic()
|
/external/llvm/test/Transforms/LoopVectorize/ARM/ |
D | mul-cast-vect.ll | 25 ; ASM: vmul.i32 72 ; ASM: vmul.i32 89 ; ASM: vmul.i32 106 ; ASM: vmul.i32
|
/external/libvpx/libvpx/vp8/encoder/arm/neon/ |
D | fastquantizeb_neon.asm | 98 vmul.s16 q2, q6, q4 ; x * Dequant 99 vmul.s16 q3, q7, q5 107 vmul.s16 q12, q6, q10 ; x2 * Dequant 108 vmul.s16 q13, q7, q11
|