/frameworks/av/media/libstagefright/codecs/amrwbenc/src/asm/ARMV7/ |
D | convolve_neon.s | 41 ADD r5, r3, #1 @ i = n + 1 45 SUB r5, r5, #1 50 CMP r5, #0 57 SUBS r5, r5, #4 63 VMOV.S32 r5, D20[0] 64 ADD r5, r5, r8 65 ADD r5, r11, r5, LSL #1 66 MOV r5, r5, LSR #16 @extract_h(s) 68 STRH r5, [r2], #2 @y[n] 73 ADD r5, r3, #1 [all …]
|
D | cor_h_vec_neon.s | 31 @r5 ---- cor_1[] 48 MOV r5, #0 @L_sum1 = 0 59 MLA r5, r12, r8, r5 65 MLA r5, r12, r14, r5 67 MOV r5, r5, LSL #2 @L_sum1 = (L_sum1 << 2) 69 ADD r9, r5, r14 70 MOV r5, r9, ASR #16 76 MUL r12, r5, r10 78 MOV r5, r12, ASR #15 86 ADD r5, r5, r10 [all …]
|
D | Norm_Corr_neon.s | 31 @ r5 --- t_max 59 ADD r5, r0, r11, LSL #1 @get the &exc[k] 63 MOV r0, r5 165 VMOV.S32 r5, D22[0] 167 @r5 --- L_tmp, r6 --- L_tmp1 169 ADD r5, r10, r5, LSL #1 @L_tmp = (L_tmp << 1) + 1 172 CLZ r10, r5 173 CMP r5, #0 174 RSBLT r11, r5, #0 178 MOV r5, r5, LSL r10 @L_tmp = (L_tmp << exp) [all …]
|
/frameworks/av/media/libstagefright/codecs/mp3dec/src/asm/ |
D | pvmp3_mdct_18_gcc.s | 51 mov r5,r0 92 mov r0,r5 @@ r0 = vec 94 add r0,r5,#0x24 @@ r0 = &vec[9] 97 ldr r0,[r5,#0x20] 98 ldr r2,[r5,#0x40] 99 str r0,[r5,#0x40] 100 ldr r0,[r5,#0x1c] 101 ldr r3,[r5,#0x38] 102 str r0,[r5,#0x38] 103 ldr r1,[r5,#0x18] [all …]
|
D | pvmp3_polyphase_filter_window_gcc.s | 69 ldr r5,[r3] 73 smlal r2,r9,lr,r5 77 smlal r5,r11,r2,r5 78 smull r6,r5,r2,r6 79 sub r9,r9,r5 80 ldr r5,[r1,#8] 83 smlal r6,r9,r5,r7 84 smull r6,r2,r5,r8 85 ldr r5,[r1,#0xc] 87 smlal r8,r9,r5,r8 [all …]
|
D | pvmp3_dct_16_gcc.s | 77 smull r4,r5,lr,r4 187 sub lr,r5,lr 189 add r3,r5,r3 190 smull r5,lr,r4,lr 192 ldr r5,[sp,#0x10] 194 sub r5,r5,r6 196 mov r5,r5,lsl #1 197 smull r7,r5,r6,r5 215 rsb r5,r5,#0 231 sub r4,lr,r5 [all …]
|
D | pvmp3_dct_9_gcc.s | 52 ldr r5,[r0,#8] 54 add r12,r4,r5 55 sub r4,r4,r5 56 ldr r5,[r0, #0x14] 59 add r6,r5,r7 60 sub r5,r5,r7 118 mov r1,r5,lsl #1 133 add r4,r5,r4 148 smull r5,lr,r4,lr 154 smull r5,lr,r7,r1 [all …]
|
/frameworks/av/media/libstagefright/codecs/on2/h264dec/omxdl/arm_neon/vc/m4p10/src_gcc/ |
D | armVCM4P10_Average_4x_Align_unsafe_s.S | 38 UHSUB8 r5,r10,r12 40 EOR r5,r5,r7 41 STR r5,[r2],r3 49 UHSUB8 r5,r10,r12 52 EOR r5,r5,r7 53 STR r5,[r2],r3 66 LDR r5,[r0,#4] 73 ORR r11,r11,r5,LSL #16 74 UHSUB8 r5,r10,r12 76 EOR r5,r5,r7 [all …]
|
D | armVCM4P10_InterpolateLuma_DiagCopy_unsafe_s.S | 37 LDR r5,[r0,#4] 41 UQSUB16 r5,r5,r6 45 USAT16 r5,#13,r5 49 AND r5,r12,r5,LSR #5 52 ORR r10,r4,r5,LSL #8 69 LDR r5,[r0,#4] 73 UQSUB16 r5,r5,r6 77 USAT16 r5,#13,r5 81 AND r5,r12,r5,LSR #5 84 ORR r10,r4,r5,LSL #8 [all …]
|
D | omxVCM4P10_PredictIntra_4x4_s.S | 43 LDRD r4,r5,[sp,#0x50] 54 ADD r11,r3,r5 55 ADD r12,r5,r5 63 ADD r11,r3,r5 64 ADD r12,r5,r5 88 ADD r11,r3,r5 89 ADD r12,r5,r5 97 ADD r11,r3,r5 98 ADD r12,r5,r5 109 ADD r11,r3,r5 [all …]
|
D | armVCM4P10_InterpolateLuma_Copy_unsafe_s.S | 41 LDR r5,[r0],r1 44 STR r5,[r2],r3 50 LDR r5,[r0,#4] 55 ORR r4,r4,r5,LSL #24 59 LDR r5,[r0,#4] 65 ORR r4,r4,r5,LSL #24 72 LDR r5,[r0,#4] 77 ORR r4,r4,r5,LSL #16 82 LDR r5,[r0,#4] 87 ORR r4,r4,r5,LSL #16 [all …]
|
D | armVCM4P10_UnpackBlock4x4_s.S | 35 MOV r5,#0 37 STRD r4,r5,[r1,#0] 38 STRD r4,r5,[r1,#8] 39 STRD r4,r5,[r1,#0x10] 40 STRD r4,r5,[r1,#0x18] 43 LDRSBNE r5,[r2,#1] 47 ORRNE r4,r4,r5,LSL #8
|
D | omxVCM4P10_TransformDequantChromaDCFromPair_s.S | 39 ldrsbne r5, [r9, #1] 43 orrne r4, r4, r5, lsl #8 50 ldr r5, .LarmVCM4P10_QPDivTable 51 P0: add r5, pc 56 ldrsb r9, [r5, r2] 58 sadd16 r5, r3, r4 61 vmov d0, r5, r6
|
/frameworks/av/media/libstagefright/codecs/amrwbenc/src/asm/ARMV5E/ |
D | cor_h_vec_opt.s | 30 @r5 ---- cor_1[] 47 MOV r5, #0 @L_sum1 = 0 58 MLA r5, r12, r8, r5 64 MLA r5, r12, r14, r5 66 MOV r5, r5, LSL #2 @L_sum1 = (L_sum1 << 2) 68 ADD r9, r5, r14 69 MOV r5, r9, ASR #16 75 MUL r12, r5, r10 77 MOV r5, r12, ASR #15 85 ADD r5, r5, r10 [all …]
|
D | Norm_Corr_opt.s | 31 @ r5 --- t_max 59 ADD r5, r0, r11, LSL #1 @get the &exc[k] 63 MOV r0, r5 72 MOV r5, #64 89 SUBS r5, r5, #8 104 MOV r5, #0 @L_tmp = 0 116 SMLABB r5, r10, r11, r5 @L_tmp += xn[i] * excf[i] 117 SMLATT r5, r10, r11, r5 @L_tmp += xn[i+1] * excf[i+1] 122 SMLABB r5, r10, r11, r5 123 SMLATT r5, r10, r11, r5 [all …]
|
D | convolve_opt.s | 40 ADD r5, r3, #1 @ i = n + 1 44 SUB r5, r5, #1 48 CMP r5, #0 61 SUBS r5, r5, #4 68 ADD r5, r11, r8, LSL #1 69 MOV r5, r5, LSR #16 @extract_h(s) 71 STRH r5, [r2], #2 @y[n] 75 ADD r5, r3, #1 83 SUB r5, r5, #2 87 CMP r5, #0 [all …]
|
D | syn_filt_opt.s | 40 MOV r5, r13 @ copy yy = y_buf address 56 STRH r6, [r5], #2 57 STRH r7, [r5], #2 58 STRH r8, [r5], #2 59 STRH r9, [r5], #2 60 STRH r10, [r5], #2 61 STRH r11, [r5], #2 62 STRH r12, [r5], #2 63 STRH r14, [r5], #2 74 STRH r6, [r5], #2 [all …]
|
D | scale_sig_opt.s | 44 LDRSH r5, [r4] @load x[i] 45 MOV r12, r5, LSL r10 46 TEQ r5, r12, ASR r10 47 EORNE r12, r8, r5, ASR #31 57 LDRSH r5, [r4] @load x[i] 58 MOV r6, r5, LSL #16 @L_tmp = x[i] << 16 59 MOV r5, r6, ASR r7 @L_tmp >>= exp 60 QADD r11, r5, r9
|
/frameworks/av/media/libstagefright/codecs/m4v_h263/dec/src/ |
D | idct.cpp | 131 int32 r0, r1, r2, r3, r4, r5, r6, r7, r8; /* butterfly nodes */ in idct_intra() local 154 r5 = blk[B_SIZE * 7 + i]; in idct_intra() 158 if (!(r1 | r2 | r3 | r4 | r5 | r6 | r7)) in idct_intra() 182 r8 = W7 * (r4 + r5); in idct_intra() 186 r5 = (r8 - (W1 + W7) * r5); in idct_intra() 201 r6 = r5 + r7; in idct_intra() 202 r5 -= r7; in idct_intra() 209 r2 = (181 * (r4 + r5) + 128) >> 8; /* rounding */ in idct_intra() 210 r4 = (181 * (r4 - r5) + 128) >> 8; in idct_intra() 242 r5 = tmpBLK32[7+(i<<3)]; in idct_intra() [all …]
|
/frameworks/av/media/libstagefright/codecs/aacenc/src/asm/ARMV5E/ |
D | band_nrg_v5.s | 34 mov r5, r2, asr #16 36 cmp r5, #0 73 cmp r4, r5 98 mov r5, r4, lsl #1 100 ldrsh r10, [r2, r5] 101 add r5, r2, r5 103 ldrsh r11, [r5, #2] 120 add r5, r8, r9 124 smull r5, r3, r5, r5 130 add r5, r12, r14 [all …]
|
D | R4R8First_v5.s | 47 add r3, r5, r7 50 sub r5, r5, r7 58 add r2, r0, r5 61 sub r0, r0, r5 102 add r3, r5, r7 @ i2 = buf[5] + buf[7]@ 105 sub r5, r5, r7 @ i3 = buf[5] - buf[7]@ 110 sub r8, r8, r2 @ r5 = (r0 - r2) >> 1@ 113 sub r2, r0, r5 @ r6 = (r1 - i3) >> 1@ 116 add r0, r0, r5 @ r7 = (r1 + i3) >> 1@ 149 add r1, r3, r5 @ i0 = buf[ 9] + buf[11]@ [all …]
|
D | Radix4FFT_v5.s | 65 smulwb r5, r10, r8 @ L_mpy_wx(sinx, t0) 70 sub r3, r3, r5 @ r3 = L_mpy_wx(cosx, t1) - L_mpy_wx(sinx, t0)@ 82 ldrd r10, [r14, #0] @ r4 = xptr[0]@ r5 = xptr[1]@ 86 smulwt r5, r11, r8 @ L_mpy_wx(cosx, t1) 92 sub r5, r5, r7 @ r5 = L_mpy_wx(cosx, t1) - L_mpy_wx(sinx, t0)@ 104 mov r11, r5 @ t1 = r5@ 106 sub r7, r7, r3 @ r5 = L_mpy_wx(cosx, t1) - L_mpy_wx(sinx, t0)@ 110 sub r5, r7, r11 @ r5 = r7 - t1@ 118 add r10, r0, r5 @ xptr[0] = r0 + r5@ 130 sub r10, r0, r5 @ xptr[0] = r0 - r5@
|
/frameworks/native/opengl/libagl/ |
D | iterators.S | 51 stmfd sp!, {r4, r5, r6, r7, r8, lr} 61 ldr r5, [r0, #0] 63 smull r8, lr, r4, r5 64 ldr r5, [r0, #8] 67 smlal r8, lr, r3, r5 76 umull r6, r5, r3, r6 78 mla r7, r3, r7, r5 88 ldmfd sp!, {r4, r5, r6, r7, r8, pc}
|
/frameworks/rs/cpu_ref/ |
D | rsCpuIntrinsics_neon_YuvToRGB.S | 89 movw r5, #((16 * 149 + (128 >> 1) + 128 * 204) >> 1) 90 vdup.i16 q13, r5 91 movw r5, #((-16 * 149 + 128 * 50 + 128 * 104) >> 0) 92 vdup.i16 q14, r5 93 movw r5, #((16 * 149 + (128 << 2) + 128 * 254) >> 1) 94 vdup.i16 q15, r5 227 push {r4,r5} 228 ldr r5, [sp, #8] 233 add r0, r5, LSL #2 234 add r1, r5 [all …]
|
/frameworks/av/media/libstagefright/codecs/aacenc/src/asm/ARMV7/ |
D | Radix4FFT_v7.s | 39 mov r5, r2, lsl #1 42 mov r5, r5, lsl #2 44 rsbeq r12, r5, r5, lsl #2 47 rsb r12, r5, r5, lsl #2 61 add r8, r8, r5 @ xptr += step@ 72 add r8, r8, r5 @ xptr += step@ 83 add r8, r8, r5 105 VSUB.S32 Q5, Q7, Q9 @ r5 = r7 - t1@ 109 VADD.S32 Q8, Q0, Q5 @ xptr[0] = r0 + r5@ 114 sub r8, r8, r5 @ xptr -= step@ [all …]
|