/external/libmpeg2/common/arm/ |
D | impeg2_format_conv.s | 139 mov r6, r5 145 sub r6, r6, #16 146 cmp r6, #15 149 cmp r6, #0 155 rsb r6, r6, #16 156 sub r0, r0, r6 157 sub r3, r3, r6 189 mov r6, r5 198 sub r6, r6, #8 199 cmp r6, #7 [all …]
|
/external/libvpx/libvpx/vp8/common/arm/armv6/ |
D | vp8_variance_halfpixvar16x16_v_armv6.asm | 41 ldr r6, [r9, #0] ; load 4 src pixels from next row 45 mvn r6, r6 46 uhsub8 r4, r4, r6 49 usub8 r6, r4, r5 ; calculate difference 51 sel r7, r6, lr ; select bytes with positive difference 52 usub8 r6, r5, r4 ; calculate difference with reversed operands 54 sel r6, r6, lr ; select bytes with negative difference 58 usad8 r5, r6, lr ; calculate sum of negative differences 59 orr r6, r6, r7 ; differences of all 4 pixels 65 uxtb16 r5, r6 ; byte (two pixels) to halfwords [all …]
|
D | vp8_variance_halfpixvar16x16_h_armv6.asm | 40 ldr r6, [r0, #1] ; load 4 src pixels with 1 byte offset 44 mvn r6, r6 45 uhsub8 r4, r4, r6 48 usub8 r6, r4, r5 ; calculate difference 50 sel r7, r6, lr ; select bytes with positive difference 51 usub8 r6, r5, r4 ; calculate difference with reversed operands 53 sel r6, r6, lr ; select bytes with negative difference 57 usad8 r5, r6, lr ; calculate sum of negative differences 58 orr r6, r6, r7 ; differences of all 4 pixels 64 uxtb16 r5, r6 ; byte (two pixels) to halfwords [all …]
|
D | vp8_variance_halfpixvar16x16_hv_armv6.asm | 41 ldr r6, [r0, #1] ; load source pixels b, row N 46 mvn r6, r6 47 uhsub8 r4, r4, r6 59 usub8 r6, r4, r5 ; calculate difference 61 sel r7, r6, lr ; select bytes with positive difference 62 usub8 r6, r5, r4 ; calculate difference with reversed operands 64 sel r6, r6, lr ; select bytes with negative difference 68 usad8 r5, r6, lr ; calculate sum of negative differences 69 orr r6, r6, r7 ; differences of all 4 pixels 75 uxtb16 r5, r6 ; byte (two pixels) to halfwords [all …]
|
D | dequant_idct_v6.asm | 32 smulbb r6, r4, r5 38 strh r6, [r0], #2 41 smulbb r6, r4, r5 49 strh r6, [r0], #2 60 ldr r6, [r0, #8] 65 smulwt r9, r3, r6 66 smulwb r7, r3, r6 67 smulwt r10, r4, r6 68 smulwb r8, r4, r6 72 uadd16 r6, r6, r7 [all …]
|
D | idct_v6.asm | 38 ldr r6, [r0, #(4*2)] ; i5 | i4 42 smulbt r9, r5, r6 ; (ip[5] * cospi8sqrt2minus1) >> 16 43 smulbb r7, r5, r6 ; (ip[4] * cospi8sqrt2minus1) >> 16 44 smulwt r10, r4, r6 ; (ip[5] * sinpi8sqrt2) >> 16 45 smulwb r8, r4, r6 ; (ip[4] * sinpi8sqrt2) >> 16 50 uadd16 r6, r6, r7 ; 5c+5 | 4c+4 64 uadd16 r6, r6, r10 ; d 68 uadd16 r9, r10, r6 ; a+d 69 usub16 r10, r10, r6 ; a-d 70 uadd16 r6, r8, r7 ; b+c [all …]
|
D | loopfilter_v6.asm | 68 ldr r6, [sp, #36] ; load thresh address 77 ldrb r3, [r6] ; thresh 90 uqsub8 r6, r9, r10 ; p3 - p2 95 orr r6, r6, r7 ; abs (p3-p2) 97 uqsub8 lr, r6, r2 ; compare to limit. lr: vp8_filter_mask 99 uqsub8 r6, r11, r12 ; p1 - p0 104 orr r6, r6, r7 ; abs (p1-p0) 105 uqsub8 r7, r6, r2 ; compare to limit 106 uqsub8 r8, r6, r3 ; compare to thresh -- save r8 for later 109 uqsub8 r6, r11, r10 ; p1 - q1 [all …]
|
D | vp8_variance16x16_armv6.asm | 43 usub8 r6, r4, r5 ; calculate difference 45 sel r7, r6, lr ; select bytes with positive difference 48 sel r6, r9, lr ; select bytes with negative difference 52 usad8 r5, r6, lr ; calculate sum of negative differences 53 orr r6, r6, r7 ; differences of all 4 pixels 59 uxtb16 r5, r6 ; byte (two pixels) to halfwords 60 uxtb16 r10, r6, ror #8 ; another two pixels to halfwords 68 usub8 r6, r4, r5 ; calculate difference 69 sel r7, r6, lr ; select bytes with positive difference 71 sel r6, r9, lr ; select bytes with negative difference [all …]
|
/external/boringssl/linux-arm/crypto/sha/ |
D | sha1-armv4-large.S | 23 stmdb sp!,{r4,r5,r6,r7,r8,r9,r10,r11,r12,lr} 25 ldmia r0,{r3,r4,r5,r6,r7} 31 mov r6,r6,ror#30 41 eor r10,r5,r6 @ F_xx_xx 48 eor r10,r5,r6 @ F_xx_xx 56 eor r10,r10,r6,ror#2 @ F_00_19(B,C,D) 63 add r6,r8,r6,ror#2 @ E+=K_00_19 68 add r6,r6,r7,ror#27 @ E+=ROR(A,27) 72 add r6,r8,r6,ror#2 @ E+=K_00_19 74 add r6,r6,r7,ror#27 @ E+=ROR(A,27) [all …]
|
/external/llvm/test/MC/ARM/ |
D | basic-arm-instructions.s | 68 adc r4, r5, r6 70 adc r4, r5, r6, lsl #1 71 adc r4, r5, r6, lsl #31 72 adc r4, r5, r6, lsr #1 73 adc r4, r5, r6, lsr #31 74 adc r4, r5, r6, lsr #32 75 adc r4, r5, r6, asr #1 76 adc r4, r5, r6, asr #31 77 adc r4, r5, r6, asr #32 78 adc r4, r5, r6, ror #1 [all …]
|
D | neon-vld-vst-align.s | 49 vld1.8 {d0}, [r4], r6 50 vld1.8 {d0}, [r4:16], r6 51 vld1.8 {d0}, [r4:32], r6 52 vld1.8 {d0}, [r4:64], r6 53 vld1.8 {d0}, [r4:128], r6 54 vld1.8 {d0}, [r4:256], r6 56 @ CHECK: vld1.8 {d0}, [r4], r6 @ encoding: [0x24,0xf9,0x06,0x07] 58 @ CHECK-ERRORS: vld1.8 {d0}, [r4:16], r6 61 @ CHECK-ERRORS: vld1.8 {d0}, [r4:32], r6 63 @ CHECK: vld1.8 {d0}, [r4:64], r6 @ encoding: [0x24,0xf9,0x16,0x07] [all …]
|
/external/libhevc/decoder/arm/ |
D | ihevcd_fmt_conv_420sp_to_420p.s | 97 @ LDR r6,[sp,#56] @//Load u2_strideuv 113 MOV r6,r8 @// Copying width 117 SUB r6,r6,#16 120 CMP r6,#16 122 CMP r6,#0 128 RSB r6,r6,#16 129 SUB r0,r0,r6 130 SUB r2,r2,r6 164 MOV r6,r8 @// Copying width 168 SUB r6,r6,#16 [all …]
|
D | ihevcd_fmt_conv_420sp_to_420sp.s | 105 MOV r6,r8 @// Copying width 109 SUB r6,r6,#32 118 CMP r6,#32 120 CMP r6,#0 126 RSB r6,r6,#32 127 SUB r0,r0,r6 128 SUB R2,R2,r6 161 MOV r6,r8 @// Copying width 166 SUB r6,r6,#16 171 CMP r6,#16 [all …]
|
/external/tremolo/Tremolo/ |
D | mdctARM.s | 63 LDMDB r2!,{r5,r6,r7,r12} 66 MOV r6, r6, ASR #9 @ r6 = (*--r)>>9 80 MOV r14,r6, ASR #15 82 EORNE r6, r4, r14,ASR #31 83 STRH r6, [r0], r3 125 LDR r6, [r2],#8 130 RSB r6, r6, #0 135 MOV r6, r6, ASR #9 @ r6 = (-*l)>>9 148 MOV r14,r6, ASR #15 150 EORNE r6, r4, r14,ASR #31 [all …]
|
D | mdctLARM.s | 61 LDMDB r2!,{r5,r6,r7,r12} 64 MOV r6, r6, ASR #9 @ r6 = (*--r)>>9 78 MOV r14,r6, ASR #15 80 EORNE r6, r4, r14,ASR #31 81 STRH r6, [r0], r3 123 LDR r6, [r2],#8 128 RSB r6, r6, #0 133 MOV r6, r6, ASR #9 @ r6 = (-*l)>>9 146 MOV r14,r6, ASR #15 148 EORNE r6, r4, r14,ASR #31 [all …]
|
D | bitwiseARM.s | 67 STMFD r13!,{r5,r6} 79 LDRLT r6,[r3] @ r6 = ptr[1] 82 ORRLT r10,r10,r6,LSL r12 @ r10= first bitsLeftInSeg bits+crap 97 LDMIA r11,{r6,r12,r14} @ r6 = buffer 100 LDR r6,[r6] @ r6 = buffer->data 103 ADD r6,r6,r12 @ r6 = buffer->data+begin 105 LDRB r12,[r6],#1 @ r12= *buffer 116 LDMFD r13!,{r5,r6,r10,r11,PC} 125 LDMFD r13!,{r5,r6,r10,r11,PC} 130 LDMFD r13!,{r5,r6,r10,r11,PC} [all …]
|
/external/libavc/encoder/arm/ |
D | ih264e_fmt_conv.s | 85 mov r6, r5 91 sub r6, r6, #16 92 cmp r6, #15 95 cmp r6, #0 101 rsb r6, r6, #16 102 sub r0, r0, r6 103 sub r3, r3, r6 136 mov r6, r5 145 sub r6, r6, #8 146 cmp r6, #7 [all …]
|
D | ih264e_evaluate_intra4x4_modes_a9q.s | 112 @r6 = u4_intra_mode, 160 subs r6, r1, #0 162 lslne r6, r0, #2 164 moveq r6, r0 @ 166 add r9, r6, r9 168 subs r6, r11, r9 183 subs r6, r1, #1 186 lslne r6, r0, #2 190 moveq r6, r0 @ 191 add r9, r6, r9 [all …]
|
/external/jpeg/ |
D | armv6_idct.S | 71 stmdb sp!, {r4, r5, r6, r7, r8, r9, r10, r11, r12, r14} 92 ldrsh r6, [r14, #96] 102 orreqs r8, r6, r7 112 mla r6, r11, r6, r2 120 rsb r2, r6, r2, lsl #1 157 add r0, r0, r6 158 sub r2, r2, r6 159 sub r6, r0, r6, lsl #1 191 add r6, r6, r5 193 sub r5, r6, r5, lsl #1 [all …]
|
/external/libhevc/common/arm/ |
D | ihevc_intra_pred_luma_mode_18_34.s | 127 moveq r6,#1 128 movne r6,#-1 133 vld1.8 {d0},[r8],r6 135 vld1.8 {d1},[r8],r6 137 vld1.8 {d2},[r8],r6 138 vld1.8 {d3},[r8],r6 140 vld1.8 {d4},[r8],r6 141 vld1.8 {d5},[r8],r6 142 vld1.8 {d6},[r8],r6 144 vld1.8 {d7},[r8],r6 [all …]
|
/external/llvm/test/MC/Disassembler/ARM/ |
D | basic-arm-instructions.txt | 43 # CHECK: adc r4, r5, r6 45 # CHECK: adc r4, r5, r6, lsl #1 46 # CHECK: adc r4, r5, r6, lsl #31 47 # CHECK: adc r4, r5, r6, lsr #1 48 # CHECK: adc r4, r5, r6, lsr #31 49 # CHECK: adc r4, r5, r6, lsr #32 50 # CHECK: adc r4, r5, r6, asr #1 51 # CHECK: adc r4, r5, r6, asr #31 52 # CHECK: adc r4, r5, r6, asr #32 53 # CHECK: adc r4, r5, r6, ror #1 [all …]
|
/external/boringssl/src/crypto/chacha/ |
D | chacha_vec_arm.S | 64 push {r4, r5, r6, r7, r8, r9, r10, fp, lr} 77 add r6, r7, #144 98 stmia r6!, {r0, r1} 99 mov r6, r5 116 stmia r6!, {r0, r1, r2, r3} 174 ldr r6, [r3, #28] 179 str r6, [r7, #120] 181 ldr r6, [r7, #60] 191 str r6, [r7, #124] 209 mov r6, r3 [all …]
|
/external/libvpx/libvpx/vp9/common/arm/neon/ |
D | vp9_avg_neon.asm | 19 push {r4-r6, lr} 21 mov r6, r2 39 vld1.8 {q8-q9}, [r6@128]! 40 vld1.8 {q10-q11}, [r6@128], r4 49 pop {r4-r6, pc} 54 vld1.8 {q8-q9}, [r6@128], r3 55 vld1.8 {q10-q11}, [r6@128], r3 60 pld [r6] 62 pld [r6, r3] 68 pop {r4-r6, pc} [all …]
|
/external/libvpx/libvpx/vp8/encoder/arm/neon/ |
D | subtract_neon.asm | 32 ldr r6, [r0, #vp8_block_src_stride] 36 vld1.8 {d0}, [r3], r6 ;load src 38 vld1.8 {d2}, [r3], r6 40 vld1.8 {d4}, [r3], r6 42 vld1.8 {d6}, [r3], r6 72 mov r6, #32 ; "diff" stride x2 94 vst1.16 {q8}, [r0], r6 ;store diff 95 vst1.16 {q9}, [r5], r6 96 vst1.16 {q10}, [r0], r6 97 vst1.16 {q11}, [r5], r6 [all …]
|
/external/libvpx/libvpx/vp8/encoder/arm/armv5te/ |
D | vp8_packtokens_mbrow_armv5.asm | 50 ldr r6, _VP8_COMMON_MBrows_ 53 ldr r5, [r4, r6] ; load up mb_rows 81 ldrb r6, [r1, #tokenextra_token] ; t 84 add r4, r4, r6, lsl #3 ; a = vp8_coef_encodings + t 89 ldr r6, [r4, #vp8_token_value] ; v 101 lsl r12, r6, r4 ; r12 = v << 32 - n 112 mul r6, r4, r7 ; ((range-1) * pp[i>>1])) 120 add r4, r7, r6, lsr #8 ; 1 + (((range-1) * pp[i>>1]) >> 8) 126 clz r6, r4 127 sub r6, r6, #24 ; shift [all …]
|