/external/libhevc/common/arm/ |
D | ihevc_itrans_recon_32x32.s | 216 vmlal.s16 q12,d9,d0[3] @// y1 * cos1 + y3 * cos3(part of b0) 217 vmlal.s16 q13,d9,d2[1] @// y1 * cos3 - y3 * sin1(part of b1) 218 vmlal.s16 q14,d9,d3[3] @// y1 * sin3 - y3 * cos1(part of b2) 219 vmlal.s16 q15,d9,d5[1] @// y1 * sin1 - y3 * sin3(part of b3) 226 vmlal.s16 q10,d11,d0[2] 230 vmlal.s16 q11,d11,d1[2] 233 vmlal.s16 q8,d11,d2[2] 236 vmlal.s16 q9,d11,d3[2] 251 vmlal.s16 q12,d14,d1[1] 252 vmlal.s16 q13,d14,d3[3] [all …]
|
D | ihevc_inter_pred_filters_luma_vert_w16inp.s | 150 vmlal.s16 q4,d0,d22 @mul_res1 = vmlal_u8(mul_res1, src_tmp1, coeffabs_0)@ 152 vmlal.s16 q4,d2,d24 @mul_res1 = vmlal_u8(mul_res1, src_tmp3, coeffabs_2)@ 154 vmlal.s16 q4,d3,d25 @mul_res1 = vmlal_u8(mul_res1, src_tmp4, coeffabs_3)@ 156 vmlal.s16 q4,d4,d26 @mul_res1 = vmlal_u8(mul_res1, src_tmp1, coeffabs_4)@ 158 vmlal.s16 q4,d5,d27 @mul_res1 = vmlal_u8(mul_res1, src_tmp2, coeffabs_5)@ 159 vmlal.s16 q4,d6,d28 @mul_res1 = vmlal_u8(mul_res1, src_tmp3, coeffabs_6)@ 160 vmlal.s16 q4,d7,d29 @mul_res1 = vmlal_u8(mul_res1, src_tmp4, coeffabs_7)@ 166 vmlal.s16 q5,d1,d22 @mul_res2 = vmlal_u8(mul_res2, src_tmp2, coeffabs_0)@ 168 vmlal.s16 q5,d3,d24 @mul_res2 = vmlal_u8(mul_res2, src_tmp4, coeffabs_2)@ 170 vmlal.s16 q5,d4,d25 @mul_res2 = vmlal_u8(mul_res2, src_tmp1, coeffabs_3)@ [all …]
|
D | ihevc_inter_pred_luma_vert_w16inp_w16out.s | 160 vmlal.s16 q4,d0,d22 @mul_res1 = vmlal_u8(mul_res1, src_tmp1, coeffabs_0)@ 162 vmlal.s16 q4,d2,d24 @mul_res1 = vmlal_u8(mul_res1, src_tmp3, coeffabs_2)@ 164 vmlal.s16 q4,d3,d25 @mul_res1 = vmlal_u8(mul_res1, src_tmp4, coeffabs_3)@ 166 vmlal.s16 q4,d4,d26 @mul_res1 = vmlal_u8(mul_res1, src_tmp1, coeffabs_4)@ 168 vmlal.s16 q4,d5,d27 @mul_res1 = vmlal_u8(mul_res1, src_tmp2, coeffabs_5)@ 169 vmlal.s16 q4,d6,d28 @mul_res1 = vmlal_u8(mul_res1, src_tmp3, coeffabs_6)@ 170 vmlal.s16 q4,d7,d29 @mul_res1 = vmlal_u8(mul_res1, src_tmp4, coeffabs_7)@ 176 vmlal.s16 q5,d1,d22 @mul_res2 = vmlal_u8(mul_res2, src_tmp2, coeffabs_0)@ 178 vmlal.s16 q5,d3,d24 @mul_res2 = vmlal_u8(mul_res2, src_tmp4, coeffabs_2)@ 180 vmlal.s16 q5,d4,d25 @mul_res2 = vmlal_u8(mul_res2, src_tmp1, coeffabs_3)@ [all …]
|
D | ihevc_itrans_recon_16x16.s | 247 vmlal.s16 q12,d7,d0[3] @// y1 * cos1 + y3 * cos3(part of b0) 248 vmlal.s16 q13,d7,d2[1] @// y1 * cos3 - y3 * sin1(part of b1) 249 vmlal.s16 q14,d7,d3[3] @// y1 * sin3 - y3 * cos1(part of b2) 258 vmlal.s16 q6,d11,d0[2] 260 vmlal.s16 q7,d11,d1[2] 262 vmlal.s16 q8,d11,d2[2] 264 vmlal.s16 q9,d11,d3[2] 269 vmlal.s16 q12,d8,d1[1] 270 vmlal.s16 q13,d8,d3[3] 275 vmlal.s16 q12,d9,d1[3] [all …]
|
D | ihevc_intra_pred_luma_planar.s | 188 vmlal.u8 q6, d5, d0 @(1)(row+1) * src[nt-1] 191 vmlal.u8 q6, d8, d1 @(1)(col+1) * src[3nt+1] 194 vmlal.u8 q6, d6, d3 @(1)(nt-1-row) * src[2nt+1+col] 197 vmlal.u8 q6, d9, d20 @(1)(nt-1-col) * src[2nt-1-row] 205 vmlal.u8 q15, d5, d0 @(2) 208 vmlal.u8 q15, d8, d1 @(2) 210 vmlal.u8 q15, d6, d3 @(2) 211 vmlal.u8 q15, d9, d21 @(2) 219 vmlal.u8 q14, d5, d0 @(3) 222 vmlal.u8 q14, d8, d1 @(3) [all …]
|
D | ihevc_inter_pred_chroma_vert_w16inp_w16out.s | 144 vmlal.s16 q0,d2,d13 146 vmlal.s16 q4,d3,d13 149 vmlal.s16 q0,d3,d14 150 vmlal.s16 q4,d6,d14 151 vmlal.s16 q0,d6,d15 152 vmlal.s16 q4,d2,d15 188 vmlal.s16 q15,d1,d13 189 vmlal.s16 q15,d2,d14 191 vmlal.s16 q15,d3,d15 197 vmlal.s16 q14,d2,d13 [all …]
|
D | ihevc_inter_pred_chroma_vert_w16inp.s | 144 vmlal.s16 q0,d2,d13 146 vmlal.s16 q4,d3,d13 149 vmlal.s16 q0,d3,d14 150 vmlal.s16 q4,d6,d14 151 vmlal.s16 q0,d6,d15 152 vmlal.s16 q4,d2,d15 189 vmlal.s16 q15,d1,d13 190 vmlal.s16 q15,d2,d14 192 vmlal.s16 q15,d3,d15 197 vmlal.s16 q14,d2,d13 [all …]
|
D | ihevc_intra_pred_chroma_planar.s | 182 vmlal.u8 q6, d6, d10 @(nt-1-row) * src[2nt+1+col] 184 vmlal.u8 q6, d8, d1 @(col+1) * src[3nt+1] 186 vmlal.u8 q6, d30, d4 @(nt-1-col) * src[2nt-1-row] 192 vmlal.u8 q14,d6,d11 196 vmlal.u8 q14,d31,d4 198 vmlal.u8 q14,d9,d1 203 vmlal.u8 q13, d19, d10 @(nt-1-row) * src[2nt+1+col] 205 vmlal.u8 q13, d8, d1 @(col+1) * src[3nt+1] 207 vmlal.u8 q13, d30, d3 @(nt-1-col) * src[2nt-1-row] 216 vmlal.u8 q12,d19,d11 [all …]
|
D | ihevc_inter_pred_filters_luma_vert.s | 164 vmlal.u8 q4,d3,d25 @mul_res1 = vmlal_u8(mul_res1, src_tmp4, coeffabs_3)@ 166 vmlal.u8 q4,d4,d26 @mul_res1 = vmlal_u8(mul_res1, src_tmp1, coeffabs_4)@ 170 vmlal.u8 q4,d6,d28 @mul_res1 = vmlal_u8(mul_res1, src_tmp3, coeffabs_6)@ 185 vmlal.u8 q5,d4,d25 @mul_res2 = vmlal_u8(mul_res2, src_tmp1, coeffabs_3)@ 187 vmlal.u8 q5,d5,d26 @mul_res2 = vmlal_u8(mul_res2, src_tmp2, coeffabs_4)@ 192 vmlal.u8 q5,d7,d28 @mul_res2 = vmlal_u8(mul_res2, src_tmp4, coeffabs_6)@ 206 vmlal.u8 q6,d5,d25 207 vmlal.u8 q6,d6,d26 209 vmlal.u8 q6,d16,d28 220 vmlal.u8 q7,d6,d25 [all …]
|
D | ihevc_itrans_recon_4x4_ttype1.s | 144 vmlal.s16 q3,d0,d4[0] @74 * pi2_src[1] + 29 * pi2_src[0] 145 vmlal.s16 q3,d3,d4[1] @74 * pi2_src[1] + 29 * pi2_src[0] + 55 * pi2_src[3] 146 …vmlal.s16 q3,d2,d4[3] @pi2_out[0] = 29* pi2_src[0] + 74 * pi2_src[1] + 84* pi2_s… 149 vmlal.s16 q4,d0,d4[1] @74 * pi2_src[1] + 55 * pi2_src[0] 155 …vmlal.s16 q5,d3,d4[2] @pi2_out[2] = 74 * pi2_src[0] - 74 * pi2_src[2] + 74 * pi2… 160 …vmlal.s16 q6,d0,d4[3] @pi2_out[3] = 84 * pi2_src[0] - 74 * pi2_src[1] + 55 * pi2… 183 vmlal.s16 q3,d14,d4[0] @74 * pi2_src[1] + 29 * pi2_src[0] 184 vmlal.s16 q3,d17,d4[1] @74 * pi2_src[1] + 29 * pi2_src[0] + 55 * pi2_src[3] 185 …vmlal.s16 q3,d16,d4[3] @pi2_out[0] = 29* pi2_src[0] + 74 * pi2_src[1] + 84* pi2_s… 188 vmlal.s16 q4,d14,d4[1] @74 * pi2_src[1] + 55 * pi2_src[0] [all …]
|
D | ihevc_itrans_recon_8x8.s | 200 vmlal.s16 q12,d7,d0[3] @// y1 * cos1 + y3 * cos3(part of b0) 234 vmlal.s16 q12,d14,d1[1] @// y1 * cos1 + y3 * cos3 + y5 * sin3(part of b0) 236 vmlal.s16 q14,d14,d1[3] @// y1 * sin3 - y3 * cos1 + y5 * sin1(part of b2) 237 vmlal.s16 q15,d14,d0[3] @// y1 * sin1 - y3 * sin3 + y5 * cos3(part of b3) 240 vmlal.s16 q3,d11,d1[2] @// d0 = y2 * cos2 + y6 * sin2(part of a0 and a1) 245 …vmlal.s16 q12,d15,d1[3] @// b0 = y1 * cos1 + y3 * cos3 + y5 * sin3 + y7 * sin1(par… 247 …vmlal.s16 q14,d15,d0[3] @// b2 = y1 * sin3 - y3 * cos1 + y5 * sin1 + y7 * cos3(par… 307 vmlal.s16 q12,d7,d0[3] @// y1 * cos1 + y3 * cos3(part of b0) 356 vmlal.s16 q12,d9,d0[3] @// y1 * cos1 + y3 * cos3(part of b0) 367 vmlal.s16 q12,d16,d1[1] @// y1 * cos1 + y3 * cos3 + y5 * sin3(part of b0) [all …]
|
/external/boringssl/src/crypto/curve25519/asm/ |
D | x25519-asm-arm.S | 331 vmlal.s32 q12,d11,d1 332 vmlal.s32 q12,d12,d0 333 vmlal.s32 q12,d13,d23 334 vmlal.s32 q12,d16,d22 335 vmlal.s32 q12,d7,d21 337 vmlal.s32 q10,d4,d1 338 vmlal.s32 q10,d13,d0 339 vmlal.s32 q10,d6,d23 340 vmlal.s32 q10,d17,d22 342 vmlal.s32 q13,d11,d3 [all …]
|
/external/libavc/encoder/arm/ |
D | ih264e_half_pel.s | 119 vmlal.u8 q4, d31, d1 @// a0 + a5 + 20a2 (column1,row0) 121 vmlal.u8 q5, d30, d1 @// a0 + a5 + 20a2 (column2,row0) 123 vmlal.u8 q6, d29, d1 @// a0 + a5 + 20a2 (column3,row0) 125 vmlal.u8 q7, d28, d1 @// a0 + a5 + 20a2 (column1,row1) 128 vmlal.u8 q8, d27, d1 @// a0 + a5 + 20a2 (column2,row1) 130 vmlal.u8 q9, d26, d1 @// a0 + a5 + 20a2 (column3,row1) 132 vmlal.u8 q4, d31, d1 @// a0 + a5 + 20a2 + 20a3 (column1,row0) 134 vmlal.u8 q5, d30, d1 @// a0 + a5 + 20a2 + 20a3 (column2,row0) 136 vmlal.u8 q6, d29, d1 @// a0 + a5 + 20a2 + 20a3 (column3,row0) 138 vmlal.u8 q7, d28, d1 @// a0 + a5 + 20a2 + 20a3 (column1,row1) [all …]
|
/external/boringssl/src/crypto/poly1305/ |
D | poly1305_arm_asm.S | 463 # asm 1: vmlal.u32 <r4=reg128#16,<x01=reg128#9%bot,<z34=reg128#6%top 464 # asm 2: vmlal.u32 <r4=q15,<x01=d16,<z34=d11 465 vmlal.u32 q15,d16,d11 478 # asm 1: vmlal.u32 <r4=reg128#16,<x01=reg128#9%top,<z34=reg128#6%bot 479 # asm 2: vmlal.u32 <r4=q15,<x01=d17,<z34=d10 480 vmlal.u32 q15,d17,d10 493 # asm 1: vmlal.u32 <r4=reg128#16,<x23=reg128#10%bot,<z12=reg128#3%top 494 # asm 2: vmlal.u32 <r4=q15,<x23=d18,<z12=d5 495 vmlal.u32 q15,d18,d5 503 # asm 1: vmlal.u32 <r4=reg128#16,<x23=reg128#10%top,<z12=reg128#3%bot [all …]
|
/external/libavc/common/arm/ |
D | ih264_inter_pred_chroma_a9q.s | 149 vmlal.u8 q5, d5, d30 150 vmlal.u8 q5, d3, d29 151 vmlal.u8 q5, d8, d31 157 vmlal.u8 q6, d1, d28 158 vmlal.u8 q6, d9, d31 159 vmlal.u8 q6, d4, d29 176 vmlal.u8 q5, d5, d30 177 vmlal.u8 q5, d3, d29 178 vmlal.u8 q5, d8, d31 182 vmlal.u8 q6, d1, d28 [all …]
|
D | ih264_inter_pred_luma_horz_qpel_vert_qpel_a9q.s | 149 vmlal.u8 q12, d4, d30 150 vmlal.u8 q12, d6, d30 160 vmlal.u8 q14, d20, d30 161 vmlal.u8 q14, d21, d30 166 vmlal.u8 q12, d5, d30 167 vmlal.u8 q12, d7, d30 180 vmlal.u8 q12, d20, d30 181 vmlal.u8 q12, d21, d30 186 vmlal.u8 q8, d6, d30 187 vmlal.u8 q8, d8, d30 [all …]
|
D | ih264_inter_pred_filters_luma_horz_a9q.s | 135 vmlal.u8 q4, d31, d1 @// a0 + a5 + 20a2 (column1,row0) 137 vmlal.u8 q5, d30, d1 @// a0 + a5 + 20a2 (column2,row0) 139 vmlal.u8 q7, d28, d1 @// a0 + a5 + 20a2 (column1,row1) 141 vmlal.u8 q8, d27, d1 @// a0 + a5 + 20a2 (column2,row1) 143 vmlal.u8 q4, d31, d1 @// a0 + a5 + 20a2 + 20a3 (column1,row0) 145 vmlal.u8 q5, d30, d1 @// a0 + a5 + 20a2 + 20a3 (column2,row0) 147 vmlal.u8 q7, d28, d1 @// a0 + a5 + 20a2 + 20a3 (column1,row1) 149 vmlal.u8 q8, d27, d1 @// a0 + a5 + 20a2 + 20a3 (column2,row1) 190 vmlal.u8 q7, d25, d1 @// a0 + a5 + 20a2 (column1,row1) 191 vmlal.u8 q7, d24, d1 @// a0 + a5 + 20a2 + 20a3 (column1,row1) [all …]
|
D | ih264_inter_pred_luma_horz_hpel_vert_hpel_a9q.s | 166 vmlal.s16 q1, d22, d0[0] @// a0 + a5 + 20a2 (column1) 167 vmlal.s16 q15, d23, d0[0] @// a0 + a5 + 20a2 (column1) 172 vmlal.s16 q1, d20, d0[0] @// a0 + a5 + 20a2 + 20a3 (column1) 173 vmlal.s16 q15, d21, d0[0] @// a0 + a5 + 20a2 + 20a3 (column1) 185 vmlal.s16 q1, d22, d0[0] @// a0 + a5 + 20a2 (column2) 186 vmlal.s16 q15, d23, d0[0] @// a0 + a5 + 20a2 (column2) 189 vmlal.s16 q1, d20, d0[0] @// a0 + a5 + 20a2 + 20a3 (column2) 190 vmlal.s16 q15, d21, d0[0] @// a0 + a5 + 20a2 + 20a3 (column2) 225 vmlal.s16 q3, d22, d0[0] @// a0 + a5 + 20a2 (column1) 226 vmlal.s16 q15, d23, d0[0] @// a0 + a5 + 20a2 (column1) [all …]
|
D | ih264_inter_pred_luma_horz_qpel_a9q.s | 142 vmlal.u8 q4, d31, d1 @// a0 + a5 + 20a2 (column1,row0) 144 vmlal.u8 q5, d30, d1 @// a0 + a5 + 20a2 (column2,row0) 146 vmlal.u8 q7, d28, d1 @// a0 + a5 + 20a2 (column1,row1) 148 vmlal.u8 q8, d27, d1 @// a0 + a5 + 20a2 (column2,row1) 150 vmlal.u8 q4, d31, d1 @// a0 + a5 + 20a2 + 20a3 (column1,row0) 152 vmlal.u8 q5, d30, d1 @// a0 + a5 + 20a2 + 20a3 (column2,row0) 154 vmlal.u8 q7, d28, d1 @// a0 + a5 + 20a2 + 20a3 (column1,row1) 156 vmlal.u8 q8, d27, d1 @// a0 + a5 + 20a2 + 20a3 (column2,row1) 202 vmlal.u8 q7, d25, d1 @// a0 + a5 + 20a2 (column1,row1) 203 vmlal.u8 q7, d24, d1 @// a0 + a5 + 20a2 + 20a3 (column1,row1) [all …]
|
/external/llvm/test/MC/ARM/ |
D | neont2-mul-accum-encoding.s | 26 vmlal.s8 q8, d19, d18 27 vmlal.s16 q8, d19, d18 28 vmlal.s32 q8, d19, d18 29 vmlal.u8 q8, d19, d18 30 vmlal.u16 q8, d19, d18 31 vmlal.u32 q8, d19, d18 32 vmlal.s32 q0, d5, d10[0] 34 @ CHECK: vmlal.s8 q8, d19, d18 @ encoding: [0xc3,0xef,0xa2,0x08] 35 @ CHECK: vmlal.s16 q8, d19, d18 @ encoding: [0xd3,0xef,0xa2,0x08] 36 @ CHECK: vmlal.s32 q8, d19, d18 @ encoding: [0xe3,0xef,0xa2,0x08] [all …]
|
D | neon-mul-accum-encoding.s | 24 vmlal.s8 q8, d19, d18 25 vmlal.s16 q8, d19, d18 26 vmlal.s32 q8, d19, d18 27 vmlal.u8 q8, d19, d18 28 vmlal.u16 q8, d19, d18 29 vmlal.u32 q8, d19, d18 31 @ CHECK: vmlal.s8 q8, d19, d18 @ encoding: [0xa2,0x08,0xc3,0xf2] 32 @ CHECK: vmlal.s16 q8, d19, d18 @ encoding: [0xa2,0x08,0xd3,0xf2] 33 @ CHECK: vmlal.s32 q8, d19, d18 @ encoding: [0xa2,0x08,0xe3,0xf2] 34 @ CHECK: vmlal.u8 q8, d19, d18 @ encoding: [0xa2,0x08,0xc3,0xf3] [all …]
|
/external/swiftshader/third_party/LLVM/test/MC/ARM/ |
D | neon-mul-encoding.s | 22 vmlal.s8 q8, d19, d18 23 vmlal.s16 q8, d19, d18 24 vmlal.s32 q8, d19, d18 25 vmlal.u8 q8, d19, d18 26 vmlal.u16 q8, d19, d18 27 vmlal.u32 q8, d19, d18 29 @ CHECK: vmlal.s8 q8, d19, d18 @ encoding: [0xa2,0x08,0xc3,0xf2] 30 @ CHECK: vmlal.s16 q8, d19, d18 @ encoding: [0xa2,0x08,0xd3,0xf2] 31 @ CHECK: vmlal.s32 q8, d19, d18 @ encoding: [0xa2,0x08,0xe3,0xf2] 32 @ CHECK: vmlal.u8 q8, d19, d18 @ encoding: [0xa2,0x08,0xc3,0xf3] [all …]
|
D | neon-mul-accum-encoding.s | 19 @ CHECK: vmlal.s8 q8, d19, d18 @ encoding: [0xa2,0x08,0xc3,0xf2] 20 vmlal.s8 q8, d19, d18 21 @ CHECK: vmlal.s16 q8, d19, d18 @ encoding: [0xa2,0x08,0xd3,0xf2] 22 vmlal.s16 q8, d19, d18 23 @ CHECK: vmlal.s32 q8, d19, d18 @ encoding: [0xa2,0x08,0xe3,0xf2] 24 vmlal.s32 q8, d19, d18 25 @ CHECK: vmlal.u8 q8, d19, d18 @ encoding: [0xa2,0x08,0xc3,0xf3] 26 vmlal.u8 q8, d19, d18 27 @ CHECK: vmlal.u16 q8, d19, d18 @ encoding: [0xa2,0x08,0xd3,0xf3] 28 vmlal.u16 q8, d19, d18 [all …]
|
D | neont2-mul-accum-encoding.s | 24 vmlal.s8 q8, d19, d18 25 vmlal.s16 q8, d19, d18 26 vmlal.s32 q8, d19, d18 27 vmlal.u8 q8, d19, d18 28 vmlal.u16 q8, d19, d18 29 vmlal.u32 q8, d19, d18 31 @ CHECK: vmlal.s8 q8, d19, d18 @ encoding: [0xc3,0xef,0xa2,0x08] 32 @ CHECK: vmlal.s16 q8, d19, d18 @ encoding: [0xd3,0xef,0xa2,0x08] 33 @ CHECK: vmlal.s32 q8, d19, d18 @ encoding: [0xe3,0xef,0xa2,0x08] 34 @ CHECK: vmlal.u8 q8, d19, d18 @ encoding: [0xc3,0xff,0xa2,0x08] [all …]
|
/external/libjpeg-turbo/simd/ |
D | jsimd_arm_neon.S | 269 vmlal.s16 q6, d5, XFIX_1_175875602 273 vmlal.s16 q7, d5, XFIX_1_175875602_MINUS_0_390180644 277 vmlal.s16 q2, ROW6L, XFIX_0_541196100_MINUS_1_847759065 282 vmlal.s16 q6, ROW3L, XFIX_3_072711026_MINUS_2_562915447 294 vmlal.s16 q7, ROW1L, XFIX_1_501321110_MINUS_0_899976223 298 vmlal.s16 q5, ROW5L, XFIX_2_053119869_MINUS_2_562915447 305 vmlal.s16 q6, ROW6L, XFIX_0_541196100 320 vmlal.s16 q4, ROW7L, XFIX_0_298631336_MINUS_0_899976223 347 vmlal.s16 q6, d8, XFIX_1_175875602 350 vmlal.s16 q7, d8, XFIX_1_175875602_MINUS_0_390180644 [all …]
|