/external/libhevc/common/arm64/ |
D | ihevc_intra_pred_luma_dc.s | 237 uxtl v20.8h, v0.8b 240 add v20.8h, v20.8h , v24.8h //col 1::7 add 3dc+2 (prol) 243 sqshrun v2.8b, v20.8h,#2 //columns shx2 movn (prol) 263 bsl v20.8b, v3.8b , v16.8b //row 1 (prol) 268 st1 {v20.8b},[x2], x3 //store row 1 (prol) 280 bsl v20.8b, v3.8b , v16.8b //row 3 (prol) 285 st1 {v20.8b},[x2], x3 //store row 3 (prol) 296 bsl v20.8b, v3.8b , v16.8b //row 5 (prol) 301 st1 {v20.8b},[x2], x3 //store row 5 (prol) 314 bsl v20.8b, v3.8b , v16.8b //row 7 (prol) [all …]
|
D | ihevc_intra_pred_chroma_ver.s | 117 ld2 {v20.8b, v21.8b}, [x6],#16 //16 loads (col 0:15) 127 st2 {v20.8b, v21.8b}, [x2],#16 128 st2 {v20.8b, v21.8b}, [x5],#16 129 st2 {v20.8b, v21.8b}, [x8],#16 130 st2 {v20.8b, v21.8b}, [x10],#16 140 st2 {v20.8b, v21.8b}, [x2],#16 141 st2 {v20.8b, v21.8b}, [x5],#16 142 st2 {v20.8b, v21.8b}, [x8],#16 143 st2 {v20.8b, v21.8b}, [x10],#16 153 st2 {v20.8b, v21.8b}, [x2],#16 [all …]
|
D | ihevc_sao_edge_offset_class1.s | 154 …SUB v20.16b, v17.16b , v5.16b //sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 160 ADD v5.16b, v5.16b , v20.16b //edge_idx = vaddq_s8(edge_idx, sign_down) 163 NEG v16.16b, v20.16b //sign_up = vnegq_s8(sign_down) 177 …Uxtl v20.8h, v3.8b //pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(… 179 …SADDW v20.8h, v20.8h , v5.8b //pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0]… 181 …SMAX v20.8h, v20.8h , v2.8h //pi2_tmp_cur_row.val[0] = vmaxq_s16(pi2_tmp_cur_row.val[0… 183 …UMIN v20.8h, v20.8h , v4.8h //pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vminq_u16… 197 xtn v20.8b, v20.8h //vmovn_s16(pi2_tmp_cur_row.val[0]) 200 xtn2 v20.16b, v1.8h //vmovn_s16(pi2_tmp_cur_row.val[1]) 209 ST1 { v20.16b},[x10],x1 //vst1q_u8(pu1_src_cpy, pu1_cur_row) [all …]
|
D | ihevc_inter_pred_luma_vert_w16inp_w16out.s | 182 smull v20.4s,v2.4h,v23.4h //mul_res2 = smull_u8(src_tmp3, coeffabs_1)// 185 smlal v20.4s,v1.4h,v22.4h //mul_res2 = smlal_u8(mul_res2, src_tmp2, coeffabs_0)// 187 smlal v20.4s,v3.4h,v24.4h //mul_res2 = smlal_u8(mul_res2, src_tmp4, coeffabs_2)// 189 smlal v20.4s,v4.4h,v25.4h //mul_res2 = smlal_u8(mul_res2, src_tmp1, coeffabs_3)// 191 smlal v20.4s,v5.4h,v26.4h //mul_res2 = smlal_u8(mul_res2, src_tmp2, coeffabs_4)// 193 smlal v20.4s,v6.4h,v27.4h //mul_res2 = smlal_u8(mul_res2, src_tmp3, coeffabs_5)// 194 smlal v20.4s,v7.4h,v28.4h //mul_res2 = smlal_u8(mul_res2, src_tmp4, coeffabs_6)// 195 smlal v20.4s,v16.4h,v29.4h //mul_res2 = smlal_u8(mul_res2, src_tmp1, coeffabs_7)// 210 sub v20.4s, v20.4s, v30.4s 230 shrn v20.4h, v20.4s, #6 [all …]
|
D | ihevc_inter_pred_filters_luma_vert_w16inp.s | 171 smull v20.4s, v2.4h, v23.4h //mul_res2 = vmull_u8(src_tmp3, coeffabs_1)// 174 smlal v20.4s, v1.4h, v22.4h //mul_res2 = vmlal_u8(mul_res2, src_tmp2, coeffabs_0)// 176 smlal v20.4s, v3.4h, v24.4h //mul_res2 = vmlal_u8(mul_res2, src_tmp4, coeffabs_2)// 178 smlal v20.4s, v4.4h, v25.4h //mul_res2 = vmlal_u8(mul_res2, src_tmp1, coeffabs_3)// 180 smlal v20.4s, v5.4h, v26.4h //mul_res2 = vmlal_u8(mul_res2, src_tmp2, coeffabs_4)// 182 smlal v20.4s, v6.4h, v27.4h //mul_res2 = vmlal_u8(mul_res2, src_tmp3, coeffabs_5)// 183 smlal v20.4s, v7.4h, v28.4h //mul_res2 = vmlal_u8(mul_res2, src_tmp4, coeffabs_6)// 184 smlal v20.4s, v16.4h, v29.4h //mul_res2 = vmlal_u8(mul_res2, src_tmp1, coeffabs_7)// 199 sqshrn v20.4h, v20.4s,#6 218 sqrshrun v20.8b, v20.8h,#6 //sto_res = vqmovun_s16(sto_res_tmp)// [all …]
|
D | ihevc_sao_edge_offset_class1_chroma.s | 183 …SUB v20.16b, v19.16b , v5.16b //sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 189 ADD v5.16b, v5.16b , v20.16b //edge_idx = vaddq_s8(edge_idx, sign_down) 193 NEG v16.16b, v20.16b //sign_up = vnegq_s8(sign_down) 209 …Uxtl v20.8h, v3.8b //pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(… 215 …SADDW v20.8h, v20.8h , v5.8b //pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0]… 217 …SMAX v20.8h, v20.8h , v2.8h //pi2_tmp_cur_row.val[0] = vmaxq_s16(pi2_tmp_cur_row.val[0… 219 …UMIN v20.8h, v20.8h , v4.8h //pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vminq_u16… 243 xtn v20.8b, v20.8h //vmovn_s16(pi2_tmp_cur_row.val[0]) 246 xtn2 v20.16b, v28.8h //vmovn_s16(pi2_tmp_cur_row.val[1]) 257 ST1 { v20.16b},[x10],x1 //vst1q_u8(pu1_src_cpy, pu1_cur_row) [all …]
|
D | ihevc_inter_pred_filters_luma_vert.s | 182 umull v20.8h, v2.8b, v23.8b //mul_res2 = vmull_u8(src_tmp3, coeffabs_1)// 186 umlsl v20.8h, v1.8b, v22.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp2, coeffabs_0)// 190 umlsl v20.8h, v3.8b, v24.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp4, coeffabs_2)// 193 umlal v20.8h, v4.8b, v25.8b //mul_res2 = vmlal_u8(mul_res2, src_tmp1, coeffabs_3)// 196 umlal v20.8h, v5.8b, v26.8b //mul_res2 = vmlal_u8(mul_res2, src_tmp2, coeffabs_4)// 199 umlsl v20.8h, v6.8b, v27.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp3, coeffabs_5)// 202 umlal v20.8h, v7.8b, v28.8b //mul_res2 = vmlal_u8(mul_res2, src_tmp4, coeffabs_6)// 206 umlsl v20.8h, v16.8b, v29.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp1, coeffabs_7)// 224 sqrshrun v20.8b, v20.8h,#6 //sto_res = vqmovun_s16(sto_res_tmp)// 243 st1 {v20.8b},[x14],x6 //vst1_u8(pu1_dst_tmp,sto_res)// [all …]
|
/external/libmpeg2/common/armv8/ |
D | icv_variance_av8.s | 86 umull v20.8h, v0.8b, v0.8b 91 uaddl v21.4s, v20.4h, v22.4h 93 uaddl2 v20.4s, v20.8h, v22.8h 96 add v20.4s, v20.4s, v21.4s 98 add v20.4s, v20.4s, v22.4s 99 addp v20.4s, v20.4s, v20.4s 100 addp v20.2s, v20.2s, v20.2s 106 smov x1, v20.s[0]
|
D | ideint_spatial_filter_av8.s | 70 movi v20.8h, #0 115 uabal v20.8h, v2.8b, v4.8b 127 addp v20.8h, v20.8h, v20.8h 131 uaddlp v20.2s, v20.4h 140 mul v20.2s, v20.2s, v31.2s 149 smov x7, v20.s[0] 177 smov x7, v20.s[1]
|
/external/libhevc/decoder/arm64/ |
D | ihevcd_fmt_conv_420sp_to_rgba8888.s | 203 sMULL v20.4s, v6.4h, v0.h[0] ////(V-128)*C1 FOR R 217 sqshrn v7.4h, v20.4s,#13 ////D10 = (V-128)*C1>>13 4 16-BIT VALUES 230 UADDW v20.8h, v5.8h , v31.8b ////Q10 - HAS Y + B 239 sqxtun v20.8b, v20.8h 251 ZIP1 v27.8b, v20.8b, v21.8b 252 ZIP2 v21.8b, v20.8b, v21.8b 253 mov v20.d[0], v27.d[0] 259 mov v20.d[1], v21.d[0] 266 ZIP1 v25.8h, v20.8h, v22.8h 267 ZIP2 v19.8h, v20.8h, v22.8h [all …]
|
/external/libavc/common/armv8/ |
D | ih264_inter_pred_luma_horz_qpel_vert_hpel_av8.s | 176 uaddl v20.8h, v4.8b, v6.8b 179 mla v18.8h, v20.8h , v28.8h 181 uaddl v20.8h, v1.8b, v11.8b 183 mla v20.8h, v24.8h , v28.8h 187 mls v20.8h, v26.8h , v30.8h 192 st1 {v20.4s}, [x9], #16 193 ext v24.16b, v18.16b , v20.16b , #4 194 ext v26.16b, v18.16b , v20.16b , #6 196 ext v22.16b, v18.16b , v20.16b , #10 198 ext v24.16b, v18.16b , v20.16b , #2 [all …]
|
D | ih264_inter_pred_luma_horz_hpel_vert_hpel_av8.s | 114 uaddl v20.8h, v4.8b, v6.8b 117 mla v18.8h, v20.8h , v28.8h 119 uaddl v20.8h, v1.8b, v11.8b 121 mla v20.8h, v24.8h , v28.8h 125 mls v20.8h, v26.8h , v30.8h 130 ext v24.16b, v18.16b , v20.16b , #4 131 ext v26.16b, v18.16b , v20.16b , #6 133 ext v23.16b, v18.16b , v20.16b , #10 135 ext v24.16b, v18.16b , v20.16b , #2 136 ext v26.16b, v18.16b , v20.16b , #8 [all …]
|
D | ih264_intra_pred_luma_16x16_av8.s | 219 dup v20.16b, v0.b[5] 226 st1 {v20.16b}, [x1], x3 323 dup v20.16b, w15 330 dup v20.8h, w11 337 uqshl v0.8h, v0.8h, v20.8h 339 dup v20.16b, v0.b[0] 343 st1 { v20.16b}, [x1], x3 344 st1 { v20.16b}, [x1], x3 345 st1 { v20.16b}, [x1], x3 346 st1 { v20.16b}, [x1], x3 [all …]
|
D | ih264_inter_pred_chroma_av8.s | 159 umull v20.8h, v0.8b, v28.8b 161 umlal v20.8h, v3.8b, v29.8b 163 umlal v20.8h, v5.8b, v30.8b 165 umlal v20.8h, v8.8b, v31.8b 166 sqrshrun v26.8b, v20.8h, #6 190 umull v20.8h, v10.8b, v28.8b 191 umlal v20.8h, v13.8b, v29.8b 192 umlal v20.8h, v0.8b, v30.8b 193 umlal v20.8h, v3.8b, v31.8b 194 sqrshrun v26.8b, v20.8h, #6 [all …]
|
D | ih264_inter_pred_filters_luma_vert_av8.s | 143 uaddl v20.8h, v1.8b, v11.8b // temp4 = src[0_8] + src[5_8] 145 mla v20.8h, v18.8h , v22.8h // temp4 += temp3 * 20 153 mls v20.8h, v26.8h , v24.8h // temp4 -= temp5 * 5 161 sqrshrun v31.8b, v20.8h, #5 // dst[0_8] = CLIP_U8((temp4 +16) >> 5) 167 uaddl v20.8h, v6.8b, v0.8b 174 mls v18.8h, v20.8h , v24.8h 180 uaddl v20.8h, v8.8b, v2.8b 189 mls v14.8h, v20.8h , v24.8h 196 uaddl v20.8h, v9.8b, v7.8b // temp4 = src[0_8] + src[5_8] 205 mla v20.8h, v18.8h , v22.8h // temp4 += temp3 * 20 [all …]
|
D | ih264_resi_trans_quant_av8.s | 142 sub v20.4h, v15.4h , v16.4h //x2 = x5-x6 145 shl v22.4h, v20.4h, #1 //u_shift(x2,1,shft) 152 add v25.4h, v23.4h , v20.4h //x6 = u_shift(x3,1,shft) + x2; 176 add v20.4s, v0.4s, v23.4s 183 sshl v20.4s, v20.4s, v24.4s //shift row 1 188 xtn v20.4h, v20.4s //narrow row 1 193 neg v24.8h, v20.8h //get negative 199 cmeq v0.4h, v20.4h, #0 204 bsl v4.8b, v20.8b, v24.8b //restore sign of row 1 and 2 344 sub v20.4h, v15.4h , v16.4h //x2 = x5-x6 [all …]
|
/external/libavc/encoder/armv8/ |
D | ih264e_half_pel_av8.s | 174 …sqrshrun v20.8b, v8.8h, #5 //// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column1… 181 st1 {v20.8b, v21.8b}, [x1], #16 ////Store dest row0 307 uaddl v20.8h, v2.8b, v17.8b //// a0 + a5 (column1,row0) 309 umlal v20.8h, v8.8b, v1.8b //// a0 + a5 + 20a2 (column1,row0) 310 umlal v20.8h, v11.8b, v1.8b //// a0 + a5 + 20a2 + 20a3 (column1,row0) 311 umlsl v20.8h, v5.8b, v31.8b //// a0 + a5 + 20a2 + 20a3 - 5a1 (column1,row0) 312 umlsl v20.8h, v14.8b, v31.8b //// a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 (column1,row0) 313 mov v21.d[0], v20.d[1] 320 ext v30.8b, v20.8b , v21.8b , #4 325 ext v29.8b, v20.8b , v21.8b , #6 [all …]
|
D | ih264e_evaluate_intra16x16_modes_av8.s | 135 dup v20.8h, w11 142 uqshl v0.8h, v0.8h, v20.8h 168 dup v20.8b, v9.b[15] ///HORIZONTAL VALUE ROW=0// 178 uabdl v26.8h, v0.8b, v20.8b 189 dup v20.8b, v9.b[14] ///HORIZONTAL VALUE ROW=1// 199 uabal v26.8h, v2.8b, v20.8b 206 dup v20.8b, v9.b[13] ///HORIZONTAL VALUE ROW=2// 215 uabal v26.8h, v4.8b, v20.8b 222 dup v20.8b, v9.b[12] ///HORIZONTAL VALUE ROW=3// 231 uabal v26.8h, v6.8b, v20.8b [all …]
|
/external/libxaac/decoder/armv8/ |
D | ixheaacd_pre_twiddle.s | 201 uMULL v20.4s, v6.4h, v11.4h 206 ushR v20.4s, v20.4s, #16 213 sMLAL v20.4s, v7.4h, v11.4h 222 ADD v20.4s, v20.4s , v18.4s 224 NEG v20.4s, v20.4s 232 sshL v20.4s, v20.4s, v14.4s 264 ST2 { v20.4s, v21.4s}, [x7], x8 283 uMULL v20.4s, v6.4h, v11.4h 291 ushR v20.4s, v20.4s, #16 300 sMLAL v20.4s, v7.4h, v11.4h [all …]
|
/external/llvm/test/MC/SystemZ/ |
D | insn-good-z13.s | 25 #CHECK: vab %v18, %v3, %v20 # encoding: [0xe7,0x23,0x40,0x00,0x0a,0xf3] 31 vab %v18, %v3, %v20 37 #CHECK: vaccb %v18, %v3, %v20 # encoding: [0xe7,0x23,0x40,0x00,0x0a,0xf1] 43 vaccb %v18, %v3, %v20 63 #CHECK: vaccf %v18, %v3, %v20 # encoding: [0xe7,0x23,0x40,0x00,0x2a,0xf1] 69 vaccf %v18, %v3, %v20 75 #CHECK: vaccg %v18, %v3, %v20 # encoding: [0xe7,0x23,0x40,0x00,0x3a,0xf1] 81 vaccg %v18, %v3, %v20 87 #CHECK: vacch %v18, %v3, %v20 # encoding: [0xe7,0x23,0x40,0x00,0x1a,0xf1] 93 vacch %v18, %v3, %v20 [all …]
|
/external/llvm/test/MC/Disassembler/SystemZ/ |
D | insns-z13.txt | 17 #CHECK: vab %v18, %v3, %v20 26 #CHECK: vaccb %v18, %v3, %v20 35 #CHECK: vacccq %v3, %v20, %v5, %v22 44 #CHECK: vaccf %v18, %v3, %v20 53 #CHECK: vaccg %v18, %v3, %v20 62 #CHECK: vacch %v18, %v3, %v20 71 #CHECK: vaccq %v18, %v3, %v20 80 #CHECK: vacq %v3, %v20, %v5, %v22 89 #CHECK: vaf %v18, %v3, %v20 98 #CHECK: vag %v18, %v3, %v20 [all …]
|
/external/boringssl/ios-aarch64/crypto/fipsmodule/ |
D | ghashv8-armx64.S | 37 eor v20.16b,v3.16b,v16.16b //twisted H 38 st1 {v20.2d},[x0],#16 //store Htable[0] 41 ext v16.16b,v20.16b,v20.16b,#8 //Karatsuba pre-processing 42 pmull v0.1q,v20.1d,v20.1d 43 eor v16.16b,v16.16b,v20.16b 44 pmull2 v2.1q,v20.2d,v20.2d 76 ld1 {v20.2d,v21.2d},[x1] //load twisted H, ... 83 pmull v0.1q,v20.1d,v3.1d //H.lo·Xi.lo 85 pmull2 v2.1q,v20.2d,v3.2d //H.hi·Xi.hi 132 ld1 {v20.2d,v21.2d},[x1],#32 //load twisted H, ..., H^2 [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/MC/SystemZ/ |
D | insn-good-z13.s | 454 #CHECK: va %v18, %v3, %v20, 11 # encoding: [0xe7,0x23,0x40,0x00,0xba,0xf3] 461 va %v18, %v3, %v20, 11 467 #CHECK: vab %v18, %v3, %v20 # encoding: [0xe7,0x23,0x40,0x00,0x0a,0xf3] 473 vab %v18, %v3, %v20 496 #CHECK: vacc %v18, %v3, %v20, 11 # encoding: [0xe7,0x23,0x40,0x00,0xba,0xf1] 503 vacc %v18, %v3, %v20, 11 509 #CHECK: vaccb %v18, %v3, %v20 # encoding: [0xe7,0x23,0x40,0x00,0x0a,0xf1] 515 vaccb %v18, %v3, %v20 551 #CHECK: vaccf %v18, %v3, %v20 # encoding: [0xe7,0x23,0x40,0x00,0x2a,0xf1] 557 vaccf %v18, %v3, %v20 [all …]
|
/external/boringssl/linux-aarch64/crypto/fipsmodule/ |
D | ghashv8-armx64.S | 38 eor v20.16b,v3.16b,v16.16b //twisted H 39 st1 {v20.2d},[x0],#16 //store Htable[0] 42 ext v16.16b,v20.16b,v20.16b,#8 //Karatsuba pre-processing 43 pmull v0.1q,v20.1d,v20.1d 44 eor v16.16b,v16.16b,v20.16b 45 pmull2 v2.1q,v20.2d,v20.2d 77 ld1 {v20.2d,v21.2d},[x1] //load twisted H, ... 84 pmull v0.1q,v20.1d,v3.1d //H.lo·Xi.lo 86 pmull2 v2.1q,v20.2d,v3.2d //H.hi·Xi.hi 133 ld1 {v20.2d,v21.2d},[x1],#32 //load twisted H, ..., H^2 [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/MC/Disassembler/SystemZ/ |
D | insns-z13.txt | 545 # CHECK: va %v18, %v3, %v20, 11 554 # CHECK: vab %v18, %v3, %v20 563 # CHECK: vac %v3, %v20, %v5, %v22, 11 572 # CHECK: vacc %v18, %v3, %v20, 11 581 # CHECK: vaccb %v18, %v3, %v20 590 # CHECK: vaccc %v3, %v20, %v5, %v22, 11 599 # CHECK: vacccq %v3, %v20, %v5, %v22 608 # CHECK: vaccf %v18, %v3, %v20 617 # CHECK: vaccg %v18, %v3, %v20 626 # CHECK: vacch %v18, %v3, %v20 [all …]
|