Lines Matching refs:sn

388 void Arm32Assembler::vadds(SRegister sd, SRegister sn, SRegister sm,  in vadds()  argument
390 EmitVFPsss(cond, B21 | B20, sd, sn, sm); in vadds()
400 void Arm32Assembler::vsubs(SRegister sd, SRegister sn, SRegister sm, in vsubs() argument
402 EmitVFPsss(cond, B21 | B20 | B6, sd, sn, sm); in vsubs()
412 void Arm32Assembler::vmuls(SRegister sd, SRegister sn, SRegister sm, in vmuls() argument
414 EmitVFPsss(cond, B21, sd, sn, sm); in vmuls()
424 void Arm32Assembler::vmlas(SRegister sd, SRegister sn, SRegister sm, in vmlas() argument
426 EmitVFPsss(cond, 0, sd, sn, sm); in vmlas()
436 void Arm32Assembler::vmlss(SRegister sd, SRegister sn, SRegister sm, in vmlss() argument
438 EmitVFPsss(cond, B6, sd, sn, sm); in vmlss()
448 void Arm32Assembler::vdivs(SRegister sd, SRegister sn, SRegister sm, in vdivs() argument
450 EmitVFPsss(cond, B23, sd, sn, sm); in vdivs()
873 void Arm32Assembler::vmovsr(SRegister sn, Register rt, Condition cond) { in vmovsr() argument
874 CHECK_NE(sn, kNoSRegister); in vmovsr()
881 ((static_cast<int32_t>(sn) >> 1)*B16) | in vmovsr()
883 ((static_cast<int32_t>(sn) & 1)*B7) | B4; in vmovsr()
888 void Arm32Assembler::vmovrs(Register rt, SRegister sn, Condition cond) { in vmovrs() argument
889 CHECK_NE(sn, kNoSRegister); in vmovrs()
896 ((static_cast<int32_t>(sn) >> 1)*B16) | in vmovrs()
898 ((static_cast<int32_t>(sn) & 1)*B7) | B4; in vmovrs()
1088 SRegister sd, SRegister sn, SRegister sm) { in EmitVFPsss() argument
1090 CHECK_NE(sn, kNoSRegister); in EmitVFPsss()
1096 ((static_cast<int32_t>(sn) >> 1)*B16) | in EmitVFPsss()
1098 ((static_cast<int32_t>(sn) & 1)*B7) | in EmitVFPsss()