/external/libjpeg-turbo/ |
D | jidctint.c | 178 JLONG tmp10, tmp11, tmp12, tmp13; variable 246 tmp13 = tmp0 - tmp3; 290 wsptr[DCTSIZE * 3] = (int)DESCALE(tmp13 + tmp0, CONST_BITS - PASS1_BITS); 291 wsptr[DCTSIZE * 4] = (int)DESCALE(tmp13 - tmp0, CONST_BITS - PASS1_BITS); 348 tmp13 = tmp0 - tmp3; 404 outptr[3] = range_limit[(int)DESCALE(tmp13 + tmp0, 407 outptr[4] = range_limit[(int)DESCALE(tmp13 - tmp0, 431 JLONG tmp0, tmp1, tmp2, tmp10, tmp11, tmp12, tmp13; in jpeg_idct_7x7() local 450 tmp13 = DEQUANTIZE(inptr[DCTSIZE * 0], quantptr[DCTSIZE * 0]); in jpeg_idct_7x7() 451 tmp13 = LEFT_SHIFT(tmp13, CONST_BITS); in jpeg_idct_7x7() [all …]
|
D | jfdctflt.c | 63 FAST_FLOAT tmp10, tmp11, tmp12, tmp13; variable 84 tmp13 = tmp0 - tmp3; 91 z1 = (tmp12 + tmp13) * ((FAST_FLOAT)0.707106781); /* c4 */ 92 dataptr[2] = tmp13 + z1; /* phase 5 */ 93 dataptr[6] = tmp13 - z1; 134 tmp13 = tmp0 - tmp3; 141 z1 = (tmp12 + tmp13) * ((FAST_FLOAT)0.707106781); /* c4 */ 142 dataptr[DCTSIZE * 2] = tmp13 + z1; /* phase 5 */ 143 dataptr[DCTSIZE * 6] = tmp13 - z1;
|
D | jfdctfst.c | 120 DCTELEM tmp10, tmp11, tmp12, tmp13; variable 142 tmp13 = tmp0 - tmp3; 149 z1 = MULTIPLY(tmp12 + tmp13, FIX_0_707106781); /* c4 */ 150 dataptr[2] = tmp13 + z1; /* phase 5 */ 151 dataptr[6] = tmp13 - z1; 192 tmp13 = tmp0 - tmp3; 199 z1 = MULTIPLY(tmp12 + tmp13, FIX_0_707106781); /* c4 */ 200 dataptr[DCTSIZE * 2] = tmp13 + z1; /* phase 5 */ 201 dataptr[DCTSIZE * 6] = tmp13 - z1;
|
D | jidctflt.c | 77 FAST_FLOAT tmp10, tmp11, tmp12, tmp13; variable 136 tmp13 = tmp1 + tmp3; /* phases 5-3 */ 137 tmp12 = (tmp1 - tmp3) * ((FAST_FLOAT)1.414213562) - tmp13; /* 2*c4 */ 139 tmp0 = tmp10 + tmp13; /* phase 2 */ 140 tmp3 = tmp10 - tmp13; 199 tmp13 = wsptr[2] + wsptr[6]; 200 tmp12 = (wsptr[2] - wsptr[6]) * ((FAST_FLOAT)1.414213562) - tmp13; 202 tmp0 = tmp10 + tmp13; 203 tmp3 = tmp10 - tmp13;
|
D | jidctfst.c | 176 DCTELEM tmp10, tmp11, tmp12, tmp13; variable 235 tmp13 = tmp1 + tmp3; /* phases 5-3 */ 236 tmp12 = MULTIPLY(tmp1 - tmp3, FIX_1_414213562) - tmp13; /* 2*c4 */ 238 tmp0 = tmp10 + tmp13; /* phase 2 */ 239 tmp3 = tmp10 - tmp13; 321 tmp13 = ((DCTELEM)wsptr[2] + (DCTELEM)wsptr[6]); 323 MULTIPLY((DCTELEM)wsptr[2] - (DCTELEM)wsptr[6], FIX_1_414213562) - tmp13; 325 tmp0 = tmp10 + tmp13; 326 tmp3 = tmp10 - tmp13;
|
D | jfdctint.c | 146 JLONG tmp10, tmp11, tmp12, tmp13; variable 172 tmp13 = tmp0 - tmp3; 179 z1 = MULTIPLY(tmp12 + tmp13, FIX_0_541196100); 180 dataptr[2] = (DCTELEM)DESCALE(z1 + MULTIPLY(tmp13, FIX_0_765366865), 237 tmp13 = tmp0 - tmp3; 244 z1 = MULTIPLY(tmp12 + tmp13, FIX_0_541196100); 246 (DCTELEM)DESCALE(z1 + MULTIPLY(tmp13, FIX_0_765366865),
|
/external/llvm-project/llvm/test/CodeGen/AMDGPU/ |
D | image-schedule.ll | 25 %tmp13.0 = extractelement <3 x i32> %tmp9, i32 0 26 %tmp13.1 = extractelement <3 x i32> %tmp9, i32 1 27 …%tmp14 = call <4 x float> @llvm.amdgcn.image.load.2d.v4f32.i32(i32 15, i32 %tmp13.0, i32 %tmp13.1,… 30 …m.amdgcn.image.store.2d.v4f32.i32(<4 x float> %tmp14, i32 15, i32 %tmp13.0, i32 %tmp13.1, <8 x i32… 32 …%tmp18 = call <4 x float> @llvm.amdgcn.image.load.2d.v4f32.i32(i32 165, i32 %tmp13.0, i32 %tmp13.1… 36 …m.amdgcn.image.store.2d.v4f32.i32(<4 x float> %tmp18, i32 15, i32 %tmp13.0, i32 %tmp13.1, <8 x i32…
|
D | global_smrd_cfg.ll | 50 %tmp13 = phi i32 [ %tmp25, %bb22 ], [ 0, %bb6 ] 51 %tmp14 = srem i32 %tmp13, %arg2 56 %tmp19 = sext i32 %tmp13 to i64 67 %tmp25 = add nuw nsw i32 %tmp13, 1 96 %tmp13 = phi i32 [ %tmp25, %bb1 ], [ 0, %bb ] 97 %tmp14 = srem i32 %tmp13, %arg2 101 %tmp19 = sext i32 %tmp13 to i64 104 %tmp25 = add nuw nsw i32 %tmp13, 1
|
/external/llvm-project/llvm/test/Transforms/LowerExpectIntrinsic/ |
D | phi_merge.ll | 22 %tmp13 = icmp sgt i32 %arg1, %tmp12 23 br i1 %tmp13, label %bb14, label %bb18 24 ; CHECK: br i1 %tmp13, {{.*}}!prof [[WEIGHT]] 65 %tmp13 = icmp sgt i32 %arg1, %tmp12 66 br i1 %tmp13, label %bb14, label %bb18 67 ; CHECK: br i1 %tmp13 108 %tmp13 = icmp sgt i32 %arg1, %tmp12 109 br i1 %tmp13, label %bb14, label %bb18 110 ; CHECK: br i1 %tmp13, {{.*}}!prof [[WEIGHT]] 149 %tmp13 = icmp sgt i32 %arg1, %tmp12 [all …]
|
/external/llvm/test/CodeGen/ARM/ |
D | ifcvt2.ll | 9 %tmp13 = add i32 %b, %a 13 %tmp15 = add i32 %tmp13, %c 18 ret i32 %tmp13 32 %tmp13 = add i32 %b, %a 36 %tmp15 = add i32 %tmp13, %c 41 ret i32 %tmp13
|
D | lsr-code-insertion.ll | 30 %tmp13 = getelementptr i32, i32* %mc, i32 %k.069.0 ; <i32*> [#uses=5] 31 store i32 %tmp10, i32* %tmp13 41 store i32 %tmp24, i32* %tmp13 45 %tmp39 = load i32, i32* %tmp13 ; <i32> [#uses=1] 49 store i32 %tmp44, i32* %tmp13 54 store i32 -987654321, i32* %tmp13
|
/external/libyuv/files/source/ |
D | rotate_mmi.cc | 28 uint64_t tmp7, tmp8, tmp9, tmp10, tmp11, tmp12, tmp13; in TransposeWx8_MMI() local 142 [tmp12] "=&f"(tmp12), [tmp13] "=&f"(tmp13), [dst] "+&r"(dst), in TransposeWx8_MMI() 157 uint64_t tmp7, tmp8, tmp9, tmp10, tmp11, tmp12, tmp13; in TransposeUVWx8_MMI() local 279 [tmp12] "=&f"(tmp12), [tmp13] "=&f"(tmp13), [dst_a] "+&r"(dst_a), in TransposeUVWx8_MMI()
|
/external/llvm-project/llvm/test/CodeGen/ARM/ |
D | ifcvt2.ll | 9 %tmp13 = add i32 %b, %a 13 %tmp15 = add i32 %tmp13, %c 18 ret i32 %tmp13 32 %tmp13 = add i32 %b, %a 36 %tmp15 = add i32 %tmp13, %c 41 ret i32 %tmp13
|
D | lsr-code-insertion.ll | 30 %tmp13 = getelementptr i32, i32* %mc, i32 %k.069.0 ; <i32*> [#uses=5] 31 store i32 %tmp10, i32* %tmp13 41 store i32 %tmp24, i32* %tmp13 45 %tmp39 = load i32, i32* %tmp13 ; <i32> [#uses=1] 49 store i32 %tmp44, i32* %tmp13 54 store i32 -987654321, i32* %tmp13
|
/external/llvm/test/CodeGen/X86/ |
D | 2006-08-07-CycleInDAG.ll | 12 %tmp13.i12.i = tail call double @ldexp( double 0.000000e+00, i32 0 ) ; <double> [#uses=1] 13 %tmp13.i13.i = fptrunc double %tmp13.i12.i to float ; <float> [#uses=1] 17 %tmp13.i7 = mul i32 %tmp11.i, %n.i ; <i32> [#uses=1] 18 %tmp.i8 = tail call i8* @calloc( i32 %tmp13.i7, i32 4 ) ; <i8*> [#uses=0] 25 %tmp165.i = fpext float %tmp13.i13.i to double ; <double> [#uses=0]
|
D | vec_logical.ll | 36 %tmp13 = bitcast <4 x i32> %tmp11 to <4 x float> 37 ret <4 x float> %tmp13 54 %tmp13 = bitcast <2 x i64> %tmp11 to <2 x double> 55 ret <2 x double> %tmp13 80 %tmp13 = xor <4 x i32> %tmp11, < i32 -1, i32 -1, i32 -1, i32 -1 > 81 %tmp14 = and <4 x i32> %tmp12, %tmp13
|
/external/llvm-project/llvm/test/CodeGen/X86/ |
D | 2006-08-07-CycleInDAG.ll | 12 %tmp13.i12.i = tail call double @ldexp( double 0.000000e+00, i32 0 ) ; <double> [#uses=1] 13 %tmp13.i13.i = fptrunc double %tmp13.i12.i to float ; <float> [#uses=1] 17 %tmp13.i7 = mul i32 %tmp11.i, %n.i ; <i32> [#uses=1] 18 %tmp.i8 = tail call i8* @calloc( i32 %tmp13.i7, i32 4 ) ; <i8*> [#uses=0] 25 %tmp165.i = fpext float %tmp13.i13.i to double ; <double> [#uses=0]
|
/external/llvm/test/Transforms/NaryReassociate/ |
D | pr24301.ll | 7 %tmp13 = add i32 %tmp4, -128 ; deleted 8 %tmp14 = add i32 %tmp13, 8 ; => %tmp5 + -128 10 ; do not rewrite %tmp23 against %tmp13 because %tmp13 is already deleted
|
/external/libjpeg-turbo/simd/arm/ |
D | jfdctfst-neon.c | 100 int16x8_t tmp13 = vsubq_s16(tmp0, tmp3); in jsimd_fdct_ifast_neon() local 107 int16x8_t z1 = vqdmulhq_lane_s16(vaddq_s16(tmp12, tmp13), consts, 2); in jsimd_fdct_ifast_neon() 108 col2 = vaddq_s16(tmp13, z1); /* phase 5 */ in jsimd_fdct_ifast_neon() 109 col6 = vsubq_s16(tmp13, z1); in jsimd_fdct_ifast_neon() 174 tmp13 = vsubq_s16(tmp0, tmp3); in jsimd_fdct_ifast_neon() 181 z1 = vqdmulhq_lane_s16(vaddq_s16(tmp12, tmp13), consts, 2); in jsimd_fdct_ifast_neon() 182 row2 = vaddq_s16(tmp13, z1); /* phase 5 */ in jsimd_fdct_ifast_neon() 183 row6 = vsubq_s16(tmp13, z1); in jsimd_fdct_ifast_neon()
|
D | jidctfst-neon.c | 135 int16x4_t tmp13 = vadd_s16(tmp1, tmp3); /* phases 5-3 */ in jsimd_idct_ifast_neon() local 139 tmp12 = vsub_s16(tmp12, tmp13); in jsimd_idct_ifast_neon() 141 tmp0 = vadd_s16(tmp10, tmp13); /* phase 2 */ in jsimd_idct_ifast_neon() 142 tmp3 = vsub_s16(tmp10, tmp13); in jsimd_idct_ifast_neon() 210 int16x4_t tmp13 = vadd_s16(tmp1, tmp3); /* phases 5-3 */ in jsimd_idct_ifast_neon() local 214 tmp12 = vsub_s16(tmp12, tmp13); in jsimd_idct_ifast_neon() 216 tmp0 = vadd_s16(tmp10, tmp13); /* phase 2 */ in jsimd_idct_ifast_neon() 217 tmp3 = vsub_s16(tmp10, tmp13); in jsimd_idct_ifast_neon() 280 int16x8_t tmp13 = vaddq_s16(tmp1, tmp3); /* phases 5-3 */ in jsimd_idct_ifast_neon() local 284 tmp12 = vsubq_s16(tmp12, tmp13); in jsimd_idct_ifast_neon() [all …]
|
/external/llvm/test/CodeGen/PowerPC/ |
D | ppcf128-3.ll | 7 %tmp13 = call i32 (...) @__inline_isfinite( ppc_fp128 %tmp1112 ) nounwind ; <i32> [#uses=0] 14 %tmp13 = call i32 (...) @__inline_isfinite( ppc_fp128 %tmp1112 ) nounwind ; <i32> [#uses=0] 21 %tmp13 = call i32 (...) @__inline_isfinite( ppc_fp128 %tmp1112 ) nounwind ; <i32> [#uses=0] 28 %tmp13 = call i32 (...) @__inline_isfinite( ppc_fp128 %tmp1112 ) nounwind ; <i32> [#uses=0]
|
/external/llvm-project/llvm/test/CodeGen/PowerPC/ |
D | ppcf128-3.ll | 7 %tmp13 = call i32 (...) @__inline_isfinite( ppc_fp128 %tmp1112 ) nounwind ; <i32> [#uses=0] 14 %tmp13 = call i32 (...) @__inline_isfinite( ppc_fp128 %tmp1112 ) nounwind ; <i32> [#uses=0] 21 %tmp13 = call i32 (...) @__inline_isfinite( ppc_fp128 %tmp1112 ) nounwind ; <i32> [#uses=0] 28 %tmp13 = call i32 (...) @__inline_isfinite( ppc_fp128 %tmp1112 ) nounwind ; <i32> [#uses=0]
|
/external/llvm-project/llvm/test/Transforms/IndVarSimplify/X86/ |
D | deterministic-scev-verify.ll | 13 ; In this test, adding -verify-indvars causes %tmp13 to not be optimized away. 26 %tmp8 = add nuw nsw i32 %tmp13, 1 31 %tmp13 = phi i32 [ 2, %tmp0 ], [ %tmp8, %tmp7 ] 32 %tmp14 = phi i32 [ 1, %tmp0 ], [ %tmp13, %tmp7 ]
|
/external/llvm-project/polly/test/ScopInfo/ |
D | multidim_fortran_2d.ll | 43 %tmp13 = phi i64 [ %tmp30, %bb28 ], [ 1, %bb11 ] 53 %tmp19 = mul i64 %tmp13, %tmp7 56 %tmp22 = add i64 %tmp18, %tmp13 68 %tmp29 = icmp eq i64 %tmp13, %tmp9 69 %tmp30 = add i64 %tmp13, 1
|
/external/llvm-project/llvm/test/Transforms/LoopIdiom/ |
D | int_sideeffect.ll | 14 %i.02 = phi i64 [ 0, %bb7.lr.ph ], [ %tmp13, %bb7 ] 17 %tmp13 = add i64 %i.02, 1 18 %tmp6 = icmp ult i64 %tmp13, %n
|