/external/v8/test/mjsunit/es6/ |
D | iteration-semantics.js | 80 function fold(cons, seed, iterable) { function 176 assertEquals(45, fold(sum, 0, integers_until(10))); 178 assertEquals([1, 2, 3], fold(append, [], values(1, 2, 3))); 180 assertEquals(45, fold(sum, 0, take(integers_from(0), 10))); 182 assertEquals(90, fold(sum, 0, take(skip_every(integers_from(0), 2), 10))); 192 fold(append, [], 205 fold(append, [], 210 fold(append, [], 230 fold(sum, 0, remove_next_after(integers_until(10), 10))); 233 fold(sum, 0, poison_next_after(integers_until(10), 10))); [all …]
|
/external/llvm/test/Analysis/ScalarEvolution/ |
D | latch-dominating-conditions.ll | 41 %does.not.fold.to.true = icmp slt i8 %idx, 42 42 ; CHECK: call void @side_effect(i1 %does.not.fold.to.true) 43 call void @side_effect(i1 %does.not.fold.to.true) 48 ; CHECK: call void @side_effect(i1 %does.not.fold.to.true) 49 call void @side_effect(i1 %does.not.fold.to.true)
|
/external/opencv3/3rdparty/jinja2/ |
D | optimizer.py | 54 def fold(self, node): member in Optimizer 67 visit_Filter = visit_Test = visit_CondExpr = fold 68 del fold
|
/external/llvm/test/Transforms/CodeGenPrepare/ |
D | basic.ll | 7 ; objectsize should fold to a constant, which causes the branch to fold to an 8 ; uncond branch. Next, we fold the control flow alltogether.
|
/external/llvm/test/Assembler/ |
D | ConstantExprFoldCast.ll | 1 ; This test checks to make sure that constant exprs fold in some simple situations 6 @A = global i32* bitcast (i8* null to i32*) ; Cast null -> fold 7 @B = global i32** bitcast (i32** @A to i32**) ; Cast to same type -> fold
|
D | ConstantExprNoFold.ll | 1 ; This test checks to make sure that constant exprs don't fold in some simple 13 ; Don't fold this. @A might really be allocated next to @B, in which case the 20 ; Don't fold this completely away either. In theory this could be simplified
|
/external/llvm/test/CodeGen/AMDGPU/ |
D | operand-folding.ll | 5 define void @fold_sgpr(i32 addrspace(1)* %out, i32 %fold) { 7 %tmp0 = icmp ne i32 %fold, 0 12 %offset = add i32 %fold, %id 25 %fold = add i32 3, 2 31 %val = or i32 %id, %fold
|
/external/llvm/test/CodeGen/AArch64/ |
D | fast-isel-folding.ll | 3 ; Test that we don't fold the shift. 17 ; Test that we don't fold the sign-extend. 31 ; Test that we don't fold the sign-extend.
|
/external/llvm/test/CodeGen/X86/ |
D | pr2656.ll | 9 ; We can not fold either stack load into an 'xor' instruction because that 11 ; We can fold the 16-byte constant load into either 'xor' instruction, 37 ; We can not fold the load from the stack into the 'and' instruction because 39 ; We can fold the load of the constant because it is a 16-byte vector constant.
|
D | 2005-01-17-CycleInDAG.ll | 1 ; This testcase was distilled from 132.ijpeg. Bsaically we cannot fold the 4 ; that we do not fold the load into the sub.
|
D | dagcombine-shifts.ll | 3 ; fold (shl (zext (lshr (A, X))), X) -> (zext (shl (lshr (A, X)), X)) 7 ; This will help DAGCombiner to identify and then fold the sequence 96 ; Don't fold the pattern if we use arithmetic shifts. 137 ; Don't fold the pattern if there is more than one use of the 190 ; Verify also that we correctly fold the shl-shr sequence into an
|
D | and-or-fold.ll | 4 ; The dag combiner should fold together (x&127)|(y&16711680) -> (x|y)&c1 17 ; <rdar://problem/7529774> The optimizer shouldn't fold this into (and (or, C), D)
|
D | lower-bitcast.ll | 11 ; FIXME: Ideally we should be able to fold the entire body of @test1 into a 72 ; Ideally, we should fold that sequence into a single paddd. This is fixed with 109 ; FIXME: Ideally we should be able to fold the entire body of @test6 into a 152 ; FIXME: Ideally we should be able to fold the entire body of @test8 into a
|
D | narrow-shl-load.ll | 6 ; DAGCombiner should fold this code in finite time. 33 ; DAG Combiner can't fold this into a load of the 1'th byte.
|
D | logical-load-fold.ll | 4 ; Although we have the ability to fold an unaligned load with AVX 6 ; can not fold the load under any circumstances in these test
|
/external/opencv/ml/src/ |
D | mlestimate.cpp | 108 int k, fold; 115 fold = ++crVal->current_fold; 117 if (fold >= crVal->folds_all) 119 if (fold == crVal->folds_all) 127 k = crVal->folds[fold + 1] - crVal->folds[fold]; 128 crVal->sampleIdxTrain->data.i = crVal->sampleIdxAll + crVal->folds[fold + 1]; 130 crVal->sampleIdxEval->data.i = crVal->sampleIdxAll + crVal->folds[fold];
|
/external/llvm/test/Transforms/InstCombine/ |
D | fold-calls.ll | 3 ; This shouldn't fold, because sin(inf) is invalid. 11 ; This should fold.
|
D | 2004-02-23-ShiftShiftOverflow.ll | 4 ; Do not fold into shr X, 34, as this uses undefined behavior! 11 ; Do not fold into shl X, 34, as this uses undefined behavior!
|
D | fold-sqrt-sqrtf.ll | 4 ; Instcombine tries to fold (fptrunc (sqrt (fpext x))) -> (sqrtf x), but this 5 ; shouldn't fold when sqrtf isn't available.
|
D | sitofp.ll | 115 ; This can't fold because the 25-bit input doesn't fit in the mantissa. 145 ; This can fold because the 25-bit input is signed and we disard the sign bit. 154 ; This can't fold because the 26-bit input won't fit the mantissa 165 ; This can fold because the 54-bit output is signed and we disard the sign bit. 174 ; This can't fold because the 55-bit output won't fit the mantissa
|
/external/llvm/test/Analysis/GlobalsModRef/ |
D | nonescaping-noalias.ll | 11 ; Ensure that we can fold a store to a load of a global across a store to 28 ; Ensure that we can fold a store to a load of a global across a store to 47 ; Ensure that we can fold a store to a load of a global across a store to 68 ; Ensure that we can fold a store to a load of a global across a store to 102 ; Ensure that we can fold a store to a load of a global across a store to
|
/external/llvm/test/CodeGen/ARM/ |
D | 2012-03-26-FoldImmBug.ll | 5 ; fold the immediate by breaking it into two parts and fold them into the
|
/external/llvm/test/Transforms/LoopStrengthReduce/X86/ |
D | no_superflous_induction_vars.ll | 33 %fold = add i64 %1, 1 34 %n.mod.vf = and i64 %fold, 7
|
/external/llvm/test/Transforms/IndVarSimplify/ |
D | iv-fold.ll | 5 ; Indvars should be able to fold IV increments into shr when low bits are zero. 31 ; Invdars should not fold an increment into shr unless 2^shiftBits is
|
/external/llvm/test/Transforms/InstSimplify/ |
D | fast-math.ll | 55 ; Dont fold this 79 ; Dont fold 97 ; Dont fold
|