/external/tensorflow/tensorflow/python/keras/layers/ |
D | normalization_test.py | 177 fused=[True, False])) 179 def test_batchnorm_mixed_precision_does_not_overflow(self, fused): argument 183 fused=fused, 253 self.assertTrue(layer.fused) 263 self.assertFalse(layer.fused) 277 self.assertEqual(norm.fused, True) 279 norm = normalization.BatchNormalization(fused=False) 280 self.assertEqual(norm.fused, False) 283 self.assertEqual(norm.fused, False) 286 self.assertEqual(norm.fused, True) [all …]
|
D | normalization.py | 183 fused=None, argument 214 if fused: 218 elif fused is None and not self._fused_can_be_used(): 219 fused = False 220 elif fused is None: 221 fused = True 224 self.fused = fused 336 if self.fused in (None, True): 343 if self.fused is None: 344 self.fused = ndims == 4 [all …]
|
/external/tensorflow/tensorflow/compiler/xla/service/ |
D | multi_output_fusion.cc | 141 HloInstruction* fused = instr2; in Fuse() local 144 if (fused->opcode() == HloOpcode::kFusion) { in Fuse() 145 std::swap(remaining, fused); in Fuse() 147 if (fused->IsMultiOutputFusion()) { in Fuse() 148 std::swap(remaining, fused); in Fuse() 150 if (fused->opcode() == HloOpcode::kFusion) { in Fuse() 151 remaining->MergeFusionInstructionIntoMultiOutput(fused); in Fuse() 153 remaining->FuseInstructionIntoMultiOutput(fused); in Fuse() 154 CHECK_EQ(0, fused->user_count()); in Fuse() 155 TF_CHECK_OK(computation()->RemoveInstruction(fused)); in Fuse() [all …]
|
D | logical_buffer_analysis.cc | 36 for (auto* fused : instruction->fused_instructions()) { in GatherFusionInstructions() local 37 if (fused->opcode() == HloOpcode::kFusion) { in GatherFusionInstructions() 38 GatherFusionInstructions(fused, fusion_instructions); in GatherFusionInstructions()
|
/external/tensorflow/tensorflow/lite/g3doc/convert/ |
D | operation_fusion.md | 6 in TensorFlow to fused operations in TensorFlow Lite. This infrastructure is 8 to a corresponding fused operation in TensorFlow Lite. 14 ### What are fused operations 27 A fused operation corresponds to a single operation that subsumes all the 31 ### Benefits of fused operations 42 TensorFlow Lite has many instances of fused operations for the reasons 43 articulated above. These fused operations typically correspond to composite 45 TensorFlow that are implemented as a single fused operation in TensorFlow Lite 49 implemented in the fused LSTM operations. 51 ### Challenges with fused operations [all …]
|
/external/tensorflow/tensorflow/python/distribute/ |
D | zero_batch_test.py | 47 fused=[True, False])) 48 def testBNWithZeroBatchInputGraph(self, distribution, fused): argument 61 axis=3, epsilon=1e-3, momentum=0.9, fused=fused) 118 fused=[True, False])) 119 def testBNWithZeroBatchInput(self, distribution, fused): argument 125 axis=3, epsilon=1e-3, momentum=0.9, fused=fused) 168 fused=[True, False])) 169 def testBNWithDynamicBatchInputEager(self, distribution, fused): argument 182 axis=-1, epsilon=1e-3, momentum=0.9, fused=fused)
|
/external/llvm-project/mlir/test/IR/ |
D | invalid-locations.mlir | 77 return loc(fused<true [unknown]) // expected-error {{expected '>' after fused location metadata}} 85 return loc(fused<) // expected-error {{expected valid attribute metadata}} 92 return loc(fused<true>unknown]) // expected-error {{expected '[' in fused location}} 99 return loc(fused[unknown) // expected-error {{expected ']' in fused location}}
|
D | locations.mlir | 15 // CHECK: } loc(fused["foo", "mysource.cc":10:8]) 17 } loc(fused["foo", "mysource.cc":10:8]) 19 // CHECK: } loc(fused<"myPass">["foo", "foo2"]) 21 } loc(fused<"myPass">["foo", "foo2"])
|
D | diagnostic-handler.mlir | 6 // Emit the first available call stack in the fused location. 11 …%x = "std.constant"() {value = 100} : () -> i1 loc(fused["bar", callsite("foo"("mysource1":0:0) at…
|
D | pretty-locations.mlir | 21 } loc(fused["foo", "mysource.cc":10:8]) 25 } loc(fused<"myPass">["foo", "foo2"])
|
D | opaque_locations.mlir | 27 } loc(fused["foo", "mysource.cc":10:8]) 31 } loc(fused["foo", "mysource.cc":10:8, callsite("foo" at "mysource.cc":10:8)])
|
/external/llvm-project/llvm/test/MC/X86/ |
D | align-branch-fused.s | 1 # RUN: llvm-mc -filetype=obj -triple x86_64 --x86-align-branch-boundary=32 --x86-align-branch=fused… 3 # Exercise cases where fused instructions need to be aligned. 12 # 'cmp %rax, %rbp' is macro fused with 'jne foo', 24 # 'cmp %rax, %rbp' is fusible but can not fused with `jo foo`, 36 # The second 'cmp %rax, %rbp' is macro fused with 'jne foo'.
|
D | align-branch-bundle.s | 1 … x86_64-unknown-unknown --x86-align-branch-boundary=16 --x86-align-branch=fused+jcc --mc-relax-all… 3 # Check using option --x86-align-branch-boundary=16 --x86-align-branch=fused+jcc --mc-relax-all wit…
|
D | align-branch-align.s | 1 # RUN: llvm-mc -filetype=obj -triple x86_64 --x86-align-branch-boundary=32 --x86-align-branch=fused… 6 # Check the align directive between the macro fused pair
|
D | align-branch-general.s | 1 # RUN: llvm-mc -filetype=obj -triple x86_64 --x86-align-branch-boundary=32 --x86-align-branch=fused… 6 # to the fined options --x86-align-branch-boundary=32 --x86-align-branch=fused+jcc+jmp.
|
D | align-branch-relax-all.s | 1 # RUN: llvm-mc -filetype=obj -triple x86_64 --x86-align-branch-boundary=32 --x86-align-branch=fused… 2 # RUN: llvm-mc -filetype=obj -triple x86_64 --x86-align-branch-boundary=32 --x86-align-branch=fused…
|
/external/tensorflow/tensorflow/python/keras/legacy_tf_layers/ |
D | normalization.py | 141 fused=None, argument 164 fused=fused, 197 fused=None, argument 331 fused=fused,
|
/external/tensorflow/tensorflow/compiler/xla/service/gpu/ |
D | horizontal_input_fusion.cc | 129 HloInstruction* fused = candidates[j]; in Run() local 130 if (ShapesCompatibleForMultiOutputFusion(*fusion_anchor, *fused) && in Run() 131 !FusionWouldBeTooLarge(*fusion_anchor, *fused)) { in Run() 132 VLOG(3) << "Fuse " << fused->ToString() << " into " in Run() 134 fusion_anchor->MergeFusionInstructionIntoMultiOutput(fused); in Run()
|
D | multi_output_fusion.cc | 229 HloInstruction* fused = *j; in FuseSiblings() local 230 if (fused->opcode() == HloOpcode::kFusion) { in FuseSiblings() 231 remaining->MergeFusionInstructionIntoMultiOutput(fused); in FuseSiblings() 233 remaining->FuseInstructionIntoMultiOutput(fused); in FuseSiblings() 234 CHECK_EQ(0, fused->user_count()); in FuseSiblings() 235 TF_CHECK_OK(computation_->RemoveInstruction(fused)); in FuseSiblings()
|
/external/llvm-project/mlir/test/Transforms/ |
D | location-snapshot.mlir | 10 // TAG-NEXT: loc(fused["original", "tagged"("[[FILE]]":{{[0-9]+}}:{{[0-9]+}})]) 11 // TAG-NEXT: loc(fused["original", "tagged"("[[FILE]]":{{[0-9]+}}:{{[0-9]+}})]) 12 // TAG-NEXT: } loc(fused["original", "tagged"("[[FILE]]":{{[0-9]+}}:{{[0-9]+}})])
|
D | strip-debuginfo.mlir | 13 } loc(fused["foo", "mysource.cc":10:8]) 18 } loc(fused<"myPass">["foo", "foo2"])
|
/external/tensorflow/tensorflow/compiler/mlir/lite/quantization/tests/ |
D | import_quant_stats.mlir | 7 loc(fused["skip1", "skip2.cc":10:8, callsite("op" at "skip3.cc":10:8)]) 17 loc(fused["skip1.cc":10:8, "op", callsite("skip2" at "skip3.cc":10:8)]) 29 loc(fused["skip1.cc":10:8, "op_0", callsite("skip2" at "skip3.cc":10:8)]) 40 loc(fused["skip1.cc":10:8, "op_regex", callsite("skip2" at "skip3.cc":10:8)])
|
/external/tensorflow/tensorflow/python/keras/benchmarks/ |
D | eager_microbenchmarks_test.py | 182 ("batch_norm_fused_inf", tf.keras.layers.BatchNormalization(fused=True), 184 ("batch_norm_fused_train", tf.keras.layers.BatchNormalization(fused=True), 187 tf.keras.layers.BatchNormalization(fused=False), (1, 1, 1, 1)), 189 tf.keras.layers.BatchNormalization(fused=False), (1, 1, 1, 1),
|
/external/tensorflow/tensorflow/compiler/mlir/hlo/tests/ |
D | mhlo-fusion.mlir | 56 // relationship. Thus Reduce Op should not be fused with above two ops. 65 // Above two ops should not be fused since reduce op can not be 66 // fused with its consumer. 92 // Following op should not be fused with the above ops since reduce op can not be 93 // fused with its consumer.
|
/external/tensorflow/tensorflow/compiler/mlir/hlo/include/mlir-hlo/Dialect/mhlo/IR/ |
D | infer_fusibility_op_interface.td | 34 /*desc=*/[{If true, this op can be fused with its operands 41 /// Returns whether this op can be fused with its operands 46 /*desc=*/[{If true, this op can be fused with its consumers 53 /// Return whether this op can be fused with its consumers
|