/external/tensorflow/tensorflow/lite/toco/graph_transformations/ |
D | fuse_activation_functions.cc | 36 ac_op->type != OperatorType::kRelu1 && in Run() 94 } else if (ac_op->type == OperatorType::kRelu1) { in Run() 95 op->fused_activation_function = FusedActivationFunctionType::kRelu1; in Run()
|
D | remove_trivial_quantized_activation_func.cc | 43 case OperatorType::kRelu1: in IsTrivialUnfusedActivationFunc() 73 case FusedActivationFunctionType::kRelu1: in IsTrivialFusedActivationFunc()
|
D | propagate_activation_function_into_constants.cc | 35 ac_op->type != OperatorType::kRelu1 && in Run() 98 case OperatorType::kRelu1: { in Run()
|
D | resolve_constant_unary.cc | 135 case OperatorType::kRelu1: in Run() 320 unary_op->type == OperatorType::kRelu1 || in Run() 331 case OperatorType::kRelu1: { in Run()
|
D | unfuse_activation_functions.cc | 48 case FusedActivationFunctionType::kRelu1: in Run()
|
D | reorder_elementwise_unary.cc | 38 case OperatorType::kRelu1: in IsElementwiseOperator()
|
D | propagate_fake_quant_num_bits.cc | 110 case OperatorType::kRelu1: in DoesOpBlockBackwardPropagation()
|
D | quantize.cc | 87 OperatorType::kRelu1, in SupportsQuantization() 413 op.type == OperatorType::kRelu || op.type == OperatorType::kRelu1 || in ChooseQuantizationForOperatorOutput()
|
D | propagate_fixed_sizes.cc | 2167 case OperatorType::kRelu1: in Run()
|
/external/tensorflow/tensorflow/lite/toco/tflite/ |
D | types.cc | 227 case FusedActivationFunctionType::kRelu1: in Serialize() 244 return FusedActivationFunctionType::kRelu1; in Deserialize()
|
D | types_test.cc | 213 {FusedActivationFunctionType::kRelu1, in TEST()
|
D | op_version.cc | 238 {{OperatorType::kRelu1, 1}, "1.5.0"}, in GetMinimumRuntimeVersionForModel()
|
D | operator_test.cc | 118 CheckSimpleOperator<Relu1Operator>("RELU_N1_TO_1", OperatorType::kRelu1); in TEST_F()
|
D | operator.cc | 2005 ::tflite::BuiltinOperator_RELU_N1_TO_1, OperatorType::kRelu1)); in BuildOperatorList()
|
/external/tensorflow/tensorflow/core/kernels/neon/ |
D | depthwiseconv_float.h | 560 Ac == FusedActivationFunctionType::kRelu1, 677 } else if (Ac == FusedActivationFunctionType::kRelu1) { 695 } else if (Ac == FusedActivationFunctionType::kRelu1) { 710 } else if (Ac == FusedActivationFunctionType::kRelu1) {
|
D | types.h | 23 enum class FusedActivationFunctionType { kNone, kRelu6, kRelu1, kRelu }; enumerator
|
/external/tensorflow/tensorflow/lite/kernels/internal/ |
D | depthwiseconv_float_test.cc | 90 FusedActivationFunctionType::kRelu1, in TryTestOneDepthwiseConv()
|
D | types.h | 30 kRelu1, enumerator
|
D | common.h | 47 case FusedActivationFunctionType::kRelu1: in GetActivationMinMax()
|
/external/tensorflow/tensorflow/compiler/mlir/lite/transforms/ |
D | optimize.cc | 64 constexpr char kRelu1[] = "RELU_N1_TO_1"; variable 966 FuseFullyConnectedAndReluX<TFL::Relu1Op, kRelu1>, in runOnFunction() 978 FuseFullyConnectedAndReluX<TFL::Relu1Op, kRelu1>, in runOnFunction()
|
/external/tensorflow/tensorflow/lite/kernels/internal/reference/ |
D | legacy_reference_ops.h | 345 Ac == FusedActivationFunctionType::kRelu1, in Conv() 556 Ac == FusedActivationFunctionType::kRelu1, in FullyConnected() 1268 Ac == FusedActivationFunctionType::kRelu1, in Add() 1324 Ac == FusedActivationFunctionType::kRelu1, in BroadcastAdd() 1391 Ac == FusedActivationFunctionType::kRelu1, in BroadcastAddFivefold() 1445 Ac == FusedActivationFunctionType::kRelu1, in Add() 1615 Ac == FusedActivationFunctionType::kRelu1, in AveragePool() 1706 Ac == FusedActivationFunctionType::kRelu1, in MaxPool()
|
/external/tensorflow/tensorflow/lite/micro/kernels/ceva/ |
D | types.h | 109 kRelu1, enumerator 214 enum class FusedActivationFunctionType : uint8_t { kNone, kRelu6, kRelu1, kRelu };
|
/external/tensorflow/tensorflow/lite/toco/ |
D | model.h | 76 kRelu1, enumerator 773 Relu1Operator() : Operator(OperatorType::kRelu1) {}
|
D | dump_graphviz.cc | 367 case FusedActivationFunctionType::kRelu1: in GetOpAttributes()
|
/external/tensorflow/tensorflow/lite/kernels/internal/optimized/ |
D | legacy_optimized_ops.h | 1783 Ac == FusedActivationFunctionType::kRelu1, in FullyConnected() 2797 Ac == FusedActivationFunctionType::kRelu1, in Conv() 2824 Ac == FusedActivationFunctionType::kRelu1, in Conv() 2877 Ac == FusedActivationFunctionType::kRelu1, in ConvAsGemm() 3520 Ac == FusedActivationFunctionType::kRelu1, in Add() 3589 Ac == FusedActivationFunctionType::kRelu1, in BroadcastAdd() 3628 Ac == FusedActivationFunctionType::kRelu1, in BroadcastAddFivefold() 3683 Ac == FusedActivationFunctionType::kRelu1, in Add() 3837 Ac == FusedActivationFunctionType::kRelu1, in AveragePool() 3928 Ac == FusedActivationFunctionType::kRelu1, in MaxPool()
|