Home
last modified time | relevance | path

Searched refs:input_max (Results 1 – 25 of 35) sorted by relevance

12

/external/tensorflow/tensorflow/core/kernels/
Dquantization_utils_test.cc35 float input_max, float output_min, float output_max, in TestRequantizeMany() argument
43 QuantizedToFloat(values_quantized[value_index], input_min, input_max), in TestRequantizeMany()
55 input_max, output_min, output_max, in TestRequantizeMany()
59 *eigen_device, i_tensor, input_min, input_max, output_min, output_max, in TestRequantizeMany()
71 << ", input_max=" << input_max << ", output_min=" << output_min in TestRequantizeMany()
76 void TestRequantizeMany8To32Bit(float input_min, float input_max, in TestRequantizeMany8To32Bit() argument
85 QuantizedToFloat(values_quantized[value_index], input_min, input_max), in TestRequantizeMany8To32Bit()
96 input_max, output_min, output_max, in TestRequantizeMany8To32Bit()
107 << ", input_max=" << input_max << ", output_min=" << output_min in TestRequantizeMany8To32Bit()
231 const float input_max = ranges[range_index][1]; in TestRequantizeManyInNewRangeEigenVsNonEigen() local
[all …]
Dquantized_activation_ops_test.cc46 const float input_max = 127.0f; in TEST_F() local
52 FloatTensorToQuantized<quint8>(input_float, input_min, input_max); in TEST_F()
59 AddInputFromArray<float>(TensorShape({1}), {input_max}); in TEST_F()
77 const float input_max = 127.0f; in TEST_F() local
83 FloatTensorToQuantized<quint8>(input_float, input_min, input_max); in TEST_F()
90 AddInputFromArray<float>(TensorShape({1}), {input_max}); in TEST_F()
Dquantized_pooling_ops_test.cc52 const float input_max = 255.0f; in TEST_F() local
62 FloatTensorToQuantized<quint8>(input_float, input_min, input_max); in TEST_F()
73 AddInputFromArray<float>(TensorShape({1}), {input_max}); in TEST_F()
97 const float input_max = 255.0f; in TEST_F() local
107 FloatTensorToQuantized<quint8>(input_float, input_min, input_max); in TEST_F()
118 AddInputFromArray<float>(TensorShape({1}), {input_max}); in TEST_F()
Dquantized_bias_add_op_test.cc52 const float input_max = 60.0f; in TEST_F() local
59 FloatTensorToQuantized<quint8>(input_float, input_min, input_max); in TEST_F()
78 AddInputFromArray<float>(TensorShape({1}), {input_max}); in TEST_F()
102 const float input_max = 2006.27f; in TEST_F() local
119 FloatTensorToQuantized<quint8>(input_float, input_min, input_max); in TEST_F()
160 AddInputFromArray<float>(TensorShape({1}), {input_max}); in TEST_F()
Dquantized_bias_add_op.cc42 const float input_max = context->input(3).flat<float>()(0); in Compute() local
71 GetOutputMinAndMaxForQuantizedAdd(input_min, input_max, bias_min, in Compute()
75 bias_ui8_array.size(), input_min, input_max, in Compute()
81 input_max, bias, bias_min, bias_max, output, &total_min, &total_max); in Compute()
Dquantized_batch_norm_op.cc32 const float input_max, const Tensor& mean, in ReferenceBatchNorm() argument
57 QuantizedToFloat(input_flat(input_index), input_min, input_max); in ReferenceBatchNorm()
95 const float input_max, const Tensor& mean, in FixedPointBatchNorm() argument
151 input_max, *output_min, *output_max); in FixedPointBatchNorm()
177 const float input_max = context->input(2).flat<float>()(0); in Compute() local
212 FixedPointBatchNorm<T1, T2>(input, input_min, input_max, mean, mean_min, in Compute()
Dquantized_concat_op.cc42 const float input_max = (*input_min_and_max)[input_index].second; in Copy() local
43 if (input_min == output_min && input_max == output_max) { in Copy()
52 QuantizedToFloatStruct<T> q2f(input_min, input_max); in Copy()
88 const float input_max = input_maxes[i].flat<float>()(0); in CalculateInputAndOutputRange() local
89 input_mins_and_maxes->emplace_back(input_min, input_max); in CalculateInputAndOutputRange()
91 overall_max = std::max(overall_max, input_max); in CalculateInputAndOutputRange()
Dquantized_batch_norm_op_test.cc62 const float input_max = 127.0f; in TEST_F() local
72 FloatTensorToQuantized<quint8>(input_float, input_min, input_max); in TEST_F()
101 AddInputFromArray<float>(TensorShape({1}), {input_max}); in TEST_F()
159 const float input_max = 127.0f; in TEST_F() local
169 FloatTensorToQuantized<quint8>(input_float, input_min, input_max); in TEST_F()
198 AddInputFromArray<float>(TensorShape({1}), {input_max}); in TEST_F()
Dquantize_and_dequantize_op.h44 auto input_max = input_max_tensor->scalar<T>(); in Compute() local
47 input_max.device(d) = input.maximum(); in Compute()
50 d.memcpyDeviceToHost(&max_range, input_max.data(), sizeof(T)); in Compute()
Dquantization_utils.cc20 void GetOutputMinAndMaxForQuantizedAdd(float input_min, float input_max, in GetOutputMinAndMaxForQuantizedAdd() argument
36 std::max(input_max, std::max(-input_min, std::max(smaller_input_max, in GetOutputMinAndMaxForQuantizedAdd()
Dmeta_support.h80 float input_min, float input_max, float output_min,
100 float input_min, float input_max, float bias_min,
Dmeta_support.cc257 float input_min, float input_max, float output_min, in Requantize() argument
272 CalculateRangeScale<int32_t>(input_min, input_max); in Requantize()
348 float input_min, float input_max, float bias_min, in QuantizedBiasAdd() argument
366 CalculateRangeScale<uint8_t>(input_min, input_max); in QuantizedBiasAdd()
Dquantized_instance_norm.cc278 float input_max = context->input(2).flat<float>()(0); in Compute() local
279 float input_scale = (input_max - input_min) / 255.0f; in Compute()
281 OP_REQUIRES(context, input_min < input_max, in Compute()
284 " >= ", input_max)); in Compute()
Dquantization_utils.h796 void GetOutputMinAndMaxForQuantizedAdd(float input_min, float input_max,
806 float input_max, const Tensor& smaller_input, in QuantizedAddUsingEigen() argument
814 GetOutputMinAndMaxForQuantizedAdd(input_min, input_max, smaller_input_min, in QuantizedAddUsingEigen()
826 QuantizedToFloatStruct<T1> input_q2f(input_min, input_max); in QuantizedAddUsingEigen()
853 float input_min, float input_max, const Tensor& smaller_input, in QuantizedAdd() argument
860 GetOutputMinAndMaxForQuantizedAdd(input_min, input_max, smaller_input_min, in QuantizedAdd()
886 input_value, input_min, input_max, total_min, total_max); in QuantizedAdd()
/external/tensorflow/tensorflow/core/graph/
Dquantize_training.cc55 float input_max; member
64 input_max(max) {} in EdgeToConvert()
80 bool* range_given, float* input_min, float* input_max) { in FindType() argument
96 *input_max = 6; in FindType()
101 *input_max = 1; in FindType()
106 *input_max = 1; in FindType()
114 input_max); in FindType()
124 input_max); in FindType()
499 Node** input_max) { in MakeInputMinMax() argument
511 input_max_tensor.flat<float>()(0) = edge.input_max; in MakeInputMinMax()
[all …]
/external/tensorflow/tensorflow/compiler/tf2xla/kernels/
Dfake_quantize_ops.cc106 float input_min, input_max; in FakeQuantWithMinMaxArgsOp() local
108 OP_REQUIRES_OK(ctx, ctx->GetAttr("max", &input_max)); in FakeQuantWithMinMaxArgsOp()
109 CpuNudge(input_min, input_max, quant_min_, quant_max_, &nudged_input_min_, in FakeQuantWithMinMaxArgsOp()
154 float input_min, input_max, scale; in FakeQuantWithMinMaxArgsGradOp() local
156 OP_REQUIRES_OK(ctx, ctx->GetAttr("max", &input_max)); in FakeQuantWithMinMaxArgsGradOp()
157 CpuNudge(input_min, input_max, quant_min, quant_max, &nudged_input_min_, in FakeQuantWithMinMaxArgsGradOp()
210 xla::ComputationDataHandle input_max = ctx->Input(2); in Compile() local
214 XlaNudge(b, data_type, input_min, input_max, quant_min_, quant_max_, in Compile()
251 xla::ComputationDataHandle input_max = ctx->Input(3); in Compile() local
255 XlaNudge(b, data_type, input_min, input_max, quant_min_, quant_max_, in Compile()
Dquantize_and_dequantize_op.cc50 xla::ComputationDataHandle input_min, input_max; in Compile() local
56 input_max = XlaHelpers::FloatLiteral(b, data_type, input_max_value); in Compile()
62 input_max = in Compile()
65 xla::ComputationDataHandle m = b->Max(b->Abs(input_min), b->Abs(input_max)); in Compile()
/external/webrtc/webrtc/modules/audio_coding/neteq/
Dmerge.cc66 int16_t expanded_max, input_max; in Process() local
69 &expanded_max, &input_max); in Process()
92 expanded_max, input_max, old_length, in Process()
208 int16_t* expanded_max, int16_t* input_max) const { in SignalScaling()
213 *input_max = WebRtcSpl_MaxAbsValueW16(input, mod_input_length); in SignalScaling()
228 WebRtcSpl_NormW32(*input_max * *input_max); in SignalScaling()
310 size_t Merge::CorrelateAndPeakSearch(int16_t expanded_max, int16_t input_max, in CorrelateAndPeakSearch() argument
318 if (expanded_max * input_max > 26843546) { in CorrelateAndPeakSearch()
Dmerge.h76 int16_t* expanded_max, int16_t* input_max) const;
87 size_t CorrelateAndPeakSearch(int16_t expanded_max, int16_t input_max,
/external/tensorflow/tensorflow/compiler/tests/
Dfake_quant_ops_test.py82 def _TestOp(self, input_min, input_max, num_bits, narrow_range, argument
117 max=input_max,
180 def _TestOp(self, input_min, input_max, num_bits, narrow_range, argument
211 max=input_max,
281 def _TestOp(self, input_min, input_max, num_bits, narrow_range, argument
325 max_placeholder: input_max
386 def _TestOp(self, input_min, input_max, num_bits, narrow_range, argument
429 max_placeholder: input_max
/external/tensorflow/tensorflow/core/api_def/base_api/
Dapi_def_Requantize.pbtxt10 name: "input_max"
55 [input_min, input_max] are scalar floats that specify the range for the float
57 input_max is 1.0f, and we are dealing with quint16 quantized data, then a 0
Dapi_def_QuantizedReshape.pbtxt16 name: "input_max"
30 This value is copied from input_max.
Dapi_def_RequantizationRange.pbtxt10 name: "input_max"
33 summary: "Given a quantized tensor described by (input, input_min, input_max), outputs a"
Dapi_def_QuantizeDownAndShrinkRange.pbtxt10 name: "input_max"
44 [input_min, input_max] are scalar floats that specify the range for the float
46 input_max is 1.0f, and we are dealing with quint16 quantized data, then a 0
Dapi_def_QuantizeAndDequantizeV2.pbtxt17 name: "input_max"
57 1. m = max(abs(input_min), abs(input_max)) if range_given is true,

12