Home
last modified time | relevance | path

Searched refs:input_min (Results 1 – 25 of 44) sorted by relevance

12

/external/tensorflow/tensorflow/lite/experimental/micro/kernels/
Dfully_connected_test.cc103 std::initializer_list<uint8_t> input_data, float input_min, float input_max, in TestFullyConnectedQuantized() argument
121 CreateQuantizedTensor(input_data, input_dims, "input_tensor", input_min, in TestFullyConnectedQuantized()
266 const float input_min = -63.5f; in TF_LITE_MICRO_TEST() local
280 F2Q(1, input_min, input_max), F2Q(2, input_min, input_max), in TF_LITE_MICRO_TEST()
281 F2Q(3, input_min, input_max), F2Q(4, input_min, input_max), in TF_LITE_MICRO_TEST()
282 F2Q(5, input_min, input_max), F2Q(6, input_min, input_max), in TF_LITE_MICRO_TEST()
283 F2Q(7, input_min, input_max), F2Q(8, input_min, input_max), in TF_LITE_MICRO_TEST()
284 F2Q(-9, input_min, input_max), F2Q(-10, input_min, input_max), in TF_LITE_MICRO_TEST()
285 F2Q(1, input_min, input_max), F2Q(2, input_min, input_max), in TF_LITE_MICRO_TEST()
286 F2Q(3, input_min, input_max), F2Q(4, input_min, input_max), in TF_LITE_MICRO_TEST()
[all …]
Ddepthwise_conv_test.cc110 std::initializer_list<uint8_t> input_data, float input_min, float input_max, in TestDepthwiseConvQuantized() argument
128 CreateQuantizedTensor(input_data, input_dims, "input_tensor", input_min, in TestDepthwiseConvQuantized()
228 const float input_min = -63.5f; in TF_LITE_MICRO_TEST() local
243 F2Q(1, input_min, input_max), in TF_LITE_MICRO_TEST()
244 F2Q(2, input_min, input_max), in TF_LITE_MICRO_TEST()
245 F2Q(7, input_min, input_max), in TF_LITE_MICRO_TEST()
246 F2Q(8, input_min, input_max), in TF_LITE_MICRO_TEST()
247 F2Q(3, input_min, input_max), in TF_LITE_MICRO_TEST()
248 F2Q(4, input_min, input_max), in TF_LITE_MICRO_TEST()
249 F2Q(9, input_min, input_max), in TF_LITE_MICRO_TEST()
[all …]
Dsoftmax_test.cc91 float input_min, float input_max, in TestSoftmaxQuantized() argument
104 CreateQuantizedTensor(input_data, input_dims, "input_tensor", input_min, in TestSoftmaxQuantized()
191 const float input_min = -63.5f; in TF_LITE_MICRO_TEST() local
200 F2Q(1.0, input_min, input_max), in TF_LITE_MICRO_TEST()
201 F2Q(2.0, input_min, input_max), in TF_LITE_MICRO_TEST()
202 F2Q(3.0, input_min, input_max), in TF_LITE_MICRO_TEST()
203 F2Q(4.0, input_min, input_max), in TF_LITE_MICRO_TEST()
204 F2Q(5.0, input_min, input_max), in TF_LITE_MICRO_TEST()
206 input_min, input_max, // Input quantized range. in TF_LITE_MICRO_TEST()
/external/tensorflow/tensorflow/core/kernels/
Dquantization_utils_test.cc34 void TestRequantizeMany(Eigen::ThreadPoolDevice* eigen_device, float input_min, in TestRequantizeMany() argument
43 QuantizedToFloat(values_quantized[value_index], input_min, input_max), in TestRequantizeMany()
54 RequantizeManyInNewRange(input_array.data(), input_array.size(), input_min, in TestRequantizeMany()
59 *eigen_device, i_tensor, input_min, input_max, output_min, output_max, in TestRequantizeMany()
70 << "]=" << values_quantized[value_index] << ", input_min=" << input_min in TestRequantizeMany()
76 void TestRequantizeMany8To32Bit(float input_min, float input_max, in TestRequantizeMany8To32Bit() argument
85 QuantizedToFloat(values_quantized[value_index], input_min, input_max), in TestRequantizeMany8To32Bit()
95 RequantizeManyInNewRange(input_array.data(), input_array.size(), input_min, in TestRequantizeMany8To32Bit()
106 << "]=" << values_quantized[value_index] << ", input_min=" << input_min in TestRequantizeMany8To32Bit()
230 const float input_min = ranges[range_index][0]; in TestRequantizeManyInNewRangeEigenVsNonEigen() local
[all …]
Dquantized_activation_ops_test.cc45 const float input_min = -128.0f; in TEST_F() local
52 FloatTensorToQuantized<quint8>(input_float, input_min, input_max); in TEST_F()
58 AddInputFromArray<float>(TensorShape({1}), {input_min}); in TEST_F()
76 const float input_min = -128.0f; in TEST_F() local
83 FloatTensorToQuantized<quint8>(input_float, input_min, input_max); in TEST_F()
89 AddInputFromArray<float>(TensorShape({1}), {input_min}); in TEST_F()
Dquantized_pooling_ops_test.cc51 const float input_min = 0.0f; in TEST_F() local
62 FloatTensorToQuantized<quint8>(input_float, input_min, input_max); in TEST_F()
72 AddInputFromArray<float>(TensorShape({1}), {input_min}); in TEST_F()
96 const float input_min = 0.0f; in TEST_F() local
107 FloatTensorToQuantized<quint8>(input_float, input_min, input_max); in TEST_F()
117 AddInputFromArray<float>(TensorShape({1}), {input_min}); in TEST_F()
Dquantized_bias_add_op_test.cc51 const float input_min = 0.0f; in TEST_F() local
59 FloatTensorToQuantized<quint8>(input_float, input_min, input_max); in TEST_F()
77 AddInputFromArray<float>(TensorShape({1}), {input_min}); in TEST_F()
101 const float input_min = -2164.25f; in TEST_F() local
119 FloatTensorToQuantized<quint8>(input_float, input_min, input_max); in TEST_F()
159 AddInputFromArray<float>(TensorShape({1}), {input_min}); in TEST_F()
Dmkl_quantized_pooling_ops_test.cc81 const float input_min = 0.0f; in TEST_F() local
92 FloatTensorToQuantized<quint8>(input_float, input_min, input_max); in TEST_F()
110 AddInputFromArray<float>(TensorShape({1}), {input_min}); in TEST_F()
150 const float input_min = 0.0f; in TEST_F() local
161 FloatTensorToQuantized<quint8>(input_float, input_min, input_max); in TEST_F()
178 AddInputFromArray<float>(TensorShape({1}), {input_min}); in TEST_F()
Dmkl_requantization_range_per_channel_op.cc46 const Tensor& input_min = ctx->input(kInputMinIndex); in Compute() local
51 ctx, input_min.dim_size(0) == depth, in Compute()
53 depth, " was ", input_min.dim_size(0))); in Compute()
59 const float* input_min_data = input_min.flat<float>().data(); in Compute()
Dquantized_bias_add_op.cc41 const float input_min = context->input(2).flat<float>()(0); in Compute() local
71 GetOutputMinAndMaxForQuantizedAdd(input_min, input_max, bias_min, in Compute()
75 bias_ui8_array.size(), input_min, input_max, in Compute()
80 context->template eigen_device<CPUDevice>(), input, input_min, in Compute()
Dquantized_concat_op.cc41 const float input_min = (*input_min_and_max)[input_index].first; in Copy() local
43 if (input_min == output_min && input_max == output_max) { in Copy()
52 QuantizedToFloatStruct<T> q2f(input_min, input_max); in Copy()
87 const float input_min = input_mins[i].flat<float>()(0); in CalculateInputAndOutputRange() local
89 input_mins_and_maxes->emplace_back(input_min, input_max); in CalculateInputAndOutputRange()
90 overall_min = std::min(overall_min, input_min); in CalculateInputAndOutputRange()
Dquantized_batch_norm_op.cc31 void ReferenceBatchNorm(const Tensor& input, const float input_min, in ReferenceBatchNorm() argument
57 QuantizedToFloat(input_flat(input_index), input_min, input_max); in ReferenceBatchNorm()
94 void FixedPointBatchNorm(const Tensor& input, const float input_min, in FixedPointBatchNorm() argument
150 RequantizeInNewRange<T1, T2>(input_flat(input_index), input_min, in FixedPointBatchNorm()
176 const float input_min = context->input(1).flat<float>()(0); in Compute() local
212 FixedPointBatchNorm<T1, T2>(input, input_min, input_max, mean, mean_min, in Compute()
Dquantized_batch_norm_op_test.cc61 const float input_min = -128.0f; in TEST_F() local
72 FloatTensorToQuantized<quint8>(input_float, input_min, input_max); in TEST_F()
100 AddInputFromArray<float>(TensorShape({1}), {input_min}); in TEST_F()
158 const float input_min = -128.0f; in TEST_F() local
169 FloatTensorToQuantized<quint8>(input_float, input_min, input_max); in TEST_F()
197 AddInputFromArray<float>(TensorShape({1}), {input_min}); in TEST_F()
Dquantization_utils.cc20 void GetOutputMinAndMaxForQuantizedAdd(float input_min, float input_max, in GetOutputMinAndMaxForQuantizedAdd() argument
36 std::max(input_max, std::max(-input_min, std::max(smaller_input_max, in GetOutputMinAndMaxForQuantizedAdd()
Dmeta_support.cc257 float input_min, float input_max, float output_min, in Requantize() argument
269 params.kernel.input_range_min = input_min; in Requantize()
272 CalculateRangeScale<int32_t>(input_min, input_max); in Requantize()
348 float input_min, float input_max, float bias_min, in QuantizedBiasAdd() argument
363 params.kernel.input_range_min = input_min; in QuantizedBiasAdd()
366 CalculateRangeScale<uint8_t>(input_min, input_max); in QuantizedBiasAdd()
Dmeta_support.h80 float input_min, float input_max, float output_min,
100 float input_min, float input_max, float bias_min,
Dquantize_and_dequantize_op.h114 auto input_min = input_min_tensor->scalar<T>(); in Compute() local
117 input_min.device(d) = input.minimum(); in Compute()
119 d.memcpyDeviceToHost(&min_range, input_min.data(), sizeof(T)); in Compute()
/external/tensorflow/tensorflow/core/api_def/base_api/
Dapi_def_QuantizeAndDequantizeV2.pbtxt10 name: "input_min"
77 If range_given == False, the initial input_min, input_max will be determined
79 the specified values of input_min, input_max are used.
81 Note: If the input_min, input_max are specified, they do not need to equal the
87 [input_min, input_max] range to a range that lies within the representable
90 It determines the scale from one of input_min and input_max, then updates the
95 * if the output is signed, num_bits = 8, [input_min, input_max] = [-10.0,
98 * if the output is signed, num_bits = 8, [input_min, input_max] = [-10.0,
100 would update input_min to be 128.0 / 12.7 = -10.07874
101 * if the output is unsigned, input_min is forced to be 0, and only the
[all …]
Dapi_def_Requantize.pbtxt4 name: "input_min"
57 `[input_min, input_max]` are scalar floats that specify the range for the float
58 interpretation of the `input` data. For example, if `input_min` is -1.0f and
Dapi_def_QuantizedReshape.pbtxt10 name: "input_min"
24 This value is copied from input_min.
Dapi_def_RequantizationRange.pbtxt4 name: "input_min"
36 Given a quantized tensor described by `(input, input_min, input_max)`, outputs a
Dapi_def_QuantizeDownAndShrinkRange.pbtxt4 name: "input_min"
44 [input_min, input_max] are scalar floats that specify the range for the float
45 interpretation of the 'input' data. For example, if input_min is -1.0f and
/external/tensorflow/tensorflow/core/graph/
Dquantize_training.cc54 float input_min; member
63 input_min(min), in EdgeToConvert()
80 bool* range_given, float* input_min, float* input_max) { in FindType() argument
95 *input_min = 0; in FindType()
100 *input_min = 0; in FindType()
105 *input_min = -1; in FindType()
113 FindType(graph, edge->src(), signed_input, range_given, input_min, in FindType()
123 FindType(graph, edge->src(), signed_input, range_given, input_min, in FindType()
504 std::vector<Node*>* added_variables, Node** input_min, in MakeInputMinMax() argument
510 input_min_tensor.flat<float>()(0) = edge.input_min; in MakeInputMinMax()
[all …]
/external/tensorflow/tensorflow/compiler/tf2xla/kernels/
Dfake_quantize_ops.cc100 float input_min, input_max; in FakeQuantWithMinMaxArgsOp() local
101 OP_REQUIRES_OK(ctx, ctx->GetAttr("min", &input_min)); in FakeQuantWithMinMaxArgsOp()
103 CpuNudge(input_min, input_max, quant_min_, quant_max_, &nudged_input_min_, in FakeQuantWithMinMaxArgsOp()
148 float input_min, input_max, scale; in FakeQuantWithMinMaxArgsGradOp() local
149 OP_REQUIRES_OK(ctx, ctx->GetAttr("min", &input_min)); in FakeQuantWithMinMaxArgsGradOp()
151 CpuNudge(input_min, input_max, quant_min, quant_max, &nudged_input_min_, in FakeQuantWithMinMaxArgsGradOp()
202 xla::XlaOp input_min = ctx->Input(1); in Compile() local
207 XlaNudge(b, data_type, input_min, input_max, quant_min_, quant_max_, in Compile()
245 xla::XlaOp input_min = ctx->Input(2); in Compile() local
250 XlaNudge(b, data_type, input_min, input_max, quant_min_, quant_max_, in Compile()
/external/tensorflow/tensorflow/compiler/tests/
Dfake_quant_ops_test.py82 def _TestOp(self, input_min, input_max, num_bits, narrow_range, argument
116 min=input_min,
180 def _TestOp(self, input_min, input_max, num_bits, narrow_range, argument
210 min=input_min,
281 def _TestOp(self, input_min, input_max, num_bits, narrow_range, argument
324 min_placeholder: input_min,
386 def _TestOp(self, input_min, input_max, num_bits, narrow_range, argument
428 min_placeholder: input_min,

12