/external/tensorflow/tensorflow/lite/experimental/micro/kernels/ |
D | fully_connected_test.cc | 103 std::initializer_list<uint8_t> input_data, float input_min, float input_max, in TestFullyConnectedQuantized() argument 122 input_max), in TestFullyConnectedQuantized() 267 const float input_max = 64.0f; in TF_LITE_MICRO_TEST() local 280 F2Q(1, input_min, input_max), F2Q(2, input_min, input_max), in TF_LITE_MICRO_TEST() 281 F2Q(3, input_min, input_max), F2Q(4, input_min, input_max), in TF_LITE_MICRO_TEST() 282 F2Q(5, input_min, input_max), F2Q(6, input_min, input_max), in TF_LITE_MICRO_TEST() 283 F2Q(7, input_min, input_max), F2Q(8, input_min, input_max), in TF_LITE_MICRO_TEST() 284 F2Q(-9, input_min, input_max), F2Q(-10, input_min, input_max), in TF_LITE_MICRO_TEST() 285 F2Q(1, input_min, input_max), F2Q(2, input_min, input_max), in TF_LITE_MICRO_TEST() 286 F2Q(3, input_min, input_max), F2Q(4, input_min, input_max), in TF_LITE_MICRO_TEST() [all …]
|
D | depthwise_conv_test.cc | 110 std::initializer_list<uint8_t> input_data, float input_min, float input_max, in TestDepthwiseConvQuantized() argument 129 input_max), in TestDepthwiseConvQuantized() 229 const float input_max = 64.0f; in TF_LITE_MICRO_TEST() local 243 F2Q(1, input_min, input_max), in TF_LITE_MICRO_TEST() 244 F2Q(2, input_min, input_max), in TF_LITE_MICRO_TEST() 245 F2Q(7, input_min, input_max), in TF_LITE_MICRO_TEST() 246 F2Q(8, input_min, input_max), in TF_LITE_MICRO_TEST() 247 F2Q(3, input_min, input_max), in TF_LITE_MICRO_TEST() 248 F2Q(4, input_min, input_max), in TF_LITE_MICRO_TEST() 249 F2Q(9, input_min, input_max), in TF_LITE_MICRO_TEST() [all …]
|
D | softmax_test.cc | 91 float input_min, float input_max, in TestSoftmaxQuantized() argument 105 input_max), in TestSoftmaxQuantized() 192 const float input_max = 64.0f; in TF_LITE_MICRO_TEST() local 200 F2Q(1.0, input_min, input_max), in TF_LITE_MICRO_TEST() 201 F2Q(2.0, input_min, input_max), in TF_LITE_MICRO_TEST() 202 F2Q(3.0, input_min, input_max), in TF_LITE_MICRO_TEST() 203 F2Q(4.0, input_min, input_max), in TF_LITE_MICRO_TEST() 204 F2Q(5.0, input_min, input_max), in TF_LITE_MICRO_TEST() 206 input_min, input_max, // Input quantized range. in TF_LITE_MICRO_TEST()
|
/external/tensorflow/tensorflow/core/kernels/ |
D | quantization_utils_test.cc | 35 float input_max, float output_min, float output_max, in TestRequantizeMany() argument 43 QuantizedToFloat(values_quantized[value_index], input_min, input_max), in TestRequantizeMany() 55 input_max, output_min, output_max, in TestRequantizeMany() 59 *eigen_device, i_tensor, input_min, input_max, output_min, output_max, in TestRequantizeMany() 71 << ", input_max=" << input_max << ", output_min=" << output_min in TestRequantizeMany() 76 void TestRequantizeMany8To32Bit(float input_min, float input_max, in TestRequantizeMany8To32Bit() argument 85 QuantizedToFloat(values_quantized[value_index], input_min, input_max), in TestRequantizeMany8To32Bit() 96 input_max, output_min, output_max, in TestRequantizeMany8To32Bit() 107 << ", input_max=" << input_max << ", output_min=" << output_min in TestRequantizeMany8To32Bit() 231 const float input_max = ranges[range_index][1]; in TestRequantizeManyInNewRangeEigenVsNonEigen() local [all …]
|
D | quantized_activation_ops_test.cc | 46 const float input_max = 127.0f; in TEST_F() local 52 FloatTensorToQuantized<quint8>(input_float, input_min, input_max); in TEST_F() 59 AddInputFromArray<float>(TensorShape({1}), {input_max}); in TEST_F() 77 const float input_max = 127.0f; in TEST_F() local 83 FloatTensorToQuantized<quint8>(input_float, input_min, input_max); in TEST_F() 90 AddInputFromArray<float>(TensorShape({1}), {input_max}); in TEST_F()
|
D | quantized_pooling_ops_test.cc | 52 const float input_max = 255.0f; in TEST_F() local 62 FloatTensorToQuantized<quint8>(input_float, input_min, input_max); in TEST_F() 73 AddInputFromArray<float>(TensorShape({1}), {input_max}); in TEST_F() 97 const float input_max = 255.0f; in TEST_F() local 107 FloatTensorToQuantized<quint8>(input_float, input_min, input_max); in TEST_F() 118 AddInputFromArray<float>(TensorShape({1}), {input_max}); in TEST_F()
|
D | quantized_bias_add_op_test.cc | 52 const float input_max = 60.0f; in TEST_F() local 59 FloatTensorToQuantized<quint8>(input_float, input_min, input_max); in TEST_F() 78 AddInputFromArray<float>(TensorShape({1}), {input_max}); in TEST_F() 102 const float input_max = 2006.27f; in TEST_F() local 119 FloatTensorToQuantized<quint8>(input_float, input_min, input_max); in TEST_F() 160 AddInputFromArray<float>(TensorShape({1}), {input_max}); in TEST_F()
|
D | mkl_requantization_range_per_channel_op.cc | 47 const Tensor& input_max = ctx->input(kInputMaxIndex); in Compute() local 49 const size_t depth = input_max.NumElements(); in Compute() 55 ctx, input_max.dim_size(0) == depth, in Compute() 57 depth, " was ", input_max.dim_size(0))); in Compute() 60 const float* input_max_data = input_max.flat<float>().data(); in Compute()
|
D | mkl_quantized_pooling_ops_test.cc | 82 const float input_max = 255.0f; in TEST_F() local 92 FloatTensorToQuantized<quint8>(input_float, input_min, input_max); in TEST_F() 111 AddInputFromArray<float>(TensorShape({1}), {input_max}); in TEST_F() 151 const float input_max = 255.0f; in TEST_F() local 161 FloatTensorToQuantized<quint8>(input_float, input_min, input_max); in TEST_F() 179 AddInputFromArray<float>(TensorShape({1}), {input_max}); in TEST_F()
|
D | quantized_bias_add_op.cc | 42 const float input_max = context->input(3).flat<float>()(0); in Compute() local 71 GetOutputMinAndMaxForQuantizedAdd(input_min, input_max, bias_min, in Compute() 75 bias_ui8_array.size(), input_min, input_max, in Compute() 81 input_max, bias, bias_min, bias_max, output, &total_min, &total_max); in Compute()
|
D | quantized_concat_op.cc | 42 const float input_max = (*input_min_and_max)[input_index].second; in Copy() local 43 if (input_min == output_min && input_max == output_max) { in Copy() 52 QuantizedToFloatStruct<T> q2f(input_min, input_max); in Copy() 88 const float input_max = input_maxes[i].flat<float>()(0); in CalculateInputAndOutputRange() local 89 input_mins_and_maxes->emplace_back(input_min, input_max); in CalculateInputAndOutputRange() 91 overall_max = std::max(overall_max, input_max); in CalculateInputAndOutputRange()
|
D | quantized_batch_norm_op.cc | 32 const float input_max, const Tensor& mean, in ReferenceBatchNorm() argument 57 QuantizedToFloat(input_flat(input_index), input_min, input_max); in ReferenceBatchNorm() 95 const float input_max, const Tensor& mean, in FixedPointBatchNorm() argument 151 input_max, *output_min, *output_max); in FixedPointBatchNorm() 177 const float input_max = context->input(2).flat<float>()(0); in Compute() local 212 FixedPointBatchNorm<T1, T2>(input, input_min, input_max, mean, mean_min, in Compute()
|
D | quantized_batch_norm_op_test.cc | 62 const float input_max = 127.0f; in TEST_F() local 72 FloatTensorToQuantized<quint8>(input_float, input_min, input_max); in TEST_F() 101 AddInputFromArray<float>(TensorShape({1}), {input_max}); in TEST_F() 159 const float input_max = 127.0f; in TEST_F() local 169 FloatTensorToQuantized<quint8>(input_float, input_min, input_max); in TEST_F() 198 AddInputFromArray<float>(TensorShape({1}), {input_max}); in TEST_F()
|
D | quantization_utils.cc | 20 void GetOutputMinAndMaxForQuantizedAdd(float input_min, float input_max, in GetOutputMinAndMaxForQuantizedAdd() argument 36 std::max(input_max, std::max(-input_min, std::max(smaller_input_max, in GetOutputMinAndMaxForQuantizedAdd()
|
D | meta_support.h | 80 float input_min, float input_max, float output_min, 100 float input_min, float input_max, float bias_min,
|
D | quantize_and_dequantize_op.h | 115 auto input_max = input_max_tensor->scalar<T>(); in Compute() local 118 input_max.device(d) = input.maximum(); in Compute() 120 d.memcpyDeviceToHost(&max_range, input_max.data(), sizeof(T)); in Compute()
|
/external/tensorflow/tensorflow/core/api_def/base_api/ |
D | api_def_QuantizeAndDequantizeV2.pbtxt | 18 name: "input_max" 77 If range_given == False, the initial input_min, input_max will be determined 79 the specified values of input_min, input_max are used. 81 Note: If the input_min, input_max are specified, they do not need to equal the 87 [input_min, input_max] range to a range that lies within the representable 90 It determines the scale from one of input_min and input_max, then updates the 95 * if the output is signed, num_bits = 8, [input_min, input_max] = [-10.0, 97 would update input_max to be 127 / 12.8 = 9.921875 98 * if the output is signed, num_bits = 8, [input_min, input_max] = [-10.0, 102 specified input_max is used. [all …]
|
D | api_def_Requantize.pbtxt | 10 name: "input_max" 57 `[input_min, input_max]` are scalar floats that specify the range for the float 59 `input_max` is 1.0f, and we are dealing with `quint16` quantized data, then a 0
|
D | api_def_QuantizedReshape.pbtxt | 16 name: "input_max" 30 This value is copied from input_max.
|
D | api_def_RequantizationRange.pbtxt | 10 name: "input_max" 36 Given a quantized tensor described by `(input, input_min, input_max)`, outputs a
|
/external/tensorflow/tensorflow/core/graph/ |
D | quantize_training.cc | 55 float input_max; member 64 input_max(max) {} in EdgeToConvert() 80 bool* range_given, float* input_min, float* input_max) { in FindType() argument 96 *input_max = 6; in FindType() 101 *input_max = 1; in FindType() 106 *input_max = 1; in FindType() 114 input_max); in FindType() 124 input_max); in FindType() 505 Node** input_max) { in MakeInputMinMax() argument 517 input_max_tensor.flat<float>()(0) = edge.input_max; in MakeInputMinMax() [all …]
|
/external/webrtc/webrtc/modules/audio_coding/neteq/ |
D | merge.cc | 66 int16_t expanded_max, input_max; in Process() local 69 &expanded_max, &input_max); in Process() 92 expanded_max, input_max, old_length, in Process() 208 int16_t* expanded_max, int16_t* input_max) const { in SignalScaling() 213 *input_max = WebRtcSpl_MaxAbsValueW16(input, mod_input_length); in SignalScaling() 228 WebRtcSpl_NormW32(*input_max * *input_max); in SignalScaling() 310 size_t Merge::CorrelateAndPeakSearch(int16_t expanded_max, int16_t input_max, in CorrelateAndPeakSearch() argument 318 if (expanded_max * input_max > 26843546) { in CorrelateAndPeakSearch()
|
D | merge.h | 76 int16_t* expanded_max, int16_t* input_max) const; 87 size_t CorrelateAndPeakSearch(int16_t expanded_max, int16_t input_max,
|
/external/tensorflow/tensorflow/compiler/tf2xla/kernels/ |
D | fake_quantize_ops.cc | 100 float input_min, input_max; in FakeQuantWithMinMaxArgsOp() local 102 OP_REQUIRES_OK(ctx, ctx->GetAttr("max", &input_max)); in FakeQuantWithMinMaxArgsOp() 103 CpuNudge(input_min, input_max, quant_min_, quant_max_, &nudged_input_min_, in FakeQuantWithMinMaxArgsOp() 148 float input_min, input_max, scale; in FakeQuantWithMinMaxArgsGradOp() local 150 OP_REQUIRES_OK(ctx, ctx->GetAttr("max", &input_max)); in FakeQuantWithMinMaxArgsGradOp() 151 CpuNudge(input_min, input_max, quant_min, quant_max, &nudged_input_min_, in FakeQuantWithMinMaxArgsGradOp() 203 xla::XlaOp input_max = ctx->Input(2); in Compile() local 207 XlaNudge(b, data_type, input_min, input_max, quant_min_, quant_max_, in Compile() 246 xla::XlaOp input_max = ctx->Input(3); in Compile() local 250 XlaNudge(b, data_type, input_min, input_max, quant_min_, quant_max_, in Compile()
|
/external/tensorflow/tensorflow/compiler/tests/ |
D | fake_quant_ops_test.py | 82 def _TestOp(self, input_min, input_max, num_bits, narrow_range, argument 117 max=input_max, 180 def _TestOp(self, input_min, input_max, num_bits, narrow_range, argument 211 max=input_max, 281 def _TestOp(self, input_min, input_max, num_bits, narrow_range, argument 325 max_placeholder: input_max 386 def _TestOp(self, input_min, input_max, num_bits, narrow_range, argument 429 max_placeholder: input_max
|