/external/tensorflow/tensorflow/lite/delegates/gpu/common/tasks/ |
D | resize.cc | 43 if (op_def.IsBatchSupported()) { in GetResizeCode() 48 if (op_def.IsBatchSupported()) { in GetResizeCode() 61 if (op_def.IsBatchSupported()) { in GetResizeCode() 93 if (op_def.IsBatchSupported()) { in GetResizeCode() 117 if (op_def.IsBatchSupported()) { in GetResizeCode() 178 if (op_def.IsBatchSupported()) { in GetResize3DCode() 183 if (op_def.IsBatchSupported()) { in GetResize3DCode() 200 if (op_def.IsBatchSupported()) { in GetResize3DCode() 239 if (op_def.IsBatchSupported()) { in GetResize3DCode() 255 if (op_def.IsBatchSupported()) { in GetResize3DCode()
|
D | pooling.cc | 32 if (op_def.IsBatchSupported()) { in GetAveragePoolingKernelCode() 37 if (op_def.IsBatchSupported()) { in GetAveragePoolingKernelCode() 98 if (op_def.IsBatchSupported()) { in GetAveragePoolingKernelCode() 116 if (op_def.IsBatchSupported()) { in GetAveragePoolingKernelCode() 149 if (op_def.IsBatchSupported()) { in GetMaxPoolingKernelCode() 154 if (op_def.IsBatchSupported()) { in GetMaxPoolingKernelCode() 160 if (op_def.IsBatchSupported()) { in GetMaxPoolingKernelCode() 220 if (op_def.IsBatchSupported()) { in GetMaxPoolingKernelCode() 232 if (op_def.IsBatchSupported()) { in GetMaxPoolingKernelCode() 298 definition.IsBatchSupported() && attr.strides.w != 1; in CreatePooling() [all …]
|
D | convolution_transposed_4x4.cc | 76 if (op_def.IsBatchSupported()) { in GenerateConvolutionTransposedCode() 82 if (op_def.IsBatchSupported()) { in GenerateConvolutionTransposedCode() 149 op_def.IsBatchSupported() ? "args.dst_tensor.Batch()" : "1"; in GenerateConvolutionTransposedCode() 177 if (op_def.IsBatchSupported()) { in GenerateConvolutionTransposedCode() 186 if (op_def.IsBatchSupported()) { in GenerateConvolutionTransposedCode() 335 if (op_def.IsBatchSupported()) { in GenerateConvolutionTransposedCode() 345 if (op_def.IsBatchSupported()) { in GenerateConvolutionTransposedCode()
|
D | depthwise_conv.cc | 74 if (op_def.IsBatchSupported()) { in GenerateDepthwiseConvolutionCode() 83 if (op_def.IsBatchSupported()) { in GenerateDepthwiseConvolutionCode() 112 if (op_def.IsBatchSupported()) { in GenerateDepthwiseConvolutionCode() 189 op_def.IsBatchSupported() ? "args.dilation_x * args.src_tensor.Batch()" in GenerateDepthwiseConvolutionCode() 254 definition.IsBatchSupported() && attr.strides.w != 1; in CreateDepthwiseConvolution2D() 283 definition.IsBatchSupported() && attr.strides.w != 1; in CreateDepthwiseConvolution2DDynamicWeights() 322 definition.IsBatchSupported() && attr.strides.w != 1; in CreateDepthwiseConvolution3D()
|
D | softmax.cc | 69 if (definition.IsBatchSupported()) { in CreateSoftmax() 74 if (definition.IsBatchSupported()) { in CreateSoftmax()
|
D | conv_constants.cc | 104 if (op_def.IsBatchSupported()) { in GenerateConvolutionConstantCode() 110 if (op_def.IsBatchSupported()) { in GenerateConvolutionConstantCode() 134 if (op_def.IsBatchSupported()) { in GenerateConvolutionConstantCode() 171 op_def.IsBatchSupported() ? "args.dilation_x * args.src_tensor.Batch()" in GenerateConvolutionConstantCode() 283 definition.IsBatchSupported() && attr.strides.w != 1; in CreateConvConstants()
|
D | max_unpooling.cc | 30 if (op_def.IsBatchSupported()) { in GetMaxUnpoolingKernelCode() 36 if (op_def.IsBatchSupported()) { in GetMaxUnpoolingKernelCode() 41 if (op_def.IsBatchSupported()) { in GetMaxUnpoolingKernelCode()
|
D | concat_z.cc | 129 if (definition.IsBatchSupported()) { in CreateConcatZ() 135 if (definition.IsBatchSupported()) { in CreateConcatZ()
|
D | convolution_transposed_3x3.cc | 61 if (op_def.IsBatchSupported()) { in GenerateConvolutionTransposedCode() 67 if (op_def.IsBatchSupported()) { in GenerateConvolutionTransposedCode() 135 op_def.IsBatchSupported() ? "args.dst_tensor.Batch()" : "1"; in GenerateConvolutionTransposedCode() 154 if (op_def.IsBatchSupported()) { in GenerateConvolutionTransposedCode()
|
D | add.cc | 39 if (definition.IsBatchSupported()) { in CreateAdd()
|
D | space_to_depth.cc | 31 if (op_def.IsBatchSupported()) { in GetSpaceToDepthCode()
|
D | softmax1x1.cc | 51 if (op_def.IsBatchSupported()) { in GetSoftmaxKernelCode()
|
D | conv_metal.cc | 997 definition.IsBatchSupported() && attr.strides.w != 1; in CreateConvolutionMetal() 1001 if (definition.IsBatchSupported()) { in CreateConvolutionMetal() 1006 if (definition.IsBatchSupported()) { in CreateConvolutionMetal() 1126 if (definition.IsBatchSupported()) { in CreateConvolutionMetalWino4x4To6x6() 1131 if (definition.IsBatchSupported()) { in CreateConvolutionMetalWino4x4To6x6()
|
D | convolution_transposed_thin.cc | 79 if (op_def.IsBatchSupported()) { in GenerateConvolutionTransposedCode()
|
D | conv_buffer_1x1.cc | 171 if (op_def.IsBatchSupported()) { in GenerateConvBuffer1x1() 190 if (op_def.IsBatchSupported()) { in GenerateConvBuffer1x1()
|
D | convolution_transposed_3x3_thin.cc | 87 if (op_def.IsBatchSupported()) { in GenerateConvolutionTransposedCode()
|
D | elementwise.cc | 309 if (definition.IsBatchSupported()) { in CreateElementwiseTwoInput()
|
D | conv_powervr.cc | 248 definition_.IsBatchSupported() && stride_.x != 1; in GenerateCode() 343 if (op_def.IsBatchSupported()) { in GenerateConv() 405 if (op_def.IsBatchSupported()) { in GenerateConv()
|
D | convolution_transposed.cc | 247 if (op_def.IsBatchSupported()) { in GenerateConvolutionTransposedCode()
|
/external/tensorflow/tensorflow/lite/delegates/gpu/common/selectors/default/ |
D | fully_connected_selector.cc | 30 if (op_def.IsBatchSupported()) { in SelectFullyConnectedGeneric() 43 if (op_def.IsBatchSupported()) { in SelectFullyConnectedAdreno() 56 if (op_def.IsBatchSupported()) { in SelectFullyConnectedPowerVR() 68 if (op_def.IsBatchSupported()) { in SelectFullyConnectedMali() 88 if (op_def.IsBatchSupported() && IsConvolutionMetalSupported(op_def)) { in SelectFullyConnected()
|
/external/tensorflow/tensorflow/lite/delegates/gpu/cl/kernels/ |
D | cl_test.cc | 61 if (src_shape.b != 1 && !op_def.IsBatchSupported()) { in ExecuteGPUOperation() 74 if (dst_shape.b != 1 && !op_def.IsBatchSupported()) { in ExecuteGPUOperation() 109 if (src_shape.b != 1 && !op_def.IsBatchSupported()) { in ExecuteGPUOperation() 122 if (dst_shape.b != 1 && !op_def.IsBatchSupported()) { in ExecuteGPUOperation()
|
/external/tensorflow/tensorflow/lite/delegates/gpu/metal/kernels/ |
D | test_util.cc | 72 if (src_shape.b != 1 && !op_def.IsBatchSupported()) { in ExecuteGPUOperation() 84 if (dst_shape.b != 1 && !op_def.IsBatchSupported()) { in ExecuteGPUOperation()
|
/external/tensorflow/tensorflow/lite/delegates/gpu/common/task/ |
D | gpu_operation.cc | 61 bool OperationDef::IsBatchSupported() const { in IsBatchSupported() function in tflite::gpu::OperationDef 190 if (definition_.IsBatchSupported()) { in AssembleCode() 198 if (definition_.IsBatchSupported()) { in AssembleCode()
|
D | gpu_operation.h | 84 bool IsBatchSupported() const;
|