/external/tensorflow/tensorflow/compiler/tf2xla/ |
D | shape_util.h | 39 Status TensorShapeToXLAShape(DataType dtype, const TensorShape& tensor_shape, 44 xla::Shape TensorShapeToXLAShape(xla::PrimitiveType type, 49 Status TensorShapeToXLAShape(DataType dtype, 55 xla::Shape TensorShapeToXLAShape(xla::PrimitiveType type,
|
D | shape_util.cc | 103 Status TensorShapeToXLAShape(DataType dtype, in TensorShapeToXLAShape() function 108 *shape = TensorShapeToXLAShape(type, tensor_shape); in TensorShapeToXLAShape() 112 xla::Shape TensorShapeToXLAShape(xla::PrimitiveType type, in TensorShapeToXLAShape() function 145 Status TensorShapeToXLAShape(DataType dtype, const TensorShape& tensor_shape, in TensorShapeToXLAShape() function 149 *shape = TensorShapeToXLAShape(type, tensor_shape); in TensorShapeToXLAShape() 153 xla::Shape TensorShapeToXLAShape(xla::PrimitiveType type, in TensorShapeToXLAShape() function
|
D | literal_util.cc | 28 TF_RETURN_IF_ERROR(TensorShapeToXLAShape(host_tensor.dtype(), in HostTensorToBorrowingLiteral() 55 TF_RETURN_IF_ERROR(TensorShapeToXLAShape(host_tensor->dtype(), in HostTensorToMutableBorrowingLiteral() 79 TF_RETURN_IF_ERROR(TensorShapeToXLAShape(tensor->dtype(), tensor->shape(), in HostTensorsToBorrowingLiteralTuple()
|
D | xla_helpers.cc | 96 TensorShapeToXLAShape(index_type, output_shape, &iota_shape)); in OneHot() 133 TF_RETURN_IF_ERROR(TensorShapeToXLAShape(dtype, shape, &xla_shape)); in IdentityShapeRepresentationFn()
|
D | xla_compiler_test.cc | 310 TF_RETURN_IF_ERROR(TensorShapeToXLAShape(dt, shape, &xla_shape)); in TEST_F() 353 TF_RETURN_IF_ERROR(TensorShapeToXLAShape(dt, shape, &xla_shape)); in TEST_F() 406 TF_RETURN_IF_ERROR(TensorShapeToXLAShape(dt, shape, &xla_shape)); in TEST_F() 1102 TF_RETURN_IF_ERROR(TensorShapeToXLAShape(type, shape, &xla_shape)); in TEST_F() 1143 TF_RETURN_IF_ERROR(TensorShapeToXLAShape(type, shape, &xla_shape)); in TEST_F() 1616 TF_ASSERT_OK(TensorShapeToXLAShape(DT_INT32, TensorShape{1}, in TEST_F() 1619 TF_ASSERT_OK(TensorShapeToXLAShape(DT_INT32, TensorShape{}, &index_shape)); in TEST_F()
|
/external/tensorflow/tensorflow/core/tpu/kernels/ |
D | image_resize_ops.cc | 44 TensorShapeToXLAShape(ctx->output_xla_type(0), ctx->InputShape(0)); in GetOutputShape() 57 TensorShapeToXLAShape(ctx->output_xla_type(0), ctx->InputShape(0)); in CompileGrad() 132 TensorShapeToXLAShape(ctx->output_xla_type(0), ctx->InputShape(1)); in Compile()
|
D | outfeed_ops.cc | 38 OP_REQUIRES_OK(ctx, TensorShapeToXLAShape(dtype_, shape_, &xla_shape_)); in TpuOutfeedDequeueOp() 79 TensorShapeToXLAShape(dtypes_[i], shapes_[i], &xla_shape)); in TpuOutfeedDequeueTupleOp()
|
D | infeed_ops.cc | 245 OP_REQUIRES_OK(ctx, TensorShapeToXLAShape(dtype_, shape_, &shape)); in PrelinearizeOp() 309 TensorShapeToXLAShape(dtypes_[i], shapes_[i], &xla_shape)); in PrelinearizeTupleOp() 412 OP_REQUIRES_OK(ctx, TensorShapeToXLAShape(dtype_, shape_, &shape)); in TpuInfeedEnqueueOp() 462 TensorShapeToXLAShape(dtypes_[i], shapes_[i], &xla_shape)); in TpuInfeedEnqueueTupleOp()
|
/external/tensorflow/tensorflow/compiler/tf2xla/kernels/ |
D | random_ops.cc | 50 OP_REQUIRES_OK(ctx, TensorShapeToXLAShape(dtype, shape, &xla_shape)); in Compile() 228 TensorShapeToXLAShape(input_type(1), shape, &xla_shape)); in Compile() 266 OP_REQUIRES_OK(ctx, TensorShapeToXLAShape(dtype, shape, &xla_shape)); in Compile() 294 OP_REQUIRES_OK(ctx, TensorShapeToXLAShape(dtype, shape, &xla_shape)); in Compile() 327 OP_REQUIRES_OK(ctx, TensorShapeToXLAShape(dtype, shape, &xla_shape)); in Compile()
|
D | stateless_random_ops.cc | 163 OP_REQUIRES_OK(ctx, TensorShapeToXLAShape(rng_dtype, shape, &xla_shape)); in Compile() 218 OP_REQUIRES_OK(ctx, TensorShapeToXLAShape(dtype_, shape, &xla_shape)); in Compile() 259 OP_REQUIRES_OK(ctx, TensorShapeToXLAShape(dtype_, shape, &xla_shape)); in Compile() 300 OP_REQUIRES_OK(ctx, TensorShapeToXLAShape(rng_dtype, shape, &xla_shape)); in Compile() 352 OP_REQUIRES_OK(ctx, TensorShapeToXLAShape(rng_dtype, shape, &xla_shape)); in Compile()
|
D | stateful_random_ops.cc | 229 TF_RETURN_IF_ERROR(TensorShapeToXLAShape(rng_dtype, shape, &xla_shape)); in Compile() 273 TF_RETURN_IF_ERROR(TensorShapeToXLAShape(rng_dtype, shape, &xla_shape)); in Compile() 313 TF_RETURN_IF_ERROR(TensorShapeToXLAShape(rng_dtype, shape, &xla_shape)); in Compile() 357 TF_RETURN_IF_ERROR(TensorShapeToXLAShape(dtype_, shape, &xla_shape)); in Compile() 390 TF_RETURN_IF_ERROR(TensorShapeToXLAShape(dtype_, shape, &xla_shape)); in Compile()
|
D | conv_ops.cc | 109 TensorShapeToXLAShape(ctx->input_xla_type(1), input_tensor_shape); in Compile() 175 TensorShapeToXLAShape(ctx->input_xla_type(0), filter_tensor_shape); in Compile()
|
D | fake_param_op.cc | 37 OP_REQUIRES_OK(ctx, TensorShapeToXLAShape(dtype, tensor_shape, &shape_)); in XlaFakeParamOp()
|
D | stateless_random_ops_v2.cc | 224 OP_REQUIRES_OK(ctx, TensorShapeToXLAShape(rng_dtype, shape, &xla_shape)); in Compile() 290 OP_REQUIRES_OK(ctx, TensorShapeToXLAShape(dtype_, shape, &xla_shape)); in Compile() 337 OP_REQUIRES_OK(ctx, TensorShapeToXLAShape(dtype_, shape, &xla_shape)); in Compile() 385 OP_REQUIRES_OK(ctx, TensorShapeToXLAShape(rng_dtype, shape, &xla_shape)); in Compile() 440 OP_REQUIRES_OK(ctx, TensorShapeToXLAShape(rng_dtype, shape, &xla_shape)); in Compile()
|
D | sendrecv_ops.cc | 72 OP_REQUIRES_OK(ctx, TensorShapeToXLAShape(dtype, tensor_shape, &shape_)); in RecvOp()
|
D | spmd_manual_sharding_ops.cc | 102 auto output_shape = TensorShapeToXLAShape( in Compile()
|
D | reduce_window_op.cc | 93 TensorShapeToXLAShape(dtype, TensorShape(), &scalar_shape)); in Compile()
|
D | sort_ops.cc | 102 TensorShapeToXLAShape(DT_BOOL, TensorShape(), in Compile()
|
D | xla_reduce_op.cc | 91 context, TensorShapeToXLAShape(dtype, TensorShape(), &expected_shape)); in Compile()
|
D | xla_select_and_scatter_op.cc | 100 TensorShapeToXLAShape(dtype, TensorShape(), &scalar_shape)); in Compile()
|
D | cast_op.cc | 176 TensorShapeToXLAShape(dst_type_, input_shape); in Compile()
|
/external/tensorflow/tensorflow/core/tpu/kernels/xla/ |
D | outfeed_ops.cc | 39 ctx, TensorShapeToXLAShape(dtype_, ctx->InputShape(0), &xla_shape)); in Compile() 69 TensorShapeToXLAShape(dtypes_[i], shapes[i], &xla_shape)); in Compile()
|
D | infeed_op.cc | 93 OP_REQUIRES_OK(ctx, TensorShapeToXLAShape(dtype_, shape_, &xla_shape_)); in InfeedDequeueOp() 121 TensorShapeToXLAShape(dtypes_[i], shapes_[i], &xla_shape)); in InfeedDequeueTupleOp()
|
D | host_compute_ops.cc | 56 TensorShapeToXLAShape(dtypes[i], shapes[i], &single_xla_shape)); in MakeXlaShapes() 152 OP_REQUIRES_OK(ctx, TensorShapeToXLAShape(input_dtypes_[i], in Compile() 418 OP_REQUIRES_OK(ctx, TensorShapeToXLAShape(input_dtype_, ctx->InputShape(0), in Compile() 471 ctx, TensorShapeToXLAShape(output_dtype_, output_shape_, &xla_shape)); in Compile()
|
/external/tensorflow/tensorflow/compiler/jit/ |
D | xla_device_context.cc | 100 TF_RETURN_IF_ERROR(TensorShapeToXLAShape(dtype, shape, &xla_shape)); in XlaDeviceContext()
|