/external/tensorflow/tensorflow/compiler/mlir/tensorflow/transforms/ |
D | fold_broadcast.cc | 165 std::array<llvm::ArrayRef<int64_t>, 2> operand_shapes; in RewriteOp() local 166 operand_shapes[i] = broadcast_arg_type.getShape(); in RewriteOp() 167 operand_shapes[1 - i] = argument_type.getShape(); in RewriteOp() 172 if (!get_broadcasted_shape(operand_shapes[0], operand_shapes[1], in RewriteOp()
|
/external/tensorflow/tensorflow/compiler/xla/service/cpu/ |
D | mlir_emitter.cc | 92 llvm::ArrayRef<Shape> operand_shapes, llvm::Value *result_ptr, in EmitMlirFuncAndCall() argument 102 for (int i = 0; i != operand_shapes.size(); ++i) { in EmitMlirFuncAndCall() 105 ConvertTensorShapeToMemRefType(operand_shapes[i], mlir_builder)); in EmitMlirFuncAndCall() 135 for (int i = 0; i != operand_shapes.size(); ++i) { in EmitMlirFuncAndCall() 136 BuildViewForBuffer(&op_vals, b, operand_shapes[i], operand_ptrs[i]); in EmitMlirFuncAndCall()
|
D | mlir_emitter.h | 36 llvm::ArrayRef<Shape> operand_shapes, llvm::Value *result_ptr,
|
D | dot_op_emitter.cc | 255 Shape operand_shapes[] = {dot_info_.lhs_shape, dot_info_.rhs_shape}; in EmitLinalgMatmul() local 271 mlir_context_, b_, dot_info_.result_shape, operand_shapes, target_ptr, in EmitLinalgMatmul()
|
/external/tensorflow/tensorflow/compiler/xla/service/ |
D | shape_inference.cc | 1067 std::vector<const Shape*> operand_shapes; in InferVariadicOpShape() local 1068 operand_shapes.reserve(operands.size()); in InferVariadicOpShape() 1070 operand_shapes.push_back(&operand->shape()); in InferVariadicOpShape() 1072 return InferVariadicOpShape(opcode, operand_shapes); in InferVariadicOpShape() 1076 HloOpcode opcode, absl::Span<const Shape* const> operand_shapes) { in InferVariadicOpShape() argument 1077 for (const Shape* shape : operand_shapes) { in InferVariadicOpShape() 1083 result.mutable_tuple_shapes()->reserve(operand_shapes.size()); in InferVariadicOpShape() 1084 for (const Shape* shape : operand_shapes) { in InferVariadicOpShape() 1090 if (operand_shapes.size() == 1) { in InferVariadicOpShape() 1091 return *operand_shapes[0]; in InferVariadicOpShape() [all …]
|
D | hlo_verifier.cc | 115 std::vector<const Shape*> operand_shapes; in HandleConcatenate() local 117 operand_shapes.push_back(&operand->shape()); in HandleConcatenate() 121 operand_shapes, concatenate->concatenate_dimension())); in HandleConcatenate() 284 std::vector<const Shape*> operand_shapes; in HandleAllReduce() local 286 operand_shapes.push_back(&operand->shape()); in HandleAllReduce() 288 return CheckShape(hlo, ShapeInference::InferAllReduceShape(operand_shapes)); in HandleAllReduce() 326 std::vector<const Shape*> operand_shapes; in HandleAllToAll() local 328 operand_shapes.push_back(&operand->shape()); in HandleAllToAll() 331 ShapeInference::InferAllToAllTupleShape(operand_shapes)); in HandleAllToAll() 666 std::vector<const Shape*> operand_shapes; in HandleReduce() local [all …]
|
D | all_reduce_combiner.cc | 62 std::vector<Shape> operand_shapes; in CombineAllReduces() local 75 operand_shapes.push_back(operand->shape()); in CombineAllReduces() 83 ShapeUtil::MakeTupleShape(operand_shapes), operands, reduction, in CombineAllReduces()
|
D | shape_inference.h | 76 HloOpcode opcode, absl::Span<const Shape* const> operand_shapes); 137 absl::Span<const Shape* const> operand_shapes); 148 absl::Span<const Shape* const> operand_shapes);
|
D | hlo_creation_utils.cc | 242 std::vector<const Shape*> operand_shapes; in MakeConcatHlo() local 243 absl::c_transform(operands, std::back_inserter(operand_shapes), in MakeConcatHlo() 247 operand_shapes, dimension)); in MakeConcatHlo() 305 std::vector<const Shape*> operand_shapes; in MakeMapHlo() local 309 operand_shapes.push_back(&operand->shape()); in MakeMapHlo() 317 operand_shapes, map_computation->ComputeProgramShape(), map_dims)); in MakeMapHlo()
|
D | hlo_evaluator.cc | 2410 absl::InlinedVector<const Shape*, 1> operand_shapes; in HandleReduce() local 2412 operand_shapes.push_back(&operand->shape()); in HandleReduce() 2416 operand_shapes, dimensions_to_reduce, in HandleReduce()
|
D | hlo_instruction.cc | 554 std::vector<Shape> operand_shapes; in CreateFromProto() local 557 operand_shapes.emplace_back(shape_proto); in CreateFromProto() 561 operand_shapes, proto.backend_config()); in CreateFromProto()
|
/external/tensorflow/tensorflow/compiler/xla/service/gpu/ |
D | gpu_autotuning.proto | 13 repeated xla.ShapeProto operand_shapes = 2; field
|
D | ir_emitter_unnested.cc | 1684 std::vector<Shape>* operand_shapes, in ProcessFusionForConversion() argument 1715 operand_shapes->push_back(std::move(shape)); in ProcessFusionForConversion() 1780 absl::MakeSpan(ir_arrays).subspan(0, context.operand_shapes.size()); in EmitLoopFusionFromMlir() 1782 context.operand_shapes.size(), context.output_shapes.size()); in EmitLoopFusionFromMlir() 1785 context.operand_shapes.size() + context.output_shapes.size() + 1) { in EmitLoopFusionFromMlir() 1786 tuple_output_array = &ir_arrays[context.operand_shapes.size() + in EmitLoopFusionFromMlir() 1798 for (int i = 0; i < context.operand_shapes.size(); i++) { in EmitLoopFusionFromMlir() 2739 std::vector<Shape> operand_shapes, output_shapes; in GetOrCreateSubComputationFromRegion() local 2744 ProcessFusionForConversion(region, &operand_shapes, &output_shapes)); in GetOrCreateSubComputationFromRegion() 2765 CHECK_EQ(operand_shapes.size(), fused_computation->num_parameters()); in GetOrCreateSubComputationFromRegion() [all …]
|
D | ir_emitter_unnested.h | 81 std::vector<Shape> operand_shapes; member
|
/external/tensorflow/tensorflow/compiler/xla/client/ |
D | xla_builder.cc | 235 std::vector<Shape> operand_shapes; in GetOperandShapes() local 238 operand_shapes.push_back(*shape); in GetOperandShapes() 240 return operand_shapes; in GetOperandShapes() 828 TF_ASSIGN_OR_RETURN(const auto& operand_shapes, GetOperandShapes(operands)); in Call() 829 absl::c_transform(operand_shapes, std::back_inserter(operand_shape_ptrs), in Call() 1061 TF_ASSIGN_OR_RETURN(const auto& operand_shapes, GetOperandShapes(operands)); in ConcatInDim() 1062 absl::c_transform(operand_shapes, std::back_inserter(operand_shape_ptrs), in ConcatInDim() 1239 TF_ASSIGN_OR_RETURN(const auto& operand_shapes, GetOperandShapes(elements)); in Tuple() 1240 absl::c_transform(operand_shapes, std::back_inserter(operand_shape_ptrs), in Tuple() 2064 TF_ASSIGN_OR_RETURN(std::vector<Shape> operand_shapes, in Sort() [all …]
|
/external/tensorflow/tensorflow/compiler/xla/service/spmd/ |
D | spmd_partitioner.cc | 2184 std::vector<Shape> operand_shapes; in HandleSingleDevice() local 2187 operand_shapes.push_back(operand->shape()); in HandleSingleDevice() 2190 auto operand_shape = ShapeUtil::MakeTupleShape(operand_shapes); in HandleSingleDevice() 2206 HloInstruction::CreateGetTupleElement(operand_shapes[i], param, i))); in HandleSingleDevice()
|