/external/tensorflow/tensorflow/compiler/xla/service/ |
D | sharding_propagation_test.cc | 162 EXPECT_THAT(instruction, op::Sharding("{devices=[1,2,2,1]0,1,2,3}")); in TEST_P() 190 EXPECT_THAT(instruction, op::Sharding("{devices=[1,2,2,1]0,1,2,3}")); in TEST_F() 213 EXPECT_THAT(instruction, op::Sharding("{devices=[1,2,2,1]0,1,2,3}")); in TEST_F() 236 EXPECT_THAT(instruction, op::Sharding("{devices=[1,2,2,1]0,1,2,3}")); in TEST_F() 262 EXPECT_THAT(instruction, op::Sharding("{devices=[1,2,2,1]0,1,2,3}")); in TEST_P() 293 EXPECT_THAT(instruction, op::Sharding("{devices=[1,2,2,1]0,1,2,3}")); in TEST_P() 324 EXPECT_THAT(instruction, op::Sharding("{devices=[1,2,2,1]0,1,2,3}")); in TEST_P() 354 EXPECT_THAT(instruction, op::Sharding("{devices=[1,2,2,1]0,1,2,3}")); in TEST_P() 386 op::Sharding("{devices=[1,2,1,2]0,1,2,3 last_tile_dim_replicate}")); in TEST_P() 417 EXPECT_THAT(instruction, op::Sharding("{devices=[1,2,2]0,1,2,3}")); in TEST_P() [all …]
|
D | hlo_matchers_test.cc | 181 ::testing::Not(op::Sharding(HloSharding::AssignDevice(1)))); in TEST_F() 184 ::testing::Not(op::Sharding(HloSharding::AssignDevice(0)))); in TEST_F() 185 EXPECT_THAT(p1.get(), op::Sharding(HloSharding::AssignDevice(1))); in TEST_F() 189 op::Sharding("{{devices=[2]0,1}, {maximal device=1}, {replicated}}")); in TEST_F() 191 EXPECT_THAT(Explain(p0.get(), op::Sharding(HloSharding::AssignDevice(1))), in TEST_F() 197 EXPECT_THAT(Explain(p1.get(), op::Sharding(HloSharding::AssignDevice(0))), in TEST_F()
|
D | hlo_matchers.h | 403 inline ::testing::Matcher<const ::xla::HloInstruction*> Sharding( in Sharding() function 409 inline ::testing::Matcher<const ::xla::HloInstruction*> Sharding( in Sharding() function
|
D | hlo.proto | 202 // Sharding for kDomain instructions.
|
/external/tensorflow/tensorflow/compiler/xla/experimental/xla_sharding/ |
D | xla_sharding.py | 28 class Sharding(object): class 46 return Sharding( 56 return Sharding( 68 return Sharding( 96 return Sharding( 121 return Sharding( 158 return Sharding( 262 return Sharding.replicate().apply_to_tensor( 273 return Sharding.assign_device(device).apply_to_tensor( 292 return Sharding.tile(tile_assignment).apply_to_tensor( [all …]
|
/external/tensorflow/tensorflow/core/protobuf/tpu/ |
D | compile_metadata.proto | 40 // Sharding is allowed if host training loop exists.
|
D | tpu_embedding_configuration.proto | 45 // Sharding strategy of the embedding tables among the hosts.
|
/external/llvm-project/clang-tools-extra/clangd/unittests/ |
D | FileIndexTests.cpp | 578 TEST(FileShardedIndexTest, Sharding) { in TEST() argument
|
/external/tensorflow/tensorflow/compiler/mlir/xla/tests/translate/ |
D | export.mlir | 1048 …%0 = "mhlo.custom_call"(%arg0) {backend_config = "", call_target_name = "Sharding", mhlo.sharding … 1054 …*]] = f32[16,16] custom-call(f32[16,16] %[[ARG0]]), custom_call_target="Sharding", sharding={devic…
|
/external/tensorflow/tensorflow/compiler/mlir/xla/tests/ |
D | legalize-tf.mlir | 4813 …// CHECK-NEXT: "mhlo.custom_call"(%arg0) {backend_config = "", call_target_name = "Sharding", has_…
|