Home
last modified time | relevance | path

Searched refs:col_ctx_ (Results 1 – 13 of 13) sorted by relevance

/external/tensorflow/tensorflow/core/common_runtime/
Dpermuter.cc40 : col_ctx_(nullptr), col_params_(nullptr), done_(nullptr), counter_(0) {} in Permuter()
56 col_ctx_ = col_ctx; in InitializeCollectiveContext()
71 col_ctx_->input, CheckCounterAndCallDone()); in Run()
74 DispatchRecv(i, col_params_->instance.permutation[i], col_ctx_->output, in Run()
83 strings::StrCat(col_ctx_->exec_key, src_rank, target_rank); in DispatchSend()
85 << col_ctx_->device_name << " to_device " in DispatchSend()
88 col_ctx_->col_exec->remote_access()->PostToPeer( in DispatchSend()
91 col_ctx_->device, col_ctx_->op_ctx->op_device_context(), in DispatchSend()
92 col_ctx_->op_ctx->output_alloc_attr(0), tensor, col_ctx_->device_locality, in DispatchSend()
93 col_ctx_->op_ctx->cancellation_manager(), done); in DispatchSend()
[all …]
Dring_reducer.cc56 CHECK(col_ctx_); in Run()
60 col_ctx_->col_exec->UnblockDependencies(*col_params_); in Run()
83 VLOG(1) << "RingReducer::Run for device " << col_ctx_->device_name in Run()
90 if ((col_ctx_->input != col_ctx_->output) && in Run()
91 (DMAHelper::base(col_ctx_->input) != DMAHelper::base(col_ctx_->output))) { in Run()
98 col_ctx_->op_ctx->op_device_context(), in Run()
99 col_ctx_->op_ctx->op_device_context(), col_ctx_->device, in Run()
100 col_ctx_->device, col_ctx_->op_ctx->input_alloc_attr(0), in Run()
101 col_ctx_->op_ctx->output_alloc_attr(0), col_ctx_->input, in Run()
102 col_ctx_->output, 0 /*dev_to_dev_stream_index*/, in Run()
[all …]
Dhierarchical_tree_broadcaster.cc57 : col_ctx_(nullptr), in HierarchicalTreeBroadcaster()
190 col_ctx_ = col_ctx; in InitializeCollectiveContext()
198 CHECK(col_ctx_); in Run()
307 VLOG(1) << "Running Broadcast tree device=" << col_ctx_->device_name in RunTree()
323 DispatchRecv(si, recv_from_rank, my_rank, col_ctx_->output, in RunTree()
347 (is_source_ ? col_ctx_->input : col_ctx_->output), in RunTree()
368 << col_ctx_->device_name << " subdiv=" << si; in RunTree()
369 if (col_ctx_->input != col_ctx_->output && in RunTree()
370 (DMAHelper::base(col_ctx_->input) != in RunTree()
371 DMAHelper::base(col_ctx_->output))) { in RunTree()
[all …]
Dring_gatherer.cc64 DCHECK(col_ctx_); in Run()
87 VLOG(1) << "RingGatherer::Run for device " << col_ctx_->device_name in Run()
93 AllocatorAttributes attr = col_ctx_->op_ctx->output_alloc_attr(0); in Run()
94 ca_.reset(MakeCollectiveAdapter(col_ctx_->output, group_size_ * num_subdivs_, in Run()
95 col_ctx_->device->GetAllocator(attr), in Run()
107 col_ctx_->op_ctx->op_device_context(), in Run()
108 col_ctx_->op_ctx->op_device_context(), col_ctx_->device, in Run()
109 col_ctx_->device, col_ctx_->op_ctx->input_alloc_attr(0), in Run()
110 col_ctx_->op_ctx->output_alloc_attr(0), col_ctx_->input, &alias_chunk, in Run()
141 col_ctx_->device->tensorflow_gpu_device_info(); in RunAsyncParts()
[all …]
Dring_alg.cc102 col_ctx_(nullptr), in RingAlg()
247 col_ctx_ = col_ctx; in InitializeCollectiveContext()
256 col_ctx_->op_ctx->device()->tensorflow_gpu_device_info(); in TensorDebugString()
260 &tensor, "" /*tensor_name*/, col_ctx_->device, &cpu_tensor); in TensorDebugString()
287 if (col_ctx_->op_ctx->cancellation_manager() == nullptr || in StartAbort()
288 (!col_ctx_->op_ctx->cancellation_manager()->IsCancelled() && in StartAbort()
289 !col_ctx_->op_ctx->cancellation_manager()->IsCancelling())) { in StartAbort()
290 col_ctx_->col_exec->StartAbort(s); in StartAbort()
298 ca_->ConsumeFinalValue(col_ctx_->output); in Finish()
384 string send_buf_key = RingAlgBufKey(name_, col_ctx_->exec_key, in DispatchSend()
[all …]
Dpermuter.h67 std::shared_ptr<CollectiveContext> col_ctx_;
Dhierarchical_tree_broadcaster.h84 std::shared_ptr<CollectiveContext> col_ctx_; variable
Dring_alg.h112 std::shared_ptr<CollectiveContext> col_ctx_; variable
/external/tensorflow/tensorflow/core/kernels/
Dcollective_nccl_reducer.cc37 switch (col_ctx_->output->dtype()) { in Run()
60 DataTypeString(col_ctx_->output->dtype()))); in Run()
64 col_ctx_->device->GetAllocator(col_ctx_->op_ctx->input_alloc_attr(0)), in Run()
65 col_ctx_->output->dtype(), TensorShape({})); in Run()
66 DeviceContext* op_dev_ctx = col_ctx_->op_ctx->op_device_context(); in Run()
70 &group_size_val, col_ctx_->device, &group_size, in Run()
94 col_ctx_->nccl_communicator->Enqueue(col_ctx_, std::move(done_callback)); in Run()
118 col_ctx_->op_ctx, col_ctx_->op_params, col_ctx_->device, in Run()
119 col_params_->final_op, col_ctx_->output, &group_size); in Run()
Dcollective_nccl_broadcaster.cc27 col_ctx_->nccl_communicator->Enqueue(col_ctx_, std::move(done)); in Run()
Dcollective_nccl_gatherer.cc27 col_ctx_->nccl_communicator->Enqueue(col_ctx_, std::move(done)); in Run()
Dcollective_nccl.cc27 : type_(type), name_(name), col_ctx_(nullptr), col_params_(nullptr) {} in NcclBase()
63 col_ctx_ = col_ctx; in InitializeCollectiveContext()
Dcollective_nccl.h42 std::shared_ptr<CollectiveContext> col_ctx_; variable