/external/tensorflow/tensorflow/core/kernels/ |
D | cuda_solvers.h | 167 const string& debug_info, bool on_host); 170 bool on_host); 336 ScratchSpace(OpKernelContext* context, int64 size, bool on_host) 337 : ScratchSpace(context, TensorShape({size}), "", on_host) {} 340 bool on_host) 341 : ScratchSpace(context, TensorShape({size}), debug_info, on_host) {} 344 const string& debug_info, bool on_host) 345 : context_(context), debug_info_(debug_info), on_host_(on_host) { 347 if (on_host) { 379 bool on_host() const { return on_host_; } [all …]
|
D | ops_testutil.h | 57 const bool on_host = in SetOutputAttrs() local 59 attr.set_on_host(on_host); in SetOutputAttrs()
|
D | debug_ops.h | 76 !context->input_alloc_attr(0).on_host(); in Compute() 94 !context->input_alloc_attr(0).on_host(); in Compute()
|
/external/tensorflow/tensorflow/contrib/gdr/ |
D | gdr_worker.cc | 91 const bool on_host = send_args.alloc_attrs.on_host(); in GrpcRecvTensorAsync() local 104 on_host, [proto, done, response](const Status& s) { in GrpcRecvTensorAsync() 116 if (src_dev->tensorflow_gpu_device_info() && (!on_host)) { in GrpcRecvTensorAsync() 177 hook->prod_dev, hook->prod_ctx, hook->prod_attr.on_host(), in RecvBufAsync()
|
D | gdr_memory_manager.h | 44 Device* device, DeviceContext* device_context, bool on_host, 51 Device* device, DeviceContext* device_context, bool on_host,
|
D | gdr_rendezvous_mgr.cc | 63 const bool on_host = recv_args_.alloc_attrs.on_host(); in Start() local 66 recv_args_.device_context, on_host, in Start()
|
D | gdr_memory_manager.cc | 123 Device* device, DeviceContext* device_context, bool on_host, 128 Device* device, DeviceContext* device_context, bool on_host, 367 Device* device, DeviceContext* device_context, bool on_host, in TransportOptionsFromTensor() argument 406 if (copy && device->tensorflow_gpu_device_info() && !on_host) { in TransportOptionsFromTensor() 425 Device* device, DeviceContext* device_context, bool on_host, in TensorFromTransportOptions() argument 478 StatusCallback callback = [done, copy, device, device_context, on_host, in TensorFromTransportOptions() 492 if (copy && device->tensorflow_gpu_device_info() && !on_host) { in TensorFromTransportOptions()
|
D | gdr_collective_executor_mgr.cc | 107 to_device_ctx, to_alloc_attr.on_host(), done); in RecvFromPeer()
|
/external/tensorflow/tensorflow/core/framework/ |
D | allocator_test.cc | 49 for (bool on_host : {false, true}) { in TEST() 53 aa.set_on_host(on_host); in TEST() 56 EXPECT_EQ(on_host, aa.on_host()); in TEST()
|
D | allocator.h | 361 bool on_host() const { return value & 0x1; } in on_host() function
|
/external/tensorflow/tensorflow/core/grappler/optimizers/ |
D | gpu_swapping_kernels.cc | 32 ctx, !ctx->input_alloc_attr(0).on_host(), in ComputeAsync() 65 ctx, ctx->input_alloc_attr(0).on_host(), in ComputeAsync()
|
/external/tensorflow/tensorflow/core/common_runtime/sycl/ |
D | sycl_device.cc | 39 if (attr.on_host()) in GetAllocator() 58 if (alloc_attrs.on_host()) { in MakeTensorFromProto()
|
/external/tensorflow/tensorflow/core/distributed_runtime/rpc/ |
D | grpc_worker_service.cc | 491 const bool on_host = send_args.alloc_attrs.on_host(); in GrpcRecvTensorAsync() local 494 if (src_dev->tensorflow_gpu_device_info() && (!on_host)) { in GrpcRecvTensorAsync() 583 const bool on_host = in RecvBufAsync() local 585 hook->prod_attr.on_host(); in RecvBufAsync() 586 if ((!on_host) && (num_bytes > 0)) { in RecvBufAsync()
|
/external/tensorflow/tensorflow/core/common_runtime/ |
D | collective_rma_local.cc | 95 src_attr.on_host() ? DEVICE_CPU : src_dev->attributes().device_type()); in MemCpyAsync() 97 dst_attr.on_host() ? DEVICE_CPU : dst_dev->attributes().device_type()); in MemCpyAsync()
|
D | rendezvous_mgr.cc | 69 (send_args.alloc_attrs.on_host() || parsed.src.type == "CPU"); in SameWorkerRecvDone() 71 (recv_args.alloc_attrs.on_host() || parsed.dst.type == "CPU"); in SameWorkerRecvDone()
|
D | copy_tensor.cc | 259 src_alloc_attr.on_host() ? DEVICE_CPU : src->attributes().device_type()); in ViaDMA() 261 dst_alloc_attr.on_host() ? DEVICE_CPU : dst->attributes().device_type()); in ViaDMA()
|
D | partitioning_utils_test.cc | 166 ASSERT_EQ(expected[i], actual[i].on_host()) << " at index " << i; in CheckAlloc()
|
/external/tensorflow/tensorflow/core/kernels/data/ |
D | single_threaded_executor.cc | 129 bool on_host = op_kernel->output_memory_types()[out] == HOST_MEMORY; in Initialize() local 130 if (on_host) { in Initialize() 132 h.set_on_host(on_host); in Initialize()
|
D | dataset_test_base.cc | 201 const bool on_host = in CreateOpKernelContext() local 203 attr.set_on_host(on_host); in CreateOpKernelContext()
|
/external/tensorflow/tensorflow/core/common_runtime/gpu/ |
D | gpu_device_factory.cc | 45 if (attr.on_host()) { in GetAllocator()
|
/external/tensorflow/tensorflow/core/distributed_runtime/ |
D | base_rendezvous_mgr.cc | 236 (send_args.alloc_attrs.on_host() || parsed.src.type == "CPU"); in SameWorkerRecvDone() 238 (recv_args.alloc_attrs.on_host() || parsed.dst.type == "CPU"); in SameWorkerRecvDone()
|
D | tensor_coding.cc | 47 if (alloc_attrs_.on_host() || da.device_type() == "CPU") { in InitAlloc()
|
/external/tensorflow/tensorflow/compiler/jit/ |
D | xla_device.cc | 246 if (attr.on_host()) { in GetAllocatorLocked() 452 if (alloc_attrs.on_host()) { in MakeTensorFromProto()
|
/external/tensorflow/tensorflow/contrib/verbs/ |
D | rdma.cc | 1060 const bool on_host = send_args.alloc_attrs.on_host(); in RecvHandler() local 1061 if (src_dev_->tensorflow_gpu_device_info() && !on_host) { in RecvHandler() 1566 bool on_host = recv_args_.alloc_attrs.on_host(); in AllocateTensorsAsync() local 1567 if (dst_dev_->tensorflow_gpu_device_info() && !on_host && in AllocateTensorsAsync()
|
/external/tensorflow/tensorflow/contrib/mpi/ |
D | mpi_rendezvous_mgr.cc | 242 (!send_args.alloc_attrs.on_host())) { in AddRequest()
|