/external/tensorflow/tensorflow/compiler/jit/ |
D | xla_device_context.cc | 57 void XlaTransferManager::CopyCPUTensorToDevice(const Tensor* cpu_tensor, in CopyCPUTensorToDevice() argument 61 if (cpu_tensor->NumElements() > 0) { in CopyCPUTensorToDevice() 63 << reinterpret_cast<const void*>(cpu_tensor->tensor_data().data()) in CopyCPUTensorToDevice() 67 << " " << cpu_tensor->NumElements(); in CopyCPUTensorToDevice() 69 void* src_ptr = const_cast<void*>(DMAHelper::base(cpu_tensor)); in CopyCPUTensorToDevice() 70 const int64 total_bytes = cpu_tensor->TotalBytes(); in CopyCPUTensorToDevice() 95 Tensor* cpu_tensor, in CopyDeviceTensorToCPU() argument 102 << reinterpret_cast<const void*>(cpu_tensor->tensor_data().data()) in CopyDeviceTensorToCPU() 105 const int64 total_bytes = cpu_tensor->TotalBytes(); in CopyDeviceTensorToCPU() 108 void* dst_ptr = DMAHelper::base(cpu_tensor); in CopyDeviceTensorToCPU() [all …]
|
D | xla_device_context.h | 54 void CopyCPUTensorToDevice(const Tensor* cpu_tensor, Device* device, 58 Tensor* cpu_tensor, StatusCallback done); 74 void CopyCPUTensorToDevice(const Tensor* cpu_tensor, Device* device, 79 Tensor* cpu_tensor, StatusCallback done) override;
|
/external/tensorflow/tensorflow/core/debug/ |
D | debug_gateway.cc | 84 Tensor cpu_tensor(cpu_allocator, src_tensor->dtype(), src_tensor->shape()); in CopyTensor() local 98 src_tensor, "TensorCopy", device, &cpu_tensor, in CopyTensor() 99 [node_name, cpu_tensor, copy_done_cb](const Status& s) { in CopyTensor() 101 copy_done_cb(&cpu_tensor); in CopyTensor() 111 cpu_tensor.UnsafeCopyFromInternal(*src_tensor, src_tensor->dtype(), in CopyTensor() 114 copy_done_cb(&cpu_tensor); in CopyTensor()
|
/external/tensorflow/tensorflow/core/common_runtime/gpu/ |
D | gpu_util_platform_specific.cc | 25 void GPUDeviceContext::CopyCPUTensorToDevice(const Tensor* cpu_tensor, in CopyCPUTensorToDevice() argument 29 GPUUtil::CopyCPUTensorToGPU(cpu_tensor, this, device, device_tensor, done); in CopyCPUTensorToDevice() 34 Device* device, Tensor* cpu_tensor, in CopyDeviceTensorToCPU() argument 36 GPUUtil::CopyGPUTensorToCPU(device, this, device_tensor, cpu_tensor, done); in CopyDeviceTensorToCPU()
|
D | gpu_util.cc | 261 const Tensor* gpu_tensor, Tensor* cpu_tensor, in CopyGPUTensorToCPU() argument 266 Status s = PrepareCopy(gpu_device, device_context, *gpu_tensor, cpu_tensor, in CopyGPUTensorToCPU() 287 void* dst_ptr = GetBase(cpu_tensor); in CopyGPUTensorToCPU() 304 void GPUUtil::CopyCPUTensorToGPU(const Tensor* cpu_tensor, in CopyCPUTensorToGPU() argument 311 Status s = PrepareCopy(gpu_device, device_context, *cpu_tensor, gpu_tensor, in CopyCPUTensorToGPU() 328 const int64 total_bytes = cpu_tensor->TotalBytes(); in CopyCPUTensorToGPU() 331 void* src_ptr = GetBase(cpu_tensor); in CopyCPUTensorToGPU() 337 TensorReference input_ref(*cpu_tensor); in CopyCPUTensorToGPU()
|
D | gpu_util.h | 51 const Tensor* gpu_tensor, Tensor* cpu_tensor, 91 static void CopyCPUTensorToGPU(const Tensor* cpu_tensor,
|
/external/tensorflow/tensorflow/core/common_runtime/sycl/ |
D | sycl_device_context.cc | 25 void SYCLDeviceContext::CopyCPUTensorToDevice(const Tensor *cpu_tensor, in CopyCPUTensorToDevice() argument 29 const int64 total_bytes = cpu_tensor->TotalBytes(); in CopyCPUTensorToDevice() 31 const void *src_ptr = DMAHelper::base(cpu_tensor); in CopyCPUTensorToDevice() 33 switch (cpu_tensor->dtype()) { in CopyCPUTensorToDevice() 105 Tensor *cpu_tensor, in CopyDeviceTensorToCPU() argument 110 void *dst_ptr = DMAHelper::base(cpu_tensor); in CopyDeviceTensorToCPU()
|
D | sycl_device_context.h | 34 void CopyCPUTensorToDevice(const Tensor *cpu_tensor, Device *device, 39 Device *device, Tensor *cpu_tensor,
|
/external/tensorflow/tensorflow/core/kernels/ |
D | stack_ops.cc | 295 Tensor* cpu_tensor = in ComputeAsync() local 298 &tensor, "StackPush", device, cpu_tensor, in ComputeAsync() 299 [cpu_tensor, stack, ctx, done](const Status& s) { in ComputeAsync() 303 ctx->SetStatus(stack->Push({*cpu_tensor, alloc_attrs, true})); in ComputeAsync() 306 ctx->set_output(0, *cpu_tensor); in ComputeAsync() 309 delete cpu_tensor; in ComputeAsync() 414 Tensor* cpu_tensor = &value.tensor; in ComputeAsync() local 417 new Tensor(gpu_allocator, cpu_tensor->dtype(), cpu_tensor->shape()); in ComputeAsync() 419 cpu_tensor, device, device_tensor, in ComputeAsync()
|
D | dynamic_partition_op_gpu.cu.cc | 279 Tensor cpu_tensor; in ComputeAsync() local 286 &cpu_tensor, alloc_attr), in ComputeAsync() 292 ->ThenMemcpy(cpu_tensor.flat<int32>().data(), wrapped, in ComputeAsync() 304 partition_ref, cpu_tensor, done]() { in ComputeAsync() 306 this->AllocateOutputs(c, &data, &partitions, &cpu_tensor, &outputs, done); in ComputeAsync()
|
/external/tensorflow/tensorflow/contrib/eager/python/examples/notebooks/ |
D | 1_basics.ipynb | 429 "cpu_tensor = tf.random_normal([SIZE, SIZE])\n", 432 " gpu_tensor = cpu_tensor.gpu()\n", 455 "%time tf.matmul(cpu_tensor, cpu_tensor)" 503 "cpu_tensor = tf.random_normal([SIZE, SIZE])\n", 505 "%time tf.matmul(cpu_tensor, cpu_tensor)\n", 509 " gpu_tensor = cpu_tensor.gpu()\n",
|
/external/tensorflow/tensorflow/core/common_runtime/ |
D | copy_tensor.cc | 280 Tensor* cpu_tensor = in ViaDMA() local 283 [cpu_tensor](StatusCallback done_, in ViaDMA() 286 delete cpu_tensor; in ViaDMA() 291 [delete_and_done, recv_dev_context, cpu_tensor, cpu_allocator, in ViaDMA() 299 CopyHostToDevice(cpu_tensor, cpu_allocator, out_allocator, edge_name, in ViaDMA() 305 cpu_tensor, send_dev_context, in ViaDMA()
|
D | gpu_device_context.h | 55 void CopyCPUTensorToDevice(const Tensor* cpu_tensor, Device* device, 60 Device* device, Tensor* cpu_tensor,
|
/external/tensorflow/tensorflow/core/framework/ |
D | device_base.h | 78 virtual void CopyCPUTensorToDevice(const Tensor* cpu_tensor, Device* device, in CopyCPUTensorToDevice() argument 89 Tensor* cpu_tensor, StatusCallback done) { in CopyDeviceTensorToCPU() argument
|
/external/tensorflow/tensorflow/python/eager/ |
D | ops_test.py | 272 cpu_tensor = constant_op.constant(1.0) 273 gpu_tensor = cpu_tensor.gpu() 274 self.assertAllEqual(cpu_tensor + gpu_tensor, 2.0)
|