Home
last modified time | relevance | path

Searched refs:cpu_tensor (Results 1 – 15 of 15) sorted by relevance

/external/tensorflow/tensorflow/compiler/jit/
Dxla_device_context.cc57 void XlaTransferManager::CopyCPUTensorToDevice(const Tensor* cpu_tensor, in CopyCPUTensorToDevice() argument
61 if (cpu_tensor->NumElements() > 0) { in CopyCPUTensorToDevice()
63 << reinterpret_cast<const void*>(cpu_tensor->tensor_data().data()) in CopyCPUTensorToDevice()
67 << " " << cpu_tensor->NumElements(); in CopyCPUTensorToDevice()
69 void* src_ptr = const_cast<void*>(DMAHelper::base(cpu_tensor)); in CopyCPUTensorToDevice()
70 const int64 total_bytes = cpu_tensor->TotalBytes(); in CopyCPUTensorToDevice()
95 Tensor* cpu_tensor, in CopyDeviceTensorToCPU() argument
102 << reinterpret_cast<const void*>(cpu_tensor->tensor_data().data()) in CopyDeviceTensorToCPU()
105 const int64 total_bytes = cpu_tensor->TotalBytes(); in CopyDeviceTensorToCPU()
108 void* dst_ptr = DMAHelper::base(cpu_tensor); in CopyDeviceTensorToCPU()
[all …]
Dxla_device_context.h54 void CopyCPUTensorToDevice(const Tensor* cpu_tensor, Device* device,
58 Tensor* cpu_tensor, StatusCallback done);
74 void CopyCPUTensorToDevice(const Tensor* cpu_tensor, Device* device,
79 Tensor* cpu_tensor, StatusCallback done) override;
/external/tensorflow/tensorflow/core/debug/
Ddebug_gateway.cc84 Tensor cpu_tensor(cpu_allocator, src_tensor->dtype(), src_tensor->shape()); in CopyTensor() local
98 src_tensor, "TensorCopy", device, &cpu_tensor, in CopyTensor()
99 [node_name, cpu_tensor, copy_done_cb](const Status& s) { in CopyTensor()
101 copy_done_cb(&cpu_tensor); in CopyTensor()
111 cpu_tensor.UnsafeCopyFromInternal(*src_tensor, src_tensor->dtype(), in CopyTensor()
114 copy_done_cb(&cpu_tensor); in CopyTensor()
/external/tensorflow/tensorflow/core/common_runtime/gpu/
Dgpu_util_platform_specific.cc25 void GPUDeviceContext::CopyCPUTensorToDevice(const Tensor* cpu_tensor, in CopyCPUTensorToDevice() argument
29 GPUUtil::CopyCPUTensorToGPU(cpu_tensor, this, device, device_tensor, done); in CopyCPUTensorToDevice()
34 Device* device, Tensor* cpu_tensor, in CopyDeviceTensorToCPU() argument
36 GPUUtil::CopyGPUTensorToCPU(device, this, device_tensor, cpu_tensor, done); in CopyDeviceTensorToCPU()
Dgpu_util.cc261 const Tensor* gpu_tensor, Tensor* cpu_tensor, in CopyGPUTensorToCPU() argument
266 Status s = PrepareCopy(gpu_device, device_context, *gpu_tensor, cpu_tensor, in CopyGPUTensorToCPU()
287 void* dst_ptr = GetBase(cpu_tensor); in CopyGPUTensorToCPU()
304 void GPUUtil::CopyCPUTensorToGPU(const Tensor* cpu_tensor, in CopyCPUTensorToGPU() argument
311 Status s = PrepareCopy(gpu_device, device_context, *cpu_tensor, gpu_tensor, in CopyCPUTensorToGPU()
328 const int64 total_bytes = cpu_tensor->TotalBytes(); in CopyCPUTensorToGPU()
331 void* src_ptr = GetBase(cpu_tensor); in CopyCPUTensorToGPU()
337 TensorReference input_ref(*cpu_tensor); in CopyCPUTensorToGPU()
Dgpu_util.h51 const Tensor* gpu_tensor, Tensor* cpu_tensor,
91 static void CopyCPUTensorToGPU(const Tensor* cpu_tensor,
/external/tensorflow/tensorflow/core/common_runtime/sycl/
Dsycl_device_context.cc25 void SYCLDeviceContext::CopyCPUTensorToDevice(const Tensor *cpu_tensor, in CopyCPUTensorToDevice() argument
29 const int64 total_bytes = cpu_tensor->TotalBytes(); in CopyCPUTensorToDevice()
31 const void *src_ptr = DMAHelper::base(cpu_tensor); in CopyCPUTensorToDevice()
33 switch (cpu_tensor->dtype()) { in CopyCPUTensorToDevice()
105 Tensor *cpu_tensor, in CopyDeviceTensorToCPU() argument
110 void *dst_ptr = DMAHelper::base(cpu_tensor); in CopyDeviceTensorToCPU()
Dsycl_device_context.h34 void CopyCPUTensorToDevice(const Tensor *cpu_tensor, Device *device,
39 Device *device, Tensor *cpu_tensor,
/external/tensorflow/tensorflow/core/kernels/
Dstack_ops.cc295 Tensor* cpu_tensor = in ComputeAsync() local
298 &tensor, "StackPush", device, cpu_tensor, in ComputeAsync()
299 [cpu_tensor, stack, ctx, done](const Status& s) { in ComputeAsync()
303 ctx->SetStatus(stack->Push({*cpu_tensor, alloc_attrs, true})); in ComputeAsync()
306 ctx->set_output(0, *cpu_tensor); in ComputeAsync()
309 delete cpu_tensor; in ComputeAsync()
414 Tensor* cpu_tensor = &value.tensor; in ComputeAsync() local
417 new Tensor(gpu_allocator, cpu_tensor->dtype(), cpu_tensor->shape()); in ComputeAsync()
419 cpu_tensor, device, device_tensor, in ComputeAsync()
Ddynamic_partition_op_gpu.cu.cc279 Tensor cpu_tensor; in ComputeAsync() local
286 &cpu_tensor, alloc_attr), in ComputeAsync()
292 ->ThenMemcpy(cpu_tensor.flat<int32>().data(), wrapped, in ComputeAsync()
304 partition_ref, cpu_tensor, done]() { in ComputeAsync()
306 this->AllocateOutputs(c, &data, &partitions, &cpu_tensor, &outputs, done); in ComputeAsync()
/external/tensorflow/tensorflow/contrib/eager/python/examples/notebooks/
D1_basics.ipynb429 "cpu_tensor = tf.random_normal([SIZE, SIZE])\n",
432 " gpu_tensor = cpu_tensor.gpu()\n",
455 "%time tf.matmul(cpu_tensor, cpu_tensor)"
503 "cpu_tensor = tf.random_normal([SIZE, SIZE])\n",
505 "%time tf.matmul(cpu_tensor, cpu_tensor)\n",
509 " gpu_tensor = cpu_tensor.gpu()\n",
/external/tensorflow/tensorflow/core/common_runtime/
Dcopy_tensor.cc280 Tensor* cpu_tensor = in ViaDMA() local
283 [cpu_tensor](StatusCallback done_, in ViaDMA()
286 delete cpu_tensor; in ViaDMA()
291 [delete_and_done, recv_dev_context, cpu_tensor, cpu_allocator, in ViaDMA()
299 CopyHostToDevice(cpu_tensor, cpu_allocator, out_allocator, edge_name, in ViaDMA()
305 cpu_tensor, send_dev_context, in ViaDMA()
Dgpu_device_context.h55 void CopyCPUTensorToDevice(const Tensor* cpu_tensor, Device* device,
60 Device* device, Tensor* cpu_tensor,
/external/tensorflow/tensorflow/core/framework/
Ddevice_base.h78 virtual void CopyCPUTensorToDevice(const Tensor* cpu_tensor, Device* device, in CopyCPUTensorToDevice() argument
89 Tensor* cpu_tensor, StatusCallback done) { in CopyDeviceTensorToCPU() argument
/external/tensorflow/tensorflow/python/eager/
Dops_test.py272 cpu_tensor = constant_op.constant(1.0)
273 gpu_tensor = cpu_tensor.gpu()
274 self.assertAllEqual(cpu_tensor + gpu_tensor, 2.0)