/external/tensorflow/tensorflow/core/kernels/ |
D | eigen_benchmark.h | 39 : iters_(iters), device_(device) {} in SpatialConvolutionBenchmarksSuite() 52 static_cast<Scalar*>(device_.allocate(BufferSize(input_dims))); in SpatialConvolution() 54 static_cast<Scalar*>(device_.allocate(BufferSize(filter_dims))); in SpatialConvolution() 56 static_cast<Scalar*>(device_.allocate(BufferSize(output_dims))); in SpatialConvolution() 58 device_.memset(input_data, 123, BufferSize(input_dims)); in SpatialConvolution() 59 device_.memset(filter_data, 123, BufferSize(filter_dims)); in SpatialConvolution() 67 output.device(device_) = Eigen::SpatialConvolution(input, filter); in SpatialConvolution() 72 device_.deallocate(input_data); in SpatialConvolution() 73 device_.deallocate(filter_data); in SpatialConvolution() 74 device_.deallocate(output_data); in SpatialConvolution() [all …]
|
D | ops_testutil.h | 74 : device_(DeviceFactory::NewDevice("CPU", {}, "/job:a/replica:0/task:0")), in OpsTestBase() 76 CHECK(device_.get()) << "Could not create CPU device"; 77 allocator_ = device_->GetAllocator(AllocatorAttributes()); 104 kernel_ = CreateOpKernel(device_type_, device_.get(), allocator(), in InitOpWithGraphVersion() 142 ResourceMgr* rm = device_->resource_manager(); in AddResourceInput() 148 handle.set_device(device_->name()); in AddResourceInput() 168 params_.get()->device = device_.get(); in RunOpKernel() 178 params_.get()->resource_manager = device_.get()->resource_manager(); in RunOpKernel() 181 device_->Compute(kernel_.get(), context_.get()); in RunOpKernel() 225 std::unique_ptr<Device> device_;
|
D | scoped_allocator_ops.cc | 37 device_ = context->device(); in ScopedAllocatorOp() 51 ScopedAllocatorMgr* sam = device_->GetScopedAllocatorMgr(); in Compute() 54 "ScopedAllocatorMgr not supported on device ", device_->name())); in Compute() 82 DeviceBase* device_; member in tensorflow::ScopedAllocatorOp 101 device_ = context->device(); in ScopedAllocatorConcatOp() 163 DeviceBase* device_; member in tensorflow::ScopedAllocatorConcatOp 180 device_ = context->device(); in ScopedAllocatorSplitOp() 225 DeviceBase* device_; member in tensorflow::ScopedAllocatorSplitOp
|
D | collective_nccl_reducer_test.cc | 181 Device* dev = instances_[rank]->device_; in RunTest() 219 TF_CHECK_OK(parent_->dev_mgr_->LookupDevice(device_name_, &device_)) in DeviceInstance() 233 Tensor(device_->GetAllocator(AllocatorAttributes()), dtype, shape); in InitTensor() 237 auto* dev_info = device_->tensorflow_gpu_device_info(); in InitTensor() 240 &cpu_tensor, device_, &tensor_, in InitTensor() 246 col_params_.merge_op = GetAdd(device_); in DoReduce() 247 col_params_.final_op = GetDiv(device_); in DoReduce() 252 op_params.device = device_; in DoReduce() 261 auto* dev_info = device_->tensorflow_gpu_device_info(); in DoReduce() 276 parent_->GetCollectiveReduce(col_params_, &tensor_, device_); in DoReduce() [all …]
|
D | ops_testutil.cc | 27 CHECK(device_) << "No device provided"; in SetDevice() 29 device_ = std::move(device); in SetDevice() 36 allocator_ = device_->GetAllocator(AllocatorAttributes()); in SetDevice()
|
/external/eigen/bench/tensors/ |
D | tensor_benchmarks.h | 21 : m_(m), k_(k), n_(n), device_(device) { in BenchmarkSuite() 26 : m_(m), k_(m), n_(m), device_(device) { in BenchmarkSuite() 31 device_.deallocate(a_); in ~BenchmarkSuite() 32 device_.deallocate(b_); in ~BenchmarkSuite() 33 device_.deallocate(c_); in ~BenchmarkSuite() 40 device_.memcpy(c_, a_, m_ * m_ * sizeof(T)); in memcpy() 61 B.device(device_) = A.template cast<T>(); in typeCasting() 76 C.device(device_) = C.random(); in random() 99 C.slice(first_quadrant, quarter_sizes).device(device_) = in slicing() 101 C.slice(second_quadrant, quarter_sizes).device(device_) = in slicing() [all …]
|
/external/nos/host/generic/libnos/ |
D | NuggetClient.cpp | 42 device_name_.empty() ? nullptr : device_name_.c_str(), &device_) == 0; in Open() 48 device_.ops.close(device_.ctx); in Close() 78 uint32_t status_code = nos_call_application(&device_, appId, arg, in CallApp() 90 return open_ ? &device_ : nullptr; in Device() 94 return open_ ? &device_ : nullptr; in Device()
|
/external/tensorflow/tensorflow/core/common_runtime/ |
D | kernel_benchmark_testlib.cc | 58 device_ = in Benchmark() 60 CHECK(device_) << "Could not create a " << device << " device"; in Benchmark() 78 params.device = device_.get(); in Benchmark() 82 return CreateNonCachedKernel(device_.get(), nullptr, ndef, in Benchmark() 104 if (device_) { in ~Benchmark() 110 device_.reset(); in ~Benchmark() 134 if (!device_ || iters == 0) { in RunWithArgs() 171 TF_CHECK_OK(device_->Sync()); in RunWithArgs() 189 TF_CHECK_OK(device_->Sync()); in RunWithArgs()
|
D | graph_runner.cc | 96 device_(device_deleter_.get()) {} in GraphRunner() 97 GraphRunner::GraphRunner(Device* device) : device_(device) {} in GraphRunner() 105 if (device_ == nullptr) { in Run() 110 function_library->device()->device_type() != device_->device_type()) { in Run() 115 VLOG(1) << "Cannot run on: " << device_->device_type() in Run() 147 device_->attributes(), false /* use_function_convention */, &metadata)); in Run() 158 params.device = device_; in Run() 163 return CreateNonCachedKernel(device_, function_library, ndef, producer, in Run()
|
D | ring_reducer_test.cc | 327 Device* dev = instances_[di]->device_; in RunTest() 408 TF_CHECK_OK(parent_->dev_mgr_->LookupDevice(dev_name, &device_)) in DeviceInstance() 445 Tensor(device_->GetAllocator(AllocatorAttributes()), dtype, shape); in InitTensor() 451 auto* dev_info = device_->tensorflow_gpu_device_info(); in InitTensor() 455 &cpu_tensor, device_, &tensor_, [¬e](const Status& s) { in InitTensor() 467 GetAdd(col_params_.instance.data_type, device_type_, device_); in DoReduce() 469 GetDiv(col_params_.instance.data_type, device_type_, device_); in DoReduce() 474 op_params.device = device_; in DoReduce() 483 auto* dev_info = device_->tensorflow_gpu_device_info(); in DoReduce() 498 col_params_, &tensor_, DEVICE_CPU, device_); in DoReduce() [all …]
|
/external/tensorflow/tensorflow/core/common_runtime/eager/ |
D | kernel_and_device.cc | 104 options.target = device_ == nullptr ? "" : device_->name(); in Init() 156 device_->resource_manager()->Cleanup(name).IgnoreError(); in Run() 225 if (device_->tensorflow_gpu_device_info() != nullptr) { in Run() 226 device_context = device_->tensorflow_gpu_device_info()->default_context; in Run() 232 params.device = device_; in Run() 236 params.resource_manager = device_->resource_manager(); in Run() 278 device_->ComputeAsync(async, &context, [&done]() { done.Notify(); }); in Run() 284 if (device_->TraceUsingAnnotations()) { in Run() 286 device_->Compute(kernel_.get(), &context); in Run() 289 device_->Compute(kernel_.get(), &context); in Run() [all …]
|
D | eager_operation.h | 31 device_(nullptr), in EagerOperation() 61 tensorflow::Device* Device() const { return device_; } in Device() 63 void SetDevice(tensorflow::Device* device) { device_ = device; } in SetDevice() 75 tensorflow::Device* device_; variable
|
D | tensor_handle.cc | 53 device_(d), in TensorHandle() 68 device_(d), in TensorHandle() 88 device_(d), in TensorHandle() 106 device_(nullptr), in TensorHandle() 169 *device = device_; in TensorAndDevice() 338 strings::StrAppend(&out, "Device: ", device_ ? device_->DebugString() : "[]"); in DebugString() 340 strings::StrAppend(&out, ", Tensor: ", device_ ? "?" : tensor_.DebugString(), in DebugString()
|
D | tensor_handle.h | 84 tensorflow::Device* device() const { return device_; } in device() 124 return device_ == nullptr || in OnHostCPU() 125 (ctx_ == nullptr || ctx_->HostCPU() == device_); in OnHostCPU() 157 tensorflow::Device* const device_; variable
|
/external/tensorflow/tensorflow/cc/framework/ |
D | scope.cc | 95 device_(other.impl()->device_), in Impl() 113 device_(other.impl()->device_), in Impl() 137 device_(other.impl()->device_), in Impl() 154 device_(device), in Impl() 172 device_(other.impl()->device_), in Impl() 189 device_(other.impl()->device_), in Impl() 207 device_(other.impl()->device_), in Impl() 225 device_(other.impl()->device_), in Impl() 246 device_(other.impl()->device_), in Impl() 264 device_(other.impl()->device_), in Impl() [all …]
|
/external/eigen/unsupported/Eigen/CXX11/src/Tensor/ |
D | TensorDeviceCuda.h | 103 cudaGetDevice(&device_); in initializeDeviceProp() 107 …CudaStreamDevice(int device) : stream_(&default_stream), device_(device), scratch_(NULL), semaphor… in initializeDeviceProp() 115 : stream_(stream), device_(device), scratch_(NULL), semaphore_(NULL) { in initializeDeviceProp() 117 cudaGetDevice(&device_); in initializeDeviceProp() 124 device_ = device; in initializeDeviceProp() 137 return m_deviceProperties[device_]; in initializeDeviceProp() 140 cudaError_t err = cudaSetDevice(device_); in initializeDeviceProp() 150 cudaError_t err = cudaSetDevice(device_); in initializeDeviceProp() 178 int device_; in initializeDeviceProp() variable
|
/external/tensorflow/tensorflow/lite/kernels/ |
D | eigen_support.cc | 79 if (!device_) { in GetThreadPoolDevice() 82 device_.reset(new Eigen::ThreadPoolDevice(thread_pool_wrapper_.get(), in GetThreadPoolDevice() 85 return device_.get(); in GetThreadPoolDevice() 95 device_.reset(); in SetNumThreads() 103 std::unique_ptr<Eigen::ThreadPoolDevice> device_; member in tflite::eigen_support::__anon0b33ddad0111::LazyEigenThreadPoolHolder
|
/external/tensorflow/tensorflow/stream_executor/cuda/ |
D | cuda_gpu_executor.cc | 139 status = GpuDriver::GetDevice(device_ordinal_, &device_); in Init() 144 status = GpuDriver::CreateContext(device_ordinal_, device_, device_options, in Init() 150 return GpuDriver::GetComputeCapability(&cc_major_, &cc_minor_, device_); in Init() 923 if (!GpuDriver::GetGridLimits(&x, &y, &z, device_)) { in FillBlockDimLimit() 1042 string pci_bus_id = GpuDriver::GetPCIBusID(device_); in PopulateDeviceDescription() 1056 device_) in PopulateDeviceDescription() 1061 CU_DEVICE_ATTRIBUTE_MAX_BLOCK_DIM_X, device_) in PopulateDeviceDescription() 1064 CU_DEVICE_ATTRIBUTE_MAX_BLOCK_DIM_Y, device_) in PopulateDeviceDescription() 1067 CU_DEVICE_ATTRIBUTE_MAX_BLOCK_DIM_Z, device_) in PopulateDeviceDescription() 1072 GpuDriver::GetDeviceAttribute(CU_DEVICE_ATTRIBUTE_CLOCK_RATE, device_) in PopulateDeviceDescription() [all …]
|
/external/pdfium/xfa/fwl/ |
D | cfx_barcode_unittest.cpp | 31 device_ = std::move(device); in SetUp() 36 device_.reset(); in TearDown() 55 return barcode_->RenderDevice(device_.get(), &matrix_); in RenderDevice() 66 std::unique_ptr<CFX_RenderDevice> device_; member in BarcodeTest
|
/external/tensorflow/tensorflow/core/framework/ |
D | resource_handle.h | 39 const string& device() const { return device_; } in device() 40 void set_device(const string& device) { device_ = device; } in set_device() 76 string device_;
|
D | resource_op_kernel_test.cc | 100 DEVICE_CPU, &device_, device_.GetAllocator(AllocatorAttributes()), in CreateOp() 118 params.device = &device_; in RunOpKernel() 127 StubDevice device_; member in tensorflow::__anon1ebbe4430111::ResourceOpKernelTest
|
/external/tensorflow/tensorflow/stream_executor/rocm/ |
D | rocm_gpu_executor.cc | 152 status = GpuDriver::GetDevice(device_ordinal_, &device_); in Init() 157 status = GpuDriver::CreateContext(device_ordinal_, device_, device_options, in Init() 163 return GpuDriver::GetGpuISAVersion(&version_, device_); in Init() 825 if (!GpuDriver::GetGridLimits(&x, &y, &z, device_)) { in FillBlockDimLimit() 889 string pci_bus_id = GpuDriver::GetPCIBusID(device_); in PopulateDeviceDescription() 916 (void)GpuDriver::IsEccEnabled(device_, &ecc_enabled); in PopulateDeviceDescription() 922 (void)GpuDriver::GetDeviceTotalMemory(device_, &device_memory_size); in PopulateDeviceDescription() 934 (void)GpuDriver::GetDeviceName(device_, &device_name); in PopulateDeviceDescription() 948 GpuDriver::GetMaxSharedMemoryPerCore(device_).ValueOrDie()); in PopulateDeviceDescription() 950 GpuDriver::GetMaxSharedMemoryPerBlock(device_).ValueOrDie()); in PopulateDeviceDescription() [all …]
|
/external/v4l2_codec2/vda/ |
D | v4l2_video_decode_accelerator.cc | 44 if (device_->Ioctl(type, arg) != 0) { \ 59 if (device_->Ioctl(type, arg) != 0) \ 127 device_(device), in V4L2VideoDecodeAccelerator() 191 if (!device_->Open(V4L2Device::Type::kDecoder, input_format_fourcc_)) { in Initialize() 932 if (!device_->ClearDevicePollInterrupt()) { in ServiceDeviceTask() 1007 if (!device_->SetDevicePollInterrupt()) { in Enqueue() 1029 if (!device_->SetDevicePollInterrupt()) { in Enqueue() 1051 while (device_->Ioctl(VIDIOC_DQEVENT, &ev) == 0) { in DequeueResolutionChangeEvent() 1095 if (device_->Ioctl(VIDIOC_DQBUF, &dqbuf) != 0) { in DequeueInputBuffer() 1131 if (device_->Ioctl(VIDIOC_DQBUF, &dqbuf) != 0) { in DequeueOutputBuffer() [all …]
|
/external/tensorflow/tensorflow/core/distributed_runtime/ |
D | tensor_coding.cc | 30 device_ = nullptr; in Clear() 44 device_ = d; in InitAlloc() 50 allocator_ = device_->GetAllocator(alloc_attrs_); in InitAlloc() 61 s = device_->MakeTensorFromProto(meta_.tensor(), alloc_attrs_, &tensor_); in InitFrom() 92 device_->MakeTensorFromProto(meta_.tensor(), alloc_attrs_, &tensor_); in ParseFrom()
|
/external/tensorflow/tensorflow/compiler/tf2xla/ |
D | graph_compiler.h | 61 : device_(device), in GraphCompiler() 81 XlaCompilationDevice* device_; variable
|