/external/tensorflow/tensorflow/core/kernels/ |
D | substr_op.cc | 72 tensorflow::internal::SubtleMustCopy(pos_tensor.scalar<T>()()); in Compute() 74 tensorflow::internal::SubtleMustCopy(len_tensor.scalar<T>()()); in Compute() 102 const T pos = tensorflow::internal::SubtleMustCopy(pos_flat(i)); in Compute() 103 const T len = tensorflow::internal::SubtleMustCopy(len_flat(i)); in Compute() 181 const T pos = tensorflow::internal::SubtleMustCopy(pos_bcast(i)); in Compute() 182 const T len = tensorflow::internal::SubtleMustCopy(len_bcast(i)); in Compute() 246 tensorflow::internal::SubtleMustCopy(pos_bcast(i, j)); in Compute() 248 tensorflow::internal::SubtleMustCopy(len_bcast(i, j)); in Compute()
|
D | decode_bmp_op.cc | 66 int32 header_size_ = internal::SubtleMustCopy( in Compute() 69 int32 width_ = internal::SubtleMustCopy( in Compute() 72 int32 height_ = internal::SubtleMustCopy( in Compute() 75 int32 bpp_ = internal::SubtleMustCopy( in Compute()
|
D | scatter_functor.h | 212 const Index index = ::tensorflow::internal::SubtleMustCopy(indices(i)); 238 const Index index = ::tensorflow::internal::SubtleMustCopy(indices(i)); 272 const Index index = ::tensorflow::internal::SubtleMustCopy(indices(i)); 298 const Index index = ::tensorflow::internal::SubtleMustCopy(indices(i)); 310 const Index index = ::tensorflow::internal::SubtleMustCopy(indices(i)); 336 const Index index = ::tensorflow::internal::SubtleMustCopy(indices(i)); 368 const Index index = ::tensorflow::internal::SubtleMustCopy(indices(i)); 393 const Index index = ::tensorflow::internal::SubtleMustCopy(indices(i)); 427 const Index index = ::tensorflow::internal::SubtleMustCopy(indices(i)); 452 const Index index = ::tensorflow::internal::SubtleMustCopy(indices(i)); [all …]
|
D | segment_reduction_ops.cc | 99 ? internal::SubtleMustCopy(segment_vec(num_indices - 1)) + 1 in Compute() 125 Index out_index = internal::SubtleMustCopy(segment_vec(start)); in Compute() 136 next_index = internal::SubtleMustCopy(segment_vec(end)); in Compute() 385 Index j = internal::SubtleMustCopy(segment_ids(i)); in operator ()() 482 internal::SubtleMustCopy(num_segments.scalar<int32>()()); in Compute() 643 output_rows = internal::SubtleMustCopy(num_segments.scalar<int32>()()); in Compute() 667 ? internal::SubtleMustCopy(segment_vec(num_indices - 1)) + 1 in Compute() 699 OutputRow out_index = internal::SubtleMustCopy(segment_vec(start)); in Compute() 707 next_index = internal::SubtleMustCopy(segment_vec(end)); in Compute() 1030 internal::SubtleMustCopy(output_dim0.scalar<SegmentId>()()); in Compute() [all …]
|
D | stateless_random_ops.cc | 37 seed0 = internal::SubtleMustCopy(seed_vals(0)); in GenerateKey() 38 seed1 = internal::SubtleMustCopy(seed_vals(1)); in GenerateKey() 41 seed0 = internal::SubtleMustCopy(seed_vals(0)); in GenerateKey() 42 seed1 = internal::SubtleMustCopy(seed_vals(1)); in GenerateKey()
|
D | dynamic_partition_op.cc | 60 const int32 p = internal::SubtleMustCopy(e_partitions(i)); in ValidateAndAllocateOutputs() 113 const int32 p = internal::SubtleMustCopy(e_partitions(i)); in Compute() 141 const int32 p = internal::SubtleMustCopy(e_partitions(i)); in Compute()
|
D | sparse_tensor_dense_matmul_op.cc | 268 const Tindices m = internal::SubtleMustCopy(a_indices(i, lhs_index_a)); in Compute() 269 const Tindices k = internal::SubtleMustCopy(a_indices(i, rhs_index_a)); in Compute() 288 const Tindices m = internal::SubtleMustCopy(a_indices(i, lhs_index_a)); \ in Compute() 289 const Tindices k = internal::SubtleMustCopy(a_indices(i, rhs_index_a)); \ in Compute()
|
D | concat_op.cc | 83 internal::SubtleMustCopy(concat_dim_tensor->scalar<int32>()()); in Compute() 86 internal::SubtleMustCopy(concat_dim_tensor->scalar<int64>()()); in Compute() 315 const int64 cdim = internal::SubtleMustCopy(concat_dim.scalar<int32>()()); in Compute()
|
D | crop_and_resize_op.cc | 160 const int crop_height = internal::SubtleMustCopy(crop_size_vec(0)); in ComputeAsync() 161 const int crop_width = internal::SubtleMustCopy(crop_size_vec(1)); in ComputeAsync() 381 const int batch_size = internal::SubtleMustCopy(image_size_vec(0)); in ComputeAsync() 382 const int image_height = internal::SubtleMustCopy(image_size_vec(1)); in ComputeAsync() 383 const int image_width = internal::SubtleMustCopy(image_size_vec(2)); in ComputeAsync() 384 const int depth = internal::SubtleMustCopy(image_size_vec(3)); in ComputeAsync()
|
D | image_resizer_state.h | 96 out_height = internal::SubtleMustCopy(Svec(0)); in ValidateAndCalculateOutputSize() 97 out_width = internal::SubtleMustCopy(Svec(1)); in ValidateAndCalculateOutputSize()
|
D | sparse_xent_op.h | 73 const Index label = tensorflow::internal::SubtleMustCopy(labels_(batch)); in operator() 113 const Index label = tensorflow::internal::SubtleMustCopy(labels_(batch)); in operator()
|
D | spacetobatch_functor.h | 58 (*output)[i] = SubtleMustCopy(eigen_vec(i)); in SubtleMustCopyFlatHelper()
|
D | scatter_nd_op_cpu_impl.h | 119 const Index ix_d = internal::SubtleMustCopy(Tindices(loc, dim)); 203 const Index ix_d = internal::SubtleMustCopy(Tindices(loc, dim));
|
D | unique_op.cc | 74 axis = internal::SubtleMustCopy(axis_tensor.scalar<int32>()()); in Compute() 76 axis = internal::SubtleMustCopy(axis_tensor.scalar<int64>()()); in Compute()
|
D | sample_distorted_bounding_box_op.cc | 257 const uint64 height_raw = internal::SubtleMustCopy(image_size.flat<T>()(0)); in Compute() 258 const uint64 width_raw = internal::SubtleMustCopy(image_size.flat<T>()(1)); in Compute()
|
D | dynamic_stitch_op.cc | 265 int32 index = internal::SubtleMustCopy(indices_vec(i)); in Compute() 277 int32 index = internal::SubtleMustCopy(indices_vec(i)); in Compute()
|
D | in_topk_op.cc | 76 auto target = internal::SubtleMustCopy(targets(b)); in Compute()
|
D | sparse_tensor_dense_add_op.cc | 134 idx[d] = internal::SubtleMustCopy(indices(i, d)); in operator ()()
|
D | sparse_dense_binary_op_shared.cc | 131 idx[d] = internal::SubtleMustCopy(indices_mat(i, d)); \ in Compute()
|
D | gather_nd_op_cpu_impl.h | 59 const Index ix_i = internal::SubtleMustCopy(Tindices_(loc, i)); in GenerateIndices()
|
D | scan_ops.cc | 55 internal::SubtleMustCopy(tensor_axis.scalar<Tidx>()()); in Compute()
|
/external/tensorflow/tensorflow/core/util/ |
D | strided_slice_op.cc | 123 dense->begin[full_index] = internal::SubtleMustCopy<T>(begin_flat[i]); in BuildDenseSpec() 126 dense->end[full_index] = internal::SubtleMustCopy<T>(end_flat[i]); in BuildDenseSpec() 129 internal::SubtleMustCopy<T>(strides_flat[i]); in BuildDenseSpec()
|
/external/tensorflow/tensorflow/contrib/factorization/kernels/ |
D | masked_matmul_ops.cc | 120 int64 a_index = internal::SubtleMustCopy(indices_mat(i, 0)); in Compute() 127 int64 b_index = internal::SubtleMustCopy(indices_mat(i, 1)); in Compute()
|
/external/tensorflow/tensorflow/core/framework/ |
D | bounds_check.h | 45 EIGEN_ALWAYS_INLINE EIGEN_DEVICE_FUNC const T SubtleMustCopy(const T &x) { in SubtleMustCopy() function
|
/external/tensorflow/tensorflow/contrib/tensor_forest/kernels/ |
D | tree_utils.h | 185 int64 midi = internal::SubtleMustCopy(sparse_input_indices(mid, 0)); in FindSparseValue() 186 int64 midj = internal::SubtleMustCopy(sparse_input_indices(mid, 1)); in FindSparseValue()
|