/external/tensorflow/tensorflow/contrib/reduce_slice_ops/kernels/ |
D | reduce_slice_ops.cc | 157 #define REGISTER_CPU_SUMPROD_REDUCE_SLICE_KERNELS(type, index_type) \ argument 161 .TypeConstraint<index_type>("Tindices"), \ 162 ReduceSliceKernel<CPUDevice, type, index_type, \ 167 .TypeConstraint<index_type>("Tindices"), \ 168 ReduceSliceKernel<CPUDevice, type, index_type, \ 171 #define REGISTER_CPU_MINMAX_REDUCE_SLICE_KERNELS(type, index_type) \ argument 175 .TypeConstraint<index_type>("Tindices"), \ 176 ReduceSliceKernel<CPUDevice, type, index_type, \ 181 .TypeConstraint<index_type>("Tindices"), \ 182 ReduceSliceKernel<CPUDevice, type, index_type, \ [all …]
|
/external/tensorflow/tensorflow/compiler/xla/service/gpu/ |
D | parallel_loop_emitter.cc | 60 llvm::Type* index_type) { in EmitIndexAndSetExitBasicBlock() argument 73 CHECK_NE(index_type, nullptr); in EmitIndexAndSetExitBasicBlock() 79 block_id = b_->CreateZExtOrTrunc(block_id, index_type, "block_id"); in EmitIndexAndSetExitBasicBlock() 89 thread_id = b_->CreateZExtOrTrunc(thread_id, index_type, "thread_id"); in EmitIndexAndSetExitBasicBlock() 94 index_type, launch_dimensions_.threads_per_block()), in EmitIndexAndSetExitBasicBlock() 111 llvm::ConstantInt::get(index_type, in EmitIndexAndSetExitBasicBlock() 119 linear_index_base, llvm::ConstantInt::get(index_type, unroll_factor_), in EmitIndexAndSetExitBasicBlock() 126 b_->CreateAdd(linear_index_base, llvm::ConstantInt::get(index_type, i), in EmitIndexAndSetExitBasicBlock() 135 llvm::ConstantInt::get(index_type, ShapeUtil::ElementsIn(shape_))), in EmitIndexAndSetExitBasicBlock()
|
D | elemental_ir_emitter.cc | 372 llvm::Type* index_type = index.GetType(); in MakeElementGenerator() local 377 llvm_ir::ForLoopNest loops(IrName(hlo), b_, index_type); in MakeElementGenerator() 429 index_type); in MakeElementGenerator() 449 llvm::Type* index_type = output_index.GetType(); in MakeElementGenerator() local 452 IrArray::Index(index_type))); in MakeElementGenerator() 455 llvm_ir::ForLoopNest loops(IrName(hlo), b_, index_type); in MakeElementGenerator() 471 input_multi_index, hlo->operand(0)->shape(), index_type); in MakeElementGenerator()
|
/external/tensorflow/tensorflow/core/kernels/ |
D | segment_reduction_ops.cc | 300 #define REGISTER_CPU_KERNEL_SEGMENT(name, functor, type, index_type, \ argument 306 .TypeConstraint<index_type>("Tindices"), \ 307 SegmentReductionOp<CPUDevice, type, index_type, functor, default_value>) 309 #define REGISTER_REAL_CPU_KERNELS(type, index_type) \ argument 311 type, index_type, 0); \ 313 "SegmentMean", Eigen::internal::MeanReducer<type>, type, index_type, 0); \ 315 "SegmentProd", Eigen::internal::ProdReducer<type>, type, index_type, 1); \ 317 type, index_type, 0); \ 319 type, index_type, 0) 321 #define REGISTER_COMPLEX_CPU_KERNELS(type, index_type) \ argument [all …]
|
D | one_hot_op.cc | 128 #define REGISTER_ONE_HOT_INDEX(type, index_type) \ argument 131 .TypeConstraint<index_type>("TI") \ 134 OneHotOp<CPUDevice, type, index_type>); 172 #define REGISTER_ONE_HOT_GPU_INDEX(type, index_type) \ argument 175 .TypeConstraint<index_type>("TI") \ 178 OneHotOp<GPUDevice, type, index_type>);
|
D | scatter_nd_op.cc | 308 #define REGISTER_SCATTER_ND_KERNEL_INDEX(type, index_type, dev, name) \ argument 312 .TypeConstraint<index_type>("Tindices") \ 314 ScatterNdOp<dev##Device, type, index_type>) 316 #define REGISTER_SCATTER_ND_UPDATE_KERNEL_INDEX(type, index_type, dev, name, \ argument 322 .TypeConstraint<index_type>("Tindices"), \ 323 ScatterNdUpdateOp<dev##Device, type, index_type, op>) 325 #define REGISTER_RESOURCE_SCATTER_ND_UPDATE_KERNEL_INDEX(type, index_type, \ argument 331 .TypeConstraint<index_type>("Tindices") \ 333 ScatterNdUpdateOp<dev##Device, type, index_type, op>) 387 #define REGISTER_SCATTER_ND_TENSOR_UPDATE_TYPE_INDEX_TYPE(type, index_type, \ argument [all …]
|
D | gather_op.cc | 122 #define REGISTER_GATHER_FULL(dev, type, index_type) \ argument 126 .TypeConstraint<index_type>("Tindices"), \ 127 GatherOp<dev##Device, type, index_type>); \ 131 .TypeConstraint<index_type>("Tindices") \ 133 GatherOp<dev##Device, type, index_type>)
|
D | sparse_to_dense_op.cc | 142 #define REGISTER_KERNELS(type, index_type) \ argument 146 .TypeConstraint<index_type>("Tindices"), \ 147 SparseToDense<type, index_type>);
|
D | scatter_nd_op_test.cc | 41 void MakeOp(DataType variable_ref_type, DataType index_type) { in MakeOp() argument 44 .Input(FakeInput(index_type)) in MakeOp() 243 void MakeBenchmarkOp(const char* op, DataType index_type) { in MakeBenchmarkOp() argument 246 .Input(FakeInput(index_type)) in MakeBenchmarkOp()
|
D | gather_op_test.cc | 43 void MakeOp(DataType data_type, DataType index_type) { in MakeOp() argument 46 .Input(FakeInput(index_type)) in MakeOp() 47 .Input(FakeInput(index_type)) in MakeOp()
|
/external/libcxx/include/ |
D | span | 43 using index_type = ptrdiff_t; 51 static constexpr index_type extent = Extent; 55 constexpr span(pointer ptr, index_type count); 81 constexpr span<element_type, dynamic_extent> first(index_type count) const; 82 constexpr span<element_type, dynamic_extent> last(index_type count) const; 83 …constexpr span<element_type, dynamic_extent> subspan(index_type offset, index_type count = dynamic… 86 constexpr index_type size() const noexcept; 87 constexpr index_type size_bytes() const noexcept; 91 constexpr reference operator[](index_type idx) const; 92 constexpr reference operator()(index_type idx) const; [all …]
|
/external/tensorflow/tensorflow/compiler/tf2xla/kernels/ |
D | dynamic_slice_ops.cc | 37 DataType index_type = ctx->InputType("indices"); in Compile() local 38 CHECK(index_type == DT_INT32 || index_type == DT_INT64); in Compile() 71 DataType index_type = ctx->InputType("start_indices"); in Compile() local 72 CHECK(index_type == DT_INT32 || index_type == DT_INT64); in Compile() 73 CHECK(index_type == ctx->InputType("size_indices")); in Compile()
|
D | matrix_band_part_op.cc | 54 DataType index_type = context->input_type(1); in Compile() local 72 auto zero_index = XlaHelpers::Zero(builder, index_type); in Compile() 74 XlaHelpers::IntegerLiteral(builder, index_type, m), in Compile() 77 XlaHelpers::IntegerLiteral(builder, index_type, n), in Compile()
|
D | gather_op.cc | 32 DataType index_type, xla::XlaBuilder* builder, in XlaGather() argument 179 DataType index_type = input_type(1); in Compile() local 180 OP_REQUIRES(context, index_type == DT_INT32 || index_type == DT_INT64, in Compile() 186 /*indices_are_nd=*/false, input_type(0), index_type, in Compile()
|
D | index_ops.cc | 61 DataType index_type = output_type(0); in Compile() local 63 OP_REQUIRES_OK(ctx, DataTypeToPrimitiveType(index_type, &index_xla_type)); in Compile()
|
/external/tensorflow/tensorflow/lite/testing/nnapi_tflite_zip_tests/ |
D | test_manifest.txt | 1763 …e,constant_indices=False,dtype=tf.float32,end=[12,2,2,5],end_mask=None,index_type=tf.int32,input_s… 1764 …e,constant_indices=False,dtype=tf.float32,end=[12,2,2,5],end_mask=None,index_type=tf.int32,input_s… 1765 …one,constant_indices=False,dtype=tf.int32,end=[12,2,2,5],end_mask=None,index_type=tf.int32,input_s… 1766 …one,constant_indices=False,dtype=tf.int32,end=[12,2,2,5],end_mask=None,index_type=tf.int32,input_s… 1767 …one,constant_indices=False,dtype=tf.int64,end=[12,2,2,5],end_mask=None,index_type=tf.int32,input_s… 1768 …one,constant_indices=False,dtype=tf.int64,end=[12,2,2,5],end_mask=None,index_type=tf.int32,input_s… 1769 …ne,constant_indices=True,dtype=tf.float32,end=[12,2,2,5],end_mask=None,index_type=tf.int32,input_s… 1770 …ne,constant_indices=True,dtype=tf.float32,end=[12,2,2,5],end_mask=None,index_type=tf.int32,input_s… 1771 …None,constant_indices=True,dtype=tf.int32,end=[12,2,2,5],end_mask=None,index_type=tf.int32,input_s… 1772 …None,constant_indices=True,dtype=tf.int32,end=[12,2,2,5],end_mask=None,index_type=tf.int32,input_s… [all …]
|
/external/tensorflow/tensorflow/compiler/xla/service/llvm_ir/ |
D | loop_emitter.cc | 88 absl::string_view loop_name, llvm::Type* index_type) { in EmitIndexAndSetExitBasicBlock() argument 89 CHECK_NE(index_type, nullptr); in EmitIndexAndSetExitBasicBlock() 93 return {IrArray::Index(index_type)}; in EmitIndexAndSetExitBasicBlock() 110 IrArray::Index array_index(array_multi_index, shape_, index_type); in EmitIndexAndSetExitBasicBlock() 126 llvm::Type* index_type) { in EmitLoop() argument 127 if (index_type == nullptr) { in EmitLoop() 128 index_type = b_->getInt64Ty(); in EmitLoop() 132 EmitIndexAndSetExitBasicBlock(loop_name, index_type)) { in EmitLoop()
|
D | loop_emitter.h | 71 absl::string_view loop_name, llvm::Type* index_type); 75 llvm::Type* index_type = nullptr);
|
/external/Microsoft-GSL/include/gsl/ |
D | span | 137 using difference_type = typename Span::index_type; 144 constexpr span_iterator(const Span* span, typename Span::index_type idx) noexcept 307 using index_type = std::ptrdiff_t; 313 template <index_type Other> 321 constexpr extent_type(index_type size) { Expects(size == Ext); } 323 constexpr index_type size() const noexcept { return Ext; } 330 using index_type = std::ptrdiff_t; 332 template <index_type Other> 336 explicit constexpr extent_type(index_type size) : size_(size) { Expects(size >= 0); } 338 constexpr index_type size() const noexcept { return size_; } [all …]
|
D | string_span | 183 using index_type = typename impl_type::index_type; 198 constexpr basic_string_span(pointer ptr, index_type length) : span_(ptr, length) {} 255 template <index_type Count> 261 constexpr basic_string_span<element_type, dynamic_extent> first(index_type count) const 266 template <index_type Count> 272 constexpr basic_string_span<element_type, dynamic_extent> last(index_type count) const 277 template <index_type Offset, index_type Count> 284 subspan(index_type offset, index_type count = dynamic_extent) const 289 constexpr reference operator[](index_type idx) const { return span_[idx]; } 290 constexpr reference operator()(index_type idx) const { return span_[idx]; } [all …]
|
D | multi_span | 591 using index_type = multi_span_index<rank>; 592 using const_index_type = std::add_const_t<index_type>; 677 constexpr size_type linearize(const index_type& idx) const { return m_ranges.linearize(idx); } 679 constexpr bool contains(const index_type& idx) const noexcept 708 constexpr index_type index_bounds() const noexcept 729 return const_iterator(*this, index_type{}); 750 using index_type = multi_span_index<rank>; 751 using const_index_type = std::add_const_t<index_type>; 763 constexpr strided_bounds(const value_type (&values)[rank], index_type strides) 767 constexpr strided_bounds(const index_type& extents, const index_type& strides) noexcept [all …]
|
/external/tensorflow/tensorflow/lite/kernels/ |
D | slice_test.cc | 26 template <typename input_type, typename index_type> 45 void SetBegin(std::initializer_list<index_type> data) { in SetBegin() 46 PopulateTensor<index_type>(begin_, data); in SetBegin() 48 void SetSize(std::initializer_list<index_type> data) { in SetSize() 49 PopulateTensor<index_type>(size_, data); in SetSize()
|
/external/tensorflow/tensorflow/compiler/xla/service/cpu/ |
D | parallel_loop_emitter.cc | 34 llvm::Type* index_type) { in EmitIndexAndSetExitBasicBlock() argument 35 CHECK_NE(index_type, nullptr); in EmitIndexAndSetExitBasicBlock() 74 llvm_ir::IrArray::Index array_index(array_multi_index, shape_, index_type); in EmitIndexAndSetExitBasicBlock()
|
/external/skqp/src/sksl/ir/ |
D | SkSLIndexExpression.h | 20 static const Type& index_type(const Context& context, const Type& type) { in index_type() function 55 : INHERITED(base->fOffset, kIndex_Kind, index_type(context, base->fType)) in IndexExpression()
|
/external/skia/src/sksl/ir/ |
D | SkSLIndexExpression.h | 20 static const Type& index_type(const Context& context, const Type& type) { in index_type() function 55 : INHERITED(base->fOffset, kIndex_Kind, index_type(context, base->fType)) in IndexExpression()
|