/external/eigen/Eigen/src/Core/ |
D | DenseStorage.h | 252 Index m_cols; 254 EIGEN_DEVICE_FUNC DenseStorage() : m_rows(0), m_cols(0) {} 256 : m_data(internal::constructor_without_unaligned_array_assert()), m_rows(0), m_cols(0) {} 257 …ge(const DenseStorage& other) : m_data(other.m_data), m_rows(other.m_rows), m_cols(other.m_cols) {} 264 m_cols = other.m_cols; 268 EIGEN_DEVICE_FUNC DenseStorage(Index, Index rows, Index cols) : m_rows(rows), m_cols(cols) {} 270 …{ std::swap(m_data,other.m_data); std::swap(m_rows,other.m_rows); std::swap(m_cols,other.m_cols); } 272 EIGEN_DEVICE_FUNC Index cols() const {return m_cols;} 273 …EVICE_FUNC void conservativeResize(Index, Index rows, Index cols) { m_rows = rows; m_cols = cols; } 274 EIGEN_DEVICE_FUNC void resize(Index, Index rows, Index cols) { m_rows = rows; m_cols = cols; } [all …]
|
D | MapBase.h | 91 EIGEN_DEVICE_FUNC inline Index cols() const { return m_cols.value(); } in cols() 149 …apBase(PointerType dataPtr) : m_data(dataPtr), m_rows(RowsAtCompileTime), m_cols(ColsAtCompileTime) in MapBase() 160 m_cols(ColsAtCompileTime == Dynamic ? vecSize : Index(ColsAtCompileTime)) in MapBase() 171 : m_data(dataPtr), m_rows(rows), m_cols(cols) in MapBase() 202 const internal::variable_if_dynamic<Index, ColsAtCompileTime> m_cols; variable
|
D | CoreEvaluators.h | 1192 m_cols(replicate.nestedExpression().cols()) 1204 : col % m_cols.value(); 1214 ? (ColFactor==1 ? index : index%m_cols.value()) 1229 : col % m_cols.value(); 1239 ? (ColFactor==1 ? index : index%m_cols.value()) 1249 const variable_if_dynamic<Index, ArgType::ColsAtCompileTime> m_cols; 1449 m_cols(ReverseCol ? reverse.nestedExpression().cols() : 1) 1456 ReverseCol ? m_cols.value() - col - 1 : col); 1462 return m_argImpl.coeff(m_rows.value() * m_cols.value() - index - 1); 1469 ReverseCol ? m_cols.value() - col - 1 : col); [all …]
|
D | CwiseNullaryOp.h | 69 : m_rows(rows), m_cols(cols), m_functor(func) 80 EIGEN_STRONG_INLINE Index cols() const { return m_cols.value(); } 88 const internal::variable_if_dynamic<Index, ColsAtCompileTime> m_cols;
|
/external/tensorflow/tensorflow/lite/kernels/internal/optimized/ |
D | neon_tensor_utils.cc | 98 int m_cols, const float* vector, in NeonMatrixBatchVectorMultiplyAccumulate() argument 105 m_cols - (m_cols & (kFloatWeightsPerNeonLane - 1)); in NeonMatrixBatchVectorMultiplyAccumulate() 109 const float* vector_in_batch = vector + b * m_cols; in NeonMatrixBatchVectorMultiplyAccumulate() 127 for (int c = postamble_start; c < m_cols; c++) { in NeonMatrixBatchVectorMultiplyAccumulate() 130 matrix_row += m_cols; in NeonMatrixBatchVectorMultiplyAccumulate() 151 const int m_cols, void** shuffled_vectors_free) { in ShuffleVectors() argument 155 kWeightsPerUint32, n_batch * m_cols, shuffled_vectors_free)); in ShuffleVectors() 158 int8* shuffled_vectors_ptr = shuffled_vectors + (i * m_cols); in ShuffleVectors() 160 reinterpret_cast<const int8*>(vectors) + (i * m_cols); in ShuffleVectors() 162 reinterpret_cast<const int8*>(vectors) + ((i + 1) * m_cols); in ShuffleVectors() [all …]
|
D | tensor_utils_impl.h | 38 int m_rows, int m_cols, 43 int m_cols, const float* vector, 49 const int8_t* __restrict__ matrix, const int m_rows, const int m_cols, 53 const int8_t* __restrict__ matrix, const int m_rows, const int m_cols, 58 const float* matrix, const uint8_t* ledger, int m_rows, int m_cols, 61 const float* matrix, const uint8_t* ledger, int m_rows, int m_cols, 67 const int m_cols, const int8_t* __restrict__ vectors, 72 const int m_cols, const int8_t* __restrict__ vectors,
|
D | neon_tensor_utils.h | 28 int m_cols, const float* vector, in MatrixBatchVectorMultiplyAccumulate() argument 31 NEON_OR_PORTABLE(MatrixBatchVectorMultiplyAccumulate, matrix, m_rows, m_cols, in MatrixBatchVectorMultiplyAccumulate() 36 const int8_t* __restrict__ matrix, const int m_rows, const int m_cols, in MatrixBatchVectorMultiplyAccumulate() argument 39 NEON_OR_PORTABLE(MatrixBatchVectorMultiplyAccumulate, matrix, m_rows, m_cols, in MatrixBatchVectorMultiplyAccumulate() 45 const int m_cols, const float* vector, int n_batch, float* result, in SparseMatrixBatchVectorMultiplyAccumulate() argument 48 matrix, ledger, m_rows, m_cols, vector, n_batch, result, result_stride); in SparseMatrixBatchVectorMultiplyAccumulate() 53 const int m_cols, const int8_t* __restrict__ vectors, in SparseMatrixBatchVectorMultiplyAccumulate() argument 56 NeonSparseMatrixBatchVectorMultiplyAccumulate(matrix, ledger, m_rows, m_cols, in SparseMatrixBatchVectorMultiplyAccumulate()
|
/external/tensorflow/tensorflow/lite/kernels/internal/reference/ |
D | portable_tensor_utils.h | 41 int m_rows, int m_cols, 47 const int8_t* __restrict__ matrix, const int m_rows, const int m_cols, 52 const float* matrix, const uint8_t* ledger, int m_rows, int m_cols, 57 const int m_cols, const int8_t* __restrict__ vectors, 162 int m_cols, const float* vector, in MatrixBatchVectorMultiplyAccumulate() argument 165 PortableMatrixBatchVectorMultiplyAccumulate(matrix, m_rows, m_cols, vector, in MatrixBatchVectorMultiplyAccumulate() 170 const int8_t* __restrict__ matrix, const int m_rows, const int m_cols, in MatrixBatchVectorMultiplyAccumulate() argument 173 PortableMatrixBatchVectorMultiplyAccumulate(matrix, m_rows, m_cols, vector, in MatrixBatchVectorMultiplyAccumulate() 179 const float* matrix, const uint8_t* ledger, int m_rows, int m_cols, in SparseMatrixBatchVectorMultiplyAccumulate() argument 182 matrix, ledger, m_rows, m_cols, vector, n_batch, result, result_stride); in SparseMatrixBatchVectorMultiplyAccumulate() [all …]
|
D | portable_tensor_utils.cc | 69 int m_rows, int m_cols, in PortableMatrixBatchVectorMultiplyAccumulate() argument 78 const float* vector_in_batch = vector + b * m_cols; in PortableMatrixBatchVectorMultiplyAccumulate() 79 for (int c = 0; c < m_cols; c++) { in PortableMatrixBatchVectorMultiplyAccumulate() 89 const int8_t* __restrict__ matrix, const int m_rows, const int m_cols, in PortableMatrixBatchVectorMultiplyAccumulate() argument 93 for (batch = 0; batch < n_batch; ++batch, vectors += m_cols) { in PortableMatrixBatchVectorMultiplyAccumulate() 105 for (col = 0; col < m_cols; ++col, ++row_ptr) { in PortableMatrixBatchVectorMultiplyAccumulate() 114 const float* matrix, const uint8_t* ledger, int m_rows, int m_cols, in PortableSparseMatrixBatchVectorMultiplyAccumulate() argument 118 m_cols % kBlockSize, 0); in PortableSparseMatrixBatchVectorMultiplyAccumulate() 127 const float* vector_in_batch = vector + b * m_cols; in PortableSparseMatrixBatchVectorMultiplyAccumulate() 145 const int m_cols, const int8_t* __restrict__ vectors, in PortableSparseMatrixBatchVectorMultiplyAccumulate() argument [all …]
|
/external/eigen/debug/msvc/ |
D | eigen_autoexp_part.dat | 128 cols: $c.m_storage.m_cols, 134 …expr: ($c.m_storage.m_data)[($i % $c.m_storage.m_rows)*$c.m_storage.m_cols + (($i- $i % $c.m_stora… 135 size: ($r==1)*$c.m_storage.m_rows+($r==0)*$c.m_storage.m_cols 142 size: ($r==1)*$c.m_storage.m_rows+($r==0)*$c.m_storage.m_cols 154 $c.m_storage.m_cols, 158 size : $c.m_storage.m_rows*$c.m_storage.m_cols 172 cols: $c.m_storage.m_cols, 178 …expr: ($c.m_storage.m_data)[($i % $c.RowsAtCompileTime)*$c.m_storage.m_cols + (($i- $i % $c.RowsAt… 179 size: ($r==1)*$c.RowsAtCompileTime+($r==0)*$c.m_storage.m_cols 186 size: ($r==1)*$c.RowsAtCompileTime+($r==0)*$c.m_storage.m_cols [all …]
|
/external/eigen/Eigen/src/SVD/ |
D | SVDBase.h | 193 inline Index cols() const { return m_cols; } in cols() 236 Index m_nonzeroSingularValues, m_rows, m_cols, m_diagSize; variable 248 m_rows(-1), m_cols(-1), m_diagSize(0) in SVDBase() 281 cols == m_cols && in allocate() 288 m_cols = cols; in allocate() 301 m_diagSize = (std::min)(m_rows, m_cols); in allocate() 306 m_matrixV.resize(m_cols, m_computeFullV ? m_cols : m_computeThinV ? m_diagSize : 0); in allocate()
|
D | JacobiSVD.h | 597 using Base::m_cols; 619 cols == m_cols && 626 m_cols = cols; 644 m_diagSize = (std::min)(m_rows, m_cols); 651 m_matrixV.resize(m_cols, m_computeFullV ? m_cols 656 if(m_cols>m_rows) m_qr_precond_morecols.allocate(*this); 657 if(m_rows>m_cols) m_qr_precond_morerows.allocate(*this); 658 if(m_rows!=m_cols) m_scaledMatrix.resize(rows,cols); 681 if(m_rows!=m_cols) 692 if(m_computeFullV) m_matrixV.setIdentity(m_cols,m_cols); [all …]
|
D | JacobiSVD_LAPACKE.h | 64 …ldvt = (m_computeFullV) ? internal::convert_index<lapack_int>(m_cols) : (m_computeThinV) ? interna… 66 localV.resize(ldvt, m_cols); \ 71 …::convert_index<lapack_int>(m_rows), internal::convert_index<lapack_int>(m_cols), (LAPACKE_TYPE*)m…
|
/external/tensorflow/tensorflow/lite/kernels/internal/ |
D | tensor_utils.h | 54 int m_cols, const float* vector, 70 const float* matrix, const uint8_t* ledger, int m_rows, int m_cols, 81 const int8_t* __restrict__ matrix, const int m_rows, const int m_cols, 98 const int m_cols, const int8_t* __restrict__ vectors,
|
/external/eigen/Eigen/src/misc/ |
D | Kernel.h | 46 m_cols(m_rank==dec.cols() ? 1 : dec.cols() - m_rank) 50 inline Index cols() const { return m_cols; } 61 Index m_rank, m_cols;
|
D | Image.h | 44 m_cols(m_rank == 0 ? 1 : m_rank), 49 inline Index cols() const { return m_cols; } 61 Index m_rank, m_cols;
|