/external/eigen/Eigen/src/Core/ |
H A D | DenseStorage.h | 252 Index m_cols; member in class:Eigen::DenseStorage 254 EIGEN_DEVICE_FUNC DenseStorage() : m_rows(0), m_cols(0) {} 256 : m_data(internal::constructor_without_unaligned_array_assert()), m_rows(0), m_cols(0) {} 257 EIGEN_DEVICE_FUNC DenseStorage(const DenseStorage& other) : m_data(other.m_data), m_rows(other.m_rows), m_cols(other.m_cols) {} 264 m_cols = other.m_cols; 268 EIGEN_DEVICE_FUNC DenseStorage(Index, Index rows, Index cols) : m_rows(rows), m_cols(cols) {} 270 { std::swap(m_data,other.m_data); std::swap(m_rows,other.m_rows); std::swap(m_cols,other.m_cols); } 312 Index m_cols; member in class:Eigen::DenseStorage 342 Index m_cols; member in class:Eigen::DenseStorage 424 Index m_cols; member in class:Eigen::DenseStorage [all...] |
H A D | MapBase.h | 91 EIGEN_DEVICE_FUNC inline Index cols() const { return m_cols.value(); } 149 explicit inline MapBase(PointerType dataPtr) : m_data(dataPtr), m_rows(RowsAtCompileTime), m_cols(ColsAtCompileTime) 160 m_cols(ColsAtCompileTime == Dynamic ? vecSize : Index(ColsAtCompileTime)) 171 : m_data(dataPtr), m_rows(rows), m_cols(cols) 202 const internal::variable_if_dynamic<Index, ColsAtCompileTime> m_cols; member in class:Eigen::MapBase
|
H A D | CoreEvaluators.h | 1192 m_cols(replicate.nestedExpression().cols()) 1204 : col % m_cols.value(); 1214 ? (ColFactor==1 ? index : index%m_cols.value()) 1229 : col % m_cols.value(); 1239 ? (ColFactor==1 ? index : index%m_cols.value()) 1249 const variable_if_dynamic<Index, ArgType::ColsAtCompileTime> m_cols; member in struct:Eigen::internal::unary_evaluator 1449 m_cols(ReverseCol ? reverse.nestedExpression().cols() : 1) 1456 ReverseCol ? m_cols.value() - col - 1 : col); 1462 return m_argImpl.coeff(m_rows.value() * m_cols.value() - index - 1); 1469 ReverseCol ? m_cols 1533 const variable_if_dynamic<Index, ReverseCol ? ArgType::ColsAtCompileTime : 1> m_cols; member in struct:Eigen::internal::unary_evaluator [all...] |
H A D | CwiseNullaryOp.h | 69 : m_rows(rows), m_cols(cols), m_functor(func) 80 EIGEN_STRONG_INLINE Index cols() const { return m_cols.value(); } 88 const internal::variable_if_dynamic<Index, ColsAtCompileTime> m_cols; member in class:Eigen::CwiseNullaryOp
|
/external/eigen/Eigen/src/misc/ |
H A D | Kernel.h | 46 m_cols(m_rank==dec.cols() ? 1 : dec.cols() - m_rank) 50 inline Index cols() const { return m_cols; } 61 Index m_rank, m_cols; member in struct:Eigen::internal::kernel_retval_base
|
H A D | Image.h | 44 m_cols(m_rank == 0 ? 1 : m_rank), 49 inline Index cols() const { return m_cols; } 61 Index m_rank, m_cols; member in struct:Eigen::internal::image_retval_base
|
/external/tensorflow/tensorflow/contrib/lite/kernels/internal/optimized/ |
H A D | neon_tensor_utils.cc | 31 int m_cols, const float* vector, 38 m_cols - (m_cols & (kFloatWeightsPerNeonLane - 1)); 42 new float32x4_t[(m_cols / kFloatWeightsPerNeonLane) * 47 const float* vector_in_batch = vector + b * m_cols; 53 matrix_ptr1 = matrix + m_cols; 83 for (int c = postamble_start; c < m_cols; c++) { 88 matrix_ptr0 += kUnrollSize * m_cols; 89 matrix_ptr1 += kUnrollSize * m_cols; 106 for (int c = postamble_start; c < m_cols; 30 NeonMatrixBatchVectorMultiplyAccumulate(const float* matrix, int m_rows, int m_cols, const float* vector, int n_batch, float* result, int result_stride) argument [all...] |
H A D | tensor_utils_impl.h | 34 int m_rows, int m_cols, 39 int m_cols, const float* vector,
|
H A D | neon_tensor_utils.h | 28 int m_cols, const float* vector, 31 NEON_OR_PORTABLE(MatrixBatchVectorMultiplyAccumulate, matrix, m_rows, m_cols, 27 MatrixBatchVectorMultiplyAccumulate(const float* matrix, int m_rows, int m_cols, const float* vector, int n_batch, float* result, int result_stride) argument
|
/external/eigen/Eigen/src/SVD/ |
H A D | SVDBase.h | 193 inline Index cols() const { return m_cols; } 236 Index m_nonzeroSingularValues, m_rows, m_cols, m_diagSize; member in class:Eigen::SVDBase 248 m_rows(-1), m_cols(-1), m_diagSize(0) 281 cols == m_cols && 288 m_cols = cols; 301 m_diagSize = (std::min)(m_rows, m_cols); 306 m_matrixV.resize(m_cols, m_computeFullV ? m_cols : m_computeThinV ? m_diagSize : 0);
|
H A D | JacobiSVD_LAPACKE.h | 64 ldvt = (m_computeFullV) ? internal::convert_index<lapack_int>(m_cols) : (m_computeThinV) ? internal::convert_index<lapack_int>(m_diagSize) : 1; \ 66 localV.resize(ldvt, m_cols); \ 71 LAPACKE_##LAPACKE_PREFIX##gesvd( matrix_order, jobu, jobvt, internal::convert_index<lapack_int>(m_rows), internal::convert_index<lapack_int>(m_cols), (LAPACKE_TYPE*)m_temp.data(), lda, (LAPACKE_RTYPE*)m_singularValues.data(), u, ldu, vt, ldvt, superb.data()); \
|
H A D | JacobiSVD.h | 597 using Base::m_cols; 619 cols == m_cols && 626 m_cols = cols; 644 m_diagSize = (std::min)(m_rows, m_cols); 651 m_matrixV.resize(m_cols, m_computeFullV ? m_cols 656 if(m_cols>m_rows) m_qr_precond_morecols.allocate(*this); 657 if(m_rows>m_cols) m_qr_precond_morerows.allocate(*this); 658 if(m_rows!=m_cols) m_scaledMatrix.resize(rows,cols); 681 if(m_rows!=m_cols) [all...] |
/external/tensorflow/tensorflow/contrib/lite/kernels/internal/reference/ |
H A D | portable_tensor_utils.cc | 31 int m_rows, int m_cols, 39 const float* vector_in_batch = vector + b * m_cols; 40 for (int c = 0; c < m_cols; c++) { 30 PortableMatrixBatchVectorMultiplyAccumulate(const float* matrix, int m_rows, int m_cols, const float* vector, int n_batch, float* result, int result_stride) argument
|
H A D | portable_tensor_utils.h | 31 int m_rows, int m_cols, 107 int m_cols, const float* vector, 110 PortableMatrixBatchVectorMultiplyAccumulate(matrix, m_rows, m_cols, vector, 106 MatrixBatchVectorMultiplyAccumulate(const float* matrix, int m_rows, int m_cols, const float* vector, int n_batch, float* result, int result_stride) argument
|
/external/tensorflow/tensorflow/contrib/lite/kernels/internal/ |
H A D | tensor_utils.h | 33 int m_cols, const float* vector,
|