Searched refs:m_cols (Results 1 - 15 of 15) sorted by relevance

/external/eigen/Eigen/src/Core/
H A DDenseStorage.h252 Index m_cols; member in class:Eigen::DenseStorage
254 EIGEN_DEVICE_FUNC DenseStorage() : m_rows(0), m_cols(0) {}
256 : m_data(internal::constructor_without_unaligned_array_assert()), m_rows(0), m_cols(0) {}
257 EIGEN_DEVICE_FUNC DenseStorage(const DenseStorage& other) : m_data(other.m_data), m_rows(other.m_rows), m_cols(other.m_cols) {}
264 m_cols = other.m_cols;
268 EIGEN_DEVICE_FUNC DenseStorage(Index, Index rows, Index cols) : m_rows(rows), m_cols(cols) {}
270 { std::swap(m_data,other.m_data); std::swap(m_rows,other.m_rows); std::swap(m_cols,other.m_cols); }
312 Index m_cols; member in class:Eigen::DenseStorage
342 Index m_cols; member in class:Eigen::DenseStorage
424 Index m_cols; member in class:Eigen::DenseStorage
[all...]
H A DMapBase.h91 EIGEN_DEVICE_FUNC inline Index cols() const { return m_cols.value(); }
149 explicit inline MapBase(PointerType dataPtr) : m_data(dataPtr), m_rows(RowsAtCompileTime), m_cols(ColsAtCompileTime)
160 m_cols(ColsAtCompileTime == Dynamic ? vecSize : Index(ColsAtCompileTime))
171 : m_data(dataPtr), m_rows(rows), m_cols(cols)
202 const internal::variable_if_dynamic<Index, ColsAtCompileTime> m_cols; member in class:Eigen::MapBase
H A DCoreEvaluators.h1192 m_cols(replicate.nestedExpression().cols())
1204 : col % m_cols.value();
1214 ? (ColFactor==1 ? index : index%m_cols.value())
1229 : col % m_cols.value();
1239 ? (ColFactor==1 ? index : index%m_cols.value())
1249 const variable_if_dynamic<Index, ArgType::ColsAtCompileTime> m_cols; member in struct:Eigen::internal::unary_evaluator
1449 m_cols(ReverseCol ? reverse.nestedExpression().cols() : 1)
1456 ReverseCol ? m_cols.value() - col - 1 : col);
1462 return m_argImpl.coeff(m_rows.value() * m_cols.value() - index - 1);
1469 ReverseCol ? m_cols
1533 const variable_if_dynamic<Index, ReverseCol ? ArgType::ColsAtCompileTime : 1> m_cols; member in struct:Eigen::internal::unary_evaluator
[all...]
H A DCwiseNullaryOp.h69 : m_rows(rows), m_cols(cols), m_functor(func)
80 EIGEN_STRONG_INLINE Index cols() const { return m_cols.value(); }
88 const internal::variable_if_dynamic<Index, ColsAtCompileTime> m_cols; member in class:Eigen::CwiseNullaryOp
/external/eigen/Eigen/src/misc/
H A DKernel.h46 m_cols(m_rank==dec.cols() ? 1 : dec.cols() - m_rank)
50 inline Index cols() const { return m_cols; }
61 Index m_rank, m_cols; member in struct:Eigen::internal::kernel_retval_base
H A DImage.h44 m_cols(m_rank == 0 ? 1 : m_rank),
49 inline Index cols() const { return m_cols; }
61 Index m_rank, m_cols; member in struct:Eigen::internal::image_retval_base
/external/tensorflow/tensorflow/contrib/lite/kernels/internal/optimized/
H A Dneon_tensor_utils.cc31 int m_cols, const float* vector,
38 m_cols - (m_cols & (kFloatWeightsPerNeonLane - 1));
42 new float32x4_t[(m_cols / kFloatWeightsPerNeonLane) *
47 const float* vector_in_batch = vector + b * m_cols;
53 matrix_ptr1 = matrix + m_cols;
83 for (int c = postamble_start; c < m_cols; c++) {
88 matrix_ptr0 += kUnrollSize * m_cols;
89 matrix_ptr1 += kUnrollSize * m_cols;
106 for (int c = postamble_start; c < m_cols;
30 NeonMatrixBatchVectorMultiplyAccumulate(const float* matrix, int m_rows, int m_cols, const float* vector, int n_batch, float* result, int result_stride) argument
[all...]
H A Dtensor_utils_impl.h34 int m_rows, int m_cols,
39 int m_cols, const float* vector,
H A Dneon_tensor_utils.h28 int m_cols, const float* vector,
31 NEON_OR_PORTABLE(MatrixBatchVectorMultiplyAccumulate, matrix, m_rows, m_cols,
27 MatrixBatchVectorMultiplyAccumulate(const float* matrix, int m_rows, int m_cols, const float* vector, int n_batch, float* result, int result_stride) argument
/external/eigen/Eigen/src/SVD/
H A DSVDBase.h193 inline Index cols() const { return m_cols; }
236 Index m_nonzeroSingularValues, m_rows, m_cols, m_diagSize; member in class:Eigen::SVDBase
248 m_rows(-1), m_cols(-1), m_diagSize(0)
281 cols == m_cols &&
288 m_cols = cols;
301 m_diagSize = (std::min)(m_rows, m_cols);
306 m_matrixV.resize(m_cols, m_computeFullV ? m_cols : m_computeThinV ? m_diagSize : 0);
H A DJacobiSVD_LAPACKE.h64 ldvt = (m_computeFullV) ? internal::convert_index<lapack_int>(m_cols) : (m_computeThinV) ? internal::convert_index<lapack_int>(m_diagSize) : 1; \
66 localV.resize(ldvt, m_cols); \
71 LAPACKE_##LAPACKE_PREFIX##gesvd( matrix_order, jobu, jobvt, internal::convert_index<lapack_int>(m_rows), internal::convert_index<lapack_int>(m_cols), (LAPACKE_TYPE*)m_temp.data(), lda, (LAPACKE_RTYPE*)m_singularValues.data(), u, ldu, vt, ldvt, superb.data()); \
H A DJacobiSVD.h597 using Base::m_cols;
619 cols == m_cols &&
626 m_cols = cols;
644 m_diagSize = (std::min)(m_rows, m_cols);
651 m_matrixV.resize(m_cols, m_computeFullV ? m_cols
656 if(m_cols>m_rows) m_qr_precond_morecols.allocate(*this);
657 if(m_rows>m_cols) m_qr_precond_morerows.allocate(*this);
658 if(m_rows!=m_cols) m_scaledMatrix.resize(rows,cols);
681 if(m_rows!=m_cols)
[all...]
/external/tensorflow/tensorflow/contrib/lite/kernels/internal/reference/
H A Dportable_tensor_utils.cc31 int m_rows, int m_cols,
39 const float* vector_in_batch = vector + b * m_cols;
40 for (int c = 0; c < m_cols; c++) {
30 PortableMatrixBatchVectorMultiplyAccumulate(const float* matrix, int m_rows, int m_cols, const float* vector, int n_batch, float* result, int result_stride) argument
H A Dportable_tensor_utils.h31 int m_rows, int m_cols,
107 int m_cols, const float* vector,
110 PortableMatrixBatchVectorMultiplyAccumulate(matrix, m_rows, m_cols, vector,
106 MatrixBatchVectorMultiplyAccumulate(const float* matrix, int m_rows, int m_cols, const float* vector, int n_batch, float* result, int result_stride) argument
/external/tensorflow/tensorflow/contrib/lite/kernels/internal/
H A Dtensor_utils.h33 int m_cols, const float* vector,

Completed in 652 milliseconds