Searched defs:_Options (Results 1 - 3 of 3) sorted by relevance
/external/autotest/client/site_tests/firmware_TouchMTB/ |
H A D | firmware_constants.py | 173 class _Options(_Constant): class in inherits:_Constant 176 OPTIONS = _Options()
|
/external/eigen/Eigen/src/SparseCore/ |
H A D | SparseUtil.h | 100 enum { _Options = ((Flags&RowMajorBit)==RowMajorBit) ? RowMajor : ColMajor }; enumerator in enum:Eigen::internal::sparse_eval::__anon5841 102 typedef SparseMatrix<_Scalar, _Options, _StorageIndex> type; 115 enum { _Options = ((evaluator<T>::Flags&RowMajorBit)==RowMajorBit) ? RowMajor : ColMajor }; enumerator in enum:Eigen::internal::plain_matrix_type::__anon5842 117 typedef SparseMatrix<_Scalar, _Options, _StorageIndex> type;
|
H A D | SparseMatrix.h | 33 * \tparam _Options Union of bit flags controlling the storage scheme. Currently the only possibility 46 template<typename _Scalar, int _Options, typename _StorageIndex> 47 struct traits<SparseMatrix<_Scalar, _Options, _StorageIndex> > 58 Flags = _Options | NestByRefBit | LvalueBit | CompressedAccessBit, 63 template<typename _Scalar, int _Options, typename _StorageIndex, int DiagIndex> 64 struct traits<Diagonal<SparseMatrix<_Scalar, _Options, _StorageIndex>, DiagIndex> > 66 typedef SparseMatrix<_Scalar, _Options, _StorageIndex> MatrixType; 84 template<typename _Scalar, int _Options, typename _StorageIndex, int DiagIndex> 85 struct traits<Diagonal<const SparseMatrix<_Scalar, _Options, _StorageIndex>, DiagIndex> > 86 : public traits<Diagonal<SparseMatrix<_Scalar, _Options, _StorageInde 1192 eigen_internal_assert(data_end < m_data.allocatedSize()); StorageIndex new_end = convert_index(m_data.allocatedSize()); for(Index k=outer+1; k<=m_outerSize; ++k) if(m_outerIndex[k]==data_end) m_outerIndex[k] = new_end; } return m_data.value(p); } if(m_outerIndex[outer+1]==data_end && m_outerIndex[outer]+m_innerNonZeros[outer]==m_data.size()) { eigen_internal_assert(outer+1==m_outerSize || m_innerNonZeros[outer+1]==0); ++m_innerNonZeros[outer]; m_data.resize(m_data.size()+1); if(data_end != m_data.allocatedSize()) { eigen_internal_assert(data_end < m_data.allocatedSize()); StorageIndex new_end = convert_index(m_data.allocatedSize()); for(Index k=outer+1; k<=m_outerSize; ++k) if(m_outerIndex[k]==data_end) m_outerIndex[k] = new_end; } Index startId = m_outerIndex[outer]; Index p = m_outerIndex[outer]+m_innerNonZeros[outer]-1; while ( (p > startId) && (m_data.index(p-1) > inner) ) { m_data.index(p) = m_data.index(p-1); m_data.value(p) = m_data.value(p-1); --p; } m_data.index(p) = convert_index(inner); return (m_data.value(p) = 0); } if(m_data.size() != m_data.allocatedSize()) { m_data.resize(m_data.allocatedSize()); this->reserveInnerVectors(Array<StorageIndex,Dynamic,1>::Constant(m_outerSize, 2)); } return insertUncompressed(row,col); } template<typename _Scalar, int _Options, typename _StorageIndex> EIGEN_DONT_INLINE typename SparseMatrix<_Scalar,_Options,_StorageIndex>::Scalar& SparseMatrix<_Scalar,_Options,_StorageIndex>::insertUncompressed(Index row, Index col) { eigen_assert(!isCompressed()); const Index outer = IsRowMajor ? row : col; const StorageIndex inner = convert_index(IsRowMajor ? col : row); Index room = m_outerIndex[outer+1] - m_outerIndex[outer]; StorageIndex innerNNZ = m_innerNonZeros[outer]; if(innerNNZ>=room) argument [all...] |
Completed in 993 milliseconds