Searched defs:_Options (Results 1 - 3 of 3) sorted by relevance

/external/autotest/client/site_tests/firmware_TouchMTB/
H A Dfirmware_constants.py173 class _Options(_Constant): class in inherits:_Constant
176 OPTIONS = _Options()
/external/eigen/Eigen/src/SparseCore/
H A DSparseUtil.h100 enum { _Options = ((Flags&RowMajorBit)==RowMajorBit) ? RowMajor : ColMajor }; enumerator in enum:Eigen::internal::sparse_eval::__anon5841
102 typedef SparseMatrix<_Scalar, _Options, _StorageIndex> type;
115 enum { _Options = ((evaluator<T>::Flags&RowMajorBit)==RowMajorBit) ? RowMajor : ColMajor }; enumerator in enum:Eigen::internal::plain_matrix_type::__anon5842
117 typedef SparseMatrix<_Scalar, _Options, _StorageIndex> type;
H A DSparseMatrix.h33 * \tparam _Options Union of bit flags controlling the storage scheme. Currently the only possibility
46 template<typename _Scalar, int _Options, typename _StorageIndex>
47 struct traits<SparseMatrix<_Scalar, _Options, _StorageIndex> >
58 Flags = _Options | NestByRefBit | LvalueBit | CompressedAccessBit,
63 template<typename _Scalar, int _Options, typename _StorageIndex, int DiagIndex>
64 struct traits<Diagonal<SparseMatrix<_Scalar, _Options, _StorageIndex>, DiagIndex> >
66 typedef SparseMatrix<_Scalar, _Options, _StorageIndex> MatrixType;
84 template<typename _Scalar, int _Options, typename _StorageIndex, int DiagIndex>
85 struct traits<Diagonal<const SparseMatrix<_Scalar, _Options, _StorageIndex>, DiagIndex> >
86 : public traits<Diagonal<SparseMatrix<_Scalar, _Options, _StorageInde
1192 eigen_internal_assert(data_end < m_data.allocatedSize()); StorageIndex new_end = convert_index(m_data.allocatedSize()); for(Index k=outer+1; k<=m_outerSize; ++k) if(m_outerIndex[k]==data_end) m_outerIndex[k] = new_end; } return m_data.value(p); } if(m_outerIndex[outer+1]==data_end && m_outerIndex[outer]+m_innerNonZeros[outer]==m_data.size()) { eigen_internal_assert(outer+1==m_outerSize || m_innerNonZeros[outer+1]==0); ++m_innerNonZeros[outer]; m_data.resize(m_data.size()+1); if(data_end != m_data.allocatedSize()) { eigen_internal_assert(data_end < m_data.allocatedSize()); StorageIndex new_end = convert_index(m_data.allocatedSize()); for(Index k=outer+1; k<=m_outerSize; ++k) if(m_outerIndex[k]==data_end) m_outerIndex[k] = new_end; } Index startId = m_outerIndex[outer]; Index p = m_outerIndex[outer]+m_innerNonZeros[outer]-1; while ( (p > startId) && (m_data.index(p-1) > inner) ) { m_data.index(p) = m_data.index(p-1); m_data.value(p) = m_data.value(p-1); --p; } m_data.index(p) = convert_index(inner); return (m_data.value(p) = 0); } if(m_data.size() != m_data.allocatedSize()) { m_data.resize(m_data.allocatedSize()); this->reserveInnerVectors(Array<StorageIndex,Dynamic,1>::Constant(m_outerSize, 2)); } return insertUncompressed(row,col); } template<typename _Scalar, int _Options, typename _StorageIndex> EIGEN_DONT_INLINE typename SparseMatrix<_Scalar,_Options,_StorageIndex>::Scalar& SparseMatrix<_Scalar,_Options,_StorageIndex>::insertUncompressed(Index row, Index col) { eigen_assert(!isCompressed()); const Index outer = IsRowMajor ? row : col; const StorageIndex inner = convert_index(IsRowMajor ? col : row); Index room = m_outerIndex[outer+1] - m_outerIndex[outer]; StorageIndex innerNNZ = m_innerNonZeros[outer]; if(innerNNZ>=room) argument
[all...]

Completed in 993 milliseconds