|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
#ifndef EIGEN_SPARSE_BLOCKFORDYNAMICMATRIX_H |
|
|
#define EIGEN_SPARSE_BLOCKFORDYNAMICMATRIX_H |
|
|
|
|
|
namespace Eigen { |
|
|
|
|
|
#if 0 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
template<typename _Scalar, int _Options, typename _Index, int Size> |
|
|
class SparseInnerVectorSet<DynamicSparseMatrix<_Scalar, _Options, _Index>, Size> |
|
|
: public SparseMatrixBase<SparseInnerVectorSet<DynamicSparseMatrix<_Scalar, _Options, _Index>, Size> > |
|
|
{ |
|
|
typedef DynamicSparseMatrix<_Scalar, _Options, _Index> MatrixType; |
|
|
public: |
|
|
|
|
|
enum { IsRowMajor = internal::traits<SparseInnerVectorSet>::IsRowMajor }; |
|
|
|
|
|
EIGEN_SPARSE_PUBLIC_INTERFACE(SparseInnerVectorSet) |
|
|
class InnerIterator: public MatrixType::InnerIterator |
|
|
{ |
|
|
public: |
|
|
inline InnerIterator(const SparseInnerVectorSet& xpr, Index outer) |
|
|
: MatrixType::InnerIterator(xpr.m_matrix, xpr.m_outerStart + outer), m_outer(outer) |
|
|
{} |
|
|
inline Index row() const { return IsRowMajor ? m_outer : this->index(); } |
|
|
inline Index col() const { return IsRowMajor ? this->index() : m_outer; } |
|
|
protected: |
|
|
Index m_outer; |
|
|
}; |
|
|
|
|
|
inline SparseInnerVectorSet(const MatrixType& matrix, Index outerStart, Index outerSize) |
|
|
: m_matrix(matrix), m_outerStart(outerStart), m_outerSize(outerSize) |
|
|
{ |
|
|
eigen_assert( (outerStart>=0) && ((outerStart+outerSize)<=matrix.outerSize()) ); |
|
|
} |
|
|
|
|
|
inline SparseInnerVectorSet(const MatrixType& matrix, Index outer) |
|
|
: m_matrix(matrix), m_outerStart(outer), m_outerSize(Size) |
|
|
{ |
|
|
eigen_assert(Size!=Dynamic); |
|
|
eigen_assert( (outer>=0) && (outer<matrix.outerSize()) ); |
|
|
} |
|
|
|
|
|
template<typename OtherDerived> |
|
|
inline SparseInnerVectorSet& operator=(const SparseMatrixBase<OtherDerived>& other) |
|
|
{ |
|
|
if (IsRowMajor != ((OtherDerived::Flags&RowMajorBit)==RowMajorBit)) |
|
|
{ |
|
|
|
|
|
DynamicSparseMatrix<Scalar,IsRowMajor?RowMajorBit:0> aux(other); |
|
|
*this = aux.markAsRValue(); |
|
|
} |
|
|
else |
|
|
{ |
|
|
|
|
|
for (Index j=0; j<m_outerSize.value(); ++j) |
|
|
{ |
|
|
SparseVector<Scalar,IsRowMajor ? RowMajorBit : 0> aux(other.innerVector(j)); |
|
|
m_matrix.const_cast_derived()._data()[m_outerStart+j].swap(aux._data()); |
|
|
} |
|
|
} |
|
|
return *this; |
|
|
} |
|
|
|
|
|
inline SparseInnerVectorSet& operator=(const SparseInnerVectorSet& other) |
|
|
{ |
|
|
return operator=<SparseInnerVectorSet>(other); |
|
|
} |
|
|
|
|
|
Index nonZeros() const |
|
|
{ |
|
|
Index count = 0; |
|
|
for (Index j=0; j<m_outerSize.value(); ++j) |
|
|
count += m_matrix._data()[m_outerStart+j].size(); |
|
|
return count; |
|
|
} |
|
|
|
|
|
const Scalar& lastCoeff() const |
|
|
{ |
|
|
EIGEN_STATIC_ASSERT_VECTOR_ONLY(SparseInnerVectorSet); |
|
|
eigen_assert(m_matrix.data()[m_outerStart].size()>0); |
|
|
return m_matrix.data()[m_outerStart].vale(m_matrix.data()[m_outerStart].size()-1); |
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
EIGEN_STRONG_INLINE Index rows() const { return IsRowMajor ? m_outerSize.value() : m_matrix.rows(); } |
|
|
EIGEN_STRONG_INLINE Index cols() const { return IsRowMajor ? m_matrix.cols() : m_outerSize.value(); } |
|
|
|
|
|
protected: |
|
|
|
|
|
const typename MatrixType::Nested m_matrix; |
|
|
Index m_outerStart; |
|
|
const internal::variable_if_dynamic<Index, Size> m_outerSize; |
|
|
|
|
|
}; |
|
|
|
|
|
#endif |
|
|
|
|
|
} |
|
|
|
|
|
#endif |
|
|
|