// This file is part of Eigen, a lightweight C++ template library // for linear algebra. // // Copyright (C) 2013 Desire Nuentsa // Copyright (C) 2013 Gael Guennebaud // // This Source Code Form is subject to the terms of the Mozilla // Public License v. 2.0. If a copy of the MPL was not distributed // with this file, You can obtain one at http://mozilla.org/MPL/2.0/. #ifndef EIGEN_SPARSEBLOCKMATRIX_H #define EIGEN_SPARSEBLOCKMATRIX_H namespace Eigen { /** \ingroup SparseCore_Module * * \class BlockSparseMatrix * * \brief A versatile sparse matrix representation where each element is a block * * This class provides routines to manipulate block sparse matrices stored in a * BSR-like representation. There are two main types : * * 1. All blocks have the same number of rows and columns, called block size * in the following. In this case, if this block size is known at compile time, * it can be given as a template parameter like * \code * BlockSparseMatrix bmat(b_rows, b_cols); * \endcode * Here, bmat is a b_rows x b_cols block sparse matrix * where each coefficient is a 3x3 dense matrix. * If the block size is fixed but will be given at runtime, * \code * BlockSparseMatrix bmat(b_rows, b_cols); * bmat.setBlockSize(block_size); * \endcode * * 2. The second case is for variable-block sparse matrices. * Here each block has its own dimensions. The only restriction is that all the blocks * in a row (resp. a column) should have the same number of rows (resp. of columns). * It is thus required in this case to describe the layout of the matrix by calling * setBlockLayout(rowBlocks, colBlocks). * * In any of the previous case, the matrix can be filled by calling setFromTriplets(). * A regular sparse matrix can be converted to a block sparse matrix and vice versa. * It is obviously required to describe the block layout beforehand by calling either * setBlockSize() for fixed-size blocks or setBlockLayout for variable-size blocks. * * \tparam _Scalar The Scalar type * \tparam _BlockAtCompileTime The block layout option. It takes the following values * Dynamic : block size known at runtime * a numeric number : fixed-size block known at compile time */ template class BlockSparseMatrix; template class BlockSparseMatrixView; namespace internal { template struct traits > { typedef _Scalar Scalar; typedef _Index Index; typedef Sparse StorageKind; // FIXME Where is it used ?? typedef MatrixXpr XprKind; enum { RowsAtCompileTime = Dynamic, ColsAtCompileTime = Dynamic, MaxRowsAtCompileTime = Dynamic, MaxColsAtCompileTime = Dynamic, BlockSize = _BlockAtCompileTime, Flags = _Options | NestByRefBit | LvalueBit, CoeffReadCost = NumTraits::ReadCost, SupportedAccessPatterns = InnerRandomAccessPattern }; }; template struct traits > { typedef Ref > Scalar; typedef Ref > RealScalar; }; // Function object to sort a triplet list template struct TripletComp { typedef typename Iterator::value_type Triplet; bool operator()(const Triplet& a, const Triplet& b) { if(IsColMajor) return ((a.col() == b.col() && a.row() < b.row()) || (a.col() < b.col())); else return ((a.row() == b.row() && a.col() < b.col()) || (a.row() < b.row())); } }; } // end namespace internal /* Proxy to view the block sparse matrix as a regular sparse matrix */ template class BlockSparseMatrixView : public SparseMatrixBase { public: typedef Ref Scalar; typedef Ref RealScalar; typedef typename BlockSparseMatrixT::Index Index; typedef BlockSparseMatrixT Nested; enum { Flags = BlockSparseMatrixT::Options, Options = BlockSparseMatrixT::Options, RowsAtCompileTime = BlockSparseMatrixT::RowsAtCompileTime, ColsAtCompileTime = BlockSparseMatrixT::ColsAtCompileTime, MaxColsAtCompileTime = BlockSparseMatrixT::MaxColsAtCompileTime, MaxRowsAtCompileTime = BlockSparseMatrixT::MaxRowsAtCompileTime }; public: BlockSparseMatrixView(const BlockSparseMatrixT& spblockmat) : m_spblockmat(spblockmat) {} Index outerSize() const { return (Flags&RowMajorBit) == 1 ? this->rows() : this->cols(); } Index cols() const { return m_spblockmat.blockCols(); } Index rows() const { return m_spblockmat.blockRows(); } Scalar coeff(Index row, Index col) { return m_spblockmat.coeff(row, col); } Scalar coeffRef(Index row, Index col) { return m_spblockmat.coeffRef(row, col); } // Wrapper to iterate over all blocks class InnerIterator : public BlockSparseMatrixT::BlockInnerIterator { public: InnerIterator(const BlockSparseMatrixView& mat, Index outer) : BlockSparseMatrixT::BlockInnerIterator(mat.m_spblockmat, outer) {} }; protected: const BlockSparseMatrixT& m_spblockmat; }; // Proxy to view a regular vector as a block vector template class BlockVectorView { public: enum { BlockSize = BlockSparseMatrixT::BlockSize, ColsAtCompileTime = VectorType::ColsAtCompileTime, RowsAtCompileTime = VectorType::RowsAtCompileTime, Flags = VectorType::Flags }; typedef Ref >Scalar; typedef typename BlockSparseMatrixT::Index Index; public: BlockVectorView(const BlockSparseMatrixT& spblockmat, const VectorType& vec) : m_spblockmat(spblockmat),m_vec(vec) { } inline Index cols() const { return m_vec.cols(); } inline Index size() const { return m_spblockmat.blockRows(); } inline Scalar coeff(Index bi) const { Index startRow = m_spblockmat.blockRowsIndex(bi); Index rowSize = m_spblockmat.blockRowsIndex(bi+1) - startRow; return m_vec.middleRows(startRow, rowSize); } inline Scalar coeff(Index bi, Index j) const { Index startRow = m_spblockmat.blockRowsIndex(bi); Index rowSize = m_spblockmat.blockRowsIndex(bi+1) - startRow; return m_vec.block(startRow, j, rowSize, 1); } protected: const BlockSparseMatrixT& m_spblockmat; const VectorType& m_vec; }; template class BlockVectorReturn; // Proxy to view a regular vector as a block vector template class BlockVectorReturn { public: enum { ColsAtCompileTime = VectorType::ColsAtCompileTime, RowsAtCompileTime = VectorType::RowsAtCompileTime, Flags = VectorType::Flags }; typedef Ref > Scalar; typedef typename BlockSparseMatrixT::Index Index; public: BlockVectorReturn(const BlockSparseMatrixT& spblockmat, VectorType& vec) : m_spblockmat(spblockmat),m_vec(vec) { } inline Index size() const { return m_spblockmat.blockRows(); } inline Scalar coeffRef(Index bi) { Index startRow = m_spblockmat.blockRowsIndex(bi); Index rowSize = m_spblockmat.blockRowsIndex(bi+1) - startRow; return m_vec.middleRows(startRow, rowSize); } inline Scalar coeffRef(Index bi, Index j) { Index startRow = m_spblockmat.blockRowsIndex(bi); Index rowSize = m_spblockmat.blockRowsIndex(bi+1) - startRow; return m_vec.block(startRow, j, rowSize, 1); } protected: const BlockSparseMatrixT& m_spblockmat; VectorType& m_vec; }; // Block version of the sparse dense product template class BlockSparseTimeDenseProduct; namespace internal { template struct traits > { typedef Dense StorageKind; typedef MatrixXpr XprKind; typedef typename BlockSparseMatrixT::Scalar Scalar; typedef typename BlockSparseMatrixT::Index Index; enum { RowsAtCompileTime = Dynamic, ColsAtCompileTime = Dynamic, MaxRowsAtCompileTime = Dynamic, MaxColsAtCompileTime = Dynamic, Flags = 0, CoeffReadCost = internal::traits::CoeffReadCost }; }; } // end namespace internal template class BlockSparseTimeDenseProduct : public ProductBase, Lhs, Rhs> { public: EIGEN_PRODUCT_PUBLIC_INTERFACE(BlockSparseTimeDenseProduct) BlockSparseTimeDenseProduct(const Lhs& lhs, const Rhs& rhs) : Base(lhs,rhs) {} template void scaleAndAddTo(Dest& dest, const typename Rhs::Scalar& alpha) const { BlockVectorReturn tmpDest(m_lhs, dest); internal::sparse_time_dense_product( BlockSparseMatrixView(m_lhs), BlockVectorView(m_lhs, m_rhs), tmpDest, alpha); } private: BlockSparseTimeDenseProduct& operator=(const BlockSparseTimeDenseProduct&); }; template class BlockSparseMatrix : public SparseMatrixBase > { public: typedef _Scalar Scalar; typedef typename NumTraits::Real RealScalar; typedef _StorageIndex StorageIndex; typedef typename internal::ref_selector >::type Nested; enum { Options = _Options, Flags = Options, BlockSize=_BlockAtCompileTime, RowsAtCompileTime = Dynamic, ColsAtCompileTime = Dynamic, MaxRowsAtCompileTime = Dynamic, MaxColsAtCompileTime = Dynamic, IsVectorAtCompileTime = 0, IsColMajor = Flags&RowMajorBit ? 0 : 1 }; typedef Matrix BlockScalar; typedef Matrix BlockRealScalar; typedef typename internal::conditional<_BlockAtCompileTime==Dynamic, Scalar, BlockScalar>::type BlockScalarReturnType; typedef BlockSparseMatrix PlainObject; public: // Default constructor BlockSparseMatrix() : m_innerBSize(0),m_outerBSize(0),m_innerOffset(0),m_outerOffset(0), m_nonzerosblocks(0),m_values(0),m_blockPtr(0),m_indices(0), m_outerIndex(0),m_blockSize(BlockSize) { } /** * \brief Construct and resize * */ BlockSparseMatrix(Index brow, Index bcol) : m_innerBSize(IsColMajor ? brow : bcol), m_outerBSize(IsColMajor ? bcol : brow), m_innerOffset(0),m_outerOffset(0),m_nonzerosblocks(0), m_values(0),m_blockPtr(0),m_indices(0), m_outerIndex(0),m_blockSize(BlockSize) { } /** * \brief Copy-constructor */ BlockSparseMatrix(const BlockSparseMatrix& other) : m_innerBSize(other.m_innerBSize),m_outerBSize(other.m_outerBSize), m_nonzerosblocks(other.m_nonzerosblocks),m_nonzeros(other.m_nonzeros), m_blockPtr(0),m_blockSize(other.m_blockSize) { // should we allow copying between variable-size blocks and fixed-size blocks ?? eigen_assert(m_blockSize == BlockSize && " CAN NOT COPY BETWEEN FIXED-SIZE AND VARIABLE-SIZE BLOCKS"); std::copy(other.m_innerOffset, other.m_innerOffset+m_innerBSize+1, m_innerOffset); std::copy(other.m_outerOffset, other.m_outerOffset+m_outerBSize+1, m_outerOffset); std::copy(other.m_values, other.m_values+m_nonzeros, m_values); if(m_blockSize != Dynamic) std::copy(other.m_blockPtr, other.m_blockPtr+m_nonzerosblocks, m_blockPtr); std::copy(other.m_indices, other.m_indices+m_nonzerosblocks, m_indices); std::copy(other.m_outerIndex, other.m_outerIndex+m_outerBSize, m_outerIndex); } friend void swap(BlockSparseMatrix& first, BlockSparseMatrix& second) { std::swap(first.m_innerBSize, second.m_innerBSize); std::swap(first.m_outerBSize, second.m_outerBSize); std::swap(first.m_innerOffset, second.m_innerOffset); std::swap(first.m_outerOffset, second.m_outerOffset); std::swap(first.m_nonzerosblocks, second.m_nonzerosblocks); std::swap(first.m_nonzeros, second.m_nonzeros); std::swap(first.m_values, second.m_values); std::swap(first.m_blockPtr, second.m_blockPtr); std::swap(first.m_indices, second.m_indices); std::swap(first.m_outerIndex, second.m_outerIndex); std::swap(first.m_BlockSize, second.m_blockSize); } BlockSparseMatrix& operator=(BlockSparseMatrix other) { //Copy-and-swap paradigm ... avoid leaked data if thrown swap(*this, other); return *this; } // Destructor ~BlockSparseMatrix() { delete[] m_outerIndex; delete[] m_innerOffset; delete[] m_outerOffset; delete[] m_indices; delete[] m_blockPtr; delete[] m_values; } /** * \brief Constructor from a sparse matrix * */ template inline BlockSparseMatrix(const MatrixType& spmat) : m_blockSize(BlockSize) { EIGEN_STATIC_ASSERT((m_blockSize != Dynamic), THIS_METHOD_IS_ONLY_FOR_FIXED_SIZE); *this = spmat; } /** * \brief Assignment from a sparse matrix with the same storage order * * Convert from a sparse matrix to block sparse matrix. * \warning Before calling this function, tt is necessary to call * either setBlockLayout() (matrices with variable-size blocks) * or setBlockSize() (for fixed-size blocks). */ template inline BlockSparseMatrix& operator=(const MatrixType& spmat) { eigen_assert((m_innerBSize != 0 && m_outerBSize != 0) && "Trying to assign to a zero-size matrix, call resize() first"); eigen_assert(((MatrixType::Options&RowMajorBit) != IsColMajor) && "Wrong storage order"); typedef SparseMatrix MatrixPatternType; MatrixPatternType blockPattern(blockRows(), blockCols()); m_nonzeros = 0; // First, compute the number of nonzero blocks and their locations for(StorageIndex bj = 0; bj < m_outerBSize; ++bj) { // Browse each outer block and compute the structure std::vector nzblocksFlag(m_innerBSize,false); // Record the existing blocks blockPattern.startVec(bj); for(StorageIndex j = blockOuterIndex(bj); j < blockOuterIndex(bj+1); ++j) { typename MatrixType::InnerIterator it_spmat(spmat, j); for(; it_spmat; ++it_spmat) { StorageIndex bi = innerToBlock(it_spmat.index()); // Index of the current nonzero block if(!nzblocksFlag[bi]) { // Save the index of this nonzero block nzblocksFlag[bi] = true; blockPattern.insertBackByOuterInnerUnordered(bj, bi) = true; // Compute the total number of nonzeros (including explicit zeros in blocks) m_nonzeros += blockOuterSize(bj) * blockInnerSize(bi); } } } // end current outer block } blockPattern.finalize(); // Allocate the internal arrays setBlockStructure(blockPattern); for(StorageIndex nz = 0; nz < m_nonzeros; ++nz) m_values[nz] = Scalar(0); for(StorageIndex bj = 0; bj < m_outerBSize; ++bj) { // Now copy the values for(StorageIndex j = blockOuterIndex(bj); j < blockOuterIndex(bj+1); ++j) { // Browse the outer block column by column (for column-major matrices) typename MatrixType::InnerIterator it_spmat(spmat, j); for(; it_spmat; ++it_spmat) { StorageIndex idx = 0; // Position of this block in the column block StorageIndex bi = innerToBlock(it_spmat.index()); // Index of the current nonzero block // Go to the inner block where this element belongs to while(bi > m_indices[m_outerIndex[bj]+idx]) ++idx; // Not expensive for ordered blocks StorageIndex idxVal;// Get the right position in the array of values for this element if(m_blockSize == Dynamic) { // Offset from all blocks before ... idxVal = m_blockPtr[m_outerIndex[bj]+idx]; // ... and offset inside the block idxVal += (j - blockOuterIndex(bj)) * blockOuterSize(bj) + it_spmat.index() - m_innerOffset[bi]; } else { // All blocks before idxVal = (m_outerIndex[bj] + idx) * m_blockSize * m_blockSize; // inside the block idxVal += (j - blockOuterIndex(bj)) * m_blockSize + (it_spmat.index()%m_blockSize); } // Insert the value m_values[idxVal] = it_spmat.value(); } // end of this column } // end of this block } // end of this outer block return *this; } /** * \brief Set the nonzero block pattern of the matrix * * Given a sparse matrix describing the nonzero block pattern, * this function prepares the internal pointers for values. * After calling this function, any *nonzero* block (bi, bj) can be set * with a simple call to coeffRef(bi,bj). * * * \warning Before calling this function, tt is necessary to call * either setBlockLayout() (matrices with variable-size blocks) * or setBlockSize() (for fixed-size blocks). * * \param blockPattern Sparse matrix of boolean elements describing the block structure * * \sa setBlockLayout() \sa setBlockSize() */ template void setBlockStructure(const MatrixType& blockPattern) { resize(blockPattern.rows(), blockPattern.cols()); reserve(blockPattern.nonZeros()); // Browse the block pattern and set up the various pointers m_outerIndex[0] = 0; if(m_blockSize == Dynamic) m_blockPtr[0] = 0; for(StorageIndex nz = 0; nz < m_nonzeros; ++nz) m_values[nz] = Scalar(0); for(StorageIndex bj = 0; bj < m_outerBSize; ++bj) { //Browse each outer block //First, copy and save the indices of nonzero blocks //FIXME : find a way to avoid this ... std::vector nzBlockIdx; typename MatrixType::InnerIterator it(blockPattern, bj); for(; it; ++it) { nzBlockIdx.push_back(it.index()); } std::sort(nzBlockIdx.begin(), nzBlockIdx.end()); // Now, fill block indices and (eventually) pointers to blocks for(StorageIndex idx = 0; idx < nzBlockIdx.size(); ++idx) { StorageIndex offset = m_outerIndex[bj]+idx; // offset in m_indices m_indices[offset] = nzBlockIdx[idx]; if(m_blockSize == Dynamic) m_blockPtr[offset] = m_blockPtr[offset-1] + blockInnerSize(nzBlockIdx[idx]) * blockOuterSize(bj); // There is no blockPtr for fixed-size blocks... not needed !??? } // Save the pointer to the next outer block m_outerIndex[bj+1] = m_outerIndex[bj] + nzBlockIdx.size(); } } /** * \brief Set the number of rows and columns blocks */ inline void resize(Index brow, Index bcol) { m_innerBSize = IsColMajor ? brow : bcol; m_outerBSize = IsColMajor ? bcol : brow; } /** * \brief set the block size at runtime for fixed-size block layout * * Call this only for fixed-size blocks */ inline void setBlockSize(Index blockSize) { m_blockSize = blockSize; } /** * \brief Set the row and column block layouts, * * This function set the size of each row and column block. * So this function should be used only for blocks with variable size. * \param rowBlocks : Number of rows per row block * \param colBlocks : Number of columns per column block * \sa resize(), setBlockSize() */ inline void setBlockLayout(const VectorXi& rowBlocks, const VectorXi& colBlocks) { const VectorXi& innerBlocks = IsColMajor ? rowBlocks : colBlocks; const VectorXi& outerBlocks = IsColMajor ? colBlocks : rowBlocks; eigen_assert(m_innerBSize == innerBlocks.size() && "CHECK THE NUMBER OF ROW OR COLUMN BLOCKS"); eigen_assert(m_outerBSize == outerBlocks.size() && "CHECK THE NUMBER OF ROW OR COLUMN BLOCKS"); m_outerBSize = outerBlocks.size(); // starting index of blocks... cumulative sums m_innerOffset = new StorageIndex[m_innerBSize+1]; m_outerOffset = new StorageIndex[m_outerBSize+1]; m_innerOffset[0] = 0; m_outerOffset[0] = 0; std::partial_sum(&innerBlocks[0], &innerBlocks[m_innerBSize-1]+1, &m_innerOffset[1]); std::partial_sum(&outerBlocks[0], &outerBlocks[m_outerBSize-1]+1, &m_outerOffset[1]); // Compute the total number of nonzeros m_nonzeros = 0; for(StorageIndex bj = 0; bj < m_outerBSize; ++bj) for(StorageIndex bi = 0; bi < m_innerBSize; ++bi) m_nonzeros += outerBlocks[bj] * innerBlocks[bi]; } /** * \brief Allocate the internal array of pointers to blocks and their inner indices * * \note For fixed-size blocks, call setBlockSize() to set the block. * And For variable-size blocks, call setBlockLayout() before using this function * * \param nonzerosblocks Number of nonzero blocks. The total number of nonzeros is * is computed in setBlockLayout() for variable-size blocks * \sa setBlockSize() */ inline void reserve(const Index nonzerosblocks) { eigen_assert((m_innerBSize != 0 && m_outerBSize != 0) && "TRYING TO RESERVE ZERO-SIZE MATRICES, CALL resize() first"); //FIXME Should free if already allocated m_outerIndex = new StorageIndex[m_outerBSize+1]; m_nonzerosblocks = nonzerosblocks; if(m_blockSize != Dynamic) { m_nonzeros = nonzerosblocks * (m_blockSize * m_blockSize); m_blockPtr = 0; } else { // m_nonzeros is already computed in setBlockLayout() m_blockPtr = new StorageIndex[m_nonzerosblocks+1]; } m_indices = new StorageIndex[m_nonzerosblocks+1]; m_values = new Scalar[m_nonzeros]; } /** * \brief Fill values in a matrix from a triplet list. * * Each triplet item has a block stored in an Eigen dense matrix. * The InputIterator class should provide the functions row(), col() and value() * * \note For fixed-size blocks, call setBlockSize() before this function. * * FIXME Do not accept duplicates */ template void setFromTriplets(const InputIterator& begin, const InputIterator& end) { eigen_assert((m_innerBSize!=0 && m_outerBSize !=0) && "ZERO BLOCKS, PLEASE CALL resize() before"); /* First, sort the triplet list * FIXME This can be unnecessarily expensive since only the inner indices have to be sorted * The best approach is like in SparseMatrix::setFromTriplets() */ internal::TripletComp tripletcomp; std::sort(begin, end, tripletcomp); /* Count the number of rows and column blocks, * and the number of nonzero blocks per outer dimension */ VectorXi rowBlocks(m_innerBSize); // Size of each block row VectorXi colBlocks(m_outerBSize); // Size of each block column rowBlocks.setZero(); colBlocks.setZero(); VectorXi nzblock_outer(m_outerBSize); // Number of nz blocks per outer vector VectorXi nz_outer(m_outerBSize); // Number of nz per outer vector...for variable-size blocks nzblock_outer.setZero(); nz_outer.setZero(); for(InputIterator it(begin); it !=end; ++it) { eigen_assert(it->row() >= 0 && it->row() < this->blockRows() && it->col() >= 0 && it->col() < this->blockCols()); eigen_assert((it->value().rows() == it->value().cols() && (it->value().rows() == m_blockSize)) || (m_blockSize == Dynamic)); if(m_blockSize == Dynamic) { eigen_assert((rowBlocks[it->row()] == 0 || rowBlocks[it->row()] == it->value().rows()) && "NON CORRESPONDING SIZES FOR ROW BLOCKS"); eigen_assert((colBlocks[it->col()] == 0 || colBlocks[it->col()] == it->value().cols()) && "NON CORRESPONDING SIZES FOR COLUMN BLOCKS"); rowBlocks[it->row()] =it->value().rows(); colBlocks[it->col()] = it->value().cols(); } nz_outer(IsColMajor ? it->col() : it->row()) += it->value().rows() * it->value().cols(); nzblock_outer(IsColMajor ? it->col() : it->row())++; } // Allocate member arrays if(m_blockSize == Dynamic) setBlockLayout(rowBlocks, colBlocks); StorageIndex nzblocks = nzblock_outer.sum(); reserve(nzblocks); // Temporary markers VectorXi block_id(m_outerBSize); // To be used as a block marker during insertion // Setup outer index pointers and markers m_outerIndex[0] = 0; if (m_blockSize == Dynamic) m_blockPtr[0] = 0; for(StorageIndex bj = 0; bj < m_outerBSize; ++bj) { m_outerIndex[bj+1] = m_outerIndex[bj] + nzblock_outer(bj); block_id(bj) = m_outerIndex[bj]; if(m_blockSize==Dynamic) { m_blockPtr[m_outerIndex[bj+1]] = m_blockPtr[m_outerIndex[bj]] + nz_outer(bj); } } // Fill the matrix for(InputIterator it(begin); it!=end; ++it) { StorageIndex outer = IsColMajor ? it->col() : it->row(); StorageIndex inner = IsColMajor ? it->row() : it->col(); m_indices[block_id(outer)] = inner; StorageIndex block_size = it->value().rows()*it->value().cols(); StorageIndex nz_marker = blockPtr(block_id[outer]); memcpy(&(m_values[nz_marker]), it->value().data(), block_size * sizeof(Scalar)); if(m_blockSize == Dynamic) { m_blockPtr[block_id(outer)+1] = m_blockPtr[block_id(outer)] + block_size; } block_id(outer)++; } // An alternative when the outer indices are sorted...no need to use an array of markers // for(Index bcol = 0; bcol < m_outerBSize; ++bcol) // { // Index id = 0, id_nz = 0, id_nzblock = 0; // for(InputIterator it(begin); it!=end; ++it) // { // while (idvalue().rows()*it->value().cols(); // m_blockPtr[id_nzblock+1] = m_blockPtr[id_nzblock] + block_size; // id_nzblock++; // memcpy(&(m_values[id_nz]),it->value().data(), block_size*sizeof(Scalar)); // id_nz += block_size; // } // while(id < m_outerBSize-1) // Empty columns at the end // { // id++; // m_outerIndex[id+1]=m_outerIndex[id]; // } // } } /** * \returns the number of rows */ inline Index rows() const { // return blockRows(); return (IsColMajor ? innerSize() : outerSize()); } /** * \returns the number of cols */ inline Index cols() const { // return blockCols(); return (IsColMajor ? outerSize() : innerSize()); } inline Index innerSize() const { if(m_blockSize == Dynamic) return m_innerOffset[m_innerBSize]; else return (m_innerBSize * m_blockSize) ; } inline Index outerSize() const { if(m_blockSize == Dynamic) return m_outerOffset[m_outerBSize]; else return (m_outerBSize * m_blockSize) ; } /** \returns the number of rows grouped by blocks */ inline Index blockRows() const { return (IsColMajor ? m_innerBSize : m_outerBSize); } /** \returns the number of columns grouped by blocks */ inline Index blockCols() const { return (IsColMajor ? m_outerBSize : m_innerBSize); } inline Index outerBlocks() const { return m_outerBSize; } inline Index innerBlocks() const { return m_innerBSize; } /** \returns the block index where outer belongs to */ inline Index outerToBlock(Index outer) const { eigen_assert(outer < outerSize() && "OUTER INDEX OUT OF BOUNDS"); if(m_blockSize != Dynamic) return (outer / m_blockSize); // Integer division StorageIndex b_outer = 0; while(m_outerOffset[b_outer] <= outer) ++b_outer; return b_outer - 1; } /** \returns the block index where inner belongs to */ inline Index innerToBlock(Index inner) const { eigen_assert(inner < innerSize() && "OUTER INDEX OUT OF BOUNDS"); if(m_blockSize != Dynamic) return (inner / m_blockSize); // Integer division StorageIndex b_inner = 0; while(m_innerOffset[b_inner] <= inner) ++b_inner; return b_inner - 1; } /** *\returns a reference to the (i,j) block as an Eigen Dense Matrix */ Ref coeffRef(Index brow, Index bcol) { eigen_assert(brow < blockRows() && "BLOCK ROW INDEX OUT OF BOUNDS"); eigen_assert(bcol < blockCols() && "BLOCK nzblocksFlagCOLUMN OUT OF BOUNDS"); StorageIndex rsize = IsColMajor ? blockInnerSize(brow): blockOuterSize(bcol); StorageIndex csize = IsColMajor ? blockOuterSize(bcol) : blockInnerSize(brow); StorageIndex inner = IsColMajor ? brow : bcol; StorageIndex outer = IsColMajor ? bcol : brow; StorageIndex offset = m_outerIndex[outer]; while(offset < m_outerIndex[outer+1] && m_indices[offset] != inner) offset++; if(m_indices[offset] == inner) { return Map(&(m_values[blockPtr(offset)]), rsize, csize); } else { //FIXME the block does not exist, Insert it !!!!!!!!! eigen_assert("DYNAMIC INSERTION IS NOT YET SUPPORTED"); } } /** * \returns the value of the (i,j) block as an Eigen Dense Matrix */ Map coeff(Index brow, Index bcol) const { eigen_assert(brow < blockRows() && "BLOCK ROW INDEX OUT OF BOUNDS"); eigen_assert(bcol < blockCols() && "BLOCK COLUMN OUT OF BOUNDS"); StorageIndex rsize = IsColMajor ? blockInnerSize(brow): blockOuterSize(bcol); StorageIndex csize = IsColMajor ? blockOuterSize(bcol) : blockInnerSize(brow); StorageIndex inner = IsColMajor ? brow : bcol; StorageIndex outer = IsColMajor ? bcol : brow; StorageIndex offset = m_outerIndex[outer]; while(offset < m_outerIndex[outer+1] && m_indices[offset] != inner) offset++; if(m_indices[offset] == inner) { return Map (&(m_values[blockPtr(offset)]), rsize, csize); } else // return BlockScalar::Zero(rsize, csize); eigen_assert("NOT YET SUPPORTED"); } // Block Matrix times vector product template BlockSparseTimeDenseProduct operator*(const VecType& lhs) const { return BlockSparseTimeDenseProduct(*this, lhs); } /** \returns the number of nonzero blocks */ inline Index nonZerosBlocks() const { return m_nonzerosblocks; } /** \returns the total number of nonzero elements, including eventual explicit zeros in blocks */ inline Index nonZeros() const { return m_nonzeros; } inline BlockScalarReturnType *valuePtr() {return static_cast(m_values);} // inline Scalar *valuePtr(){ return m_values; } inline StorageIndex *innerIndexPtr() {return m_indices; } inline const StorageIndex *innerIndexPtr() const {return m_indices; } inline StorageIndex *outerIndexPtr() {return m_outerIndex; } inline const StorageIndex* outerIndexPtr() const {return m_outerIndex; } /** \brief for compatibility purposes with the SparseMatrix class */ inline bool isCompressed() const {return true;} /** * \returns the starting index of the bi row block */ inline Index blockRowsIndex(Index bi) const { return IsColMajor ? blockInnerIndex(bi) : blockOuterIndex(bi); } /** * \returns the starting index of the bj col block */ inline Index blockColsIndex(Index bj) const { return IsColMajor ? blockOuterIndex(bj) : blockInnerIndex(bj); } inline Index blockOuterIndex(Index bj) const { return (m_blockSize == Dynamic) ? m_outerOffset[bj] : (bj * m_blockSize); } inline Index blockInnerIndex(Index bi) const { return (m_blockSize == Dynamic) ? m_innerOffset[bi] : (bi * m_blockSize); } // Not needed ??? inline Index blockInnerSize(Index bi) const { return (m_blockSize == Dynamic) ? (m_innerOffset[bi+1] - m_innerOffset[bi]) : m_blockSize; } inline Index blockOuterSize(Index bj) const { return (m_blockSize == Dynamic) ? (m_outerOffset[bj+1]- m_outerOffset[bj]) : m_blockSize; } /** * \brief Browse the matrix by outer index */ class InnerIterator; // Browse column by column /** * \brief Browse the matrix by block outer index */ class BlockInnerIterator; // Browse block by block friend std::ostream & operator << (std::ostream & s, const BlockSparseMatrix& m) { for (StorageIndex j = 0; j < m.outerBlocks(); ++j) { BlockInnerIterator itb(m, j); for(; itb; ++itb) { s << "("< in the array of values */ Index blockPtr(Index id) const { if(m_blockSize == Dynamic) return m_blockPtr[id]; else return id * m_blockSize * m_blockSize; //return blockDynIdx(id, typename internal::conditional<(BlockSize==Dynamic), internal::true_type, internal::false_type>::type()); } protected: // inline Index blockDynIdx(Index id, internal::true_type) const // { // return m_blockPtr[id]; // } // inline Index blockDynIdx(Index id, internal::false_type) const // { // return id * BlockSize * BlockSize; // } // To be implemented // Insert a block at a particular location... need to make a room for that Map insert(Index brow, Index bcol); Index m_innerBSize; // Number of block rows Index m_outerBSize; // Number of block columns StorageIndex *m_innerOffset; // Starting index of each inner block (size m_innerBSize+1) StorageIndex *m_outerOffset; // Starting index of each outer block (size m_outerBSize+1) Index m_nonzerosblocks; // Total nonzeros blocks (lower than m_innerBSize x m_outerBSize) Index m_nonzeros; // Total nonzeros elements Scalar *m_values; //Values stored block column after block column (size m_nonzeros) StorageIndex *m_blockPtr; // Pointer to the beginning of each block in m_values, size m_nonzeroblocks ... null for fixed-size blocks StorageIndex *m_indices; //Inner block indices, size m_nonzerosblocks ... OK StorageIndex *m_outerIndex; // Starting pointer of each block column in m_indices (size m_outerBSize)... OK Index m_blockSize; // Size of a block for fixed-size blocks, otherwise -1 }; template class BlockSparseMatrix<_Scalar, _BlockAtCompileTime, _Options, _StorageIndex>::BlockInnerIterator { public: enum{ Flags = _Options }; BlockInnerIterator(const BlockSparseMatrix& mat, const Index outer) : m_mat(mat),m_outer(outer), m_id(mat.m_outerIndex[outer]), m_end(mat.m_outerIndex[outer+1]) { } inline BlockInnerIterator& operator++() {m_id++; return *this; } inline const Map value() const { return Map(&(m_mat.m_values[m_mat.blockPtr(m_id)]), rows(),cols()); } inline Map valueRef() { return Map(&(m_mat.m_values[m_mat.blockPtr(m_id)]), rows(),cols()); } // Block inner index inline Index index() const {return m_mat.m_indices[m_id]; } inline Index outer() const { return m_outer; } // block row index inline Index row() const {return index(); } // block column index inline Index col() const {return outer(); } // FIXME Number of rows in the current block inline Index rows() const { return (m_mat.m_blockSize==Dynamic) ? (m_mat.m_innerOffset[index()+1] - m_mat.m_innerOffset[index()]) : m_mat.m_blockSize; } // Number of columns in the current block ... inline Index cols() const { return (m_mat.m_blockSize==Dynamic) ? (m_mat.m_outerOffset[m_outer+1]-m_mat.m_outerOffset[m_outer]) : m_mat.m_blockSize;} inline operator bool() const { return (m_id < m_end); } protected: const BlockSparseMatrix<_Scalar, _BlockAtCompileTime, _Options, StorageIndex>& m_mat; const Index m_outer; Index m_id; Index m_end; }; template class BlockSparseMatrix<_Scalar, _BlockAtCompileTime, _Options, _StorageIndex>::InnerIterator { public: InnerIterator(const BlockSparseMatrix& mat, Index outer) : m_mat(mat),m_outerB(mat.outerToBlock(outer)),m_outer(outer), itb(mat, mat.outerToBlock(outer)), m_offset(outer - mat.blockOuterIndex(m_outerB)) { if (itb) { m_id = m_mat.blockInnerIndex(itb.index()); m_start = m_id; m_end = m_mat.blockInnerIndex(itb.index()+1); } } inline InnerIterator& operator++() { m_id++; if (m_id >= m_end) { ++itb; if (itb) { m_id = m_mat.blockInnerIndex(itb.index()); m_start = m_id; m_end = m_mat.blockInnerIndex(itb.index()+1); } } return *this; } inline const Scalar& value() const { return itb.value().coeff(m_id - m_start, m_offset); } inline Scalar& valueRef() { return itb.valueRef().coeff(m_id - m_start, m_offset); } inline Index index() const { return m_id; } inline Index outer() const {return m_outer; } inline Index col() const {return outer(); } inline Index row() const { return index();} inline operator bool() const { return itb; } protected: const BlockSparseMatrix& m_mat; const Index m_outer; const Index m_outerB; BlockInnerIterator itb; // Iterator through the blocks const Index m_offset; // Position of this column in the block Index m_start; // starting inner index of this block Index m_id; // current inner index in the block Index m_end; // starting inner index of the next block }; } // end namespace Eigen #endif // EIGEN_SPARSEBLOCKMATRIX_H