tomoto 0.1.0
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +7 -0
- data/CHANGELOG.md +3 -0
- data/LICENSE.txt +22 -0
- data/README.md +123 -0
- data/ext/tomoto/ext.cpp +245 -0
- data/ext/tomoto/extconf.rb +28 -0
- data/lib/tomoto.rb +12 -0
- data/lib/tomoto/ct.rb +11 -0
- data/lib/tomoto/hdp.rb +11 -0
- data/lib/tomoto/lda.rb +67 -0
- data/lib/tomoto/version.rb +3 -0
- data/vendor/EigenRand/EigenRand/Core.h +1139 -0
- data/vendor/EigenRand/EigenRand/Dists/Basic.h +111 -0
- data/vendor/EigenRand/EigenRand/Dists/Discrete.h +877 -0
- data/vendor/EigenRand/EigenRand/Dists/GammaPoisson.h +108 -0
- data/vendor/EigenRand/EigenRand/Dists/NormalExp.h +626 -0
- data/vendor/EigenRand/EigenRand/EigenRand +19 -0
- data/vendor/EigenRand/EigenRand/Macro.h +24 -0
- data/vendor/EigenRand/EigenRand/MorePacketMath.h +978 -0
- data/vendor/EigenRand/EigenRand/PacketFilter.h +286 -0
- data/vendor/EigenRand/EigenRand/PacketRandomEngine.h +624 -0
- data/vendor/EigenRand/EigenRand/RandUtils.h +413 -0
- data/vendor/EigenRand/EigenRand/doc.h +220 -0
- data/vendor/EigenRand/LICENSE +21 -0
- data/vendor/EigenRand/README.md +288 -0
- data/vendor/eigen/COPYING.BSD +26 -0
- data/vendor/eigen/COPYING.GPL +674 -0
- data/vendor/eigen/COPYING.LGPL +502 -0
- data/vendor/eigen/COPYING.MINPACK +52 -0
- data/vendor/eigen/COPYING.MPL2 +373 -0
- data/vendor/eigen/COPYING.README +18 -0
- data/vendor/eigen/Eigen/CMakeLists.txt +19 -0
- data/vendor/eigen/Eigen/Cholesky +46 -0
- data/vendor/eigen/Eigen/CholmodSupport +48 -0
- data/vendor/eigen/Eigen/Core +537 -0
- data/vendor/eigen/Eigen/Dense +7 -0
- data/vendor/eigen/Eigen/Eigen +2 -0
- data/vendor/eigen/Eigen/Eigenvalues +61 -0
- data/vendor/eigen/Eigen/Geometry +62 -0
- data/vendor/eigen/Eigen/Householder +30 -0
- data/vendor/eigen/Eigen/IterativeLinearSolvers +48 -0
- data/vendor/eigen/Eigen/Jacobi +33 -0
- data/vendor/eigen/Eigen/LU +50 -0
- data/vendor/eigen/Eigen/MetisSupport +35 -0
- data/vendor/eigen/Eigen/OrderingMethods +73 -0
- data/vendor/eigen/Eigen/PaStiXSupport +48 -0
- data/vendor/eigen/Eigen/PardisoSupport +35 -0
- data/vendor/eigen/Eigen/QR +51 -0
- data/vendor/eigen/Eigen/QtAlignedMalloc +40 -0
- data/vendor/eigen/Eigen/SPQRSupport +34 -0
- data/vendor/eigen/Eigen/SVD +51 -0
- data/vendor/eigen/Eigen/Sparse +36 -0
- data/vendor/eigen/Eigen/SparseCholesky +45 -0
- data/vendor/eigen/Eigen/SparseCore +69 -0
- data/vendor/eigen/Eigen/SparseLU +46 -0
- data/vendor/eigen/Eigen/SparseQR +37 -0
- data/vendor/eigen/Eigen/StdDeque +27 -0
- data/vendor/eigen/Eigen/StdList +26 -0
- data/vendor/eigen/Eigen/StdVector +27 -0
- data/vendor/eigen/Eigen/SuperLUSupport +64 -0
- data/vendor/eigen/Eigen/UmfPackSupport +40 -0
- data/vendor/eigen/Eigen/src/Cholesky/LDLT.h +673 -0
- data/vendor/eigen/Eigen/src/Cholesky/LLT.h +542 -0
- data/vendor/eigen/Eigen/src/Cholesky/LLT_LAPACKE.h +99 -0
- data/vendor/eigen/Eigen/src/CholmodSupport/CholmodSupport.h +639 -0
- data/vendor/eigen/Eigen/src/Core/Array.h +329 -0
- data/vendor/eigen/Eigen/src/Core/ArrayBase.h +226 -0
- data/vendor/eigen/Eigen/src/Core/ArrayWrapper.h +209 -0
- data/vendor/eigen/Eigen/src/Core/Assign.h +90 -0
- data/vendor/eigen/Eigen/src/Core/AssignEvaluator.h +935 -0
- data/vendor/eigen/Eigen/src/Core/Assign_MKL.h +178 -0
- data/vendor/eigen/Eigen/src/Core/BandMatrix.h +353 -0
- data/vendor/eigen/Eigen/src/Core/Block.h +452 -0
- data/vendor/eigen/Eigen/src/Core/BooleanRedux.h +164 -0
- data/vendor/eigen/Eigen/src/Core/CommaInitializer.h +160 -0
- data/vendor/eigen/Eigen/src/Core/ConditionEstimator.h +175 -0
- data/vendor/eigen/Eigen/src/Core/CoreEvaluators.h +1688 -0
- data/vendor/eigen/Eigen/src/Core/CoreIterators.h +127 -0
- data/vendor/eigen/Eigen/src/Core/CwiseBinaryOp.h +184 -0
- data/vendor/eigen/Eigen/src/Core/CwiseNullaryOp.h +866 -0
- data/vendor/eigen/Eigen/src/Core/CwiseTernaryOp.h +197 -0
- data/vendor/eigen/Eigen/src/Core/CwiseUnaryOp.h +103 -0
- data/vendor/eigen/Eigen/src/Core/CwiseUnaryView.h +128 -0
- data/vendor/eigen/Eigen/src/Core/DenseBase.h +611 -0
- data/vendor/eigen/Eigen/src/Core/DenseCoeffsBase.h +681 -0
- data/vendor/eigen/Eigen/src/Core/DenseStorage.h +570 -0
- data/vendor/eigen/Eigen/src/Core/Diagonal.h +260 -0
- data/vendor/eigen/Eigen/src/Core/DiagonalMatrix.h +343 -0
- data/vendor/eigen/Eigen/src/Core/DiagonalProduct.h +28 -0
- data/vendor/eigen/Eigen/src/Core/Dot.h +318 -0
- data/vendor/eigen/Eigen/src/Core/EigenBase.h +159 -0
- data/vendor/eigen/Eigen/src/Core/ForceAlignedAccess.h +146 -0
- data/vendor/eigen/Eigen/src/Core/Fuzzy.h +155 -0
- data/vendor/eigen/Eigen/src/Core/GeneralProduct.h +455 -0
- data/vendor/eigen/Eigen/src/Core/GenericPacketMath.h +593 -0
- data/vendor/eigen/Eigen/src/Core/GlobalFunctions.h +187 -0
- data/vendor/eigen/Eigen/src/Core/IO.h +225 -0
- data/vendor/eigen/Eigen/src/Core/Inverse.h +118 -0
- data/vendor/eigen/Eigen/src/Core/Map.h +171 -0
- data/vendor/eigen/Eigen/src/Core/MapBase.h +303 -0
- data/vendor/eigen/Eigen/src/Core/MathFunctions.h +1415 -0
- data/vendor/eigen/Eigen/src/Core/MathFunctionsImpl.h +101 -0
- data/vendor/eigen/Eigen/src/Core/Matrix.h +459 -0
- data/vendor/eigen/Eigen/src/Core/MatrixBase.h +529 -0
- data/vendor/eigen/Eigen/src/Core/NestByValue.h +110 -0
- data/vendor/eigen/Eigen/src/Core/NoAlias.h +108 -0
- data/vendor/eigen/Eigen/src/Core/NumTraits.h +248 -0
- data/vendor/eigen/Eigen/src/Core/PermutationMatrix.h +633 -0
- data/vendor/eigen/Eigen/src/Core/PlainObjectBase.h +1035 -0
- data/vendor/eigen/Eigen/src/Core/Product.h +186 -0
- data/vendor/eigen/Eigen/src/Core/ProductEvaluators.h +1112 -0
- data/vendor/eigen/Eigen/src/Core/Random.h +182 -0
- data/vendor/eigen/Eigen/src/Core/Redux.h +505 -0
- data/vendor/eigen/Eigen/src/Core/Ref.h +283 -0
- data/vendor/eigen/Eigen/src/Core/Replicate.h +142 -0
- data/vendor/eigen/Eigen/src/Core/ReturnByValue.h +117 -0
- data/vendor/eigen/Eigen/src/Core/Reverse.h +211 -0
- data/vendor/eigen/Eigen/src/Core/Select.h +162 -0
- data/vendor/eigen/Eigen/src/Core/SelfAdjointView.h +352 -0
- data/vendor/eigen/Eigen/src/Core/SelfCwiseBinaryOp.h +47 -0
- data/vendor/eigen/Eigen/src/Core/Solve.h +188 -0
- data/vendor/eigen/Eigen/src/Core/SolveTriangular.h +235 -0
- data/vendor/eigen/Eigen/src/Core/SolverBase.h +130 -0
- data/vendor/eigen/Eigen/src/Core/StableNorm.h +221 -0
- data/vendor/eigen/Eigen/src/Core/Stride.h +111 -0
- data/vendor/eigen/Eigen/src/Core/Swap.h +67 -0
- data/vendor/eigen/Eigen/src/Core/Transpose.h +403 -0
- data/vendor/eigen/Eigen/src/Core/Transpositions.h +407 -0
- data/vendor/eigen/Eigen/src/Core/TriangularMatrix.h +983 -0
- data/vendor/eigen/Eigen/src/Core/VectorBlock.h +96 -0
- data/vendor/eigen/Eigen/src/Core/VectorwiseOp.h +695 -0
- data/vendor/eigen/Eigen/src/Core/Visitor.h +273 -0
- data/vendor/eigen/Eigen/src/Core/arch/AVX/Complex.h +451 -0
- data/vendor/eigen/Eigen/src/Core/arch/AVX/MathFunctions.h +439 -0
- data/vendor/eigen/Eigen/src/Core/arch/AVX/PacketMath.h +637 -0
- data/vendor/eigen/Eigen/src/Core/arch/AVX/TypeCasting.h +51 -0
- data/vendor/eigen/Eigen/src/Core/arch/AVX512/MathFunctions.h +391 -0
- data/vendor/eigen/Eigen/src/Core/arch/AVX512/PacketMath.h +1316 -0
- data/vendor/eigen/Eigen/src/Core/arch/AltiVec/Complex.h +430 -0
- data/vendor/eigen/Eigen/src/Core/arch/AltiVec/MathFunctions.h +322 -0
- data/vendor/eigen/Eigen/src/Core/arch/AltiVec/PacketMath.h +1061 -0
- data/vendor/eigen/Eigen/src/Core/arch/CUDA/Complex.h +103 -0
- data/vendor/eigen/Eigen/src/Core/arch/CUDA/Half.h +674 -0
- data/vendor/eigen/Eigen/src/Core/arch/CUDA/MathFunctions.h +91 -0
- data/vendor/eigen/Eigen/src/Core/arch/CUDA/PacketMath.h +333 -0
- data/vendor/eigen/Eigen/src/Core/arch/CUDA/PacketMathHalf.h +1124 -0
- data/vendor/eigen/Eigen/src/Core/arch/CUDA/TypeCasting.h +212 -0
- data/vendor/eigen/Eigen/src/Core/arch/Default/ConjHelper.h +29 -0
- data/vendor/eigen/Eigen/src/Core/arch/Default/Settings.h +49 -0
- data/vendor/eigen/Eigen/src/Core/arch/NEON/Complex.h +490 -0
- data/vendor/eigen/Eigen/src/Core/arch/NEON/MathFunctions.h +91 -0
- data/vendor/eigen/Eigen/src/Core/arch/NEON/PacketMath.h +760 -0
- data/vendor/eigen/Eigen/src/Core/arch/SSE/Complex.h +471 -0
- data/vendor/eigen/Eigen/src/Core/arch/SSE/MathFunctions.h +562 -0
- data/vendor/eigen/Eigen/src/Core/arch/SSE/PacketMath.h +895 -0
- data/vendor/eigen/Eigen/src/Core/arch/SSE/TypeCasting.h +77 -0
- data/vendor/eigen/Eigen/src/Core/arch/ZVector/Complex.h +397 -0
- data/vendor/eigen/Eigen/src/Core/arch/ZVector/MathFunctions.h +137 -0
- data/vendor/eigen/Eigen/src/Core/arch/ZVector/PacketMath.h +945 -0
- data/vendor/eigen/Eigen/src/Core/functors/AssignmentFunctors.h +168 -0
- data/vendor/eigen/Eigen/src/Core/functors/BinaryFunctors.h +475 -0
- data/vendor/eigen/Eigen/src/Core/functors/NullaryFunctors.h +188 -0
- data/vendor/eigen/Eigen/src/Core/functors/StlFunctors.h +136 -0
- data/vendor/eigen/Eigen/src/Core/functors/TernaryFunctors.h +25 -0
- data/vendor/eigen/Eigen/src/Core/functors/UnaryFunctors.h +792 -0
- data/vendor/eigen/Eigen/src/Core/products/GeneralBlockPanelKernel.h +2156 -0
- data/vendor/eigen/Eigen/src/Core/products/GeneralMatrixMatrix.h +492 -0
- data/vendor/eigen/Eigen/src/Core/products/GeneralMatrixMatrixTriangular.h +311 -0
- data/vendor/eigen/Eigen/src/Core/products/GeneralMatrixMatrixTriangular_BLAS.h +145 -0
- data/vendor/eigen/Eigen/src/Core/products/GeneralMatrixMatrix_BLAS.h +122 -0
- data/vendor/eigen/Eigen/src/Core/products/GeneralMatrixVector.h +619 -0
- data/vendor/eigen/Eigen/src/Core/products/GeneralMatrixVector_BLAS.h +136 -0
- data/vendor/eigen/Eigen/src/Core/products/Parallelizer.h +163 -0
- data/vendor/eigen/Eigen/src/Core/products/SelfadjointMatrixMatrix.h +521 -0
- data/vendor/eigen/Eigen/src/Core/products/SelfadjointMatrixMatrix_BLAS.h +287 -0
- data/vendor/eigen/Eigen/src/Core/products/SelfadjointMatrixVector.h +260 -0
- data/vendor/eigen/Eigen/src/Core/products/SelfadjointMatrixVector_BLAS.h +118 -0
- data/vendor/eigen/Eigen/src/Core/products/SelfadjointProduct.h +133 -0
- data/vendor/eigen/Eigen/src/Core/products/SelfadjointRank2Update.h +93 -0
- data/vendor/eigen/Eigen/src/Core/products/TriangularMatrixMatrix.h +466 -0
- data/vendor/eigen/Eigen/src/Core/products/TriangularMatrixMatrix_BLAS.h +315 -0
- data/vendor/eigen/Eigen/src/Core/products/TriangularMatrixVector.h +350 -0
- data/vendor/eigen/Eigen/src/Core/products/TriangularMatrixVector_BLAS.h +255 -0
- data/vendor/eigen/Eigen/src/Core/products/TriangularSolverMatrix.h +335 -0
- data/vendor/eigen/Eigen/src/Core/products/TriangularSolverMatrix_BLAS.h +163 -0
- data/vendor/eigen/Eigen/src/Core/products/TriangularSolverVector.h +145 -0
- data/vendor/eigen/Eigen/src/Core/util/BlasUtil.h +398 -0
- data/vendor/eigen/Eigen/src/Core/util/Constants.h +547 -0
- data/vendor/eigen/Eigen/src/Core/util/DisableStupidWarnings.h +83 -0
- data/vendor/eigen/Eigen/src/Core/util/ForwardDeclarations.h +302 -0
- data/vendor/eigen/Eigen/src/Core/util/MKL_support.h +130 -0
- data/vendor/eigen/Eigen/src/Core/util/Macros.h +1001 -0
- data/vendor/eigen/Eigen/src/Core/util/Memory.h +993 -0
- data/vendor/eigen/Eigen/src/Core/util/Meta.h +534 -0
- data/vendor/eigen/Eigen/src/Core/util/NonMPL2.h +3 -0
- data/vendor/eigen/Eigen/src/Core/util/ReenableStupidWarnings.h +27 -0
- data/vendor/eigen/Eigen/src/Core/util/StaticAssert.h +218 -0
- data/vendor/eigen/Eigen/src/Core/util/XprHelper.h +821 -0
- data/vendor/eigen/Eigen/src/Eigenvalues/ComplexEigenSolver.h +346 -0
- data/vendor/eigen/Eigen/src/Eigenvalues/ComplexSchur.h +459 -0
- data/vendor/eigen/Eigen/src/Eigenvalues/ComplexSchur_LAPACKE.h +91 -0
- data/vendor/eigen/Eigen/src/Eigenvalues/EigenSolver.h +622 -0
- data/vendor/eigen/Eigen/src/Eigenvalues/GeneralizedEigenSolver.h +418 -0
- data/vendor/eigen/Eigen/src/Eigenvalues/GeneralizedSelfAdjointEigenSolver.h +226 -0
- data/vendor/eigen/Eigen/src/Eigenvalues/HessenbergDecomposition.h +374 -0
- data/vendor/eigen/Eigen/src/Eigenvalues/MatrixBaseEigenvalues.h +158 -0
- data/vendor/eigen/Eigen/src/Eigenvalues/RealQZ.h +654 -0
- data/vendor/eigen/Eigen/src/Eigenvalues/RealSchur.h +546 -0
- data/vendor/eigen/Eigen/src/Eigenvalues/RealSchur_LAPACKE.h +77 -0
- data/vendor/eigen/Eigen/src/Eigenvalues/SelfAdjointEigenSolver.h +870 -0
- data/vendor/eigen/Eigen/src/Eigenvalues/SelfAdjointEigenSolver_LAPACKE.h +87 -0
- data/vendor/eigen/Eigen/src/Eigenvalues/Tridiagonalization.h +556 -0
- data/vendor/eigen/Eigen/src/Geometry/AlignedBox.h +392 -0
- data/vendor/eigen/Eigen/src/Geometry/AngleAxis.h +247 -0
- data/vendor/eigen/Eigen/src/Geometry/EulerAngles.h +114 -0
- data/vendor/eigen/Eigen/src/Geometry/Homogeneous.h +497 -0
- data/vendor/eigen/Eigen/src/Geometry/Hyperplane.h +282 -0
- data/vendor/eigen/Eigen/src/Geometry/OrthoMethods.h +234 -0
- data/vendor/eigen/Eigen/src/Geometry/ParametrizedLine.h +195 -0
- data/vendor/eigen/Eigen/src/Geometry/Quaternion.h +814 -0
- data/vendor/eigen/Eigen/src/Geometry/Rotation2D.h +199 -0
- data/vendor/eigen/Eigen/src/Geometry/RotationBase.h +206 -0
- data/vendor/eigen/Eigen/src/Geometry/Scaling.h +170 -0
- data/vendor/eigen/Eigen/src/Geometry/Transform.h +1542 -0
- data/vendor/eigen/Eigen/src/Geometry/Translation.h +208 -0
- data/vendor/eigen/Eigen/src/Geometry/Umeyama.h +166 -0
- data/vendor/eigen/Eigen/src/Geometry/arch/Geometry_SSE.h +161 -0
- data/vendor/eigen/Eigen/src/Householder/BlockHouseholder.h +103 -0
- data/vendor/eigen/Eigen/src/Householder/Householder.h +172 -0
- data/vendor/eigen/Eigen/src/Householder/HouseholderSequence.h +470 -0
- data/vendor/eigen/Eigen/src/IterativeLinearSolvers/BasicPreconditioners.h +226 -0
- data/vendor/eigen/Eigen/src/IterativeLinearSolvers/BiCGSTAB.h +228 -0
- data/vendor/eigen/Eigen/src/IterativeLinearSolvers/ConjugateGradient.h +246 -0
- data/vendor/eigen/Eigen/src/IterativeLinearSolvers/IncompleteCholesky.h +400 -0
- data/vendor/eigen/Eigen/src/IterativeLinearSolvers/IncompleteLUT.h +462 -0
- data/vendor/eigen/Eigen/src/IterativeLinearSolvers/IterativeSolverBase.h +394 -0
- data/vendor/eigen/Eigen/src/IterativeLinearSolvers/LeastSquareConjugateGradient.h +216 -0
- data/vendor/eigen/Eigen/src/IterativeLinearSolvers/SolveWithGuess.h +115 -0
- data/vendor/eigen/Eigen/src/Jacobi/Jacobi.h +462 -0
- data/vendor/eigen/Eigen/src/LU/Determinant.h +101 -0
- data/vendor/eigen/Eigen/src/LU/FullPivLU.h +891 -0
- data/vendor/eigen/Eigen/src/LU/InverseImpl.h +415 -0
- data/vendor/eigen/Eigen/src/LU/PartialPivLU.h +611 -0
- data/vendor/eigen/Eigen/src/LU/PartialPivLU_LAPACKE.h +83 -0
- data/vendor/eigen/Eigen/src/LU/arch/Inverse_SSE.h +338 -0
- data/vendor/eigen/Eigen/src/MetisSupport/MetisSupport.h +137 -0
- data/vendor/eigen/Eigen/src/OrderingMethods/Amd.h +445 -0
- data/vendor/eigen/Eigen/src/OrderingMethods/Eigen_Colamd.h +1843 -0
- data/vendor/eigen/Eigen/src/OrderingMethods/Ordering.h +157 -0
- data/vendor/eigen/Eigen/src/PaStiXSupport/PaStiXSupport.h +678 -0
- data/vendor/eigen/Eigen/src/PardisoSupport/PardisoSupport.h +543 -0
- data/vendor/eigen/Eigen/src/QR/ColPivHouseholderQR.h +653 -0
- data/vendor/eigen/Eigen/src/QR/ColPivHouseholderQR_LAPACKE.h +97 -0
- data/vendor/eigen/Eigen/src/QR/CompleteOrthogonalDecomposition.h +562 -0
- data/vendor/eigen/Eigen/src/QR/FullPivHouseholderQR.h +676 -0
- data/vendor/eigen/Eigen/src/QR/HouseholderQR.h +409 -0
- data/vendor/eigen/Eigen/src/QR/HouseholderQR_LAPACKE.h +68 -0
- data/vendor/eigen/Eigen/src/SPQRSupport/SuiteSparseQRSupport.h +313 -0
- data/vendor/eigen/Eigen/src/SVD/BDCSVD.h +1246 -0
- data/vendor/eigen/Eigen/src/SVD/JacobiSVD.h +804 -0
- data/vendor/eigen/Eigen/src/SVD/JacobiSVD_LAPACKE.h +91 -0
- data/vendor/eigen/Eigen/src/SVD/SVDBase.h +315 -0
- data/vendor/eigen/Eigen/src/SVD/UpperBidiagonalization.h +414 -0
- data/vendor/eigen/Eigen/src/SparseCholesky/SimplicialCholesky.h +689 -0
- data/vendor/eigen/Eigen/src/SparseCholesky/SimplicialCholesky_impl.h +199 -0
- data/vendor/eigen/Eigen/src/SparseCore/AmbiVector.h +377 -0
- data/vendor/eigen/Eigen/src/SparseCore/CompressedStorage.h +258 -0
- data/vendor/eigen/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h +352 -0
- data/vendor/eigen/Eigen/src/SparseCore/MappedSparseMatrix.h +67 -0
- data/vendor/eigen/Eigen/src/SparseCore/SparseAssign.h +216 -0
- data/vendor/eigen/Eigen/src/SparseCore/SparseBlock.h +603 -0
- data/vendor/eigen/Eigen/src/SparseCore/SparseColEtree.h +206 -0
- data/vendor/eigen/Eigen/src/SparseCore/SparseCompressedBase.h +341 -0
- data/vendor/eigen/Eigen/src/SparseCore/SparseCwiseBinaryOp.h +726 -0
- data/vendor/eigen/Eigen/src/SparseCore/SparseCwiseUnaryOp.h +148 -0
- data/vendor/eigen/Eigen/src/SparseCore/SparseDenseProduct.h +320 -0
- data/vendor/eigen/Eigen/src/SparseCore/SparseDiagonalProduct.h +138 -0
- data/vendor/eigen/Eigen/src/SparseCore/SparseDot.h +98 -0
- data/vendor/eigen/Eigen/src/SparseCore/SparseFuzzy.h +29 -0
- data/vendor/eigen/Eigen/src/SparseCore/SparseMap.h +305 -0
- data/vendor/eigen/Eigen/src/SparseCore/SparseMatrix.h +1403 -0
- data/vendor/eigen/Eigen/src/SparseCore/SparseMatrixBase.h +405 -0
- data/vendor/eigen/Eigen/src/SparseCore/SparsePermutation.h +178 -0
- data/vendor/eigen/Eigen/src/SparseCore/SparseProduct.h +169 -0
- data/vendor/eigen/Eigen/src/SparseCore/SparseRedux.h +49 -0
- data/vendor/eigen/Eigen/src/SparseCore/SparseRef.h +397 -0
- data/vendor/eigen/Eigen/src/SparseCore/SparseSelfAdjointView.h +656 -0
- data/vendor/eigen/Eigen/src/SparseCore/SparseSolverBase.h +124 -0
- data/vendor/eigen/Eigen/src/SparseCore/SparseSparseProductWithPruning.h +198 -0
- data/vendor/eigen/Eigen/src/SparseCore/SparseTranspose.h +92 -0
- data/vendor/eigen/Eigen/src/SparseCore/SparseTriangularView.h +189 -0
- data/vendor/eigen/Eigen/src/SparseCore/SparseUtil.h +178 -0
- data/vendor/eigen/Eigen/src/SparseCore/SparseVector.h +478 -0
- data/vendor/eigen/Eigen/src/SparseCore/SparseView.h +253 -0
- data/vendor/eigen/Eigen/src/SparseCore/TriangularSolver.h +315 -0
- data/vendor/eigen/Eigen/src/SparseLU/SparseLU.h +773 -0
- data/vendor/eigen/Eigen/src/SparseLU/SparseLUImpl.h +66 -0
- data/vendor/eigen/Eigen/src/SparseLU/SparseLU_Memory.h +226 -0
- data/vendor/eigen/Eigen/src/SparseLU/SparseLU_Structs.h +110 -0
- data/vendor/eigen/Eigen/src/SparseLU/SparseLU_SupernodalMatrix.h +301 -0
- data/vendor/eigen/Eigen/src/SparseLU/SparseLU_Utils.h +80 -0
- data/vendor/eigen/Eigen/src/SparseLU/SparseLU_column_bmod.h +181 -0
- data/vendor/eigen/Eigen/src/SparseLU/SparseLU_column_dfs.h +179 -0
- data/vendor/eigen/Eigen/src/SparseLU/SparseLU_copy_to_ucol.h +107 -0
- data/vendor/eigen/Eigen/src/SparseLU/SparseLU_gemm_kernel.h +280 -0
- data/vendor/eigen/Eigen/src/SparseLU/SparseLU_heap_relax_snode.h +126 -0
- data/vendor/eigen/Eigen/src/SparseLU/SparseLU_kernel_bmod.h +130 -0
- data/vendor/eigen/Eigen/src/SparseLU/SparseLU_panel_bmod.h +223 -0
- data/vendor/eigen/Eigen/src/SparseLU/SparseLU_panel_dfs.h +258 -0
- data/vendor/eigen/Eigen/src/SparseLU/SparseLU_pivotL.h +137 -0
- data/vendor/eigen/Eigen/src/SparseLU/SparseLU_pruneL.h +136 -0
- data/vendor/eigen/Eigen/src/SparseLU/SparseLU_relax_snode.h +83 -0
- data/vendor/eigen/Eigen/src/SparseQR/SparseQR.h +745 -0
- data/vendor/eigen/Eigen/src/StlSupport/StdDeque.h +126 -0
- data/vendor/eigen/Eigen/src/StlSupport/StdList.h +106 -0
- data/vendor/eigen/Eigen/src/StlSupport/StdVector.h +131 -0
- data/vendor/eigen/Eigen/src/StlSupport/details.h +84 -0
- data/vendor/eigen/Eigen/src/SuperLUSupport/SuperLUSupport.h +1027 -0
- data/vendor/eigen/Eigen/src/UmfPackSupport/UmfPackSupport.h +506 -0
- data/vendor/eigen/Eigen/src/misc/Image.h +82 -0
- data/vendor/eigen/Eigen/src/misc/Kernel.h +79 -0
- data/vendor/eigen/Eigen/src/misc/RealSvd2x2.h +55 -0
- data/vendor/eigen/Eigen/src/misc/blas.h +440 -0
- data/vendor/eigen/Eigen/src/misc/lapack.h +152 -0
- data/vendor/eigen/Eigen/src/misc/lapacke.h +16291 -0
- data/vendor/eigen/Eigen/src/misc/lapacke_mangling.h +17 -0
- data/vendor/eigen/Eigen/src/plugins/ArrayCwiseBinaryOps.h +332 -0
- data/vendor/eigen/Eigen/src/plugins/ArrayCwiseUnaryOps.h +552 -0
- data/vendor/eigen/Eigen/src/plugins/BlockMethods.h +1058 -0
- data/vendor/eigen/Eigen/src/plugins/CommonCwiseBinaryOps.h +115 -0
- data/vendor/eigen/Eigen/src/plugins/CommonCwiseUnaryOps.h +163 -0
- data/vendor/eigen/Eigen/src/plugins/MatrixCwiseBinaryOps.h +152 -0
- data/vendor/eigen/Eigen/src/plugins/MatrixCwiseUnaryOps.h +85 -0
- data/vendor/eigen/README.md +3 -0
- data/vendor/eigen/bench/README.txt +55 -0
- data/vendor/eigen/bench/btl/COPYING +340 -0
- data/vendor/eigen/bench/btl/README +154 -0
- data/vendor/eigen/bench/tensors/README +21 -0
- data/vendor/eigen/blas/README.txt +6 -0
- data/vendor/eigen/demos/mandelbrot/README +10 -0
- data/vendor/eigen/demos/mix_eigen_and_c/README +9 -0
- data/vendor/eigen/demos/opengl/README +13 -0
- data/vendor/eigen/unsupported/Eigen/CXX11/src/Tensor/README.md +1760 -0
- data/vendor/eigen/unsupported/README.txt +50 -0
- data/vendor/tomotopy/LICENSE +21 -0
- data/vendor/tomotopy/README.kr.rst +375 -0
- data/vendor/tomotopy/README.rst +382 -0
- data/vendor/tomotopy/src/Labeling/FoRelevance.cpp +362 -0
- data/vendor/tomotopy/src/Labeling/FoRelevance.h +88 -0
- data/vendor/tomotopy/src/Labeling/Labeler.h +50 -0
- data/vendor/tomotopy/src/TopicModel/CT.h +37 -0
- data/vendor/tomotopy/src/TopicModel/CTModel.cpp +13 -0
- data/vendor/tomotopy/src/TopicModel/CTModel.hpp +293 -0
- data/vendor/tomotopy/src/TopicModel/DMR.h +51 -0
- data/vendor/tomotopy/src/TopicModel/DMRModel.cpp +13 -0
- data/vendor/tomotopy/src/TopicModel/DMRModel.hpp +374 -0
- data/vendor/tomotopy/src/TopicModel/DT.h +65 -0
- data/vendor/tomotopy/src/TopicModel/DTM.h +22 -0
- data/vendor/tomotopy/src/TopicModel/DTModel.cpp +15 -0
- data/vendor/tomotopy/src/TopicModel/DTModel.hpp +572 -0
- data/vendor/tomotopy/src/TopicModel/GDMR.h +37 -0
- data/vendor/tomotopy/src/TopicModel/GDMRModel.cpp +14 -0
- data/vendor/tomotopy/src/TopicModel/GDMRModel.hpp +485 -0
- data/vendor/tomotopy/src/TopicModel/HDP.h +74 -0
- data/vendor/tomotopy/src/TopicModel/HDPModel.cpp +13 -0
- data/vendor/tomotopy/src/TopicModel/HDPModel.hpp +592 -0
- data/vendor/tomotopy/src/TopicModel/HLDA.h +40 -0
- data/vendor/tomotopy/src/TopicModel/HLDAModel.cpp +13 -0
- data/vendor/tomotopy/src/TopicModel/HLDAModel.hpp +681 -0
- data/vendor/tomotopy/src/TopicModel/HPA.h +27 -0
- data/vendor/tomotopy/src/TopicModel/HPAModel.cpp +21 -0
- data/vendor/tomotopy/src/TopicModel/HPAModel.hpp +588 -0
- data/vendor/tomotopy/src/TopicModel/LDA.h +144 -0
- data/vendor/tomotopy/src/TopicModel/LDACVB0Model.hpp +442 -0
- data/vendor/tomotopy/src/TopicModel/LDAModel.cpp +13 -0
- data/vendor/tomotopy/src/TopicModel/LDAModel.hpp +1058 -0
- data/vendor/tomotopy/src/TopicModel/LLDA.h +45 -0
- data/vendor/tomotopy/src/TopicModel/LLDAModel.cpp +13 -0
- data/vendor/tomotopy/src/TopicModel/LLDAModel.hpp +203 -0
- data/vendor/tomotopy/src/TopicModel/MGLDA.h +63 -0
- data/vendor/tomotopy/src/TopicModel/MGLDAModel.cpp +17 -0
- data/vendor/tomotopy/src/TopicModel/MGLDAModel.hpp +558 -0
- data/vendor/tomotopy/src/TopicModel/PA.h +43 -0
- data/vendor/tomotopy/src/TopicModel/PAModel.cpp +13 -0
- data/vendor/tomotopy/src/TopicModel/PAModel.hpp +467 -0
- data/vendor/tomotopy/src/TopicModel/PLDA.h +17 -0
- data/vendor/tomotopy/src/TopicModel/PLDAModel.cpp +13 -0
- data/vendor/tomotopy/src/TopicModel/PLDAModel.hpp +214 -0
- data/vendor/tomotopy/src/TopicModel/SLDA.h +54 -0
- data/vendor/tomotopy/src/TopicModel/SLDAModel.cpp +17 -0
- data/vendor/tomotopy/src/TopicModel/SLDAModel.hpp +456 -0
- data/vendor/tomotopy/src/TopicModel/TopicModel.hpp +692 -0
- data/vendor/tomotopy/src/Utils/AliasMethod.hpp +169 -0
- data/vendor/tomotopy/src/Utils/Dictionary.h +80 -0
- data/vendor/tomotopy/src/Utils/EigenAddonOps.hpp +181 -0
- data/vendor/tomotopy/src/Utils/LBFGS.h +202 -0
- data/vendor/tomotopy/src/Utils/LBFGS/LineSearchBacktracking.h +120 -0
- data/vendor/tomotopy/src/Utils/LBFGS/LineSearchBracketing.h +122 -0
- data/vendor/tomotopy/src/Utils/LBFGS/Param.h +213 -0
- data/vendor/tomotopy/src/Utils/LUT.hpp +82 -0
- data/vendor/tomotopy/src/Utils/MultiNormalDistribution.hpp +69 -0
- data/vendor/tomotopy/src/Utils/PolyaGamma.hpp +200 -0
- data/vendor/tomotopy/src/Utils/PolyaGammaHybrid.hpp +672 -0
- data/vendor/tomotopy/src/Utils/ThreadPool.hpp +150 -0
- data/vendor/tomotopy/src/Utils/Trie.hpp +220 -0
- data/vendor/tomotopy/src/Utils/TruncMultiNormal.hpp +94 -0
- data/vendor/tomotopy/src/Utils/Utils.hpp +337 -0
- data/vendor/tomotopy/src/Utils/avx_gamma.h +46 -0
- data/vendor/tomotopy/src/Utils/avx_mathfun.h +736 -0
- data/vendor/tomotopy/src/Utils/exception.h +28 -0
- data/vendor/tomotopy/src/Utils/math.h +281 -0
- data/vendor/tomotopy/src/Utils/rtnorm.hpp +2690 -0
- data/vendor/tomotopy/src/Utils/sample.hpp +192 -0
- data/vendor/tomotopy/src/Utils/serializer.hpp +695 -0
- data/vendor/tomotopy/src/Utils/slp.hpp +131 -0
- data/vendor/tomotopy/src/Utils/sse_gamma.h +48 -0
- data/vendor/tomotopy/src/Utils/sse_mathfun.h +710 -0
- data/vendor/tomotopy/src/Utils/text.hpp +49 -0
- data/vendor/tomotopy/src/Utils/tvector.hpp +543 -0
- metadata +531 -0
@@ -0,0 +1,67 @@
|
|
1
|
+
// This file is part of Eigen, a lightweight C++ template library
|
2
|
+
// for linear algebra.
|
3
|
+
//
|
4
|
+
// Copyright (C) 2008-2014 Gael Guennebaud <gael.guennebaud@inria.fr>
|
5
|
+
//
|
6
|
+
// This Source Code Form is subject to the terms of the Mozilla
|
7
|
+
// Public License v. 2.0. If a copy of the MPL was not distributed
|
8
|
+
// with this file, You can obtain one at http://mozilla.org/MPL/2.0/.
|
9
|
+
|
10
|
+
#ifndef EIGEN_MAPPED_SPARSEMATRIX_H
|
11
|
+
#define EIGEN_MAPPED_SPARSEMATRIX_H
|
12
|
+
|
13
|
+
namespace Eigen {
|
14
|
+
|
15
|
+
/** \deprecated Use Map<SparseMatrix<> >
|
16
|
+
* \class MappedSparseMatrix
|
17
|
+
*
|
18
|
+
* \brief Sparse matrix
|
19
|
+
*
|
20
|
+
* \param _Scalar the scalar type, i.e. the type of the coefficients
|
21
|
+
*
|
22
|
+
* See http://www.netlib.org/linalg/html_templates/node91.html for details on the storage scheme.
|
23
|
+
*
|
24
|
+
*/
|
25
|
+
namespace internal {
|
26
|
+
template<typename _Scalar, int _Flags, typename _StorageIndex>
|
27
|
+
struct traits<MappedSparseMatrix<_Scalar, _Flags, _StorageIndex> > : traits<SparseMatrix<_Scalar, _Flags, _StorageIndex> >
|
28
|
+
{};
|
29
|
+
} // end namespace internal
|
30
|
+
|
31
|
+
template<typename _Scalar, int _Flags, typename _StorageIndex>
|
32
|
+
class MappedSparseMatrix
|
33
|
+
: public Map<SparseMatrix<_Scalar, _Flags, _StorageIndex> >
|
34
|
+
{
|
35
|
+
typedef Map<SparseMatrix<_Scalar, _Flags, _StorageIndex> > Base;
|
36
|
+
|
37
|
+
public:
|
38
|
+
|
39
|
+
typedef typename Base::StorageIndex StorageIndex;
|
40
|
+
typedef typename Base::Scalar Scalar;
|
41
|
+
|
42
|
+
inline MappedSparseMatrix(Index rows, Index cols, Index nnz, StorageIndex* outerIndexPtr, StorageIndex* innerIndexPtr, Scalar* valuePtr, StorageIndex* innerNonZeroPtr = 0)
|
43
|
+
: Base(rows, cols, nnz, outerIndexPtr, innerIndexPtr, valuePtr, innerNonZeroPtr)
|
44
|
+
{}
|
45
|
+
|
46
|
+
/** Empty destructor */
|
47
|
+
inline ~MappedSparseMatrix() {}
|
48
|
+
};
|
49
|
+
|
50
|
+
namespace internal {
|
51
|
+
|
52
|
+
template<typename _Scalar, int _Options, typename _StorageIndex>
|
53
|
+
struct evaluator<MappedSparseMatrix<_Scalar,_Options,_StorageIndex> >
|
54
|
+
: evaluator<SparseCompressedBase<MappedSparseMatrix<_Scalar,_Options,_StorageIndex> > >
|
55
|
+
{
|
56
|
+
typedef MappedSparseMatrix<_Scalar,_Options,_StorageIndex> XprType;
|
57
|
+
typedef evaluator<SparseCompressedBase<XprType> > Base;
|
58
|
+
|
59
|
+
evaluator() : Base() {}
|
60
|
+
explicit evaluator(const XprType &mat) : Base(mat) {}
|
61
|
+
};
|
62
|
+
|
63
|
+
}
|
64
|
+
|
65
|
+
} // end namespace Eigen
|
66
|
+
|
67
|
+
#endif // EIGEN_MAPPED_SPARSEMATRIX_H
|
@@ -0,0 +1,216 @@
|
|
1
|
+
// This file is part of Eigen, a lightweight C++ template library
|
2
|
+
// for linear algebra.
|
3
|
+
//
|
4
|
+
// Copyright (C) 2008-2014 Gael Guennebaud <gael.guennebaud@inria.fr>
|
5
|
+
//
|
6
|
+
// This Source Code Form is subject to the terms of the Mozilla
|
7
|
+
// Public License v. 2.0. If a copy of the MPL was not distributed
|
8
|
+
// with this file, You can obtain one at http://mozilla.org/MPL/2.0/.
|
9
|
+
|
10
|
+
#ifndef EIGEN_SPARSEASSIGN_H
|
11
|
+
#define EIGEN_SPARSEASSIGN_H
|
12
|
+
|
13
|
+
namespace Eigen {
|
14
|
+
|
15
|
+
template<typename Derived>
|
16
|
+
template<typename OtherDerived>
|
17
|
+
Derived& SparseMatrixBase<Derived>::operator=(const EigenBase<OtherDerived> &other)
|
18
|
+
{
|
19
|
+
internal::call_assignment_no_alias(derived(), other.derived());
|
20
|
+
return derived();
|
21
|
+
}
|
22
|
+
|
23
|
+
template<typename Derived>
|
24
|
+
template<typename OtherDerived>
|
25
|
+
Derived& SparseMatrixBase<Derived>::operator=(const ReturnByValue<OtherDerived>& other)
|
26
|
+
{
|
27
|
+
// TODO use the evaluator mechanism
|
28
|
+
other.evalTo(derived());
|
29
|
+
return derived();
|
30
|
+
}
|
31
|
+
|
32
|
+
template<typename Derived>
|
33
|
+
template<typename OtherDerived>
|
34
|
+
inline Derived& SparseMatrixBase<Derived>::operator=(const SparseMatrixBase<OtherDerived>& other)
|
35
|
+
{
|
36
|
+
// by default sparse evaluation do not alias, so we can safely bypass the generic call_assignment routine
|
37
|
+
internal::Assignment<Derived,OtherDerived,internal::assign_op<Scalar,typename OtherDerived::Scalar> >
|
38
|
+
::run(derived(), other.derived(), internal::assign_op<Scalar,typename OtherDerived::Scalar>());
|
39
|
+
return derived();
|
40
|
+
}
|
41
|
+
|
42
|
+
template<typename Derived>
|
43
|
+
inline Derived& SparseMatrixBase<Derived>::operator=(const Derived& other)
|
44
|
+
{
|
45
|
+
internal::call_assignment_no_alias(derived(), other.derived());
|
46
|
+
return derived();
|
47
|
+
}
|
48
|
+
|
49
|
+
namespace internal {
|
50
|
+
|
51
|
+
template<>
|
52
|
+
struct storage_kind_to_evaluator_kind<Sparse> {
|
53
|
+
typedef IteratorBased Kind;
|
54
|
+
};
|
55
|
+
|
56
|
+
template<>
|
57
|
+
struct storage_kind_to_shape<Sparse> {
|
58
|
+
typedef SparseShape Shape;
|
59
|
+
};
|
60
|
+
|
61
|
+
struct Sparse2Sparse {};
|
62
|
+
struct Sparse2Dense {};
|
63
|
+
|
64
|
+
template<> struct AssignmentKind<SparseShape, SparseShape> { typedef Sparse2Sparse Kind; };
|
65
|
+
template<> struct AssignmentKind<SparseShape, SparseTriangularShape> { typedef Sparse2Sparse Kind; };
|
66
|
+
template<> struct AssignmentKind<DenseShape, SparseShape> { typedef Sparse2Dense Kind; };
|
67
|
+
template<> struct AssignmentKind<DenseShape, SparseTriangularShape> { typedef Sparse2Dense Kind; };
|
68
|
+
|
69
|
+
|
70
|
+
template<typename DstXprType, typename SrcXprType>
|
71
|
+
void assign_sparse_to_sparse(DstXprType &dst, const SrcXprType &src)
|
72
|
+
{
|
73
|
+
typedef typename DstXprType::Scalar Scalar;
|
74
|
+
typedef internal::evaluator<DstXprType> DstEvaluatorType;
|
75
|
+
typedef internal::evaluator<SrcXprType> SrcEvaluatorType;
|
76
|
+
|
77
|
+
SrcEvaluatorType srcEvaluator(src);
|
78
|
+
|
79
|
+
const bool transpose = (DstEvaluatorType::Flags & RowMajorBit) != (SrcEvaluatorType::Flags & RowMajorBit);
|
80
|
+
const Index outerEvaluationSize = (SrcEvaluatorType::Flags&RowMajorBit) ? src.rows() : src.cols();
|
81
|
+
if ((!transpose) && src.isRValue())
|
82
|
+
{
|
83
|
+
// eval without temporary
|
84
|
+
dst.resize(src.rows(), src.cols());
|
85
|
+
dst.setZero();
|
86
|
+
dst.reserve((std::max)(src.rows(),src.cols())*2);
|
87
|
+
for (Index j=0; j<outerEvaluationSize; ++j)
|
88
|
+
{
|
89
|
+
dst.startVec(j);
|
90
|
+
for (typename SrcEvaluatorType::InnerIterator it(srcEvaluator, j); it; ++it)
|
91
|
+
{
|
92
|
+
Scalar v = it.value();
|
93
|
+
dst.insertBackByOuterInner(j,it.index()) = v;
|
94
|
+
}
|
95
|
+
}
|
96
|
+
dst.finalize();
|
97
|
+
}
|
98
|
+
else
|
99
|
+
{
|
100
|
+
// eval through a temporary
|
101
|
+
eigen_assert(( ((internal::traits<DstXprType>::SupportedAccessPatterns & OuterRandomAccessPattern)==OuterRandomAccessPattern) ||
|
102
|
+
(!((DstEvaluatorType::Flags & RowMajorBit) != (SrcEvaluatorType::Flags & RowMajorBit)))) &&
|
103
|
+
"the transpose operation is supposed to be handled in SparseMatrix::operator=");
|
104
|
+
|
105
|
+
enum { Flip = (DstEvaluatorType::Flags & RowMajorBit) != (SrcEvaluatorType::Flags & RowMajorBit) };
|
106
|
+
|
107
|
+
|
108
|
+
DstXprType temp(src.rows(), src.cols());
|
109
|
+
|
110
|
+
temp.reserve((std::max)(src.rows(),src.cols())*2);
|
111
|
+
for (Index j=0; j<outerEvaluationSize; ++j)
|
112
|
+
{
|
113
|
+
temp.startVec(j);
|
114
|
+
for (typename SrcEvaluatorType::InnerIterator it(srcEvaluator, j); it; ++it)
|
115
|
+
{
|
116
|
+
Scalar v = it.value();
|
117
|
+
temp.insertBackByOuterInner(Flip?it.index():j,Flip?j:it.index()) = v;
|
118
|
+
}
|
119
|
+
}
|
120
|
+
temp.finalize();
|
121
|
+
|
122
|
+
dst = temp.markAsRValue();
|
123
|
+
}
|
124
|
+
}
|
125
|
+
|
126
|
+
// Generic Sparse to Sparse assignment
|
127
|
+
template< typename DstXprType, typename SrcXprType, typename Functor>
|
128
|
+
struct Assignment<DstXprType, SrcXprType, Functor, Sparse2Sparse>
|
129
|
+
{
|
130
|
+
static void run(DstXprType &dst, const SrcXprType &src, const internal::assign_op<typename DstXprType::Scalar,typename SrcXprType::Scalar> &/*func*/)
|
131
|
+
{
|
132
|
+
assign_sparse_to_sparse(dst.derived(), src.derived());
|
133
|
+
}
|
134
|
+
};
|
135
|
+
|
136
|
+
// Generic Sparse to Dense assignment
|
137
|
+
template< typename DstXprType, typename SrcXprType, typename Functor>
|
138
|
+
struct Assignment<DstXprType, SrcXprType, Functor, Sparse2Dense>
|
139
|
+
{
|
140
|
+
static void run(DstXprType &dst, const SrcXprType &src, const Functor &func)
|
141
|
+
{
|
142
|
+
if(internal::is_same<Functor,internal::assign_op<typename DstXprType::Scalar,typename SrcXprType::Scalar> >::value)
|
143
|
+
dst.setZero();
|
144
|
+
|
145
|
+
internal::evaluator<SrcXprType> srcEval(src);
|
146
|
+
resize_if_allowed(dst, src, func);
|
147
|
+
internal::evaluator<DstXprType> dstEval(dst);
|
148
|
+
|
149
|
+
const Index outerEvaluationSize = (internal::evaluator<SrcXprType>::Flags&RowMajorBit) ? src.rows() : src.cols();
|
150
|
+
for (Index j=0; j<outerEvaluationSize; ++j)
|
151
|
+
for (typename internal::evaluator<SrcXprType>::InnerIterator i(srcEval,j); i; ++i)
|
152
|
+
func.assignCoeff(dstEval.coeffRef(i.row(),i.col()), i.value());
|
153
|
+
}
|
154
|
+
};
|
155
|
+
|
156
|
+
// Specialization for "dst = dec.solve(rhs)"
|
157
|
+
// NOTE we need to specialize it for Sparse2Sparse to avoid ambiguous specialization error
|
158
|
+
template<typename DstXprType, typename DecType, typename RhsType, typename Scalar>
|
159
|
+
struct Assignment<DstXprType, Solve<DecType,RhsType>, internal::assign_op<Scalar,Scalar>, Sparse2Sparse>
|
160
|
+
{
|
161
|
+
typedef Solve<DecType,RhsType> SrcXprType;
|
162
|
+
static void run(DstXprType &dst, const SrcXprType &src, const internal::assign_op<Scalar,Scalar> &)
|
163
|
+
{
|
164
|
+
Index dstRows = src.rows();
|
165
|
+
Index dstCols = src.cols();
|
166
|
+
if((dst.rows()!=dstRows) || (dst.cols()!=dstCols))
|
167
|
+
dst.resize(dstRows, dstCols);
|
168
|
+
|
169
|
+
src.dec()._solve_impl(src.rhs(), dst);
|
170
|
+
}
|
171
|
+
};
|
172
|
+
|
173
|
+
struct Diagonal2Sparse {};
|
174
|
+
|
175
|
+
template<> struct AssignmentKind<SparseShape,DiagonalShape> { typedef Diagonal2Sparse Kind; };
|
176
|
+
|
177
|
+
template< typename DstXprType, typename SrcXprType, typename Functor>
|
178
|
+
struct Assignment<DstXprType, SrcXprType, Functor, Diagonal2Sparse>
|
179
|
+
{
|
180
|
+
typedef typename DstXprType::StorageIndex StorageIndex;
|
181
|
+
typedef typename DstXprType::Scalar Scalar;
|
182
|
+
typedef Array<StorageIndex,Dynamic,1> ArrayXI;
|
183
|
+
typedef Array<Scalar,Dynamic,1> ArrayXS;
|
184
|
+
template<int Options>
|
185
|
+
static void run(SparseMatrix<Scalar,Options,StorageIndex> &dst, const SrcXprType &src, const internal::assign_op<typename DstXprType::Scalar,typename SrcXprType::Scalar> &/*func*/)
|
186
|
+
{
|
187
|
+
Index dstRows = src.rows();
|
188
|
+
Index dstCols = src.cols();
|
189
|
+
if((dst.rows()!=dstRows) || (dst.cols()!=dstCols))
|
190
|
+
dst.resize(dstRows, dstCols);
|
191
|
+
|
192
|
+
Index size = src.diagonal().size();
|
193
|
+
dst.makeCompressed();
|
194
|
+
dst.resizeNonZeros(size);
|
195
|
+
Map<ArrayXI>(dst.innerIndexPtr(), size).setLinSpaced(0,StorageIndex(size)-1);
|
196
|
+
Map<ArrayXI>(dst.outerIndexPtr(), size+1).setLinSpaced(0,StorageIndex(size));
|
197
|
+
Map<ArrayXS>(dst.valuePtr(), size) = src.diagonal();
|
198
|
+
}
|
199
|
+
|
200
|
+
template<typename DstDerived>
|
201
|
+
static void run(SparseMatrixBase<DstDerived> &dst, const SrcXprType &src, const internal::assign_op<typename DstXprType::Scalar,typename SrcXprType::Scalar> &/*func*/)
|
202
|
+
{
|
203
|
+
dst.diagonal() = src.diagonal();
|
204
|
+
}
|
205
|
+
|
206
|
+
static void run(DstXprType &dst, const SrcXprType &src, const internal::add_assign_op<typename DstXprType::Scalar,typename SrcXprType::Scalar> &/*func*/)
|
207
|
+
{ dst.diagonal() += src.diagonal(); }
|
208
|
+
|
209
|
+
static void run(DstXprType &dst, const SrcXprType &src, const internal::sub_assign_op<typename DstXprType::Scalar,typename SrcXprType::Scalar> &/*func*/)
|
210
|
+
{ dst.diagonal() -= src.diagonal(); }
|
211
|
+
};
|
212
|
+
} // end namespace internal
|
213
|
+
|
214
|
+
} // end namespace Eigen
|
215
|
+
|
216
|
+
#endif // EIGEN_SPARSEASSIGN_H
|
@@ -0,0 +1,603 @@
|
|
1
|
+
// This file is part of Eigen, a lightweight C++ template library
|
2
|
+
// for linear algebra.
|
3
|
+
//
|
4
|
+
// Copyright (C) 2008-2014 Gael Guennebaud <gael.guennebaud@inria.fr>
|
5
|
+
//
|
6
|
+
// This Source Code Form is subject to the terms of the Mozilla
|
7
|
+
// Public License v. 2.0. If a copy of the MPL was not distributed
|
8
|
+
// with this file, You can obtain one at http://mozilla.org/MPL/2.0/.
|
9
|
+
|
10
|
+
#ifndef EIGEN_SPARSE_BLOCK_H
|
11
|
+
#define EIGEN_SPARSE_BLOCK_H
|
12
|
+
|
13
|
+
namespace Eigen {
|
14
|
+
|
15
|
+
// Subset of columns or rows
|
16
|
+
template<typename XprType, int BlockRows, int BlockCols>
|
17
|
+
class BlockImpl<XprType,BlockRows,BlockCols,true,Sparse>
|
18
|
+
: public SparseMatrixBase<Block<XprType,BlockRows,BlockCols,true> >
|
19
|
+
{
|
20
|
+
typedef typename internal::remove_all<typename XprType::Nested>::type _MatrixTypeNested;
|
21
|
+
typedef Block<XprType, BlockRows, BlockCols, true> BlockType;
|
22
|
+
public:
|
23
|
+
enum { IsRowMajor = internal::traits<BlockType>::IsRowMajor };
|
24
|
+
protected:
|
25
|
+
enum { OuterSize = IsRowMajor ? BlockRows : BlockCols };
|
26
|
+
typedef SparseMatrixBase<BlockType> Base;
|
27
|
+
using Base::convert_index;
|
28
|
+
public:
|
29
|
+
EIGEN_SPARSE_PUBLIC_INTERFACE(BlockType)
|
30
|
+
|
31
|
+
inline BlockImpl(XprType& xpr, Index i)
|
32
|
+
: m_matrix(xpr), m_outerStart(convert_index(i)), m_outerSize(OuterSize)
|
33
|
+
{}
|
34
|
+
|
35
|
+
inline BlockImpl(XprType& xpr, Index startRow, Index startCol, Index blockRows, Index blockCols)
|
36
|
+
: m_matrix(xpr), m_outerStart(convert_index(IsRowMajor ? startRow : startCol)), m_outerSize(convert_index(IsRowMajor ? blockRows : blockCols))
|
37
|
+
{}
|
38
|
+
|
39
|
+
EIGEN_STRONG_INLINE Index rows() const { return IsRowMajor ? m_outerSize.value() : m_matrix.rows(); }
|
40
|
+
EIGEN_STRONG_INLINE Index cols() const { return IsRowMajor ? m_matrix.cols() : m_outerSize.value(); }
|
41
|
+
|
42
|
+
Index nonZeros() const
|
43
|
+
{
|
44
|
+
typedef internal::evaluator<XprType> EvaluatorType;
|
45
|
+
EvaluatorType matEval(m_matrix);
|
46
|
+
Index nnz = 0;
|
47
|
+
Index end = m_outerStart + m_outerSize.value();
|
48
|
+
for(Index j=m_outerStart; j<end; ++j)
|
49
|
+
for(typename EvaluatorType::InnerIterator it(matEval, j); it; ++it)
|
50
|
+
++nnz;
|
51
|
+
return nnz;
|
52
|
+
}
|
53
|
+
|
54
|
+
inline const Scalar coeff(Index row, Index col) const
|
55
|
+
{
|
56
|
+
return m_matrix.coeff(row + (IsRowMajor ? m_outerStart : 0), col + (IsRowMajor ? 0 : m_outerStart));
|
57
|
+
}
|
58
|
+
|
59
|
+
inline const Scalar coeff(Index index) const
|
60
|
+
{
|
61
|
+
return m_matrix.coeff(IsRowMajor ? m_outerStart : index, IsRowMajor ? index : m_outerStart);
|
62
|
+
}
|
63
|
+
|
64
|
+
inline const XprType& nestedExpression() const { return m_matrix; }
|
65
|
+
inline XprType& nestedExpression() { return m_matrix; }
|
66
|
+
Index startRow() const { return IsRowMajor ? m_outerStart : 0; }
|
67
|
+
Index startCol() const { return IsRowMajor ? 0 : m_outerStart; }
|
68
|
+
Index blockRows() const { return IsRowMajor ? m_outerSize.value() : m_matrix.rows(); }
|
69
|
+
Index blockCols() const { return IsRowMajor ? m_matrix.cols() : m_outerSize.value(); }
|
70
|
+
|
71
|
+
protected:
|
72
|
+
|
73
|
+
typename internal::ref_selector<XprType>::non_const_type m_matrix;
|
74
|
+
Index m_outerStart;
|
75
|
+
const internal::variable_if_dynamic<Index, OuterSize> m_outerSize;
|
76
|
+
|
77
|
+
protected:
|
78
|
+
// Disable assignment with clear error message.
|
79
|
+
// Note that simply removing operator= yields compilation errors with ICC+MSVC
|
80
|
+
template<typename T>
|
81
|
+
BlockImpl& operator=(const T&)
|
82
|
+
{
|
83
|
+
EIGEN_STATIC_ASSERT(sizeof(T)==0, THIS_SPARSE_BLOCK_SUBEXPRESSION_IS_READ_ONLY);
|
84
|
+
return *this;
|
85
|
+
}
|
86
|
+
};
|
87
|
+
|
88
|
+
|
89
|
+
/***************************************************************************
|
90
|
+
* specialization for SparseMatrix
|
91
|
+
***************************************************************************/
|
92
|
+
|
93
|
+
namespace internal {
|
94
|
+
|
95
|
+
template<typename SparseMatrixType, int BlockRows, int BlockCols>
|
96
|
+
class sparse_matrix_block_impl
|
97
|
+
: public SparseCompressedBase<Block<SparseMatrixType,BlockRows,BlockCols,true> >
|
98
|
+
{
|
99
|
+
typedef typename internal::remove_all<typename SparseMatrixType::Nested>::type _MatrixTypeNested;
|
100
|
+
typedef Block<SparseMatrixType, BlockRows, BlockCols, true> BlockType;
|
101
|
+
typedef SparseCompressedBase<Block<SparseMatrixType,BlockRows,BlockCols,true> > Base;
|
102
|
+
using Base::convert_index;
|
103
|
+
public:
|
104
|
+
enum { IsRowMajor = internal::traits<BlockType>::IsRowMajor };
|
105
|
+
EIGEN_SPARSE_PUBLIC_INTERFACE(BlockType)
|
106
|
+
protected:
|
107
|
+
typedef typename Base::IndexVector IndexVector;
|
108
|
+
enum { OuterSize = IsRowMajor ? BlockRows : BlockCols };
|
109
|
+
public:
|
110
|
+
|
111
|
+
inline sparse_matrix_block_impl(SparseMatrixType& xpr, Index i)
|
112
|
+
: m_matrix(xpr), m_outerStart(convert_index(i)), m_outerSize(OuterSize)
|
113
|
+
{}
|
114
|
+
|
115
|
+
inline sparse_matrix_block_impl(SparseMatrixType& xpr, Index startRow, Index startCol, Index blockRows, Index blockCols)
|
116
|
+
: m_matrix(xpr), m_outerStart(convert_index(IsRowMajor ? startRow : startCol)), m_outerSize(convert_index(IsRowMajor ? blockRows : blockCols))
|
117
|
+
{}
|
118
|
+
|
119
|
+
template<typename OtherDerived>
|
120
|
+
inline BlockType& operator=(const SparseMatrixBase<OtherDerived>& other)
|
121
|
+
{
|
122
|
+
typedef typename internal::remove_all<typename SparseMatrixType::Nested>::type _NestedMatrixType;
|
123
|
+
_NestedMatrixType& matrix = m_matrix;
|
124
|
+
// This assignment is slow if this vector set is not empty
|
125
|
+
// and/or it is not at the end of the nonzeros of the underlying matrix.
|
126
|
+
|
127
|
+
// 1 - eval to a temporary to avoid transposition and/or aliasing issues
|
128
|
+
Ref<const SparseMatrix<Scalar, IsRowMajor ? RowMajor : ColMajor, StorageIndex> > tmp(other.derived());
|
129
|
+
eigen_internal_assert(tmp.outerSize()==m_outerSize.value());
|
130
|
+
|
131
|
+
// 2 - let's check whether there is enough allocated memory
|
132
|
+
Index nnz = tmp.nonZeros();
|
133
|
+
Index start = m_outerStart==0 ? 0 : m_matrix.outerIndexPtr()[m_outerStart]; // starting position of the current block
|
134
|
+
Index end = m_matrix.outerIndexPtr()[m_outerStart+m_outerSize.value()]; // ending position of the current block
|
135
|
+
Index block_size = end - start; // available room in the current block
|
136
|
+
Index tail_size = m_matrix.outerIndexPtr()[m_matrix.outerSize()] - end;
|
137
|
+
|
138
|
+
Index free_size = m_matrix.isCompressed()
|
139
|
+
? Index(matrix.data().allocatedSize()) + block_size
|
140
|
+
: block_size;
|
141
|
+
|
142
|
+
Index tmp_start = tmp.outerIndexPtr()[0];
|
143
|
+
|
144
|
+
bool update_trailing_pointers = false;
|
145
|
+
if(nnz>free_size)
|
146
|
+
{
|
147
|
+
// realloc manually to reduce copies
|
148
|
+
typename SparseMatrixType::Storage newdata(m_matrix.data().allocatedSize() - block_size + nnz);
|
149
|
+
|
150
|
+
internal::smart_copy(m_matrix.valuePtr(), m_matrix.valuePtr() + start, newdata.valuePtr());
|
151
|
+
internal::smart_copy(m_matrix.innerIndexPtr(), m_matrix.innerIndexPtr() + start, newdata.indexPtr());
|
152
|
+
|
153
|
+
internal::smart_copy(tmp.valuePtr() + tmp_start, tmp.valuePtr() + tmp_start + nnz, newdata.valuePtr() + start);
|
154
|
+
internal::smart_copy(tmp.innerIndexPtr() + tmp_start, tmp.innerIndexPtr() + tmp_start + nnz, newdata.indexPtr() + start);
|
155
|
+
|
156
|
+
internal::smart_copy(matrix.valuePtr()+end, matrix.valuePtr()+end + tail_size, newdata.valuePtr()+start+nnz);
|
157
|
+
internal::smart_copy(matrix.innerIndexPtr()+end, matrix.innerIndexPtr()+end + tail_size, newdata.indexPtr()+start+nnz);
|
158
|
+
|
159
|
+
newdata.resize(m_matrix.outerIndexPtr()[m_matrix.outerSize()] - block_size + nnz);
|
160
|
+
|
161
|
+
matrix.data().swap(newdata);
|
162
|
+
|
163
|
+
update_trailing_pointers = true;
|
164
|
+
}
|
165
|
+
else
|
166
|
+
{
|
167
|
+
if(m_matrix.isCompressed())
|
168
|
+
{
|
169
|
+
// no need to realloc, simply copy the tail at its respective position and insert tmp
|
170
|
+
matrix.data().resize(start + nnz + tail_size);
|
171
|
+
|
172
|
+
internal::smart_memmove(matrix.valuePtr()+end, matrix.valuePtr() + end+tail_size, matrix.valuePtr() + start+nnz);
|
173
|
+
internal::smart_memmove(matrix.innerIndexPtr()+end, matrix.innerIndexPtr() + end+tail_size, matrix.innerIndexPtr() + start+nnz);
|
174
|
+
|
175
|
+
update_trailing_pointers = true;
|
176
|
+
}
|
177
|
+
|
178
|
+
internal::smart_copy(tmp.valuePtr() + tmp_start, tmp.valuePtr() + tmp_start + nnz, matrix.valuePtr() + start);
|
179
|
+
internal::smart_copy(tmp.innerIndexPtr() + tmp_start, tmp.innerIndexPtr() + tmp_start + nnz, matrix.innerIndexPtr() + start);
|
180
|
+
}
|
181
|
+
|
182
|
+
// update outer index pointers and innerNonZeros
|
183
|
+
if(IsVectorAtCompileTime)
|
184
|
+
{
|
185
|
+
if(!m_matrix.isCompressed())
|
186
|
+
matrix.innerNonZeroPtr()[m_outerStart] = StorageIndex(nnz);
|
187
|
+
matrix.outerIndexPtr()[m_outerStart] = StorageIndex(start);
|
188
|
+
}
|
189
|
+
else
|
190
|
+
{
|
191
|
+
StorageIndex p = StorageIndex(start);
|
192
|
+
for(Index k=0; k<m_outerSize.value(); ++k)
|
193
|
+
{
|
194
|
+
StorageIndex nnz_k = internal::convert_index<StorageIndex>(tmp.innerVector(k).nonZeros());
|
195
|
+
if(!m_matrix.isCompressed())
|
196
|
+
matrix.innerNonZeroPtr()[m_outerStart+k] = nnz_k;
|
197
|
+
matrix.outerIndexPtr()[m_outerStart+k] = p;
|
198
|
+
p += nnz_k;
|
199
|
+
}
|
200
|
+
}
|
201
|
+
|
202
|
+
if(update_trailing_pointers)
|
203
|
+
{
|
204
|
+
StorageIndex offset = internal::convert_index<StorageIndex>(nnz - block_size);
|
205
|
+
for(Index k = m_outerStart + m_outerSize.value(); k<=matrix.outerSize(); ++k)
|
206
|
+
{
|
207
|
+
matrix.outerIndexPtr()[k] += offset;
|
208
|
+
}
|
209
|
+
}
|
210
|
+
|
211
|
+
return derived();
|
212
|
+
}
|
213
|
+
|
214
|
+
inline BlockType& operator=(const BlockType& other)
|
215
|
+
{
|
216
|
+
return operator=<BlockType>(other);
|
217
|
+
}
|
218
|
+
|
219
|
+
inline const Scalar* valuePtr() const
|
220
|
+
{ return m_matrix.valuePtr(); }
|
221
|
+
inline Scalar* valuePtr()
|
222
|
+
{ return m_matrix.valuePtr(); }
|
223
|
+
|
224
|
+
inline const StorageIndex* innerIndexPtr() const
|
225
|
+
{ return m_matrix.innerIndexPtr(); }
|
226
|
+
inline StorageIndex* innerIndexPtr()
|
227
|
+
{ return m_matrix.innerIndexPtr(); }
|
228
|
+
|
229
|
+
inline const StorageIndex* outerIndexPtr() const
|
230
|
+
{ return m_matrix.outerIndexPtr() + m_outerStart; }
|
231
|
+
inline StorageIndex* outerIndexPtr()
|
232
|
+
{ return m_matrix.outerIndexPtr() + m_outerStart; }
|
233
|
+
|
234
|
+
inline const StorageIndex* innerNonZeroPtr() const
|
235
|
+
{ return isCompressed() ? 0 : (m_matrix.innerNonZeroPtr()+m_outerStart); }
|
236
|
+
inline StorageIndex* innerNonZeroPtr()
|
237
|
+
{ return isCompressed() ? 0 : (m_matrix.innerNonZeroPtr()+m_outerStart); }
|
238
|
+
|
239
|
+
bool isCompressed() const { return m_matrix.innerNonZeroPtr()==0; }
|
240
|
+
|
241
|
+
inline Scalar& coeffRef(Index row, Index col)
|
242
|
+
{
|
243
|
+
return m_matrix.coeffRef(row + (IsRowMajor ? m_outerStart : 0), col + (IsRowMajor ? 0 : m_outerStart));
|
244
|
+
}
|
245
|
+
|
246
|
+
inline const Scalar coeff(Index row, Index col) const
|
247
|
+
{
|
248
|
+
return m_matrix.coeff(row + (IsRowMajor ? m_outerStart : 0), col + (IsRowMajor ? 0 : m_outerStart));
|
249
|
+
}
|
250
|
+
|
251
|
+
inline const Scalar coeff(Index index) const
|
252
|
+
{
|
253
|
+
return m_matrix.coeff(IsRowMajor ? m_outerStart : index, IsRowMajor ? index : m_outerStart);
|
254
|
+
}
|
255
|
+
|
256
|
+
const Scalar& lastCoeff() const
|
257
|
+
{
|
258
|
+
EIGEN_STATIC_ASSERT_VECTOR_ONLY(sparse_matrix_block_impl);
|
259
|
+
eigen_assert(Base::nonZeros()>0);
|
260
|
+
if(m_matrix.isCompressed())
|
261
|
+
return m_matrix.valuePtr()[m_matrix.outerIndexPtr()[m_outerStart+1]-1];
|
262
|
+
else
|
263
|
+
return m_matrix.valuePtr()[m_matrix.outerIndexPtr()[m_outerStart]+m_matrix.innerNonZeroPtr()[m_outerStart]-1];
|
264
|
+
}
|
265
|
+
|
266
|
+
EIGEN_STRONG_INLINE Index rows() const { return IsRowMajor ? m_outerSize.value() : m_matrix.rows(); }
|
267
|
+
EIGEN_STRONG_INLINE Index cols() const { return IsRowMajor ? m_matrix.cols() : m_outerSize.value(); }
|
268
|
+
|
269
|
+
inline const SparseMatrixType& nestedExpression() const { return m_matrix; }
|
270
|
+
inline SparseMatrixType& nestedExpression() { return m_matrix; }
|
271
|
+
Index startRow() const { return IsRowMajor ? m_outerStart : 0; }
|
272
|
+
Index startCol() const { return IsRowMajor ? 0 : m_outerStart; }
|
273
|
+
Index blockRows() const { return IsRowMajor ? m_outerSize.value() : m_matrix.rows(); }
|
274
|
+
Index blockCols() const { return IsRowMajor ? m_matrix.cols() : m_outerSize.value(); }
|
275
|
+
|
276
|
+
protected:
|
277
|
+
|
278
|
+
typename internal::ref_selector<SparseMatrixType>::non_const_type m_matrix;
|
279
|
+
Index m_outerStart;
|
280
|
+
const internal::variable_if_dynamic<Index, OuterSize> m_outerSize;
|
281
|
+
|
282
|
+
};
|
283
|
+
|
284
|
+
} // namespace internal
|
285
|
+
|
286
|
+
template<typename _Scalar, int _Options, typename _StorageIndex, int BlockRows, int BlockCols>
|
287
|
+
class BlockImpl<SparseMatrix<_Scalar, _Options, _StorageIndex>,BlockRows,BlockCols,true,Sparse>
|
288
|
+
: public internal::sparse_matrix_block_impl<SparseMatrix<_Scalar, _Options, _StorageIndex>,BlockRows,BlockCols>
|
289
|
+
{
|
290
|
+
public:
|
291
|
+
typedef _StorageIndex StorageIndex;
|
292
|
+
typedef SparseMatrix<_Scalar, _Options, _StorageIndex> SparseMatrixType;
|
293
|
+
typedef internal::sparse_matrix_block_impl<SparseMatrixType,BlockRows,BlockCols> Base;
|
294
|
+
inline BlockImpl(SparseMatrixType& xpr, Index i)
|
295
|
+
: Base(xpr, i)
|
296
|
+
{}
|
297
|
+
|
298
|
+
inline BlockImpl(SparseMatrixType& xpr, Index startRow, Index startCol, Index blockRows, Index blockCols)
|
299
|
+
: Base(xpr, startRow, startCol, blockRows, blockCols)
|
300
|
+
{}
|
301
|
+
|
302
|
+
using Base::operator=;
|
303
|
+
};
|
304
|
+
|
305
|
+
template<typename _Scalar, int _Options, typename _StorageIndex, int BlockRows, int BlockCols>
|
306
|
+
class BlockImpl<const SparseMatrix<_Scalar, _Options, _StorageIndex>,BlockRows,BlockCols,true,Sparse>
|
307
|
+
: public internal::sparse_matrix_block_impl<const SparseMatrix<_Scalar, _Options, _StorageIndex>,BlockRows,BlockCols>
|
308
|
+
{
|
309
|
+
public:
|
310
|
+
typedef _StorageIndex StorageIndex;
|
311
|
+
typedef const SparseMatrix<_Scalar, _Options, _StorageIndex> SparseMatrixType;
|
312
|
+
typedef internal::sparse_matrix_block_impl<SparseMatrixType,BlockRows,BlockCols> Base;
|
313
|
+
inline BlockImpl(SparseMatrixType& xpr, Index i)
|
314
|
+
: Base(xpr, i)
|
315
|
+
{}
|
316
|
+
|
317
|
+
inline BlockImpl(SparseMatrixType& xpr, Index startRow, Index startCol, Index blockRows, Index blockCols)
|
318
|
+
: Base(xpr, startRow, startCol, blockRows, blockCols)
|
319
|
+
{}
|
320
|
+
|
321
|
+
using Base::operator=;
|
322
|
+
private:
|
323
|
+
template<typename Derived> BlockImpl(const SparseMatrixBase<Derived>& xpr, Index i);
|
324
|
+
template<typename Derived> BlockImpl(const SparseMatrixBase<Derived>& xpr);
|
325
|
+
};
|
326
|
+
|
327
|
+
//----------
|
328
|
+
|
329
|
+
/** \returns the \a outer -th column (resp. row) of the matrix \c *this if \c *this
|
330
|
+
* is col-major (resp. row-major).
|
331
|
+
*/
|
332
|
+
template<typename Derived>
|
333
|
+
typename SparseMatrixBase<Derived>::InnerVectorReturnType SparseMatrixBase<Derived>::innerVector(Index outer)
|
334
|
+
{ return InnerVectorReturnType(derived(), outer); }
|
335
|
+
|
336
|
+
/** \returns the \a outer -th column (resp. row) of the matrix \c *this if \c *this
|
337
|
+
* is col-major (resp. row-major). Read-only.
|
338
|
+
*/
|
339
|
+
template<typename Derived>
|
340
|
+
const typename SparseMatrixBase<Derived>::ConstInnerVectorReturnType SparseMatrixBase<Derived>::innerVector(Index outer) const
|
341
|
+
{ return ConstInnerVectorReturnType(derived(), outer); }
|
342
|
+
|
343
|
+
/** \returns the \a outer -th column (resp. row) of the matrix \c *this if \c *this
|
344
|
+
* is col-major (resp. row-major).
|
345
|
+
*/
|
346
|
+
template<typename Derived>
|
347
|
+
typename SparseMatrixBase<Derived>::InnerVectorsReturnType
|
348
|
+
SparseMatrixBase<Derived>::innerVectors(Index outerStart, Index outerSize)
|
349
|
+
{
|
350
|
+
return Block<Derived,Dynamic,Dynamic,true>(derived(),
|
351
|
+
IsRowMajor ? outerStart : 0, IsRowMajor ? 0 : outerStart,
|
352
|
+
IsRowMajor ? outerSize : rows(), IsRowMajor ? cols() : outerSize);
|
353
|
+
|
354
|
+
}
|
355
|
+
|
356
|
+
/** \returns the \a outer -th column (resp. row) of the matrix \c *this if \c *this
|
357
|
+
* is col-major (resp. row-major). Read-only.
|
358
|
+
*/
|
359
|
+
template<typename Derived>
|
360
|
+
const typename SparseMatrixBase<Derived>::ConstInnerVectorsReturnType
|
361
|
+
SparseMatrixBase<Derived>::innerVectors(Index outerStart, Index outerSize) const
|
362
|
+
{
|
363
|
+
return Block<const Derived,Dynamic,Dynamic,true>(derived(),
|
364
|
+
IsRowMajor ? outerStart : 0, IsRowMajor ? 0 : outerStart,
|
365
|
+
IsRowMajor ? outerSize : rows(), IsRowMajor ? cols() : outerSize);
|
366
|
+
|
367
|
+
}
|
368
|
+
|
369
|
+
/** Generic implementation of sparse Block expression.
|
370
|
+
* Real-only.
|
371
|
+
*/
|
372
|
+
template<typename XprType, int BlockRows, int BlockCols, bool InnerPanel>
|
373
|
+
class BlockImpl<XprType,BlockRows,BlockCols,InnerPanel,Sparse>
|
374
|
+
: public SparseMatrixBase<Block<XprType,BlockRows,BlockCols,InnerPanel> >, internal::no_assignment_operator
|
375
|
+
{
|
376
|
+
typedef Block<XprType, BlockRows, BlockCols, InnerPanel> BlockType;
|
377
|
+
typedef SparseMatrixBase<BlockType> Base;
|
378
|
+
using Base::convert_index;
|
379
|
+
public:
|
380
|
+
enum { IsRowMajor = internal::traits<BlockType>::IsRowMajor };
|
381
|
+
EIGEN_SPARSE_PUBLIC_INTERFACE(BlockType)
|
382
|
+
|
383
|
+
typedef typename internal::remove_all<typename XprType::Nested>::type _MatrixTypeNested;
|
384
|
+
|
385
|
+
/** Column or Row constructor
|
386
|
+
*/
|
387
|
+
inline BlockImpl(XprType& xpr, Index i)
|
388
|
+
: m_matrix(xpr),
|
389
|
+
m_startRow( (BlockRows==1) && (BlockCols==XprType::ColsAtCompileTime) ? convert_index(i) : 0),
|
390
|
+
m_startCol( (BlockRows==XprType::RowsAtCompileTime) && (BlockCols==1) ? convert_index(i) : 0),
|
391
|
+
m_blockRows(BlockRows==1 ? 1 : xpr.rows()),
|
392
|
+
m_blockCols(BlockCols==1 ? 1 : xpr.cols())
|
393
|
+
{}
|
394
|
+
|
395
|
+
/** Dynamic-size constructor
|
396
|
+
*/
|
397
|
+
inline BlockImpl(XprType& xpr, Index startRow, Index startCol, Index blockRows, Index blockCols)
|
398
|
+
: m_matrix(xpr), m_startRow(convert_index(startRow)), m_startCol(convert_index(startCol)), m_blockRows(convert_index(blockRows)), m_blockCols(convert_index(blockCols))
|
399
|
+
{}
|
400
|
+
|
401
|
+
inline Index rows() const { return m_blockRows.value(); }
|
402
|
+
inline Index cols() const { return m_blockCols.value(); }
|
403
|
+
|
404
|
+
inline Scalar& coeffRef(Index row, Index col)
|
405
|
+
{
|
406
|
+
return m_matrix.coeffRef(row + m_startRow.value(), col + m_startCol.value());
|
407
|
+
}
|
408
|
+
|
409
|
+
inline const Scalar coeff(Index row, Index col) const
|
410
|
+
{
|
411
|
+
return m_matrix.coeff(row + m_startRow.value(), col + m_startCol.value());
|
412
|
+
}
|
413
|
+
|
414
|
+
inline Scalar& coeffRef(Index index)
|
415
|
+
{
|
416
|
+
return m_matrix.coeffRef(m_startRow.value() + (RowsAtCompileTime == 1 ? 0 : index),
|
417
|
+
m_startCol.value() + (RowsAtCompileTime == 1 ? index : 0));
|
418
|
+
}
|
419
|
+
|
420
|
+
inline const Scalar coeff(Index index) const
|
421
|
+
{
|
422
|
+
return m_matrix.coeff(m_startRow.value() + (RowsAtCompileTime == 1 ? 0 : index),
|
423
|
+
m_startCol.value() + (RowsAtCompileTime == 1 ? index : 0));
|
424
|
+
}
|
425
|
+
|
426
|
+
inline const XprType& nestedExpression() const { return m_matrix; }
|
427
|
+
inline XprType& nestedExpression() { return m_matrix; }
|
428
|
+
Index startRow() const { return m_startRow.value(); }
|
429
|
+
Index startCol() const { return m_startCol.value(); }
|
430
|
+
Index blockRows() const { return m_blockRows.value(); }
|
431
|
+
Index blockCols() const { return m_blockCols.value(); }
|
432
|
+
|
433
|
+
protected:
|
434
|
+
// friend class internal::GenericSparseBlockInnerIteratorImpl<XprType,BlockRows,BlockCols,InnerPanel>;
|
435
|
+
friend struct internal::unary_evaluator<Block<XprType,BlockRows,BlockCols,InnerPanel>, internal::IteratorBased, Scalar >;
|
436
|
+
|
437
|
+
Index nonZeros() const { return Dynamic; }
|
438
|
+
|
439
|
+
typename internal::ref_selector<XprType>::non_const_type m_matrix;
|
440
|
+
const internal::variable_if_dynamic<Index, XprType::RowsAtCompileTime == 1 ? 0 : Dynamic> m_startRow;
|
441
|
+
const internal::variable_if_dynamic<Index, XprType::ColsAtCompileTime == 1 ? 0 : Dynamic> m_startCol;
|
442
|
+
const internal::variable_if_dynamic<Index, RowsAtCompileTime> m_blockRows;
|
443
|
+
const internal::variable_if_dynamic<Index, ColsAtCompileTime> m_blockCols;
|
444
|
+
|
445
|
+
protected:
|
446
|
+
// Disable assignment with clear error message.
|
447
|
+
// Note that simply removing operator= yields compilation errors with ICC+MSVC
|
448
|
+
template<typename T>
|
449
|
+
BlockImpl& operator=(const T&)
|
450
|
+
{
|
451
|
+
EIGEN_STATIC_ASSERT(sizeof(T)==0, THIS_SPARSE_BLOCK_SUBEXPRESSION_IS_READ_ONLY);
|
452
|
+
return *this;
|
453
|
+
}
|
454
|
+
|
455
|
+
};
|
456
|
+
|
457
|
+
namespace internal {
|
458
|
+
|
459
|
+
template<typename ArgType, int BlockRows, int BlockCols, bool InnerPanel>
|
460
|
+
struct unary_evaluator<Block<ArgType,BlockRows,BlockCols,InnerPanel>, IteratorBased >
|
461
|
+
: public evaluator_base<Block<ArgType,BlockRows,BlockCols,InnerPanel> >
|
462
|
+
{
|
463
|
+
class InnerVectorInnerIterator;
|
464
|
+
class OuterVectorInnerIterator;
|
465
|
+
public:
|
466
|
+
typedef Block<ArgType,BlockRows,BlockCols,InnerPanel> XprType;
|
467
|
+
typedef typename XprType::StorageIndex StorageIndex;
|
468
|
+
typedef typename XprType::Scalar Scalar;
|
469
|
+
|
470
|
+
enum {
|
471
|
+
IsRowMajor = XprType::IsRowMajor,
|
472
|
+
|
473
|
+
OuterVector = (BlockCols==1 && ArgType::IsRowMajor)
|
474
|
+
| // FIXME | instead of || to please GCC 4.4.0 stupid warning "suggest parentheses around &&".
|
475
|
+
// revert to || as soon as not needed anymore.
|
476
|
+
(BlockRows==1 && !ArgType::IsRowMajor),
|
477
|
+
|
478
|
+
CoeffReadCost = evaluator<ArgType>::CoeffReadCost,
|
479
|
+
Flags = XprType::Flags
|
480
|
+
};
|
481
|
+
|
482
|
+
typedef typename internal::conditional<OuterVector,OuterVectorInnerIterator,InnerVectorInnerIterator>::type InnerIterator;
|
483
|
+
|
484
|
+
explicit unary_evaluator(const XprType& op)
|
485
|
+
: m_argImpl(op.nestedExpression()), m_block(op)
|
486
|
+
{}
|
487
|
+
|
488
|
+
inline Index nonZerosEstimate() const {
|
489
|
+
Index nnz = m_block.nonZeros();
|
490
|
+
if(nnz<0)
|
491
|
+
return m_argImpl.nonZerosEstimate() * m_block.size() / m_block.nestedExpression().size();
|
492
|
+
return nnz;
|
493
|
+
}
|
494
|
+
|
495
|
+
protected:
|
496
|
+
typedef typename evaluator<ArgType>::InnerIterator EvalIterator;
|
497
|
+
|
498
|
+
evaluator<ArgType> m_argImpl;
|
499
|
+
const XprType &m_block;
|
500
|
+
};
|
501
|
+
|
502
|
+
template<typename ArgType, int BlockRows, int BlockCols, bool InnerPanel>
|
503
|
+
class unary_evaluator<Block<ArgType,BlockRows,BlockCols,InnerPanel>, IteratorBased>::InnerVectorInnerIterator
|
504
|
+
: public EvalIterator
|
505
|
+
{
|
506
|
+
enum { IsRowMajor = unary_evaluator::IsRowMajor };
|
507
|
+
const XprType& m_block;
|
508
|
+
Index m_end;
|
509
|
+
public:
|
510
|
+
|
511
|
+
EIGEN_STRONG_INLINE InnerVectorInnerIterator(const unary_evaluator& aEval, Index outer)
|
512
|
+
: EvalIterator(aEval.m_argImpl, outer + (IsRowMajor ? aEval.m_block.startRow() : aEval.m_block.startCol())),
|
513
|
+
m_block(aEval.m_block),
|
514
|
+
m_end(IsRowMajor ? aEval.m_block.startCol()+aEval.m_block.blockCols() : aEval.m_block.startRow()+aEval.m_block.blockRows())
|
515
|
+
{
|
516
|
+
while( (EvalIterator::operator bool()) && (EvalIterator::index() < (IsRowMajor ? m_block.startCol() : m_block.startRow())) )
|
517
|
+
EvalIterator::operator++();
|
518
|
+
}
|
519
|
+
|
520
|
+
inline StorageIndex index() const { return EvalIterator::index() - convert_index<StorageIndex>(IsRowMajor ? m_block.startCol() : m_block.startRow()); }
|
521
|
+
inline Index outer() const { return EvalIterator::outer() - (IsRowMajor ? m_block.startRow() : m_block.startCol()); }
|
522
|
+
inline Index row() const { return EvalIterator::row() - m_block.startRow(); }
|
523
|
+
inline Index col() const { return EvalIterator::col() - m_block.startCol(); }
|
524
|
+
|
525
|
+
inline operator bool() const { return EvalIterator::operator bool() && EvalIterator::index() < m_end; }
|
526
|
+
};
|
527
|
+
|
528
|
+
template<typename ArgType, int BlockRows, int BlockCols, bool InnerPanel>
|
529
|
+
class unary_evaluator<Block<ArgType,BlockRows,BlockCols,InnerPanel>, IteratorBased>::OuterVectorInnerIterator
|
530
|
+
{
|
531
|
+
enum { IsRowMajor = unary_evaluator::IsRowMajor };
|
532
|
+
const unary_evaluator& m_eval;
|
533
|
+
Index m_outerPos;
|
534
|
+
const Index m_innerIndex;
|
535
|
+
Index m_end;
|
536
|
+
EvalIterator m_it;
|
537
|
+
public:
|
538
|
+
|
539
|
+
EIGEN_STRONG_INLINE OuterVectorInnerIterator(const unary_evaluator& aEval, Index outer)
|
540
|
+
: m_eval(aEval),
|
541
|
+
m_outerPos( (IsRowMajor ? aEval.m_block.startCol() : aEval.m_block.startRow()) ),
|
542
|
+
m_innerIndex(IsRowMajor ? aEval.m_block.startRow() : aEval.m_block.startCol()),
|
543
|
+
m_end(IsRowMajor ? aEval.m_block.startCol()+aEval.m_block.blockCols() : aEval.m_block.startRow()+aEval.m_block.blockRows()),
|
544
|
+
m_it(m_eval.m_argImpl, m_outerPos)
|
545
|
+
{
|
546
|
+
EIGEN_UNUSED_VARIABLE(outer);
|
547
|
+
eigen_assert(outer==0);
|
548
|
+
|
549
|
+
while(m_it && m_it.index() < m_innerIndex) ++m_it;
|
550
|
+
if((!m_it) || (m_it.index()!=m_innerIndex))
|
551
|
+
++(*this);
|
552
|
+
}
|
553
|
+
|
554
|
+
inline StorageIndex index() const { return convert_index<StorageIndex>(m_outerPos - (IsRowMajor ? m_eval.m_block.startCol() : m_eval.m_block.startRow())); }
|
555
|
+
inline Index outer() const { return 0; }
|
556
|
+
inline Index row() const { return IsRowMajor ? 0 : index(); }
|
557
|
+
inline Index col() const { return IsRowMajor ? index() : 0; }
|
558
|
+
|
559
|
+
inline Scalar value() const { return m_it.value(); }
|
560
|
+
inline Scalar& valueRef() { return m_it.valueRef(); }
|
561
|
+
|
562
|
+
inline OuterVectorInnerIterator& operator++()
|
563
|
+
{
|
564
|
+
// search next non-zero entry
|
565
|
+
while(++m_outerPos<m_end)
|
566
|
+
{
|
567
|
+
// Restart iterator at the next inner-vector:
|
568
|
+
m_it.~EvalIterator();
|
569
|
+
::new (&m_it) EvalIterator(m_eval.m_argImpl, m_outerPos);
|
570
|
+
// search for the key m_innerIndex in the current outer-vector
|
571
|
+
while(m_it && m_it.index() < m_innerIndex) ++m_it;
|
572
|
+
if(m_it && m_it.index()==m_innerIndex) break;
|
573
|
+
}
|
574
|
+
return *this;
|
575
|
+
}
|
576
|
+
|
577
|
+
inline operator bool() const { return m_outerPos < m_end; }
|
578
|
+
};
|
579
|
+
|
580
|
+
template<typename _Scalar, int _Options, typename _StorageIndex, int BlockRows, int BlockCols>
|
581
|
+
struct unary_evaluator<Block<SparseMatrix<_Scalar, _Options, _StorageIndex>,BlockRows,BlockCols,true>, IteratorBased>
|
582
|
+
: evaluator<SparseCompressedBase<Block<SparseMatrix<_Scalar, _Options, _StorageIndex>,BlockRows,BlockCols,true> > >
|
583
|
+
{
|
584
|
+
typedef Block<SparseMatrix<_Scalar, _Options, _StorageIndex>,BlockRows,BlockCols,true> XprType;
|
585
|
+
typedef evaluator<SparseCompressedBase<XprType> > Base;
|
586
|
+
explicit unary_evaluator(const XprType &xpr) : Base(xpr) {}
|
587
|
+
};
|
588
|
+
|
589
|
+
template<typename _Scalar, int _Options, typename _StorageIndex, int BlockRows, int BlockCols>
|
590
|
+
struct unary_evaluator<Block<const SparseMatrix<_Scalar, _Options, _StorageIndex>,BlockRows,BlockCols,true>, IteratorBased>
|
591
|
+
: evaluator<SparseCompressedBase<Block<const SparseMatrix<_Scalar, _Options, _StorageIndex>,BlockRows,BlockCols,true> > >
|
592
|
+
{
|
593
|
+
typedef Block<const SparseMatrix<_Scalar, _Options, _StorageIndex>,BlockRows,BlockCols,true> XprType;
|
594
|
+
typedef evaluator<SparseCompressedBase<XprType> > Base;
|
595
|
+
explicit unary_evaluator(const XprType &xpr) : Base(xpr) {}
|
596
|
+
};
|
597
|
+
|
598
|
+
} // end namespace internal
|
599
|
+
|
600
|
+
|
601
|
+
} // end namespace Eigen
|
602
|
+
|
603
|
+
#endif // EIGEN_SPARSE_BLOCK_H
|