Please, help us to better know about our user community by answering the following short survey: https://forms.gle/wpyrxWi18ox9Z5ae9
Eigen  3.3.9
SparseAssign.h
1 // This file is part of Eigen, a lightweight C++ template library
2 // for linear algebra.
3 //
4 // Copyright (C) 2008-2014 Gael Guennebaud <gael.guennebaud@inria.fr>
5 //
6 // This Source Code Form is subject to the terms of the Mozilla
7 // Public License v. 2.0. If a copy of the MPL was not distributed
8 // with this file, You can obtain one at http://mozilla.org/MPL/2.0/.
9 
10 #ifndef EIGEN_SPARSEASSIGN_H
11 #define EIGEN_SPARSEASSIGN_H
12 
13 namespace Eigen {
14 
15 template<typename Derived>
16 template<typename OtherDerived>
17 Derived& SparseMatrixBase<Derived>::operator=(const EigenBase<OtherDerived> &other)
18 {
19  internal::call_assignment_no_alias(derived(), other.derived());
20  return derived();
21 }
22 
23 template<typename Derived>
24 template<typename OtherDerived>
25 Derived& SparseMatrixBase<Derived>::operator=(const ReturnByValue<OtherDerived>& other)
26 {
27  // TODO use the evaluator mechanism
28  other.evalTo(derived());
29  return derived();
30 }
31 
32 template<typename Derived>
33 template<typename OtherDerived>
34 inline Derived& SparseMatrixBase<Derived>::operator=(const SparseMatrixBase<OtherDerived>& other)
35 {
36  // by default sparse evaluation do not alias, so we can safely bypass the generic call_assignment routine
37  internal::Assignment<Derived,OtherDerived,internal::assign_op<Scalar,typename OtherDerived::Scalar> >
38  ::run(derived(), other.derived(), internal::assign_op<Scalar,typename OtherDerived::Scalar>());
39  return derived();
40 }
41 
42 template<typename Derived>
43 inline Derived& SparseMatrixBase<Derived>::operator=(const Derived& other)
44 {
45  internal::call_assignment_no_alias(derived(), other.derived());
46  return derived();
47 }
48 
49 namespace internal {
50 
51 template<>
52 struct storage_kind_to_evaluator_kind<Sparse> {
53  typedef IteratorBased Kind;
54 };
55 
56 template<>
57 struct storage_kind_to_shape<Sparse> {
58  typedef SparseShape Shape;
59 };
60 
61 struct Sparse2Sparse {};
62 struct Sparse2Dense {};
63 
64 template<> struct AssignmentKind<SparseShape, SparseShape> { typedef Sparse2Sparse Kind; };
65 template<> struct AssignmentKind<SparseShape, SparseTriangularShape> { typedef Sparse2Sparse Kind; };
66 template<> struct AssignmentKind<DenseShape, SparseShape> { typedef Sparse2Dense Kind; };
67 template<> struct AssignmentKind<DenseShape, SparseTriangularShape> { typedef Sparse2Dense Kind; };
68 
69 
70 template<typename DstXprType, typename SrcXprType>
71 void assign_sparse_to_sparse(DstXprType &dst, const SrcXprType &src)
72 {
73  typedef typename DstXprType::Scalar Scalar;
74  typedef internal::evaluator<DstXprType> DstEvaluatorType;
75  typedef internal::evaluator<SrcXprType> SrcEvaluatorType;
76 
77  SrcEvaluatorType srcEvaluator(src);
78 
79  const bool transpose = (DstEvaluatorType::Flags & RowMajorBit) != (SrcEvaluatorType::Flags & RowMajorBit);
80  const Index outerEvaluationSize = (SrcEvaluatorType::Flags&RowMajorBit) ? src.rows() : src.cols();
81  if ((!transpose) && src.isRValue())
82  {
83  // eval without temporary
84  dst.resize(src.rows(), src.cols());
85  dst.setZero();
86  dst.reserve((std::max)(src.rows(),src.cols())*2);
87  for (Index j=0; j<outerEvaluationSize; ++j)
88  {
89  dst.startVec(j);
90  for (typename SrcEvaluatorType::InnerIterator it(srcEvaluator, j); it; ++it)
91  {
92  Scalar v = it.value();
93  dst.insertBackByOuterInner(j,it.index()) = v;
94  }
95  }
96  dst.finalize();
97  }
98  else
99  {
100  // eval through a temporary
101  eigen_assert(( ((internal::traits<DstXprType>::SupportedAccessPatterns & OuterRandomAccessPattern)==OuterRandomAccessPattern) ||
102  (!((DstEvaluatorType::Flags & RowMajorBit) != (SrcEvaluatorType::Flags & RowMajorBit)))) &&
103  "the transpose operation is supposed to be handled in SparseMatrix::operator=");
104 
105  enum { Flip = (DstEvaluatorType::Flags & RowMajorBit) != (SrcEvaluatorType::Flags & RowMajorBit) };
106 
107 
108  DstXprType temp(src.rows(), src.cols());
109 
110  temp.reserve((std::max)(src.rows(),src.cols())*2);
111  for (Index j=0; j<outerEvaluationSize; ++j)
112  {
113  temp.startVec(j);
114  for (typename SrcEvaluatorType::InnerIterator it(srcEvaluator, j); it; ++it)
115  {
116  Scalar v = it.value();
117  temp.insertBackByOuterInner(Flip?it.index():j,Flip?j:it.index()) = v;
118  }
119  }
120  temp.finalize();
121 
122  dst = temp.markAsRValue();
123  }
124 }
125 
126 // Generic Sparse to Sparse assignment
127 template< typename DstXprType, typename SrcXprType, typename Functor>
128 struct Assignment<DstXprType, SrcXprType, Functor, Sparse2Sparse>
129 {
130  static void run(DstXprType &dst, const SrcXprType &src, const internal::assign_op<typename DstXprType::Scalar,typename SrcXprType::Scalar> &/*func*/)
131  {
132  assign_sparse_to_sparse(dst.derived(), src.derived());
133  }
134 };
135 
136 // Generic Sparse to Dense assignment
137 template< typename DstXprType, typename SrcXprType, typename Functor>
138 struct Assignment<DstXprType, SrcXprType, Functor, Sparse2Dense>
139 {
140  static void run(DstXprType &dst, const SrcXprType &src, const Functor &func)
141  {
142  if(internal::is_same<Functor,internal::assign_op<typename DstXprType::Scalar,typename SrcXprType::Scalar> >::value)
143  dst.setZero();
144 
145  internal::evaluator<SrcXprType> srcEval(src);
146  resize_if_allowed(dst, src, func);
147  internal::evaluator<DstXprType> dstEval(dst);
148 
149  const Index outerEvaluationSize = (internal::evaluator<SrcXprType>::Flags&RowMajorBit) ? src.rows() : src.cols();
150  for (Index j=0; j<outerEvaluationSize; ++j)
151  for (typename internal::evaluator<SrcXprType>::InnerIterator i(srcEval,j); i; ++i)
152  func.assignCoeff(dstEval.coeffRef(i.row(),i.col()), i.value());
153  }
154 };
155 
156 // Specialization for "dst = dec.solve(rhs)"
157 // NOTE we need to specialize it for Sparse2Sparse to avoid ambiguous specialization error
158 template<typename DstXprType, typename DecType, typename RhsType, typename Scalar>
159 struct Assignment<DstXprType, Solve<DecType,RhsType>, internal::assign_op<Scalar,Scalar>, Sparse2Sparse>
160 {
161  typedef Solve<DecType,RhsType> SrcXprType;
162  static void run(DstXprType &dst, const SrcXprType &src, const internal::assign_op<Scalar,Scalar> &)
163  {
164  Index dstRows = src.rows();
165  Index dstCols = src.cols();
166  if((dst.rows()!=dstRows) || (dst.cols()!=dstCols))
167  dst.resize(dstRows, dstCols);
168 
169  src.dec()._solve_impl(src.rhs(), dst);
170  }
171 };
172 
173 struct Diagonal2Sparse {};
174 
175 template<> struct AssignmentKind<SparseShape,DiagonalShape> { typedef Diagonal2Sparse Kind; };
176 
177 template< typename DstXprType, typename SrcXprType, typename Functor>
178 struct Assignment<DstXprType, SrcXprType, Functor, Diagonal2Sparse>
179 {
180  typedef typename DstXprType::StorageIndex StorageIndex;
181  typedef typename DstXprType::Scalar Scalar;
182  typedef Array<StorageIndex,Dynamic,1> ArrayXI;
183  typedef Array<Scalar,Dynamic,1> ArrayXS;
184  template<int Options>
185  static void run(SparseMatrix<Scalar,Options,StorageIndex> &dst, const SrcXprType &src, const internal::assign_op<typename DstXprType::Scalar,typename SrcXprType::Scalar> &/*func*/)
186  {
187  Index dstRows = src.rows();
188  Index dstCols = src.cols();
189  if((dst.rows()!=dstRows) || (dst.cols()!=dstCols))
190  dst.resize(dstRows, dstCols);
191 
192  Index size = src.diagonal().size();
193  dst.makeCompressed();
194  dst.resizeNonZeros(size);
195  Map<ArrayXI>(dst.innerIndexPtr(), size).setLinSpaced(0,StorageIndex(size)-1);
196  Map<ArrayXI>(dst.outerIndexPtr(), size+1).setLinSpaced(0,StorageIndex(size));
197  Map<ArrayXS>(dst.valuePtr(), size) = src.diagonal();
198  }
199 
200  template<typename DstDerived>
201  static void run(SparseMatrixBase<DstDerived> &dst, const SrcXprType &src, const internal::assign_op<typename DstXprType::Scalar,typename SrcXprType::Scalar> &/*func*/)
202  {
203  dst.diagonal() = src.diagonal();
204  }
205 
206  static void run(DstXprType &dst, const SrcXprType &src, const internal::add_assign_op<typename DstXprType::Scalar,typename SrcXprType::Scalar> &/*func*/)
207  { dst.diagonal() += src.diagonal(); }
208 
209  static void run(DstXprType &dst, const SrcXprType &src, const internal::sub_assign_op<typename DstXprType::Scalar,typename SrcXprType::Scalar> &/*func*/)
210  { dst.diagonal() -= src.diagonal(); }
211 };
212 } // end namespace internal
213 
214 } // end namespace Eigen
215 
216 #endif // EIGEN_SPARSEASSIGN_H
Eigen
Namespace containing all symbols from the Eigen library.
Definition: Core:309
Eigen::RowMajorBit
const unsigned int RowMajorBit
Definition: Constants.h:61
Eigen::Index
EIGEN_DEFAULT_DENSE_INDEX_TYPE Index
The Index type as used for the API.
Definition: Meta.h:33