import pandas as pd import numpy as np import matplotlib.pyplot as plt import seaborn as sns from sklearn.decomposition import PCA from sklearn.manifold import TSNE from sklearn.feature_selection import SelectKBest, f_regression, RFE from sklearn.ensemble import RandomForestRegressor from sklearn.cluster import KMeans from sklearn.preprocessing import StandardScaler from scipy.stats import pearsonr, spearmanr import warnings warnings.filterwarnings('ignore') # 设置中文字体 plt.rcParams['font.sans-serif'] = ['SimHei', 'DejaVu Sans'] plt.rcParams['axes.unicode_minus'] = False # ===== Configuration ===== class Config: TRAIN_PATH = "/AI4M/users/mjzhang/workspace/DRW/new_data/train.parquet" TEST_PATH = "/AI4M/users/mjzhang/workspace/DRW/new_data/test.parquet" SUBMISSION_PATH = "/AI4M/users/mjzhang/workspace/DRW/new_data/sample_submission.csv" LABEL_COLUMN = "label" def load_data(): """加载数据""" print("正在加载数据...") train_df = pd.read_parquet(Config.TRAIN_PATH, columns=Config.FEATURES + [Config.LABEL_COLUMN]) print(f"数据加载完成,训练集形状: {train_df.shape}") return train_df def feature_engineering(df): """特征工程(与原始代码保持一致)""" # Original features df['volume_weighted_sell'] = df['sell_qty'] * df['volume'] df['buy_sell_ratio'] = df['buy_qty'] / (df['sell_qty'] + 1e-8) df['selling_pressure'] = df['sell_qty'] / (df['volume'] + 1e-8) df['effective_spread_proxy'] = np.abs(df['buy_qty'] - df['sell_qty']) / (df['volume'] + 1e-8) # New robust features df['log_volume'] = np.log1p(df['volume']) df['bid_ask_imbalance'] = (df['bid_qty'] - df['ask_qty']) / (df['bid_qty'] + df['ask_qty'] + 1e-8) df['order_flow_imbalance'] = (df['buy_qty'] - df['sell_qty']) / (df['buy_qty'] + df['sell_qty'] + 1e-8) df['liquidity_ratio'] = (df['bid_qty'] + df['ask_qty']) / (df['volume'] + 1e-8) # Handle infinities and NaN df = df.replace([np.inf, -np.inf], np.nan) # For each column, replace NaN with median for robustness for col in df.columns: if df[col].isna().any(): median_val = df[col].median() df[col] = df[col].fillna(median_val if not pd.isna(median_val) else 0) return df def analyze_feature_label_correlation(df, features, label_col): """分析特征与标签的相关性""" print("\n=== 特征与标签相关性分析 ===") correlations = [] for feature in features: if feature in df.columns: # Pearson相关系数 pearson_corr, pearson_p = pearsonr(df[feature], df[label_col]) # Spearman相关系数 spearman_corr, spearman_p = spearmanr(df[feature], df[label_col]) correlations.append({ 'feature': feature, 'pearson_corr': pearson_corr, 'pearson_p': pearson_p, 'spearman_corr': spearman_corr, 'spearman_p': spearman_p, 'abs_pearson': abs(pearson_corr), 'abs_spearman': abs(spearman_corr) }) # 转换为DataFrame并排序 corr_df = pd.DataFrame(correlations) corr_df = corr_df.sort_values('abs_pearson', ascending=False) print("\n特征与标签相关性排序(按Pearson相关系数绝对值):") print("=" * 80) print(f"{'特征名':<20} {'Pearson':<10} {'P值':<10} {'Spearman':<10} {'P值':<10}") print("=" * 80) for _, row in corr_df.iterrows(): print(f"{row['feature']:<20} {row['pearson_corr']:<10.4f} {row['pearson_p']:<10.4f} " f"{row['spearman_corr']:<10.4f} {row['spearman_p']:<10.4f}") return corr_df def analyze_feature_correlations(df, features, top_n=10): """分析特征间的相关性""" print(f"\n=== 特征间相关性分析(显示前{top_n}个最强相关) ===") # 计算相关性矩阵 feature_df = df[features] corr_matrix = feature_df.corr() # 获取上三角矩阵的相关性 upper_tri = corr_matrix.where(np.triu(np.ones(corr_matrix.shape), k=1).astype(bool)) # 转换为长格式并排序 correlations = [] for i in range(len(features)): for j in range(i+1, len(features)): feat1, feat2 = features[i], features[j] corr_val = corr_matrix.loc[feat1, feat2] if not pd.isna(corr_val): correlations.append({ 'feature1': feat1, 'feature2': feat2, 'correlation': corr_val, 'abs_correlation': abs(corr_val) }) corr_df = pd.DataFrame(correlations) corr_df = corr_df.sort_values('abs_correlation', ascending=False) print(f"\n特征间相关性排序(前{top_n}个):") print("=" * 60) print(f"{'特征1':<15} {'特征2':<15} {'相关性':<10} {'绝对值':<10}") print("=" * 60) for _, row in corr_df.head(top_n).iterrows(): print(f"{row['feature1']:<15} {row['feature2']:<15} {row['correlation']:<10.4f} {row['abs_correlation']:<10.4f}") return corr_matrix, corr_df def plot_correlation_heatmap(corr_matrix, title="特征相关性热力图", figsize=(12, 10)): """绘制相关性热力图""" plt.figure(figsize=figsize) mask = np.triu(np.ones_like(corr_matrix, dtype=bool)) sns.heatmap(corr_matrix, mask=mask, annot=True, cmap='RdBu_r', center=0, square=True, linewidths=0.5, cbar_kws={"shrink": .8}) plt.title(title, fontsize=16, pad=20) plt.tight_layout() plt.savefig(f"{title.replace(' ', '_')}.png", dpi=300, bbox_inches='tight') plt.show() def plot_feature_label_correlation(corr_df, title="特征与标签相关性", figsize=(12, 8)): """绘制特征与标签相关性条形图""" plt.figure(figsize=figsize) # 选择前20个特征 top_features = corr_df.head(20) x = range(len(top_features)) pearson_vals = top_features['pearson_corr'].values spearman_vals = top_features['spearman_corr'].values # 创建条形图 width = 0.35 plt.bar([i - width/2 for i in x], pearson_vals, width, label='Pearson', alpha=0.8) plt.bar([i + width/2 for i in x], spearman_vals, width, label='Spearman', alpha=0.8) plt.xlabel('特征') plt.ylabel('相关系数') plt.title(title) plt.xticks(x, top_features['feature'], rotation=45, ha='right') plt.legend() plt.grid(True, alpha=0.3) plt.tight_layout() plt.savefig(f"{title.replace(' ', '_')}.png", dpi=300, bbox_inches='tight') plt.show() def analyze_feature_importance_by_correlation(corr_df, threshold=0.01): """基于相关性分析特征重要性""" print(f"\n=== 基于相关性的特征重要性分析(P值阈值: {threshold}) ===") # 筛选显著相关的特征 significant_features = corr_df[corr_df['pearson_p'] < threshold].copy() significant_features = significant_features.sort_values('abs_pearson', ascending=False) print(f"\n显著相关的特征数量: {len(significant_features)}") print("\n高重要性特征(|相关系数| > 0.05):") high_importance = significant_features[significant_features['abs_pearson'] > 0.05] for _, row in high_importance.iterrows(): print(f"- {row['feature']}: Pearson={row['pearson_corr']:.4f}, P={row['pearson_p']:.4f}") return significant_features def find_high_correlations(corr_matrix, threshold=0.8): high_corr_pairs = [] for i in range(len(corr_matrix.columns)): for j in range(i + 1, len(corr_matrix.columns)): if abs(corr_matrix.iloc[i, j]) > threshold: high_corr_pairs += [{ 'feature1': corr_matrix.columns[i], 'feature2': corr_matrix.columns[j], 'correlation': corr_matrix.iloc[i, j] }] return pd.DataFrame(high_corr_pairs) def main(): """主函数""" print("开始特征相关性分析...") # 加载数据 train = pd.read_parquet(Config.TRAIN_PATH) anonymized_features = list(train.columns[list(range(5, train.shape[1] - 1))]) target = 'label' fig, axes = plt.subplots(2, 3, figsize=(15, 10)) fig.suptitle('Distribution of Random 6 Anonymized Features', fontsize=16) import random for i, feature in enumerate(random.sample(list(anonymized_features), 6)): row, col = i // 3, i % 3 train[feature].hist(bins=50, ax=axes[row, col], alpha=0.7) axes[row, col].set_title(f'{feature}') axes[row, col].set_xlabel('Value') axes[row, col].set_ylabel('Frequency') plt.tight_layout() plt.show() breakpoint() # heatmap correlation between features subset_features = anonymized_features[:50] correlation_matrix = train[subset_features + [target]].corr() # Plot correlation heatmap plt.figure(figsize=(32, 20)) sns.heatmap(correlation_matrix, cmap='coolwarm', center=0, square=True, fmt='.2f',annot=True, cbar_kws={'shrink': 0.8}) plt.title('Correlation Matrix - First 50 Anonymized Features + Target') plt.tight_layout() plt.show() breakpoint() high_corr_df = find_high_correlations(correlation_matrix, threshold=0.98) print(f"High correlation pairs (|corr| > 0.98):") print(high_corr_df.sort_values('correlation', key=abs, ascending=False).head(50)) # Correlation with target variable target_correlations = train[anonymized_features + [target]].corr()[target].abs().sort_values(ascending=False) print(f"\nTop 20 features most correlated with target:") print(target_correlations.head(51)[1:]) # Exclude target itself breakpoint() from sklearn.decomposition import IncrementalPCA X_numeric = train[anonymized_features].select_dtypes(include=[np.number]).iloc[:, :200] # 2. Fill missing values with column median X_clean = X_numeric.fillna(X_numeric.median()) # 3. Downcast to float32 to save memory X_clean = X_clean.astype(np.float32) # 4. Standardize features scaler = StandardScaler() X_scaled = scaler.fit_transform(X_clean) # 5. Perform Incremental PCA print("Performing Incremental PCA...") ipca = IncrementalPCA(n_components=50, batch_size=200) X_pca = ipca.fit_transform(X_scaled) # 6. Explained variance analysis explained_variance_ratio = ipca.explained_variance_ratio_ cumsum_variance = np.cumsum(explained_variance_ratio) n_components_90 = np.argmax(cumsum_variance >= 0.90) + 1 n_components_95 = np.argmax(cumsum_variance >= 0.95) + 1 print(f"Number of components needed for 90% variance: {n_components_90}") print(f"Number of components needed for 95% variance: {n_components_95}") breakpoint() # t-SNE visualization (on a sample for computational efficiency) print("Performing t-SNE on sample data...") sample_size = min(5000, len(X_clean)) sample_indices = np.random.choice(len(X_clean), sample_size, replace=False) X_sample = X_scaled[sample_indices] y = train[target] # <-- Replace 'target' with your actual label column name y_sample = y.iloc[sample_indices] # Use PCA first to reduce dimensions before t-SNE pca_pre = PCA(n_components=50) X_pca_sample = pca_pre.fit_transform(X_sample) tsne = TSNE(n_components=2, random_state=42, perplexity=30) X_tsne = tsne.fit_transform(X_pca_sample) breakpoint() from sklearn.feature_selection import SelectKBest, f_regression # 使用单变量的 f回归判断,其实就是 pearson排序 print("\n" + "="*50) print("4. FEATURE SELECTION") print("="*50) # Univariate feature selection print("Performing univariate feature selection...") selector_univariate = SelectKBest(score_func=f_regression, k=100) X_selected_univariate = selector_univariate.fit_transform(X_clean, y) # Get feature scores anonymized_features = X_clean.columns # <-- Make sure this matches the full feature list feature_scores = pd.DataFrame({ 'feature': anonymized_features, 'score': selector_univariate.scores_ }).sort_values('score', ascending=False) print("Top 20 features by univariate selection:") print(feature_scores.head(20)) breakpoint() from sklearn.ensemble import RandomForestRegressor print("\nTraining Random Forest for feature importance (with speed-up)...") rf = RandomForestRegressor( n_estimators=100, # Reduce trees from 100 → 50 for faster training max_depth=10, # Limit depth to prevent very deep trees max_features='sqrt', # Use sqrt of features at each split (default) n_jobs=-1, # Use all CPU cores random_state=42 ) y = train[target] rf.fit(X_clean, y) feature_importance = pd.DataFrame({ 'feature': anonymized_features, 'importance': rf.feature_importances_ }).sort_values('importance', ascending=False) print("Top 20 features by Random Forest importance:") print(feature_importance.head(20)) breakpoint() from sklearn.linear_model import Lasso, LassoCV from sklearn.feature_selection import RFE # lasso + rfe 相比于 f回归能考虑特征之间的相关性 # lasso alpha 0.01比较小,应该是设置越大,惩罚越大,回归系数压缩到 0的特征越多,但是如果没有,rfe也会迭代删除最弱的特征 # lasso 的系数排序不等于 rfe的排序。lasso有时对弱相关特征选择不稳定,rfe 更鲁棒。ref是反向过程,记录是最后被保留的排名,不是当前有多大的系数 print("\nPerforming Recursive Feature Elimination with Lasso...") # Initialize Lasso with a small alpha (regularization strength) and enough iterations lasso_cv = LassoCV(cv=5, max_iter=10000, random_state=42) lasso_cv.fit(X_clean, y) print(f"Best alpha via CV: {lasso_cv.alpha_}") lasso = Lasso(alpha=lasso_cv.alpha_, max_iter=10000, random_state=42) # RFE with step=50 (removes 50 features at a time) to speed up rfe = RFE(estimator=lasso, n_features_to_select=100, step=50) # Fit RFE on your data rfe.fit(X_clean, y) # Collect feature selection results rfe_features = pd.DataFrame({ 'feature': anonymized_features, 'selected': rfe.support_, 'ranking': rfe.ranking_ }).sort_values('ranking') selected_features_rfe = rfe_features[rfe_features['selected']]['feature'].tolist() print(f"RFE selected {len(selected_features_rfe)} features") print("Top 20 RFE selected features:") print(rfe_features.head(20)) breakpoint() # K-means clustering on PCA-reduced data print("Performing K-means clustering...") n_clusters_range = range(2, 11) inertias = [] for k in n_clusters_range: kmeans = KMeans(n_clusters=k, random_state=42) kmeans.fit(X_pca[:, :50]) # Use first 50 PCA components inertias.append(kmeans.inertia_) # Apply optimal clustering optimal_k = 5 # You can adjust based on elbow curve kmeans_final = KMeans(n_clusters=optimal_k, random_state=42) clusters = kmeans_final.fit_predict(X_pca[:, :50]) # Analyze clusters cluster_analysis = pd.DataFrame({ 'cluster': clusters, 'target': y }) print(f"\nCluster analysis with {optimal_k} clusters:") cluster_stats = cluster_analysis.groupby('cluster')['target'].agg(['count', 'mean', 'std']) print(cluster_stats) print(f"1. Dataset contains {890} anonymized features") print(f"2. {n_components_90} components explain 90% of variance, {n_components_95} explain 95%") print(f"3. Top correlated feature with target: {target_correlations.index[1]} (corr: {target_correlations.iloc[1]:.4f})") print(f"4. {len(high_corr_df)} feature pairs have high correlation (>0.7)") print(f"5. Optimal number of clusters appears to be around {optimal_k}") # Save important features for further analysis important_features = { 'top_univariate': feature_scores.head(50)['feature'].tolist(), 'top_rf_importance': feature_importance.head(50)['feature'].tolist(), 'rfe_selected': selected_features_rfe, 'high_target_corr': target_correlations.head(50).index[1:].tolist() } common_features = set(important_features['top_univariate']) & \ set(important_features['top_rf_importance']) & \ set(important_features['rfe_selected']) print(f" Common Features selected by all 3 methods: {(common_features)}") if __name__ == "__main__": main()