| | import pandas as pd |
| | import numpy as np |
| | import matplotlib.pyplot as plt |
| | import seaborn as sns |
| | from sklearn.decomposition import PCA |
| | from sklearn.manifold import TSNE |
| | from sklearn.feature_selection import SelectKBest, f_regression, RFE |
| | from sklearn.ensemble import RandomForestRegressor |
| | from sklearn.cluster import KMeans |
| | from sklearn.preprocessing import StandardScaler |
| | from scipy.stats import pearsonr, spearmanr |
| | import warnings |
| | warnings.filterwarnings('ignore') |
| |
|
| | |
| | plt.rcParams['font.sans-serif'] = ['SimHei', 'DejaVu Sans'] |
| | plt.rcParams['axes.unicode_minus'] = False |
| |
|
| | |
| | class Config: |
| | TRAIN_PATH = "/AI4M/users/mjzhang/workspace/DRW/new_data/train.parquet" |
| | TEST_PATH = "/AI4M/users/mjzhang/workspace/DRW/new_data/test.parquet" |
| | SUBMISSION_PATH = "/AI4M/users/mjzhang/workspace/DRW/new_data/sample_submission.csv" |
| | |
| | LABEL_COLUMN = "label" |
| |
|
| | def load_data(): |
| | """加载数据""" |
| | print("正在加载数据...") |
| | train_df = pd.read_parquet(Config.TRAIN_PATH, columns=Config.FEATURES + [Config.LABEL_COLUMN]) |
| | print(f"数据加载完成,训练集形状: {train_df.shape}") |
| | return train_df |
| |
|
| | def feature_engineering(df): |
| | """特征工程(与原始代码保持一致)""" |
| | |
| | df['volume_weighted_sell'] = df['sell_qty'] * df['volume'] |
| | df['buy_sell_ratio'] = df['buy_qty'] / (df['sell_qty'] + 1e-8) |
| | df['selling_pressure'] = df['sell_qty'] / (df['volume'] + 1e-8) |
| | df['effective_spread_proxy'] = np.abs(df['buy_qty'] - df['sell_qty']) / (df['volume'] + 1e-8) |
| | |
| | |
| | df['log_volume'] = np.log1p(df['volume']) |
| | df['bid_ask_imbalance'] = (df['bid_qty'] - df['ask_qty']) / (df['bid_qty'] + df['ask_qty'] + 1e-8) |
| | df['order_flow_imbalance'] = (df['buy_qty'] - df['sell_qty']) / (df['buy_qty'] + df['sell_qty'] + 1e-8) |
| | df['liquidity_ratio'] = (df['bid_qty'] + df['ask_qty']) / (df['volume'] + 1e-8) |
| | |
| | |
| | df = df.replace([np.inf, -np.inf], np.nan) |
| | |
| | |
| | for col in df.columns: |
| | if df[col].isna().any(): |
| | median_val = df[col].median() |
| | df[col] = df[col].fillna(median_val if not pd.isna(median_val) else 0) |
| | |
| | return df |
| |
|
| | def analyze_feature_label_correlation(df, features, label_col): |
| | """分析特征与标签的相关性""" |
| | print("\n=== 特征与标签相关性分析 ===") |
| | |
| | correlations = [] |
| | for feature in features: |
| | if feature in df.columns: |
| | |
| | pearson_corr, pearson_p = pearsonr(df[feature], df[label_col]) |
| | |
| | spearman_corr, spearman_p = spearmanr(df[feature], df[label_col]) |
| | |
| | correlations.append({ |
| | 'feature': feature, |
| | 'pearson_corr': pearson_corr, |
| | 'pearson_p': pearson_p, |
| | 'spearman_corr': spearman_corr, |
| | 'spearman_p': spearman_p, |
| | 'abs_pearson': abs(pearson_corr), |
| | 'abs_spearman': abs(spearman_corr) |
| | }) |
| | |
| | |
| | corr_df = pd.DataFrame(correlations) |
| | corr_df = corr_df.sort_values('abs_pearson', ascending=False) |
| | |
| | print("\n特征与标签相关性排序(按Pearson相关系数绝对值):") |
| | print("=" * 80) |
| | print(f"{'特征名':<20} {'Pearson':<10} {'P值':<10} {'Spearman':<10} {'P值':<10}") |
| | print("=" * 80) |
| | |
| | for _, row in corr_df.iterrows(): |
| | print(f"{row['feature']:<20} {row['pearson_corr']:<10.4f} {row['pearson_p']:<10.4f} " |
| | f"{row['spearman_corr']:<10.4f} {row['spearman_p']:<10.4f}") |
| | |
| | return corr_df |
| |
|
| | def analyze_feature_correlations(df, features, top_n=10): |
| | """分析特征间的相关性""" |
| | print(f"\n=== 特征间相关性分析(显示前{top_n}个最强相关) ===") |
| | |
| | |
| | feature_df = df[features] |
| | corr_matrix = feature_df.corr() |
| | |
| | |
| | upper_tri = corr_matrix.where(np.triu(np.ones(corr_matrix.shape), k=1).astype(bool)) |
| | |
| | |
| | correlations = [] |
| | for i in range(len(features)): |
| | for j in range(i+1, len(features)): |
| | feat1, feat2 = features[i], features[j] |
| | corr_val = corr_matrix.loc[feat1, feat2] |
| | if not pd.isna(corr_val): |
| | correlations.append({ |
| | 'feature1': feat1, |
| | 'feature2': feat2, |
| | 'correlation': corr_val, |
| | 'abs_correlation': abs(corr_val) |
| | }) |
| | |
| | corr_df = pd.DataFrame(correlations) |
| | corr_df = corr_df.sort_values('abs_correlation', ascending=False) |
| | |
| | print(f"\n特征间相关性排序(前{top_n}个):") |
| | print("=" * 60) |
| | print(f"{'特征1':<15} {'特征2':<15} {'相关性':<10} {'绝对值':<10}") |
| | print("=" * 60) |
| | |
| | for _, row in corr_df.head(top_n).iterrows(): |
| | print(f"{row['feature1']:<15} {row['feature2']:<15} {row['correlation']:<10.4f} {row['abs_correlation']:<10.4f}") |
| | |
| | return corr_matrix, corr_df |
| |
|
| | def plot_correlation_heatmap(corr_matrix, title="特征相关性热力图", figsize=(12, 10)): |
| | """绘制相关性热力图""" |
| | plt.figure(figsize=figsize) |
| | mask = np.triu(np.ones_like(corr_matrix, dtype=bool)) |
| | sns.heatmap(corr_matrix, mask=mask, annot=True, cmap='RdBu_r', center=0, |
| | square=True, linewidths=0.5, cbar_kws={"shrink": .8}) |
| | plt.title(title, fontsize=16, pad=20) |
| | plt.tight_layout() |
| | plt.savefig(f"{title.replace(' ', '_')}.png", dpi=300, bbox_inches='tight') |
| | plt.show() |
| |
|
| | def plot_feature_label_correlation(corr_df, title="特征与标签相关性", figsize=(12, 8)): |
| | """绘制特征与标签相关性条形图""" |
| | plt.figure(figsize=figsize) |
| | |
| | |
| | top_features = corr_df.head(20) |
| | |
| | x = range(len(top_features)) |
| | pearson_vals = top_features['pearson_corr'].values |
| | spearman_vals = top_features['spearman_corr'].values |
| | |
| | |
| | width = 0.35 |
| | plt.bar([i - width/2 for i in x], pearson_vals, width, label='Pearson', alpha=0.8) |
| | plt.bar([i + width/2 for i in x], spearman_vals, width, label='Spearman', alpha=0.8) |
| | |
| | plt.xlabel('特征') |
| | plt.ylabel('相关系数') |
| | plt.title(title) |
| | plt.xticks(x, top_features['feature'], rotation=45, ha='right') |
| | plt.legend() |
| | plt.grid(True, alpha=0.3) |
| | plt.tight_layout() |
| | plt.savefig(f"{title.replace(' ', '_')}.png", dpi=300, bbox_inches='tight') |
| | plt.show() |
| |
|
| | def analyze_feature_importance_by_correlation(corr_df, threshold=0.01): |
| | """基于相关性分析特征重要性""" |
| | print(f"\n=== 基于相关性的特征重要性分析(P值阈值: {threshold}) ===") |
| | |
| | |
| | significant_features = corr_df[corr_df['pearson_p'] < threshold].copy() |
| | significant_features = significant_features.sort_values('abs_pearson', ascending=False) |
| | |
| | print(f"\n显著相关的特征数量: {len(significant_features)}") |
| | print("\n高重要性特征(|相关系数| > 0.05):") |
| | high_importance = significant_features[significant_features['abs_pearson'] > 0.05] |
| | |
| | for _, row in high_importance.iterrows(): |
| | print(f"- {row['feature']}: Pearson={row['pearson_corr']:.4f}, P={row['pearson_p']:.4f}") |
| | |
| | return significant_features |
| |
|
| | def find_high_correlations(corr_matrix, threshold=0.8): |
| | high_corr_pairs = [] |
| | for i in range(len(corr_matrix.columns)): |
| | for j in range(i + 1, len(corr_matrix.columns)): |
| | if abs(corr_matrix.iloc[i, j]) > threshold: |
| | high_corr_pairs += [{ |
| | 'feature1': corr_matrix.columns[i], |
| | 'feature2': corr_matrix.columns[j], |
| | 'correlation': corr_matrix.iloc[i, j] |
| | }] |
| | return pd.DataFrame(high_corr_pairs) |
| |
|
| |
|
| |
|
| | def main(): |
| | """主函数""" |
| | print("开始特征相关性分析...") |
| | |
| | |
| | train = pd.read_parquet(Config.TRAIN_PATH) |
| | anonymized_features = list(train.columns[list(range(5, train.shape[1] - 1))]) |
| | target = 'label' |
| | fig, axes = plt.subplots(2, 3, figsize=(15, 10)) |
| | fig.suptitle('Distribution of Random 6 Anonymized Features', fontsize=16) |
| | import random |
| | for i, feature in enumerate(random.sample(list(anonymized_features), 6)): |
| | row, col = i // 3, i % 3 |
| | train[feature].hist(bins=50, ax=axes[row, col], alpha=0.7) |
| | axes[row, col].set_title(f'{feature}') |
| | axes[row, col].set_xlabel('Value') |
| | axes[row, col].set_ylabel('Frequency') |
| | plt.tight_layout() |
| | plt.show() |
| | breakpoint() |
| | |
| | |
| | subset_features = anonymized_features[:50] |
| | correlation_matrix = train[subset_features + [target]].corr() |
| |
|
| | |
| | plt.figure(figsize=(32, 20)) |
| | sns.heatmap(correlation_matrix, cmap='coolwarm', center=0, |
| | square=True, fmt='.2f',annot=True, cbar_kws={'shrink': 0.8}) |
| | plt.title('Correlation Matrix - First 50 Anonymized Features + Target') |
| | plt.tight_layout() |
| | plt.show() |
| | breakpoint() |
| | |
| | high_corr_df = find_high_correlations(correlation_matrix, threshold=0.98) |
| | print(f"High correlation pairs (|corr| > 0.98):") |
| | print(high_corr_df.sort_values('correlation', key=abs, ascending=False).head(50)) |
| |
|
| | |
| | target_correlations = train[anonymized_features + [target]].corr()[target].abs().sort_values(ascending=False) |
| | print(f"\nTop 20 features most correlated with target:") |
| | print(target_correlations.head(51)[1:]) |
| | breakpoint() |
| | |
| | from sklearn.decomposition import IncrementalPCA |
| | X_numeric = train[anonymized_features].select_dtypes(include=[np.number]).iloc[:, :200] |
| | |
| | X_clean = X_numeric.fillna(X_numeric.median()) |
| | |
| | X_clean = X_clean.astype(np.float32) |
| | |
| | scaler = StandardScaler() |
| | X_scaled = scaler.fit_transform(X_clean) |
| |
|
| | |
| | print("Performing Incremental PCA...") |
| | ipca = IncrementalPCA(n_components=50, batch_size=200) |
| | X_pca = ipca.fit_transform(X_scaled) |
| |
|
| | |
| | explained_variance_ratio = ipca.explained_variance_ratio_ |
| | cumsum_variance = np.cumsum(explained_variance_ratio) |
| | n_components_90 = np.argmax(cumsum_variance >= 0.90) + 1 |
| | n_components_95 = np.argmax(cumsum_variance >= 0.95) + 1 |
| |
|
| | print(f"Number of components needed for 90% variance: {n_components_90}") |
| | print(f"Number of components needed for 95% variance: {n_components_95}") |
| | breakpoint() |
| | |
| | |
| | print("Performing t-SNE on sample data...") |
| | sample_size = min(5000, len(X_clean)) |
| | sample_indices = np.random.choice(len(X_clean), sample_size, replace=False) |
| | X_sample = X_scaled[sample_indices] |
| |
|
| | y = train[target] |
| | y_sample = y.iloc[sample_indices] |
| |
|
| | |
| | pca_pre = PCA(n_components=50) |
| | X_pca_sample = pca_pre.fit_transform(X_sample) |
| |
|
| | tsne = TSNE(n_components=2, random_state=42, perplexity=30) |
| | X_tsne = tsne.fit_transform(X_pca_sample) |
| | |
| | |
| | breakpoint() |
| | from sklearn.feature_selection import SelectKBest, f_regression |
| | |
| |
|
| | print("\n" + "="*50) |
| | print("4. FEATURE SELECTION") |
| | print("="*50) |
| |
|
| | |
| | print("Performing univariate feature selection...") |
| | selector_univariate = SelectKBest(score_func=f_regression, k=100) |
| | X_selected_univariate = selector_univariate.fit_transform(X_clean, y) |
| |
|
| | |
| | anonymized_features = X_clean.columns |
| | feature_scores = pd.DataFrame({ |
| | 'feature': anonymized_features, |
| | 'score': selector_univariate.scores_ |
| | }).sort_values('score', ascending=False) |
| |
|
| | print("Top 20 features by univariate selection:") |
| | print(feature_scores.head(20)) |
| | |
| | breakpoint() |
| | |
| | from sklearn.ensemble import RandomForestRegressor |
| |
|
| | print("\nTraining Random Forest for feature importance (with speed-up)...") |
| |
|
| | rf = RandomForestRegressor( |
| | n_estimators=100, |
| | max_depth=10, |
| | max_features='sqrt', |
| | n_jobs=-1, |
| | random_state=42 |
| | ) |
| |
|
| | y = train[target] |
| | rf.fit(X_clean, y) |
| |
|
| | feature_importance = pd.DataFrame({ |
| | 'feature': anonymized_features, |
| | 'importance': rf.feature_importances_ |
| | }).sort_values('importance', ascending=False) |
| |
|
| | print("Top 20 features by Random Forest importance:") |
| | print(feature_importance.head(20)) |
| | breakpoint() |
| | |
| | from sklearn.linear_model import Lasso, LassoCV |
| | from sklearn.feature_selection import RFE |
| |
|
| | |
| | |
| | |
| |
|
| | print("\nPerforming Recursive Feature Elimination with Lasso...") |
| |
|
| | |
| | lasso_cv = LassoCV(cv=5, max_iter=10000, random_state=42) |
| | lasso_cv.fit(X_clean, y) |
| | print(f"Best alpha via CV: {lasso_cv.alpha_}") |
| | lasso = Lasso(alpha=lasso_cv.alpha_, max_iter=10000, random_state=42) |
| |
|
| | |
| | rfe = RFE(estimator=lasso, n_features_to_select=100, step=50) |
| |
|
| | |
| | rfe.fit(X_clean, y) |
| |
|
| | |
| | rfe_features = pd.DataFrame({ |
| | 'feature': anonymized_features, |
| | 'selected': rfe.support_, |
| | 'ranking': rfe.ranking_ |
| | }).sort_values('ranking') |
| |
|
| | selected_features_rfe = rfe_features[rfe_features['selected']]['feature'].tolist() |
| | print(f"RFE selected {len(selected_features_rfe)} features") |
| | print("Top 20 RFE selected features:") |
| | print(rfe_features.head(20)) |
| | breakpoint() |
| | |
| | |
| | print("Performing K-means clustering...") |
| | n_clusters_range = range(2, 11) |
| | inertias = [] |
| |
|
| | for k in n_clusters_range: |
| | kmeans = KMeans(n_clusters=k, random_state=42) |
| | kmeans.fit(X_pca[:, :50]) |
| | inertias.append(kmeans.inertia_) |
| | |
| | |
| | optimal_k = 5 |
| | kmeans_final = KMeans(n_clusters=optimal_k, random_state=42) |
| | clusters = kmeans_final.fit_predict(X_pca[:, :50]) |
| |
|
| | |
| | cluster_analysis = pd.DataFrame({ |
| | 'cluster': clusters, |
| | 'target': y |
| | }) |
| |
|
| | print(f"\nCluster analysis with {optimal_k} clusters:") |
| | cluster_stats = cluster_analysis.groupby('cluster')['target'].agg(['count', 'mean', 'std']) |
| | print(cluster_stats) |
| | |
| | print(f"1. Dataset contains {890} anonymized features") |
| | print(f"2. {n_components_90} components explain 90% of variance, {n_components_95} explain 95%") |
| | print(f"3. Top correlated feature with target: {target_correlations.index[1]} (corr: {target_correlations.iloc[1]:.4f})") |
| | print(f"4. {len(high_corr_df)} feature pairs have high correlation (>0.7)") |
| | print(f"5. Optimal number of clusters appears to be around {optimal_k}") |
| |
|
| | |
| | important_features = { |
| | 'top_univariate': feature_scores.head(50)['feature'].tolist(), |
| | 'top_rf_importance': feature_importance.head(50)['feature'].tolist(), |
| | 'rfe_selected': selected_features_rfe, |
| | 'high_target_corr': target_correlations.head(50).index[1:].tolist() |
| | } |
| |
|
| | common_features = set(important_features['top_univariate']) & \ |
| | set(important_features['top_rf_importance']) & \ |
| | set(important_features['rfe_selected']) |
| | print(f" Common Features selected by all 3 methods: {(common_features)}") |
| |
|
| | if __name__ == "__main__": |
| | main() |