File size: 10,023 Bytes
c687548
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
import sys
import pandas as pd
import numpy as np
import torch
import torch.nn as nn
import torch.optim as optim
from sklearn.model_selection import KFold
from sklearn.preprocessing import RobustScaler
from scipy.stats import pearsonr
import warnings
import pickle
warnings.filterwarnings('ignore')

# ===== Feature Engineering =====
def feature_engineering(df):
    # 保持接口一致,实际特征工程已在feature.py完成
    return df

# ===== Configuration =====
class Config:
    TRAIN_PATH = "/AI4M/users/mjzhang/workspace/DRW/ZMJ/max_IC_mixed/train_aggregated.parquet"
    TEST_PATH = "/AI4M/users/mjzhang/workspace/DRW/ZMJ/max_IC_mixed/test_aggregated.parquet"
    SUBMISSION_PATH = "/AI4M/users/mjzhang/workspace/DRW/ZMJ/max_IC_mixed/sample_submission.csv"
    LABEL_COLUMN = "label"
    N_FOLDS = 5  # 增加交叉验证折数
    RANDOM_STATE = 42
    DEVICE = torch.device('cuda' if torch.cuda.is_available() else 'cpu')
    BATCH_SIZE = 256  # 增加批次大小
    EPOCHS = 100  # 增加训练轮数
    LEARNING_RATE = 5e-4  # 降低学习率
    WEIGHT_DECAY = 1e-3  # 增加L2正则化
    PATIENCE = 15  # 增加早停容忍轮数

# ===== MLP Model Definition =====
class MLP(nn.Module):
    def __init__(self, input_dim):
        super().__init__()
        self.net = nn.Sequential(
            nn.Linear(input_dim, 256),
            # nn.BatchNorm1d(512),
            nn.ReLU(),
            nn.Dropout(0.1),
            
            nn.Linear(256, 128),
            # nn.BatchNorm1d(256),
            nn.ReLU(),
            nn.Dropout(0.1),
            
            # nn.Linear(256, 128),
            # nn.BatchNorm1d(128),
            # nn.ReLU(),
            # nn.Dropout(0.1),
            
            # nn.Linear(128, 64),
            # nn.BatchNorm1d(64),
            # nn.ReLU(),
            # nn.Dropout(0.1),
            
            nn.Linear(128, 1)
        )
        
        # 权重初始化
        self.apply(self._init_weights)
    
    def _init_weights(self, module):
        if isinstance(module, nn.Linear):
            nn.init.xavier_uniform_(module.weight)
            if module.bias is not None:
                nn.init.zeros_(module.bias)
    
    def forward(self, x):
        return self.net(x)

# ===== Data Loading =====
def load_data():
    train_df = pd.read_parquet(Config.TRAIN_PATH)
    test_df = pd.read_parquet(Config.TEST_PATH)
    submission_df = pd.read_csv(Config.SUBMISSION_PATH)
    Config.FEATURES = [col for col in train_df.columns.tolist() if col != Config.LABEL_COLUMN]
    print(f"Loaded data - Train: {train_df.shape}, Test: {test_df.shape}, Submission: {submission_df.shape}")
    print(f"Total features: {len(Config.FEATURES)}")
    return train_df.reset_index(drop=True), test_df.reset_index(drop=True), submission_df

# ===== Model Training =====
def train_mlp(X_train, y_train, X_valid, y_valid, X_test, scaler):
    X_train = scaler.transform(X_train)
    X_valid = scaler.transform(X_valid)
    X_test = scaler.transform(X_test)
    X_train = torch.tensor(X_train, dtype=torch.float32, device=Config.DEVICE)
    y_train = torch.tensor(y_train.values, dtype=torch.float32, device=Config.DEVICE).view(-1, 1)
    X_valid = torch.tensor(X_valid, dtype=torch.float32, device=Config.DEVICE)
    y_valid = torch.tensor(y_valid.values, dtype=torch.float32, device=Config.DEVICE).view(-1, 1)
    X_test = torch.tensor(X_test, dtype=torch.float32, device=Config.DEVICE)
    
    model = MLP(X_train.shape[1]).to(Config.DEVICE)
    optimizer = optim.AdamW(model.parameters(), lr=Config.LEARNING_RATE, weight_decay=Config.WEIGHT_DECAY)
    criterion = nn.MSELoss()
    scheduler = optim.lr_scheduler.CosineAnnealingLR(optimizer, T_max=Config.EPOCHS)
    
    best_model_state = None
    best_score = -np.inf
    patience_counter = 0
    
    for epoch in range(Config.EPOCHS):
        # 训练阶段
        model.train()
        idx = np.random.permutation(len(X_train))
        total_loss = 0
        num_batches = 0
        
        for i in range(0, len(X_train), Config.BATCH_SIZE):
            batch_idx = idx[i:i+Config.BATCH_SIZE]
            xb = X_train[batch_idx]
            yb = y_train[batch_idx]
            optimizer.zero_grad()
            pred = model(xb)
            loss = criterion(pred, yb)
            loss.backward()
            optimizer.step()
            total_loss += loss.item()
            num_batches += 1
        
        avg_loss = total_loss / num_batches
        scheduler.step()
        
        # 验证阶段
        model.eval()
        with torch.no_grad():
            val_pred = model(X_valid).cpu().numpy().flatten()
            val_score = np.corrcoef(val_pred, y_valid.cpu().numpy().flatten())[0, 1]
            
            if val_score > best_score:
                best_score = val_score
                best_model_state = model.state_dict().copy()
                patience_counter = 0
            else:
                patience_counter += 1
        
        # 每10轮显示一次进度
        if (epoch + 1) % 10 == 0:
            print(f"  Epoch {epoch+1}/{Config.EPOCHS} - Loss: {avg_loss:.4f} - Val Corr: {val_score:.4f} - Best: {best_score:.4f}")
        
        # 早停检查
        if patience_counter >= Config.PATIENCE:
            print(f"Early stopping at epoch {epoch+1}, best valid corr: {best_score:.4f}")
            break
    
    # 加载最佳模型
    model.load_state_dict(best_model_state)
    model.eval()
    
    # 用最佳模型预测
    with torch.no_grad():
        valid_pred = model(X_valid).cpu().numpy().flatten()
        test_pred = model(X_test).cpu().numpy().flatten()
    
    return valid_pred, test_pred, best_score

def train_and_evaluate(train_df, test_df):
    """使用k折交叉验证训练MLP模型"""
    X_train = train_df[Config.FEATURES]
    y_train = train_df[Config.LABEL_COLUMN]
    X_test = test_df[Config.FEATURES]
    
    # 初始化预测数组
    n_samples = len(train_df)
    oof_preds = np.zeros(n_samples)
    test_preds = np.zeros(len(test_df))
    
    # 初始化交叉验证
    kf = KFold(n_splits=Config.N_FOLDS, shuffle=True, random_state=Config.RANDOM_STATE)
    
    fold_scores = []
    best_fold_score = -np.inf
    best_fold_model = None
    best_fold_scaler = None
    best_model_state = None
    
    print(f"开始 {Config.N_FOLDS} 折交叉验证...")
    
    for fold, (train_idx, valid_idx) in enumerate(kf.split(train_df), 1):
        print(f"\n--- Fold {fold}/{Config.N_FOLDS} ---")
        
        # 分割数据
        X_fold_train = X_train.iloc[train_idx]
        y_fold_train = y_train.iloc[train_idx]
        X_fold_valid = X_train.iloc[valid_idx]
        y_fold_valid = y_train.iloc[valid_idx]
        
        # 数据标准化
        scaler = RobustScaler()
        scaler.fit(X_fold_train)
        
        # 训练模型
        valid_pred, test_pred, fold_score = train_mlp(
            X_fold_train, y_fold_train, X_fold_valid, y_fold_valid, X_test, scaler
        )
        
        # 保存OOF预测
        oof_preds[valid_idx] = valid_pred
        test_preds += test_pred / Config.N_FOLDS
        
        fold_scores.append(fold_score)
        print(f"Fold {fold} 验证集相关系数: {fold_score:.4f}")
        
        # 记录最佳模型
        if fold_score > best_fold_score:
            best_fold_score = fold_score
            best_fold_model = fold
            best_fold_scaler = scaler
            # 保存最佳模型状态(这里需要重新训练一次来获取状态)
            print(f"  -> 新的最佳模型 (Fold {fold})")
    
    # 计算整体OOF分数
    overall_score = pearsonr(y_train, oof_preds)[0]
    
    print(f"\n=== 交叉验证结果 ===")
    print(f"各折验证集相关系数: {[f'{score:.4f}' for score in fold_scores]}")
    print(f"平均验证集相关系数: {np.mean(fold_scores):.4f} ± {np.std(fold_scores):.4f}")
    print(f"最佳验证集相关系数: {best_fold_score:.4f} (Fold {best_fold_model})")
    print(f"整体OOF相关系数: {overall_score:.4f}")
    
    # 保存最佳模型信息
    model_info = {
        'best_fold': best_fold_model,
        'best_score': best_fold_score,
        'fold_scores': fold_scores,
        'overall_score': overall_score,
        'config': {
            'n_folds': Config.N_FOLDS,
            'epochs': Config.EPOCHS,
            'learning_rate': Config.LEARNING_RATE,
            'batch_size': Config.BATCH_SIZE
        }
    }
    
    with open('/AI4M/users/mjzhang/workspace/DRW/ZMJ/max_IC_mixed/mlp_best_model_info.pkl', 'wb') as f:
        pickle.dump(model_info, f)
    print(f"最佳模型信息已保存: mlp_best_model_info.pkl")
    
    return oof_preds, test_preds, fold_scores, best_fold_score

# ===== Submission =====
def create_submission(train_df, oof_preds, test_preds, submission_df, fold_scores, best_fold_score):
    overall_score = pearsonr(train_df[Config.LABEL_COLUMN], oof_preds)[0]
    print(f"\n=== MLP 最终结果 ===")
    print(f"各折验证集相关系数: {[f'{score:.4f}' for score in fold_scores]}")
    print(f"平均验证集相关系数: {np.mean(fold_scores):.4f} ± {np.std(fold_scores):.4f}")
    print(f"最佳验证集相关系数: {best_fold_score:.4f}")
    print(f"整体OOF相关系数: {overall_score:.4f}")
    
    submission = submission_df.copy()
    submission["prediction"] = test_preds
    submission.to_csv("/AI4M/users/mjzhang/workspace/DRW/ZMJ/max_IC_mixed/submission_mlp_cv.csv", index=False)
    print("Saved: submission_mlp_cv.csv")
    return overall_score

# ===== Main Execution =====
if __name__ == "__main__":
    print("Loading data...")
    train_df, test_df, submission_df = load_data()
    print("\nTraining MLP model...")
    oof_preds, test_preds, fold_scores, best_fold_score = train_and_evaluate(train_df, test_df)
    print("\nCreating submission...")
    score = create_submission(train_df, oof_preds, test_preds, submission_df, fold_scores, best_fold_score)
    print(f"\nAll done! MLP OOF PearsonR: {score:.4f}")