| | import sys |
| | import pandas as pd |
| | import numpy as np |
| | from sklearn.model_selection import KFold |
| | from xgboost import XGBRegressor |
| | from lightgbm import LGBMRegressor |
| | from sklearn.linear_model import ( |
| | HuberRegressor, RANSACRegressor, TheilSenRegressor, |
| | Lasso, ElasticNet, Ridge |
| | ) |
| | from sklearn.cross_decomposition import PLSRegression |
| | from sklearn.preprocessing import StandardScaler, RobustScaler |
| | from sklearn.ensemble import RandomForestRegressor |
| | from scipy.stats import pearsonr |
| | import warnings |
| | warnings.filterwarnings('ignore') |
| |
|
| | |
| | def feature_engineering(df): |
| | """Original features plus new robust features""" |
| | |
| | df['volume_weighted_sell'] = df['sell_qty'] * df['volume'] |
| | df['buy_sell_ratio'] = df['buy_qty'] / (df['sell_qty'] + 1e-8) |
| | df['selling_pressure'] = df['sell_qty'] / (df['volume'] + 1e-8) |
| | df['effective_spread_proxy'] = np.abs(df['buy_qty'] - df['sell_qty']) / (df['volume'] + 1e-8) |
| | |
| | |
| | df['log_volume'] = np.log1p(df['volume']) |
| | df['bid_ask_imbalance'] = (df['bid_qty'] - df['ask_qty']) / (df['bid_qty'] + df['ask_qty'] + 1e-8) |
| | df['order_flow_imbalance'] = (df['buy_qty'] - df['sell_qty']) / (df['buy_qty'] + df['sell_qty'] + 1e-8) |
| | df['liquidity_ratio'] = (df['bid_qty'] + df['ask_qty']) / (df['volume'] + 1e-8) |
| | |
| | |
| | df = df.replace([np.inf, -np.inf], np.nan) |
| | |
| | |
| | for col in df.columns: |
| | if df[col].isna().any(): |
| | median_val = df[col].median() |
| | df[col] = df[col].fillna(median_val if not pd.isna(median_val) else 0) |
| | |
| | return df |
| |
|
| | |
| | class Config: |
| | ORIGIN_TRAIN_PATH = "/AI4M/users/mjzhang/workspace/DRW/data/train.parquet" |
| | ORIGIN_TEST_PATH = "/AI4M/users/mjzhang/workspace/DRW/data/test.parquet" |
| | TRAIN_PATH = "/AI4M/users/mjzhang/workspace/DRW/ZMJ/threshold_6_29/train_final.parquet" |
| | TEST_PATH = "/AI4M/users/mjzhang/workspace/DRW/ZMJ/threshold_6_29/test_final.parquet" |
| | SUBMISSION_PATH = "/AI4M/users/mjzhang/workspace/DRW/ZMJ/threshold_6_29/sample_submission.csv" |
| | |
| | |
| | FEATURES = [ |
| | "X863", "X856", "X598", "X862", "X385", "X852", "X603", "X860", "X674", |
| | "X415", "X345", "X855", "X174", "X302", "X178", "X168", "X612", |
| | "buy_qty", "sell_qty", "volume", "X888", "X421", "X333", |
| | "bid_qty", "ask_qty" |
| | ] |
| | |
| | MERGE = False |
| | LABEL_COLUMN = "label" |
| | N_FOLDS = 5 |
| | RANDOM_STATE = 42 |
| |
|
| | |
| | |
| | XGB_PARAMS = { |
| | "tree_method": "hist", |
| | "device": "gpu", |
| | "colsample_bylevel": 0.3, |
| | "colsample_bynode": 0.25, |
| | "colsample_bytree": 0.5, |
| | "gamma": 2.5, |
| | "learning_rate": 0.015, |
| | "max_depth": 12, |
| | "max_leaves": 8, |
| | "min_child_weight": 25, |
| | "n_estimators": 2000, |
| | "subsample": 0.7, |
| | "reg_alpha": 50, |
| | "reg_lambda": 100, |
| | "verbosity": 0, |
| | "random_state": Config.RANDOM_STATE, |
| | "n_jobs": -1 |
| | } |
| |
|
| | |
| | LGBM_PARAMS = { |
| | "n_estimators": 800, |
| | "learning_rate": 0.02, |
| | "num_leaves": 25, |
| | "min_child_samples": 80, |
| | "subsample": 0.7, |
| | "colsample_bytree": 0.6, |
| | "reg_alpha": 15, |
| | "reg_lambda": 20, |
| | "min_split_gain": 0.1, |
| | "random_state": Config.RANDOM_STATE, |
| | "device": "gpu", |
| | "verbosity": -1, |
| | "n_jobs": -1 |
| | } |
| |
|
| | |
| | LEARNERS = [ |
| | {"name": "xgb_baseline", "Estimator": XGBRegressor, "params": XGB_PARAMS, "need_scale": False}, |
| | {"name": "lgbm", "Estimator": LGBMRegressor, "params": LGBM_PARAMS, "need_scale": False}, |
| | {"name": "huber", "Estimator": HuberRegressor, "params": {"epsilon": 1.35, "alpha": 0.005, "max_iter": 1000}, "need_scale": True}, |
| | {"name": "ransac", "Estimator": RANSACRegressor, "params": {"min_samples": 0.8, "max_trials": 150, "random_state": Config.RANDOM_STATE}, "need_scale": True}, |
| | {"name": "theilsen", "Estimator": TheilSenRegressor, "params": {"max_subpopulation": 15000, "random_state": Config.RANDOM_STATE}, "need_scale": True}, |
| | {"name": "lasso", "Estimator": Lasso, "params": {"alpha": 0.0005, "max_iter": 1500, "tol": 1e-4}, "need_scale": True}, |
| | {"name": "elasticnet", "Estimator": ElasticNet, "params": {"alpha": 0.0005, "l1_ratio": 0.3, "max_iter": 1500, "tol": 1e-4}, "need_scale": True}, |
| | {"name": "pls", "Estimator": PLSRegression, "params": {"n_components": 30}, "need_scale": True}, |
| | ] |
| |
|
| | |
| | def create_time_decay_weights(n: int, decay: float = 0.9) -> np.ndarray: |
| | """Create time decay weights for more recent data importance""" |
| | positions = np.arange(n) |
| | normalized = positions / (n - 1) |
| | weights = decay ** (1.0 - normalized) |
| | return weights * n / weights.sum() |
| |
|
| | def merge_origin_and_df(origin_df, df, features): |
| | """ |
| | 合并 origin_df 和 df,仅使用 origin_df 中指定的列,并保留一个 label 列。 |
| | |
| | 参数: |
| | origin_df (pd.DataFrame): 原始数据 DataFrame。 |
| | df (pd.DataFrame): 处理后的数据 DataFrame。 |
| | features (list): 需要从 origin_df 中提取的列名列表。 |
| | |
| | 返回: |
| | pd.DataFrame: 合并后的 DataFrame。 |
| | """ |
| | |
| | assert origin_df.index.equals(df.index), "两个 DataFrame 的索引必须一致" |
| |
|
| | |
| | origin_selected = origin_df[features] |
| |
|
| | |
| | |
| |
|
| | |
| | merged_df = pd.concat([origin_selected, df], axis=1) |
| |
|
| | return merged_df |
| |
|
| | def load_data(): |
| | """Load and preprocess data""" |
| | origin_train_df = pd.read_parquet(Config.ORIGIN_TRAIN_PATH) |
| | origin_test_df = pd.read_parquet(Config.ORIGIN_TEST_PATH) |
| | train_df = pd.read_parquet(Config.TRAIN_PATH) |
| | test_df = pd.read_parquet(Config.TEST_PATH) |
| | submission_df = pd.read_csv(Config.SUBMISSION_PATH) |
| | |
| | Config.AGGREGATE_FEATURES = [col for col in train_df.columns.tolist() if col != 'label'] |
| | |
| | if Config.MERGE == True: |
| | |
| | origin_train_df = feature_engineering(origin_train_df) |
| | origin_test_df = feature_engineering(origin_test_df) |
| | |
| | |
| | engineered_features = [ |
| | "volume_weighted_sell", "buy_sell_ratio", "selling_pressure", |
| | "effective_spread_proxy", "log_volume", "bid_ask_imbalance", |
| | "order_flow_imbalance", "liquidity_ratio" |
| | ] |
| | Config.FEATURES = list(set(Config.FEATURES + engineered_features)) |
| | merged_train_df = merge_origin_and_df(origin_train_df, train_df, Config.FEATURES) |
| | merged_test_df = merge_origin_and_df(origin_test_df, test_df, Config.FEATURES) |
| | Config.FEATURES = [col for col in merged_train_df.columns.tolist() if col != 'label'] |
| | else: |
| | Config.FEATURES = Config.AGGREGATE_FEATURES |
| | merged_train_df = train_df |
| | merged_test_df = test_df |
| | |
| | print(f"Loaded data - Train: {merged_train_df.shape}, Test: {merged_test_df.shape}, Submission: {submission_df.shape}") |
| | print(f"Total features: {len(Config.FEATURES)}") |
| | |
| | return merged_train_df.reset_index(drop=True), merged_test_df.reset_index(drop=True), submission_df |
| |
|
| | |
| | def get_model_slices(n_samples: int): |
| | """Define different data slices for training""" |
| | return [ |
| | {"name": "full_data", "cutoff": 0}, |
| | {"name": "last_75pct", "cutoff": int(0.25 * n_samples)}, |
| | {"name": "last_50pct", "cutoff": int(0.50 * n_samples)}, |
| | ] |
| |
|
| | def train_single_model(X_train, y_train, X_valid, y_valid, X_test, learner, sample_weights=None): |
| | """Train a single model with appropriate scaling if needed""" |
| | if learner["need_scale"]: |
| | scaler = RobustScaler() |
| | X_train_scaled = scaler.fit_transform(X_train) |
| | X_valid_scaled = scaler.transform(X_valid) |
| | X_test_scaled = scaler.transform(X_test) |
| | else: |
| | X_train_scaled = X_train |
| | X_valid_scaled = X_valid |
| | X_test_scaled = X_test |
| | |
| | model = learner["Estimator"](**learner["params"]) |
| | |
| | |
| | if learner["name"] in ["xgb_baseline", "lgbm"]: |
| | if learner["name"] == "xgb_baseline": |
| | model.fit(X_train_scaled, y_train, sample_weight=sample_weights, |
| | eval_set=[(X_valid_scaled, y_valid)], verbose=False) |
| | else: |
| | model.fit(X_train_scaled, y_train, sample_weight=sample_weights, |
| | eval_set=[(X_valid_scaled, y_valid)], callbacks=[]) |
| | elif learner["name"] in ["huber", "lasso", "elasticnet"]: |
| | model.fit(X_train_scaled, y_train, sample_weight=sample_weights) |
| | else: |
| | |
| | model.fit(X_train_scaled, y_train) |
| | |
| | valid_pred = model.predict(X_valid_scaled) |
| | test_pred = model.predict(X_test_scaled) |
| | |
| | return valid_pred, test_pred |
| |
|
| | def train_and_evaluate(train_df, test_df): |
| | """Train all models with cross-validation""" |
| | n_samples = len(train_df) |
| | model_slices = get_model_slices(n_samples) |
| | |
| | |
| | oof_preds = { |
| | learner["name"]: {s["name"]: np.zeros(n_samples) for s in model_slices} |
| | for learner in LEARNERS |
| | } |
| | test_preds = { |
| | learner["name"]: {s["name"]: np.zeros(len(test_df)) for s in model_slices} |
| | for learner in LEARNERS |
| | } |
| | |
| | full_weights = create_time_decay_weights(n_samples) |
| | kf = KFold(n_splits=Config.N_FOLDS, shuffle=False) |
| | |
| | for fold, (train_idx, valid_idx) in enumerate(kf.split(train_df), start=1): |
| | print(f"\n--- Fold {fold}/{Config.N_FOLDS} ---") |
| | X_valid = train_df.iloc[valid_idx][Config.FEATURES] |
| | y_valid = train_df.iloc[valid_idx][Config.LABEL_COLUMN] |
| | X_test = test_df[Config.FEATURES] |
| | |
| | for s in model_slices: |
| | cutoff = s["cutoff"] |
| | slice_name = s["name"] |
| | subset = train_df.iloc[cutoff:].reset_index(drop=True) |
| | rel_idx = train_idx[train_idx >= cutoff] - cutoff |
| | |
| | if len(rel_idx) == 0: |
| | continue |
| | |
| | X_train = subset.iloc[rel_idx][Config.FEATURES] |
| | y_train = subset.iloc[rel_idx][Config.LABEL_COLUMN] |
| | sw = create_time_decay_weights(len(subset))[rel_idx] if cutoff > 0 else full_weights[train_idx] |
| | |
| | print(f" Training slice: {slice_name}, samples: {len(X_train)}") |
| | |
| | for learner in LEARNERS: |
| | try: |
| | valid_pred, test_pred = train_single_model( |
| | X_train, y_train, X_valid, y_valid, X_test, learner, sw |
| | ) |
| | |
| | |
| | mask = valid_idx >= cutoff |
| | if mask.any(): |
| | idxs = valid_idx[mask] |
| | X_valid_subset = train_df.iloc[idxs][Config.FEATURES] |
| | if learner["need_scale"]: |
| | scaler = RobustScaler() |
| | scaler.fit(X_train) |
| | valid_pred_subset = learner["Estimator"](**learner["params"]).fit( |
| | scaler.transform(X_train), y_train |
| | ).predict(scaler.transform(X_valid_subset)) |
| | oof_preds[learner["name"]][slice_name][idxs] = valid_pred_subset |
| | else: |
| | oof_preds[learner["name"]][slice_name][idxs] = valid_pred[mask] |
| | |
| | if cutoff > 0 and (~mask).any(): |
| | oof_preds[learner["name"]][slice_name][valid_idx[~mask]] = \ |
| | oof_preds[learner["name"]]["full_data"][valid_idx[~mask]] |
| | |
| | test_preds[learner["name"]][slice_name] += test_pred |
| | |
| | except Exception as e: |
| | print(f" Error training {learner['name']}: {str(e)}") |
| | continue |
| | |
| | |
| | for learner_name in test_preds: |
| | for slice_name in test_preds[learner_name]: |
| | test_preds[learner_name][slice_name] /= Config.N_FOLDS |
| | |
| | return oof_preds, test_preds, model_slices |
| |
|
| | |
| | def create_submissions(train_df, oof_preds, test_preds, submission_df): |
| | """Create multiple submission files for different strategies""" |
| | all_submissions = {} |
| | |
| | |
| | if "xgb_baseline" in oof_preds: |
| | xgb_oof = np.mean(list(oof_preds["xgb_baseline"].values()), axis=0) |
| | xgb_test = np.mean(list(test_preds["xgb_baseline"].values()), axis=0) |
| | xgb_score = pearsonr(train_df[Config.LABEL_COLUMN], xgb_oof)[0] |
| | print(f"\nXGBoost Baseline Score: {xgb_score:.4f}") |
| | |
| | submission_xgb = submission_df.copy() |
| | submission_xgb["prediction"] = xgb_test |
| | submission_xgb.to_csv("/AI4M/users/mjzhang/workspace/DRW/ZMJ/threshold_6_30/submission_xgb_baseline.csv", index=False) |
| | all_submissions["xgb_baseline"] = xgb_score |
| | |
| | |
| | robust_methods = ["huber", "ransac", "theilsen"] |
| | robust_oof_list = [] |
| | robust_test_list = [] |
| | |
| | for method in robust_methods: |
| | if method in oof_preds: |
| | method_oof = np.mean(list(oof_preds[method].values()), axis=0) |
| | method_test = np.mean(list(test_preds[method].values()), axis=0) |
| | method_score = pearsonr(train_df[Config.LABEL_COLUMN], method_oof)[0] |
| | print(f"{method.upper()} Score: {method_score:.4f}") |
| | |
| | if not np.isnan(method_score): |
| | robust_oof_list.append(method_oof) |
| | robust_test_list.append(method_test) |
| | |
| | if robust_oof_list: |
| | robust_oof = np.mean(robust_oof_list, axis=0) |
| | robust_test = np.mean(robust_test_list, axis=0) |
| | robust_score = pearsonr(train_df[Config.LABEL_COLUMN], robust_oof)[0] |
| | print(f"\nRobust Ensemble Score: {robust_score:.4f}") |
| | |
| | submission_robust = submission_df.copy() |
| | submission_robust["prediction"] = robust_test |
| | submission_robust.to_csv("/AI4M/users/mjzhang/workspace/DRW/ZMJ/threshold_6_30/submission_robust_ensemble.csv", index=False) |
| | all_submissions["robust_ensemble"] = robust_score |
| | |
| | |
| | regularized_methods = ["lasso", "elasticnet"] |
| | reg_oof_list = [] |
| | reg_test_list = [] |
| | |
| | for method in regularized_methods: |
| | if method in oof_preds: |
| | method_oof = np.mean(list(oof_preds[method].values()), axis=0) |
| | method_test = np.mean(list(test_preds[method].values()), axis=0) |
| | method_score = pearsonr(train_df[Config.LABEL_COLUMN], method_oof)[0] |
| | print(f"{method.upper()} Score: {method_score:.4f}") |
| | |
| | if not np.isnan(method_score): |
| | reg_oof_list.append(method_oof) |
| | reg_test_list.append(method_test) |
| | |
| | if reg_oof_list: |
| | reg_oof = np.mean(reg_oof_list, axis=0) |
| | reg_test = np.mean(reg_test_list, axis=0) |
| | reg_score = pearsonr(train_df[Config.LABEL_COLUMN], reg_oof)[0] |
| | print(f"\nRegularized Ensemble Score: {reg_score:.4f}") |
| | |
| | submission_reg = submission_df.copy() |
| | submission_reg["prediction"] = reg_test |
| | submission_reg.to_csv("/AI4M/users/mjzhang/workspace/DRW/ZMJ/threshold_6_30/submission_regularized_ensemble.csv", index=False) |
| | all_submissions["regularized_ensemble"] = reg_score |
| | |
| | |
| | tree_methods = ["xgb_baseline", "lgbm"] |
| | tree_oof_list = [] |
| | tree_test_list = [] |
| | |
| | for method in tree_methods: |
| | if method in oof_preds: |
| | method_oof = np.mean(list(oof_preds[method].values()), axis=0) |
| | method_test = np.mean(list(test_preds[method].values()), axis=0) |
| | tree_oof_list.append(method_oof) |
| | tree_test_list.append(method_test) |
| | |
| | if tree_oof_list: |
| | tree_oof = np.mean(tree_oof_list, axis=0) |
| | tree_test = np.mean(tree_test_list, axis=0) |
| | tree_score = pearsonr(train_df[Config.LABEL_COLUMN], tree_oof)[0] |
| | print(f"\nTree Ensemble Score: {tree_score:.4f}") |
| | |
| | submission_tree = submission_df.copy() |
| | submission_tree["prediction"] = tree_test |
| | submission_tree.to_csv("/AI4M/users/mjzhang/workspace/DRW/ZMJ/threshold_6_30/submission_tree_ensemble.csv", index=False) |
| | all_submissions["tree_ensemble"] = tree_score |
| | |
| | |
| | all_oof_scores = {} |
| | all_oof_preds = {} |
| | all_test_preds = {} |
| | |
| | for learner_name in oof_preds: |
| | learner_oof = np.mean(list(oof_preds[learner_name].values()), axis=0) |
| | learner_test = np.mean(list(test_preds[learner_name].values()), axis=0) |
| | score = pearsonr(train_df[Config.LABEL_COLUMN], learner_oof)[0] |
| | |
| | if not np.isnan(score) and score > 0: |
| | all_oof_scores[learner_name] = score |
| | all_oof_preds[learner_name] = learner_oof |
| | all_test_preds[learner_name] = learner_test |
| | |
| | |
| | if all_oof_scores: |
| | total_score = sum(all_oof_scores.values()) |
| | weights = {k: v/total_score for k, v in all_oof_scores.items()} |
| | |
| | weighted_oof = sum(weights[k] * all_oof_preds[k] for k in weights) |
| | weighted_test = sum(weights[k] * all_test_preds[k] for k in weights) |
| | weighted_score = pearsonr(train_df[Config.LABEL_COLUMN], weighted_oof)[0] |
| | |
| | print(f"\nWeighted Full Ensemble Score: {weighted_score:.4f}") |
| | print("Weights:", {k: f"{v:.3f}" for k, v in weights.items()}) |
| | |
| | submission_weighted = submission_df.copy() |
| | submission_weighted["prediction"] = weighted_test |
| | submission_weighted.to_csv("/AI4M/users/mjzhang/workspace/DRW/ZMJ/threshold_6_30/submission_weighted_ensemble.csv", index=False) |
| | all_submissions["weighted_ensemble"] = weighted_score |
| | |
| | |
| | simple_oof = np.mean(list(all_oof_preds.values()), axis=0) |
| | simple_test = np.mean(list(all_test_preds.values()), axis=0) |
| | simple_score = pearsonr(train_df[Config.LABEL_COLUMN], simple_oof)[0] |
| | |
| | print(f"\nSimple Full Ensemble Score: {simple_score:.4f}") |
| | |
| | submission_simple = submission_df.copy() |
| | submission_simple["prediction"] = simple_test |
| | submission_simple.to_csv("/AI4M/users/mjzhang/workspace/DRW/ZMJ/threshold_6_30/submission_simple_ensemble.csv", index=False) |
| | all_submissions["simple_ensemble"] = simple_score |
| | |
| | |
| | print("\n" + "="*50) |
| | print("SUBMISSION SUMMARY:") |
| | print("="*50) |
| | for name, score in sorted(all_submissions.items(), key=lambda x: x[1], reverse=True): |
| | print(f"{name:25s}: {score:.4f}") |
| | |
| | return all_submissions |
| |
|
| | |
| | if __name__ == "__main__": |
| | print("Loading data...") |
| | train_df, test_df, submission_df = load_data() |
| | |
| | print("\nTraining models...") |
| | oof_preds, test_preds, model_slices = train_and_evaluate(train_df, test_df) |
| | |
| | print("\nCreating submissions...") |
| | submission_scores = create_submissions(train_df, oof_preds, test_preds, submission_df) |
| | |
| | print("\nAll submissions created successfully!") |
| | print("Files created:") |
| | print("- submission_xgb_baseline.csv (original baseline)") |
| | print("- submission_robust_ensemble.csv (Huber + RANSAC + TheilSen)") |
| | print("- submission_regularized_ensemble.csv (Lasso + ElasticNet)") |
| | print("- submission_tree_ensemble.csv (XGBoost + LightGBM)") |
| | print("- submission_weighted_ensemble.csv (weighted by performance)") |
| | print("- submission_simple_ensemble.csv (simple average)") |