| |
| """Untitled17.ipynb |
| |
| Automatically generated by Colab. |
| |
| Original file is located at |
| https://colab.research.google.com/drive/1GwdSjrwh3f6QCzOa8KHr_XkWy0KmZvdV |
| """ |
|
|
| import pandas as pd |
| import numpy as np |
| from pathlib import Path |
| from math import log2 |
| from tqdm import tqdm |
| from sklearn.ensemble import GradientBoostingClassifier |
| from sklearn.preprocessing import StandardScaler |
| from sklearn.metrics import roc_auc_score, average_precision_score, classification_report |
| import warnings |
| warnings.filterwarnings('ignore') |
|
|
| PATHS = { |
| 'data_frozen': BASE_PATH / 'data' / 'frozen', |
| 'features': BASE_PATH / 'features', |
| 'models': BASE_PATH / 'models', |
| 'results': BASE_PATH / 'results', |
| } |
|
|
| for path in PATHS.values(): |
| path.mkdir(parents=True, exist_ok=True) |
|
|
| def shannon_entropy(seq): |
| """Calculer l'entropie de Shannon d'une séquence""" |
| if not seq or len(seq) == 0: |
| return 0.0 |
| probs = [seq.count(aa)/len(seq) for aa in set(seq)] |
| return -sum(p * log2(p) for p in probs if p > 0) |
|
|
| df_features_full = pd.read_parquet(PATHS['features'] / 'features_classical_full.parquet') |
| df_features_strict = pd.read_parquet(PATHS['features'] / 'features_classical_mito_strict.parquet') |
|
|
| print(f" Features chargées (full): {len(df_features_full):,}") |
| print(f" Features chargées (strict): {len(df_features_strict):,}") |
|
|
| uniprot_file = BASE_PATH / 'data' / 'raw' / 'uniprot_human_reviewed.parquet' |
| df_uniprot = pd.read_parquet(uniprot_file) |
| seq_dict = dict(zip(df_uniprot['accession'], df_uniprot['sequence'])) |
|
|
| print(f" Séquences: {len(seq_dict):,}") |
|
|
| def add_entropy_feature(df, seq_dict, window=15): |
| """Ajouter la feature d'entropie locale""" |
| entropies = [] |
|
|
| for _, row in tqdm(df.iterrows(), total=len(df), desc="Entropie"): |
| seq = seq_dict.get(row['uniprot_acc'], '') |
| pos = row['position'] |
|
|
| if seq and 0 <= pos < len(seq): |
| start = max(0, pos - window) |
| end = min(len(seq), pos + window + 1) |
| local_seq = seq[start:end] |
| entropies.append(shannon_entropy(local_seq)) |
| else: |
| entropies.append(0.0) |
|
|
| df['local_sequence_entropy'] = entropies |
| return df |
|
|
| print("\n Ajout de l'entropie locale...") |
| df_features_full = add_entropy_feature(df_features_full, seq_dict) |
| df_features_strict = add_entropy_feature(df_features_strict, seq_dict) |
|
|
| df_features_full.to_parquet(PATHS['features'] / 'features_classical_full.parquet') |
| df_features_strict.to_parquet(PATHS['features'] / 'features_classical_mito_strict.parquet') |
|
|
|
|
|
|
|
|
| id_cols = ['mutation_idx', 'uniprot_acc', 'gene_symbol', 'position', 'wt_aa', 'mut_aa', 'label'] |
| feature_cols = [c for c in df_features_full.columns if c not in id_cols] |
|
|
| print(f" Features: {len(feature_cols)}") |
|
|
| X_full = df_features_full[feature_cols].values |
| y_full = df_features_full['label'].values |
|
|
| X_strict = df_features_strict[feature_cols].values |
| y_strict = df_features_strict['label'].values |
|
|
| print(f" X shape: {X_full.shape}") |
| print(f" y: {np.sum(y_full==1)} pathogènes, {np.sum(y_full==0)} bénins") |
|
|
| print(f" X shape: {X_strict.shape}") |
| print(f" y: {np.sum(y_strict==1)} pathogènes, {np.sum(y_strict==0)} bénins") |
|
|
| X_full = np.nan_to_num(X_full, nan=0.0, posinf=0.0, neginf=0.0) |
| X_strict = np.nan_to_num(X_strict, nan=0.0, posinf=0.0, neginf=0.0) |
|
|
|
|
|
|
| from sklearn.model_selection import train_test_split |
|
|
| X_train, X_test, y_train, y_test = train_test_split( |
| X_full, y_full, test_size=0.2, random_state=42, stratify=y_full |
| ) |
|
|
| print(f" Train: {len(X_train)} ({np.sum(y_train==1)} patho)") |
| print(f" Test: {len(X_test)} ({np.sum(y_test==1)} patho)") |
|
|
| scaler = StandardScaler() |
| X_train_scaled = scaler.fit_transform(X_train) |
| X_test_scaled = scaler.transform(X_test) |
|
|
|
|
| model = GradientBoostingClassifier( |
| n_estimators=300, |
| max_depth=5, |
| learning_rate=0.05, |
| min_samples_leaf=10, |
| subsample=0.8, |
| random_state=42, |
| verbose=0 |
| ) |
|
|
| model.fit(X_train_scaled, y_train) |
| y_pred_proba = model.predict_proba(X_test_scaled)[:, 1] |
| y_pred = model.predict(X_test_scaled) |
|
|
| auc_roc = roc_auc_score(y_test, y_pred_proba) |
| auc_pr = average_precision_score(y_test, y_pred_proba) |
|
|
| print(f" AUC-ROC: {auc_roc:.4f}") |
| print(f" AUC-PR: {auc_pr:.4f}") |
|
|
| print(classification_report(y_test, y_pred, target_names=['Bénin', 'Pathogène'])) |
|
|
|
|
| proteins = df_features_full['uniprot_acc'].unique() |
| print(f" Protéines uniques: {len(proteins)}") |
|
|
| lpocv_results = [] |
| proteins_evaluated = 0 |
|
|
|
|
| for protein in tqdm(proteins, desc="LPOCV"): |
| test_mask = df_features_full['uniprot_acc'] == protein |
| train_mask = ~test_mask |
|
|
| if test_mask.sum() < 2: |
| continue |
|
|
| X_train_lpo = X_full[train_mask] |
| y_train_lpo = y_full[train_mask] |
| X_test_lpo = X_full[test_mask] |
| y_test_lpo = y_full[test_mask] |
|
|
| if len(np.unique(y_test_lpo)) < 2: |
| pass |
|
|
| scaler_lpo = StandardScaler() |
| X_train_lpo_scaled = scaler_lpo.fit_transform(X_train_lpo) |
| X_test_lpo_scaled = scaler_lpo.transform(X_test_lpo) |
|
|
| model_lpo = GradientBoostingClassifier( |
| n_estimators=100, |
| max_depth=4, |
| learning_rate=0.1, |
| min_samples_leaf=10, |
| random_state=42, |
| verbose=0 |
| ) |
|
|
| model_lpo.fit(X_train_lpo_scaled, y_train_lpo) |
|
|
| y_pred_lpo = model_lpo.predict_proba(X_test_lpo_scaled)[:, 1] |
|
|
| for i, (pred, true) in enumerate(zip(y_pred_lpo, y_test_lpo)): |
| lpocv_results.append({ |
| 'protein': protein, |
| 'y_true': true, |
| 'y_pred_proba': pred, |
| }) |
|
|
| proteins_evaluated += 1 |
|
|
| df_lpocv = pd.DataFrame(lpocv_results) |
|
|
|
|
| if len(df_lpocv) > 0 and len(df_lpocv['y_true'].unique()) > 1: |
| auc_roc_lpocv = roc_auc_score(df_lpocv['y_true'], df_lpocv['y_pred_proba']) |
| auc_pr_lpocv = average_precision_score(df_lpocv['y_true'], df_lpocv['y_pred_proba']) |
|
|
| print(f"\n 📊 RÉSULTATS LPOCV:") |
| print(f" AUC-ROC: {auc_roc_lpocv:.4f}") |
| print(f" AUC-PR: {auc_pr_lpocv:.4f}") |
| else: |
| auc_roc_lpocv = 0 |
| auc_pr_lpocv = 0 |
|
|
|
|
| scaler_final = StandardScaler() |
| X_full_scaled = scaler_final.fit_transform(X_full) |
|
|
| model_final = GradientBoostingClassifier( |
| n_estimators=300, |
| max_depth=5, |
| learning_rate=0.05, |
| min_samples_leaf=10, |
| subsample=0.8, |
| random_state=42, |
| verbose=0 |
| ) |
|
|
| model_final.fit(X_full_scaled, y_full) |
|
|
| importances = model_final.feature_importances_ |
| importance_df = pd.DataFrame({ |
| 'feature': feature_cols, |
| 'importance': importances |
| }).sort_values('importance', ascending=False) |
| for i, row in importance_df.head(20).iterrows(): |
| print(f" {row['importance']:.4f} {row['feature']}") |
|
|
| importance_df.to_csv(PATHS['results'] / 'feature_importances_classical.csv', index=False) |
|
|
|
|
| import pickle |
|
|
| model_data = { |
| 'model': model_final, |
| 'scaler': scaler_final, |
| 'feature_cols': feature_cols, |
| 'metrics': { |
| 'auc_roc_split': auc_roc, |
| 'auc_pr_split': auc_pr, |
| 'auc_roc_lpocv': auc_roc_lpocv, |
| 'auc_pr_lpocv': auc_pr_lpocv, |
| }, |
| 'n_samples': len(X_full), |
| 'n_features': len(feature_cols), |
| } |
|
|
| with open(PATHS['models'] / 'model_classical_baseline.pkl', 'wb') as f: |
| pickle.dump(model_data, f) |
|
|
|
|
| df_lpocv.to_parquet(PATHS['results'] / 'lpocv_predictions.parquet') |