| |
|
| | """CGMacros Data Preprocessing for Glucose Prediction
|
| |
|
| | Modified to predict glucose levels at 30, 60, and 120 minutes after meals with two input versions:
|
| | - **Raw version**: 60 timesteps (1-minute resolution)
|
| | - **Binned version**: 12 timesteps (5-minute bins)
|
| |
|
| | Usage:
|
| | 1. Update BASE_DIR below to point to your CGMacros dataset directory
|
| | 2. Ensure directory contains: bio.csv and participant folders (CGMacros-XXX/)
|
| | 3. Run: python glucose_prediction_preprocessing.py
|
| | """
|
| |
|
| | import os
|
| | import argparse
|
| | import pickle
|
| | import numpy as np
|
| | import pandas as pd
|
| | from glob import glob
|
| | from tqdm import tqdm
|
| | from scipy.stats import skew
|
| |
|
| |
|
| | BASE_DIR = "./CGMacros"
|
| | SAVE_DIR = os.path.join(BASE_DIR, "Prediction")
|
| | os.makedirs(SAVE_DIR, exist_ok=True)
|
| |
|
| |
|
| | WINDOW_SIZE = 60
|
| | BIN_SIZE = 5
|
| | N_BINS = WINDOW_SIZE // BIN_SIZE
|
| | TARGET_HORIZONS = [30, 60, 120]
|
| | CGM_SOURCES = ["Libre GL", "Dexcom GL"]
|
| |
|
| | print(f"Window size: {WINDOW_SIZE} minutes")
|
| | print(f"Number of bins: {N_BINS}")
|
| | print(f"Target horizons: {TARGET_HORIZONS} minutes")
|
| |
|
| | """## Load Participant Metadata"""
|
| |
|
| |
|
| | bio_df = pd.read_csv(os.path.join(BASE_DIR, "bio.csv"))
|
| | bio_df["participant_id"] = bio_df["subject"].apply(lambda x: f"CGMacros-{int(x):03d}")
|
| |
|
| |
|
| | def classify_hba1c(hba1c):
|
| | if pd.isna(hba1c):
|
| | return -1
|
| | elif hba1c < 5.7:
|
| | return 0
|
| | elif hba1c <= 6.4:
|
| | return 1
|
| | else:
|
| | return 2
|
| |
|
| | bio_df["Diagnosis"] = bio_df["A1c PDL (Lab)"].apply(classify_hba1c)
|
| |
|
| |
|
| | bio_static_map = bio_df.set_index("participant_id")[["Age", "Gender", "BMI", "A1c PDL (Lab)"]].to_dict("index")
|
| | diagnosis_map = dict(zip(bio_df["participant_id"], bio_df["Diagnosis"]))
|
| |
|
| | print(f"Loaded {len(bio_df)} participants")
|
| | print(f"\nDiagnosis distribution:")
|
| | print(bio_df["Diagnosis"].value_counts().sort_index())
|
| | print("\n0=Healthy, 1=Pre-diabetes, 2=Type 2 Diabetes, -1=Unknown")
|
| |
|
| | """## Helper Functions"""
|
| |
|
| | def load_participant_file(folder):
|
| | """Load and preprocess participant data"""
|
| | csv_file = glob(os.path.join(folder, "*.csv"))[0]
|
| | print(f"Loading: {csv_file}")
|
| | df = pd.read_csv(csv_file)
|
| | df["timestamp"] = pd.to_datetime(df["Timestamp"])
|
| |
|
| |
|
| | if "HR" in df.columns:
|
| | df["HR"] = df["HR"].ffill().bfill()
|
| | if df["HR"].isnull().any():
|
| | df["HR"] = df["HR"].fillna(df["HR"].median())
|
| | else:
|
| | df["HR"] = np.nan
|
| |
|
| |
|
| | if "METs" in df.columns:
|
| | pass
|
| | elif "Intensity" in df.columns:
|
| | df["METs"] = df["Intensity"].map({0: 10, 1: 30}).fillna(10)
|
| | else:
|
| | df["METs"] = 10
|
| |
|
| |
|
| | if "Calories (Activity)" in df.columns:
|
| | pass
|
| | elif "Steps" in df.columns:
|
| | df["Calories (Activity)"] = df["Steps"] * 0.05
|
| | else:
|
| | df["Calories (Activity)"] = 0.0
|
| |
|
| | return df
|
| |
|
| | def get_static_features(pid, row):
|
| | """Extract static features for a meal"""
|
| | meta = bio_static_map.get(pid, {"Age": np.nan, "Gender": "Unknown", "BMI": np.nan, "A1c PDL (Lab)": np.nan})
|
| |
|
| |
|
| | timestamp = row["timestamp"]
|
| | hour = timestamp.hour
|
| | hour_sin = np.sin(2 * np.pi * hour / 24)
|
| | hour_cos = np.cos(2 * np.pi * hour / 24)
|
| |
|
| |
|
| | is_morning = 1 if 6 <= hour < 12 else 0
|
| | is_evening = 1 if 18 <= hour < 24 else 0
|
| | is_weekend = 1 if timestamp.dayofweek >= 5 else 0
|
| |
|
| |
|
| | meal_type = row.get("Meal Type", "").lower()
|
| | is_breakfast = 1 if meal_type == "breakfast" else 0
|
| | is_lunch = 1 if meal_type == "lunch" else 0
|
| | is_dinner = 1 if meal_type == "dinner" else 0
|
| |
|
| | return [
|
| |
|
| | meta["Age"],
|
| | 1 if str(meta["Gender"]).lower().startswith("m") else 0,
|
| | meta["BMI"],
|
| | meta["A1c PDL (Lab)"],
|
| |
|
| | row.get("Calories", np.nan),
|
| | row.get("Carbs", np.nan),
|
| | row.get("Protein", np.nan),
|
| | row.get("Fat", np.nan),
|
| | row.get("Fiber", np.nan),
|
| |
|
| | hour_sin,
|
| | hour_cos,
|
| | is_morning,
|
| | is_evening,
|
| | is_weekend,
|
| |
|
| | is_breakfast,
|
| | is_lunch,
|
| | is_dinner
|
| | ]
|
| |
|
| | def bin_sequence(sequence, n_bins):
|
| | """Average sequence into n_bins of equal size"""
|
| | bin_size = len(sequence) // n_bins
|
| | binned = []
|
| | for i in range(n_bins):
|
| | start_idx = i * bin_size
|
| | end_idx = start_idx + bin_size
|
| | bin_data = sequence[start_idx:end_idx].mean(axis=0)
|
| | binned.append(bin_data)
|
| | return np.array(binned)
|
| |
|
| | def extract_meal_windows(df, pid, cgm_col):
|
| | """Extract meal windows with 60 min before meal and targets at 30, 60, 120 min after"""
|
| | output_X_raw, output_X_binned = [], []
|
| | output_y, output_static, output_pid, output_diagnosis = [], [], [], []
|
| | required_cols = [cgm_col, "HR", "Calories (Activity)", "METs"]
|
| |
|
| |
|
| | diagnosis_label = diagnosis_map.get(pid, -1)
|
| |
|
| | for _, row in df.iterrows():
|
| | if pd.isnull(row["Meal Type"]) or row["Meal Type"] not in ["breakfast", "lunch", "dinner"]:
|
| | continue
|
| |
|
| | meal_time = row["timestamp"]
|
| |
|
| |
|
| | start_before = meal_time - pd.Timedelta(minutes=WINDOW_SIZE)
|
| | end_before = meal_time
|
| | segment_before = df[(df["timestamp"] >= start_before) & (df["timestamp"] < end_before)]
|
| |
|
| |
|
| | if len(segment_before) < WINDOW_SIZE or segment_before[required_cols].isnull().any().any():
|
| | continue
|
| |
|
| |
|
| | targets = []
|
| | valid_targets = True
|
| | for horizon in TARGET_HORIZONS:
|
| | target_time = meal_time + pd.Timedelta(minutes=horizon)
|
| |
|
| | target_window = df[
|
| | (df["timestamp"] >= target_time - pd.Timedelta(minutes=2)) &
|
| | (df["timestamp"] <= target_time + pd.Timedelta(minutes=2))
|
| | ]
|
| |
|
| | tw_nonan = target_window[target_window[cgm_col].notna()]
|
| | if tw_nonan.empty:
|
| | valid_targets = False
|
| | break
|
| |
|
| | closest_idx = (tw_nonan["timestamp"] - target_time).abs().idxmin()
|
| | targets.append(tw_nonan.loc[closest_idx, cgm_col])
|
| |
|
| | if not valid_targets:
|
| | continue
|
| |
|
| |
|
| | x_seq_raw = segment_before[required_cols].values
|
| | x_seq_binned = bin_sequence(x_seq_raw, N_BINS)
|
| |
|
| |
|
| | s_feat = get_static_features(pid, row)
|
| |
|
| | output_X_raw.append(x_seq_raw)
|
| | output_X_binned.append(x_seq_binned)
|
| | output_y.append(targets)
|
| | output_static.append(s_feat)
|
| | output_pid.append(pid)
|
| | output_diagnosis.append(diagnosis_label)
|
| |
|
| | return output_X_raw, output_X_binned, output_y, output_static, output_pid, output_diagnosis
|
| |
|
| | """## Process All Participants"""
|
| |
|
| |
|
| | for cgm_col in CGM_SOURCES:
|
| | cgm_type = "Libre" if cgm_col == "Libre GL" else "Dexcom"
|
| | print(f"\n{'='*60}")
|
| | print(f"Processing {cgm_type} data...")
|
| | print(f"{'='*60}")
|
| |
|
| | all_X_raw, all_X_binned = [], []
|
| | all_y, all_static, all_pid, all_diagnosis = [], [], [], []
|
| |
|
| | for folder in tqdm(glob(os.path.join(BASE_DIR, "CGMacros-0*/"))):
|
| | folder_name = os.path.basename(folder.rstrip("/"))
|
| | df = load_participant_file(folder)
|
| |
|
| | X_raw, X_binned, y, static, pids, diagnosis = extract_meal_windows(df, folder_name, cgm_col)
|
| |
|
| | all_X_raw.extend(X_raw)
|
| | all_X_binned.extend(X_binned)
|
| | all_y.extend(y)
|
| | all_static.extend(static)
|
| | all_pid.extend(pids)
|
| | all_diagnosis.extend(diagnosis)
|
| |
|
| |
|
| | np.savez_compressed(
|
| | os.path.join(SAVE_DIR, f"{cgm_type.lower()}_raw_prediction.npz"),
|
| | X=np.array(all_X_raw),
|
| | static=np.array(all_static),
|
| | y=np.array(all_y),
|
| | participant_id=np.array(all_pid),
|
| | diagnosis=np.array(all_diagnosis)
|
| | )
|
| | print(f"\nSaved: {cgm_type}_raw — {len(all_X_raw)} samples")
|
| | print(f" X shape: {np.array(all_X_raw).shape}")
|
| | print(f" static shape: {np.array(all_static).shape}")
|
| | print(f" y shape: {np.array(all_y).shape}")
|
| | print(f" diagnosis distribution: {np.unique(all_diagnosis, return_counts=True)}")
|
| |
|
| |
|
| | np.savez_compressed(
|
| | os.path.join(SAVE_DIR, f"{cgm_type.lower()}_binned_prediction.npz"),
|
| | X=np.array(all_X_binned),
|
| | static=np.array(all_static),
|
| | y=np.array(all_y),
|
| | participant_id=np.array(all_pid),
|
| | diagnosis=np.array(all_diagnosis)
|
| | )
|
| | print(f"\nSaved: {cgm_type}_binned — {len(all_X_binned)} samples")
|
| | print(f" X shape: {np.array(all_X_binned).shape}")
|
| | print(f" static shape: {np.array(all_static).shape}")
|
| | print(f" y shape: {np.array(all_y).shape}")
|
| | print(f" diagnosis distribution: {np.unique(all_diagnosis, return_counts=True)}")
|
| |
|
| | """## Summary"""
|
| |
|
| | print("\n" + "="*60)
|
| | print("Output files (4 total):")
|
| | print("="*60)
|
| | print("\nEach .npz file contains:")
|
| | print("- X: (n_samples, seq_len, 4) — time series of CGM, HR, Calories, METs")
|
| | print(" * raw version: seq_len = 60 (1 minute resolution)")
|
| | print(" * binned version: seq_len = 12 (5 minute bins)")
|
| | print("\n- static: (n_samples, 17) — static features per meal:")
|
| | print(" Demographic: [Age, Gender (0/1), BMI, HbA1c]")
|
| | print(" Macronutrients: [Calories, Carbs, Protein, Fat, Fiber]")
|
| | print(" Time of meal: [hour_sin, hour_cos, is_morning, is_evening, is_weekend, is_breakfast, is_lunch, is_dinner]")
|
| | print("\n- y: (n_samples, 3) — glucose levels at [30min, 60min, 120min] after meal")
|
| | print("\n- participant_id: for grouped CV")
|
| | print("\n- diagnosis: (n_samples,) — diagnosis category for stratified analysis")
|
| | print(" * 0 = Healthy (HbA1c < 5.7)")
|
| | print(" * 1 = Pre-diabetes (5.7 ≤ HbA1c ≤ 6.4)")
|
| | print(" * 2 = Type 2 Diabetes (HbA1c > 6.4)")
|
| | print(" * -1 = Unknown")
|
| | print("\nFiles saved to:", SAVE_DIR)
|
| |
|
| | """## Verification"""
|
| |
|
| |
|
| | sample_file = os.path.join(SAVE_DIR, "libre_raw_prediction.npz")
|
| | if os.path.exists(sample_file):
|
| | data = np.load(sample_file)
|
| | print("\nSample file inspection:")
|
| | print(f"X shape: {data['X'].shape}")
|
| | print(f"static shape: {data['static'].shape}")
|
| | print(f"y shape: {data['y'].shape}")
|
| | print(f"participant_id shape: {data['participant_id'].shape}")
|
| | print(f"diagnosis shape: {data['diagnosis'].shape}")
|
| | print(f"\nSample X (first 3 timesteps):\n{data['X'][0][:3]}")
|
| | print(f"\nSample static features (17 features):")
|
| | static_names = ['Age', 'Gender', 'BMI', 'HbA1c',
|
| | 'Calories', 'Carbs', 'Protein', 'Fat', 'Fiber',
|
| | 'hour_sin', 'hour_cos', 'is_morning', 'is_evening', 'is_weekend',
|
| | 'is_breakfast', 'is_lunch', 'is_dinner']
|
| | for i, (name, val) in enumerate(zip(static_names, data['static'][0])):
|
| | print(f" {i}. {name}: {val}")
|
| | print(f"\nSample y (targets at 30, 60, 120 min):\n{data['y'][0]}")
|
| | print(f"\nSample participant_id: {data['participant_id'][0]}")
|
| | print(f"Sample diagnosis: {data['diagnosis'][0]} (0=Healthy, 1=Pre-diabetes, 2=T2D, -1=Unknown)")
|
| |
|
| | """## Normalization"""
|
| |
|
| |
|
| |
|
| |
|
| | INDICES = {
|
| | 'age': 0, 'gender': 1, 'bmi': 2, 'hba1c': 3,
|
| | 'calories': 4, 'carbs': 5, 'protein': 6, 'fat': 7, 'fiber': 8,
|
| | 'hour_sin': 9, 'hour_cos': 10,
|
| | 'is_morning': 11, 'is_evening': 12, 'is_weekend': 13,
|
| | 'is_breakfast': 14, 'is_lunch': 15, 'is_dinner': 16
|
| | }
|
| |
|
| | def normalize_data(data_file):
|
| | """Normalize data following the suggested approach"""
|
| | print(f"\nProcessing: {os.path.basename(data_file)}")
|
| |
|
| |
|
| | data = np.load(data_file)
|
| | X = data['X'].copy()
|
| | static = data['static'].copy()
|
| | y = data['y'].copy()
|
| | participant_ids = data['participant_id']
|
| | diagnosis = data['diagnosis']
|
| |
|
| | print(f" Samples: {X.shape[0]}")
|
| |
|
| |
|
| | X_norm = X.copy()
|
| | static_norm = static.copy()
|
| |
|
| |
|
| | norm_params = {'per_subject': {}, 'global': {}}
|
| |
|
| |
|
| |
|
| |
|
| | print(" Normalizing time series per subject...")
|
| | unique_subjects = np.unique(participant_ids)
|
| |
|
| | for subject in unique_subjects:
|
| | subject_mask = participant_ids == subject
|
| | subject_X = X[subject_mask]
|
| | subject_X_flat = subject_X.reshape(-1, 4)
|
| |
|
| |
|
| | cgm_mean = np.nanmean(subject_X_flat[:, 0])
|
| | cgm_std = np.nanstd(subject_X_flat[:, 0])
|
| | if cgm_std > 0:
|
| | X_norm[subject_mask, :, 0] = (X[subject_mask, :, 0] - cgm_mean) / cgm_std
|
| |
|
| |
|
| | hr_mean = np.nanmean(subject_X_flat[:, 1])
|
| | hr_std = np.nanstd(subject_X_flat[:, 1])
|
| | if hr_std > 0:
|
| | X_norm[subject_mask, :, 1] = (X[subject_mask, :, 1] - hr_mean) / hr_std
|
| |
|
| |
|
| | cal_min = np.nanmin(subject_X_flat[:, 2])
|
| | cal_max = np.nanmax(subject_X_flat[:, 2])
|
| | if cal_max > cal_min:
|
| | X_norm[subject_mask, :, 2] = (X[subject_mask, :, 2] - cal_min) / (cal_max - cal_min)
|
| |
|
| |
|
| | mets_min = np.nanmin(subject_X_flat[:, 3])
|
| | mets_max = np.nanmax(subject_X_flat[:, 3])
|
| | if mets_max > mets_min:
|
| | X_norm[subject_mask, :, 3] = (X[subject_mask, :, 3] - mets_min) / (mets_max - mets_min)
|
| |
|
| |
|
| | norm_params['per_subject'][subject] = {
|
| | 'cgm_mean': cgm_mean, 'cgm_std': cgm_std,
|
| | 'hr_mean': hr_mean, 'hr_std': hr_std,
|
| | 'cal_min': cal_min, 'cal_max': cal_max,
|
| | 'mets_min': mets_min, 'mets_max': mets_max
|
| | }
|
| |
|
| |
|
| |
|
| |
|
| | print(" Normalizing static features globally...")
|
| |
|
| |
|
| | age_mean = np.nanmean(static[:, INDICES['age']])
|
| | age_std = np.nanstd(static[:, INDICES['age']])
|
| | if age_std > 0:
|
| | static_norm[:, INDICES['age']] = (static[:, INDICES['age']] - age_mean) / age_std
|
| | norm_params['global']['age'] = {'mean': float(age_mean), 'std': float(age_std)}
|
| |
|
| |
|
| | bmi_mean = np.nanmean(static[:, INDICES['bmi']])
|
| | bmi_std = np.nanstd(static[:, INDICES['bmi']])
|
| | if bmi_std > 0:
|
| | static_norm[:, INDICES['bmi']] = (static[:, INDICES['bmi']] - bmi_mean) / bmi_std
|
| | norm_params['global']['bmi'] = {'mean': float(bmi_mean), 'std': float(bmi_std)}
|
| |
|
| |
|
| | hba1c_mean = np.nanmean(static[:, INDICES['hba1c']])
|
| | hba1c_std = np.nanstd(static[:, INDICES['hba1c']])
|
| | if hba1c_std > 0:
|
| | static_norm[:, INDICES['hba1c']] = (static[:, INDICES['hba1c']] - hba1c_mean) / hba1c_std
|
| | norm_params['global']['hba1c'] = {'mean': float(hba1c_mean), 'std': float(hba1c_std)}
|
| |
|
| |
|
| | macro_features = ['calories', 'carbs', 'protein', 'fat', 'fiber']
|
| | for feat in macro_features:
|
| | idx = INDICES[feat]
|
| | values = static[:, idx]
|
| | values_clean = values[~np.isnan(values)]
|
| |
|
| | if len(values_clean) > 0:
|
| | skewness = skew(values_clean)
|
| |
|
| | if abs(skewness) > 1.0:
|
| | print(f" {feat} is skewed ({skewness:.2f}), applying log(x+1) transform")
|
| | values_transformed = np.log1p(values)
|
| | mean_val = np.nanmean(values_transformed)
|
| | std_val = np.nanstd(values_transformed)
|
| | if std_val > 0:
|
| | static_norm[:, idx] = (values_transformed - mean_val) / std_val
|
| | norm_params['global'][feat] = {
|
| | 'mean': float(mean_val), 'std': float(std_val), 'log_transformed': True
|
| | }
|
| | else:
|
| | mean_val = np.nanmean(values)
|
| | std_val = np.nanstd(values)
|
| | if std_val > 0:
|
| | static_norm[:, idx] = (values - mean_val) / std_val
|
| | norm_params['global'][feat] = {
|
| | 'mean': float(mean_val), 'std': float(std_val), 'log_transformed': False
|
| | }
|
| |
|
| | print(f" Normalization complete!")
|
| | print(f" X range: [{X_norm.min():.3f}, {X_norm.max():.3f}]")
|
| | print(f" Static range: [{static_norm.min():.3f}, {static_norm.max():.3f}]")
|
| | print(f" y unchanged: [{y.min():.1f}, {y.max():.1f}] mg/dL")
|
| |
|
| | return X_norm, static_norm, y, participant_ids, diagnosis, norm_params
|
| |
|
| |
|
| |
|
| |
|
| | OUTPUT_DIR_NORM = os.path.join(SAVE_DIR, "..", "Prediction_Normalized")
|
| | os.makedirs(OUTPUT_DIR_NORM, exist_ok=True)
|
| |
|
| | all_norm_params = {}
|
| | CGM_TYPES = ["libre", "dexcom"]
|
| | VERSIONS = ["raw", "binned"]
|
| |
|
| | for cgm_type in CGM_TYPES:
|
| | for version in VERSIONS:
|
| | filename = f"{cgm_type}_{version}_prediction.npz"
|
| | input_file = os.path.join(SAVE_DIR, filename)
|
| |
|
| | if not os.path.exists(input_file):
|
| | print(f" File not found: {filename}")
|
| | continue
|
| |
|
| |
|
| | X_norm, static_norm, y, pids, diagnosis, norm_params = normalize_data(input_file)
|
| |
|
| |
|
| | output_file = os.path.join(OUTPUT_DIR_NORM, filename)
|
| | np.savez_compressed(
|
| | output_file,
|
| | X=X_norm,
|
| | static=static_norm,
|
| | y=y,
|
| | participant_id=pids,
|
| | diagnosis=diagnosis
|
| | )
|
| | print(f" Saved: {output_file}\n")
|
| |
|
| |
|
| | all_norm_params[f"{cgm_type}_{version}"] = norm_params
|
| |
|
| |
|
| | params_file = os.path.join(OUTPUT_DIR_NORM, "normalization_params.pkl")
|
| | with open(params_file, 'wb') as f:
|
| | pickle.dump(all_norm_params, f)
|
| | print(f"Normalization parameters saved to: {params_file}")
|
| |
|
| | print("\n" + "="*60)
|
| | print("NORMALIZATION COMPLETE!")
|
| | print("="*60)
|
| | print(f"\nNormalized files saved to: {OUTPUT_DIR_NORM}")
|
| | print("\nFiles created:")
|
| | for cgm_type in CGM_TYPES:
|
| | for version in VERSIONS:
|
| | print(f" - {cgm_type}_{version}_prediction.npz")
|
| | print(" - normalization_params.pkl")
|
| | print("\nTarget glucose (y) kept in mg/dL for evaluation.") |