Engine Predictive Maintenance Model
Model Overview
This is a Tuned Random Forest Classifier trained for predictive engine maintenance with SMOTE oversampling to handle class imbalance and achieve high recall for failure detection.
Model Details
- Model Type: Random Forest Classifier with SMOTE Pipeline
- Framework: scikit-learn, imbalanced-learn
- Task: Binary Classification (Engine Condition: Good/Failing)
- Input Features: 14 engineered sensor features (RPM, pressure, temperature, etc.)
- Output: Probability of engine failure (0-1)
Model Performance
Test Set Metrics
| Metric | Score |
|---|---|
| Accuracy | 0.6340 |
| Precision | 0.7456 |
| Recall | 0.6366 |
| F1 Score | 0.6868 |
| F2 Score | 0.6558 |
| ROC-AUC | 0.6893 |
| Brier Score | 0.2195 |
Key Insights
- High Recall (0.6366): Detects ~64% of actual failures
- Competitive Precision (0.7456): ~75% of predictions are correct
- Strong AUC (0.6893): Good discrimination between failure and non-failure cases
Intended Use
This model is designed for:
- Predictive Maintenance: Identify engines at risk of failure before breakdown
- Condition Monitoring: Support data-driven maintenance decision-making
- Fleet Management: Optimize maintenance scheduling and resource allocation
- Risk Assessment: Provide failure probability scores for maintenance prioritization
Limitations
- Trained on historical engine data with specific sensor configurations
- Performance may vary with new sensor types or operating conditions
- Model requires regular retraining with updated failure data
- Does not capture temporal degradation patterns (time-series)
- Assumes consistent sensor calibration and operating conditions
Training Data
- Dataset: Engine Predictive Maintenance Dataset
- Total Samples: 19,581 engines
- Training Samples: 13,674 (70%)
- Test Samples: 3,907 (20%)
- Features: 14 engineered features (6 raw + 8 derived)
- Class Distribution: Imbalanced (Good: ~63%, Failure: ~37%)
Training Procedure
- Data preprocessing and feature engineering
- Train-test split (70-20-10)
- SMOTE oversampling on training data to handle class imbalance
- Hyperparameter tuning via GridSearchCV with 5-fold cross-validation
- Model evaluation on held-out test set
Hyperparameters
- n_estimators: 400
- max_depth: 12
- min_samples_leaf: 4
- SMOTE k_neighbors: 5
- Random state: 42
Recommendations
- Threshold Tuning: Adjust decision threshold based on cost of false positives vs. false negatives
- Continuous Monitoring: Track model performance in production and retrain quarterly with new data
- Feature Importance: Use SHAP or feature importance analysis to identify critical sensors
- Ensemble Approaches: Consider combining with other models (XGBoost, LightGBM) for robust predictions
- Domain Expertise: Combine predictions with expert knowledge for final maintenance decisions
Citation
If you use this model, please cite:
@misc{predictive-maintenance-model-2026,
title={Engine Predictive Maintenance Model},
author={GreatLearning Capstone Team},
year={2026},
howpublished={Hugging Face Hub},
url={https://huggingface.co/models/nilanjanadevc/engine-predictive-maintenance-model}
}
License
This model is released under the MIT License. See LICENSE file for details.
Contact & Support
For questions or issues:
- GitHub: Check repository
- Hugging Face: @nilanjanadevc
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support