MC-MLIPs
Molecular Crystals Database for Machine Learning Interatomic Potentials
Fine-tuned MACE models for polymorphic molecular crystals, trained using the AMLP framework.
Models
| Compound | CSD Code | Energy MAE (meV/atom) |
Force MAE (meV/Γ ) |
Model | Dataset |
|---|---|---|---|---|---|
| Resorcinol | RESORA | 1.568 | 3.903 | β model | β train β valid |
| Durene | DURENE | 1.647 | 5.193 | β model | β train β valid |
| Coumarin | COUMAR | 1.670 | 4.296 | β model | β train β valid |
| Benzamide | BZAMID | 0.713 | 8.786 | β model | β train β valid |
| Niacinamide | NICOAM | 1.513 | 7.207 | β model | β train β valid |
| Nicotinamide | NICOAC | 1.201 | 5.824 | β model | β train β valid |
| Isonicotinamide | EHOWIH | 1.912 | 10.809 | β model | β train β valid |
| Pyrazinamide | PYRIZIN | 1.634 | 6.732 | β model | β train β valid |
| Benzoic acid | BENZAC | 1.329 | 7.897 | β model | β train β valid |
| Acridine | ACRDIN | 3.700 | 8.300 | β model | β train β valid |
| Mean | 1.689 | 6.895 |
Training Protocol
- Foundation Model: MACE-MP-0 (
mace-mh-1-omol-1%) - Reference Data: DFT (PBE-D4) optimizations + AIMD trajectories (25-500K)
- DFT Settings: VASP, 750 eV cutoff, EDIFF = 10β»β· eV
Two-stage training:
- Initial: LR = 2Γ10β»Β³, energy weight = 100, force weight = 10
- SWA (epoch 200+): LR = 5Γ10β»β΅, force weight = 100
Early stopping with patience = 75 epochs. All models trained in float64.
Validation
All models validated for:
- Energy conservation: NVE drift < 10β»β΅ over 25 ps
- Thermal stability: NVT stable up to 600K
- Structural integrity: RDFs and Pβ order parameters preserved
Usage
We recommend using the AMLP-Analysis module (amlpa.py) for running simulations with these models:
python3 amlpa.py structure.xyz config.yaml
In your config.yaml, point to the downloaded model:
model_paths:
- 'path/to/model.model'
device: 'gpu'
gpus: ['cuda:0']
Alternatively, you can use the models directly via the MACE calculator:
from mace.calculators import MACECalculator
calc = MACECalculator(model_paths="path/to/model.model", device="cuda")
atoms.calc = calc
For full configuration options (MD, geometry optimization, RDF analysis, etc.), refer to the AMLP documentation.
Repository Structure
MC-MLIPs/
βββ models/ -> Trained MACE model files
βββ datasets/ -> Training/validation HDF5 files
Citation
If you use these models, please cite:
@article{lahouari2026mcmlips,
title={Benchmarking Fine-Tuned MACE Interatomic Potentials for Polymorphic Molecular Crystals},
author={Lahouari, Adam and Ai, Shen and Han, Jihye and Hoffstadt, Jillian and H{\"o}llmer, Philipp and Infante, Charlotte and Jain, Pulkita and Kadam, Sangram and Martirossyan, Maya M. and McCune, Amara and Newton, Hypatia and Paul, Shlok J. and Pena, Willmor and Raghoonanan, Jonathan and Sahu, Sumon and Tan, Oliver and Vergara, Andrea and Rogal, Jutta and Tuckerman, Mark E.},
journal={arXiv preprint arXiv:XXXXX},
year={2026}
}
@article{lahouari2026amlp,
title={Automated Machine Learning Pipeline: Large Language Models-Assisted Automated Data set Generation for Training Machine-Learned Interatomic Potentials},
author={Lahouari, Adam and Rogal, Jutta and Tuckerman, Mark E.},
journal={Journal of Chemical Theory and Computation},
volume={22},
number={1},
pages={305--317},
year={2026},
publisher={American Chemical Society},
doi={10.1021/acs.jctc.5c01610},
url={https://doi.org/10.1021/acs.jctc.5c01610}
}
License
MIT License
Acknowledgments
- MACE development team
- NYU High Performance Computing
Inference Providers NEW
This model isn't deployed by any Inference Provider. π Ask for provider support