--- language: - en license: mit library_name: transformers tags: - materials-science - crystallography - generative-ai - inverse-design - chemistry - photovoltaics datasets: - c-bone/mpdb-slme-full base_model: c-bone/CrystaLLM-pi_base pipeline_tag: text-generation --- # Model Card for CrystaLLM-pi_SLME ## Model Details ### Model Description **CrystaLLM-pi_SLME** is a conditional generative model designed for the discovery of high-performance photovoltaic materials. It is a fine-tuned version of the `CrystaLLM-pi` framework, based on a GPT-2 decoder-only architecture. This variant employs the **Property-Key-Value (PKV)** attention mechanism to condition the generation of Crystallographic Information Files (CIFs) on the **Spectroscopic Limited Maximum Efficiency (SLME)** metric. The model generates crystal structures based on a single target scalar property: 1. **SLME** (%) - A theoretical maximum efficiency metric for photovoltaic absorbers. - **Developed by:** Bone et al. (University College London) - **Model type:** Autoregressive Transformer with Prefix Attention Conditioning - **Language(s):** CIF (Crystallographic Information File) syntax - **License:** MIT - **Finetuned from model:** `c-bone/CrystaLLM-pi_base` ### Model Sources - **Repository:** [GitHub: CrystaLLM-pi](https://github.com/C-Bone-UCL/CrystaLLM-pi) - **Paper:** [Discovery and recovery of crystalline materials with property-conditioned transformers (arXiv:2511.21299)](https://arxiv.org/abs/2511.21299) - **Dataset:** [HuggingFace: c-bone/mpdb-slme-full](https://huggingface.co/datasets/c-bone/mpdb-slme-full) ## Uses ### Direct Use The model is intended for the exploration of chemical space for new photovoltaic candidates. Users can condition generation on high SLME values (e.g., >25%) to discover novel materials with optimal optical and electronic properties for solar energy conversion. ### Out-of-Scope Use - **Large Unit Cells:** Context window limit applies (~1024 tokens). - **Production Deployment:** Generated structures are theoretical predictions. Verification via Hybrid-DFT calculations and experimental synthesis is required. ## Bias, Risks, and Limitations - **Implicit Learning:** The model was not explicitly trained on band gap data, but implicitly learned to target the optimal Shockley-Queisser range (1.2-1.4 eV) via the SLME metric. It may be less effective at targeting SLME values driven by mechanisms outside the primary training distribution. - **Data Scarcity:** The model was fine-tuned on a relatively small dataset (~5.3K materials). ## How to Get Started with the Model For instructions on how to load and run generation with this model, please refer to the `_load_and_generate.py` script in the [CrystaLLM-pi GitHub Repository](https://github.com/C-Bone-UCL/CrystaLLM-pi). ## Training Details ### Training Data The model was fine-tuned on the **MP SLME** dataset, containing inorganic structures labeled with their calculated Spectroscopic Limited Maximum Efficiency. - **Source:** Materials Project / Derived from Walker and Butler (via `c-bone/mpdb-slme-full`) - **Preprocessing:** CIFs are augmented, tokenized, and SLME values are normalized. ### Training Procedure - **Architecture:** GPT-2 with Property-Key-Value (PKV) encoder layers. (~38.7M parameters) - **Mechanism:** Prefix Tuning (PKV) is used to inject the SLME target directly into the attention mechanism. ## Evaluation ### Metrics The model is evaluated based on: 1. **Hit-Rate:** Fraction of generated materials with predicted SLME values near the target. 2. **VSUN:** Validity, Stability, Uniqueness, and Novelty of the generated candidates. ### Results The model successfully generated stable, novel candidates (e.g., $Rb_2(NbBr_3)_3$) with high predicted efficiencies, demonstrating the ability to map complex structure-property relationships from limited data. ## Citation ```bibtex @misc{bone2025discoveryrecoverycrystallinematerials, title={Discovery and recovery of crystalline materials with property-conditioned transformers}, author={Cyprien Bone and Matthew Walker and Kuangdai Leng and Luis M. Antunes and Ricardo Grau-Crespo and Amil Aligayev and Javier Dominguez and Keith T. Butler}, year={2025}, eprint={2511.21299}, archivePrefix={arXiv}, primaryClass={cond-mat.mtrl-sci}, url={[https://arxiv.org/abs/2511.21299](https://arxiv.org/abs/2511.21299)}, }