File size: 4,435 Bytes
2c9f973
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
---
language:
- en
license: mit
library_name: transformers
tags:
- materials-science
- crystallography
- generative-ai
- inverse-design
- chemistry
- photovoltaics
datasets:
- c-bone/mpdb-slme-full
base_model: c-bone/CrystaLLM-pi_base
pipeline_tag: text-generation
---

# Model Card for CrystaLLM-pi_SLME

## Model Details

### Model Description

**CrystaLLM-pi_SLME** is a conditional generative model designed for the discovery of high-performance photovoltaic materials. It is a fine-tuned version of the `CrystaLLM-pi` framework, based on a GPT-2 decoder-only architecture. This variant employs the **Property-Key-Value (PKV)** attention mechanism to condition the generation of Crystallographic Information Files (CIFs) on the **Spectroscopic Limited Maximum Efficiency (SLME)** metric.

The model generates crystal structures based on a single target scalar property:
1. **SLME** (%) - A theoretical maximum efficiency metric for photovoltaic absorbers.

- **Developed by:** Bone et al. (University College London)
- **Model type:** Autoregressive Transformer with Prefix Attention Conditioning
- **Language(s):** CIF (Crystallographic Information File) syntax
- **License:** MIT
- **Finetuned from model:** `c-bone/CrystaLLM-pi_base`

### Model Sources

- **Repository:** [GitHub: CrystaLLM-pi](https://github.com/C-Bone-UCL/CrystaLLM-pi)
- **Paper:** [Discovery and recovery of crystalline materials with property-conditioned transformers (arXiv:2511.21299)](https://arxiv.org/abs/2511.21299)
- **Dataset:** [HuggingFace: c-bone/mpdb-slme-full](https://huggingface.co/datasets/c-bone/mpdb-slme-full)

## Uses

### Direct Use

The model is intended for the exploration of chemical space for new photovoltaic candidates. Users can condition generation on high SLME values (e.g., >25%) to discover novel materials with optimal optical and electronic properties for solar energy conversion.

### Out-of-Scope Use

- **Large Unit Cells:** Context window limit applies (~1024 tokens).
- **Production Deployment:** Generated structures are theoretical predictions. Verification via Hybrid-DFT calculations and experimental synthesis is required.

## Bias, Risks, and Limitations

- **Implicit Learning:** The model was not explicitly trained on band gap data, but implicitly learned to target the optimal Shockley-Queisser range (1.2-1.4 eV) via the SLME metric. It may be less effective at targeting SLME values driven by mechanisms outside the primary training distribution.
- **Data Scarcity:** The model was fine-tuned on a relatively small dataset (~5.3K materials).

## How to Get Started with the Model

For instructions on how to load and run generation with this model, please refer to the `_load_and_generate.py` script in the [CrystaLLM-pi GitHub Repository](https://github.com/C-Bone-UCL/CrystaLLM-pi).

## Training Details

### Training Data

The model was fine-tuned on the **MP SLME** dataset, containing inorganic structures labeled with their calculated Spectroscopic Limited Maximum Efficiency.

- **Source:** Materials Project / Derived from Walker and Butler (via `c-bone/mpdb-slme-full`)
- **Preprocessing:** CIFs are augmented, tokenized, and SLME values are normalized.

### Training Procedure

- **Architecture:** GPT-2 with Property-Key-Value (PKV) encoder layers. (~38.7M parameters)
- **Mechanism:** Prefix Tuning (PKV) is used to inject the SLME target directly into the attention mechanism.

## Evaluation

### Metrics

The model is evaluated based on:
1. **Hit-Rate:** Fraction of generated materials with predicted SLME values near the target.
2. **VSUN:** Validity, Stability, Uniqueness, and Novelty of the generated candidates.

### Results

The model successfully generated stable, novel candidates (e.g., $Rb_2(NbBr_3)_3$) with high predicted efficiencies, demonstrating the ability to map complex structure-property relationships from limited data.

## Citation

```bibtex
@misc{bone2025discoveryrecoverycrystallinematerials,
      title={Discovery and recovery of crystalline materials with property-conditioned transformers}, 
      author={Cyprien Bone and Matthew Walker and Kuangdai Leng and Luis M. Antunes and Ricardo Grau-Crespo and Amil Aligayev and Javier Dominguez and Keith T. Butler},
      year={2025},
      eprint={2511.21299},
      archivePrefix={arXiv},
      primaryClass={cond-mat.mtrl-sci},
      url={[https://arxiv.org/abs/2511.21299](https://arxiv.org/abs/2511.21299)}, 
}