Graph Machine Learning
equiformer_v3 / README.md
nielsr's picture
nielsr HF Staff
Add pipeline tag and improve model card discovery
087f70d verified
|
raw
history blame
4.82 kB
---
license: mit
pipeline_tag: graph-ml
---
<h1 align="center" style="font-size: 24px;">EquiformerV3:<br>Scaling Efficient, Expressive, and General SE(3)-Equivariant Graph Attention Transformers</h1>
<p align="center">
<a href="https://github.com/atomicarchitects/equiformer_v3" style="color: #1a73e8; font-weight: bold; font-size: 20px;">Code</a> |
<a href="https://arxiv.org/abs/2604.09130" style="color: #1a73e8; font-weight: bold; font-size: 20px;">Paper</a>
</p>
This repository contains the checkpoints for **EquiformerV3**, the third generation of the $SE(3)$-equivariant graph attention Transformer. EquiformerV3 is designed to advance efficiency, expressivity, and generality in 3D atomistic modeling.
Building on EquiformerV2, this version introduces software optimizations achieving a $1.75\times$ speedup, structural improvements like equivariant merged layer normalization and smooth-cutoff attention, and SwiGLU-$S^2$ activations to incorporate many-body interactions while preserving strict equivariance. EquiformerV3 achieves state-of-the-art results on benchmarks including OC20, OMat24, and Matbench Discovery.
Please refer to the [official GitHub repository](https://github.com/atomicarchitects/equiformer_v3) for detailed instructions on environment setup and usage.
<p align="center">
<img width="50%" height="50%" src="https://cdn-uploads.huggingface.co/production/uploads/64948a4a8d5ff0dd776655fe/03TPndezDyUw4FcfTBk4n.png">
</p>
## Checkpoints
### MPtrj
<table>
<tr style="background-color: #f0f0f0;">
<td><strong>Model</strong></td>
<td><strong>Training data</strong></td>
<td><strong>Checkpoint</strong></td>
</tr>
<tr>
<td>EquiformerV3</td>
<td>MPtrj</td>
<td>
<a href="https://huggingface.co/yilunliao/equiformer_v3/blob/main/checkpoint/mptrj_gradient.pt">
mptrj_gradient.pt
</a>
</td>
</tr>
</table>
### OMat24 → MPtrj and sAlex
Training consists of (1) direct pre-training on OMat24, (2) gradient fine-tuning on OMat24 initialized from (1), and (3) gradient fine-tuning on MPtrj and sAlex initialized from (2).
<table>
<tr style="background-color: #f0f0f0;">
<td><strong>Model</strong></td>
<td><strong>Training data</strong></td>
<td><strong>Config</strong></td>
<td><strong>Checkpoint</strong></td>
</tr>
<tr>
<td>EquiformerV3 (direct pre-training)</td>
<td>OMat24</td>
<td>
<a href="https://github.com/atomicarchitects/equiformer_v3/blob/main/experimental/configs/omat24/omat24/experiments/direct/equiformer_v3_N%407_L%404_attn-hidden%4032_rbf%4064_max-neighbors%40300_attn-grid%4014-8_ffn-grid%4014_use-gate-force-head_merge-layer-norm_epochs%404-bs%40512-wd%401e-3-beta2%400.98-eps%401e-6_dens-p%400.5-std%400.025-r%400.5-0.75-w%401-no-stress-max-f%402.5_no-amp.yml">
omat24_direct.yml
</a>
</td>
<td>
<a href="https://huggingface.co/yilunliao/equiformer_v3/blob/main/checkpoint/omat24_direct.pt">
omat24_direct.pt
</a>
</td>
</tr>
<tr>
<td>EquiformerV3 (gradient fine-tuning)</td>
<td>OMat24</td>
<td>
<a href="https://github.com/atomicarchitects/equiformer_v3/blob/main/experimental/configs/omat24/omat24/experiments/gradient/equiformer_v3_grad-finetune_N%407_L%404_attn-hidden%4032_rbf%4064_max-neighbors%40300_attn-grid%4014-8_ffn-grid%4014_merge-layer-norm_lr%400-1e-4-epochs%402-bs%40512-wd%401e-3-beta2%400.98-eps%401e-6_pt-reg-dens-ft-no-reg.yml">
omat24_gradient.yml
</a>
</td>
<td>
<a href="https://huggingface.co/yilunliao/equiformer_v3/blob/main/checkpoint/omat24_gradient.pt">
omat24_gradient.pt
</a>
</td>
</tr>
<tr>
<td>EquiformerV3 (gradient fine-tuning)</td>
<td>MPtrj and sAlex</td>
<td>
<a href="https://github.com/atomicarchitects/equiformer_v3/blob/main/experimental/configs/omat24/salex_mptrj/experiments/gradient/equiformer_v3_grad-finetune_N%407_L%404_attn-hidden%4032_rbf%4064_max-neighbors%40300_attn-grid%4014-8_ffn-grid%4014_attn-eps%401e-8_lr%400-5e-5-warmup%400.1-epochs%402-mptrj-salex-ratio%408-bs%40256-wd%401e-3-beta2%400.98-eps%401e-6_pt-reg-dens-ft-no-reg-lr%401e-4.yml">
mptrj-salex_gradient.yml
</a>
</td>
<td>
<a href="https://huggingface.co/yilunliao/equiformer_v3/blob/main/checkpoint/omat24-mptrj-salex_gradient.pt">
omat24-mptrj-salex_gradient.pt
</a>
</td>
</tr>
</table>
## Citation
If you find this work helpful, please consider citing:
```bibtex
@article{equiformer_v3,
title={EquiformerV3: Scaling Efficient, Expressive, and General SE(3)-Equivariant Graph Attention Transformers},
author={Yi-Lun Liao and Alexander J. Hoffman and Sabrina C. Shen and Alexandre Duval and Sam Walton Norwood and Tess Smidt},
journal={arXiv preprint arXiv:2604.09130},
year={2026}
}
```