Add pipeline tag and improve model card discovery (#1)
Browse files- Add pipeline tag and improve model card discovery (087f70d8d9666596ea992b77b4802cd578f1cbff)
Co-authored-by: Niels Rogge <nielsr@users.noreply.huggingface.co>
README.md
CHANGED
|
@@ -1,87 +1,28 @@
|
|
| 1 |
---
|
| 2 |
license: mit
|
|
|
|
| 3 |
---
|
| 4 |
|
| 5 |
<h1 align="center" style="font-size: 24px;">EquiformerV3:<br>Scaling Efficient, Expressive, and General SE(3)-Equivariant Graph Attention Transformers</h1>
|
| 6 |
|
| 7 |
-
<
|
| 8 |
-
# **[Code](https://github.com/atomicarchitects/equiformer_v3)** | **[Paper]()**
|
| 9 |
-
-->
|
| 10 |
-
|
| 11 |
<a href="https://github.com/atomicarchitects/equiformer_v3" style="color: #1a73e8; font-weight: bold; font-size: 20px;">Code</a> |
|
| 12 |
<a href="https://arxiv.org/abs/2604.09130" style="color: #1a73e8; font-weight: bold; font-size: 20px;">Paper</a>
|
| 13 |
-
|
| 14 |
-
This repository contains the checkpoints of the work "EquiformerV3: Scaling Efficient, Expressive, and General SE(3)-Equivariant Graph Attention Transformers".
|
| 15 |
-
Please refer to the [code](https://github.com/atomicarchitects/equiformer_v3) for detailed description of usage.
|
| 16 |
-
|
| 17 |
-
<p align="center">
|
| 18 |
-
<img width="50%" height="50%" src="https://cdn-uploads.huggingface.co/production/uploads/64948a4a8d5ff0dd776655fe/03TPndezDyUw4FcfTBk4n.png"?
|
| 19 |
</p>
|
| 20 |
|
|
|
|
| 21 |
|
|
|
|
| 22 |
|
|
|
|
| 23 |
|
| 24 |
-
|
| 25 |
-
<
|
| 26 |
-
|
| 27 |
-
-->
|
| 28 |
-
0. [MPtrj](#mptrj)
|
| 29 |
-
0. [OMat24 → MPtrj and sAlex](#oam)
|
| 30 |
-
|
| 31 |
-
|
| 32 |
-
<!--
|
| 33 |
-
<h2 id="oc20">OC20</h2>
|
| 34 |
-
|
| 35 |
-
<table>
|
| 36 |
-
<tr style="background-color: #f0f0f0;">
|
| 37 |
-
<td><strong>Model</strong></td>
|
| 38 |
-
<td><strong>Training data</strong></td>
|
| 39 |
-
<td><strong>Config</strong></td>
|
| 40 |
-
<td><strong>Checkpoint</strong></td>
|
| 41 |
-
</tr>
|
| 42 |
-
|
| 43 |
-
<tr>
|
| 44 |
-
<td>EquiformerV3 (91M)</td>
|
| 45 |
-
<td>OC20 S2EF-2M</td>
|
| 46 |
-
<td><a href="https://github.com/atomicarchitects/equiformer_v3/blob/main/experimental/configs/oc20/2M/equiformer_v3/experiments/base_N%408-L%406-C%40128-attn-hidden%4064-ffn%40512-envelope-num-rbf%40128_merge-layer-norm_gates2-gridmlp_use-gate-force-head_wd%401e-3-grad-clip%40100_lin-ref-e%404.yml">base.yml</a></td>
|
| 47 |
-
<td></td>
|
| 48 |
-
</tr>
|
| 49 |
-
</table>
|
| 50 |
-
-->
|
| 51 |
|
|
|
|
| 52 |
|
| 53 |
-
|
| 54 |
-
<!--
|
| 55 |
-
Training consists of (1) direct pre-training and (2) gradient fine-tuning initialized from (1).
|
| 56 |
-
<table>
|
| 57 |
-
<tr style="background-color: #f0f0f0;">
|
| 58 |
-
<td><strong>Model</strong></td>
|
| 59 |
-
<td><strong>Training data</strong></td>
|
| 60 |
-
<td><strong>Config</strong></td>
|
| 61 |
-
<td><strong>Checkpoint</strong></td>
|
| 62 |
-
</tr>
|
| 63 |
-
<tr>
|
| 64 |
-
<td>EquiformerV3 (direct pre-training)</td>
|
| 65 |
-
<td>MPtrj</td>
|
| 66 |
-
<td>
|
| 67 |
-
<a href="https://github.com/atomicarchitects/equiformer_v3/blob/main/experimental/configs/omat24/mptrj/experiments/direct/equiformer_v3_N%407_L%404_attn-hidden%4032_rbf%4010_max-neighbors%40300_attn-grid%4014-8_ffn-grid%4014_use-gate-force-head_merge-layer-norm_epochs%4070-bs%40512-wd%401e-3-beta2%400.95_dens-p%400.5-std%400.025-r%400.5-w%4010-strict-max-r%400.75-no-stress.yml">
|
| 68 |
-
direct.yml
|
| 69 |
-
</a>
|
| 70 |
-
</td>
|
| 71 |
-
<td></td>
|
| 72 |
-
</tr>
|
| 73 |
-
<tr>
|
| 74 |
-
<td>EquiformerV3 (gradient fine-tuning)</td>
|
| 75 |
-
<td>MPtrj</td>
|
| 76 |
-
<td>
|
| 77 |
-
<a href="https://github.com/atomicarchitects/equiformer_v3/blob/main/experimental/configs/omat24/mptrj/experiments/gradient/equiformer_v3_grad-finetune_N%407_L%404_attn-hidden%4032_rbf%4010_max-neighbors%40300_attn-grid%4014-8_ffn-grid%4014_pt-reg-dens-no-stress-strict-max-r%400.75-ft-no-reg_lr%400-5e-5-epochs%4010-bs%4064x8-wd%401e-3-beta2%400.95.yml">
|
| 78 |
-
gradient.yml
|
| 79 |
-
</a>
|
| 80 |
-
</td>
|
| 81 |
-
<td></td>
|
| 82 |
-
</tr>
|
| 83 |
-
</table>
|
| 84 |
-
-->
|
| 85 |
<table>
|
| 86 |
<tr style="background-color: #f0f0f0;">
|
| 87 |
<td><strong>Model</strong></td>
|
|
@@ -99,11 +40,9 @@ Training consists of (1) direct pre-training and (2) gradient fine-tuning initia
|
|
| 99 |
</tr>
|
| 100 |
</table>
|
| 101 |
|
|
|
|
|
|
|
| 102 |
|
| 103 |
-
<h2 id="oam">OMat24 → MPtrj and sAlex</h2>
|
| 104 |
-
Training consists of (1) direct pre-training on OMat24,
|
| 105 |
-
(2) gradient fine-tuning on OMat24 initialized from (1), and
|
| 106 |
-
(3) gradient fine-tuning on MPtrj and sAlex initialized from (2).
|
| 107 |
<table>
|
| 108 |
<tr style="background-color: #f0f0f0;">
|
| 109 |
<td><strong>Model</strong></td>
|
|
@@ -153,4 +92,17 @@ Training consists of (1) direct pre-training on OMat24,
|
|
| 153 |
</a>
|
| 154 |
</td>
|
| 155 |
</tr>
|
| 156 |
-
</table>
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
---
|
| 2 |
license: mit
|
| 3 |
+
pipeline_tag: graph-ml
|
| 4 |
---
|
| 5 |
|
| 6 |
<h1 align="center" style="font-size: 24px;">EquiformerV3:<br>Scaling Efficient, Expressive, and General SE(3)-Equivariant Graph Attention Transformers</h1>
|
| 7 |
|
| 8 |
+
<p align="center">
|
|
|
|
|
|
|
|
|
|
| 9 |
<a href="https://github.com/atomicarchitects/equiformer_v3" style="color: #1a73e8; font-weight: bold; font-size: 20px;">Code</a> |
|
| 10 |
<a href="https://arxiv.org/abs/2604.09130" style="color: #1a73e8; font-weight: bold; font-size: 20px;">Paper</a>
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 11 |
</p>
|
| 12 |
|
| 13 |
+
This repository contains the checkpoints for **EquiformerV3**, the third generation of the $SE(3)$-equivariant graph attention Transformer. EquiformerV3 is designed to advance efficiency, expressivity, and generality in 3D atomistic modeling.
|
| 14 |
|
| 15 |
+
Building on EquiformerV2, this version introduces software optimizations achieving a $1.75\times$ speedup, structural improvements like equivariant merged layer normalization and smooth-cutoff attention, and SwiGLU-$S^2$ activations to incorporate many-body interactions while preserving strict equivariance. EquiformerV3 achieves state-of-the-art results on benchmarks including OC20, OMat24, and Matbench Discovery.
|
| 16 |
|
| 17 |
+
Please refer to the [official GitHub repository](https://github.com/atomicarchitects/equiformer_v3) for detailed instructions on environment setup and usage.
|
| 18 |
|
| 19 |
+
<p align="center">
|
| 20 |
+
<img width="50%" height="50%" src="https://cdn-uploads.huggingface.co/production/uploads/64948a4a8d5ff0dd776655fe/03TPndezDyUw4FcfTBk4n.png">
|
| 21 |
+
</p>
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 22 |
|
| 23 |
+
## Checkpoints
|
| 24 |
|
| 25 |
+
### MPtrj
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 26 |
<table>
|
| 27 |
<tr style="background-color: #f0f0f0;">
|
| 28 |
<td><strong>Model</strong></td>
|
|
|
|
| 40 |
</tr>
|
| 41 |
</table>
|
| 42 |
|
| 43 |
+
### OMat24 → MPtrj and sAlex
|
| 44 |
+
Training consists of (1) direct pre-training on OMat24, (2) gradient fine-tuning on OMat24 initialized from (1), and (3) gradient fine-tuning on MPtrj and sAlex initialized from (2).
|
| 45 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 46 |
<table>
|
| 47 |
<tr style="background-color: #f0f0f0;">
|
| 48 |
<td><strong>Model</strong></td>
|
|
|
|
| 92 |
</a>
|
| 93 |
</td>
|
| 94 |
</tr>
|
| 95 |
+
</table>
|
| 96 |
+
|
| 97 |
+
## Citation
|
| 98 |
+
|
| 99 |
+
If you find this work helpful, please consider citing:
|
| 100 |
+
|
| 101 |
+
```bibtex
|
| 102 |
+
@article{equiformer_v3,
|
| 103 |
+
title={EquiformerV3: Scaling Efficient, Expressive, and General SE(3)-Equivariant Graph Attention Transformers},
|
| 104 |
+
author={Yi-Lun Liao and Alexander J. Hoffman and Sabrina C. Shen and Alexandre Duval and Sam Walton Norwood and Tess Smidt},
|
| 105 |
+
journal={arXiv preprint arXiv:2604.09130},
|
| 106 |
+
year={2026}
|
| 107 |
+
}
|
| 108 |
+
```
|