Graph Machine Learning
yilunliao nielsr HF Staff commited on
Commit
633adcd
·
1 Parent(s): c04da5a

Add pipeline tag and improve model card discovery (#1)

Browse files

- Add pipeline tag and improve model card discovery (087f70d8d9666596ea992b77b4802cd578f1cbff)


Co-authored-by: Niels Rogge <nielsr@users.noreply.huggingface.co>

Files changed (1) hide show
  1. README.md +26 -74
README.md CHANGED
@@ -1,87 +1,28 @@
1
  ---
2
  license: mit
 
3
  ---
4
 
5
  <h1 align="center" style="font-size: 24px;">EquiformerV3:<br>Scaling Efficient, Expressive, and General SE(3)-Equivariant Graph Attention Transformers</h1>
6
 
7
- <!--
8
- # **[Code](https://github.com/atomicarchitects/equiformer_v3)** | **[Paper]()**
9
- -->
10
-
11
  <a href="https://github.com/atomicarchitects/equiformer_v3" style="color: #1a73e8; font-weight: bold; font-size: 20px;">Code</a> |
12
  <a href="https://arxiv.org/abs/2604.09130" style="color: #1a73e8; font-weight: bold; font-size: 20px;">Paper</a>
13
-
14
- This repository contains the checkpoints of the work "EquiformerV3: Scaling Efficient, Expressive, and General SE(3)-Equivariant Graph Attention Transformers".
15
- Please refer to the [code](https://github.com/atomicarchitects/equiformer_v3) for detailed description of usage.
16
-
17
- <p align="center">
18
- <img width="50%" height="50%" src="https://cdn-uploads.huggingface.co/production/uploads/64948a4a8d5ff0dd776655fe/03TPndezDyUw4FcfTBk4n.png"?
19
  </p>
20
 
 
21
 
 
22
 
 
23
 
24
- ## Content ##
25
- <!--
26
- 0. [OC20](#oc20)
27
- -->
28
- 0. [MPtrj](#mptrj)
29
- 0. [OMat24 → MPtrj and sAlex](#oam)
30
-
31
-
32
- <!--
33
- <h2 id="oc20">OC20</h2>
34
-
35
- <table>
36
- <tr style="background-color: #f0f0f0;">
37
- <td><strong>Model</strong></td>
38
- <td><strong>Training data</strong></td>
39
- <td><strong>Config</strong></td>
40
- <td><strong>Checkpoint</strong></td>
41
- </tr>
42
-
43
- <tr>
44
- <td>EquiformerV3 (91M)</td>
45
- <td>OC20 S2EF-2M</td>
46
- <td><a href="https://github.com/atomicarchitects/equiformer_v3/blob/main/experimental/configs/oc20/2M/equiformer_v3/experiments/base_N%408-L%406-C%40128-attn-hidden%4064-ffn%40512-envelope-num-rbf%40128_merge-layer-norm_gates2-gridmlp_use-gate-force-head_wd%401e-3-grad-clip%40100_lin-ref-e%404.yml">base.yml</a></td>
47
- <td></td>
48
- </tr>
49
- </table>
50
- -->
51
 
 
52
 
53
- <h2 id="mptrj">MPtrj</h2>
54
- <!--
55
- Training consists of (1) direct pre-training and (2) gradient fine-tuning initialized from (1).
56
- <table>
57
- <tr style="background-color: #f0f0f0;">
58
- <td><strong>Model</strong></td>
59
- <td><strong>Training data</strong></td>
60
- <td><strong>Config</strong></td>
61
- <td><strong>Checkpoint</strong></td>
62
- </tr>
63
- <tr>
64
- <td>EquiformerV3 (direct pre-training)</td>
65
- <td>MPtrj</td>
66
- <td>
67
- <a href="https://github.com/atomicarchitects/equiformer_v3/blob/main/experimental/configs/omat24/mptrj/experiments/direct/equiformer_v3_N%407_L%404_attn-hidden%4032_rbf%4010_max-neighbors%40300_attn-grid%4014-8_ffn-grid%4014_use-gate-force-head_merge-layer-norm_epochs%4070-bs%40512-wd%401e-3-beta2%400.95_dens-p%400.5-std%400.025-r%400.5-w%4010-strict-max-r%400.75-no-stress.yml">
68
- direct.yml
69
- </a>
70
- </td>
71
- <td></td>
72
- </tr>
73
- <tr>
74
- <td>EquiformerV3 (gradient fine-tuning)</td>
75
- <td>MPtrj</td>
76
- <td>
77
- <a href="https://github.com/atomicarchitects/equiformer_v3/blob/main/experimental/configs/omat24/mptrj/experiments/gradient/equiformer_v3_grad-finetune_N%407_L%404_attn-hidden%4032_rbf%4010_max-neighbors%40300_attn-grid%4014-8_ffn-grid%4014_pt-reg-dens-no-stress-strict-max-r%400.75-ft-no-reg_lr%400-5e-5-epochs%4010-bs%4064x8-wd%401e-3-beta2%400.95.yml">
78
- gradient.yml
79
- </a>
80
- </td>
81
- <td></td>
82
- </tr>
83
- </table>
84
- -->
85
  <table>
86
  <tr style="background-color: #f0f0f0;">
87
  <td><strong>Model</strong></td>
@@ -99,11 +40,9 @@ Training consists of (1) direct pre-training and (2) gradient fine-tuning initia
99
  </tr>
100
  </table>
101
 
 
 
102
 
103
- <h2 id="oam">OMat24 → MPtrj and sAlex</h2>
104
- Training consists of (1) direct pre-training on OMat24,
105
- (2) gradient fine-tuning on OMat24 initialized from (1), and
106
- (3) gradient fine-tuning on MPtrj and sAlex initialized from (2).
107
  <table>
108
  <tr style="background-color: #f0f0f0;">
109
  <td><strong>Model</strong></td>
@@ -153,4 +92,17 @@ Training consists of (1) direct pre-training on OMat24,
153
  </a>
154
  </td>
155
  </tr>
156
- </table>
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: mit
3
+ pipeline_tag: graph-ml
4
  ---
5
 
6
  <h1 align="center" style="font-size: 24px;">EquiformerV3:<br>Scaling Efficient, Expressive, and General SE(3)-Equivariant Graph Attention Transformers</h1>
7
 
8
+ <p align="center">
 
 
 
9
  <a href="https://github.com/atomicarchitects/equiformer_v3" style="color: #1a73e8; font-weight: bold; font-size: 20px;">Code</a> |
10
  <a href="https://arxiv.org/abs/2604.09130" style="color: #1a73e8; font-weight: bold; font-size: 20px;">Paper</a>
 
 
 
 
 
 
11
  </p>
12
 
13
+ This repository contains the checkpoints for **EquiformerV3**, the third generation of the $SE(3)$-equivariant graph attention Transformer. EquiformerV3 is designed to advance efficiency, expressivity, and generality in 3D atomistic modeling.
14
 
15
+ Building on EquiformerV2, this version introduces software optimizations achieving a $1.75\times$ speedup, structural improvements like equivariant merged layer normalization and smooth-cutoff attention, and SwiGLU-$S^2$ activations to incorporate many-body interactions while preserving strict equivariance. EquiformerV3 achieves state-of-the-art results on benchmarks including OC20, OMat24, and Matbench Discovery.
16
 
17
+ Please refer to the [official GitHub repository](https://github.com/atomicarchitects/equiformer_v3) for detailed instructions on environment setup and usage.
18
 
19
+ <p align="center">
20
+ <img width="50%" height="50%" src="https://cdn-uploads.huggingface.co/production/uploads/64948a4a8d5ff0dd776655fe/03TPndezDyUw4FcfTBk4n.png">
21
+ </p>
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
22
 
23
+ ## Checkpoints
24
 
25
+ ### MPtrj
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
26
  <table>
27
  <tr style="background-color: #f0f0f0;">
28
  <td><strong>Model</strong></td>
 
40
  </tr>
41
  </table>
42
 
43
+ ### OMat24 → MPtrj and sAlex
44
+ Training consists of (1) direct pre-training on OMat24, (2) gradient fine-tuning on OMat24 initialized from (1), and (3) gradient fine-tuning on MPtrj and sAlex initialized from (2).
45
 
 
 
 
 
46
  <table>
47
  <tr style="background-color: #f0f0f0;">
48
  <td><strong>Model</strong></td>
 
92
  </a>
93
  </td>
94
  </tr>
95
+ </table>
96
+
97
+ ## Citation
98
+
99
+ If you find this work helpful, please consider citing:
100
+
101
+ ```bibtex
102
+ @article{equiformer_v3,
103
+ title={EquiformerV3: Scaling Efficient, Expressive, and General SE(3)-Equivariant Graph Attention Transformers},
104
+ author={Yi-Lun Liao and Alexander J. Hoffman and Sabrina C. Shen and Alexandre Duval and Sam Walton Norwood and Tess Smidt},
105
+ journal={arXiv preprint arXiv:2604.09130},
106
+ year={2026}
107
+ }
108
+ ```