psiformer_torch / README.md
jorgemunozl's picture
Create README.md
dfa2a34 verified
---
license: apache-2.0
datasets:
- bigai/TongSIM-Asset
language:
- en
metrics:
- exact_match
new_version: zai-org/GLM-4.7
pipeline_tag: reinforcement-learning
library_name: transformers
tags:
- physics
- chemistry
- deepmind
---
# PsiFormer Checkpoint: Hydrogen → Oxygen
This repository contains pretrained **PsiFormer** checkpoints for electronic-structure modeling across atomic systems ranging from **Hydrogen (Z=1)** to **Oxygen (Z=8)**.
The model is designed for **variational quantum Monte Carlo (VMC)**–style wavefunction modeling, with a Transformer-based architecture that captures electron–electron correlations efficiently and scalably.
---
## Model Overview
- **Architecture**: PsiFormer (Transformer-based wavefunction ansatz)
- **Task**: Electronic wavefunction approximation
- **Method**: Variational Monte Carlo (VMC)
- **Atomic range**: Hydrogen → Oxygen
- **Framework**: PyTorch
- **Precision**: FP32 (unless otherwise specified)
The model outputs parameters of a many-body wavefunction that can be used to estimate ground-state energies and other observables via Monte Carlo sampling.
---
## Training Details
- **Systems**: Isolated atoms with atomic numbers Z = 1–8
- **Electrons**: Corresponding neutral configurations
- **Optimization**: Stochastic gradient–based optimization of variational energy
- **Sampling**: Metropolis–Hastings MCMC
- **Objective**: Minimize the expectation value of the Hamiltonian
Exact hyperparameters (learning rate, batch size, number of walkers, etc.) should be considered checkpoint-specific and are documented in the accompanying configuration files when available.
---
## Intended Use
This checkpoint is intended for:
- Initializing PsiFormer models for light atoms
- Transfer learning to larger atoms or small molecules
- Benchmarking neural quantum states
- Research and educational purposes in computational quantum physics
It is **not** intended for production chemistry workflows without further validation.
---
## Example Usage
```python
import torch
from psiformer import PsiFormer
model = PsiFormer(...)
state_dict = torch.load("psiformer_h_to_o.pt", map_location="cpu")
model.load_state_dict(state_dict)
model.eval()
````
Refer to the PsiFormer repository for full examples including sampling and energy evaluation.
---
## Limitations
* Trained only on **isolated atoms**, not molecules
* Accuracy degrades outside the Z = 1–8 range
* Performance depends strongly on sampling quality and optimization setup
* No relativistic or spin–orbit effects included
---
## Citation
If you use this checkpoint in academic work, please cite the corresponding PsiFormer paper or repository.
```bibtex
@misc{psiformer,
title={PsiFormer: Transformer-based Neural Quantum States},
author={...},
year={202X}
}
```
---
## License
Specify the license here (e.g. MIT, Apache 2.0, custom research license).
---
## Contact
For questions, issues, or collaborations, please open an issue in the main PsiFormer repository.