|
|
--- |
|
|
license: apache-2.0 |
|
|
library_name: pytorch |
|
|
tags: |
|
|
- materials-science |
|
|
- crystal-structures |
|
|
- solid-state-batteries |
|
|
- representation-learning |
|
|
- screening |
|
|
model-index: |
|
|
- name: SSB Screening Model (RTX6000x2) |
|
|
results: |
|
|
- task: |
|
|
type: text-classification |
|
|
name: Screening Proxy (3-class) |
|
|
metrics: |
|
|
- type: accuracy |
|
|
value: 0.8118937 |
|
|
- type: f1 |
|
|
value: 0.8060277 |
|
|
- type: precision |
|
|
value: 0.7671543 |
|
|
- type: recall |
|
|
value: 0.8694215 |
|
|
- type: val_loss |
|
|
value: 0.2856999 |
|
|
--- |
|
|
|
|
|
# SSB Screening Model (RTX6000x2) |
|
|
|
|
|
## Model Summary |
|
|
This model is a lightweight MLP classifier trained on NPZ-encoded inorganic crystal structure features for solid-state battery (SSB) screening proxies. It is intended to prioritize candidate structures, not to replace DFT or experimental validation. |
|
|
|
|
|
- **Architecture**: MLP (input_dim=144, hidden_dims=[512, 256, 128], dropout variable by sweep) |
|
|
- **Output**: 3-class classification proxy for screening tasks |
|
|
- **Training Regime**: supervised training on curated NPZ dataset with class-weighted loss |
|
|
- **Best checkpoint**: `checkpoint_epoch45.pt` (lowest observed val_loss in logs) |
|
|
|
|
|
## Intended Use |
|
|
- **Primary**: ranking/prioritization of SSB electrolyte candidates |
|
|
- **Not intended**: absolute property prediction or experimental ground truth replacement |
|
|
|
|
|
## Training Data |
|
|
- **Dataset**: `ssb_npz_v1` (curated NPZ features) |
|
|
- **Split**: 80/10/10 (train/val/test) |
|
|
- **Features**: composition + lattice + derived scalar statistics (144-dim) |
|
|
|
|
|
## Evaluation |
|
|
Metrics from the latest run summary: |
|
|
- **Val loss**: 0.2857 |
|
|
- **Val accuracy**: 0.8119 |
|
|
- **Holdout accuracy**: 0.8096 |
|
|
- **F1**: 0.8060 |
|
|
- **Precision**: 0.7672 |
|
|
- **Recall**: 0.8694 |
|
|
|
|
|
## Limitations |
|
|
- The model is a proxy classifier; it does not predict ground-truth physical properties. |
|
|
- Performance is tied to the training distribution of `ssb_npz_v1`. |
|
|
- Chemical regimes underrepresented in the training set may be poorly ranked. |
|
|
|
|
|
## Training Configuration (abridged) |
|
|
- Optimizer: AdamW |
|
|
- LR: sweep (best around ~3e-4) |
|
|
- Weight decay: sweep (0.005–0.02) |
|
|
- Scheduler: cosine |
|
|
- Batch size: sweep (128–512) |
|
|
- Epochs: sweep (20–60) |
|
|
- Gradient accumulation: sweep (1–4) |
|
|
|
|
|
## Citation |
|
|
If you use this model, please cite the dataset and training pipeline from the Nexa_compute repository. |
|
|
|