File size: 2,419 Bytes
a2abe63 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 |
---
license: apache-2.0
library_name: pytorch
tags:
- materials-science
- crystal-structures
- solid-state-batteries
- representation-learning
- screening
model-index:
- name: SSB Screening Model (RTX6000x2)
results:
- task:
type: text-classification
name: Screening Proxy (3-class)
metrics:
- type: accuracy
value: 0.8118937
- type: f1
value: 0.8060277
- type: precision
value: 0.7671543
- type: recall
value: 0.8694215
- type: val_loss
value: 0.2856999
---
# SSB Screening Model (RTX6000x2)
## Model Summary
This model is a lightweight MLP classifier trained on NPZ-encoded inorganic crystal structure features for solid-state battery (SSB) screening proxies. It is intended to prioritize candidate structures, not to replace DFT or experimental validation.
- **Architecture**: MLP (input_dim=144, hidden_dims=[512, 256, 128], dropout variable by sweep)
- **Output**: 3-class classification proxy for screening tasks
- **Training Regime**: supervised training on curated NPZ dataset with class-weighted loss
- **Best checkpoint**: `checkpoint_epoch45.pt` (lowest observed val_loss in logs)
## Intended Use
- **Primary**: ranking/prioritization of SSB electrolyte candidates
- **Not intended**: absolute property prediction or experimental ground truth replacement
## Training Data
- **Dataset**: `ssb_npz_v1` (curated NPZ features)
- **Split**: 80/10/10 (train/val/test)
- **Features**: composition + lattice + derived scalar statistics (144-dim)
## Evaluation
Metrics from the latest run summary:
- **Val loss**: 0.2857
- **Val accuracy**: 0.8119
- **Holdout accuracy**: 0.8096
- **F1**: 0.8060
- **Precision**: 0.7672
- **Recall**: 0.8694
## Limitations
- The model is a proxy classifier; it does not predict ground-truth physical properties.
- Performance is tied to the training distribution of `ssb_npz_v1`.
- Chemical regimes underrepresented in the training set may be poorly ranked.
## Training Configuration (abridged)
- Optimizer: AdamW
- LR: sweep (best around ~3e-4)
- Weight decay: sweep (0.005–0.02)
- Scheduler: cosine
- Batch size: sweep (128–512)
- Epochs: sweep (20–60)
- Gradient accumulation: sweep (1–4)
## Citation
If you use this model, please cite the dataset and training pipeline from the Nexa_compute repository.
|