metadata
license: apache-2.0
library_name: pytorch
tags:
- materials-science
- crystal-structures
- solid-state-batteries
- representation-learning
- screening
model-index:
- name: SSB Screening Model (RTX6000x2)
results:
- task:
type: text-classification
name: Screening Proxy (3-class)
metrics:
- type: accuracy
value: 0.8118937
- type: f1
value: 0.8060277
- type: precision
value: 0.7671543
- type: recall
value: 0.8694215
- type: val_loss
value: 0.2856999
SSB Screening Model (RTX6000x2)
Model Summary
This model is a lightweight MLP classifier trained on NPZ-encoded inorganic crystal structure features for solid-state battery (SSB) screening proxies. It is intended to prioritize candidate structures, not to replace DFT or experimental validation.
- Architecture: MLP (input_dim=144, hidden_dims=[512, 256, 128], dropout variable by sweep)
- Output: 3-class classification proxy for screening tasks
- Training Regime: supervised training on curated NPZ dataset with class-weighted loss
- Best checkpoint:
checkpoint_epoch45.pt(lowest observed val_loss in logs)
Intended Use
- Primary: ranking/prioritization of SSB electrolyte candidates
- Not intended: absolute property prediction or experimental ground truth replacement
Training Data
- Dataset:
ssb_npz_v1(curated NPZ features) - Split: 80/10/10 (train/val/test)
- Features: composition + lattice + derived scalar statistics (144-dim)
Evaluation
Metrics from the latest run summary:
- Val loss: 0.2857
- Val accuracy: 0.8119
- Holdout accuracy: 0.8096
- F1: 0.8060
- Precision: 0.7672
- Recall: 0.8694
Limitations
- The model is a proxy classifier; it does not predict ground-truth physical properties.
- Performance is tied to the training distribution of
ssb_npz_v1. - Chemical regimes underrepresented in the training set may be poorly ranked.
Training Configuration (abridged)
- Optimizer: AdamW
- LR: sweep (best around ~3e-4)
- Weight decay: sweep (0.005–0.02)
- Scheduler: cosine
- Batch size: sweep (128–512)
- Epochs: sweep (20–60)
- Gradient accumulation: sweep (1–4)
Citation
If you use this model, please cite the dataset and training pipeline from the Nexa_compute repository.