| --- |
| license: apache-2.0 |
| library_name: pytorch |
| tags: |
| - materials-science |
| - crystal-structures |
| - solid-state-batteries |
| - representation-learning |
| - screening |
| model-index: |
| - name: SSB Screening Model (RTX6000x2) |
| results: |
| - task: |
| type: text-classification |
| name: Screening Proxy (3-class) |
| metrics: |
| - type: accuracy |
| value: 0.8118937 |
| - type: f1 |
| value: 0.8060277 |
| - type: precision |
| value: 0.7671543 |
| - type: recall |
| value: 0.8694215 |
| - type: val_loss |
| value: 0.2856999 |
| --- |
| |
| # SSB Screening Model (RTX6000x2) |
|
|
| ## Model Summary |
| This model is a lightweight MLP classifier trained on NPZ-encoded inorganic crystal structure features for solid-state battery (SSB) screening proxies. It is intended to prioritize candidate structures, not to replace DFT or experimental validation. |
|
|
| - **Architecture**: MLP (input_dim=144, hidden_dims=[512, 256, 128], dropout variable by sweep) |
| - **Output**: 3-class classification proxy for screening tasks |
| - **Training Regime**: supervised training on curated NPZ dataset with class-weighted loss |
| - **Best checkpoint**: `checkpoint_epoch45.pt` (lowest observed val_loss in logs) |
| |
| ## Intended Use |
| - **Primary**: ranking/prioritization of SSB electrolyte candidates |
| - **Not intended**: absolute property prediction or experimental ground truth replacement |
| |
| ## Training Data |
| - **Dataset**: `ssb_npz_v1` (curated NPZ features) |
| - **Split**: 80/10/10 (train/val/test) |
| - **Features**: composition + lattice + derived scalar statistics (144-dim) |
| |
| ## Evaluation |
| Metrics from the latest run summary: |
| - **Val loss**: 0.2857 |
| - **Val accuracy**: 0.8119 |
| - **Holdout accuracy**: 0.8096 |
| - **F1**: 0.8060 |
| - **Precision**: 0.7672 |
| - **Recall**: 0.8694 |
| |
| ## Limitations |
| - The model is a proxy classifier; it does not predict ground-truth physical properties. |
| - Performance is tied to the training distribution of `ssb_npz_v1`. |
| - Chemical regimes underrepresented in the training set may be poorly ranked. |
| |
| ## Training Configuration (abridged) |
| - Optimizer: AdamW |
| - LR: sweep (best around ~3e-4) |
| - Weight decay: sweep (0.005–0.02) |
| - Scheduler: cosine |
| - Batch size: sweep (128–512) |
| - Epochs: sweep (20–60) |
| - Gradient accumulation: sweep (1–4) |
| |
| ## Citation |
| If you use this model, please cite the dataset and training pipeline from the Nexa_compute repository. |
|
|