File size: 2,419 Bytes
a2abe63
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
---
license: apache-2.0
library_name: pytorch
tags:
  - materials-science
  - crystal-structures
  - solid-state-batteries
  - representation-learning
  - screening
model-index:
  - name: SSB Screening Model (RTX6000x2)
    results:
      - task:
          type: text-classification
          name: Screening Proxy (3-class)
        metrics:
          - type: accuracy
            value: 0.8118937
          - type: f1
            value: 0.8060277
          - type: precision
            value: 0.7671543
          - type: recall
            value: 0.8694215
          - type: val_loss
            value: 0.2856999
---

# SSB Screening Model (RTX6000x2)

## Model Summary
This model is a lightweight MLP classifier trained on NPZ-encoded inorganic crystal structure features for solid-state battery (SSB) screening proxies. It is intended to prioritize candidate structures, not to replace DFT or experimental validation.

- **Architecture**: MLP (input_dim=144, hidden_dims=[512, 256, 128], dropout variable by sweep)
- **Output**: 3-class classification proxy for screening tasks
- **Training Regime**: supervised training on curated NPZ dataset with class-weighted loss
- **Best checkpoint**: `checkpoint_epoch45.pt` (lowest observed val_loss in logs)

## Intended Use
- **Primary**: ranking/prioritization of SSB electrolyte candidates
- **Not intended**: absolute property prediction or experimental ground truth replacement

## Training Data
- **Dataset**: `ssb_npz_v1` (curated NPZ features)
- **Split**: 80/10/10 (train/val/test)
- **Features**: composition + lattice + derived scalar statistics (144-dim)

## Evaluation
Metrics from the latest run summary:
- **Val loss**: 0.2857
- **Val accuracy**: 0.8119
- **Holdout accuracy**: 0.8096
- **F1**: 0.8060
- **Precision**: 0.7672
- **Recall**: 0.8694

## Limitations
- The model is a proxy classifier; it does not predict ground-truth physical properties.
- Performance is tied to the training distribution of `ssb_npz_v1`.
- Chemical regimes underrepresented in the training set may be poorly ranked.

## Training Configuration (abridged)
- Optimizer: AdamW
- LR: sweep (best around ~3e-4)
- Weight decay: sweep (0.005–0.02)
- Scheduler: cosine
- Batch size: sweep (128–512)
- Epochs: sweep (20–60)
- Gradient accumulation: sweep (1–4)

## Citation
If you use this model, please cite the dataset and training pipeline from the Nexa_compute repository.