Dataset Viewer
Auto-converted to Parquet Duplicate
candidate_id
stringlengths
4
10
tier
stringclasses
3 values
true_label
int64
0
2
mp-1184785
hard_negative
0
mp-541637
hard_negative
0
mp-1063755
hard_negative
0
mp-20360
hard_negative
0
mp-1184704
hard_negative
0
mp-571512
hard_negative
0
mp-1045826
general_inorganic
1
mp-1120743
general_inorganic
1
mp-10709
hard_negative
0
mp-1188006
hard_negative
0
mp-998916
hard_negative
0
mp-1002230
hard_negative
0
mp-2281971
hard_negative
0
mp-976588
hard_negative
0
mp-972912
hard_negative
0
mp-2814
hard_negative
0
mp-1225044
hard_negative
0
mp-978964
hard_negative
0
mp-985301
hard_negative
0
mp-626
hard_negative
0
mp-1185137
hard_negative
0
mp-1088
hard_negative
0
mp-867815
hard_negative
0
mp-862796
hard_negative
0
mp-1018732
hard_negative
0
mp-980663
hard_negative
0
mp-1225704
hard_negative
0
mp-581942
hard_negative
0
mp-1038767
hard_negative
0
mp-1016268
hard_negative
0
mp-1183254
hard_negative
0
mp-680196
general_inorganic
1
mp-974355
hard_negative
0
mp-1071768
hard_negative
0
mp-567359
general_inorganic
1
mp-568348
general_inorganic
1
mp-568705
hard_negative
0
mp-8093
hard_negative
0
mp-1184888
hard_negative
0
mp-1214755
hard_negative
0
mp-23221
general_inorganic
1
mp-1183473
hard_negative
0
mp-1014138
hard_negative
0
mp-1062055
hard_negative
0
mp-999203
hard_negative
0
mp-1540935
general_inorganic
1
mp-1215200
hard_negative
0
mp-12055
hard_negative
0
mp-862851
hard_negative
0
mp-541055
general_inorganic
1
mp-1187215
hard_negative
0
mp-1190042
hard_negative
0
mp-53
hard_negative
0
mp-1186913
hard_negative
0
mp-1104937
hard_negative
0
mp-1186708
hard_negative
0
mp-570539
general_inorganic
1
mp-1094636
hard_negative
0
mp-1215848
hard_negative
0
mp-1185983
hard_negative
0
mp-1197771
hard_negative
0
mp-754197
hard_negative
0
mp-1226533
hard_negative
0
mp-1018075
hard_negative
0
mp-1238909
hard_negative
0
mp-571501
hard_negative
0
mp-2848
hard_negative
0
mp-974603
general_inorganic
1
mp-1187486
hard_negative
0
mp-1409
hard_negative
0
mp-979988
hard_negative
0
mp-1184140
hard_negative
0
mp-862604
hard_negative
0
mp-8939
hard_negative
0
mp-1104805
hard_negative
0
mp-1209131
hard_negative
0
mp-28860
general_inorganic
1
mp-1094034
hard_negative
0
mp-570604
hard_negative
0
mp-1094295
hard_negative
0
mp-1217189
hard_negative
0
mp-1214
hard_negative
0
mp-1224812
hard_negative
0
mp-973065
hard_negative
0
mp-1185922
hard_negative
0
mp-1185771
hard_negative
0
mp-644514
general_inorganic
1
mp-1008820
hard_negative
0
mp-19762
hard_negative
0
mp-1192251
hard_negative
0
mp-21295
hard_negative
0
mp-1225502
hard_negative
0
mp-10192
hard_negative
0
mp-30667
hard_negative
0
mp-1186537
hard_negative
0
mp-1220372
hard_negative
0
mp-570071
hard_negative
0
mp-975882
hard_negative
0
mp-1525632
general_inorganic
1
mp-1183116
hard_negative
0
End of preview. Expand in Data Studio

SSB Pilots Combined (Sanitized)

This dataset packages two production-style pilot outputs for solid-state battery (SSB) candidate screening:

  • inference_pilot/: one full scoring run across a mixed candidate pool.
  • tuning_pilot/: multi-run tuning and repeatability artifacts used to evaluate ranking stability and lift.

The goal is prioritization, not direct property prediction. Use this to shortlist candidates for downstream DFT/experimental validation when budget is limited.

Why this matters

Materials discovery pipelines usually face extreme class imbalance and tight validation capacity. This dataset shows how ranking-based screening can increase hit density in top-K selections versus random sampling, with explicit uncertainty and repeatability context.

What is included

1) Inference pilot (inference_pilot/)

Core files:

  • mixed_inference_input.jsonl: input candidates with candidate_id, composition, and features.
  • ranked_candidates.csv: scored and ranked output table (primary file for downstream usage).
  • evaluation_lookup.csv: labeled lookup used for post-hoc evaluation on the reference subset.
  • top_005permille.csv, top_010permille.csv, top_050permille.csv, top_100permille.csv: precomputed shortlists.
  • screening_metrics.json: machine-readable metrics and confidence sweeps.
  • screening_report.md: human-readable report.

Inference pilot context (from screening_metrics.json):

  • Run timestamp: 2026-02-01T23:00:49.099228+00:00
  • Total scored: 33,000
  • Eval subset size: 13,000
  • Positives in eval subset: 19 (prevalence 0.00146)
  • Throughput: 35,645 candidates/s (CPU inference in this pilot)

Top-K enrichment highlights on labeled eval subset:

  • Top 0.5% (k=165): enrichment 40.25x
  • Top 1% (k=330): precision@k 0.0241, recall@k 0.1053, enrichment 16.49x
  • Top 5% (k=1650): enrichment 5.60x
  • Top 10% (k=3300): enrichment 3.43x

2) Tuning pilot (tuning_pilot/)

Contains multiple timestamped run folders (for example ssb_tuning_20260201T234218Z) with:

  • screening_card.md and screening_card.json: run-level summary and recommended policy.
  • repeatability_summary.json: multi-trial aggregate metrics.
  • trial_*/ and optional flagship/: per-trial scored outputs, shortlists, and metric summaries.

Flagship tuning snapshot (ssb_tuning_20260201T234218Z/screening_card.json):

  • Screened N: 113,000
  • Top 1% lift: 11.54x ± 0.99
  • Top 5% lift: 7.56x ± 0.24
  • Expected hits per 50 experiments: 28.4
  • Repeatability (Jaccard mean): top1 0.207, top5 0.210

Data schema (primary table)

inference_pilot/ranked_candidates.csv columns:

  • candidate_id: unique candidate identifier.
  • composition: composition string.
  • predicted_class: numeric class ID from classifier.
  • predicted_label: human-readable class label.
  • confidence_score: max predicted class probability.
  • positive_probability: probability for positive screening target.
  • rank: 1-based rank (lower is better).
  • rank_percentile: normalized rank in [0, 1].
  • is_eval_ref: whether candidate belongs to labeled evaluation subset.
  • true_label: reference label (mostly for eval subset).
  • tier: source tier metadata.

Recommended usage workflow

  1. Load inference_pilot/ranked_candidates.csv.
  2. Select a top-K budget (for example top 1% or top 5%).
  3. Optionally apply a confidence threshold guided by inference_pilot/screening_metrics.json.
  4. Pull diversity-aware alternatives from tuning pilot shortlists if execution constraints require broader chemistry coverage.
  5. Send shortlisted candidates to downstream simulation or experimental validation.

Minimal example (Python)

import pandas as pd

ranked = pd.read_csv("inference_pilot/ranked_candidates.csv")

# Example: top 1%
k = max(1, int(0.01 * len(ranked)))
shortlist = ranked.nsmallest(k, "rank")

print(shortlist[["candidate_id", "composition", "positive_probability", "confidence_score"]].head(10))

How to interpret metrics correctly

  • lift/enrichment is measured against random selection at the same cutoff.
  • Reported precision/recall values are computed on the labeled reference subset, not guaranteed for every external pool.
  • Class prevalence can shift strongly across new candidate pools; re-benchmark on your own labeled slice before operational deployment.

Sanitization and privacy

This public release is sanitized:

  • Removed all *.log files.
  • Removed run_manifest.json files.
  • Retained only structured artifacts required for analysis (.csv, .json, .jsonl, .md, .parquet).

Intended use

  • Candidate ranking and triage.
  • Policy comparison for top-K vs confidence-threshold filtering.
  • Repeatability analysis of screening behavior across runs.

Out-of-scope use

  • Direct claims of experimental property prediction without further validation.
  • Safety-critical decisions without independent verification.

Provenance

  • Upstream dataset reference: Allanatrix/SSB_Dataset
  • Upstream model reference: Allanatrix/SSB_Screening_Model

Limitations

  • Labels are proxy labels for screening, not full experimental truth.
  • Performance may degrade under domain shift.
  • Repeatability metrics summarize pilot conditions and may differ under different feature pipelines or candidate distributions.
Downloads last month
21