SiatBioInf's picture
Update README.md
55afedd verified
metadata
license: cc-by-4.0
language:
  - en
task_categories:
  - other
tags:
  - biology
  - single-cell
  - scRNA-seq
  - h5ad
  - cancer
  - tumor
  - benchmark
  - bioinformatics
size_categories:
  - 1M<n<10M

SingleCell-Unseen-Benchmark

Overview

SingleCell-Unseen-Benchmark is a large-scale unseen single-cell transcriptomic benchmark designed to systematically evaluate foundation models on cell identification and cell type tracing tasks.
The benchmark covers tumor, stem, neural, and normal cell populations, with a particular emphasis on unseen data distributions, including rare cell types, cross-dataset generalization, and heterogeneous tumor states.

In addition to curated datasets, this repository provides standardized benchmark results for multiple single-cell foundation models, enabling transparent and reproducible comparison.


Dataset Collection

Tumor Cells

  • Source: GEO
  • Cancer types: 21
  • Samples: 2,225
  • Cells: 1,645,662
  • Cell states: Primary tumors, metastases, circulating tumor cells (CTCs)

Stem Cells

  • Source: CELLxGENE
  • Datasets: 5
  • Cells: 325,092
  • Stem cell types: 4

Neural Cells

  • Source: CELLxGENE
  • Datasets: 1
  • Cells: 423,707
  • Neural cell types: 6

Normal Cells

  • Source: CELLxGENE
  • Datasets: 7
  • Cells: 1,838,991
  • Normal cell types: 10

Preprocessing

  • All genes were mapped to HGNC symbols
  • Cells with fewer than 200 detected genes were removed
  • Expression matrices are stored in AnnData (.h5ad) format

Cell Type and Malignancy Annotation Strategy

Tumor cells derived from GEO were re-identified using a consensus workflow:

  1. Lineage-level screening based on CancerSCEM 2.0 marker genes
  2. Malignancy confirmation using inferCNV

CELLxGENE-derived datasets retain their original annotations.

This strategy ensures consistent tumor labeling while minimizing dataset-specific bias.


Downstream Benchmark Tasks

The benchmark evaluates foundation models across multiple biologically meaningful tasks:

Category Task Prediction Type
Tumor Tumor cell identification Binary
Tumor Primary site tracing Multi-class
Stem Stem cell identification Binary
Stem Stem cell subtype classification Multi-class
Neural Neural cell identification Binary
Neural Neural cell subtype classification Multi-class

Models take high-dimensional cell embeddings as input and perform prediction using lightweight downstream classifiers, isolating representation quality from classifier complexity.


Benchmark Models

The following single-cell foundation models are evaluated:

  • Geneformer
  • scFoundation
  • scGPT
  • UCE
  • scLONG

Evaluation Metrics

  • Binary classification tasks

    • Accuracy
    • Precision
    • Recall
    • F1-score
  • Multi-class classification tasks

    • Accuracy
    • Macro-Precision
    • Macro-Recall
    • Macro-F1

Data Format and Access

Data Files

All datasets are provided in AnnData (.h5ad) format.

Note
.h5ad files are not natively supported by the Hugging Face Dataset Viewer.
Users are expected to download the files and load them locally using standard single-cell analysis tools such as Scanpy or Seurat.

Benchmark Results

In addition to raw datasets, we provide complete benchmark evaluation results under the results/ directory.

Design Rationale

  • by_model/
    Provides a model-centric view, facilitating analysis of how a single model performs across different tasks.

  • by_task/
    Provides a task-centric view, enabling direct comparison of multiple models on the same task.

Both views contain identical information and are provided to improve usability, clarity, and reproducibility.


Intended Use

This benchmark is intended for:

  • Evaluating generalization and robustness of single-cell foundation models
  • Studying tumor cell identification and origin tracing under unseen conditions
  • Benchmarking representation quality across diverse biological contexts

The dataset is not intended for clinical decision-making.


Citation

If you use this dataset or benchmark in your work, please cite:

Contact

For questions, issues, or suggestions, please open an issue on the Hugging Face repository.