|
|
--- |
|
|
license: apache-2.0 |
|
|
library_name: transformers |
|
|
pipeline_tag: text-classification |
|
|
tags: |
|
|
- scibert |
|
|
- data-paper-classification |
|
|
- scholarly-papers |
|
|
- binary-classification |
|
|
base_model: allenai/scibert_scivocab_uncased |
|
|
metrics: |
|
|
- accuracy |
|
|
- f1 |
|
|
model-index: |
|
|
- name: scibert-data-paper |
|
|
results: |
|
|
- task: |
|
|
type: text-classification |
|
|
name: Data Paper Classification |
|
|
metrics: |
|
|
- name: Edge Case Accuracy |
|
|
type: accuracy |
|
|
value: 1 |
|
|
- name: Mean Confidence |
|
|
type: accuracy |
|
|
value: 0.94 |
|
|
--- |
|
|
|
|
|
# SciBERT Data-Paper Classifier |
|
|
|
|
|
A fine-tuned [SciBERT](https://huggingface.co/allenai/scibert_scivocab_uncased) model for binary classification of scholarly papers as **data papers** (datasets, databases, atlases, benchmarks) vs **non-data papers** (methods, reviews, surveys, clinical trials). |
|
|
|
|
|
Built for the [DataRank Portal](https://github.com/zehrakorkusuz/sindex-portal) — a data-sharing influence engine using Personalized PageRank on citation graphs. |
|
|
|
|
|
## Usage |
|
|
|
|
|
```python |
|
|
from transformers import pipeline |
|
|
|
|
|
clf = pipeline("text-classification", model="zehralx/scibert-data-paper", top_k=None, device=-1) |
|
|
result = clf("MIMIC-III, a freely accessible critical care database") |
|
|
# [{'label': 'LABEL_1', 'score': 0.9519}, {'label': 'LABEL_0', 'score': 0.0481}] |
|
|
# LABEL_1 = data paper, LABEL_0 = not data paper |
|
|
``` |
|
|
|
|
|
## Model Details |
|
|
|
|
|
| Property | Value | |
|
|
|----------|-------| |
|
|
| Base model | `allenai/scibert_scivocab_uncased` | |
|
|
| Architecture | BertForSequenceClassification (12 layers, 768 hidden, 12 heads) | |
|
|
| Parameters | ~110M | |
|
|
| Max tokens | 512 | |
|
|
| Output | Binary: `data_paper` (1) / `not_data_paper` (0) | |
|
|
| Inference | CPU (no GPU required) | |
|
|
|
|
|
|
|
|
|
|
|
## Training |
|
|
|
|
|
[Train Data](https://www.kaggle.com/datasets/zehrakorkusuz/labeling-4k-datasets-with-gemini-flash-2-0) |
|
|
|
|
|
Two-phase continued fine-tuning: |
|
|
|
|
|
1. **Phase 1**: 5 epochs, learning rate 2e-5 |
|
|
2. **Phase 2**: 3 epochs, learning rate 5e-6 (lower LR for refinement) |
|
|
|
|
|
| Hyperparameter | Value | |
|
|
|----------------|-------| |
|
|
| Batch size | 24 | |
|
|
| Label smoothing | 0.1 | |
|
|
| Edge case weight | 5x | |
|
|
| Mixed precision | FP16 | |
|
|
|
|
|
## Evaluation |
|
|
|
|
|
Tested on 38 curated edge cases spanning diverse categories: |
|
|
|
|
|
| Category | Examples | Correctly classified | |
|
|
|----------|----------|---------------------| |
|
|
| Data papers | UniProt, GTEx, ImageNet, TCGA, MIMIC-III, UK Biobank | All | |
|
|
| Non-data papers | Methods, reviews, surveys, perspectives, protocols | All | |
|
|
|
|
|
- **Edge case accuracy**: 100% (38/38) |
|
|
- **Confidence range**: 0.80 - 0.96 |
|
|
- **Mean confidence**: 0.94 |
|
|
|
|
|
## Input Format |
|
|
|
|
|
Concatenated `title + abstract`, truncated to 512 tokens. The model works well with title-only input when abstracts are unavailable. |
|
|
|
|
|
## Limitations |
|
|
|
|
|
- Trained primarily on biomedical/life sciences papers; may underperform on other domains |
|
|
- Binary classification only (no multi-class dataset subtypes) |
|
|
- Confidence may be lower for interdisciplinary papers that mix methods and data contributions |
|
|
|
|
|
## Citation |
|
|
|
|
|
```bibtex |
|
|
@misc{scibert-data-paper-2026, |
|
|
title={SciBERT Data-Paper Classifier}, |
|
|
author={Zehra Korkusuz, Kuan-Lin Huang}, |
|
|
year={2026}, |
|
|
url={https://huggingface.co/zehralx/scibert-data-paper} |
|
|
} |
|
|
``` |