prediction_task / README.md
sumuks's picture
Upload README.md with huggingface_hub
860a45b verified
---
tags:
- ocr
- peer-review
- classification
license: other
language:
- en
---
# Popper Reviews — Private Prediction Subset
## Dataset Summary
This repository exposes an 80/20 train/test split tailored for acceptance prediction tasks. Each example contains:
- `paper_text`: OCR’d manuscript text.
- `anonymized_paper_text`: the same text with the author block removed (starts at the abstract).
- `decision_label`: normalized `accept`/`reject` outcome.
- `decision_text`: original decision string when available.
- `average_review_score`: mean of numeric reviewer ratings extracted from the Popper review JSON files.
Source corpora: Popper’s ICLR, TMLR, and Nature review dumps. Only papers with an explicit accept/reject decision are included. Reference lists are removed from `anonymized_paper_text` to focus on the manuscript narrative.
## Splits
| Split | Records |
| --- | --- |
| train | 1 884 |
| test | 472 |
Splits are stratified with an 80/20 ratio using a fixed random seed (42).
## Usage
```python
from datasets import load_dataset
data = load_dataset("popper-spiralworks/prediction_task", split="train", token=token)
print(data[0]["decision_label"], data[0]["average_review_score"])
```
## Processing Notes
- OCR text comes from DeepSeek-OCR via Popper (`metadata.backend = deepseek` when available).
- Average scores are computed by parsing the numeric prefix of each reviewer `rating` field.
- Non-numeric or missing ratings are ignored during averaging.
- Additional review metadata and reviewer comments are available in the public dataset [`sumuks/research_papers_with_reviews_ocr`](https://huggingface.co/datasets/sumuks/research_papers_with_reviews_ocr).
## Attribution
When using this dataset, please credit the original venues (ICLR, TMLR, Nature) and cite the Popper project. Access to this repository is restricted to the Popper Spiralworks collaboration.