Neooooo's picture
Update README.md
22396fe verified
---
dataset_info:
features:
- name: title
dtype: string
- name: keywords
sequence: string
- name: messages
list:
- name: content
dtype: string
- name: role
dtype: string
splits:
- name: train
num_bytes: 1851098097.8341584
num_examples: 145064
- name: test
num_bytes: 78063099.39124106
num_examples: 6653
download_size: 626249553
dataset_size: 1929161197.2253995
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
---
# structured_paper_summarization
A **151 k‑example** dataset of chat‐style *prompt → structured abstract* pairs, built from ~19 000 research papers across business, management, information‑systems and social‑science domains. Each example shows the full paper (body text) being summarised into a five‑section Emerald‑style structured abstract (Purpose, Design/methodology/approach, Findings, Practical implications, Originality/value).
---
## Why this dataset?
Large‑language models (LLMs) frequently struggle to:
1. **Condense long scientific prose** into factual, concise summaries.
2. **Follow rigid output structures** (e.g. subsection headings).
This dataset targets both challenges simultaneously, enabling fine‑tuning or instruction‑tuning of LLMs that must output *structured* scholarly abstracts.
---
## At a glance
| Split | Rows | Size (compressed) |
|-------|------|-------------------|
| train | **145 067** | 626 MB |
| test | **6 650** | 29 MB |
| **Total** | **151 717** | ≈655 MB |
<sup>Counts taken from the Hugging Face viewer on 2025‑04‑29.</sup>
---
## Data schema
```text
{
title: string # Paper title
keywords: list[string] # Author‑supplied keywords (0‑23)
messages: list[dict] length ≥ 2 # ChatML‑style conversation
}
```
### `messages` format
Each list contains alternating dictionaries with:
- `role`: either `"user"` or `"assistant"`.
- `content`: UTF‑8 text.
Typical pattern (2 items):
```jsonc
[
{
"role": "user",
"content": "Summarize the following paper into structured abstract.\n\n<full paper text>"
},
{
"role": "assistant",
"content": "Purpose: …\nDesign/methodology/approach: …\nFindings: …\nPractical implications: …\nOriginality/value: …"
}
]
```
Some papers are longer and may be truncated to ~8 k tokens.
---
## Loading the data
```python
from datasets import load_dataset
ds_train = load_dataset(
"Neooooo/structured_paper_summarization", split="train"
)
print(ds_train[0]["messages"][1]["content"][:500])
```
The dataset is stored as Apache **Parquet** with streaming support; the example above requires ~5 s to start iterating with no local download.
---
## Suggested use‑cases
* **Instruction‑tuning** chat LLMs for long‑document summarisation.
* Research on **controlled text generation** and output formatting.
* Training **retrieval‑augmented systems** that must cite sections of the source paper.
---
## Source & construction
1. Full‑text articles were collected via institutional access to the *Emerald Insight* corpus (open‑access + subscription).
2. The canonical *structured abstract* supplied by each journal was extracted as ground truth.
3. The article’s main body was embedded into a prompt of the form shown above.
4. Data were converted to Hugging Face `datasets` ➜ auto‑parquet.
No additional manual cleaning was performed; typos and OCR artefacts may persist.
---
## Licensing & acceptable use
The article texts are **copyright their original publishers/authors** and are redistributed here *solely for non‑commercial research*. By using this dataset you agree to:
- **Not** redistribute the raw paper texts.
- Cite the original articles in any derivative work.
- Abide by Emerald’s usage policy and your local copyright laws.
The **metadata & structured abstracts** are released under **CC BY‑NC 4.0**. For commercial licensing, please contact the original rights‑holders.
---
## Citation
If you use this dataset, please cite:
```text
@dataset{hu_2025_structured_prompts,
author = {Xingyu Hu},
title = {structured_paper_summarization},
year = 2025,
url = {https://huggingface.co/datasets/Neooooo/structured_paper_summarization},
note = {Version 1.0}
}
```
---
## Contributions
Feel free to open PRs to:
- Fix metadata errors.
- Provide additional splits (validation, domain‑specific subsets).
- Add scripts for evaluation or preprocessing.
---
*Happy summarising!*