--- dataset_info: features: - name: title dtype: string - name: keywords sequence: string - name: messages list: - name: content dtype: string - name: role dtype: string splits: - name: train num_bytes: 1851098097.8341584 num_examples: 145064 - name: test num_bytes: 78063099.39124106 num_examples: 6653 download_size: 626249553 dataset_size: 1929161197.2253995 configs: - config_name: default data_files: - split: train path: data/train-* - split: test path: data/test-* --- # structured_paper_summarization A **151 k‑example** dataset of chat‐style *prompt → structured abstract* pairs, built from ~19 000 research papers across business, management, information‑systems and social‑science domains. Each example shows the full paper (body text) being summarised into a five‑section Emerald‑style structured abstract (Purpose, Design/methodology/approach, Findings, Practical implications, Originality/value). --- ## Why this dataset? Large‑language models (LLMs) frequently struggle to: 1. **Condense long scientific prose** into factual, concise summaries. 2. **Follow rigid output structures** (e.g. subsection headings). This dataset targets both challenges simultaneously, enabling fine‑tuning or instruction‑tuning of LLMs that must output *structured* scholarly abstracts. --- ## At a glance | Split | Rows | Size (compressed) | |-------|------|-------------------| | train | **145 067** | 626 MB | | test | **6 650** | 29 MB | | **Total** | **151 717** | ≈655 MB | Counts taken from the Hugging Face viewer on 2025‑04‑29. --- ## Data schema ```text { title: string # Paper title keywords: list[string] # Author‑supplied keywords (0‑23) messages: list[dict] length ≥ 2 # ChatML‑style conversation } ``` ### `messages` format Each list contains alternating dictionaries with: - `role`: either `"user"` or `"assistant"`. - `content`: UTF‑8 text. Typical pattern (2 items): ```jsonc [ { "role": "user", "content": "Summarize the following paper into structured abstract.\n\n" }, { "role": "assistant", "content": "Purpose: …\nDesign/methodology/approach: …\nFindings: …\nPractical implications: …\nOriginality/value: …" } ] ``` Some papers are longer and may be truncated to ~8 k tokens. --- ## Loading the data ```python from datasets import load_dataset ds_train = load_dataset( "Neooooo/structured_paper_summarization", split="train" ) print(ds_train[0]["messages"][1]["content"][:500]) ``` The dataset is stored as Apache **Parquet** with streaming support; the example above requires ~5 s to start iterating with no local download. --- ## Suggested use‑cases * **Instruction‑tuning** chat LLMs for long‑document summarisation. * Research on **controlled text generation** and output formatting. * Training **retrieval‑augmented systems** that must cite sections of the source paper. --- ## Source & construction 1. Full‑text articles were collected via institutional access to the *Emerald Insight* corpus (open‑access + subscription). 2. The canonical *structured abstract* supplied by each journal was extracted as ground truth. 3. The article’s main body was embedded into a prompt of the form shown above. 4. Data were converted to Hugging Face `datasets` ➜ auto‑parquet. No additional manual cleaning was performed; typos and OCR artefacts may persist. --- ## Licensing & acceptable use The article texts are **copyright their original publishers/authors** and are redistributed here *solely for non‑commercial research*. By using this dataset you agree to: - **Not** redistribute the raw paper texts. - Cite the original articles in any derivative work. - Abide by Emerald’s usage policy and your local copyright laws. The **metadata & structured abstracts** are released under **CC BY‑NC 4.0**. For commercial licensing, please contact the original rights‑holders. --- ## Citation If you use this dataset, please cite: ```text @dataset{hu_2025_structured_prompts, author = {Xingyu Hu}, title = {structured_paper_summarization}, year = 2025, url = {https://huggingface.co/datasets/Neooooo/structured_paper_summarization}, note = {Version 1.0} } ``` --- ## Contributions Feel free to open PRs to: - Fix metadata errors. - Provide additional splits (validation, domain‑specific subsets). - Add scripts for evaluation or preprocessing. --- *Happy summarising!*