Datasets:

Modalities:
Text
Formats:
csv
Languages:
English
Size:
< 1K
ArXiv:
Libraries:
Datasets
pandas
License:

Improve dataset card: Add task categories, tags, and update paper title

#3
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +9 -1
README.md CHANGED
@@ -6,9 +6,17 @@ size_categories:
6
  - n<1K
7
  task_categories:
8
  - summarization
 
 
 
 
 
 
 
 
9
  ---
10
 
11
- PlainFact-summary is a high-quality human-annotated dataset designed for Plain Language Summarization tasks, along with [PlainQAFact](https://github.com/zhiwenyou103/PlainQAFact) factuality evaluation framework, as described in [PlainQAFact: Automatic Factuality Evaluation Metric for Biomedical Plain Language Summaries Generation](https://huggingface.co/papers/2503.08890). It is collected from the [Cochrane database](https://www.cochranelibrary.com/) sampled from CELLS dataset ([Guo et al., 2024](https://doi.org/10.1016/j.jbi.2023.104580)).
12
  In addition to using all factual plain language summaries, we also generate contrasting non-factual examples for each plain language summary. These contrasting examples are perturbed using GPT-4o, following the perturbation criteria for faithfulness introduced in APPLS ([Guo et al., 2024](https://aclanthology.org/2024.emnlp-main.519/)).
13
 
14
  We also provided a sentence-level version [PlainFact](https://huggingface.co/datasets/uzw/PlainFact) that split the summaries into sentences with fine-grained explanation annotations. In total, we have 200 plain language summary-abstract pairs.
 
6
  - n<1K
7
  task_categories:
8
  - summarization
9
+ - question-answering
10
+ - text-classification
11
+ tags:
12
+ - biomedical
13
+ - medical-domain
14
+ - factual-consistency
15
+ - plain-language-summarization
16
+ - evaluation
17
  ---
18
 
19
+ PlainFact-summary is a high-quality human-annotated dataset designed for Plain Language Summarization tasks, along with [PlainQAFact](https://github.com/zhiwenyou103/PlainQAFact) factuality evaluation framework, as described in [PlainQAFact: Retrieval-augmented Factual Consistency Evaluation Metric for Biomedical Plain Language Summarization](https://huggingface.co/papers/2503.08890). It is collected from the [Cochrane database](https://www.cochranelibrary.com/) sampled from CELLS dataset ([Guo et al., 2024](https://doi.org/10.1016/j.jbi.2023.104580)).
20
  In addition to using all factual plain language summaries, we also generate contrasting non-factual examples for each plain language summary. These contrasting examples are perturbed using GPT-4o, following the perturbation criteria for faithfulness introduced in APPLS ([Guo et al., 2024](https://aclanthology.org/2024.emnlp-main.519/)).
21
 
22
  We also provided a sentence-level version [PlainFact](https://huggingface.co/datasets/uzw/PlainFact) that split the summaries into sentences with fine-grained explanation annotations. In total, we have 200 plain language summary-abstract pairs.