sparrow8i8's picture
Upload README.md with huggingface_hub
87dc28f verified
---
dataset_info:
- config_name: paw_edits
features:
- name: text
dtype: string
- name: page
dtype: string
- name: section_name
dtype: string
- name: is_paw_edit
dtype: bool
- config_name: queries
features:
- name: query
dtype: string
- name: type
dtype: string
license: cc-by-4.0
task_categories:
- text-classification
tags:
- animal-welfare
- wikipedia
- data-attribution
- llm-training
---
# PAW Wikipedia Attribution Dataset
Data for the paper "Small edits, large models: How Wikipedia advocacy shapes LLM values" (Navas, Brazilek & Gnauck, 2026).
## Files
### paw_edits.csv (236 rows)
Wikipedia sections used in training data attribution experiments. Each row is a text chunk from Wikipedia.
| Column | Description |
|--------|-------------|
| `text` | The Wikipedia section text |
| `page` | Wikipedia article title |
| `section_name` | Section heading |
| `is_paw_edit` | Whether this section was added/edited by Pro-Animal Wikipedians |
118 PAW-edited sections (animal welfare content added by advocacy editors) paired with 118 control sections from the same Wikipedia articles.
### queries.csv (170 rows)
Queries used to measure training data influence.
| Column | Description |
|--------|-------------|
| `query` | The query text |
| `type` | `animal_welfare` (80 queries) or `general` (90 queries) |
Animal welfare queries ask about animal welfare at specific companies/by specific politicians. General queries ask about the same entities without mentioning animal welfare.
## Citation
```bibtex
@article{navas2026small,
title={Small edits, large models: How Wikipedia advocacy shapes LLM values},
author={Navas, M and Brazilek, J and Gnauck, A},
year={2026}
}
```