| ---
|
| dataset_info:
|
| - config_name: paw_edits
|
| features:
|
| - name: text
|
| dtype: string
|
| - name: page
|
| dtype: string
|
| - name: section_name
|
| dtype: string
|
| - name: is_paw_edit
|
| dtype: bool
|
| - config_name: queries
|
| features:
|
| - name: query
|
| dtype: string
|
| - name: type
|
| dtype: string
|
| license: cc-by-4.0
|
| task_categories:
|
| - text-classification
|
| tags:
|
| - animal-welfare
|
| - wikipedia
|
| - data-attribution
|
| - llm-training
|
| ---
|
|
|
| # PAW Wikipedia Attribution Dataset
|
|
|
| Data for the paper "Small edits, large models: How Wikipedia advocacy shapes LLM values" (Navas, Brazilek & Gnauck, 2026).
|
|
|
| ## Files
|
|
|
| ### paw_edits.csv (236 rows)
|
| Wikipedia sections used in training data attribution experiments. Each row is a text chunk from Wikipedia.
|
|
|
| | Column | Description |
|
| |--------|-------------|
|
| | `text` | The Wikipedia section text |
|
| | `page` | Wikipedia article title |
|
| | `section_name` | Section heading |
|
| | `is_paw_edit` | Whether this section was added/edited by Pro-Animal Wikipedians |
|
|
|
| 118 PAW-edited sections (animal welfare content added by advocacy editors) paired with 118 control sections from the same Wikipedia articles.
|
|
|
| ### queries.csv (170 rows)
|
| Queries used to measure training data influence.
|
|
|
| | Column | Description |
|
| |--------|-------------|
|
| | `query` | The query text |
|
| | `type` | `animal_welfare` (80 queries) or `general` (90 queries) |
|
|
|
| Animal welfare queries ask about animal welfare at specific companies/by specific politicians. General queries ask about the same entities without mentioning animal welfare.
|
|
|
| ## Citation
|
|
|
| ```bibtex
|
| @article{navas2026small,
|
| title={Small edits, large models: How Wikipedia advocacy shapes LLM values},
|
| author={Navas, M and Brazilek, J and Gnauck, A},
|
| year={2026}
|
| }
|
| ```
|
|
|