|
|
--- |
|
|
dataset: |
|
|
- name: USSocialIssues_2016ElectionSurvey |
|
|
tags: |
|
|
- survey |
|
|
- public-opinion |
|
|
- crime |
|
|
- welfare |
|
|
- united-states |
|
|
- election-2016 |
|
|
- panel-data |
|
|
license: mit |
|
|
--- |
|
|
|
|
|
|
|
|
# Survey on Social Issues in the United States (2016 Election Study) |
|
|
|
|
|
## Overview |
|
|
|
|
|
This data product contains individual-level responses to an online survey experiment conducted in the run-up to—and immediately after—the 2016 U.S. presidential election by [Connor Jerzak](https://connorjerzak.com/), Rebecca Goldstein, and Yanilda María González. |
|
|
|
|
|
Survey began on **12 September 2016** and continued through **mid-November 2016**, giving researchers a before/after snapshot of attitudes shaped by a highly salient national campaign. |
|
|
|
|
|
**Key design features** |
|
|
|
|
|
* **Crime-framing experiment.** Respondents read a mock police-blotter story with experimentally varied details (suspect race, number of break-ins, presence/absence of racial information) before answering questions about policy, crime perceptions, and social spending. |
|
|
* **Demographics & ideology.** Over 50 items capture party identification, vote choice, income, employment security, family status, education, racial identity, and core value trade-offs. |
|
|
* **Panel structure.** A subset of respondents was re-contacted after Election Day, enabling within-person analyses of opinion change. |
|
|
|
|
|
## File Manifest |
|
|
|
|
|
| File | Description | |
|
|
|------|-------------| |
|
|
| `survey_results.csv` | Clean, respondent-level dataset (wide format). Each column corresponds to a survey variable prefixed by its original Qualtrics question ID. | |
|
|
| `Oct21_survey.pdf` | Archived survey instrument, including consent form and full questionnaire. | |
|
|
|
|
|
## Quick Start (R) |
|
|
|
|
|
```r |
|
|
library(tidyverse) |
|
|
df <- read_csv("survey_results.csv") |
|
|
|
|
|
# Recode experimental treatment |
|
|
# Q42 == "No" -> Control |
|
|
# Q42 == "Yes" & Q43 gives race |
|
|
df <- df %>% |
|
|
mutate(treat = case_when( |
|
|
Q42 == "No" ~ "Control", |
|
|
Q43 == "Black" ~ "Black", |
|
|
Q43 == "White" ~ "White" |
|
|
)) |
|
|
|
|
|
# Estimate effect of racial cue on support for longer sentences |
|
|
lm(long_sentences ~ treat + party_id + age, data = df) |
|
|
``` |
|
|
|
|
|
## Variable Highlights |
|
|
|
|
|
* **Safety perceptions:** `Q2`–`Q4`, `Q37`, `Q39` |
|
|
* **Crime policy preferences:** `Q11`, `Q12` |
|
|
* **Redistribution & welfare attitudes:** `Q8`, `Q9`, `Q46`–`Q51` |
|
|
* **2016 vote intention & choice:** `Q41`, `Q44`, `Q45` |
|
|
* **Economic security:** `Q29`–`Q32` |
|
|
* **Child-rearing values:** `Q33`–`Q36` |
|
|
|
|
|
See `Oct21_survey.pdf` for exact wording and response options. |
|
|
|
|
|
## Possible Use Cases |
|
|
|
|
|
1. **Election-season opinion dynamics** – analyze the before/after panel to examine how campaign events (debates, the Comey letter, Election Day) shifted perceptions of crime, policing, or redistribution. |
|
|
2. **Stereotype activation & policy support** – estimate causal effects of suspect-race cues on punitive crime policies or welfare attitudes. |
|
|
3. **Replication exercises** – reproduce classic findings from ANES or GSS items using a contemporary MTurk sample; ideal for teaching regression, causal inference, or text analysis (e.g., coding open-ended crime causes in `Q10`). |
|
|
4. **Value trade-off scaling** – model latent moral or parenting value dimensions with the paired choice items (`Q33`–`Q36`). |
|
|
5. **Small-N machine-learning demos** – demonstrate text classification, topic modeling, or mixed-effects models on a manageable survey. |
|
|
|
|
|
## Sampling & Fieldwork |
|
|
|
|
|
Respondents were recruited via **Amazon Mechanical Turk**. Each wave paid \$0.25 and took ~5 minutes. The instrument included an informed-consent screen and was approved by the Harvard CUHS IRB. IP geo-coordinates (rounded to 3 decimals) were recorded for coarse location checks; no personally identifying information is included. |
|
|
|
|
|
| Wave | Dates | N (unique) | Notes | |
|
|
|------|-------|------------|-------| |
|
|
| Pre-Election | 12 Sep – 04 Nov 2016 | 449 | Prior to Election Day | |
|
|
| Post-Election | 09 Nov – 15 Nov 2016 | 546 | Post Election Dad | |
|
|
|
|
|
## Data Quality Notes |
|
|
|
|
|
* **Non-probability sample.** MTurk respondents skew younger, more educated, and more politically engaged than the general U.S. adult population. |
|
|
* **Attention checks.** Various items (e.g., number of break-ins retention check) facilitate quality screening. |
|
|
* **Missing values.** Skipped or invalid responses are coded `NA`. |
|
|
|
|
|
|
|
|
## Citation |
|
|
|
|
|
``` |
|
|
@misc{SocialIssuesSurvey2016, |
|
|
author = {Jerzak, Connor and Rebecca Goldstein and Yanilda González}, |
|
|
title = {{USSocialIssues\_2016ElectionSurvey}}, |
|
|
year = {2025}, |
|
|
publisher = {Hugging Face}, |
|
|
doi = {10.57967/hf/5892}, |
|
|
url = {https://huggingface.co/datasets/cjerzak/USSocialIssues_2016ElectionSurvey}, |
|
|
howpublished = {Hugging Face Dataset} |
|
|
} |
|
|
``` |
|
|
|
|
|
|
|
|
|