| --- |
| language: |
| - en |
| license: apache-2.0 |
| library_name: transformers |
| tags: |
| - incivility |
| metrics: |
| - f1 |
| widget: |
| - text: "Be careful around those DemocRats." |
| example_title: "Namecall" |
| - text: "Be careful around those Democrats." |
| example_title: "No Namecall" |
| --- |
| |
| # Model Card for roberta-base-namecalling |
|
|
| This is a [roBERTa-base](https://huggingface.co/roberta-base) model fine-tuned on ~12K social media posts annotated for the presence or absence of namecalling. |
|
|
| # How to Get Started with the Model |
|
|
| You can use this model directly with a pipeline for text classification: |
|
|
| ```python |
| >>> import transformers |
| >>> model_name = "civility-lab/roberta-base-namecalling" |
| >>> classifier = transformers.TextClassificationPipeline( |
| ... tokenizer=transformers.AutoTokenizer.from_pretrained(model_name), |
| ... model=transformers.AutoModelForSequenceClassification.from_pretrained(model_name)) |
| >>> classifier("Be careful around those Democrats.") |
| [{'label': 'not-namecalling', 'score': 0.9995089769363403}] |
| >>> classifier("Be careful around those DemocRats.") |
| [{'label': 'namecalling', 'score': 0.996940016746521}] |
| ``` |
|
|
| # Model Details |
|
|
| This is a 2023 update of the model built by [Ozler et al. (2020)](https://aclanthology.org/2020.alw-1.4/) incorporating data from [Rains et al. (2021)](https://doi.org/10.1093/hcr/hqab009) and using a more recent version of the transformers library. |
|
|
| - **Developed by:** |
| [Steven Bethard](https://bethard.github.io/), |
| [Kate Kenski](https://comm.arizona.edu/user/kate-kenski), |
| [Steve Rains](https://comm.arizona.edu/user/steve-rains), |
| [Yotam Shmargad](https://www.yotamshmargad.com/), |
| [Kevin Coe](https://faculty.utah.edu/u0915886-Kevin_Coe/) |
| - **Language:** en |
| - **License:** apache-2.0 |
| - **Parent Model:** roberta-base |
| - **Resources for more information:** |
| - [GitHub Repo](https://github.com/clulab/incivility) |
| - Kadir Bulut Ozler; Kate Kenski; Steve Rains; Yotam Shmargad; Kevin Coe; and Steven Bethard. [Fine-tuning for multi-domain and multi-label uncivil language detection](https://aclanthology.org/2020.alw-1.4/). In Proceedings of the Fourth Workshop on Online Abuse and Harms, pages 28–33, Online, November 2020. Association for Computational Linguistics |
| - Stephen A Rains; Yotam Shmargad; Kevin Coe; Kate Kenski; and Steven Bethard. [Assessing the Russian Troll Efforts to Sow Discord on Twitter during the 2016 U.S. Election](https://doi.org/10.1093/hcr/hqab009). Human Communication Research, 47(4): 477-486. 08 2021. |
| - Stephen A Rains; Jake Harwood; Yotam Shmargad; Kate Kenski; Kevin Coe; and Steven Bethard. [Engagement with partisan Russian troll tweets during the 2016 U.S. presidential election: a social identity perspective](https://doi.org/10.1093/joc/jqac037). Journal of Communication, 73(1): 38-48. 02 2023. |
|
|
| # Uses |
|
|
| The model is intended to be used for text classification, taking as input social media posts and predicting as output whether the post contains namecalling. |
|
|
| It is not intended to generate namecalling, and it should not be used as part of any incivility generation model. |
|
|
| # Training Details |
|
|
| The model was trained on data from four sources: comments on the Arizona Daily Star website from 2011, Russian troll Tweets from 2012-2018, Tucson politician Tweets from 2018, and US presidential primary Tweets from 2019. |
| Each dataset was annotated for the presence of namecalling following the approach of [Coe et al. (2014)](https://doi.org/10.1111/jcom.12104) and split into training, development, and test partitions. |
|
|
| The [roberta-base](https://huggingface.co/roberta-base) model was fine-tuned on the combined training partitions from all four datasets, with texts tokenized using the standard [roberta-base](https://huggingface.co/roberta-base) tokenizer. |
| |
| # Evaluation |
|
|
| The model was evaluated on the test partition of each of the datasets. It achieves the following F1 scores: |
|
|
| - 0.58 F1 on Arizona Daily Star comments |
| - 0.71 F1 on Russian troll Tweets |
| - 0.71 F1 on Tucson politician Tweets |
| - 0.81 F1 on US presidential primary Tweets |
|
|
| # Limitations and Biases |
|
|
| The human coders and their trainers were mostly [Western, educated, industrialized, rich and democratic (WEIRD)](https://www.nature.com/articles/466029a), which may have shaped how they evaluated incivility. |
| The trained models will reflect such biases. |
|
|
| # Environmental Impact |
|
|
| - **Hardware Type:** Tesla V100S-PCIE-32GB |
| - **Hours used:** 22 |
| - **HPC Provider:** <https://hpc.arizona.edu/> |
| - **Carbon Emitted:** 2.85 kg CO2 (estimated by [ML CO2 Impact](https://mlco2.github.io/impact#compute)) |
|
|
| # Citation |
|
|
| ```bibtex |
| @inproceedings{ozler-etal-2020-fine, |
| title = "Fine-tuning for multi-domain and multi-label uncivil language detection", |
| author = "Ozler, Kadir Bulut and |
| Kenski, Kate and |
| Rains, Steve and |
| Shmargad, Yotam and |
| Coe, Kevin and |
| Bethard, Steven", |
| booktitle = "Proceedings of the Fourth Workshop on Online Abuse and Harms", |
| month = nov, |
| year = "2020", |
| address = "Online", |
| publisher = "Association for Computational Linguistics", |
| url = "https://aclanthology.org/2020.alw-1.4", |
| doi = "10.18653/v1/2020.alw-1.4", |
| pages = "28--33", |
| } |
| ``` |
|
|