| --- |
| datasets: |
| - tweet_eval |
| metrics: |
| - f1 |
| - accuracy |
| model-index: |
| - name: cardiffnlp/roberta-base-offensive |
| results: |
| - task: |
| type: text-classification |
| name: Text Classification |
| dataset: |
| name: tweet_eval |
| type: offensive |
| split: test |
| metrics: |
| - name: Micro F1 (tweet_eval/offensive) |
| type: micro_f1_tweet_eval/offensive |
| value: 0.8441860465116279 |
| - name: Macro F1 (tweet_eval/offensive) |
| type: micro_f1_tweet_eval/offensive |
| value: 0.8038468085106383 |
| - name: Accuracy (tweet_eval/offensive) |
| type: accuracy_tweet_eval/offensive |
| value: 0.8441860465116279 |
| pipeline_tag: text-classification |
| widget: |
| - text: Get the all-analog Classic Vinyl Edition of "Takin Off" Album from {@herbiehancock@} via {@bluenoterecords@} link below {{URL}} |
| example_title: "topic_classification 1" |
| - text: Yes, including Medicare and social security saving👍 |
| example_title: "sentiment 1" |
| - text: All two of them taste like ass. |
| example_title: "offensive 1" |
| - text: If you wanna look like a badass, have drama on social media |
| example_title: "irony 1" |
| - text: Whoever just unfollowed me you a bitch |
| example_title: "hate 1" |
| - text: I love swimming for the same reason I love meditating...the feeling of weightlessness. |
| example_title: "emotion 1" |
| - text: Beautiful sunset last night from the pontoon @TupperLakeNY |
| example_title: "emoji 1" |
| --- |
| # cardiffnlp/roberta-base-offensive |
|
|
| This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on the |
| [`tweet_eval (offensive)`](https://huggingface.co/datasets/tweet_eval) |
| via [`tweetnlp`](https://github.com/cardiffnlp/tweetnlp). |
| Training split is `train` and parameters have been tuned on the validation split `validation`. |
|
|
| Following metrics are achieved on the test split `test` ([link](https://huggingface.co/cardiffnlp/roberta-base-offensive/raw/main/metric.json)). |
|
|
| - F1 (micro): 0.8441860465116279 |
| - F1 (macro): 0.8038468085106383 |
| - Accuracy: 0.8441860465116279 |
|
|
| ### Usage |
| Install tweetnlp via pip. |
| ```shell |
| pip install tweetnlp |
| ``` |
| Load the model in python. |
| ```python |
| import tweetnlp |
| model = tweetnlp.Classifier("cardiffnlp/roberta-base-offensive", max_length=128) |
| model.predict('Get the all-analog Classic Vinyl Edition of "Takin Off" Album from {@herbiehancock@} via {@bluenoterecords@} link below {{URL}}') |
| ``` |
|
|
|
|
|
|
| ### Reference |
|
|
| |
| ``` |
| @inproceedings{camacho-collados-etal-2022-tweetnlp, |
| title={{T}weet{NLP}: {C}utting-{E}dge {N}atural {L}anguage {P}rocessing for {S}ocial {M}edia}, |
| author={Camacho-Collados, Jose and Rezaee, Kiamehr and Riahi, Talayeh and Ushio, Asahi and Loureiro, Daniel and Antypas, Dimosthenis and Boisson, Joanne and Espinosa-Anke, Luis and Liu, Fangyu and Mart{'\i}nez-C{'a}mara, Eugenio and others}, |
| author = "Ushio, Asahi and |
| Camacho-Collados, Jose", |
| booktitle = "Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing: System Demonstrations", |
| month = nov, |
| year = "2022", |
| address = "Abu Dhabi, U.A.E.", |
| publisher = "Association for Computational Linguistics", |
| } |
| ``` |
|
|