Update README.md
Browse files
README.md
CHANGED
|
@@ -21,8 +21,6 @@ widget:
|
|
| 21 |
---
|
| 22 |
|
| 23 |
|
| 24 |
-
# TurkishBERTweet
|
| 25 |
-
|
| 26 |
#### Table of contents
|
| 27 |
1. [Introduction](#introduction)
|
| 28 |
2. [Main results](#results)
|
|
@@ -36,7 +34,7 @@ widget:
|
|
| 36 |
- [HateSpeech Detection](#hs_lora)
|
| 37 |
|
| 38 |
4. [Citation](#citation)
|
| 39 |
-
# <a name="introduction"></a> TurkishBERTweet
|
| 40 |
|
| 41 |
|
| 42 |
# <a name="results"></a> Main Results
|
|
@@ -48,13 +46,13 @@ widget:
|
|
| 48 |
# <a name="trainedModels"></a> Model
|
| 49 |
Model | #params | Arch. | Max length | Pre-training data
|
| 50 |
---|---|---|---|---
|
| 51 |
-
`VRLLab/TurkishBERTweet` | 163M | base | 128 | 894M Turkish Tweets (uncased)
|
| 52 |
|
| 53 |
# <a name="loraAdapter"></a> Lora Adapters
|
| 54 |
Model | train f1 | dev f1 | test f1 | Dataset Size
|
| 55 |
---|---|---|---|---
|
| 56 |
-
`VRLLab/TurkishBERTweet-Lora-SA` | 0.799 | 0.687 | 0.692 | 42,476 Turkish Tweets
|
| 57 |
-
`VRLLab/TurkishBERTweet-Lora-HS` | 0.915 | 0.796 | 0.831 | 4,683 Turkish Tweets
|
| 58 |
# <a name="usage2"></a> Example usage
|
| 59 |
|
| 60 |
|
|
@@ -203,13 +201,14 @@ Yes : kasmayin artik ya kac kere tanik olduk bu azgin tehlikeli “multecilerin
|
|
| 203 |
# <a name="citation"></a> Citation
|
| 204 |
```bibtex
|
| 205 |
@article{najafi2022TurkishBERTweet,
|
| 206 |
-
title={TurkishBERTweet
|
| 207 |
author={Najafi, Ali and Varol, Onur},
|
| 208 |
-
journal={arXiv preprint },
|
| 209 |
year={2023}
|
| 210 |
}
|
| 211 |
```
|
| 212 |
|
|
|
|
| 213 |
## Acknowledgments
|
| 214 |
We thank [Fatih Amasyali](https://avesis.yildiz.edu.tr/amasyali) for providing access to Tweet Sentiment datasets from Kemik group.
|
| 215 |
This material is based upon work supported by the Google Cloud Research Credits program with the award GCP19980904. We also thank TUBITAK (121C220 and 222N311) for funding this project.
|
|
|
|
| 21 |
---
|
| 22 |
|
| 23 |
|
|
|
|
|
|
|
| 24 |
#### Table of contents
|
| 25 |
1. [Introduction](#introduction)
|
| 26 |
2. [Main results](#results)
|
|
|
|
| 34 |
- [HateSpeech Detection](#hs_lora)
|
| 35 |
|
| 36 |
4. [Citation](#citation)
|
| 37 |
+
# <a name="introduction"></a> TurkishBERTweet: Fast and Reliable Large Language Model for Social Media Analysis
|
| 38 |
|
| 39 |
|
| 40 |
# <a name="results"></a> Main Results
|
|
|
|
| 46 |
# <a name="trainedModels"></a> Model
|
| 47 |
Model | #params | Arch. | Max length | Pre-training data
|
| 48 |
---|---|---|---|---
|
| 49 |
+
[`VRLLab/TurkishBERTweet`](https://huggingface.co/VRLLab/TurkishBERTweet) | 163M | base | 128 | 894M Turkish Tweets (uncased)
|
| 50 |
|
| 51 |
# <a name="loraAdapter"></a> Lora Adapters
|
| 52 |
Model | train f1 | dev f1 | test f1 | Dataset Size
|
| 53 |
---|---|---|---|---
|
| 54 |
+
[`VRLLab/TurkishBERTweet-Lora-SA`](https://huggingface.co/VRLLab/TurkishBERTweet-Lora-SA) | 0.799 | 0.687 | 0.692 | 42,476 Turkish Tweets
|
| 55 |
+
[`VRLLab/TurkishBERTweet-Lora-HS`](https://huggingface.co/VRLLab/TurkishBERTweet-Lora-HS) | 0.915 | 0.796 | 0.831 | 4,683 Turkish Tweets
|
| 56 |
# <a name="usage2"></a> Example usage
|
| 57 |
|
| 58 |
|
|
|
|
| 201 |
# <a name="citation"></a> Citation
|
| 202 |
```bibtex
|
| 203 |
@article{najafi2022TurkishBERTweet,
|
| 204 |
+
title={TurkishBERTweet: Fast and Reliable Large Language Model for Social Media Analysis},
|
| 205 |
author={Najafi, Ali and Varol, Onur},
|
| 206 |
+
journal={arXiv preprint 2311.18063},
|
| 207 |
year={2023}
|
| 208 |
}
|
| 209 |
```
|
| 210 |
|
| 211 |
+
|
| 212 |
## Acknowledgments
|
| 213 |
We thank [Fatih Amasyali](https://avesis.yildiz.edu.tr/amasyali) for providing access to Tweet Sentiment datasets from Kemik group.
|
| 214 |
This material is based upon work supported by the Google Cloud Research Credits program with the award GCP19980904. We also thank TUBITAK (121C220 and 222N311) for funding this project.
|