Update README.md
Browse files
README.md
CHANGED
|
@@ -12,21 +12,25 @@ should probably proofread and complete it, then remove this comment. -->
|
|
| 12 |
|
| 13 |
# Clickbait1
|
| 14 |
|
| 15 |
-
This model is a fine-tuned version of [microsoft/Multilingual-MiniLM-L12-H384](https://huggingface.co/microsoft/Multilingual-MiniLM-L12-H384) on the
|
| 16 |
It achieves the following results on the evaluation set:
|
| 17 |
- Loss: 0.0260
|
| 18 |
|
| 19 |
## Model description
|
| 20 |
|
| 21 |
-
|
|
|
|
|
|
|
| 22 |
|
| 23 |
## Intended uses & limitations
|
| 24 |
|
| 25 |
-
|
|
|
|
|
|
|
| 26 |
|
| 27 |
## Training and evaluation data
|
| 28 |
|
| 29 |
-
|
| 30 |
|
| 31 |
## Training procedure
|
| 32 |
|
|
|
|
| 12 |
|
| 13 |
# Clickbait1
|
| 14 |
|
| 15 |
+
This model is a fine-tuned version of [microsoft/Multilingual-MiniLM-L12-H384](https://huggingface.co/microsoft/Multilingual-MiniLM-L12-H384) on the [Webis-Clickbait-17](https://zenodo.org/record/5530410) dataset.
|
| 16 |
It achieves the following results on the evaluation set:
|
| 17 |
- Loss: 0.0260
|
| 18 |
|
| 19 |
## Model description
|
| 20 |
|
| 21 |
+
MiniLM is a distilled model from the paper "MiniLM: Deep Self-Attention Distillation for Task-Agnostic Compression of Pre-Trained Transformers".
|
| 22 |
+
|
| 23 |
+
We fine tune this model to evaluate (regression) the clickbait level of a title news.
|
| 24 |
|
| 25 |
## Intended uses & limitations
|
| 26 |
|
| 27 |
+
Model was designed to test the possibilities of Transformers in this cas of NLP problem (like in the paper "Predicting Clickbait Strength in Online Social Media" by Indurthi Vijayasaradhi, Syed Bakhtiyar, Gupta Manish, Varma Vasudeva).
|
| 28 |
+
|
| 29 |
+
The model wa trained in english.
|
| 30 |
|
| 31 |
## Training and evaluation data
|
| 32 |
|
| 33 |
+
We train the model with the official training data of the chalenge, plus another set that was available after the end of the challenge.
|
| 34 |
|
| 35 |
## Training procedure
|
| 36 |
|