caush commited on
Commit
43f8fe5
·
1 Parent(s): c6ae772

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -4
README.md CHANGED
@@ -12,21 +12,25 @@ should probably proofread and complete it, then remove this comment. -->
12
 
13
  # Clickbait1
14
 
15
- This model is a fine-tuned version of [microsoft/Multilingual-MiniLM-L12-H384](https://huggingface.co/microsoft/Multilingual-MiniLM-L12-H384) on the None dataset.
16
  It achieves the following results on the evaluation set:
17
  - Loss: 0.0260
18
 
19
  ## Model description
20
 
21
- More information needed
 
 
22
 
23
  ## Intended uses & limitations
24
 
25
- More information needed
 
 
26
 
27
  ## Training and evaluation data
28
 
29
- More information needed
30
 
31
  ## Training procedure
32
 
 
12
 
13
  # Clickbait1
14
 
15
+ This model is a fine-tuned version of [microsoft/Multilingual-MiniLM-L12-H384](https://huggingface.co/microsoft/Multilingual-MiniLM-L12-H384) on the [Webis-Clickbait-17](https://zenodo.org/record/5530410) dataset.
16
  It achieves the following results on the evaluation set:
17
  - Loss: 0.0260
18
 
19
  ## Model description
20
 
21
+ MiniLM is a distilled model from the paper "MiniLM: Deep Self-Attention Distillation for Task-Agnostic Compression of Pre-Trained Transformers".
22
+
23
+ We fine tune this model to evaluate (regression) the clickbait level of a title news.
24
 
25
  ## Intended uses & limitations
26
 
27
+ Model was designed to test the possibilities of Transformers in this cas of NLP problem (like in the paper "Predicting Clickbait Strength in Online Social Media" by Indurthi Vijayasaradhi, Syed Bakhtiyar, Gupta Manish, Varma Vasudeva).
28
+
29
+ The model wa trained in english.
30
 
31
  ## Training and evaluation data
32
 
33
+ We train the model with the official training data of the chalenge, plus another set that was available after the end of the challenge.
34
 
35
  ## Training procedure
36