Shadman-Rohan commited on
Commit
b26f658
·
1 Parent(s): fbf74a0

update model card README.md

Browse files
Files changed (1) hide show
  1. README.md +18 -18
README.md CHANGED
@@ -15,11 +15,9 @@ model-index:
15
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
16
  should probably proofread and complete it, then remove this comment. -->
17
 
18
- # Coding challenge
19
- The challenge involved building a fake news classifier using the huggingface library.
20
-
21
- This final model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an fake-and-real-news dataset. The link to the dataset is https://www.kaggle.com/datasets/clmentbisaillon/fake-and-real-news-dataset.
22
 
 
23
  It achieves the following results on the evaluation set:
24
  - Loss: 0.0000
25
  - Accuracy: 1.0
@@ -29,15 +27,17 @@ It achieves the following results on the evaluation set:
29
 
30
  ## Model description
31
 
32
- Finetuned Distilbert
33
-
 
 
 
 
34
  ## Training and evaluation data
35
 
36
- The training data was split into train-dev-test in the ratio 80-10-10.
37
 
38
  ## Training procedure
39
- The title and text of each news story was concatenated to form each datapoint. Then a model was finetuned to perform single label classification on each datapoint. The final prediction is the class with the highest probability.
40
-
41
 
42
  ### Training hyperparameters
43
 
@@ -53,16 +53,16 @@ The following hyperparameters were used during training:
53
 
54
  ### Training results
55
 
56
- | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | Precision | Recall |
57
- |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|:---------:|:------:|
58
- | 0.0503 | 1.0 | 1956 | 0.0025 | 0.9995 | 0.9995 | 0.9995 | 0.9995 |
59
- | 0.001 | 2.0 | 3912 | 0.0001 | 1.0 | 1.0 | 1.0 | 1.0 |
60
- | 0.0007 | 3.0 | 5868 | 0.0000 | 1.0 | 1.0 | 1.0 | 1.0 |
61
 
62
 
63
  ### Framework versions
64
 
65
- - Transformers 4.18.0
66
- - Pytorch 1.10.0+cu111
67
- - Datasets 2.1.0
68
- - Tokenizers 0.12.1
 
15
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
16
  should probably proofread and complete it, then remove this comment. -->
17
 
18
+ # FakevsRealNews
 
 
 
19
 
20
+ This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset.
21
  It achieves the following results on the evaluation set:
22
  - Loss: 0.0000
23
  - Accuracy: 1.0
 
27
 
28
  ## Model description
29
 
30
+ More information needed
31
+
32
+ ## Intended uses & limitations
33
+
34
+ More information needed
35
+
36
  ## Training and evaluation data
37
 
38
+ More information needed
39
 
40
  ## Training procedure
 
 
41
 
42
  ### Training hyperparameters
43
 
 
53
 
54
  ### Training results
55
 
56
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | Precision | Recall |
57
+ |:-------------:|:-----:|:----:|:---------------:|:--------:|:---:|:---------:|:------:|
58
+ | 0.0554 | 1.0 | 1956 | 0.0000 | 1.0 | 1.0 | 1.0 | 1.0 |
59
+ | 0.0006 | 2.0 | 3912 | 0.0000 | 1.0 | 1.0 | 1.0 | 1.0 |
60
+ | 0.0 | 3.0 | 5868 | 0.0000 | 1.0 | 1.0 | 1.0 | 1.0 |
61
 
62
 
63
  ### Framework versions
64
 
65
+ - Transformers 4.27.2
66
+ - Pytorch 1.13.1+cu116
67
+ - Datasets 2.10.1
68
+ - Tokenizers 0.13.2