ptran74 commited on
Commit
2f16418
1 Parent(s): 03eae2b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +46 -6
README.md CHANGED
@@ -13,25 +13,60 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # DSPFirst-Finetuning-4
15
 
16
- This model is a fine-tuned version of [ahotrod/electra_large_discriminator_squad2_512](https://huggingface.co/ahotrod/electra_large_discriminator_squad2_512) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
  - Loss: 0.9028
19
  - Exact: 66.9843
20
  - F1: 74.2286
21
 
22
- ## Model description
23
-
24
- More information needed
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
25
 
26
  ## Intended uses & limitations
27
 
28
- More information needed
 
29
 
30
  ## Training and evaluation data
31
 
32
- More information needed
 
 
 
33
 
34
  ## Training procedure
 
 
 
35
 
36
  ### Training hyperparameters
37
 
@@ -46,6 +81,11 @@ The following hyperparameters were used during training:
46
  - lr_scheduler_type: linear
47
  - num_epochs: 10
48
 
 
 
 
 
 
49
  ### Training results
50
 
51
  | Training Loss | Epoch | Step | Validation Loss | Exact | F1 |
 
13
 
14
  # DSPFirst-Finetuning-4
15
 
16
+ This model is a fine-tuned version of [ahotrod/electra_large_discriminator_squad2_512](https://huggingface.co/ahotrod/electra_large_discriminator_squad2_512) on a generated Questions and Answers dataset from the DSPFirst textbook based on the SQuAD 2.0 format.<br />
17
  It achieves the following results on the evaluation set:
18
  - Loss: 0.9028
19
  - Exact: 66.9843
20
  - F1: 74.2286
21
 
22
+ ## More accurate metrics:
23
+
24
+ ### Before fine-tuning:
25
+
26
+ ```
27
+ "exact": 57.006726457399104,
28
+ "f1": 61.997705120754276
29
+ ```
30
+
31
+ ### After fine-tuning:
32
+
33
+ ```
34
+ "exact": 66.98430493273543,
35
+ "f1": 74.2285867775556
36
+ ```
37
+
38
+ # Dataset
39
+ A visualization of the dataset can be found [here](https://github.gatech.edu/pages/VIP-ITS/textbook_SQuAD_explore/explore/textbookv1.0/textbook/).<br />
40
+ The split between train and test is 70% and 30% respectively.
41
+ ```
42
+ DatasetDict({
43
+ train: Dataset({
44
+ features: ['id', 'title', 'context', 'question', 'answers'],
45
+ num_rows: 4160
46
+ })
47
+ test: Dataset({
48
+ features: ['id', 'title', 'context', 'question', 'answers'],
49
+ num_rows: 1784
50
+ })
51
+ })
52
+ ```
53
 
54
  ## Intended uses & limitations
55
 
56
+ This model is fine-tuned to answer questions from the DSPFirst textbook. I'm not really sure what I am doing so you should review before using it.<br />
57
+ Also, you should improve the Dataset either by using a **better generated questions and answers model** (currently using https://github.com/patil-suraj/question_generation) or perform **data augmentation** to increase dataset size.
58
 
59
  ## Training and evaluation data
60
 
61
+ - `batch_size` of 6 results in 14.82 GB VRAM
62
+ - Utilizes `gradient_accumulation_steps` to get total batch size to 514 (batch size should be at least 256)
63
+ - 4.52 GB RAM
64
+ - 30% of the total questions is dedicated for evaluating.
65
 
66
  ## Training procedure
67
+ - The model was trained from Google Colab
68
+ - Utilizes Tesla P100 16GB, took 6.3 hours to train
69
+ - `load_best_model_at_end` is enabled in TrainingArguments
70
 
71
  ### Training hyperparameters
72
 
 
81
  - lr_scheduler_type: linear
82
  - num_epochs: 10
83
 
84
+ ### Model hyperparameters
85
+
86
+ - hidden_dropout_prob: 0.36
87
+ - attention_probs_dropout_prob = 0.36
88
+
89
  ### Training results
90
 
91
  | Training Loss | Epoch | Step | Validation Loss | Exact | F1 |