ptran74 commited on
Commit
736f20b
1 Parent(s): a090629

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +20 -1
README.md CHANGED
@@ -13,12 +13,27 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # DSPFirst-Finetuning-3
15
 
16
- This model is a fine-tuned version of [ahotrod/electra_large_discriminator_squad2_512](https://huggingface.co/ahotrod/electra_large_discriminator_squad2_512) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
  - Loss: 0.9996
19
  - Exact: 63.9193
20
  - F1: 72.1090
21
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
22
  ## Model description
23
 
24
  More information needed
@@ -46,6 +61,10 @@ The following hyperparameters were used during training:
46
  - lr_scheduler_type: linear
47
  - num_epochs: 3
48
 
 
 
 
 
49
  ### Training results
50
 
51
  | Training Loss | Epoch | Step | Validation Loss | Exact | F1 |
 
13
 
14
  # DSPFirst-Finetuning-3
15
 
16
+ This model is a fine-tuned version of [ahotrod/electra_large_discriminator_squad2_512](https://huggingface.co/ahotrod/electra_large_discriminator_squad2_512) on a generated Questions and Answers dataset from the DSPFirst textbook based on the SQuAD 2.0 format.
17
  It achieves the following results on the evaluation set:
18
  - Loss: 0.9996
19
  - Exact: 63.9193
20
  - F1: 72.1090
21
 
22
+ # Dataset
23
+ A visualization of the dataset can be found [here](https://github.gatech.edu/pages/VIP-ITS/textbook_SQuAD_explore/explore/textbookv1.0/textbook/). The split between train and test is 80% and 20% respectively.
24
+ ```
25
+ DatasetDict({
26
+ train: Dataset({
27
+ features: ['id', 'title', 'context', 'question', 'answers'],
28
+ num_rows: 4755
29
+ })
30
+ test: Dataset({
31
+ features: ['id', 'title', 'context', 'question', 'answers'],
32
+ num_rows: 1189
33
+ })
34
+ })
35
+ ```
36
+
37
  ## Model description
38
 
39
  More information needed
 
61
  - lr_scheduler_type: linear
62
  - num_epochs: 3
63
 
64
+ ### Model hyperparameters
65
+ - hidden_dropout_prob: 0.35
66
+ - attention_probs_dropout_prob = 0.35
67
+
68
  ### Training results
69
 
70
  | Training Loss | Epoch | Step | Validation Loss | Exact | F1 |