pilotj commited on
Commit
adbe1c3
·
verified ·
1 Parent(s): 4eedeef

pilotj/distil-bert-tweaking

Browse files
Files changed (1) hide show
  1. README.md +21 -10
README.md CHANGED
@@ -1,26 +1,21 @@
1
  ---
2
- base_model: pilotj/distilbert-base-uncased-fibe-full-finetuned
3
  library_name: transformers
 
4
  tags:
5
  - generated_from_trainer
6
  model-index:
7
- - name: distil-bert-final-version
8
  results: []
9
  ---
10
 
11
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
  should probably proofread and complete it, then remove this comment. -->
13
 
14
- # distil-bert-final-version
15
 
16
  This model is a fine-tuned version of [pilotj/distilbert-base-uncased-fibe-full-finetuned](https://huggingface.co/pilotj/distilbert-base-uncased-fibe-full-finetuned) on the None dataset.
17
  It achieves the following results on the evaluation set:
18
- - eval_loss: 0.4621
19
- - eval_runtime: 244.5245
20
- - eval_samples_per_second: 106.946
21
- - eval_steps_per_second: 0.838
22
- - epoch: 0.4331
23
- - step: 9000
24
 
25
  ## Model description
26
 
@@ -46,7 +41,23 @@ The following hyperparameters were used during training:
46
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
  - lr_scheduler_type: linear
48
  - lr_scheduler_warmup_ratio: 0.5
49
- - num_epochs: 1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
50
 
51
  ### Framework versions
52
 
 
1
  ---
 
2
  library_name: transformers
3
+ base_model: pilotj/distilbert-base-uncased-fibe-full-finetuned
4
  tags:
5
  - generated_from_trainer
6
  model-index:
7
+ - name: distil-bert-tweaking
8
  results: []
9
  ---
10
 
11
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
  should probably proofread and complete it, then remove this comment. -->
13
 
14
+ # distil-bert-tweaking
15
 
16
  This model is a fine-tuned version of [pilotj/distilbert-base-uncased-fibe-full-finetuned](https://huggingface.co/pilotj/distilbert-base-uncased-fibe-full-finetuned) on the None dataset.
17
  It achieves the following results on the evaluation set:
18
+ - Loss: 0.4871
 
 
 
 
 
19
 
20
  ## Model description
21
 
 
41
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
  - lr_scheduler_type: linear
43
  - lr_scheduler_warmup_ratio: 0.5
44
+ - num_epochs: 2
45
+
46
+ ### Training results
47
+
48
+ | Training Loss | Epoch | Step | Validation Loss |
49
+ |:-------------:|:------:|:----:|:---------------:|
50
+ | 0.3542 | 0.1905 | 500 | 0.4708 |
51
+ | 0.418 | 0.3810 | 1000 | 0.4670 |
52
+ | 0.4303 | 0.5714 | 1500 | 0.4833 |
53
+ | 0.4404 | 0.7619 | 2000 | 0.5177 |
54
+ | 0.4589 | 0.9524 | 2500 | 0.5050 |
55
+ | 0.3911 | 1.1429 | 3000 | 0.5887 |
56
+ | 0.3654 | 1.3333 | 3500 | 0.5227 |
57
+ | 0.3415 | 1.5238 | 4000 | 0.5158 |
58
+ | 0.3284 | 1.7143 | 4500 | 0.5228 |
59
+ | 0.3178 | 1.9048 | 5000 | 0.4871 |
60
+
61
 
62
  ### Framework versions
63