snork-maiden commited on
Commit
46d7af1
·
1 Parent(s): b152615

update model card README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -17
README.md CHANGED
@@ -13,8 +13,6 @@ should probably proofread and complete it, then remove this comment. -->
13
  # t5-base
14
 
15
  This model is a fine-tuned version of [t5-base](https://huggingface.co/t5-base) on the None dataset.
16
- It achieves the following results on the evaluation set:
17
- - Loss: 0.0191
18
 
19
  ## Model description
20
 
@@ -35,28 +33,19 @@ More information needed
35
  The following hyperparameters were used during training:
36
  - learning_rate: 0.0001
37
  - train_batch_size: 8
38
- - eval_batch_size: 8
39
  - seed: 42
40
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
41
  - lr_scheduler_type: linear
42
- - num_epochs: 20
43
 
44
  ### Training results
45
 
46
- | Training Loss | Epoch | Step | Validation Loss |
47
- |:-------------:|:-----:|:-----:|:---------------:|
48
- | 0.0089 | 2.67 | 2000 | 0.0124 |
49
- | 0.0059 | 5.33 | 4000 | 0.0131 |
50
- | 0.005 | 8.0 | 6000 | 0.0146 |
51
- | 0.0032 | 10.67 | 8000 | 0.0167 |
52
- | 0.0027 | 13.33 | 10000 | 0.0177 |
53
- | 0.0019 | 16.0 | 12000 | 0.0183 |
54
- | 0.0014 | 18.67 | 14000 | 0.0191 |
55
 
56
 
57
  ### Framework versions
58
 
59
- - Transformers 4.27.3
60
- - Pytorch 1.13.1+cu116
61
- - Datasets 2.10.1
62
- - Tokenizers 0.13.2
 
13
  # t5-base
14
 
15
  This model is a fine-tuned version of [t5-base](https://huggingface.co/t5-base) on the None dataset.
 
 
16
 
17
  ## Model description
18
 
 
33
  The following hyperparameters were used during training:
34
  - learning_rate: 0.0001
35
  - train_batch_size: 8
36
+ - eval_batch_size: 32
37
  - seed: 42
38
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
39
  - lr_scheduler_type: linear
40
+ - num_epochs: 10
41
 
42
  ### Training results
43
 
 
 
 
 
 
 
 
 
 
44
 
45
 
46
  ### Framework versions
47
 
48
+ - Transformers 4.28.1
49
+ - Pytorch 2.0.0+cu118
50
+ - Datasets 2.11.0
51
+ - Tokenizers 0.13.3