qBob commited on
Commit
da35e56
·
1 Parent(s): 69415cb

update model card README.md

Browse files
Files changed (1) hide show
  1. README.md +14 -11
README.md CHANGED
@@ -16,12 +16,12 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  This model is a fine-tuned version of [ainize/bart-base-cnn](https://huggingface.co/ainize/bart-base-cnn) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
- - Loss: 0.0135
20
- - Rouge1: 81.7278
21
- - Rouge2: 80.2103
22
- - Rougel: 81.7237
23
- - Rougelsum: 81.7112
24
- - Gen Len: 19.434
25
 
26
  ## Model description
27
 
@@ -41,24 +41,27 @@ More information needed
41
 
42
  The following hyperparameters were used during training:
43
  - learning_rate: 2e-05
44
- - train_batch_size: 2
45
- - eval_batch_size: 2
46
  - seed: 42
47
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
48
  - lr_scheduler_type: linear
49
- - num_epochs: 1
50
  - mixed_precision_training: Native AMP
51
 
52
  ### Training results
53
 
54
  | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
55
  |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:|
56
- | 0.0121 | 1.0 | 500 | 0.0135 | 81.7278 | 80.2103 | 81.7237 | 81.7112 | 19.434 |
 
 
 
57
 
58
 
59
  ### Framework versions
60
 
61
  - Transformers 4.21.1
62
- - Pytorch 1.12.0+cu113
63
  - Datasets 2.4.0
64
  - Tokenizers 0.12.1
 
16
 
17
  This model is a fine-tuned version of [ainize/bart-base-cnn](https://huggingface.co/ainize/bart-base-cnn) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
+ - Loss: 0.0025
20
+ - Rouge1: 81.4214
21
+ - Rouge2: 80.2027
22
+ - Rougel: 81.4202
23
+ - Rougelsum: 81.4241
24
+ - Gen Len: 19.3962
25
 
26
  ## Model description
27
 
 
41
 
42
  The following hyperparameters were used during training:
43
  - learning_rate: 2e-05
44
+ - train_batch_size: 8
45
+ - eval_batch_size: 8
46
  - seed: 42
47
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
48
  - lr_scheduler_type: linear
49
+ - num_epochs: 4
50
  - mixed_precision_training: Native AMP
51
 
52
  ### Training results
53
 
54
  | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
55
  |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:|
56
+ | 0.0071 | 1.0 | 2365 | 0.0039 | 81.3664 | 80.0861 | 81.3601 | 81.3667 | 19.3967 |
57
+ | 0.0033 | 2.0 | 4730 | 0.0029 | 81.3937 | 80.1548 | 81.3902 | 81.3974 | 19.3961 |
58
+ | 0.0018 | 3.0 | 7095 | 0.0029 | 81.3838 | 80.1404 | 81.385 | 81.3878 | 19.3965 |
59
+ | 0.001 | 4.0 | 9460 | 0.0025 | 81.4214 | 80.2027 | 81.4202 | 81.4241 | 19.3962 |
60
 
61
 
62
  ### Framework versions
63
 
64
  - Transformers 4.21.1
65
+ - Pytorch 1.12.1+cu113
66
  - Datasets 2.4.0
67
  - Tokenizers 0.12.1