JYumeko commited on
Commit
10e1b54
·
1 Parent(s): ee16616

update model card README.md

Browse files
Files changed (1) hide show
  1. README.md +25 -16
README.md CHANGED
@@ -2,8 +2,8 @@
2
  license: apache-2.0
3
  tags:
4
  - generated_from_trainer
5
- datasets:
6
- - scientific_lay_summarisation
7
  model-index:
8
  - name: summarization_model
9
  results: []
@@ -14,19 +14,14 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  # summarization_model
16
 
17
- This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on the scientific_lay_summarisation dataset.
18
  It achieves the following results on the evaluation set:
19
- - eval_loss: 2.7911
20
- - eval_rouge1: 0.1518
21
- - eval_rouge2: 0.0308
22
- - eval_rougeL: 0.1126
23
- - eval_rougeLsum: 0.1127
24
- - eval_gen_len: 19.0
25
- - eval_runtime: 16.8648
26
- - eval_samples_per_second: 14.29
27
- - eval_steps_per_second: 0.949
28
- - epoch: 5.51
29
- - step: 1500
30
 
31
  ## Model description
32
 
@@ -51,12 +46,26 @@ The following hyperparameters were used during training:
51
  - seed: 42
52
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
53
  - lr_scheduler_type: linear
54
- - num_epochs: 32
55
  - mixed_precision_training: Native AMP
56
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
57
  ### Framework versions
58
 
59
- - Transformers 4.27.4
60
  - Pytorch 2.0.0+cu118
61
  - Datasets 2.11.0
62
  - Tokenizers 0.13.3
 
2
  license: apache-2.0
3
  tags:
4
  - generated_from_trainer
5
+ metrics:
6
+ - rouge
7
  model-index:
8
  - name: summarization_model
9
  results: []
 
14
 
15
  # summarization_model
16
 
17
+ This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on the None dataset.
18
  It achieves the following results on the evaluation set:
19
+ - Loss: 0.1359
20
+ - Rouge1: 0.1813
21
+ - Rouge2: 0.1114
22
+ - Rougel: 0.1616
23
+ - Rougelsum: 0.1617
24
+ - Gen Len: 19.0
 
 
 
 
 
25
 
26
  ## Model description
27
 
 
46
  - seed: 42
47
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
48
  - lr_scheduler_type: linear
49
+ - num_epochs: 8
50
  - mixed_precision_training: Native AMP
51
 
52
+ ### Training results
53
+
54
+ | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
55
+ |:-------------:|:-----:|:-----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:|
56
+ | 0.2358 | 1.0 | 1635 | 0.1719 | 0.1758 | 0.1033 | 0.1554 | 0.1554 | 19.0 |
57
+ | 0.2043 | 2.0 | 3270 | 0.1574 | 0.1764 | 0.1046 | 0.1561 | 0.1561 | 19.0 |
58
+ | 0.191 | 3.0 | 4905 | 0.1505 | 0.1778 | 0.1069 | 0.1577 | 0.1578 | 19.0 |
59
+ | 0.178 | 4.0 | 6540 | 0.1448 | 0.1797 | 0.1093 | 0.1597 | 0.1597 | 19.0 |
60
+ | 0.1734 | 5.0 | 8175 | 0.1406 | 0.1804 | 0.1102 | 0.1605 | 0.1604 | 19.0 |
61
+ | 0.1681 | 6.0 | 9810 | 0.1376 | 0.1811 | 0.111 | 0.1613 | 0.1613 | 19.0 |
62
+ | 0.1665 | 7.0 | 11445 | 0.1365 | 0.1815 | 0.1114 | 0.1618 | 0.1618 | 19.0 |
63
+ | 0.1643 | 8.0 | 13080 | 0.1359 | 0.1813 | 0.1114 | 0.1616 | 0.1617 | 19.0 |
64
+
65
+
66
  ### Framework versions
67
 
68
+ - Transformers 4.28.0
69
  - Pytorch 2.0.0+cu118
70
  - Datasets 2.11.0
71
  - Tokenizers 0.13.3