krm commited on
Commit
7dca35c
·
1 Parent(s): a1c77a2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +13 -10
README.md CHANGED
@@ -15,8 +15,8 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  # BARTkrame-abstract
17
 
18
- This model is a fine-tuned version of [krm/BARTkrame-abstract](https://huggingface.co/krm/BARTkrame-abstract) on the None dataset.
19
- It achieves the following results on the evaluation set:
20
  - Loss: 2.4196
21
  - Rouge1: 0.2703
22
  - Rouge2: 0.1334
@@ -25,15 +25,18 @@ It achieves the following results on the evaluation set:
25
 
26
  ## Model description
27
 
28
- More information needed
29
 
30
  ## Intended uses & limitations
31
 
32
  More information needed
33
 
34
  ## Training and evaluation data
 
35
 
36
- More information needed
 
 
37
 
38
  ## Training procedure
39
 
@@ -44,18 +47,18 @@ The following hyperparameters were used during training:
44
  - train_batch_size: 4
45
  - eval_batch_size: 4
46
  - seed: 42
47
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
48
  - lr_scheduler_type: linear
49
- - num_epochs: 4
50
 
51
  ### Training results
52
 
53
  | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum |
54
  |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|
55
- | 0.1316 | 1.0 | 1250 | 2.3251 | 0.2505 | 0.1158 | 0.2150 | 0.2184 |
56
- | 0.0894 | 2.0 | 2500 | 2.3467 | 0.2526 | 0.1073 | 0.2067 | 0.2124 |
57
- | 0.045 | 3.0 | 3750 | 2.3742 | 0.2593 | 0.1211 | 0.2281 | 0.2308 |
58
- | 0.0242 | 4.0 | 5000 | 2.4196 | 0.2703 | 0.1334 | 0.2392 | 0.2419 |
59
 
60
 
61
  ### Framework versions
 
15
 
16
  # BARTkrame-abstract
17
 
18
+ This model is a fine-tuned version of [krm/BARTkrame-abstract](https://huggingface.co/krm/BARTkrame-abstract) on the [krm/for-ULPGL-Dissertation](https://huggingface.co/datasets/krm/for-ULPGL-Dissertation) dataset.
19
+ It achieves (15/10/2022) the following results on the evaluation set:
20
  - Loss: 2.4196
21
  - Rouge1: 0.2703
22
  - Rouge2: 0.1334
 
25
 
26
  ## Model description
27
 
28
+ This model is primarly a finetuned version of [moussaKam/mbarthez](https://huggingface.co/moussaKam/mbarthez).
29
 
30
  ## Intended uses & limitations
31
 
32
  More information needed
33
 
34
  ## Training and evaluation data
35
+ We have used the [krm/for-ULPGL-Dissertation](https://huggingface.co/datasets/krm/for-ULPGL-Dissertation) dataset reduced to :
36
 
37
+ > **Training data :** **5000** samples taken at random with *seed=42*.
38
+
39
+ > **Validation data :** **100** samples taken at random with *seed=42*.
40
 
41
  ## Training procedure
42
 
 
47
  - train_batch_size: 4
48
  - eval_batch_size: 4
49
  - seed: 42
50
+ - optimizer: Adam with betas=(0.9, 0.999) and epsilon=1e-08
51
  - lr_scheduler_type: linear
52
+ - num_epochs: 12
53
 
54
  ### Training results
55
 
56
  | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum |
57
  |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|
58
+ | 0.1316 | 9.0 | 1250 | 2.3251 | 0.2505 | 0.1158 | 0.2150 | 0.2184 |
59
+ | 0.0894 | 10.0 | 2500 | 2.3467 | 0.2526 | 0.1073 | 0.2067 | 0.2124 |
60
+ | 0.045 | 11.0 | 3750 | 2.3742 | 0.2593 | 0.1211 | 0.2281 | 0.2308 |
61
+ | 0.0242 | 12.0 | 5000 | 2.4196 | 0.2703 | 0.1334 | 0.2392 | 0.2419 |
62
 
63
 
64
  ### Framework versions