angarcme commited on
Commit
dfa0e3a
·
verified ·
1 Parent(s): a94751c

Training complete

Browse files
Files changed (2) hide show
  1. README.md +9 -13
  2. generation_config.json +3 -8
README.md CHANGED
@@ -1,7 +1,7 @@
1
  ---
2
  library_name: transformers
3
- license: mit
4
- base_model: facebook/mbart-large-50
5
  tags:
6
  - simplification
7
  - generated_from_trainer
@@ -10,10 +10,6 @@ metrics:
10
  model-index:
11
  - name: mbart-neutralization
12
  results: []
13
- language:
14
- - es
15
- datasets:
16
- - somosnlp-hackathon-2022/neutral-es
17
  ---
18
 
19
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -21,11 +17,11 @@ should probably proofread and complete it, then remove this comment. -->
21
 
22
  # mbart-neutralization
23
 
24
- This model is a fine-tuned version of [facebook/mbart-large-50](https://huggingface.co/facebook/mbart-large-50) on somosnlp-hackathon-2022/neutral-es dataset.
25
  It achieves the following results on the evaluation set:
26
- - Loss: 0.0161
27
- - Bleu: 98.4007
28
- - Gen Len: 18.5625
29
 
30
  ## Model description
31
 
@@ -56,8 +52,8 @@ The following hyperparameters were used during training:
56
 
57
  | Training Loss | Epoch | Step | Validation Loss | Bleu | Gen Len |
58
  |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|
59
- | No log | 1.0 | 440 | 0.0411 | 97.9038 | 18.7396 |
60
- | 0.2238 | 2.0 | 880 | 0.0161 | 98.4007 | 18.5625 |
61
 
62
 
63
  ### Framework versions
@@ -65,4 +61,4 @@ The following hyperparameters were used during training:
65
  - Transformers 4.51.2
66
  - Pytorch 2.10.0+cu128
67
  - Datasets 4.0.0
68
- - Tokenizers 0.21.4
 
1
  ---
2
  library_name: transformers
3
+ license: apache-2.0
4
+ base_model: google/mt5-small
5
  tags:
6
  - simplification
7
  - generated_from_trainer
 
10
  model-index:
11
  - name: mbart-neutralization
12
  results: []
 
 
 
 
13
  ---
14
 
15
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
17
 
18
  # mbart-neutralization
19
 
20
+ This model is a fine-tuned version of [google/mt5-small](https://huggingface.co/google/mt5-small) on an unknown dataset.
21
  It achieves the following results on the evaluation set:
22
+ - Loss: 0.3279
23
+ - Bleu: 72.1306
24
+ - Gen Len: 17.1875
25
 
26
  ## Model description
27
 
 
52
 
53
  | Training Loss | Epoch | Step | Validation Loss | Bleu | Gen Len |
54
  |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|
55
+ | No log | 1.0 | 440 | 0.3966 | 66.1686 | 16.4583 |
56
+ | 0.9189 | 2.0 | 880 | 0.3279 | 72.1306 | 17.1875 |
57
 
58
 
59
  ### Framework versions
 
61
  - Transformers 4.51.2
62
  - Pytorch 2.10.0+cu128
63
  - Datasets 4.0.0
64
+ - Tokenizers 0.21.4
generation_config.json CHANGED
@@ -1,11 +1,6 @@
1
  {
2
- "bos_token_id": 0,
3
- "decoder_start_token_id": 2,
4
- "early_stopping": true,
5
- "eos_token_id": 2,
6
- "forced_eos_token_id": 2,
7
- "max_length": 200,
8
- "num_beams": 5,
9
- "pad_token_id": 1,
10
  "transformers_version": "4.51.2"
11
  }
 
1
  {
2
+ "decoder_start_token_id": 0,
3
+ "eos_token_id": 1,
4
+ "pad_token_id": 0,
 
 
 
 
 
5
  "transformers_version": "4.51.2"
6
  }