Floyd93 commited on
Commit
ece2a87
·
verified ·
1 Parent(s): 846c689

End of training

Browse files
README.md CHANGED
@@ -1,8 +1,8 @@
1
  ---
2
- license: apache-2.0
3
- base_model: t5-small
4
  tags:
5
  - generated_from_trainer
 
 
6
  model-index:
7
  - name: Translation_Grammer_Jan_2024
8
  results: []
@@ -13,16 +13,11 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # Translation_Grammer_Jan_2024
15
 
16
- This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on the None dataset.
17
  It achieves the following results on the evaluation set:
18
- - eval_loss: 0.0269
19
- - eval_bleu: 19.9909
20
- - eval_gen_len: 18.2914
21
- - eval_runtime: 864.0001
22
- - eval_samples_per_second: 115.741
23
- - eval_steps_per_second: 3.617
24
- - epoch: 20.0
25
- - step: 250000
26
 
27
  ## Model description
28
 
@@ -42,16 +37,24 @@ More information needed
42
 
43
  The following hyperparameters were used during training:
44
  - learning_rate: 2e-05
45
- - train_batch_size: 32
46
- - eval_batch_size: 32
47
  - seed: 42
48
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
  - lr_scheduler_type: linear
50
- - num_epochs: 30
 
 
 
 
 
 
 
 
51
 
52
  ### Framework versions
53
 
54
- - Transformers 4.35.2
55
- - Pytorch 2.1.0+cu121
56
  - Datasets 2.16.1
57
- - Tokenizers 0.15.1
 
1
  ---
 
 
2
  tags:
3
  - generated_from_trainer
4
+ metrics:
5
+ - bleu
6
  model-index:
7
  - name: Translation_Grammer_Jan_2024
8
  results: []
 
13
 
14
  # Translation_Grammer_Jan_2024
15
 
16
+ This model was trained from scratch on the None dataset.
17
  It achieves the following results on the evaluation set:
18
+ - Loss: 0.0440
19
+ - Bleu: 20.0
20
+ - Gen Len: 18.2937
 
 
 
 
 
21
 
22
  ## Model description
23
 
 
37
 
38
  The following hyperparameters were used during training:
39
  - learning_rate: 2e-05
40
+ - train_batch_size: 128
41
+ - eval_batch_size: 128
42
  - seed: 42
43
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
44
  - lr_scheduler_type: linear
45
+ - num_epochs: 1
46
+ - mixed_precision_training: Native AMP
47
+
48
+ ### Training results
49
+
50
+ | Training Loss | Epoch | Step | Validation Loss | Bleu | Gen Len |
51
+ |:-------------:|:-----:|:-----:|:---------------:|:----:|:-------:|
52
+ | 0.0618 | 1.0 | 12530 | 0.0440 | 20.0 | 18.2937 |
53
+
54
 
55
  ### Framework versions
56
 
57
+ - Transformers 4.36.2
58
+ - Pytorch 2.1.1+cu118
59
  - Datasets 2.16.1
60
+ - Tokenizers 0.15.0
generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "decoder_start_token_id": 0,
3
  "eos_token_id": 1,
4
  "pad_token_id": 0,
5
- "transformers_version": "4.35.2"
6
  }
 
2
  "decoder_start_token_id": 0,
3
  "eos_token_id": 1,
4
  "pad_token_id": 0,
5
+ "transformers_version": "4.36.2"
6
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:12c7a3655e2e34bc538d7ca5506d00eb5556e5419411cb460d90b8cc3cf76c74
3
  size 242041896
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fbd5a2128e7aeed55890a1adea86e427487f0eb2ceb022f55cade0a0b09896e9
3
  size 242041896
runs/Jan31_00-14-06_sirius-1.lyon.grid5000.fr/events.out.tfevents.1706656447.sirius-1.lyon.grid5000.fr.10797.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:79bbc1643d245c034f8d32e8359559da321fb42261ab6206e674665b6fb10cb5
3
- size 9421
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8d03af83b2fbc4ca3490bafd96dc5d5a97bcd2ec812bd1c2c587c1a35fa9e9f
3
+ size 10145