Yirui091 commited on
Commit
a3d58ce
·
verified ·
1 Parent(s): 006190d

Yirui091/lab1_random

Browse files
Files changed (2) hide show
  1. README.md +7 -7
  2. generation_config.json +3 -4
README.md CHANGED
@@ -23,7 +23,7 @@ model-index:
23
  metrics:
24
  - name: Bleu
25
  type: bleu
26
- value: 6.636251305602728
27
  ---
28
 
29
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -33,9 +33,9 @@ should probably proofread and complete it, then remove this comment. -->
33
 
34
  This model is a fine-tuned version of [Helsinki-NLP/opus-mt-en-fr](https://huggingface.co/Helsinki-NLP/opus-mt-en-fr) on the kde4 dataset.
35
  It achieves the following results on the evaluation set:
36
- - Loss: 5.2567
37
- - Model Preparation Time: 0.0113
38
- - Bleu: 6.6363
39
 
40
  ## Model description
41
 
@@ -58,7 +58,7 @@ The following hyperparameters were used during training:
58
  - train_batch_size: 16
59
  - eval_batch_size: 32
60
  - seed: 42
61
- - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
62
  - lr_scheduler_type: linear
63
  - training_steps: 5000
64
  - mixed_precision_training: Native AMP
@@ -69,7 +69,7 @@ The following hyperparameters were used during training:
69
 
70
  ### Framework versions
71
 
72
- - Transformers 4.57.0
73
  - Pytorch 2.10.0+cu128
74
  - Datasets 2.18.0
75
- - Tokenizers 0.22.2
 
23
  metrics:
24
  - name: Bleu
25
  type: bleu
26
+ value: 6.9589599042040895
27
  ---
28
 
29
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
33
 
34
  This model is a fine-tuned version of [Helsinki-NLP/opus-mt-en-fr](https://huggingface.co/Helsinki-NLP/opus-mt-en-fr) on the kde4 dataset.
35
  It achieves the following results on the evaluation set:
36
+ - Loss: 5.2435
37
+ - Model Preparation Time: 0.0032
38
+ - Bleu: 6.9590
39
 
40
  ## Model description
41
 
 
58
  - train_batch_size: 16
59
  - eval_batch_size: 32
60
  - seed: 42
61
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
62
  - lr_scheduler_type: linear
63
  - training_steps: 5000
64
  - mixed_precision_training: Native AMP
 
69
 
70
  ### Framework versions
71
 
72
+ - Transformers 4.45.2
73
  - Pytorch 2.10.0+cu128
74
  - Datasets 2.18.0
75
+ - Tokenizers 0.20.3
generation_config.json CHANGED
@@ -4,13 +4,12 @@
4
  59513
5
  ]
6
  ],
 
7
  "decoder_start_token_id": 59513,
8
- "eos_token_id": [
9
- 0
10
- ],
11
  "forced_eos_token_id": 0,
12
  "max_length": 512,
13
  "num_beams": 4,
14
  "pad_token_id": 59513,
15
- "transformers_version": "4.57.0"
16
  }
 
4
  59513
5
  ]
6
  ],
7
+ "bos_token_id": 0,
8
  "decoder_start_token_id": 59513,
9
+ "eos_token_id": 0,
 
 
10
  "forced_eos_token_id": 0,
11
  "max_length": 512,
12
  "num_beams": 4,
13
  "pad_token_id": 59513,
14
+ "transformers_version": "4.45.2"
15
  }