Yirui091 commited on
Commit
0fdbdaa
·
verified ·
1 Parent(s): 0fe4d3a

Yirui091/lab1_finetuning

Browse files
Files changed (2) hide show
  1. README.md +9 -9
  2. generation_config.json +3 -4
README.md CHANGED
@@ -9,7 +9,7 @@ datasets:
9
  metrics:
10
  - bleu
11
  model-index:
12
- - name: marian-finetuned-kde4-en-to-fr
13
  results:
14
  - task:
15
  name: Sequence-to-sequence Language Modeling
@@ -23,19 +23,19 @@ model-index:
23
  metrics:
24
  - name: Bleu
25
  type: bleu
26
- value: 48.93253657516627
27
  ---
28
 
29
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
30
  should probably proofread and complete it, then remove this comment. -->
31
 
32
- # marian-finetuned-kde4-en-to-fr
33
 
34
  This model is a fine-tuned version of [Helsinki-NLP/opus-mt-en-fr](https://huggingface.co/Helsinki-NLP/opus-mt-en-fr) on the kde4 dataset.
35
  It achieves the following results on the evaluation set:
36
- - Loss: 1.0256
37
- - Model Preparation Time: 0.0267
38
- - Bleu: 48.9325
39
 
40
  ## Model description
41
 
@@ -58,7 +58,7 @@ The following hyperparameters were used during training:
58
  - train_batch_size: 16
59
  - eval_batch_size: 32
60
  - seed: 42
61
- - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
62
  - lr_scheduler_type: linear
63
  - training_steps: 5000
64
  - mixed_precision_training: Native AMP
@@ -69,7 +69,7 @@ The following hyperparameters were used during training:
69
 
70
  ### Framework versions
71
 
72
- - Transformers 4.57.0
73
  - Pytorch 2.10.0+cu128
74
  - Datasets 2.18.0
75
- - Tokenizers 0.22.2
 
9
  metrics:
10
  - bleu
11
  model-index:
12
+ - name: lab1_finetuning
13
  results:
14
  - task:
15
  name: Sequence-to-sequence Language Modeling
 
23
  metrics:
24
  - name: Bleu
25
  type: bleu
26
+ value: 48.52734685570538
27
  ---
28
 
29
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
30
  should probably proofread and complete it, then remove this comment. -->
31
 
32
+ # lab1_finetuning
33
 
34
  This model is a fine-tuned version of [Helsinki-NLP/opus-mt-en-fr](https://huggingface.co/Helsinki-NLP/opus-mt-en-fr) on the kde4 dataset.
35
  It achieves the following results on the evaluation set:
36
+ - Loss: 1.0255
37
+ - Model Preparation Time: 0.003
38
+ - Bleu: 48.5273
39
 
40
  ## Model description
41
 
 
58
  - train_batch_size: 16
59
  - eval_batch_size: 32
60
  - seed: 42
61
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
62
  - lr_scheduler_type: linear
63
  - training_steps: 5000
64
  - mixed_precision_training: Native AMP
 
69
 
70
  ### Framework versions
71
 
72
+ - Transformers 4.45.2
73
  - Pytorch 2.10.0+cu128
74
  - Datasets 2.18.0
75
+ - Tokenizers 0.20.3
generation_config.json CHANGED
@@ -4,14 +4,13 @@
4
  59513
5
  ]
6
  ],
 
7
  "decoder_start_token_id": 59513,
8
- "eos_token_id": [
9
- 0
10
- ],
11
  "forced_eos_token_id": 0,
12
  "max_length": 512,
13
  "num_beams": 4,
14
  "pad_token_id": 59513,
15
  "renormalize_logits": true,
16
- "transformers_version": "4.57.0"
17
  }
 
4
  59513
5
  ]
6
  ],
7
+ "bos_token_id": 0,
8
  "decoder_start_token_id": 59513,
9
+ "eos_token_id": 0,
 
 
10
  "forced_eos_token_id": 0,
11
  "max_length": 512,
12
  "num_beams": 4,
13
  "pad_token_id": 59513,
14
  "renormalize_logits": true,
15
+ "transformers_version": "4.45.2"
16
  }