SOULAMA commited on
Commit
cc69bfa
·
verified ·
1 Parent(s): 9942700

Create Readme.md

Browse files
Files changed (1) hide show
  1. last-checkpoint/Readme.md +89 -0
last-checkpoint/Readme.md ADDED
@@ -0,0 +1,89 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - fr
5
+ - en
6
+ metrics:
7
+ - chrf
8
+ - sacrebleu
9
+ base_model:
10
+ - google-t5/t5-small
11
+ pipeline_tag: translation
12
+ ---
13
+
14
+ ---
15
+ library_name: transformers
16
+ license: apache-2.0
17
+ base_model: t5-small
18
+ tags:
19
+ - generated_from_trainer
20
+ - translation
21
+ - t5
22
+ - english-to-french
23
+ model-index:
24
+ - name: t5-fra-eng
25
+ results:
26
+ - task: translation
27
+ dataset: Custom English-French dataset https://huggingface.co/datasets/SOULAMA/eng-fra-dataset
28
+
29
+ ---
30
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
31
+ should probably proofread and complete it, then remove this comment. -->
32
+
33
+ # t5-eng-fra
34
+
35
+ This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on a French-English translation dataset.
36
+
37
+ ## Model description
38
+
39
+ `t5-fra-eng` is a sequence-to-sequence model designed to translate French text into English. It was fine-tuned from `t5-small` using Hugging Face’s `Seq2SeqTrainer` on a custom French-English dataset. The model leverages T5’s transformer architecture and the text-to-text paradigm for translation tasks.
40
+
41
+ ## Intended uses & limitations
42
+
43
+ **Intended uses:**
44
+ - Automatic translation of English text into French.
45
+ - Machine translation research and experimentation.
46
+
47
+ **Limitations:**
48
+ - The model may produce incorrect translations for idiomatic expressions or rare vocabulary.
49
+ - Not suitable for legal, medical, or critical translations without human verification.
50
+ - Performance depends on the quality and size of the fine-tuning dataset.
51
+
52
+ ## Training and evaluation data
53
+
54
+ - Dataset: Custom French-English parallel sentences.
55
+ - Split: 80% training, 20% test from dataset.
56
+ - Split 80% training, 20% test from trainset
57
+ - Data preprocessing: Text normalized, tokenized using `t5-small` tokenizer, maximum input length = 128 tokens, maximum target length = 128 tokens.
58
+
59
+ ## Training procedure
60
+
61
+ ### Training hyperparameters
62
+
63
+ The following hyperparameters were used during training:
64
+
65
+ - learning_rate: 2e-5
66
+ - per_device_train_batch_size: 16
67
+ - per_device_eval_batch_size: 16
68
+ - num_train_epochs: 3
69
+ - weight_decay: 0.01
70
+ - optimizer: AdamW (betas=(0.9, 0.999), epsilon=1e-8)
71
+ - lr_scheduler_type: linear
72
+ - mixed_precision_training: Native AMP
73
+ - seed: 42
74
+ - evaluation_strategy: epoch
75
+ - save_total_limit: 3
76
+ - predict_with_generate: True
77
+
78
+ ### Training results
79
+
80
+ - 'eval_loss': 0.5946913957595825,
81
+ - 'eval_bleu': 42.4875053753255,
82
+ - 'eval_chrf': 61.82547115972855,}
83
+
84
+ ### Framework versions
85
+
86
+ - Transformers 4.57.3
87
+ - PyTorch 2.6.0+cu126
88
+ - Datasets 3.6.0
89
+ - Tokenizers 0.22.1