kavyamanohar commited on
Commit
e53a6ef
·
verified ·
1 Parent(s): c25989c

End of training

Browse files
README.md ADDED
@@ -0,0 +1,81 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ license: apache-2.0
4
+ base_model: parambharat/whisper-small-ta
5
+ tags:
6
+ - generated_from_trainer
7
+ model-index:
8
+ - name: MTF-ta-en-translation
9
+ results: []
10
+ ---
11
+
12
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
+ should probably proofread and complete it, then remove this comment. -->
14
+
15
+ # MTF-ta-en-translation
16
+
17
+ This model is a fine-tuned version of [parambharat/whisper-small-ta](https://huggingface.co/parambharat/whisper-small-ta) on an unknown dataset.
18
+ It achieves the following results on the evaluation set:
19
+ - Loss: 0.1324
20
+ - Bleu Score: 0.0299
21
+
22
+ ## Model description
23
+
24
+ More information needed
25
+
26
+ ## Intended uses & limitations
27
+
28
+ More information needed
29
+
30
+ ## Training and evaluation data
31
+
32
+ More information needed
33
+
34
+ ## Training procedure
35
+
36
+ ### Training hyperparameters
37
+
38
+ The following hyperparameters were used during training:
39
+ - learning_rate: 1e-05
40
+ - train_batch_size: 8
41
+ - eval_batch_size: 8
42
+ - seed: 42
43
+ - gradient_accumulation_steps: 2
44
+ - total_train_batch_size: 16
45
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
46
+ - lr_scheduler_type: linear
47
+ - training_steps: 5000
48
+ - mixed_precision_training: Native AMP
49
+
50
+ ### Training results
51
+
52
+ | Training Loss | Epoch | Step | Validation Loss | Bleu Score |
53
+ |:-------------:|:-------:|:----:|:---------------:|:----------:|
54
+ | 0.0939 | 2.9412 | 250 | 0.0864 | 0.0302 |
55
+ | 0.0227 | 5.8824 | 500 | 0.0998 | 0.0301 |
56
+ | 0.0048 | 8.8235 | 750 | 0.1083 | 0.0340 |
57
+ | 0.001 | 11.7647 | 1000 | 0.1132 | 0.0312 |
58
+ | 0.0005 | 14.7059 | 1250 | 0.1164 | 0.0308 |
59
+ | 0.0003 | 17.6471 | 1500 | 0.1189 | 0.0322 |
60
+ | 0.0002 | 20.5882 | 1750 | 0.1208 | 0.0311 |
61
+ | 0.0002 | 23.5294 | 2000 | 0.1225 | 0.0307 |
62
+ | 0.0002 | 26.4706 | 2250 | 0.1242 | 0.0334 |
63
+ | 0.0001 | 29.4118 | 2500 | 0.1256 | 0.0321 |
64
+ | 0.0001 | 32.3529 | 2750 | 0.1268 | 0.0327 |
65
+ | 0.0001 | 35.2941 | 3000 | 0.1277 | 0.0324 |
66
+ | 0.0001 | 38.2353 | 3250 | 0.1286 | 0.0311 |
67
+ | 0.0001 | 41.1765 | 3500 | 0.1295 | 0.0309 |
68
+ | 0.0001 | 44.1176 | 3750 | 0.1302 | 0.0311 |
69
+ | 0.0001 | 47.0588 | 4000 | 0.1308 | 0.0310 |
70
+ | 0.0001 | 50.0 | 4250 | 0.1314 | 0.0313 |
71
+ | 0.0001 | 52.9412 | 4500 | 0.1320 | 0.0298 |
72
+ | 0.0001 | 55.8824 | 4750 | 0.1323 | 0.0299 |
73
+ | 0.0 | 58.8235 | 5000 | 0.1324 | 0.0299 |
74
+
75
+
76
+ ### Framework versions
77
+
78
+ - Transformers 4.48.3
79
+ - Pytorch 2.6.0+cu124
80
+ - Datasets 3.2.0
81
+ - Tokenizers 0.21.0
generation_config.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "begin_suppress_tokens": [
3
+ 220,
4
+ 50257
5
+ ],
6
+ "bos_token_id": 50257,
7
+ "decoder_start_token_id": 50258,
8
+ "eos_token_id": 50257,
9
+ "forced_decoder_ids": [
10
+ [
11
+ 1,
12
+ 50287
13
+ ],
14
+ [
15
+ 2,
16
+ 50358
17
+ ],
18
+ [
19
+ 3,
20
+ 50363
21
+ ]
22
+ ],
23
+ "max_length": 448,
24
+ "pad_token_id": 50257,
25
+ "transformers_version": "4.48.3",
26
+ "use_cache": false
27
+ }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0e8a0f5bd2ccb2b83f0a9ed317150f5fa5223fe018b0077267057441672a75d5
3
  size 966995080
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1406aee65c0f2115ef1544e9d398bebdebc3986310b8f975aa6a0d47ee9ad9c8
3
  size 966995080
runs/Feb22_11-57-24_kudsit-dgxserver/events.out.tfevents.1740205644.kudsit-dgxserver.1352970.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ac345bbd11b9c539fab02f41ed325c648cf0b0ebbebbea5a2824fdc79085f61b
3
- size 16234
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:615a95b0822d4f1b71c9e0c7b8e541adb33bf9cafab23016c9f43bf2b5aacb4b
3
+ size 17124