Kimang18 commited on
Commit
24c890d
·
verified ·
1 Parent(s): 3ff8cf5

End of training

Browse files
README.md ADDED
@@ -0,0 +1,83 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ language:
4
+ - khm
5
+ license: cc-by-nc-4.0
6
+ base_model: facebook/mms-1b-fl102
7
+ tags:
8
+ - generated_from_trainer
9
+ datasets:
10
+ - seanghay/khmer-mpwt-speech
11
+ metrics:
12
+ - wer
13
+ model-index:
14
+ - name: mms-1b-finetuned-khm
15
+ results:
16
+ - task:
17
+ name: Automatic Speech Recognition
18
+ type: automatic-speech-recognition
19
+ dataset:
20
+ name: seanghay/khmer-mpwt-speech
21
+ type: seanghay/khmer-mpwt-speech
22
+ metrics:
23
+ - name: Wer
24
+ type: wer
25
+ value: 0.627763041556145
26
+ ---
27
+
28
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
29
+ should probably proofread and complete it, then remove this comment. -->
30
+
31
+ # mms-1b-finetuned-khm
32
+
33
+ This model is a fine-tuned version of [facebook/mms-1b-fl102](https://huggingface.co/facebook/mms-1b-fl102) on the seanghay/khmer-mpwt-speech dataset.
34
+ It achieves the following results on the evaluation set:
35
+ - Loss: 0.5778
36
+ - Wer: 0.6278
37
+
38
+ ## Model description
39
+
40
+ More information needed
41
+
42
+ ## Intended uses & limitations
43
+
44
+ More information needed
45
+
46
+ ## Training and evaluation data
47
+
48
+ More information needed
49
+
50
+ ## Training procedure
51
+
52
+ ### Training hyperparameters
53
+
54
+ The following hyperparameters were used during training:
55
+ - learning_rate: 0.001
56
+ - train_batch_size: 8
57
+ - eval_batch_size: 8
58
+ - seed: 42
59
+ - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
60
+ - lr_scheduler_type: linear
61
+ - lr_scheduler_warmup_steps: 100
62
+ - num_epochs: 4
63
+ - mixed_precision_training: Native AMP
64
+
65
+ ### Training results
66
+
67
+ | Training Loss | Epoch | Step | Validation Loss | Wer |
68
+ |:-------------:|:------:|:----:|:---------------:|:------:|
69
+ | 4.8181 | 0.5587 | 100 | 1.0119 | 0.9584 |
70
+ | 1.1491 | 1.1173 | 200 | 0.7723 | 0.7366 |
71
+ | 0.9737 | 1.6760 | 300 | 0.6707 | 0.7366 |
72
+ | 0.8471 | 2.2346 | 400 | 0.6370 | 0.7019 |
73
+ | 0.8114 | 2.7933 | 500 | 0.5789 | 0.6863 |
74
+ | 0.6951 | 3.3520 | 600 | 0.5461 | 0.6464 |
75
+ | 0.6791 | 3.9106 | 700 | 0.5406 | 0.6672 |
76
+
77
+
78
+ ### Framework versions
79
+
80
+ - Transformers 4.57.0.dev0
81
+ - Pytorch 2.8.0+cu126
82
+ - Datasets 4.0.0
83
+ - Tokenizers 0.22.1
adapter.khm.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d4d9789e1b5513ebe7565dcbf8e1a285e186bbedac32e9dab3df7d71e90c181
3
+ size 9070120
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:85766f6e495ea262a34b337f6c096e3be97bd517e5261347570d64375c5c8e16
3
- size 3863597792
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17512882362f3b59fb489d06549340b84be532718846df076e357a6e2666cde4
3
+ size 3859162496
runs/Oct09_06-48-07_8c60993979e6/events.out.tfevents.1759994865.8c60993979e6.926.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ad1f50c097dbc0f83bac363379043526c7de3ce03f9872f16a4dc0654269220
3
+ size 406