Priyanship commited on
Commit
90771c6
·
verified ·
1 Parent(s): 7014f84

End of training

Browse files
README.md CHANGED
@@ -9,18 +9,18 @@ model-index:
9
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
10
  should probably proofread and complete it, then remove this comment. -->
11
 
12
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/priyanshipal/huggingface/runs/g7aicg5h)
13
  # eval_cache
14
 
15
  This model was trained from scratch on an unknown dataset.
16
  It achieves the following results on the evaluation set:
17
- - eval_loss: 4.3269
18
- - eval_model_preparation_time: 0.0045
19
- - eval_cer: 0.6025
20
- - eval_wer: 0.8793
21
- - eval_runtime: 36.0147
22
- - eval_samples_per_second: 15.882
23
- - eval_steps_per_second: 1.0
24
  - step: 0
25
 
26
  ## Model description
 
9
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
10
  should probably proofread and complete it, then remove this comment. -->
11
 
12
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/priyanshipal/huggingface/runs/bhazmy67)
13
  # eval_cache
14
 
15
  This model was trained from scratch on an unknown dataset.
16
  It achieves the following results on the evaluation set:
17
+ - eval_loss: 2.2164
18
+ - eval_model_preparation_time: 0.0046
19
+ - eval_cer: 0.4677
20
+ - eval_wer: 0.5669
21
+ - eval_runtime: 40.5673
22
+ - eval_samples_per_second: 14.1
23
+ - eval_steps_per_second: 0.887
24
  - step: 0
25
 
26
  ## Model description
all_results.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "eval_cer": 0.6025458851391355,
3
- "eval_loss": 4.326878547668457,
4
- "eval_model_preparation_time": 0.0045,
5
- "eval_runtime": 36.0147,
6
  "eval_samples": 572,
7
- "eval_samples_per_second": 15.882,
8
- "eval_steps_per_second": 1.0,
9
- "eval_wer": 0.8793330408073716
10
  }
 
1
  {
2
+ "eval_cer": 0.4676731793960924,
3
+ "eval_loss": 2.216402530670166,
4
+ "eval_model_preparation_time": 0.0046,
5
+ "eval_runtime": 40.5673,
6
  "eval_samples": 572,
7
+ "eval_samples_per_second": 14.1,
8
+ "eval_steps_per_second": 0.887,
9
+ "eval_wer": 0.5669153137340939
10
  }
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "/scratch/elec/puhe/p/palp3/MUCS/indicwav2vec_outputs/pd_warmup_2000/s400_shuff42",
3
  "activation_dropout": 0.0,
4
  "adapter_attn_dim": null,
5
  "adapter_kernel_size": 3,
 
1
  {
2
+ "_name_or_path": "/scratch/elec/puhe/p/palp3/MUCS/indicwav2vec_outputs/pd_warmup_2000/s300_shuff500",
3
  "activation_dropout": 0.0,
4
  "adapter_attn_dim": null,
5
  "adapter_kernel_size": 3,
eval_results.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "eval_cer": 0.6025458851391355,
3
- "eval_loss": 4.326878547668457,
4
- "eval_model_preparation_time": 0.0045,
5
- "eval_runtime": 36.0147,
6
  "eval_samples": 572,
7
- "eval_samples_per_second": 15.882,
8
- "eval_steps_per_second": 1.0,
9
- "eval_wer": 0.8793330408073716
10
  }
 
1
  {
2
+ "eval_cer": 0.4676731793960924,
3
+ "eval_loss": 2.216402530670166,
4
+ "eval_model_preparation_time": 0.0046,
5
+ "eval_runtime": 40.5673,
6
  "eval_samples": 572,
7
+ "eval_samples_per_second": 14.1,
8
+ "eval_steps_per_second": 0.887,
9
+ "eval_wer": 0.5669153137340939
10
  }
evalonlyhindi_indicwav2vec_MUCS_warmup2000_s300shuff500_2143808.out ADDED
@@ -0,0 +1,162 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0
  0%| | 0/36 [00:00<?, ?it/s]
1
  6%|▌ | 2/36 [00:01<00:28, 1.20it/s]
2
  8%|▊ | 3/36 [00:03<00:34, 1.04s/it]
3
  11%|█ | 4/36 [00:04<00:42, 1.32s/it]
4
  14%|█▍ | 5/36 [00:06<00:43, 1.40s/it]
5
  17%|█▋ | 6/36 [00:07<00:41, 1.39s/it]
6
  19%|█▉ | 7/36 [00:08<00:35, 1.23s/it]
7
  22%|██▏ | 8/36 [00:09<00:27, 1.02it/s]
8
  25%|██▌ | 9/36 [00:09<00:22, 1.19it/s]
9
  28%|██▊ | 10/36 [00:10<00:21, 1.24it/s]
10
  31%|███ | 11/36 [00:11<00:20, 1.21it/s]
11
  33%|███▎ | 12/36 [00:11<00:19, 1.23it/s]
12
  36%|███▌ | 13/36 [00:12<00:17, 1.29it/s]
13
  39%|███▉ | 14/36 [00:13<00:14, 1.48it/s]
14
  42%|████▏ | 15/36 [00:13<00:12, 1.67it/s]
15
  44%|████▍ | 16/36 [00:13<00:10, 1.82it/s]
16
  47%|████▋ | 17/36 [00:14<00:09, 1.92it/s]
17
  50%|█████ | 18/36 [00:14<00:09, 1.85it/s]
18
  53%|█████▎ | 19/36 [00:15<00:09, 1.77it/s]
19
  56%|█████▌ | 20/36 [00:16<00:08, 1.83it/s]
20
  58%|█████▊ | 21/36 [00:16<00:07, 1.97it/s]
21
  61%|██████ | 22/36 [00:17<00:07, 1.97it/s]
22
  64%|██████▍ | 23/36 [00:17<00:06, 1.93it/s]
23
  67%|██████▋ | 24/36 [00:18<00:06, 1.91it/s]
24
  69%|██████▉ | 25/36 [00:18<00:06, 1.82it/s]
25
  72%|███████▏ | 26/36 [00:19<00:05, 1.85it/s]
26
  75%|███████▌ | 27/36 [00:19<00:04, 1.95it/s]
27
  78%|███████▊ | 28/36 [00:20<00:04, 1.62it/s]
28
  81%|████████ | 29/36 [00:22<00:06, 1.08it/s]
29
  83%|████████▎ | 30/36 [00:23<00:06, 1.08s/it]
30
  86%|████████▌ | 31/36 [00:25<00:06, 1.34s/it]
31
  89%|████████▉ | 32/36 [00:26<00:04, 1.12s/it]
32
  92%|█████████▏| 33/36 [00:26<00:02, 1.04it/s]
33
  94%|█████████▍| 34/36 [00:27<00:01, 1.21it/s]
34
  97%|█████████▋| 35/36 [00:27<00:00, 1.36it/s]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb: Currently logged in as: priyanshi-pal (priyanshipal). Use `wandb login --relogin` to force relogin
2
+ wandb: wandb version 0.17.7 is available! To upgrade, please run:
3
+ wandb: $ pip install wandb --upgrade
4
+ wandb: Tracking run with wandb version 0.17.6
5
+ wandb: Run data is saved locally in /scratch/elec/t405-puhe/p/palp3/MUCS/wandb/run-20240822_161651-bhazmy67
6
+ wandb: Run `wandb offline` to turn off syncing.
7
+ wandb: Syncing run eval_pd2000_s300_shuff500_hindi
8
+ wandb: ⭐️ View project at https://wandb.ai/priyanshipal/huggingface
9
+ wandb: 🚀 View run at https://wandb.ai/priyanshipal/huggingface/runs/bhazmy67
10
+ /scratch/work/palp3/myenv/lib/python3.11/site-packages/transformers/training_args.py:1525: FutureWarning: `evaluation_strategy` is deprecated and will be removed in version 4.46 of 🤗 Transformers. Use `eval_strategy` instead
11
+ warnings.warn(
12
+
13
+ /scratch/work/palp3/myenv/lib/python3.11/site-packages/transformers/models/auto/configuration_auto.py:957: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
14
+ warnings.warn(
15
+ /scratch/work/palp3/myenv/lib/python3.11/site-packages/transformers/models/auto/feature_extraction_auto.py:329: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
16
+ warnings.warn(
17
+ /scratch/work/palp3/myenv/lib/python3.11/site-packages/accelerate/accelerator.py:488: FutureWarning: `torch.cuda.amp.GradScaler(args...)` is deprecated. Please use `torch.amp.GradScaler('cuda', args...)` instead.
18
+ self.scaler = torch.cuda.amp.GradScaler(**kwargs)
19
+ max_steps is given, it will override any value given in num_train_epochs
20
+ Wav2Vec2CTCTokenizer(name_or_path='', vocab_size=149, model_max_length=1000000000000000019884624838656, is_fast=False, padding_side='right', truncation_side='right', special_tokens={'bos_token': '<s>', 'eos_token': '</s>', 'unk_token': '[UNK]', 'pad_token': '[PAD]'}, clean_up_tokenization_spaces=True), added_tokens_decoder={
21
+ 147: AddedToken("[UNK]", rstrip=True, lstrip=True, single_word=False, normalized=False, special=False),
22
+ 148: AddedToken("[PAD]", rstrip=True, lstrip=True, single_word=False, normalized=False, special=False),
23
+ 149: AddedToken("<s>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
24
+ 150: AddedToken("</s>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
25
+ }
26
+ CHECK MODEL PARAMS Wav2Vec2ForCTC(
27
+ (wav2vec2): Wav2Vec2Model(
28
+ (feature_extractor): Wav2Vec2FeatureEncoder(
29
+ (conv_layers): ModuleList(
30
+ (0): Wav2Vec2LayerNormConvLayer(
31
+ (conv): Conv1d(1, 512, kernel_size=(10,), stride=(5,))
32
+ (layer_norm): LayerNorm((512,), eps=1e-05, elementwise_affine=True)
33
+ (activation): GELUActivation()
34
+ )
35
+ (1-4): 4 x Wav2Vec2LayerNormConvLayer(
36
+ (conv): Conv1d(512, 512, kernel_size=(3,), stride=(2,))
37
+ (layer_norm): LayerNorm((512,), eps=1e-05, elementwise_affine=True)
38
+ (activation): GELUActivation()
39
+ )
40
+ (5-6): 2 x Wav2Vec2LayerNormConvLayer(
41
+ (conv): Conv1d(512, 512, kernel_size=(2,), stride=(2,))
42
+ (layer_norm): LayerNorm((512,), eps=1e-05, elementwise_affine=True)
43
+ (activation): GELUActivation()
44
+ )
45
+ )
46
+ )
47
+ (feature_projection): Wav2Vec2FeatureProjection(
48
+ (layer_norm): LayerNorm((512,), eps=1e-05, elementwise_affine=True)
49
+ (projection): Linear(in_features=512, out_features=1024, bias=True)
50
+ (dropout): Dropout(p=0.3, inplace=False)
51
+ )
52
+ (encoder): Wav2Vec2EncoderStableLayerNorm(
53
+ (pos_conv_embed): Wav2Vec2PositionalConvEmbedding(
54
+ (conv): ParametrizedConv1d(
55
+ 1024, 1024, kernel_size=(128,), stride=(1,), padding=(64,), groups=16
56
+ (parametrizations): ModuleDict(
57
+ (weight): ParametrizationList(
58
+ (0): _WeightNorm()
59
+ )
60
+ )
61
+ )
62
+ (padding): Wav2Vec2SamePadLayer()
63
+ (activation): GELUActivation()
64
+ )
65
+ (layer_norm): LayerNorm((1024,), eps=1e-05, elementwise_affine=True)
66
+ (dropout): Dropout(p=0.2, inplace=False)
67
+ (layers): ModuleList(
68
+ (0-23): 24 x Wav2Vec2EncoderLayerStableLayerNorm(
69
+ (attention): Wav2Vec2SdpaAttention(
70
+ (k_proj): Linear(in_features=1024, out_features=1024, bias=True)
71
+ (v_proj): Linear(in_features=1024, out_features=1024, bias=True)
72
+ (q_proj): Linear(in_features=1024, out_features=1024, bias=True)
73
+ (out_proj): Linear(in_features=1024, out_features=1024, bias=True)
74
+ )
75
+ (dropout): Dropout(p=0.2, inplace=False)
76
+ (layer_norm): LayerNorm((1024,), eps=1e-05, elementwise_affine=True)
77
+ (feed_forward): Wav2Vec2FeedForward(
78
+ (intermediate_dropout): Dropout(p=0.0, inplace=False)
79
+ (intermediate_dense): Linear(in_features=1024, out_features=4096, bias=True)
80
+ (intermediate_act_fn): GELUActivation()
81
+ (output_dense): Linear(in_features=4096, out_features=1024, bias=True)
82
+ (output_dropout): Dropout(p=0.2, inplace=False)
83
+ )
84
+ (final_layer_norm): LayerNorm((1024,), eps=1e-05, elementwise_affine=True)
85
+ )
86
+ )
87
+ )
88
+ )
89
+ (dropout): Dropout(p=0.0, inplace=False)
90
+ (lm_head): Linear(in_features=1024, out_features=151, bias=True)
91
+ )
92
+ check the eval set length 572
93
+ 08/22/2024 16:17:04 - INFO - __main__ - *** Evaluate ***
94
+ /scratch/work/palp3/myenv/lib/python3.11/site-packages/transformers/models/wav2vec2/processing_wav2vec2.py:157: UserWarning: `as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your labels by using the argument `text` of the regular `__call__` method (either in the same call as your audio inputs, or in a separate call.
95
+ warnings.warn(
96
+
97
  0%| | 0/36 [00:00<?, ?it/s]
98
  6%|▌ | 2/36 [00:01<00:28, 1.20it/s]
99
  8%|▊ | 3/36 [00:03<00:34, 1.04s/it]
100
  11%|█ | 4/36 [00:04<00:42, 1.32s/it]
101
  14%|█▍ | 5/36 [00:06<00:43, 1.40s/it]
102
  17%|█▋ | 6/36 [00:07<00:41, 1.39s/it]
103
  19%|█▉ | 7/36 [00:08<00:35, 1.23s/it]
104
  22%|██▏ | 8/36 [00:09<00:27, 1.02it/s]
105
  25%|██▌ | 9/36 [00:09<00:22, 1.19it/s]
106
  28%|██▊ | 10/36 [00:10<00:21, 1.24it/s]
107
  31%|███ | 11/36 [00:11<00:20, 1.21it/s]
108
  33%|███▎ | 12/36 [00:11<00:19, 1.23it/s]
109
  36%|███▌ | 13/36 [00:12<00:17, 1.29it/s]
110
  39%|███▉ | 14/36 [00:13<00:14, 1.48it/s]
111
  42%|████▏ | 15/36 [00:13<00:12, 1.67it/s]
112
  44%|████▍ | 16/36 [00:13<00:10, 1.82it/s]
113
  47%|████▋ | 17/36 [00:14<00:09, 1.92it/s]
114
  50%|█████ | 18/36 [00:14<00:09, 1.85it/s]
115
  53%|█████▎ | 19/36 [00:15<00:09, 1.77it/s]
116
  56%|█████▌ | 20/36 [00:16<00:08, 1.83it/s]
117
  58%|█████▊ | 21/36 [00:16<00:07, 1.97it/s]
118
  61%|██████ | 22/36 [00:17<00:07, 1.97it/s]
119
  64%|██████▍ | 23/36 [00:17<00:06, 1.93it/s]
120
  67%|██████▋ | 24/36 [00:18<00:06, 1.91it/s]
121
  69%|██████▉ | 25/36 [00:18<00:06, 1.82it/s]
122
  72%|███████▏ | 26/36 [00:19<00:05, 1.85it/s]
123
  75%|███████▌ | 27/36 [00:19<00:04, 1.95it/s]
124
  78%|███████▊ | 28/36 [00:20<00:04, 1.62it/s]
125
  81%|████████ | 29/36 [00:22<00:06, 1.08it/s]
126
  83%|████████▎ | 30/36 [00:23<00:06, 1.08s/it]
127
  86%|████████▌ | 31/36 [00:25<00:06, 1.34s/it]
128
  89%|████████▉ | 32/36 [00:26<00:04, 1.12s/it]
129
  92%|█████████▏| 33/36 [00:26<00:02, 1.04it/s]
130
  94%|█████████▍| 34/36 [00:27<00:01, 1.21it/s]
131
  97%|█████████▋| 35/36 [00:27<00:00, 1.36it/s]
132
+ Printing predictions for a few samples:
133
+ Sample 1:
134
+ Reference: हम उनका उपयोग ऐसे ही कर सकते हैं या आवश्यकता अनुसार कुछ बदलाव करके उपयोग कर सकते हैं
135
+ ######
136
+
137
+
138
+ Prediction: mpl lauts मजद हहम उनका उपयोग ैसे ही कर सकते हं
139
+
140
+
141
+
142
+ Sample 2:
143
+ Reference: अतः शीर्षक इस तरह से जोड़ सकते हैं
144
+ ######
145
+
146
+
147
+ Prediction: अ ीर
148
+
149
+
150
+
151
+ Sample 3:
152
+ Reference: प्रेसेंटेशन के अंत में आपने स्लाइड की एक कॉपी बना ली है
153
+ ######
154
+
155
+
156
+ Prediction: prntation के अंत में पन
157
+
158
+
159
+
160
+ Sample 4:
161
+ Reference: चलिए अब फोंट्स और फोंट्स को फॉर्मेट करने के कुछ तरीके देखते हैं
162
+ ######
163
+
164
+
165
+ Prediction: क cop ब चलिए fonts और fonts को format करने के कुछ तरीके ंं
166
+
167
+
168
+
169
+ Sample 5:
170
+ Reference: यह एक डायलॉग बॉक्स खोलेगा जिसमें हम अपनी आवश्यकतानुसार फॉन्ट स्टाइल और साइज़ सेट कर सकते हैं
171
+ ######
172
+
173
+
174
+ Prediction: द कंयह एक dialog boxस खोलेगा जिसमें हम अपनी व्यक
175
+
176
+
177
+
178
+ last Reference string यह स्क्रिप्ट लता द्वारा अनुवादित है आईआईटी मुंबई की ओर से मैं रवि कुमार अब आपसे विदा लेता हूँहमसे जुड़ने के लिए धन्यवाद
179
+
180
+
181
+ last prediction string lता द्वारा अनुवादित है आईआईटी मुमंबई की ओर से मैं रवि कुमार अब आपसे विदा लेता हूँ हमसे जड़ने के लिए धन्यवाद
182
+ ***** eval metrics *****
183
+ eval_cer = 0.4677
184
+ eval_loss = 2.2164
185
+ eval_model_preparation_time = 0.0046
186
+ eval_runtime = 0:00:40.56
187
+ eval_samples = 572
188
+ eval_samples_per_second = 14.1
189
+ eval_steps_per_second = 0.887
190
+ eval_wer = 0.5669
191
+
192
+
193
+
194
+
195
+
196
+
197
+
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a1652bc6136498220e8e10b3cacbaabad0bb1a38be329d9722f49a67f271753b
3
  size 1262426580
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7eb487bdd3e20589cbadfa9613598ab0ecfa0a02e0d5db447642ffc85a0bb960
3
  size 1262426580
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fc40ea5915cba28a074a9b6adf1b3f7b7943525a87ed03413e8be115e383ad12
3
  size 5432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:112f6e6b2734038b6cbd8a832070a364a365778c258759d59c07634dc399dde4
3
  size 5432