Priyanship commited on
Commit
7014f84
·
verified ·
1 Parent(s): 8bc3971

End of training

Browse files
README.md ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - generated_from_trainer
4
+ model-index:
5
+ - name: eval_cache
6
+ results: []
7
+ ---
8
+
9
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
10
+ should probably proofread and complete it, then remove this comment. -->
11
+
12
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/priyanshipal/huggingface/runs/g7aicg5h)
13
+ # eval_cache
14
+
15
+ This model was trained from scratch on an unknown dataset.
16
+ It achieves the following results on the evaluation set:
17
+ - eval_loss: 4.3269
18
+ - eval_model_preparation_time: 0.0045
19
+ - eval_cer: 0.6025
20
+ - eval_wer: 0.8793
21
+ - eval_runtime: 36.0147
22
+ - eval_samples_per_second: 15.882
23
+ - eval_steps_per_second: 1.0
24
+ - step: 0
25
+
26
+ ## Model description
27
+
28
+ More information needed
29
+
30
+ ## Intended uses & limitations
31
+
32
+ More information needed
33
+
34
+ ## Training and evaluation data
35
+
36
+ More information needed
37
+
38
+ ## Training procedure
39
+
40
+ ### Training hyperparameters
41
+
42
+ The following hyperparameters were used during training:
43
+ - learning_rate: 0.0006
44
+ - train_batch_size: 16
45
+ - eval_batch_size: 16
46
+ - seed: 300
47
+ - gradient_accumulation_steps: 2
48
+ - total_train_batch_size: 32
49
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
50
+ - lr_scheduler_type: linear
51
+ - lr_scheduler_warmup_steps: 500
52
+ - training_steps: 1000
53
+ - mixed_precision_training: Native AMP
54
+
55
+ ### Framework versions
56
+
57
+ - Transformers 4.43.1
58
+ - Pytorch 2.4.0
59
+ - Datasets 2.20.0
60
+ - Tokenizers 0.19.1
all_results.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_cer": 0.6025458851391355,
3
+ "eval_loss": 4.326878547668457,
4
+ "eval_model_preparation_time": 0.0045,
5
+ "eval_runtime": 36.0147,
6
+ "eval_samples": 572,
7
+ "eval_samples_per_second": 15.882,
8
+ "eval_steps_per_second": 1.0,
9
+ "eval_wer": 0.8793330408073716
10
+ }
config.json ADDED
@@ -0,0 +1,109 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/scratch/elec/puhe/p/palp3/MUCS/indicwav2vec_outputs/pd_warmup_2000/s400_shuff42",
3
+ "activation_dropout": 0.0,
4
+ "adapter_attn_dim": null,
5
+ "adapter_kernel_size": 3,
6
+ "adapter_stride": 2,
7
+ "add_adapter": false,
8
+ "apply_spec_augment": true,
9
+ "architectures": [
10
+ "Wav2Vec2ForCTC"
11
+ ],
12
+ "attention_dropout": 0.3,
13
+ "bos_token_id": 1,
14
+ "classifier_proj_size": 256,
15
+ "codevector_dim": 256,
16
+ "contrastive_logits_temperature": 0.1,
17
+ "conv_bias": true,
18
+ "conv_dim": [
19
+ 512,
20
+ 512,
21
+ 512,
22
+ 512,
23
+ 512,
24
+ 512,
25
+ 512
26
+ ],
27
+ "conv_kernel": [
28
+ 10,
29
+ 3,
30
+ 3,
31
+ 3,
32
+ 3,
33
+ 2,
34
+ 2
35
+ ],
36
+ "conv_stride": [
37
+ 5,
38
+ 2,
39
+ 2,
40
+ 2,
41
+ 2,
42
+ 2,
43
+ 2
44
+ ],
45
+ "ctc_loss_reduction": "mean",
46
+ "ctc_zero_infinity": false,
47
+ "diversity_loss_weight": 0.1,
48
+ "do_stable_layer_norm": true,
49
+ "eos_token_id": 2,
50
+ "feat_extract_activation": "gelu",
51
+ "feat_extract_dropout": 0.0,
52
+ "feat_extract_norm": "layer",
53
+ "feat_proj_dropout": 0.3,
54
+ "feat_quantizer_dropout": 0.0,
55
+ "final_dropout": 0.0,
56
+ "hidden_act": "gelu",
57
+ "hidden_dropout": 0.2,
58
+ "hidden_dropout_prob": 0.1,
59
+ "hidden_size": 1024,
60
+ "initializer_range": 0.02,
61
+ "intermediate_size": 4096,
62
+ "layer_norm_eps": 1e-05,
63
+ "layerdrop": 0.0,
64
+ "mask_feature_length": 10,
65
+ "mask_feature_min_masks": 0,
66
+ "mask_feature_prob": 0.0,
67
+ "mask_time_length": 10,
68
+ "mask_time_min_masks": 2,
69
+ "mask_time_prob": 0.05,
70
+ "model_type": "wav2vec2",
71
+ "num_adapter_layers": 3,
72
+ "num_attention_heads": 16,
73
+ "num_codevector_groups": 2,
74
+ "num_codevectors_per_group": 320,
75
+ "num_conv_pos_embedding_groups": 16,
76
+ "num_conv_pos_embeddings": 128,
77
+ "num_feat_extract_layers": 7,
78
+ "num_hidden_layers": 24,
79
+ "num_negatives": 100,
80
+ "output_hidden_size": 1024,
81
+ "pad_token_id": 148,
82
+ "proj_codevector_dim": 256,
83
+ "tdnn_dilation": [
84
+ 1,
85
+ 2,
86
+ 3,
87
+ 1,
88
+ 1
89
+ ],
90
+ "tdnn_dim": [
91
+ 512,
92
+ 512,
93
+ 512,
94
+ 512,
95
+ 1500
96
+ ],
97
+ "tdnn_kernel": [
98
+ 5,
99
+ 3,
100
+ 3,
101
+ 1,
102
+ 1
103
+ ],
104
+ "torch_dtype": "float32",
105
+ "transformers_version": "4.43.1",
106
+ "use_weighted_layer_sum": false,
107
+ "vocab_size": 151,
108
+ "xvector_output_dim": 512
109
+ }
eval_results.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_cer": 0.6025458851391355,
3
+ "eval_loss": 4.326878547668457,
4
+ "eval_model_preparation_time": 0.0045,
5
+ "eval_runtime": 36.0147,
6
+ "eval_samples": 572,
7
+ "eval_samples_per_second": 15.882,
8
+ "eval_steps_per_second": 1.0,
9
+ "eval_wer": 0.8793330408073716
10
+ }
evalonlyhindi_indicwav2vec_MUCS_warmup2000_s400shuff42_2143185.out ADDED
@@ -0,0 +1,304 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0
  0%| | 0/36 [00:00<?, ?it/s]
1
  6%|▌ | 2/36 [00:01<00:24, 1.38it/s]
2
  8%|▊ | 3/36 [00:02<00:33, 1.02s/it]
3
  11%|█ | 4/36 [00:04<00:42, 1.34s/it]
4
  14%|█▍ | 5/36 [00:06<00:44, 1.43s/it]
5
  17%|█▋ | 6/36 [00:07<00:42, 1.43s/it]
6
  19%|█▉ | 7/36 [00:08<00:36, 1.26s/it]
7
  22%|██▏ | 8/36 [00:09<00:28, 1.00s/it]
8
  25%|██▌ | 9/36 [00:09<00:23, 1.14it/s]
9
  28%|██▊ | 10/36 [00:10<00:22, 1.17it/s]
10
  31%|███ | 11/36 [00:11<00:21, 1.14it/s]
11
  33%|███▎ | 12/36 [00:12<00:20, 1.15it/s]
12
  36%|███▌ | 13/36 [00:12<00:18, 1.26it/s]
13
  39%|███▉ | 14/36 [00:13<00:15, 1.44it/s]
14
  42%|████▏ | 15/36 [00:13<00:12, 1.62it/s]
15
  44%|████▍ | 16/36 [00:14<00:11, 1.76it/s]
16
  47%|████▋ | 17/36 [00:14<00:10, 1.86it/s]
17
  50%|█████ | 18/36 [00:15<00:10, 1.79it/s]
18
  53%|█████▎ | 19/36 [00:16<00:09, 1.72it/s]
19
  56%|█████▌ | 20/36 [00:16<00:09, 1.77it/s]
20
  58%|█████▊ | 21/36 [00:16<00:07, 1.91it/s]
21
  61%|██████ | 22/36 [00:17<00:07, 1.91it/s]
22
  64%|██████▍ | 23/36 [00:18<00:06, 1.88it/s]
23
  67%|██████▋ | 24/36 [00:18<00:06, 1.85it/s]
24
  69%|██████▉ | 25/36 [00:19<00:06, 1.77it/s]
25
  72%|███████▏ | 26/36 [00:19<00:05, 1.80it/s]
26
  75%|███████▌ | 27/36 [00:20<00:04, 1.89it/s]
27
  78%|███████▊ | 28/36 [00:21<00:05, 1.55it/s]
28
  81%|████████ | 29/36 [00:22<00:06, 1.03it/s]
29
  83%|████████▎ | 30/36 [00:24<00:06, 1.13s/it]
30
  86%|████████▌ | 31/36 [00:26<00:06, 1.39s/it]
31
  89%|████████▉ | 32/36 [00:26<00:04, 1.16s/it]
32
  92%|█████████▏| 33/36 [00:27<00:02, 1.00it/s]
33
  94%|█████████▍| 34/36 [00:28<00:01, 1.17it/s]
34
  97%|█████████▋| 35/36 [00:28<00:00, 1.31it/s]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb: Currently logged in as: priyanshi-pal (priyanshipal). Use `wandb login --relogin` to force relogin
2
+ wandb: wandb version 0.17.7 is available! To upgrade, please run:
3
+ wandb: $ pip install wandb --upgrade
4
+ wandb: Tracking run with wandb version 0.17.6
5
+ wandb: Run data is saved locally in /scratch/elec/t405-puhe/p/palp3/MUCS/wandb/run-20240822_153818-g7aicg5h
6
+ wandb: Run `wandb offline` to turn off syncing.
7
+ wandb: Syncing run eval_pd2000_s300_shuff100_hindi
8
+ wandb: ⭐️ View project at https://wandb.ai/priyanshipal/huggingface
9
+ wandb: 🚀 View run at https://wandb.ai/priyanshipal/huggingface/runs/g7aicg5h
10
+ /scratch/work/palp3/myenv/lib/python3.11/site-packages/transformers/training_args.py:1525: FutureWarning: `evaluation_strategy` is deprecated and will be removed in version 4.46 of 🤗 Transformers. Use `eval_strategy` instead
11
+ warnings.warn(
12
+
13
+ /scratch/work/palp3/myenv/lib/python3.11/site-packages/transformers/models/auto/configuration_auto.py:957: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
14
+ warnings.warn(
15
+ /scratch/work/palp3/myenv/lib/python3.11/site-packages/transformers/models/auto/feature_extraction_auto.py:329: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
16
+ warnings.warn(
17
+ Wav2Vec2CTCTokenizer(name_or_path='', vocab_size=149, model_max_length=1000000000000000019884624838656, is_fast=False, padding_side='right', truncation_side='right', special_tokens={'bos_token': '<s>', 'eos_token': '</s>', 'unk_token': '[UNK]', 'pad_token': '[PAD]'}, clean_up_tokenization_spaces=True), added_tokens_decoder={
18
+ 147: AddedToken("[UNK]", rstrip=True, lstrip=True, single_word=False, normalized=False, special=False),
19
+ 148: AddedToken("[PAD]", rstrip=True, lstrip=True, single_word=False, normalized=False, special=False),
20
+ 149: AddedToken("<s>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
21
+ 150: AddedToken("</s>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
22
+ }
23
+ CHECK MODEL PARAMS Wav2Vec2ForCTC(
24
+ (wav2vec2): Wav2Vec2Model(
25
+ (feature_extractor): Wav2Vec2FeatureEncoder(
26
+ (conv_layers): ModuleList(
27
+ (0): Wav2Vec2LayerNormConvLayer(
28
+ (conv): Conv1d(1, 512, kernel_size=(10,), stride=(5,))
29
+ (layer_norm): LayerNorm((512,), eps=1e-05, elementwise_affine=True)
30
+ (activation): GELUActivation()
31
+ )
32
+ (1-4): 4 x Wav2Vec2LayerNormConvLayer(
33
+ (conv): Conv1d(512, 512, kernel_size=(3,), stride=(2,))
34
+ (layer_norm): LayerNorm((512,), eps=1e-05, elementwise_affine=True)
35
+ (activation): GELUActivation()
36
+ )
37
+ (5-6): 2 x Wav2Vec2LayerNormConvLayer(
38
+ (conv): Conv1d(512, 512, kernel_size=(2,), stride=(2,))
39
+ (layer_norm): LayerNorm((512,), eps=1e-05, elementwise_affine=True)
40
+ (activation): GELUActivation()
41
+ )
42
+ )
43
+ )
44
+ (feature_projection): Wav2Vec2FeatureProjection(
45
+ (layer_norm): LayerNorm((512,), eps=1e-05, elementwise_affine=True)
46
+ (projection): Linear(in_features=512, out_features=1024, bias=True)
47
+ (dropout): Dropout(p=0.3, inplace=False)
48
+ )
49
+ (encoder): Wav2Vec2EncoderStableLayerNorm(
50
+ (pos_conv_embed): Wav2Vec2PositionalConvEmbedding(
51
+ (conv): ParametrizedConv1d(
52
+ 1024, 1024, kernel_size=(128,), stride=(1,), padding=(64,), groups=16
53
+ (parametrizations): ModuleDict(
54
+ (weight): ParametrizationList(
55
+ (0): _WeightNorm()
56
+ )
57
+ )
58
+ )
59
+ (padding): Wav2Vec2SamePadLayer()
60
+ (activation): GELUActivation()
61
+ )
62
+ (layer_norm): LayerNorm((1024,), eps=1e-05, elementwise_affine=True)
63
+ (dropout): Dropout(p=0.2, inplace=False)
64
+ (layers): ModuleList(
65
+ (0-23): 24 x Wav2Vec2EncoderLayerStableLayerNorm(
66
+ (attention): Wav2Vec2SdpaAttention(
67
+ (k_proj): Linear(in_features=1024, out_features=1024, bias=True)
68
+ (v_proj): Linear(in_features=1024, out_features=1024, bias=True)
69
+ (q_proj): Linear(in_features=1024, out_features=1024, bias=True)
70
+ (out_proj): Linear(in_features=1024, out_features=1024, bias=True)
71
+ )
72
+ (dropout): Dropout(p=0.2, inplace=False)
73
+ (layer_norm): LayerNorm((1024,), eps=1e-05, elementwise_affine=True)
74
+ (feed_forward): Wav2Vec2FeedForward(
75
+ (intermediate_dropout): Dropout(p=0.0, inplace=False)
76
+ (intermediate_dense): Linear(in_features=1024, out_features=4096, bias=True)
77
+ (intermediate_act_fn): GELUActivation()
78
+ (output_dense): Linear(in_features=4096, out_features=1024, bias=True)
79
+ (output_dropout): Dropout(p=0.2, inplace=False)
80
+ )
81
+ (final_layer_norm): LayerNorm((1024,), eps=1e-05, elementwise_affine=True)
82
+ )
83
+ )
84
+ )
85
+ )
86
+ (dropout): Dropout(p=0.0, inplace=False)
87
+ (lm_head): Linear(in_features=1024, out_features=151, bias=True)
88
+ )
89
+
90
+ /scratch/work/palp3/myenv/lib/python3.11/site-packages/accelerate/accelerator.py:488: FutureWarning: `torch.cuda.amp.GradScaler(args...)` is deprecated. Please use `torch.amp.GradScaler('cuda', args...)` instead.
91
+ self.scaler = torch.cuda.amp.GradScaler(**kwargs)
92
+ max_steps is given, it will override any value given in num_train_epochs
93
+ check the eval set length 572
94
+ 08/22/2024 15:38:41 - INFO - __main__ - *** Evaluate ***
95
+ /scratch/work/palp3/myenv/lib/python3.11/site-packages/transformers/models/wav2vec2/processing_wav2vec2.py:157: UserWarning: `as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your labels by using the argument `text` of the regular `__call__` method (either in the same call as your audio inputs, or in a separate call.
96
+ warnings.warn(
97
+
98
  0%| | 0/36 [00:00<?, ?it/s]
99
  6%|▌ | 2/36 [00:01<00:24, 1.38it/s]
100
  8%|▊ | 3/36 [00:02<00:33, 1.02s/it]
101
  11%|█ | 4/36 [00:04<00:42, 1.34s/it]
102
  14%|█▍ | 5/36 [00:06<00:44, 1.43s/it]
103
  17%|█▋ | 6/36 [00:07<00:42, 1.43s/it]
104
  19%|█▉ | 7/36 [00:08<00:36, 1.26s/it]
105
  22%|██▏ | 8/36 [00:09<00:28, 1.00s/it]
106
  25%|██▌ | 9/36 [00:09<00:23, 1.14it/s]
107
  28%|██▊ | 10/36 [00:10<00:22, 1.17it/s]
108
  31%|███ | 11/36 [00:11<00:21, 1.14it/s]
109
  33%|███▎ | 12/36 [00:12<00:20, 1.15it/s]
110
  36%|███▌ | 13/36 [00:12<00:18, 1.26it/s]
111
  39%|███▉ | 14/36 [00:13<00:15, 1.44it/s]
112
  42%|████▏ | 15/36 [00:13<00:12, 1.62it/s]
113
  44%|████▍ | 16/36 [00:14<00:11, 1.76it/s]
114
  47%|████▋ | 17/36 [00:14<00:10, 1.86it/s]
115
  50%|█████ | 18/36 [00:15<00:10, 1.79it/s]
116
  53%|█████▎ | 19/36 [00:16<00:09, 1.72it/s]
117
  56%|█████▌ | 20/36 [00:16<00:09, 1.77it/s]
118
  58%|█████▊ | 21/36 [00:16<00:07, 1.91it/s]
119
  61%|██████ | 22/36 [00:17<00:07, 1.91it/s]
120
  64%|██████▍ | 23/36 [00:18<00:06, 1.88it/s]
121
  67%|██████▋ | 24/36 [00:18<00:06, 1.85it/s]
122
  69%|██████▉ | 25/36 [00:19<00:06, 1.77it/s]
123
  72%|███████▏ | 26/36 [00:19<00:05, 1.80it/s]
124
  75%|███████▌ | 27/36 [00:20<00:04, 1.89it/s]
125
  78%|███████▊ | 28/36 [00:21<00:05, 1.55it/s]
126
  81%|████████ | 29/36 [00:22<00:06, 1.03it/s]
127
  83%|████████▎ | 30/36 [00:24<00:06, 1.13s/it]
128
  86%|████████▌ | 31/36 [00:26<00:06, 1.39s/it]
129
  89%|████████▉ | 32/36 [00:26<00:04, 1.16s/it]
130
  92%|█████████▏| 33/36 [00:27<00:02, 1.00it/s]
131
  94%|█████████▍| 34/36 [00:28<00:01, 1.17it/s]
132
  97%|█████████▋| 35/36 [00:28<00:00, 1.31it/s]
133
+ Printing predictions for a few samples:
134
+ Sample 1:
135
+ Reference: हम उनका उपयोग ऐसे ही कर सकते हैं या आवश्यकता अनुसार कुछ बदलाव करके उपयोग कर सकते हैं
136
+ ######
137
+
138
+
139
+ Prediction: उुampallwउdes मोजूदहमउनकाउपयोग से कयहव्यतनुहै
140
+
141
+
142
+
143
+ Sample 2:
144
+ Reference: अतः शीर्षक इस तरह से जोड़ सकते हैं
145
+ ######
146
+
147
+
148
+ Prediction: liअतीर्षैंैंैं
149
+
150
+
151
+
152
+ Sample 3:
153
+ Reference: प्रेसेंटेशन के अंत में आपने स्लाइड की एक कॉपी बना ली है
154
+ ######
155
+
156
+
157
+ Prediction: पपregentatinकेअनत मेंआपन ैं
158
+
159
+
160
+
161
+ Sample 4:
162
+ Reference: चलिए अब फोंट्स और फोंट्स को फॉर्मेट करने के कुछ तरीके देखते हैं
163
+ ######
164
+
165
+
166
+ Prediction: कएक coपpy बलfonऔरfonts को foरr्mat करनेकेकुष तरीकेदेैंैंैं
167
+
168
+
169
+
170
+ Sample 5:
171
+ Reference: यह एक डायलॉग बॉक्स खोलेगा जिसमें हम अपनी आवश्यकतानुसार फॉन्ट स्टाइल और साइज़ सेट कर सकते हैं
172
+ ######
173
+
174
+
175
+ Prediction: पकोबदलयह एक dlog बox्s खोलेगजिसमें हम अपनी आवष्यकतहैंै
176
+
177
+
178
+
179
+ last Reference string यह स्क्रिप्ट लता द्वारा अनुवादित है आईआईटी मुंबई की ओर से मैं रवि कुमार अब आपसे विदा लेता हूँहमसे जुड़ने के लिए धन्यवाद
180
+
181
+
182
+ last prediction string लतद्वअनुवदआtमुmबकी औेमैं रविकुअब आसे वदलतमसेजुड़नेकेलिएदधन्यवाद
183
+ ***** eval metrics *****
184
+ eval_cer = 0.6025
185
+ eval_loss = 4.3269
186
+ eval_model_preparation_time = 0.0045
187
+ eval_runtime = 0:00:36.01
188
+ eval_samples = 572
189
+ eval_samples_per_second = 15.882
190
+ eval_steps_per_second = 1.0
191
+ eval_wer = 0.8793
192
+
193
+
194
+
195
+
196
+
197
+
198
+
199
+
200
+
201
+
202
+
203
+
204
+
205
+
206
+
207
+
208
+
209
+
210
+
211
+
212
+
213
+
214
+
215
+
216
+
217
+
218
+
219
+
220
+
221
+
222
+
223
+
224
+
225
+
226
+
227
+
228
+
229
+
230
+
231
+
232
+
233
+
234
+
235
+
236
+
237
+
238
+
239
+
240
+
241
+
242
+
243
+
244
+
245
+
246
+
247
+
248
+
249
+
250
+
251
+
252
+
253
+
254
+
255
+
256
+
257
+
258
+
259
+
260
+
261
+
262
+
263
+
264
+
265
+
266
+
267
+
268
+
269
+
270
+
271
+
272
+
273
+
274
+
275
+
276
+
277
+
278
+
279
+
280
+
281
+
282
+
283
+
284
+
285
+
286
+
287
+
288
+
289
+
290
+
291
+
292
+
293
+
294
+
295
+
296
+
297
+
298
+
299
+
300
+
301
+
302
+
303
+
304
+
305
+
306
+
307
+
308
+
309
+
310
+
311
+
312
+
313
+
314
+
315
+
316
+
317
+
318
+
319
+
320
+
321
+
322
+
323
+
324
+
325
+
326
+
327
+
328
+
329
+
330
+
331
+
332
+
333
+
334
+
335
+
336
+
337
+
338
+
339
+
json/default-b60d5edd0f197c71/0.0.0/7483f22a71512872c377524b97484f6d20c275799bb9e7cd8fb3198178d8220a.incomplete_info.lock ADDED
File without changes
json/default-b60d5edd0f197c71/0.0.0/7483f22a71512872c377524b97484f6d20c275799bb9e7cd8fb3198178d8220a/dataset_info.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"description": "", "citation": "", "homepage": "", "license": "", "features": {"audio_id": {"dtype": "string", "_type": "Value"}, "audio_paths": {"dtype": "string", "_type": "Value"}, "transcriptions": {"dtype": "string", "_type": "Value"}}, "builder_name": "json", "dataset_name": "json", "config_name": "default", "version": {"version_str": "0.0.0", "major": 0, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 159776, "num_examples": 572, "dataset_name": "json"}}, "download_checksums": {"/m/triton/scratch/elec/puhe/p/palp3/MUCS/mucs_language_segregated_data/MUCS_test_languagesep_data.json": {"num_bytes": 1275544, "checksum": null}}, "download_size": 1275544, "dataset_size": 159776, "size_in_bytes": 1435320}
json/default-b60d5edd0f197c71/0.0.0/7483f22a71512872c377524b97484f6d20c275799bb9e7cd8fb3198178d8220a/json-train.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f67abacdef72568fdccdc70e54cc80c5aca72a7c217cf5c533dcdd9be5ca20b5
3
+ size 160568
json/default-b60d5edd0f197c71/0.0.0/7483f22a71512872c377524b97484f6d20c275799bb9e7cd8fb3198178d8220a_builder.lock ADDED
File without changes
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1652bc6136498220e8e10b3cacbaabad0bb1a38be329d9722f49a67f271753b
3
+ size 1262426580
preprocessor_config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_normalize": true,
3
+ "feature_extractor_type": "Wav2Vec2FeatureExtractor",
4
+ "feature_size": 1,
5
+ "padding_side": "right",
6
+ "padding_value": 0,
7
+ "processor_class": "Wav2Vec2Processor",
8
+ "return_attention_mask": true,
9
+ "sampling_rate": 16000
10
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc40ea5915cba28a074a9b6adf1b3f7b7943525a87ed03413e8be115e383ad12
3
+ size 5432