Maverick1713 commited on
Commit
5052d49
·
verified ·
1 Parent(s): de9b411

End of training

Browse files
README.md CHANGED
@@ -1,9 +1,11 @@
1
  ---
2
  library_name: transformers
3
  license: apache-2.0
4
- base_model: Maverick1713/wav2vec2-Pratyush-poly
5
  tags:
6
  - generated_from_trainer
 
 
7
  model-index:
8
  - name: wav2vec2-Pratyush-poly
9
  results: []
@@ -14,7 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  # wav2vec2-Pratyush-poly
16
 
17
- This model is a fine-tuned version of [Maverick1713/wav2vec2-Pratyush-poly](https://huggingface.co/Maverick1713/wav2vec2-Pratyush-poly) on an unknown dataset.
18
 
19
  ## Model description
20
 
@@ -34,17 +36,21 @@ More information needed
34
 
35
  The following hyperparameters were used during training:
36
  - learning_rate: 0.0002
37
- - train_batch_size: 2
38
  - eval_batch_size: 8
39
  - seed: 42
40
  - gradient_accumulation_steps: 4
41
- - total_train_batch_size: 8
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: linear
44
  - lr_scheduler_warmup_steps: 15
45
- - num_epochs: 1
46
  - mixed_precision_training: Native AMP
47
 
 
 
 
 
48
  ### Framework versions
49
 
50
  - Transformers 4.44.2
 
1
  ---
2
  library_name: transformers
3
  license: apache-2.0
4
+ base_model: facebook/wav2vec2-base
5
  tags:
6
  - generated_from_trainer
7
+ datasets:
8
+ - minds14
9
  model-index:
10
  - name: wav2vec2-Pratyush-poly
11
  results: []
 
16
 
17
  # wav2vec2-Pratyush-poly
18
 
19
+ This model is a fine-tuned version of [facebook/wav2vec2-base](https://huggingface.co/facebook/wav2vec2-base) on the minds14 dataset.
20
 
21
  ## Model description
22
 
 
36
 
37
  The following hyperparameters were used during training:
38
  - learning_rate: 0.0002
39
+ - train_batch_size: 6
40
  - eval_batch_size: 8
41
  - seed: 42
42
  - gradient_accumulation_steps: 4
43
+ - total_train_batch_size: 24
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
46
  - lr_scheduler_warmup_steps: 15
47
+ - num_epochs: 5
48
  - mixed_precision_training: Native AMP
49
 
50
+ ### Training results
51
+
52
+
53
+
54
  ### Framework versions
55
 
56
  - Transformers 4.44.2
added_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "</s>": 30,
3
+ "<s>": 29
4
+ }
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "Maverick1713/wav2vec2-Pratyush-poly",
3
  "activation_dropout": 0.0,
4
  "adapter_attn_dim": null,
5
  "adapter_kernel_size": 3,
@@ -87,7 +87,7 @@
87
  "num_hidden_layers": 12,
88
  "num_negatives": 100,
89
  "output_hidden_size": 768,
90
- "pad_token_id": 0,
91
  "proj_codevector_dim": 256,
92
  "tdnn_dilation": [
93
  1,
 
1
  {
2
+ "_name_or_path": "facebook/wav2vec2-base",
3
  "activation_dropout": 0.0,
4
  "adapter_attn_dim": null,
5
  "adapter_kernel_size": 3,
 
87
  "num_hidden_layers": 12,
88
  "num_negatives": 100,
89
  "output_hidden_size": 768,
90
+ "pad_token_id": 28,
91
  "proj_codevector_dim": 256,
92
  "tdnn_dilation": [
93
  1,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:98be48ae61471ab9ff302d0d84dd8a7140aea2cb44ab43aeee5969e778e6f3f5
3
  size 377611120
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1e13fb95fdb279024c5dacd770c9cab7678f141fcfbf599ff297b709abbde26
3
  size 377611120
runs/Oct20_04-22-30_15f36d18ffd2/events.out.tfevents.1729398168.15f36d18ffd2.372.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf9fbbd315e9f1dd7b5cc60ea3f41bcd655bfc52fe6926fd04c379054ff1d587
3
+ size 6636
runs/Oct20_04-27-17_15f36d18ffd2/events.out.tfevents.1729398452.15f36d18ffd2.372.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:838773f08a7c04aada026b6177ae9515465736e081297a6db49e8ddcfc9c1666
3
+ size 6984
special_tokens_map.json CHANGED
@@ -1,30 +1,6 @@
1
  {
2
- "bos_token": {
3
- "content": "<s>",
4
- "lstrip": true,
5
- "normalized": false,
6
- "rstrip": true,
7
- "single_word": false
8
- },
9
- "eos_token": {
10
- "content": "</s>",
11
- "lstrip": true,
12
- "normalized": false,
13
- "rstrip": true,
14
- "single_word": false
15
- },
16
- "pad_token": {
17
- "content": "<pad>",
18
- "lstrip": true,
19
- "normalized": false,
20
- "rstrip": true,
21
- "single_word": false
22
- },
23
- "unk_token": {
24
- "content": "<unk>",
25
- "lstrip": true,
26
- "normalized": false,
27
- "rstrip": true,
28
- "single_word": false
29
- }
30
  }
 
1
  {
2
+ "bos_token": "<s>",
3
+ "eos_token": "</s>",
4
+ "pad_token": "[PAD]",
5
+ "unk_token": "[UNK]"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6
  }
tokenizer_config.json CHANGED
@@ -1,49 +1,47 @@
1
  {
2
  "added_tokens_decoder": {
3
- "0": {
4
- "content": "<pad>",
5
  "lstrip": true,
6
  "normalized": false,
7
  "rstrip": true,
8
  "single_word": false,
9
  "special": false
10
  },
11
- "1": {
12
- "content": "<s>",
13
  "lstrip": true,
14
  "normalized": false,
15
  "rstrip": true,
16
  "single_word": false,
17
  "special": false
18
  },
19
- "2": {
20
- "content": "</s>",
21
- "lstrip": true,
22
  "normalized": false,
23
- "rstrip": true,
24
  "single_word": false,
25
- "special": false
26
  },
27
- "3": {
28
- "content": "<unk>",
29
- "lstrip": true,
30
  "normalized": false,
31
- "rstrip": true,
32
  "single_word": false,
33
- "special": false
34
  }
35
  },
36
  "bos_token": "<s>",
37
  "clean_up_tokenization_spaces": true,
38
  "do_lower_case": false,
39
- "do_normalize": true,
40
  "eos_token": "</s>",
41
  "model_max_length": 1000000000000000019884624838656,
42
- "pad_token": "<pad>",
43
  "replace_word_delimiter_char": " ",
44
- "return_attention_mask": false,
45
  "target_lang": null,
46
  "tokenizer_class": "Wav2Vec2CTCTokenizer",
47
- "unk_token": "<unk>",
48
  "word_delimiter_token": "|"
49
  }
 
1
  {
2
  "added_tokens_decoder": {
3
+ "27": {
4
+ "content": "[UNK]",
5
  "lstrip": true,
6
  "normalized": false,
7
  "rstrip": true,
8
  "single_word": false,
9
  "special": false
10
  },
11
+ "28": {
12
+ "content": "[PAD]",
13
  "lstrip": true,
14
  "normalized": false,
15
  "rstrip": true,
16
  "single_word": false,
17
  "special": false
18
  },
19
+ "29": {
20
+ "content": "<s>",
21
+ "lstrip": false,
22
  "normalized": false,
23
+ "rstrip": false,
24
  "single_word": false,
25
+ "special": true
26
  },
27
+ "30": {
28
+ "content": "</s>",
29
+ "lstrip": false,
30
  "normalized": false,
31
+ "rstrip": false,
32
  "single_word": false,
33
+ "special": true
34
  }
35
  },
36
  "bos_token": "<s>",
37
  "clean_up_tokenization_spaces": true,
38
  "do_lower_case": false,
 
39
  "eos_token": "</s>",
40
  "model_max_length": 1000000000000000019884624838656,
41
+ "pad_token": "[PAD]",
42
  "replace_word_delimiter_char": " ",
 
43
  "target_lang": null,
44
  "tokenizer_class": "Wav2Vec2CTCTokenizer",
45
+ "unk_token": "[UNK]",
46
  "word_delimiter_token": "|"
47
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d352379c537aa5e83a90ec8a30067b1ab5546db1b2e9ecc1c67805521009c4b2
3
  size 5176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:63f42e23f64dc36d207b70089444decf77046474a2ce41321568e54b76d41121
3
  size 5176
vocab.json CHANGED
@@ -1,34 +1,31 @@
1
  {
2
- "'": 27,
3
- "</s>": 2,
4
- "<pad>": 0,
5
- "<s>": 1,
6
- "<unk>": 3,
7
- "A": 7,
8
- "B": 24,
9
- "C": 19,
10
- "D": 14,
11
- "E": 5,
12
- "F": 20,
13
- "G": 21,
14
- "H": 11,
15
- "I": 10,
16
- "J": 29,
17
- "K": 26,
18
- "L": 15,
19
- "M": 17,
20
- "N": 9,
21
- "O": 8,
22
- "P": 23,
23
- "Q": 30,
24
- "R": 13,
25
- "S": 12,
26
- "T": 6,
27
- "U": 16,
28
- "V": 25,
29
- "W": 18,
30
- "X": 28,
31
- "Y": 22,
32
- "Z": 31,
33
- "|": 4
34
  }
 
1
  {
2
+ "[PAD]": 28,
3
+ "[UNK]": 27,
4
+ "a": 16,
5
+ "b": 25,
6
+ "c": 23,
7
+ "d": 13,
8
+ "e": 5,
9
+ "f": 21,
10
+ "g": 15,
11
+ "h": 20,
12
+ "i": 22,
13
+ "j": 6,
14
+ "k": 12,
15
+ "l": 24,
16
+ "m": 18,
17
+ "n": 9,
18
+ "o": 0,
19
+ "p": 11,
20
+ "q": 1,
21
+ "r": 26,
22
+ "s": 8,
23
+ "t": 14,
24
+ "u": 7,
25
+ "v": 2,
26
+ "w": 17,
27
+ "x": 19,
28
+ "y": 10,
29
+ "z": 4,
30
+ "|": 3
 
 
 
31
  }