henriklied commited on
Commit
ad760ee
·
verified ·
1 Parent(s): dbce9de

Training in progress, step 100

Browse files
config.json CHANGED
@@ -1,47 +1,31 @@
1
  {
2
- "_name_or_path": "NbAiLab/nb-whisper-small",
3
  "activation_dropout": 0.1,
4
  "activation_function": "gelu",
5
  "alignment_heads": [
6
  [
7
- 5,
8
- 3
9
  ],
10
  [
11
- 5,
12
- 9
13
- ],
14
- [
15
- 8,
16
- 0
17
- ],
18
- [
19
- 8,
20
  4
21
  ],
22
  [
23
- 8,
24
- 7
25
  ],
26
  [
27
- 8,
28
- 8
29
  ],
30
  [
31
- 9,
32
  0
33
  ],
34
  [
35
- 9,
36
- 7
37
- ],
38
- [
39
- 9,
40
- 9
41
- ],
42
- [
43
- 10,
44
- 5
45
  ]
46
  ],
47
  "apply_spec_augment": false,
@@ -52,17 +36,17 @@
52
  "begin_suppress_tokens": null,
53
  "bos_token_id": 50257,
54
  "classifier_proj_size": 256,
55
- "d_model": 768,
56
- "decoder_attention_heads": 12,
57
- "decoder_ffn_dim": 3072,
58
  "decoder_layerdrop": 0,
59
- "decoder_layers": 12,
60
  "decoder_start_token_id": 50258,
61
  "dropout": 0,
62
- "encoder_attention_heads": 12,
63
- "encoder_ffn_dim": 3072,
64
  "encoder_layerdrop": 0,
65
- "encoder_layers": 12,
66
  "eos_token_id": 50257,
67
  "forced_decoder_ids": [
68
  [
@@ -192,7 +176,7 @@
192
  "max_target_positions": 448,
193
  "median_filter_width": 7,
194
  "model_type": "whisper",
195
- "num_hidden_layers": 12,
196
  "num_mel_bins": 80,
197
  "pad_token_id": 50257,
198
  "scale_embedding": false,
 
1
  {
2
+ "_name_or_path": "NbAiLab/nb-whisper-medium",
3
  "activation_dropout": 0.1,
4
  "activation_function": "gelu",
5
  "alignment_heads": [
6
  [
7
+ 13,
8
+ 15
9
  ],
10
  [
11
+ 15,
 
 
 
 
 
 
 
 
12
  4
13
  ],
14
  [
15
+ 15,
16
+ 15
17
  ],
18
  [
19
+ 16,
20
+ 1
21
  ],
22
  [
23
+ 20,
24
  0
25
  ],
26
  [
27
+ 23,
28
+ 4
 
 
 
 
 
 
 
 
29
  ]
30
  ],
31
  "apply_spec_augment": false,
 
36
  "begin_suppress_tokens": null,
37
  "bos_token_id": 50257,
38
  "classifier_proj_size": 256,
39
+ "d_model": 1024,
40
+ "decoder_attention_heads": 16,
41
+ "decoder_ffn_dim": 4096,
42
  "decoder_layerdrop": 0,
43
+ "decoder_layers": 24,
44
  "decoder_start_token_id": 50258,
45
  "dropout": 0,
46
+ "encoder_attention_heads": 16,
47
+ "encoder_ffn_dim": 4096,
48
  "encoder_layerdrop": 0,
49
+ "encoder_layers": 24,
50
  "eos_token_id": 50257,
51
  "forced_decoder_ids": [
52
  [
 
176
  "max_target_positions": 448,
177
  "median_filter_width": 7,
178
  "model_type": "whisper",
179
+ "num_hidden_layers": 24,
180
  "num_mel_bins": 80,
181
  "pad_token_id": 50257,
182
  "scale_embedding": false,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ae381d94105f53a29b31bb65123236fc3bd1524a5cba89b459eb1d929ca8c885
3
- size 966995080
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:541adf965e2c7c048c009227fd43a0481f15bda1f978dac68ebe07dfc9dd3dca
3
+ size 3055544304
runs/Jan15_11-23-44_frail-song-speaks-fin-01/events.out.tfevents.1736940226.frail-song-speaks-fin-01.14827.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c14475f9e5a5dbae6915750dc75535051f01fe7f38dbcd49c01c5e572ffb1b0f
3
- size 14139
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ae0363b8a4abf2f8aaa73ba2295fb4876e3531a4a4850cf21f8cbca507a7271
3
+ size 17679
runs/Jan15_11-43-17_frail-song-speaks-fin-01/events.out.tfevents.1736941400.frail-song-speaks-fin-01.17384.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:09c9970af042eb8dae20ead900bde685024c8b71babb935ec61b041d8e4c3171
3
+ size 11987
runs/Jan15_11-50-54_frail-song-speaks-fin-01/events.out.tfevents.1736941857.frail-song-speaks-fin-01.19147.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d794284234563c3b10c8f6d883e89a3a823a826fb4f9e22543b122dfcd8babf
3
+ size 11920
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b8dd7c79dc6592da4cbf0c963b2950a4f27245f9e731393914999a91a155eeac
3
  size 5496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d7954e0dbd81bd8f65cadfb27b7416a982cce6a3f8f770eb6a8a7fa311101dc9
3
  size 5496