IonGrozea commited on
Commit
ed9461a
·
verified ·
1 Parent(s): 0dc91d1

Training in progress, epoch 1

Browse files
Files changed (4) hide show
  1. config.json +8 -8
  2. generation_config.json +26 -10
  3. model.safetensors +2 -2
  4. training_args.bin +1 -1
config.json CHANGED
@@ -9,18 +9,18 @@
9
  "begin_suppress_tokens": null,
10
  "bos_token_id": 50257,
11
  "classifier_proj_size": 256,
12
- "d_model": 384,
13
- "decoder_attention_heads": 6,
14
- "decoder_ffn_dim": 1536,
15
  "decoder_layerdrop": 0.0,
16
- "decoder_layers": 4,
17
  "decoder_start_token_id": 50258,
18
  "dropout": 0.0,
19
  "dtype": "float32",
20
- "encoder_attention_heads": 6,
21
- "encoder_ffn_dim": 1536,
22
  "encoder_layerdrop": 0.0,
23
- "encoder_layers": 4,
24
  "eos_token_id": 50257,
25
  "forced_decoder_ids": [
26
  [
@@ -49,7 +49,7 @@
49
  "max_target_positions": 448,
50
  "median_filter_width": 7,
51
  "model_type": "whisper",
52
- "num_hidden_layers": 4,
53
  "num_mel_bins": 80,
54
  "pad_token_id": 50257,
55
  "scale_embedding": false,
 
9
  "begin_suppress_tokens": null,
10
  "bos_token_id": 50257,
11
  "classifier_proj_size": 256,
12
+ "d_model": 768,
13
+ "decoder_attention_heads": 12,
14
+ "decoder_ffn_dim": 3072,
15
  "decoder_layerdrop": 0.0,
16
+ "decoder_layers": 12,
17
  "decoder_start_token_id": 50258,
18
  "dropout": 0.0,
19
  "dtype": "float32",
20
+ "encoder_attention_heads": 12,
21
+ "encoder_ffn_dim": 3072,
22
  "encoder_layerdrop": 0.0,
23
+ "encoder_layers": 12,
24
  "eos_token_id": 50257,
25
  "forced_decoder_ids": [
26
  [
 
49
  "max_target_positions": 448,
50
  "median_filter_width": 7,
51
  "model_type": "whisper",
52
+ "num_hidden_layers": 12,
53
  "num_mel_bins": 80,
54
  "pad_token_id": 50257,
55
  "scale_embedding": false,
generation_config.json CHANGED
@@ -1,27 +1,43 @@
1
  {
2
  "alignment_heads": [
3
  [
4
- 2,
5
- 2
 
 
 
 
6
  ],
7
  [
8
- 3,
9
  0
10
  ],
11
  [
12
- 3,
13
- 2
14
  ],
15
  [
16
- 3,
17
- 3
18
  ],
19
  [
20
- 3,
21
- 4
 
 
 
 
 
 
 
 
 
 
 
 
22
  ],
23
  [
24
- 3,
25
  5
26
  ]
27
  ],
 
1
  {
2
  "alignment_heads": [
3
  [
4
+ 5,
5
+ 3
6
+ ],
7
+ [
8
+ 5,
9
+ 9
10
  ],
11
  [
12
+ 8,
13
  0
14
  ],
15
  [
16
+ 8,
17
+ 4
18
  ],
19
  [
20
+ 8,
21
+ 7
22
  ],
23
  [
24
+ 8,
25
+ 8
26
+ ],
27
+ [
28
+ 9,
29
+ 0
30
+ ],
31
+ [
32
+ 9,
33
+ 7
34
+ ],
35
+ [
36
+ 9,
37
+ 9
38
  ],
39
  [
40
+ 10,
41
  5
42
  ]
43
  ],
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d563460e05d9374542048a75ecf93c0ce522e85e8d270432bd709a56779275d1
3
- size 151061672
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e7580e62bb03164905ffa6be95a9863a9a5abf54533e773b214f846a58a411c
3
+ size 966995080
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ff39a4739f850346834362756ffe52cda757cf6512ee98bf419f5373eda8d90c
3
  size 6033
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e6d12cc6233cb8c46314e1437089e0fcbdfc8242f4a0d436fc52d3c37d80a9e
3
  size 6033