g-shuhei commited on
Commit
a7155e2
·
verified ·
1 Parent(s): 37bdad4

s3://sagemaker-whisper-testing/output/base/en/sagemaker-whispertra-250719-0535-009-335b8026/output/model.tar.gz

Browse files
Files changed (3) hide show
  1. config.json +10 -10
  2. generation_config.json +8 -63
  3. model.safetensors +2 -2
config.json CHANGED
@@ -13,17 +13,17 @@
13
  ],
14
  "bos_token_id": 50257,
15
  "classifier_proj_size": 256,
16
- "d_model": 768,
17
- "decoder_attention_heads": 12,
18
- "decoder_ffn_dim": 3072,
19
  "decoder_layerdrop": 0.0,
20
- "decoder_layers": 12,
21
  "decoder_start_token_id": 50257,
22
- "dropout": 0.014221097611095367,
23
- "encoder_attention_heads": 12,
24
- "encoder_ffn_dim": 3072,
25
  "encoder_layerdrop": 0.0,
26
- "encoder_layers": 12,
27
  "eos_token_id": 50256,
28
  "forced_decoder_ids": null,
29
  "init_std": 0.02,
@@ -39,7 +39,7 @@
39
  "max_target_positions": 448,
40
  "median_filter_width": 7,
41
  "model_type": "whisper",
42
- "num_hidden_layers": 12,
43
  "num_mel_bins": 80,
44
  "pad_token_id": 50256,
45
  "scale_embedding": false,
@@ -47,5 +47,5 @@
47
  "transformers_version": "4.45.2",
48
  "use_cache": false,
49
  "use_weighted_layer_sum": false,
50
- "vocab_size": 51869
51
  }
 
13
  ],
14
  "bos_token_id": 50257,
15
  "classifier_proj_size": 256,
16
+ "d_model": 512,
17
+ "decoder_attention_heads": 8,
18
+ "decoder_ffn_dim": 2048,
19
  "decoder_layerdrop": 0.0,
20
+ "decoder_layers": 6,
21
  "decoder_start_token_id": 50257,
22
+ "dropout": 0.005637927715008337,
23
+ "encoder_attention_heads": 8,
24
+ "encoder_ffn_dim": 2048,
25
  "encoder_layerdrop": 0.0,
26
+ "encoder_layers": 6,
27
  "eos_token_id": 50256,
28
  "forced_decoder_ids": null,
29
  "init_std": 0.02,
 
39
  "max_target_positions": 448,
40
  "median_filter_width": 7,
41
  "model_type": "whisper",
42
+ "num_hidden_layers": 6,
43
  "num_mel_bins": 80,
44
  "pad_token_id": 50256,
45
  "scale_embedding": false,
 
47
  "transformers_version": "4.45.2",
48
  "use_cache": false,
49
  "use_weighted_layer_sum": false,
50
+ "vocab_size": 51867
51
  }
generation_config.json CHANGED
@@ -1,80 +1,24 @@
1
  {
2
  "alignment_heads": [
3
  [
4
- 6,
5
- 6
6
- ],
7
- [
8
- 7,
9
- 0
10
- ],
11
- [
12
- 7,
13
  3
14
  ],
15
  [
16
- 7,
17
- 8
18
- ],
19
- [
20
- 8,
21
- 2
22
- ],
23
- [
24
- 8,
25
- 5
26
- ],
27
- [
28
- 8,
29
  7
30
  ],
31
  [
32
- 9,
33
- 0
34
- ],
35
- [
36
- 9,
37
- 4
38
- ],
39
- [
40
- 9,
41
- 8
42
- ],
43
- [
44
- 9,
45
- 10
46
- ],
47
- [
48
- 10,
49
- 0
50
- ],
51
- [
52
- 10,
53
  1
54
  ],
55
  [
56
- 10,
57
- 2
58
- ],
59
- [
60
- 10,
61
- 3
62
- ],
63
- [
64
- 10,
65
- 6
66
- ],
67
- [
68
- 10,
69
- 11
70
- ],
71
- [
72
- 11,
73
- 2
74
  ],
75
  [
76
- 11,
77
- 4
78
  ]
79
  ],
80
  "begin_suppress_tokens": [
@@ -96,6 +40,7 @@
96
  "no_timestamps_token_id": 50362,
97
  "pad_token_id": 50256,
98
  "prev_sot_token_id": 50360,
 
99
  "suppress_tokens": [],
100
  "transformers_version": "4.45.2"
101
  }
 
1
  {
2
  "alignment_heads": [
3
  [
4
+ 3,
 
 
 
 
 
 
 
 
5
  3
6
  ],
7
  [
8
+ 4,
 
 
 
 
 
 
 
 
 
 
 
 
9
  7
10
  ],
11
  [
12
+ 5,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
  1
14
  ],
15
  [
16
+ 5,
17
+ 5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
18
  ],
19
  [
20
+ 5,
21
+ 7
22
  ]
23
  ],
24
  "begin_suppress_tokens": [
 
40
  "no_timestamps_token_id": 50362,
41
  "pad_token_id": 50256,
42
  "prev_sot_token_id": 50360,
43
+ "return_timestamps": false,
44
  "suppress_tokens": [],
45
  "transformers_version": "4.45.2"
46
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4f5f0cc1ff7933348c1407e177adff55828a68aa562bcfc58dbabdec04743d3b
3
- size 967007368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3cdbd67a72b9653850afc5e2cc9d35849d30cf1b6eb719ed33703944005329c6
3
+ size 290408032