rbcurzon commited on
Commit
a4f4d51
·
verified ·
1 Parent(s): df979be

Training in progress, step 1000

Browse files
config.json CHANGED
@@ -9,17 +9,17 @@
9
  "begin_suppress_tokens": null,
10
  "bos_token_id": 50257,
11
  "classifier_proj_size": 256,
12
- "d_model": 1024,
13
- "decoder_attention_heads": 16,
14
- "decoder_ffn_dim": 4096,
15
  "decoder_layerdrop": 0.0,
16
- "decoder_layers": 24,
17
  "decoder_start_token_id": 50258,
18
  "dropout": 0.0,
19
- "encoder_attention_heads": 16,
20
- "encoder_ffn_dim": 4096,
21
  "encoder_layerdrop": 0.0,
22
- "encoder_layers": 24,
23
  "eos_token_id": 50257,
24
  "forced_decoder_ids": null,
25
  "init_std": 0.02,
@@ -35,12 +35,12 @@
35
  "max_target_positions": 448,
36
  "median_filter_width": 7,
37
  "model_type": "whisper",
38
- "num_hidden_layers": 24,
39
  "num_mel_bins": 80,
40
  "pad_token_id": 50257,
41
  "scale_embedding": false,
42
  "torch_dtype": "float32",
43
- "transformers_version": "4.54.0.dev0",
44
  "use_cache": true,
45
  "use_weighted_layer_sum": false,
46
  "vocab_size": 51865
 
9
  "begin_suppress_tokens": null,
10
  "bos_token_id": 50257,
11
  "classifier_proj_size": 256,
12
+ "d_model": 768,
13
+ "decoder_attention_heads": 12,
14
+ "decoder_ffn_dim": 3072,
15
  "decoder_layerdrop": 0.0,
16
+ "decoder_layers": 12,
17
  "decoder_start_token_id": 50258,
18
  "dropout": 0.0,
19
+ "encoder_attention_heads": 12,
20
+ "encoder_ffn_dim": 3072,
21
  "encoder_layerdrop": 0.0,
22
+ "encoder_layers": 12,
23
  "eos_token_id": 50257,
24
  "forced_decoder_ids": null,
25
  "init_std": 0.02,
 
35
  "max_target_positions": 448,
36
  "median_filter_width": 7,
37
  "model_type": "whisper",
38
+ "num_hidden_layers": 12,
39
  "num_mel_bins": 80,
40
  "pad_token_id": 50257,
41
  "scale_embedding": false,
42
  "torch_dtype": "float32",
43
+ "transformers_version": "4.56.0.dev0",
44
  "use_cache": true,
45
  "use_weighted_layer_sum": false,
46
  "vocab_size": 51865
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:40da61b3ab3b9af8292cb0714ddbad95b5293ea7934d6c43da6d73550ea9a50b
3
- size 3055544304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e8e481793b3fbcde1209edba28064c7640b165d3a76068f1b0f1e409f940624
3
+ size 966995080
runs/Aug06_03-07-52_4853604300ab/events.out.tfevents.1754450181.4853604300ab.433.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45d429bc46bd3f9edc21a3e954b7e47da69e6e2277356dfc2a91017c5eaab391
3
+ size 4184
runs/Aug06_03-22-08_4853604300ab/events.out.tfevents.1754450534.4853604300ab.752.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:972f9bf6b8cd5a441961e99e82763abce3bc9f02b2a4fc0873fb235fec7ee4bd
3
+ size 6620
runs/Aug06_03-22-53_4853604300ab/events.out.tfevents.1754450579.4853604300ab.883.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b041350b7f17cd309406bb091628cd1ca13ad0b77380d8c68d70eb9425b92202
3
+ size 4184
runs/Aug06_03-24-42_4853604300ab/events.out.tfevents.1754450689.4853604300ab.1011.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6d405a9a793ec16c82cee69b8fdee8ed9e966e7599db0a89267867b27223ec1a
3
+ size 6646
runs/Aug06_03-26-40_4853604300ab/events.out.tfevents.1754450806.4853604300ab.1137.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ce10a00912407760c4a91267d78eb229b4501bae074de872262468bf4f5277d
3
+ size 4184
runs/Aug06_03-27-27_4853604300ab/events.out.tfevents.1754450852.4853604300ab.1267.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a4064ff2114dd059035c54989e886e0891877aab30978435c154fbc2e958b3e
3
+ size 15361
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d9c2bfbdf226de45fd131e214ebdf62a43da8c122186cc86818ec3e5096223fe
3
- size 5969
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:04f1a1f5cb52d8174af744e758af0335f3a5584f7b4871be0eb5526b0ce7019a
3
+ size 5560