maliced commited on
Commit
ed0b40d
·
verified ·
1 Parent(s): f0f6f3d

Training in progress, step 3000

Browse files
config.json CHANGED
@@ -1,65 +1,65 @@
1
- {
2
- "architectures": [
3
- "MDDTransformer"
4
- ],
5
- "batch_first": true,
6
- "d_model": 32,
7
- "dim_feedforward": 256,
8
- "hidden_dropout_prob": 0.1,
9
- "input_channels": 2,
10
- "max_source_positions": 3000,
11
- "model_type": "mdd_transformer",
12
- "num_attention_heads_decoder": 2,
13
- "num_attention_heads_encoder": 4,
14
- "num_classes": 43,
15
- "num_cross_attention_heads": 2,
16
- "num_decoder_layers": 2,
17
- "num_encoder_layers": 2,
18
- "num_mel_bins": 80,
19
- "projector_activation": "softmax",
20
- "tokens": [
21
- "SIL",
22
- "AA",
23
- "AE",
24
- "AH",
25
- "AO",
26
- "AW",
27
- "AX",
28
- "AY",
29
- "B",
30
- "CH",
31
- "D",
32
- "DH",
33
- "EH",
34
- "ER",
35
- "EY",
36
- "F",
37
- "G",
38
- "HH",
39
- "IH",
40
- "IY",
41
- "JH",
42
- "K",
43
- "L",
44
- "M",
45
- "N",
46
- "NG",
47
- "OW",
48
- "OY",
49
- "P",
50
- "R",
51
- "S",
52
- "SH",
53
- "T",
54
- "TH",
55
- "UH",
56
- "UW",
57
- "V",
58
- "W",
59
- "Y",
60
- "Z",
61
- "ZH"
62
- ],
63
- "torch_dtype": "float32",
64
- "transformers_version": "4.52.3"
65
- }
 
1
+ {
2
+ "architectures": [
3
+ "MDDTransformer"
4
+ ],
5
+ "batch_first": true,
6
+ "d_model": 32,
7
+ "dim_feedforward": 256,
8
+ "hidden_dropout_prob": 0.1,
9
+ "input_channels": 2,
10
+ "max_source_positions": 3000,
11
+ "model_type": "mdd_transformer",
12
+ "num_attention_heads_decoder": 2,
13
+ "num_attention_heads_encoder": 4,
14
+ "num_classes": 43,
15
+ "num_cross_attention_heads": 2,
16
+ "num_decoder_layers": 2,
17
+ "num_encoder_layers": 2,
18
+ "num_mel_bins": 80,
19
+ "projector_activation": "softmax",
20
+ "tokens": [
21
+ "SIL",
22
+ "AA",
23
+ "AE",
24
+ "AH",
25
+ "AO",
26
+ "AW",
27
+ "AX",
28
+ "AY",
29
+ "B",
30
+ "CH",
31
+ "D",
32
+ "DH",
33
+ "EH",
34
+ "ER",
35
+ "EY",
36
+ "F",
37
+ "G",
38
+ "HH",
39
+ "IH",
40
+ "IY",
41
+ "JH",
42
+ "K",
43
+ "L",
44
+ "M",
45
+ "N",
46
+ "NG",
47
+ "OW",
48
+ "OY",
49
+ "P",
50
+ "R",
51
+ "S",
52
+ "SH",
53
+ "T",
54
+ "TH",
55
+ "UH",
56
+ "UW",
57
+ "V",
58
+ "W",
59
+ "Y",
60
+ "Z",
61
+ "ZH"
62
+ ],
63
+ "torch_dtype": "float32",
64
+ "transformers_version": "4.52.1"
65
+ }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:acb6c2720f128e70a251512f5cd473d941768a9a23392d2e1fbf5a3cbf3d9316
3
  size 1009168
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c32fb015b0128d6693e9b9ce40320be2d806123c8d52b1c60061b9e4723de63
3
  size 1009168
runs/May30_16-09-27_malice-laptop/events.out.tfevents.1748614169.malice-laptop.9668.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:56136bcf4a8315a25f265d2633ef3b92b5fa7999548ce7f0268f9b0b3d5b2216
3
- size 53503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3cc29015bbf985928f32e0111286a99b08ed0f23a664cb5790f098fc4682a43f
3
+ size 63143
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bf82f87052e26ded05e89558f5bd204fdbae5947b5556fbda36f3aec8b45dc45
3
  size 5905
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1a20204c0ce3ff99c686cea7ac55c8099195049bc2c55c54b6423d4179bf1dc
3
  size 5905