Siphh commited on
Commit
85f5ec9
·
1 Parent(s): 68137fd

Training in progress, step 2000

Browse files
config.json CHANGED
@@ -25,20 +25,7 @@
25
  "encoder_layerdrop": 0.0,
26
  "encoder_layers": 12,
27
  "eos_token_id": 50257,
28
- "forced_decoder_ids": [
29
- [
30
- 1,
31
- 50259
32
- ],
33
- [
34
- 2,
35
- 50359
36
- ],
37
- [
38
- 3,
39
- 50363
40
- ]
41
- ],
42
  "init_std": 0.02,
43
  "is_encoder_decoder": true,
44
  "mask_feature_length": 10,
@@ -56,94 +43,7 @@
56
  "num_mel_bins": 80,
57
  "pad_token_id": 50257,
58
  "scale_embedding": false,
59
- "suppress_tokens": [
60
- 1,
61
- 2,
62
- 7,
63
- 8,
64
- 9,
65
- 10,
66
- 14,
67
- 25,
68
- 26,
69
- 27,
70
- 28,
71
- 29,
72
- 31,
73
- 58,
74
- 59,
75
- 60,
76
- 61,
77
- 62,
78
- 63,
79
- 90,
80
- 91,
81
- 92,
82
- 93,
83
- 359,
84
- 503,
85
- 522,
86
- 542,
87
- 873,
88
- 893,
89
- 902,
90
- 918,
91
- 922,
92
- 931,
93
- 1350,
94
- 1853,
95
- 1982,
96
- 2460,
97
- 2627,
98
- 3246,
99
- 3253,
100
- 3268,
101
- 3536,
102
- 3846,
103
- 3961,
104
- 4183,
105
- 4667,
106
- 6585,
107
- 6647,
108
- 7273,
109
- 9061,
110
- 9383,
111
- 10428,
112
- 10929,
113
- 11938,
114
- 12033,
115
- 12331,
116
- 12562,
117
- 13793,
118
- 14157,
119
- 14635,
120
- 15265,
121
- 15618,
122
- 16553,
123
- 16604,
124
- 18362,
125
- 18956,
126
- 20075,
127
- 21675,
128
- 22520,
129
- 26130,
130
- 26161,
131
- 26435,
132
- 28279,
133
- 29464,
134
- 31650,
135
- 32302,
136
- 32470,
137
- 36865,
138
- 42863,
139
- 47425,
140
- 49870,
141
- 50254,
142
- 50258,
143
- 50360,
144
- 50361,
145
- 50362
146
- ],
147
  "torch_dtype": "float32",
148
  "transformers_version": "4.35.2",
149
  "use_cache": true,
 
25
  "encoder_layerdrop": 0.0,
26
  "encoder_layers": 12,
27
  "eos_token_id": 50257,
28
+ "forced_decoder_ids": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
29
  "init_std": 0.02,
30
  "is_encoder_decoder": true,
31
  "mask_feature_length": 10,
 
43
  "num_mel_bins": 80,
44
  "pad_token_id": 50257,
45
  "scale_embedding": false,
46
+ "suppress_tokens": [],
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
47
  "torch_dtype": "float32",
48
  "transformers_version": "4.35.2",
49
  "use_cache": true,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b60f0393eae28a941e30c22f28c492e0c34c9faae25d09c9a75b478646375762
3
  size 966995080
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3d9d5a7c0fb35fdcbf4e91f7480e5f9b59864cad1c459c54c3a7f16303965e7
3
  size 966995080
runs/Dec05_12-56-00_f4ded5a40cc4/events.out.tfevents.1701780980.f4ded5a40cc4.291.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:689eb966adf396ffdf288130e0cdb40ecb92d79f5397d287cbe5ae7c29b67060
3
- size 12589
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d81baf58979b43943dc6dc3799a38f2281a9c1a797dfad5aea46de5100a455fa
3
+ size 13688
runs/Dec05_21-05-18_1959d114e87c/events.out.tfevents.1701810433.1959d114e87c.475.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1bdea9d04d19097e95a5805116a95849758354f2cb5216d2a2e6c71fadeef339
3
+ size 11666
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:111a073b0419d5bece36540bc5680ff2491f4ff8484e1ca4da1a7055a68f10bf
3
  size 4728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb08d15d2ed4dd39ae42c09b1c78b6f4f38a2452597c92e3c628d28f44e7c887
3
  size 4728