ryanscottbarrett commited on
Commit
87cf355
·
verified ·
1 Parent(s): 524a386

Upload folder using huggingface_hub

Browse files
braille256_vocab.json ADDED
@@ -0,0 +1,263 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "[PAD]": 0,
3
+ "[UNK]": 1,
4
+ "[BOS]": 2,
5
+ "[EOS]": 3,
6
+ "[MASK]": 4,
7
+ "⠀": 5,
8
+ "⠁": 6,
9
+ "⠂": 7,
10
+ "⠃": 8,
11
+ "⠄": 9,
12
+ "⠅": 10,
13
+ "⠆": 11,
14
+ "⠇": 12,
15
+ "⠈": 13,
16
+ "⠉": 14,
17
+ "⠊": 15,
18
+ "⠋": 16,
19
+ "⠌": 17,
20
+ "⠍": 18,
21
+ "⠎": 19,
22
+ "⠏": 20,
23
+ "⠐": 21,
24
+ "⠑": 22,
25
+ "⠒": 23,
26
+ "⠓": 24,
27
+ "⠔": 25,
28
+ "⠕": 26,
29
+ "⠖": 27,
30
+ "⠗": 28,
31
+ "⠘": 29,
32
+ "⠙": 30,
33
+ "⠚": 31,
34
+ "⠛": 32,
35
+ "⠜": 33,
36
+ "⠝": 34,
37
+ "⠞": 35,
38
+ "⠟": 36,
39
+ "⠠": 37,
40
+ "⠡": 38,
41
+ "⠢": 39,
42
+ "⠣": 40,
43
+ "⠤": 41,
44
+ "⠥": 42,
45
+ "⠦": 43,
46
+ "⠧": 44,
47
+ "⠨": 45,
48
+ "⠩": 46,
49
+ "⠪": 47,
50
+ "⠫": 48,
51
+ "⠬": 49,
52
+ "⠭": 50,
53
+ "⠮": 51,
54
+ "⠯": 52,
55
+ "⠰": 53,
56
+ "⠱": 54,
57
+ "⠲": 55,
58
+ "⠳": 56,
59
+ "⠴": 57,
60
+ "⠵": 58,
61
+ "⠶": 59,
62
+ "⠷": 60,
63
+ "⠸": 61,
64
+ "⠹": 62,
65
+ "⠺": 63,
66
+ "⠻": 64,
67
+ "⠼": 65,
68
+ "⠽": 66,
69
+ "⠾": 67,
70
+ "⠿": 68,
71
+ "⡀": 69,
72
+ "⡁": 70,
73
+ "⡂": 71,
74
+ "⡃": 72,
75
+ "⡄": 73,
76
+ "⡅": 74,
77
+ "⡆": 75,
78
+ "⡇": 76,
79
+ "⡈": 77,
80
+ "⡉": 78,
81
+ "⡊": 79,
82
+ "⡋": 80,
83
+ "⡌": 81,
84
+ "⡍": 82,
85
+ "⡎": 83,
86
+ "⡏": 84,
87
+ "⡐": 85,
88
+ "⡑": 86,
89
+ "⡒": 87,
90
+ "⡓": 88,
91
+ "⡔": 89,
92
+ "⡕": 90,
93
+ "⡖": 91,
94
+ "⡗": 92,
95
+ "⡘": 93,
96
+ "⡙": 94,
97
+ "⡚": 95,
98
+ "⡛": 96,
99
+ "⡜": 97,
100
+ "⡝": 98,
101
+ "⡞": 99,
102
+ "⡟": 100,
103
+ "⡠": 101,
104
+ "⡡": 102,
105
+ "⡢": 103,
106
+ "⡣": 104,
107
+ "⡤": 105,
108
+ "⡥": 106,
109
+ "⡦": 107,
110
+ "⡧": 108,
111
+ "⡨": 109,
112
+ "⡩": 110,
113
+ "⡪": 111,
114
+ "⡫": 112,
115
+ "⡬": 113,
116
+ "⡭": 114,
117
+ "⡮": 115,
118
+ "⡯": 116,
119
+ "⡰": 117,
120
+ "⡱": 118,
121
+ "⡲": 119,
122
+ "⡳": 120,
123
+ "⡴": 121,
124
+ "⡵": 122,
125
+ "⡶": 123,
126
+ "⡷": 124,
127
+ "⡸": 125,
128
+ "⡹": 126,
129
+ "⡺": 127,
130
+ "⡻": 128,
131
+ "⡼": 129,
132
+ "⡽": 130,
133
+ "⡾": 131,
134
+ "⡿": 132,
135
+ "⢀": 133,
136
+ "⢁": 134,
137
+ "⢂": 135,
138
+ "⢃": 136,
139
+ "⢄": 137,
140
+ "⢅": 138,
141
+ "⢆": 139,
142
+ "⢇": 140,
143
+ "⢈": 141,
144
+ "⢉": 142,
145
+ "⢊": 143,
146
+ "⢋": 144,
147
+ "⢌": 145,
148
+ "⢍": 146,
149
+ "⢎": 147,
150
+ "⢏": 148,
151
+ "⢐": 149,
152
+ "⢑": 150,
153
+ "⢒": 151,
154
+ "⢓": 152,
155
+ "⢔": 153,
156
+ "⢕": 154,
157
+ "⢖": 155,
158
+ "⢗": 156,
159
+ "⢘": 157,
160
+ "⢙": 158,
161
+ "⢚": 159,
162
+ "⢛": 160,
163
+ "⢜": 161,
164
+ "⢝": 162,
165
+ "⢞": 163,
166
+ "⢟": 164,
167
+ "⢠": 165,
168
+ "⢡": 166,
169
+ "⢢": 167,
170
+ "⢣": 168,
171
+ "⢤": 169,
172
+ "⢥": 170,
173
+ "⢦": 171,
174
+ "⢧": 172,
175
+ "⢨": 173,
176
+ "⢩": 174,
177
+ "⢪": 175,
178
+ "⢫": 176,
179
+ "⢬": 177,
180
+ "⢭": 178,
181
+ "⢮": 179,
182
+ "⢯": 180,
183
+ "⢰": 181,
184
+ "⢱": 182,
185
+ "⢲": 183,
186
+ "⢳": 184,
187
+ "⢴": 185,
188
+ "⢵": 186,
189
+ "⢶": 187,
190
+ "⢷": 188,
191
+ "⢸": 189,
192
+ "⢹": 190,
193
+ "⢺": 191,
194
+ "⢻": 192,
195
+ "⢼": 193,
196
+ "⢽": 194,
197
+ "⢾": 195,
198
+ "⢿": 196,
199
+ "⣀": 197,
200
+ "⣁": 198,
201
+ "⣂": 199,
202
+ "⣃": 200,
203
+ "⣄": 201,
204
+ "⣅": 202,
205
+ "⣆": 203,
206
+ "⣇": 204,
207
+ "⣈": 205,
208
+ "⣉": 206,
209
+ "⣊": 207,
210
+ "⣋": 208,
211
+ "⣌": 209,
212
+ "⣍": 210,
213
+ "⣎": 211,
214
+ "⣏": 212,
215
+ "⣐": 213,
216
+ "⣑": 214,
217
+ "⣒": 215,
218
+ "⣓": 216,
219
+ "⣔": 217,
220
+ "⣕": 218,
221
+ "⣖": 219,
222
+ "⣗": 220,
223
+ "⣘": 221,
224
+ "⣙": 222,
225
+ "⣚": 223,
226
+ "⣛": 224,
227
+ "⣜": 225,
228
+ "⣝": 226,
229
+ "⣞": 227,
230
+ "⣟": 228,
231
+ "⣠": 229,
232
+ "⣡": 230,
233
+ "⣢": 231,
234
+ "⣣": 232,
235
+ "⣤": 233,
236
+ "⣥": 234,
237
+ "⣦": 235,
238
+ "⣧": 236,
239
+ "⣨": 237,
240
+ "⣩": 238,
241
+ "⣪": 239,
242
+ "⣫": 240,
243
+ "⣬": 241,
244
+ "⣭": 242,
245
+ "⣮": 243,
246
+ "⣯": 244,
247
+ "⣰": 245,
248
+ "⣱": 246,
249
+ "⣲": 247,
250
+ "⣳": 248,
251
+ "⣴": 249,
252
+ "⣵": 250,
253
+ "⣶": 251,
254
+ "⣷": 252,
255
+ "⣸": 253,
256
+ "⣹": 254,
257
+ "⣺": 255,
258
+ "⣻": 256,
259
+ "⣼": 257,
260
+ "⣽": 258,
261
+ "⣾": 259,
262
+ "⣿": 260
263
+ }
config.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Braille256Model"
4
+ ],
5
+ "dtype": "float32",
6
+ "hidden_size": 512,
7
+ "intermediate_size": 2048,
8
+ "model_type": "braille256",
9
+ "num_attention_heads": 8,
10
+ "num_hidden_layers": 8,
11
+ "transformers_version": "4.57.1",
12
+ "use_dot_pattern_init": true
13
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1add3cd68c802584c00480d9d070aa40e5040b506f1d53c659c396584664b56b
3
+ size 102504944
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "[BOS]",
3
+ "eos_token": "[EOS]",
4
+ "mask_token": "[MASK]",
5
+ "pad_token": "[PAD]",
6
+ "unk_token": "[UNK]"
7
+ }
tokenizer_config.json ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "[BOS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "[EOS]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "4": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "bos_token": "[BOS]",
45
+ "clean_up_tokenization_spaces": false,
46
+ "eos_token": "[EOS]",
47
+ "extra_special_tokens": {},
48
+ "mask_token": "[MASK]",
49
+ "model_max_length": 1000000000000000019884624838656,
50
+ "pad_token": "[PAD]",
51
+ "tokenizer_class": "Braille256Tokenizer",
52
+ "unk_token": "[UNK]"
53
+ }
training_config.json ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "output_dir": "models/braille256_scaled/braille256_scaled_64M",
3
+ "learning_rate": 0.0003,
4
+ "weight_decay": 0.01,
5
+ "warmup_steps": 1000,
6
+ "max_steps": 10000,
7
+ "per_device_train_batch_size": 16,
8
+ "per_device_eval_batch_size": 16,
9
+ "gradient_accumulation_steps": 2,
10
+ "max_seq_length": 512,
11
+ "logging_steps": 100,
12
+ "eval_steps": 500,
13
+ "save_steps": 1000,
14
+ "track_emergent_patterns": true,
15
+ "pattern_analysis_steps": 1000,
16
+ "fp16": true,
17
+ "dataloader_num_workers": 4
18
+ }