Daehyun Ahn commited on
Commit
1830c1b
·
verified ·
1 Parent(s): 8f63130

initial commit

Browse files
added_tokens.json ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "\t\t": 50294,
3
+ "\t\t\t": 50293,
4
+ "\t\t\t\t": 50292,
5
+ "\t\t\t\t\t": 50291,
6
+ "\t\t\t\t\t\t": 50290,
7
+ "\t\t\t\t\t\t\t": 50289,
8
+ "\t\t\t\t\t\t\t\t": 50288,
9
+ "\t\t\t\t\t\t\t\t\t": 50287,
10
+ " ": 50286,
11
+ " ": 50285,
12
+ " ": 50284,
13
+ " ": 50283,
14
+ " ": 50282,
15
+ " ": 50281,
16
+ " ": 50280,
17
+ " ": 50279,
18
+ " ": 50278,
19
+ " ": 50277,
20
+ " ": 50276,
21
+ " ": 50275,
22
+ " ": 50274,
23
+ " ": 50273,
24
+ " ": 50272,
25
+ " ": 50271,
26
+ " ": 50270,
27
+ " ": 50269,
28
+ " ": 50268,
29
+ " ": 50267,
30
+ " ": 50266,
31
+ " ": 50265,
32
+ " ": 50264,
33
+ " ": 50263,
34
+ " ": 50262,
35
+ " ": 50261,
36
+ " ": 50260,
37
+ " ": 50259,
38
+ " ": 50258,
39
+ " ": 50257
40
+ }
config.json CHANGED
@@ -1,33 +1,30 @@
1
  {
2
  "_name_or_path": "squeezebits/dummy_slm",
3
  "architectures": [
4
- "OlmoeForCausalLM"
5
  ],
6
- "attention_bias": false,
7
  "attention_dropout": 0.0,
8
- "clip_qkv": null,
9
- "eos_token_id": 50279,
10
- "hidden_act": "silu",
11
- "hidden_size": 512,
 
12
  "initializer_range": 0.02,
13
- "intermediate_size": 256,
14
- "max_position_embeddings": 4096,
15
- "model_type": "olmoe",
16
- "norm_topk_prob": false,
17
- "num_attention_heads": 16,
18
- "num_experts": 2,
19
- "num_experts_per_tok": 1,
20
  "num_hidden_layers": 1,
21
- "num_key_value_heads": 16,
22
- "output_router_logits": false,
23
- "pad_token_id": 1,
24
- "rms_norm_eps": 1e-05,
25
  "rope_scaling": null,
26
  "rope_theta": 10000.0,
27
- "router_aux_loss_coef": 0.01,
28
  "tie_word_embeddings": false,
29
  "torch_dtype": "bfloat16",
30
  "transformers_version": "4.45.2",
31
  "use_cache": true,
32
- "vocab_size": 50304
33
  }
 
1
  {
2
  "_name_or_path": "squeezebits/dummy_slm",
3
  "architectures": [
4
+ "PhiForCausalLM"
5
  ],
 
6
  "attention_dropout": 0.0,
7
+ "bos_token_id": null,
8
+ "embd_pdrop": 0.0,
9
+ "eos_token_id": null,
10
+ "hidden_act": "gelu_new",
11
+ "hidden_size": 128,
12
  "initializer_range": 0.02,
13
+ "intermediate_size": 512,
14
+ "layer_norm_eps": 1e-05,
15
+ "max_position_embeddings": 2048,
16
+ "model_type": "phi",
17
+ "num_attention_heads": 4,
 
 
18
  "num_hidden_layers": 1,
19
+ "num_key_value_heads": 4,
20
+ "partial_rotary_factor": 0.5,
21
+ "qk_layernorm": false,
22
+ "resid_pdrop": 0.0,
23
  "rope_scaling": null,
24
  "rope_theta": 10000.0,
 
25
  "tie_word_embeddings": false,
26
  "torch_dtype": "bfloat16",
27
  "transformers_version": "4.45.2",
28
  "use_cache": true,
29
+ "vocab_size": 51200
30
  }
generation_config.json CHANGED
@@ -1,6 +1,4 @@
1
  {
2
  "_from_model_config": true,
3
- "eos_token_id": 50279,
4
- "pad_token_id": 1,
5
  "transformers_version": "4.45.2"
6
  }
 
1
  {
2
  "_from_model_config": true,
 
 
3
  "transformers_version": "4.45.2"
4
  }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7149ba62ce234bb5f9495d899d26b184b933b61cf2669f626b0f04a72ca88dd9
3
- size 106701808
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d022963075ff5ad51a150eb0dbcd08ab5032e9ed7ae5651a962f86887c0fd5e6
3
+ size 26715312
special_tokens_map.json CHANGED
@@ -1,4 +1,11 @@
1
  {
 
 
 
 
 
 
 
2
  "eos_token": {
3
  "content": "<|endoftext|>",
4
  "lstrip": false,
@@ -6,8 +13,8 @@
6
  "rstrip": false,
7
  "single_word": false
8
  },
9
- "pad_token": {
10
- "content": "<|padding|>",
11
  "lstrip": false,
12
  "normalized": false,
13
  "rstrip": false,
 
1
  {
2
+ "bos_token": {
3
+ "content": "<|endoftext|>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
  "eos_token": {
10
  "content": "<|endoftext|>",
11
  "lstrip": false,
 
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
+ "unk_token": {
17
+ "content": "<|endoftext|>",
18
  "lstrip": false,
19
  "normalized": false,
20
  "rstrip": false,
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1,25 +1,71 @@
1
  {
2
- "add_bos_token": false,
3
- "add_eos_token": false,
4
  "add_prefix_space": false,
5
  "added_tokens_decoder": {
6
- "0": {
7
- "content": "|||IP_ADDRESS|||",
 
 
 
 
 
 
 
 
8
  "lstrip": false,
9
  "normalized": true,
10
  "rstrip": false,
11
  "single_word": false,
12
  "special": false
13
  },
14
- "1": {
15
- "content": "<|padding|>",
16
  "lstrip": false,
17
- "normalized": false,
18
  "rstrip": false,
19
  "single_word": false,
20
- "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
21
  },
22
- "50254": {
 
 
 
 
 
 
 
 
23
  "content": " ",
24
  "lstrip": false,
25
  "normalized": true,
@@ -27,7 +73,7 @@
27
  "single_word": false,
28
  "special": false
29
  },
30
- "50255": {
31
  "content": " ",
32
  "lstrip": false,
33
  "normalized": true,
@@ -35,7 +81,7 @@
35
  "single_word": false,
36
  "special": false
37
  },
38
- "50256": {
39
  "content": " ",
40
  "lstrip": false,
41
  "normalized": true,
@@ -43,7 +89,7 @@
43
  "single_word": false,
44
  "special": false
45
  },
46
- "50257": {
47
  "content": " ",
48
  "lstrip": false,
49
  "normalized": true,
@@ -51,7 +97,7 @@
51
  "single_word": false,
52
  "special": false
53
  },
54
- "50258": {
55
  "content": " ",
56
  "lstrip": false,
57
  "normalized": true,
@@ -59,7 +105,7 @@
59
  "single_word": false,
60
  "special": false
61
  },
62
- "50259": {
63
  "content": " ",
64
  "lstrip": false,
65
  "normalized": true,
@@ -67,7 +113,7 @@
67
  "single_word": false,
68
  "special": false
69
  },
70
- "50260": {
71
  "content": " ",
72
  "lstrip": false,
73
  "normalized": true,
@@ -75,7 +121,7 @@
75
  "single_word": false,
76
  "special": false
77
  },
78
- "50261": {
79
  "content": " ",
80
  "lstrip": false,
81
  "normalized": true,
@@ -83,7 +129,7 @@
83
  "single_word": false,
84
  "special": false
85
  },
86
- "50262": {
87
  "content": " ",
88
  "lstrip": false,
89
  "normalized": true,
@@ -91,7 +137,7 @@
91
  "single_word": false,
92
  "special": false
93
  },
94
- "50263": {
95
  "content": " ",
96
  "lstrip": false,
97
  "normalized": true,
@@ -99,7 +145,7 @@
99
  "single_word": false,
100
  "special": false
101
  },
102
- "50264": {
103
  "content": " ",
104
  "lstrip": false,
105
  "normalized": true,
@@ -107,7 +153,7 @@
107
  "single_word": false,
108
  "special": false
109
  },
110
- "50265": {
111
  "content": " ",
112
  "lstrip": false,
113
  "normalized": true,
@@ -115,7 +161,7 @@
115
  "single_word": false,
116
  "special": false
117
  },
118
- "50266": {
119
  "content": " ",
120
  "lstrip": false,
121
  "normalized": true,
@@ -123,7 +169,7 @@
123
  "single_word": false,
124
  "special": false
125
  },
126
- "50267": {
127
  "content": " ",
128
  "lstrip": false,
129
  "normalized": true,
@@ -131,7 +177,7 @@
131
  "single_word": false,
132
  "special": false
133
  },
134
- "50268": {
135
  "content": " ",
136
  "lstrip": false,
137
  "normalized": true,
@@ -139,7 +185,7 @@
139
  "single_word": false,
140
  "special": false
141
  },
142
- "50269": {
143
  "content": " ",
144
  "lstrip": false,
145
  "normalized": true,
@@ -147,7 +193,7 @@
147
  "single_word": false,
148
  "special": false
149
  },
150
- "50270": {
151
  "content": " ",
152
  "lstrip": false,
153
  "normalized": true,
@@ -155,7 +201,7 @@
155
  "single_word": false,
156
  "special": false
157
  },
158
- "50271": {
159
  "content": " ",
160
  "lstrip": false,
161
  "normalized": true,
@@ -163,7 +209,7 @@
163
  "single_word": false,
164
  "special": false
165
  },
166
- "50272": {
167
  "content": " ",
168
  "lstrip": false,
169
  "normalized": true,
@@ -171,7 +217,7 @@
171
  "single_word": false,
172
  "special": false
173
  },
174
- "50273": {
175
  "content": " ",
176
  "lstrip": false,
177
  "normalized": true,
@@ -179,7 +225,7 @@
179
  "single_word": false,
180
  "special": false
181
  },
182
- "50274": {
183
  "content": " ",
184
  "lstrip": false,
185
  "normalized": true,
@@ -187,7 +233,7 @@
187
  "single_word": false,
188
  "special": false
189
  },
190
- "50275": {
191
  "content": " ",
192
  "lstrip": false,
193
  "normalized": true,
@@ -195,7 +241,7 @@
195
  "single_word": false,
196
  "special": false
197
  },
198
- "50276": {
199
  "content": " ",
200
  "lstrip": false,
201
  "normalized": true,
@@ -203,36 +249,76 @@
203
  "single_word": false,
204
  "special": false
205
  },
206
- "50277": {
207
- "content": "|||EMAIL_ADDRESS|||",
208
  "lstrip": false,
209
  "normalized": true,
210
  "rstrip": false,
211
  "single_word": false,
212
  "special": false
213
  },
214
- "50278": {
215
- "content": "|||PHONE_NUMBER|||",
216
  "lstrip": false,
217
  "normalized": true,
218
  "rstrip": false,
219
  "single_word": false,
220
  "special": false
221
  },
222
- "50279": {
223
- "content": "<|endoftext|>",
224
  "lstrip": false,
225
- "normalized": false,
226
  "rstrip": false,
227
  "single_word": false,
228
- "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
229
  }
230
  },
231
- "bos_token": null,
232
  "clean_up_tokenization_spaces": true,
233
  "eos_token": "<|endoftext|>",
234
- "model_max_length": 1000000000000000019884624838656,
235
- "pad_token": "<|padding|>",
236
- "tokenizer_class": "GPTNeoXTokenizer",
237
- "unk_token": null
238
  }
 
1
  {
 
 
2
  "add_prefix_space": false,
3
  "added_tokens_decoder": {
4
+ "50256": {
5
+ "content": "<|endoftext|>",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "50257": {
13
+ "content": " ",
14
  "lstrip": false,
15
  "normalized": true,
16
  "rstrip": false,
17
  "single_word": false,
18
  "special": false
19
  },
20
+ "50258": {
21
+ "content": " ",
22
  "lstrip": false,
23
+ "normalized": true,
24
  "rstrip": false,
25
  "single_word": false,
26
+ "special": false
27
+ },
28
+ "50259": {
29
+ "content": " ",
30
+ "lstrip": false,
31
+ "normalized": true,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": false
35
+ },
36
+ "50260": {
37
+ "content": " ",
38
+ "lstrip": false,
39
+ "normalized": true,
40
+ "rstrip": false,
41
+ "single_word": false,
42
+ "special": false
43
+ },
44
+ "50261": {
45
+ "content": " ",
46
+ "lstrip": false,
47
+ "normalized": true,
48
+ "rstrip": false,
49
+ "single_word": false,
50
+ "special": false
51
+ },
52
+ "50262": {
53
+ "content": " ",
54
+ "lstrip": false,
55
+ "normalized": true,
56
+ "rstrip": false,
57
+ "single_word": false,
58
+ "special": false
59
  },
60
+ "50263": {
61
+ "content": " ",
62
+ "lstrip": false,
63
+ "normalized": true,
64
+ "rstrip": false,
65
+ "single_word": false,
66
+ "special": false
67
+ },
68
+ "50264": {
69
  "content": " ",
70
  "lstrip": false,
71
  "normalized": true,
 
73
  "single_word": false,
74
  "special": false
75
  },
76
+ "50265": {
77
  "content": " ",
78
  "lstrip": false,
79
  "normalized": true,
 
81
  "single_word": false,
82
  "special": false
83
  },
84
+ "50266": {
85
  "content": " ",
86
  "lstrip": false,
87
  "normalized": true,
 
89
  "single_word": false,
90
  "special": false
91
  },
92
+ "50267": {
93
  "content": " ",
94
  "lstrip": false,
95
  "normalized": true,
 
97
  "single_word": false,
98
  "special": false
99
  },
100
+ "50268": {
101
  "content": " ",
102
  "lstrip": false,
103
  "normalized": true,
 
105
  "single_word": false,
106
  "special": false
107
  },
108
+ "50269": {
109
  "content": " ",
110
  "lstrip": false,
111
  "normalized": true,
 
113
  "single_word": false,
114
  "special": false
115
  },
116
+ "50270": {
117
  "content": " ",
118
  "lstrip": false,
119
  "normalized": true,
 
121
  "single_word": false,
122
  "special": false
123
  },
124
+ "50271": {
125
  "content": " ",
126
  "lstrip": false,
127
  "normalized": true,
 
129
  "single_word": false,
130
  "special": false
131
  },
132
+ "50272": {
133
  "content": " ",
134
  "lstrip": false,
135
  "normalized": true,
 
137
  "single_word": false,
138
  "special": false
139
  },
140
+ "50273": {
141
  "content": " ",
142
  "lstrip": false,
143
  "normalized": true,
 
145
  "single_word": false,
146
  "special": false
147
  },
148
+ "50274": {
149
  "content": " ",
150
  "lstrip": false,
151
  "normalized": true,
 
153
  "single_word": false,
154
  "special": false
155
  },
156
+ "50275": {
157
  "content": " ",
158
  "lstrip": false,
159
  "normalized": true,
 
161
  "single_word": false,
162
  "special": false
163
  },
164
+ "50276": {
165
  "content": " ",
166
  "lstrip": false,
167
  "normalized": true,
 
169
  "single_word": false,
170
  "special": false
171
  },
172
+ "50277": {
173
  "content": " ",
174
  "lstrip": false,
175
  "normalized": true,
 
177
  "single_word": false,
178
  "special": false
179
  },
180
+ "50278": {
181
  "content": " ",
182
  "lstrip": false,
183
  "normalized": true,
 
185
  "single_word": false,
186
  "special": false
187
  },
188
+ "50279": {
189
  "content": " ",
190
  "lstrip": false,
191
  "normalized": true,
 
193
  "single_word": false,
194
  "special": false
195
  },
196
+ "50280": {
197
  "content": " ",
198
  "lstrip": false,
199
  "normalized": true,
 
201
  "single_word": false,
202
  "special": false
203
  },
204
+ "50281": {
205
  "content": " ",
206
  "lstrip": false,
207
  "normalized": true,
 
209
  "single_word": false,
210
  "special": false
211
  },
212
+ "50282": {
213
  "content": " ",
214
  "lstrip": false,
215
  "normalized": true,
 
217
  "single_word": false,
218
  "special": false
219
  },
220
+ "50283": {
221
  "content": " ",
222
  "lstrip": false,
223
  "normalized": true,
 
225
  "single_word": false,
226
  "special": false
227
  },
228
+ "50284": {
229
  "content": " ",
230
  "lstrip": false,
231
  "normalized": true,
 
233
  "single_word": false,
234
  "special": false
235
  },
236
+ "50285": {
237
  "content": " ",
238
  "lstrip": false,
239
  "normalized": true,
 
241
  "single_word": false,
242
  "special": false
243
  },
244
+ "50286": {
245
  "content": " ",
246
  "lstrip": false,
247
  "normalized": true,
 
249
  "single_word": false,
250
  "special": false
251
  },
252
+ "50287": {
253
+ "content": "\t\t\t\t\t\t\t\t\t",
254
  "lstrip": false,
255
  "normalized": true,
256
  "rstrip": false,
257
  "single_word": false,
258
  "special": false
259
  },
260
+ "50288": {
261
+ "content": "\t\t\t\t\t\t\t\t",
262
  "lstrip": false,
263
  "normalized": true,
264
  "rstrip": false,
265
  "single_word": false,
266
  "special": false
267
  },
268
+ "50289": {
269
+ "content": "\t\t\t\t\t\t\t",
270
  "lstrip": false,
271
+ "normalized": true,
272
  "rstrip": false,
273
  "single_word": false,
274
+ "special": false
275
+ },
276
+ "50290": {
277
+ "content": "\t\t\t\t\t\t",
278
+ "lstrip": false,
279
+ "normalized": true,
280
+ "rstrip": false,
281
+ "single_word": false,
282
+ "special": false
283
+ },
284
+ "50291": {
285
+ "content": "\t\t\t\t\t",
286
+ "lstrip": false,
287
+ "normalized": true,
288
+ "rstrip": false,
289
+ "single_word": false,
290
+ "special": false
291
+ },
292
+ "50292": {
293
+ "content": "\t\t\t\t",
294
+ "lstrip": false,
295
+ "normalized": true,
296
+ "rstrip": false,
297
+ "single_word": false,
298
+ "special": false
299
+ },
300
+ "50293": {
301
+ "content": "\t\t\t",
302
+ "lstrip": false,
303
+ "normalized": true,
304
+ "rstrip": false,
305
+ "single_word": false,
306
+ "special": false
307
+ },
308
+ "50294": {
309
+ "content": "\t\t",
310
+ "lstrip": false,
311
+ "normalized": true,
312
+ "rstrip": false,
313
+ "single_word": false,
314
+ "special": false
315
  }
316
  },
317
+ "bos_token": "<|endoftext|>",
318
  "clean_up_tokenization_spaces": true,
319
  "eos_token": "<|endoftext|>",
320
+ "model_max_length": 2048,
321
+ "return_token_type_ids": false,
322
+ "tokenizer_class": "CodeGenTokenizer",
323
+ "unk_token": "<|endoftext|>"
324
  }
vocab.json ADDED
The diff for this file is too large to render. See raw diff