rawsun00001 commited on
Commit
f656411
·
verified ·
1 Parent(s): 9b3d705

Banking SMS JSON parser

Browse files
README.md ADDED
@@ -0,0 +1,207 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: gpt2
3
+ library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - base_model:adapter:gpt2
7
+ - lora
8
+ - transformers
9
+ ---
10
+
11
+ # Model Card for Model ID
12
+
13
+ <!-- Provide a quick summary of what the model is/does. -->
14
+
15
+
16
+
17
+ ## Model Details
18
+
19
+ ### Model Description
20
+
21
+ <!-- Provide a longer summary of what this model is. -->
22
+
23
+
24
+
25
+ - **Developed by:** [More Information Needed]
26
+ - **Funded by [optional]:** [More Information Needed]
27
+ - **Shared by [optional]:** [More Information Needed]
28
+ - **Model type:** [More Information Needed]
29
+ - **Language(s) (NLP):** [More Information Needed]
30
+ - **License:** [More Information Needed]
31
+ - **Finetuned from model [optional]:** [More Information Needed]
32
+
33
+ ### Model Sources [optional]
34
+
35
+ <!-- Provide the basic links for the model. -->
36
+
37
+ - **Repository:** [More Information Needed]
38
+ - **Paper [optional]:** [More Information Needed]
39
+ - **Demo [optional]:** [More Information Needed]
40
+
41
+ ## Uses
42
+
43
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
44
+
45
+ ### Direct Use
46
+
47
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
48
+
49
+ [More Information Needed]
50
+
51
+ ### Downstream Use [optional]
52
+
53
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
54
+
55
+ [More Information Needed]
56
+
57
+ ### Out-of-Scope Use
58
+
59
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
60
+
61
+ [More Information Needed]
62
+
63
+ ## Bias, Risks, and Limitations
64
+
65
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
66
+
67
+ [More Information Needed]
68
+
69
+ ### Recommendations
70
+
71
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
72
+
73
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
74
+
75
+ ## How to Get Started with the Model
76
+
77
+ Use the code below to get started with the model.
78
+
79
+ [More Information Needed]
80
+
81
+ ## Training Details
82
+
83
+ ### Training Data
84
+
85
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
86
+
87
+ [More Information Needed]
88
+
89
+ ### Training Procedure
90
+
91
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
92
+
93
+ #### Preprocessing [optional]
94
+
95
+ [More Information Needed]
96
+
97
+
98
+ #### Training Hyperparameters
99
+
100
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
101
+
102
+ #### Speeds, Sizes, Times [optional]
103
+
104
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
105
+
106
+ [More Information Needed]
107
+
108
+ ## Evaluation
109
+
110
+ <!-- This section describes the evaluation protocols and provides the results. -->
111
+
112
+ ### Testing Data, Factors & Metrics
113
+
114
+ #### Testing Data
115
+
116
+ <!-- This should link to a Dataset Card if possible. -->
117
+
118
+ [More Information Needed]
119
+
120
+ #### Factors
121
+
122
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
123
+
124
+ [More Information Needed]
125
+
126
+ #### Metrics
127
+
128
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
129
+
130
+ [More Information Needed]
131
+
132
+ ### Results
133
+
134
+ [More Information Needed]
135
+
136
+ #### Summary
137
+
138
+
139
+
140
+ ## Model Examination [optional]
141
+
142
+ <!-- Relevant interpretability work for the model goes here -->
143
+
144
+ [More Information Needed]
145
+
146
+ ## Environmental Impact
147
+
148
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
149
+
150
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
151
+
152
+ - **Hardware Type:** [More Information Needed]
153
+ - **Hours used:** [More Information Needed]
154
+ - **Cloud Provider:** [More Information Needed]
155
+ - **Compute Region:** [More Information Needed]
156
+ - **Carbon Emitted:** [More Information Needed]
157
+
158
+ ## Technical Specifications [optional]
159
+
160
+ ### Model Architecture and Objective
161
+
162
+ [More Information Needed]
163
+
164
+ ### Compute Infrastructure
165
+
166
+ [More Information Needed]
167
+
168
+ #### Hardware
169
+
170
+ [More Information Needed]
171
+
172
+ #### Software
173
+
174
+ [More Information Needed]
175
+
176
+ ## Citation [optional]
177
+
178
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
179
+
180
+ **BibTeX:**
181
+
182
+ [More Information Needed]
183
+
184
+ **APA:**
185
+
186
+ [More Information Needed]
187
+
188
+ ## Glossary [optional]
189
+
190
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
191
+
192
+ [More Information Needed]
193
+
194
+ ## More Information [optional]
195
+
196
+ [More Information Needed]
197
+
198
+ ## Model Card Authors [optional]
199
+
200
+ [More Information Needed]
201
+
202
+ ## Model Card Contact
203
+
204
+ [More Information Needed]
205
+ ### Framework versions
206
+
207
+ - PEFT 0.16.0
adapter_config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "gpt2",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": true,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 32,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.05,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "qalora_group_size": 16,
24
+ "r": 8,
25
+ "rank_pattern": {},
26
+ "revision": null,
27
+ "target_modules": [
28
+ "c_attn",
29
+ "c_proj"
30
+ ],
31
+ "task_type": "CAUSAL_LM",
32
+ "trainable_token_indices": null,
33
+ "use_dora": false,
34
+ "use_qalora": false,
35
+ "use_rslora": false
36
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8200861e36292130ed7c83c78864188927bf97fa70cec9dea941e226b342304e
3
+ size 3253104
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json CHANGED
@@ -1,125 +1,6 @@
1
  {
2
- "additional_special_tokens": [
3
- "<extra_id_0>",
4
- "<extra_id_1>",
5
- "<extra_id_2>",
6
- "<extra_id_3>",
7
- "<extra_id_4>",
8
- "<extra_id_5>",
9
- "<extra_id_6>",
10
- "<extra_id_7>",
11
- "<extra_id_8>",
12
- "<extra_id_9>",
13
- "<extra_id_10>",
14
- "<extra_id_11>",
15
- "<extra_id_12>",
16
- "<extra_id_13>",
17
- "<extra_id_14>",
18
- "<extra_id_15>",
19
- "<extra_id_16>",
20
- "<extra_id_17>",
21
- "<extra_id_18>",
22
- "<extra_id_19>",
23
- "<extra_id_20>",
24
- "<extra_id_21>",
25
- "<extra_id_22>",
26
- "<extra_id_23>",
27
- "<extra_id_24>",
28
- "<extra_id_25>",
29
- "<extra_id_26>",
30
- "<extra_id_27>",
31
- "<extra_id_28>",
32
- "<extra_id_29>",
33
- "<extra_id_30>",
34
- "<extra_id_31>",
35
- "<extra_id_32>",
36
- "<extra_id_33>",
37
- "<extra_id_34>",
38
- "<extra_id_35>",
39
- "<extra_id_36>",
40
- "<extra_id_37>",
41
- "<extra_id_38>",
42
- "<extra_id_39>",
43
- "<extra_id_40>",
44
- "<extra_id_41>",
45
- "<extra_id_42>",
46
- "<extra_id_43>",
47
- "<extra_id_44>",
48
- "<extra_id_45>",
49
- "<extra_id_46>",
50
- "<extra_id_47>",
51
- "<extra_id_48>",
52
- "<extra_id_49>",
53
- "<extra_id_50>",
54
- "<extra_id_51>",
55
- "<extra_id_52>",
56
- "<extra_id_53>",
57
- "<extra_id_54>",
58
- "<extra_id_55>",
59
- "<extra_id_56>",
60
- "<extra_id_57>",
61
- "<extra_id_58>",
62
- "<extra_id_59>",
63
- "<extra_id_60>",
64
- "<extra_id_61>",
65
- "<extra_id_62>",
66
- "<extra_id_63>",
67
- "<extra_id_64>",
68
- "<extra_id_65>",
69
- "<extra_id_66>",
70
- "<extra_id_67>",
71
- "<extra_id_68>",
72
- "<extra_id_69>",
73
- "<extra_id_70>",
74
- "<extra_id_71>",
75
- "<extra_id_72>",
76
- "<extra_id_73>",
77
- "<extra_id_74>",
78
- "<extra_id_75>",
79
- "<extra_id_76>",
80
- "<extra_id_77>",
81
- "<extra_id_78>",
82
- "<extra_id_79>",
83
- "<extra_id_80>",
84
- "<extra_id_81>",
85
- "<extra_id_82>",
86
- "<extra_id_83>",
87
- "<extra_id_84>",
88
- "<extra_id_85>",
89
- "<extra_id_86>",
90
- "<extra_id_87>",
91
- "<extra_id_88>",
92
- "<extra_id_89>",
93
- "<extra_id_90>",
94
- "<extra_id_91>",
95
- "<extra_id_92>",
96
- "<extra_id_93>",
97
- "<extra_id_94>",
98
- "<extra_id_95>",
99
- "<extra_id_96>",
100
- "<extra_id_97>",
101
- "<extra_id_98>",
102
- "<extra_id_99>"
103
- ],
104
- "eos_token": {
105
- "content": "</s>",
106
- "lstrip": false,
107
- "normalized": false,
108
- "rstrip": false,
109
- "single_word": false
110
- },
111
- "pad_token": {
112
- "content": "<pad>",
113
- "lstrip": false,
114
- "normalized": false,
115
- "rstrip": false,
116
- "single_word": false
117
- },
118
- "unk_token": {
119
- "content": "<unk>",
120
- "lstrip": false,
121
- "normalized": false,
122
- "rstrip": false,
123
- "single_word": false
124
- }
125
  }
 
1
  {
2
+ "bos_token": "<|endoftext|>",
3
+ "eos_token": "<|endoftext|>",
4
+ "pad_token": "<|endoftext|>",
5
+ "unk_token": "<|endoftext|>"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6
  }
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1,940 +1,21 @@
1
  {
2
- "add_prefix_space": null,
3
  "added_tokens_decoder": {
4
- "0": {
5
- "content": "<pad>",
6
  "lstrip": false,
7
- "normalized": false,
8
- "rstrip": false,
9
- "single_word": false,
10
- "special": true
11
- },
12
- "1": {
13
- "content": "</s>",
14
- "lstrip": false,
15
- "normalized": false,
16
- "rstrip": false,
17
- "single_word": false,
18
- "special": true
19
- },
20
- "2": {
21
- "content": "<unk>",
22
- "lstrip": false,
23
- "normalized": false,
24
- "rstrip": false,
25
- "single_word": false,
26
- "special": true
27
- },
28
- "32000": {
29
- "content": "<extra_id_99>",
30
- "lstrip": false,
31
- "normalized": false,
32
- "rstrip": false,
33
- "single_word": false,
34
- "special": true
35
- },
36
- "32001": {
37
- "content": "<extra_id_98>",
38
- "lstrip": false,
39
- "normalized": false,
40
- "rstrip": false,
41
- "single_word": false,
42
- "special": true
43
- },
44
- "32002": {
45
- "content": "<extra_id_97>",
46
- "lstrip": false,
47
- "normalized": false,
48
- "rstrip": false,
49
- "single_word": false,
50
- "special": true
51
- },
52
- "32003": {
53
- "content": "<extra_id_96>",
54
- "lstrip": false,
55
- "normalized": false,
56
- "rstrip": false,
57
- "single_word": false,
58
- "special": true
59
- },
60
- "32004": {
61
- "content": "<extra_id_95>",
62
- "lstrip": false,
63
- "normalized": false,
64
- "rstrip": false,
65
- "single_word": false,
66
- "special": true
67
- },
68
- "32005": {
69
- "content": "<extra_id_94>",
70
- "lstrip": false,
71
- "normalized": false,
72
- "rstrip": false,
73
- "single_word": false,
74
- "special": true
75
- },
76
- "32006": {
77
- "content": "<extra_id_93>",
78
- "lstrip": false,
79
- "normalized": false,
80
- "rstrip": false,
81
- "single_word": false,
82
- "special": true
83
- },
84
- "32007": {
85
- "content": "<extra_id_92>",
86
- "lstrip": false,
87
- "normalized": false,
88
- "rstrip": false,
89
- "single_word": false,
90
- "special": true
91
- },
92
- "32008": {
93
- "content": "<extra_id_91>",
94
- "lstrip": false,
95
- "normalized": false,
96
- "rstrip": false,
97
- "single_word": false,
98
- "special": true
99
- },
100
- "32009": {
101
- "content": "<extra_id_90>",
102
- "lstrip": false,
103
- "normalized": false,
104
- "rstrip": false,
105
- "single_word": false,
106
- "special": true
107
- },
108
- "32010": {
109
- "content": "<extra_id_89>",
110
- "lstrip": false,
111
- "normalized": false,
112
- "rstrip": false,
113
- "single_word": false,
114
- "special": true
115
- },
116
- "32011": {
117
- "content": "<extra_id_88>",
118
- "lstrip": false,
119
- "normalized": false,
120
- "rstrip": false,
121
- "single_word": false,
122
- "special": true
123
- },
124
- "32012": {
125
- "content": "<extra_id_87>",
126
- "lstrip": false,
127
- "normalized": false,
128
- "rstrip": false,
129
- "single_word": false,
130
- "special": true
131
- },
132
- "32013": {
133
- "content": "<extra_id_86>",
134
- "lstrip": false,
135
- "normalized": false,
136
- "rstrip": false,
137
- "single_word": false,
138
- "special": true
139
- },
140
- "32014": {
141
- "content": "<extra_id_85>",
142
- "lstrip": false,
143
- "normalized": false,
144
- "rstrip": false,
145
- "single_word": false,
146
- "special": true
147
- },
148
- "32015": {
149
- "content": "<extra_id_84>",
150
- "lstrip": false,
151
- "normalized": false,
152
- "rstrip": false,
153
- "single_word": false,
154
- "special": true
155
- },
156
- "32016": {
157
- "content": "<extra_id_83>",
158
- "lstrip": false,
159
- "normalized": false,
160
- "rstrip": false,
161
- "single_word": false,
162
- "special": true
163
- },
164
- "32017": {
165
- "content": "<extra_id_82>",
166
- "lstrip": false,
167
- "normalized": false,
168
- "rstrip": false,
169
- "single_word": false,
170
- "special": true
171
- },
172
- "32018": {
173
- "content": "<extra_id_81>",
174
- "lstrip": false,
175
- "normalized": false,
176
- "rstrip": false,
177
- "single_word": false,
178
- "special": true
179
- },
180
- "32019": {
181
- "content": "<extra_id_80>",
182
- "lstrip": false,
183
- "normalized": false,
184
- "rstrip": false,
185
- "single_word": false,
186
- "special": true
187
- },
188
- "32020": {
189
- "content": "<extra_id_79>",
190
- "lstrip": false,
191
- "normalized": false,
192
- "rstrip": false,
193
- "single_word": false,
194
- "special": true
195
- },
196
- "32021": {
197
- "content": "<extra_id_78>",
198
- "lstrip": false,
199
- "normalized": false,
200
- "rstrip": false,
201
- "single_word": false,
202
- "special": true
203
- },
204
- "32022": {
205
- "content": "<extra_id_77>",
206
- "lstrip": false,
207
- "normalized": false,
208
- "rstrip": false,
209
- "single_word": false,
210
- "special": true
211
- },
212
- "32023": {
213
- "content": "<extra_id_76>",
214
- "lstrip": false,
215
- "normalized": false,
216
- "rstrip": false,
217
- "single_word": false,
218
- "special": true
219
- },
220
- "32024": {
221
- "content": "<extra_id_75>",
222
- "lstrip": false,
223
- "normalized": false,
224
- "rstrip": false,
225
- "single_word": false,
226
- "special": true
227
- },
228
- "32025": {
229
- "content": "<extra_id_74>",
230
- "lstrip": false,
231
- "normalized": false,
232
- "rstrip": false,
233
- "single_word": false,
234
- "special": true
235
- },
236
- "32026": {
237
- "content": "<extra_id_73>",
238
- "lstrip": false,
239
- "normalized": false,
240
- "rstrip": false,
241
- "single_word": false,
242
- "special": true
243
- },
244
- "32027": {
245
- "content": "<extra_id_72>",
246
- "lstrip": false,
247
- "normalized": false,
248
- "rstrip": false,
249
- "single_word": false,
250
- "special": true
251
- },
252
- "32028": {
253
- "content": "<extra_id_71>",
254
- "lstrip": false,
255
- "normalized": false,
256
- "rstrip": false,
257
- "single_word": false,
258
- "special": true
259
- },
260
- "32029": {
261
- "content": "<extra_id_70>",
262
- "lstrip": false,
263
- "normalized": false,
264
- "rstrip": false,
265
- "single_word": false,
266
- "special": true
267
- },
268
- "32030": {
269
- "content": "<extra_id_69>",
270
- "lstrip": false,
271
- "normalized": false,
272
- "rstrip": false,
273
- "single_word": false,
274
- "special": true
275
- },
276
- "32031": {
277
- "content": "<extra_id_68>",
278
- "lstrip": false,
279
- "normalized": false,
280
- "rstrip": false,
281
- "single_word": false,
282
- "special": true
283
- },
284
- "32032": {
285
- "content": "<extra_id_67>",
286
- "lstrip": false,
287
- "normalized": false,
288
- "rstrip": false,
289
- "single_word": false,
290
- "special": true
291
- },
292
- "32033": {
293
- "content": "<extra_id_66>",
294
- "lstrip": false,
295
- "normalized": false,
296
- "rstrip": false,
297
- "single_word": false,
298
- "special": true
299
- },
300
- "32034": {
301
- "content": "<extra_id_65>",
302
- "lstrip": false,
303
- "normalized": false,
304
- "rstrip": false,
305
- "single_word": false,
306
- "special": true
307
- },
308
- "32035": {
309
- "content": "<extra_id_64>",
310
- "lstrip": false,
311
- "normalized": false,
312
- "rstrip": false,
313
- "single_word": false,
314
- "special": true
315
- },
316
- "32036": {
317
- "content": "<extra_id_63>",
318
- "lstrip": false,
319
- "normalized": false,
320
- "rstrip": false,
321
- "single_word": false,
322
- "special": true
323
- },
324
- "32037": {
325
- "content": "<extra_id_62>",
326
- "lstrip": false,
327
- "normalized": false,
328
- "rstrip": false,
329
- "single_word": false,
330
- "special": true
331
- },
332
- "32038": {
333
- "content": "<extra_id_61>",
334
- "lstrip": false,
335
- "normalized": false,
336
- "rstrip": false,
337
- "single_word": false,
338
- "special": true
339
- },
340
- "32039": {
341
- "content": "<extra_id_60>",
342
- "lstrip": false,
343
- "normalized": false,
344
- "rstrip": false,
345
- "single_word": false,
346
- "special": true
347
- },
348
- "32040": {
349
- "content": "<extra_id_59>",
350
- "lstrip": false,
351
- "normalized": false,
352
- "rstrip": false,
353
- "single_word": false,
354
- "special": true
355
- },
356
- "32041": {
357
- "content": "<extra_id_58>",
358
- "lstrip": false,
359
- "normalized": false,
360
- "rstrip": false,
361
- "single_word": false,
362
- "special": true
363
- },
364
- "32042": {
365
- "content": "<extra_id_57>",
366
- "lstrip": false,
367
- "normalized": false,
368
- "rstrip": false,
369
- "single_word": false,
370
- "special": true
371
- },
372
- "32043": {
373
- "content": "<extra_id_56>",
374
- "lstrip": false,
375
- "normalized": false,
376
- "rstrip": false,
377
- "single_word": false,
378
- "special": true
379
- },
380
- "32044": {
381
- "content": "<extra_id_55>",
382
- "lstrip": false,
383
- "normalized": false,
384
- "rstrip": false,
385
- "single_word": false,
386
- "special": true
387
- },
388
- "32045": {
389
- "content": "<extra_id_54>",
390
- "lstrip": false,
391
- "normalized": false,
392
- "rstrip": false,
393
- "single_word": false,
394
- "special": true
395
- },
396
- "32046": {
397
- "content": "<extra_id_53>",
398
- "lstrip": false,
399
- "normalized": false,
400
- "rstrip": false,
401
- "single_word": false,
402
- "special": true
403
- },
404
- "32047": {
405
- "content": "<extra_id_52>",
406
- "lstrip": false,
407
- "normalized": false,
408
- "rstrip": false,
409
- "single_word": false,
410
- "special": true
411
- },
412
- "32048": {
413
- "content": "<extra_id_51>",
414
- "lstrip": false,
415
- "normalized": false,
416
- "rstrip": false,
417
- "single_word": false,
418
- "special": true
419
- },
420
- "32049": {
421
- "content": "<extra_id_50>",
422
- "lstrip": false,
423
- "normalized": false,
424
- "rstrip": false,
425
- "single_word": false,
426
- "special": true
427
- },
428
- "32050": {
429
- "content": "<extra_id_49>",
430
- "lstrip": false,
431
- "normalized": false,
432
- "rstrip": false,
433
- "single_word": false,
434
- "special": true
435
- },
436
- "32051": {
437
- "content": "<extra_id_48>",
438
- "lstrip": false,
439
- "normalized": false,
440
- "rstrip": false,
441
- "single_word": false,
442
- "special": true
443
- },
444
- "32052": {
445
- "content": "<extra_id_47>",
446
- "lstrip": false,
447
- "normalized": false,
448
- "rstrip": false,
449
- "single_word": false,
450
- "special": true
451
- },
452
- "32053": {
453
- "content": "<extra_id_46>",
454
- "lstrip": false,
455
- "normalized": false,
456
- "rstrip": false,
457
- "single_word": false,
458
- "special": true
459
- },
460
- "32054": {
461
- "content": "<extra_id_45>",
462
- "lstrip": false,
463
- "normalized": false,
464
- "rstrip": false,
465
- "single_word": false,
466
- "special": true
467
- },
468
- "32055": {
469
- "content": "<extra_id_44>",
470
- "lstrip": false,
471
- "normalized": false,
472
- "rstrip": false,
473
- "single_word": false,
474
- "special": true
475
- },
476
- "32056": {
477
- "content": "<extra_id_43>",
478
- "lstrip": false,
479
- "normalized": false,
480
- "rstrip": false,
481
- "single_word": false,
482
- "special": true
483
- },
484
- "32057": {
485
- "content": "<extra_id_42>",
486
- "lstrip": false,
487
- "normalized": false,
488
- "rstrip": false,
489
- "single_word": false,
490
- "special": true
491
- },
492
- "32058": {
493
- "content": "<extra_id_41>",
494
- "lstrip": false,
495
- "normalized": false,
496
- "rstrip": false,
497
- "single_word": false,
498
- "special": true
499
- },
500
- "32059": {
501
- "content": "<extra_id_40>",
502
- "lstrip": false,
503
- "normalized": false,
504
- "rstrip": false,
505
- "single_word": false,
506
- "special": true
507
- },
508
- "32060": {
509
- "content": "<extra_id_39>",
510
- "lstrip": false,
511
- "normalized": false,
512
- "rstrip": false,
513
- "single_word": false,
514
- "special": true
515
- },
516
- "32061": {
517
- "content": "<extra_id_38>",
518
- "lstrip": false,
519
- "normalized": false,
520
- "rstrip": false,
521
- "single_word": false,
522
- "special": true
523
- },
524
- "32062": {
525
- "content": "<extra_id_37>",
526
- "lstrip": false,
527
- "normalized": false,
528
- "rstrip": false,
529
- "single_word": false,
530
- "special": true
531
- },
532
- "32063": {
533
- "content": "<extra_id_36>",
534
- "lstrip": false,
535
- "normalized": false,
536
- "rstrip": false,
537
- "single_word": false,
538
- "special": true
539
- },
540
- "32064": {
541
- "content": "<extra_id_35>",
542
- "lstrip": false,
543
- "normalized": false,
544
- "rstrip": false,
545
- "single_word": false,
546
- "special": true
547
- },
548
- "32065": {
549
- "content": "<extra_id_34>",
550
- "lstrip": false,
551
- "normalized": false,
552
- "rstrip": false,
553
- "single_word": false,
554
- "special": true
555
- },
556
- "32066": {
557
- "content": "<extra_id_33>",
558
- "lstrip": false,
559
- "normalized": false,
560
- "rstrip": false,
561
- "single_word": false,
562
- "special": true
563
- },
564
- "32067": {
565
- "content": "<extra_id_32>",
566
- "lstrip": false,
567
- "normalized": false,
568
- "rstrip": false,
569
- "single_word": false,
570
- "special": true
571
- },
572
- "32068": {
573
- "content": "<extra_id_31>",
574
- "lstrip": false,
575
- "normalized": false,
576
- "rstrip": false,
577
- "single_word": false,
578
- "special": true
579
- },
580
- "32069": {
581
- "content": "<extra_id_30>",
582
- "lstrip": false,
583
- "normalized": false,
584
- "rstrip": false,
585
- "single_word": false,
586
- "special": true
587
- },
588
- "32070": {
589
- "content": "<extra_id_29>",
590
- "lstrip": false,
591
- "normalized": false,
592
- "rstrip": false,
593
- "single_word": false,
594
- "special": true
595
- },
596
- "32071": {
597
- "content": "<extra_id_28>",
598
- "lstrip": false,
599
- "normalized": false,
600
- "rstrip": false,
601
- "single_word": false,
602
- "special": true
603
- },
604
- "32072": {
605
- "content": "<extra_id_27>",
606
- "lstrip": false,
607
- "normalized": false,
608
- "rstrip": false,
609
- "single_word": false,
610
- "special": true
611
- },
612
- "32073": {
613
- "content": "<extra_id_26>",
614
- "lstrip": false,
615
- "normalized": false,
616
- "rstrip": false,
617
- "single_word": false,
618
- "special": true
619
- },
620
- "32074": {
621
- "content": "<extra_id_25>",
622
- "lstrip": false,
623
- "normalized": false,
624
- "rstrip": false,
625
- "single_word": false,
626
- "special": true
627
- },
628
- "32075": {
629
- "content": "<extra_id_24>",
630
- "lstrip": false,
631
- "normalized": false,
632
- "rstrip": false,
633
- "single_word": false,
634
- "special": true
635
- },
636
- "32076": {
637
- "content": "<extra_id_23>",
638
- "lstrip": false,
639
- "normalized": false,
640
- "rstrip": false,
641
- "single_word": false,
642
- "special": true
643
- },
644
- "32077": {
645
- "content": "<extra_id_22>",
646
- "lstrip": false,
647
- "normalized": false,
648
- "rstrip": false,
649
- "single_word": false,
650
- "special": true
651
- },
652
- "32078": {
653
- "content": "<extra_id_21>",
654
- "lstrip": false,
655
- "normalized": false,
656
- "rstrip": false,
657
- "single_word": false,
658
- "special": true
659
- },
660
- "32079": {
661
- "content": "<extra_id_20>",
662
- "lstrip": false,
663
- "normalized": false,
664
- "rstrip": false,
665
- "single_word": false,
666
- "special": true
667
- },
668
- "32080": {
669
- "content": "<extra_id_19>",
670
- "lstrip": false,
671
- "normalized": false,
672
- "rstrip": false,
673
- "single_word": false,
674
- "special": true
675
- },
676
- "32081": {
677
- "content": "<extra_id_18>",
678
- "lstrip": false,
679
- "normalized": false,
680
- "rstrip": false,
681
- "single_word": false,
682
- "special": true
683
- },
684
- "32082": {
685
- "content": "<extra_id_17>",
686
- "lstrip": false,
687
- "normalized": false,
688
- "rstrip": false,
689
- "single_word": false,
690
- "special": true
691
- },
692
- "32083": {
693
- "content": "<extra_id_16>",
694
- "lstrip": false,
695
- "normalized": false,
696
- "rstrip": false,
697
- "single_word": false,
698
- "special": true
699
- },
700
- "32084": {
701
- "content": "<extra_id_15>",
702
- "lstrip": false,
703
- "normalized": false,
704
- "rstrip": false,
705
- "single_word": false,
706
- "special": true
707
- },
708
- "32085": {
709
- "content": "<extra_id_14>",
710
- "lstrip": false,
711
- "normalized": false,
712
- "rstrip": false,
713
- "single_word": false,
714
- "special": true
715
- },
716
- "32086": {
717
- "content": "<extra_id_13>",
718
- "lstrip": false,
719
- "normalized": false,
720
- "rstrip": false,
721
- "single_word": false,
722
- "special": true
723
- },
724
- "32087": {
725
- "content": "<extra_id_12>",
726
- "lstrip": false,
727
- "normalized": false,
728
- "rstrip": false,
729
- "single_word": false,
730
- "special": true
731
- },
732
- "32088": {
733
- "content": "<extra_id_11>",
734
- "lstrip": false,
735
- "normalized": false,
736
- "rstrip": false,
737
- "single_word": false,
738
- "special": true
739
- },
740
- "32089": {
741
- "content": "<extra_id_10>",
742
- "lstrip": false,
743
- "normalized": false,
744
- "rstrip": false,
745
- "single_word": false,
746
- "special": true
747
- },
748
- "32090": {
749
- "content": "<extra_id_9>",
750
- "lstrip": false,
751
- "normalized": false,
752
- "rstrip": false,
753
- "single_word": false,
754
- "special": true
755
- },
756
- "32091": {
757
- "content": "<extra_id_8>",
758
- "lstrip": false,
759
- "normalized": false,
760
- "rstrip": false,
761
- "single_word": false,
762
- "special": true
763
- },
764
- "32092": {
765
- "content": "<extra_id_7>",
766
- "lstrip": false,
767
- "normalized": false,
768
- "rstrip": false,
769
- "single_word": false,
770
- "special": true
771
- },
772
- "32093": {
773
- "content": "<extra_id_6>",
774
- "lstrip": false,
775
- "normalized": false,
776
- "rstrip": false,
777
- "single_word": false,
778
- "special": true
779
- },
780
- "32094": {
781
- "content": "<extra_id_5>",
782
- "lstrip": false,
783
- "normalized": false,
784
- "rstrip": false,
785
- "single_word": false,
786
- "special": true
787
- },
788
- "32095": {
789
- "content": "<extra_id_4>",
790
- "lstrip": false,
791
- "normalized": false,
792
- "rstrip": false,
793
- "single_word": false,
794
- "special": true
795
- },
796
- "32096": {
797
- "content": "<extra_id_3>",
798
- "lstrip": false,
799
- "normalized": false,
800
- "rstrip": false,
801
- "single_word": false,
802
- "special": true
803
- },
804
- "32097": {
805
- "content": "<extra_id_2>",
806
- "lstrip": false,
807
- "normalized": false,
808
- "rstrip": false,
809
- "single_word": false,
810
- "special": true
811
- },
812
- "32098": {
813
- "content": "<extra_id_1>",
814
- "lstrip": false,
815
- "normalized": false,
816
- "rstrip": false,
817
- "single_word": false,
818
- "special": true
819
- },
820
- "32099": {
821
- "content": "<extra_id_0>",
822
- "lstrip": false,
823
- "normalized": false,
824
  "rstrip": false,
825
  "single_word": false,
826
  "special": true
827
  }
828
  },
829
- "additional_special_tokens": [
830
- "<extra_id_0>",
831
- "<extra_id_1>",
832
- "<extra_id_2>",
833
- "<extra_id_3>",
834
- "<extra_id_4>",
835
- "<extra_id_5>",
836
- "<extra_id_6>",
837
- "<extra_id_7>",
838
- "<extra_id_8>",
839
- "<extra_id_9>",
840
- "<extra_id_10>",
841
- "<extra_id_11>",
842
- "<extra_id_12>",
843
- "<extra_id_13>",
844
- "<extra_id_14>",
845
- "<extra_id_15>",
846
- "<extra_id_16>",
847
- "<extra_id_17>",
848
- "<extra_id_18>",
849
- "<extra_id_19>",
850
- "<extra_id_20>",
851
- "<extra_id_21>",
852
- "<extra_id_22>",
853
- "<extra_id_23>",
854
- "<extra_id_24>",
855
- "<extra_id_25>",
856
- "<extra_id_26>",
857
- "<extra_id_27>",
858
- "<extra_id_28>",
859
- "<extra_id_29>",
860
- "<extra_id_30>",
861
- "<extra_id_31>",
862
- "<extra_id_32>",
863
- "<extra_id_33>",
864
- "<extra_id_34>",
865
- "<extra_id_35>",
866
- "<extra_id_36>",
867
- "<extra_id_37>",
868
- "<extra_id_38>",
869
- "<extra_id_39>",
870
- "<extra_id_40>",
871
- "<extra_id_41>",
872
- "<extra_id_42>",
873
- "<extra_id_43>",
874
- "<extra_id_44>",
875
- "<extra_id_45>",
876
- "<extra_id_46>",
877
- "<extra_id_47>",
878
- "<extra_id_48>",
879
- "<extra_id_49>",
880
- "<extra_id_50>",
881
- "<extra_id_51>",
882
- "<extra_id_52>",
883
- "<extra_id_53>",
884
- "<extra_id_54>",
885
- "<extra_id_55>",
886
- "<extra_id_56>",
887
- "<extra_id_57>",
888
- "<extra_id_58>",
889
- "<extra_id_59>",
890
- "<extra_id_60>",
891
- "<extra_id_61>",
892
- "<extra_id_62>",
893
- "<extra_id_63>",
894
- "<extra_id_64>",
895
- "<extra_id_65>",
896
- "<extra_id_66>",
897
- "<extra_id_67>",
898
- "<extra_id_68>",
899
- "<extra_id_69>",
900
- "<extra_id_70>",
901
- "<extra_id_71>",
902
- "<extra_id_72>",
903
- "<extra_id_73>",
904
- "<extra_id_74>",
905
- "<extra_id_75>",
906
- "<extra_id_76>",
907
- "<extra_id_77>",
908
- "<extra_id_78>",
909
- "<extra_id_79>",
910
- "<extra_id_80>",
911
- "<extra_id_81>",
912
- "<extra_id_82>",
913
- "<extra_id_83>",
914
- "<extra_id_84>",
915
- "<extra_id_85>",
916
- "<extra_id_86>",
917
- "<extra_id_87>",
918
- "<extra_id_88>",
919
- "<extra_id_89>",
920
- "<extra_id_90>",
921
- "<extra_id_91>",
922
- "<extra_id_92>",
923
- "<extra_id_93>",
924
- "<extra_id_94>",
925
- "<extra_id_95>",
926
- "<extra_id_96>",
927
- "<extra_id_97>",
928
- "<extra_id_98>",
929
- "<extra_id_99>"
930
- ],
931
  "clean_up_tokenization_spaces": false,
932
- "eos_token": "</s>",
933
- "extra_ids": 100,
934
  "extra_special_tokens": {},
935
- "model_max_length": 512,
936
- "pad_token": "<pad>",
937
- "sp_model_kwargs": {},
938
- "tokenizer_class": "T5Tokenizer",
939
- "unk_token": "<unk>"
940
  }
 
1
  {
2
+ "add_prefix_space": false,
3
  "added_tokens_decoder": {
4
+ "50256": {
5
+ "content": "<|endoftext|>",
6
  "lstrip": false,
7
+ "normalized": true,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8
  "rstrip": false,
9
  "single_word": false,
10
  "special": true
11
  }
12
  },
13
+ "bos_token": "<|endoftext|>",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14
  "clean_up_tokenization_spaces": false,
15
+ "eos_token": "<|endoftext|>",
 
16
  "extra_special_tokens": {},
17
+ "model_max_length": 1024,
18
+ "pad_token": "<|endoftext|>",
19
+ "tokenizer_class": "GPT2Tokenizer",
20
+ "unk_token": "<|endoftext|>"
 
21
  }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9538f56e7c1d87f8eecd637a615c4b9c1952cdae230d3bfb968f79668bcdb4b4
3
+ size 5304
vocab.json ADDED
The diff for this file is too large to render. See raw diff