klapinski commited on
Commit
f5e83af
·
verified ·
1 Parent(s): eaeb3dc

Training in progress, step 240

Browse files
Files changed (4) hide show
  1. config.json +2 -2
  2. model.safetensors +2 -2
  3. tokenizer.json +58 -57
  4. tokenizer_config.json +0 -1
config.json CHANGED
@@ -40,7 +40,7 @@
40
  "tie_word_embeddings": true,
41
  "type_vocab_size": 2,
42
  "use_cache": true,
43
- "vocab_size": 52
44
  },
45
  "decoder_start_token_id": 2,
46
  "dtype": "float32",
@@ -82,7 +82,7 @@
82
  "tie_word_embeddings": true,
83
  "type_vocab_size": 2,
84
  "use_cache": true,
85
- "vocab_size": 52
86
  },
87
  "eos_token_id": 0,
88
  "is_encoder_decoder": true,
 
40
  "tie_word_embeddings": true,
41
  "type_vocab_size": 2,
42
  "use_cache": true,
43
+ "vocab_size": 53
44
  },
45
  "decoder_start_token_id": 2,
46
  "dtype": "float32",
 
82
  "tie_word_embeddings": true,
83
  "type_vocab_size": 2,
84
  "use_cache": true,
85
+ "vocab_size": 53
86
  },
87
  "eos_token_id": 0,
88
  "is_encoder_decoder": true,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f9a11dc519ba013d3273e7400725b08d4b102d2332e41fc6b127b67b85b1b23b
3
- size 31205552
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77a8f63649c7f6a66cfd8839e31bbfa950c6e0bf36d123948012024aa845ab9f
3
+ size 31207604
tokenizer.json CHANGED
@@ -113,42 +113,43 @@
113
  "7": 13,
114
  "8": 14,
115
  "9": 15,
116
- "10": 16,
117
- "99": 17,
118
- "98": 18,
119
- "11": 19,
120
- "97": 20,
121
- "12": 21,
122
- "96": 22,
123
- "13": 23,
124
- "95": 24,
125
- "14": 25,
126
- "15": 26,
127
- "94": 27,
128
- "93": 28,
129
- "16": 29,
130
- "17": 30,
131
- "92": 31,
132
  "18": 32,
133
- "91": 33,
134
- "90": 34,
135
- "19": 35,
136
- "20": 36,
137
- "89": 37,
138
- "21": 38,
139
- "88": 39,
140
- "87": 40,
141
- "22": 41,
142
- "23": 42,
143
- "86": 43,
144
- "85": 44,
145
- "24": 45,
146
- "25": 46,
147
- "84": 47,
148
- "83": 48,
149
- "26": 49,
150
- "27": 50,
151
- "82": 51
 
152
  },
153
  "merges": [
154
  [
@@ -156,43 +157,51 @@
156
  "0"
157
  ],
158
  [
159
- "9",
160
- "9"
161
  ],
162
  [
163
- "9",
164
- "8"
165
  ],
166
  [
167
  "1",
168
- "1"
169
  ],
170
  [
171
- "9",
172
- "7"
173
  ],
174
  [
175
  "1",
176
- "2"
177
  ],
178
  [
179
  "9",
180
- "6"
181
  ],
182
  [
183
  "1",
184
- "3"
185
  ],
186
  [
187
  "9",
188
- "5"
189
  ],
190
  [
191
- "1",
192
- "4"
 
 
 
 
193
  ],
194
  [
195
  "1",
 
 
 
 
196
  "5"
197
  ],
198
  [
@@ -205,20 +214,12 @@
205
  ],
206
  [
207
  "1",
208
- "6"
209
- ],
210
- [
211
- "1",
212
- "7"
213
  ],
214
  [
215
  "9",
216
  "2"
217
  ],
218
- [
219
- "1",
220
- "8"
221
- ],
222
  [
223
  "9",
224
  "1"
 
113
  "7": 13,
114
  "8": 14,
115
  "9": 15,
116
+ "=": 16,
117
+ "10": 17,
118
+ "11": 18,
119
+ "12": 19,
120
+ "13": 20,
121
+ "14": 21,
122
+ "15": 22,
123
+ "99": 23,
124
+ "16": 24,
125
+ "98": 25,
126
+ "97": 26,
127
+ "96": 27,
128
+ "17": 28,
129
+ "95": 29,
130
+ "94": 30,
131
+ "93": 31,
132
  "18": 32,
133
+ "92": 33,
134
+ "91": 34,
135
+ "90": 35,
136
+ "19": 36,
137
+ "20": 37,
138
+ "89": 38,
139
+ "21": 39,
140
+ "88": 40,
141
+ "87": 41,
142
+ "22": 42,
143
+ "23": 43,
144
+ "86": 44,
145
+ "85": 45,
146
+ "24": 46,
147
+ "25": 47,
148
+ "84": 48,
149
+ "83": 49,
150
+ "26": 50,
151
+ "27": 51,
152
+ "82": 52
153
  },
154
  "merges": [
155
  [
 
157
  "0"
158
  ],
159
  [
160
+ "1",
161
+ "1"
162
  ],
163
  [
164
+ "1",
165
+ "2"
166
  ],
167
  [
168
  "1",
169
+ "3"
170
  ],
171
  [
172
+ "1",
173
+ "4"
174
  ],
175
  [
176
  "1",
177
+ "5"
178
  ],
179
  [
180
  "9",
181
+ "9"
182
  ],
183
  [
184
  "1",
185
+ "6"
186
  ],
187
  [
188
  "9",
189
+ "8"
190
  ],
191
  [
192
+ "9",
193
+ "7"
194
+ ],
195
+ [
196
+ "9",
197
+ "6"
198
  ],
199
  [
200
  "1",
201
+ "7"
202
+ ],
203
+ [
204
+ "9",
205
  "5"
206
  ],
207
  [
 
214
  ],
215
  [
216
  "1",
217
+ "8"
 
 
 
 
218
  ],
219
  [
220
  "9",
221
  "2"
222
  ],
 
 
 
 
223
  [
224
  "9",
225
  "1"
tokenizer_config.json CHANGED
@@ -2,7 +2,6 @@
2
  "backend": "tokenizers",
3
  "cls_token": "[CLS]",
4
  "eos_token": "[EOS]",
5
- "is_local": false,
6
  "model_max_length": 1000000000000000019884624838656,
7
  "pad_token": "[PAD]",
8
  "tokenizer_class": "TokenizersBackend"
 
2
  "backend": "tokenizers",
3
  "cls_token": "[CLS]",
4
  "eos_token": "[EOS]",
 
5
  "model_max_length": 1000000000000000019884624838656,
6
  "pad_token": "[PAD]",
7
  "tokenizer_class": "TokenizersBackend"