Sayan01 commited on
Commit
683dbb9
·
verified ·
1 Parent(s): 2c8e580

Upload tokenizer

Browse files
added_tokens.json CHANGED
@@ -36,5 +36,6 @@
36
  " ": 50260,
37
  " ": 50259,
38
  " ": 50258,
39
- " ": 50257
 
40
  }
 
36
  " ": 50260,
37
  " ": 50259,
38
  " ": 50258,
39
+ " ": 50257,
40
+ "[PAD]": 50295
41
  }
special_tokens_map.json CHANGED
@@ -13,6 +13,13 @@
13
  "rstrip": false,
14
  "single_word": false
15
  },
 
 
 
 
 
 
 
16
  "unk_token": {
17
  "content": "<|endoftext|>",
18
  "lstrip": false,
 
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
+ "pad_token": {
17
+ "content": "[PAD]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
  "unk_token": {
24
  "content": "<|endoftext|>",
25
  "lstrip": false,
tokenizer.json CHANGED
@@ -353,6 +353,15 @@
353
  "rstrip": false,
354
  "normalized": true,
355
  "special": false
 
 
 
 
 
 
 
 
 
356
  }
357
  ],
358
  "normalizer": null,
 
353
  "rstrip": false,
354
  "normalized": true,
355
  "special": false
356
+ },
357
+ {
358
+ "id": 50295,
359
+ "content": "[PAD]",
360
+ "single_word": false,
361
+ "lstrip": false,
362
+ "rstrip": false,
363
+ "normalized": false,
364
+ "special": true
365
  }
366
  ],
367
  "normalizer": null,
tokenizer_config.json CHANGED
@@ -312,12 +312,21 @@
312
  "rstrip": false,
313
  "single_word": false,
314
  "special": false
 
 
 
 
 
 
 
 
315
  }
316
  },
317
  "bos_token": "<|endoftext|>",
318
  "clean_up_tokenization_spaces": true,
319
  "eos_token": "<|endoftext|>",
320
  "model_max_length": 2048,
 
321
  "tokenizer_class": "CodeGenTokenizer",
322
  "unk_token": "<|endoftext|>"
323
  }
 
312
  "rstrip": false,
313
  "single_word": false,
314
  "special": false
315
+ },
316
+ "50295": {
317
+ "content": "[PAD]",
318
+ "lstrip": false,
319
+ "normalized": false,
320
+ "rstrip": false,
321
+ "single_word": false,
322
+ "special": true
323
  }
324
  },
325
  "bos_token": "<|endoftext|>",
326
  "clean_up_tokenization_spaces": true,
327
  "eos_token": "<|endoftext|>",
328
  "model_max_length": 2048,
329
+ "pad_token": "[PAD]",
330
  "tokenizer_class": "CodeGenTokenizer",
331
  "unk_token": "<|endoftext|>"
332
  }