Sayan01 commited on
Commit
5bca0ab
·
verified ·
1 Parent(s): bfd7cba

Upload tokenizer

Browse files
added_tokens.json CHANGED
@@ -36,6 +36,5 @@
36
  " ": 50260,
37
  " ": 50259,
38
  " ": 50258,
39
- " ": 50257,
40
- "[PAD]": 50295
41
  }
 
36
  " ": 50260,
37
  " ": 50259,
38
  " ": 50258,
39
+ " ": 50257
 
40
  }
special_tokens_map.json CHANGED
@@ -13,13 +13,6 @@
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
- "pad_token": {
17
- "content": "[PAD]",
18
- "lstrip": false,
19
- "normalized": false,
20
- "rstrip": false,
21
- "single_word": false
22
- },
23
  "unk_token": {
24
  "content": "<|endoftext|>",
25
  "lstrip": false,
 
13
  "rstrip": false,
14
  "single_word": false
15
  },
 
 
 
 
 
 
 
16
  "unk_token": {
17
  "content": "<|endoftext|>",
18
  "lstrip": false,
tokenizer.json CHANGED
@@ -353,15 +353,6 @@
353
  "rstrip": false,
354
  "normalized": true,
355
  "special": false
356
- },
357
- {
358
- "id": 50295,
359
- "content": "[PAD]",
360
- "single_word": false,
361
- "lstrip": false,
362
- "rstrip": false,
363
- "normalized": false,
364
- "special": true
365
  }
366
  ],
367
  "normalizer": null,
 
353
  "rstrip": false,
354
  "normalized": true,
355
  "special": false
 
 
 
 
 
 
 
 
 
356
  }
357
  ],
358
  "normalizer": null,
tokenizer_config.json CHANGED
@@ -312,21 +312,12 @@
312
  "rstrip": false,
313
  "single_word": false,
314
  "special": false
315
- },
316
- "50295": {
317
- "content": "[PAD]",
318
- "lstrip": false,
319
- "normalized": false,
320
- "rstrip": false,
321
- "single_word": false,
322
- "special": true
323
  }
324
  },
325
  "bos_token": "<|endoftext|>",
326
  "clean_up_tokenization_spaces": true,
327
  "eos_token": "<|endoftext|>",
328
  "model_max_length": 2048,
329
- "pad_token": "[PAD]",
330
  "tokenizer_class": "CodeGenTokenizer",
331
  "unk_token": "<|endoftext|>"
332
  }
 
312
  "rstrip": false,
313
  "single_word": false,
314
  "special": false
 
 
 
 
 
 
 
 
315
  }
316
  },
317
  "bos_token": "<|endoftext|>",
318
  "clean_up_tokenization_spaces": true,
319
  "eos_token": "<|endoftext|>",
320
  "model_max_length": 2048,
 
321
  "tokenizer_class": "CodeGenTokenizer",
322
  "unk_token": "<|endoftext|>"
323
  }