codebyzeb commited on
Commit
d6aebef
·
verified ·
1 Parent(s): b620488

Delete frequency_16000/fw57M_Entropy_frequency_16000

Browse files
frequency_16000/fw57M_Entropy_frequency_16000/merges.txt DELETED
The diff for this file is too large to render. See raw diff
 
frequency_16000/fw57M_Entropy_frequency_16000/merges_data.csv DELETED
The diff for this file is too large to render. See raw diff
 
frequency_16000/fw57M_Entropy_frequency_16000/special_tokens_map.json DELETED
@@ -1,4 +0,0 @@
1
- {
2
- "eos_token": "<|endoftext|>",
3
- "pad_token": "<|padding|>"
4
- }
 
 
 
 
 
frequency_16000/fw57M_Entropy_frequency_16000/tokenizer.json DELETED
The diff for this file is too large to render. See raw diff
 
frequency_16000/fw57M_Entropy_frequency_16000/tokenizer_config.json DELETED
@@ -1,29 +0,0 @@
1
- {
2
- "add_prefix_space": true,
3
- "added_tokens_decoder": {
4
- "0": {
5
- "content": "<|padding|>",
6
- "lstrip": false,
7
- "normalized": false,
8
- "rstrip": false,
9
- "single_word": false,
10
- "special": true
11
- },
12
- "1": {
13
- "content": "<|endoftext|>",
14
- "lstrip": false,
15
- "normalized": false,
16
- "rstrip": false,
17
- "single_word": false,
18
- "special": true
19
- }
20
- },
21
- "bos_token": null,
22
- "clean_up_tokenization_spaces": false,
23
- "eos_token": "<|endoftext|>",
24
- "extra_special_tokens": {},
25
- "model_max_length": 1000000000000000019884624838656,
26
- "pad_token": "<|padding|>",
27
- "tokenizer_class": "PreTrainedTokenizer",
28
- "unk_token": null
29
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
frequency_16000/fw57M_Entropy_frequency_16000/vocab.json DELETED
The diff for this file is too large to render. See raw diff