josem7 commited on
Commit
faf9faf
·
1 Parent(s): 2a0e120

Upload tokenizer_config.json with huggingface_hub

Browse files
Files changed (1) hide show
  1. tokenizer_config.json +6 -62
tokenizer_config.json CHANGED
@@ -24,39 +24,7 @@
24
  "single_word": false,
25
  "special": true
26
  },
27
- "32007": {
28
- "content": "▁<PRE>",
29
- "lstrip": true,
30
- "normalized": false,
31
- "rstrip": true,
32
- "single_word": false,
33
- "special": true
34
- },
35
- "32008": {
36
- "content": "▁<SUF>",
37
- "lstrip": true,
38
- "normalized": false,
39
- "rstrip": true,
40
- "single_word": false,
41
- "special": true
42
- },
43
- "32009": {
44
- "content": "▁<MID>",
45
- "lstrip": true,
46
- "normalized": false,
47
- "rstrip": true,
48
- "single_word": false,
49
- "special": true
50
- },
51
- "32010": {
52
- "content": "▁<EOT>",
53
- "lstrip": true,
54
- "normalized": false,
55
- "rstrip": true,
56
- "single_word": false,
57
- "special": true
58
- },
59
- "32016": {
60
  "content": "<PAD>",
61
  "lstrip": false,
62
  "normalized": false,
@@ -66,48 +34,24 @@
66
  }
67
  },
68
  "additional_special_tokens": [
69
- "▁<PRE>",
70
- "▁<SUF>",
71
- "▁<MID>",
72
- "▁<EOT>",
73
- "▁<PRE>",
74
- "▁<MID>",
75
- "▁<SUF>",
76
- "▁<EOT>",
77
  "<unk>",
78
  "<s>",
79
  "</s>",
80
- "<PRE>",
81
- "▁<SUF>",
82
- "▁<MID>",
83
- "▁<EOT>",
84
- "<PAD>",
85
- "▁<PRE>",
86
- "▁<MID>",
87
- "▁<SUF>",
88
- "▁<EOT>",
89
- "▁<PRE>",
90
- "▁<MID>",
91
- "▁<SUF>",
92
- "▁<EOT>"
93
  ],
94
  "bos_token": "<s>",
95
  "clean_up_tokenization_spaces": false,
96
  "eos_token": "</s>",
97
- "eot_token": "▁<EOT>",
98
- "fill_token": "<FILL_ME>",
99
- "legacy": null,
100
  "max_length": 4096,
101
- "middle_token": "▁<MID>",
102
  "model_max_length": 1000000000000000019884624838656,
103
  "pad_token": "<PAD>",
104
- "prefix_token": "▁<PRE>",
105
  "sp_model_kwargs": {},
106
  "stride": 0,
107
- "suffix_token": "▁<SUF>",
108
- "tokenizer_class": "CodeLlamaTokenizer",
109
  "truncation_side": "right",
110
  "truncation_strategy": "longest_first",
111
  "unk_token": "<unk>",
112
- "use_default_system_prompt": false
113
  }
 
24
  "single_word": false,
25
  "special": true
26
  },
27
+ "32000": {
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
28
  "content": "<PAD>",
29
  "lstrip": false,
30
  "normalized": false,
 
34
  }
35
  },
36
  "additional_special_tokens": [
 
 
 
 
 
 
 
 
37
  "<unk>",
38
  "<s>",
39
  "</s>",
40
+ "<PAD>"
 
 
 
 
 
 
 
 
 
 
 
 
41
  ],
42
  "bos_token": "<s>",
43
  "clean_up_tokenization_spaces": false,
44
  "eos_token": "</s>",
45
+ "legacy": false,
 
 
46
  "max_length": 4096,
 
47
  "model_max_length": 1000000000000000019884624838656,
48
  "pad_token": "<PAD>",
49
+ "padding_side": "right",
50
  "sp_model_kwargs": {},
51
  "stride": 0,
52
+ "tokenizer_class": "LlamaTokenizer",
 
53
  "truncation_side": "right",
54
  "truncation_strategy": "longest_first",
55
  "unk_token": "<unk>",
56
+ "use_default_system_prompt": true
57
  }