ItsMeDevRoland commited on
Commit
9350fc3
·
verified ·
1 Parent(s): c9fb356

Upload model trained with Unsloth

Browse files

Upload model trained with Unsloth 2x faster

Files changed (3) hide show
  1. special_tokens_map.json +4 -33
  2. tokenizer.json +3 -8
  3. tokenizer_config.json +10 -14
special_tokens_map.json CHANGED
@@ -1,35 +1,6 @@
1
  {
2
- "additional_special_tokens": [
3
- "<unk>",
4
- "<s>",
5
- "</s>"
6
- ],
7
- "bos_token": {
8
- "content": "<s>",
9
- "lstrip": false,
10
- "normalized": false,
11
- "rstrip": false,
12
- "single_word": false
13
- },
14
- "eos_token": {
15
- "content": "</s>",
16
- "lstrip": false,
17
- "normalized": false,
18
- "rstrip": false,
19
- "single_word": false
20
- },
21
- "pad_token": {
22
- "content": "<unk>",
23
- "lstrip": false,
24
- "normalized": false,
25
- "rstrip": false,
26
- "single_word": false
27
- },
28
- "unk_token": {
29
- "content": "<unk>",
30
- "lstrip": false,
31
- "normalized": false,
32
- "rstrip": false,
33
- "single_word": false
34
- }
35
  }
 
1
  {
2
+ "bos_token": "<s>",
3
+ "eos_token": "<|im_end|>",
4
+ "pad_token": "<unk>",
5
+ "unk_token": "<unk>"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6
  }
tokenizer.json CHANGED
@@ -1,11 +1,6 @@
1
  {
2
  "version": "1.0",
3
- "truncation": {
4
- "direction": "Left",
5
- "max_length": 2048,
6
- "strategy": "LongestFirst",
7
- "stride": 0
8
- },
9
  "padding": null,
10
  "added_tokens": [
11
  {
@@ -28,7 +23,7 @@
28
  },
29
  {
30
  "id": 2,
31
- "content": "</s>",
32
  "single_word": false,
33
  "lstrip": false,
34
  "rstrip": false,
@@ -143,7 +138,7 @@
143
  "vocab": {
144
  "<unk>": 0,
145
  "<s>": 1,
146
- "</s>": 2,
147
  "<0x00>": 3,
148
  "<0x01>": 4,
149
  "<0x02>": 5,
 
1
  {
2
  "version": "1.0",
3
+ "truncation": null,
 
 
 
 
 
4
  "padding": null,
5
  "added_tokens": [
6
  {
 
23
  },
24
  {
25
  "id": 2,
26
+ "content": "<|im_end|>",
27
  "single_word": false,
28
  "lstrip": false,
29
  "rstrip": false,
 
138
  "vocab": {
139
  "<unk>": 0,
140
  "<s>": 1,
141
+ "<|im_end|>": 2,
142
  "<0x00>": 3,
143
  "<0x01>": 4,
144
  "<0x02>": 5,
tokenizer_config.json CHANGED
@@ -20,7 +20,7 @@
20
  "special": true
21
  },
22
  "2": {
23
- "content": "</s>",
24
  "lstrip": false,
25
  "normalized": false,
26
  "rstrip": false,
@@ -28,23 +28,19 @@
28
  "special": true
29
  }
30
  },
31
- "additional_special_tokens": [
32
- "<unk>",
33
- "<s>",
34
- "</s>"
35
- ],
36
  "bos_token": "<s>",
 
37
  "clean_up_tokenization_spaces": false,
38
- "eos_token": "</s>",
39
  "extra_special_tokens": {},
40
  "legacy": true,
41
- "model_max_length": 32768,
 
42
  "pad_token": "<unk>",
43
- "padding_side": "right",
44
- "sp_model_kwargs": {},
45
- "spaces_between_special_tokens": false,
46
- "tokenizer_class": "LlamaTokenizer",
47
- "truncation_side": "left",
48
  "unk_token": "<unk>",
49
- "use_default_system_prompt": true
50
  }
 
20
  "special": true
21
  },
22
  "2": {
23
+ "content": "<|im_end|>",
24
  "lstrip": false,
25
  "normalized": false,
26
  "rstrip": false,
 
28
  "special": true
29
  }
30
  },
 
 
 
 
 
31
  "bos_token": "<s>",
32
+ "chat_template": "{% if 'role' in messages[0] %}{% for message in messages %}{% if message['role'] == 'user' %}{{'<|im_start|>user\n' + message['content'] + '<|im_end|>\n'}}{% elif message['role'] == 'assistant' %}{{'<|im_start|>assistant\n' + message['content'] + '<|im_end|>\n' }}{% else %}{{ '<|im_start|>system\n' + message['content'] + '<|im_end|>\n' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}{% else %}{% for message in messages %}{% if message['from'] == 'human' %}{{'<|im_start|>user\n' + message['value'] + '<|im_end|>\n'}}{% elif message['from'] == 'gpt' %}{{'<|im_start|>assistant\n' + message['value'] + '<|im_end|>\n' }}{% else %}{{ '<|im_start|>system\n' + message['value'] + '<|im_end|>\n' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}{% endif %}",
33
  "clean_up_tokenization_spaces": false,
34
+ "eos_token": "<|im_end|>",
35
  "extra_special_tokens": {},
36
  "legacy": true,
37
+ "max_length": 1024,
38
+ "model_max_length": 1000000000000000019884624838656,
39
  "pad_token": "<unk>",
40
+ "stride": 0,
41
+ "tokenizer_class": "LlamaTokenizerFast",
42
+ "truncation_side": "right",
43
+ "truncation_strategy": "longest_first",
 
44
  "unk_token": "<unk>",
45
+ "use_default_system_prompt": false
46
  }