ConicCat commited on
Commit
6687ede
·
verified ·
1 Parent(s): 4774e65

Upload tokenizer

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
chat_template.jinja ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {% if messages[0]['role'] == 'system' %}
2
+ {% set loop_messages = messages[1:] %}
3
+ {% set system_message = messages[0]['content'].strip() + '
4
+ ' %}
5
+ {% else %}
6
+ {% set loop_messages = messages %}
7
+ {% set system_message = '' %}
8
+ {% endif %}
9
+ {% for message in loop_messages %}
10
+ {% if loop.index0 == 0 %}
11
+ {% set content = system_message + message['content'] %}
12
+ {% else %}
13
+ {% set content = message['content'] %}
14
+ {% endif %}
15
+ {% if message['role'] == 'user' %}
16
+ {{ '### Instruction:
17
+ ' + content.strip() + '
18
+
19
+ '}}
20
+ {% elif message['role'] == 'assistant' %}
21
+ {% generation %}
22
+ {{ '### Response:
23
+ ' + content.strip() + '
24
+
25
+ ' }}
26
+ {% endgeneration %}
27
+ {% endif %}
28
+ {% endfor %}
29
+ {% if add_generation_prompt %}
30
+ {{ '### Response:
31
+ ' }}
32
+ {% endif %}
special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<pad>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<unk>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bb201fb226cde11f66c3cf51c5344fb37b1611f00c21e75c324546d854eff2e1
3
+ size 17078480
tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff