luisra commited on
Commit
91129c8
·
verified ·
1 Parent(s): fc2535e

Add files using upload-large-folder tool

Browse files
Files changed (44) hide show
  1. README.md +27 -7
  2. chat_template.jinja +43 -0
  3. config.json +6 -4
  4. generation_config.json +1 -1
  5. model-00001-of-00107.safetensors +3 -0
  6. model-00003-of-00107.safetensors +3 -0
  7. model-00007-of-00107.safetensors +3 -0
  8. model-00008-of-00107.safetensors +3 -0
  9. model-00009-of-00107.safetensors +3 -0
  10. model-00011-of-00107.safetensors +3 -0
  11. model-00015-of-00107.safetensors +3 -0
  12. model-00030-of-00107.safetensors +3 -0
  13. model-00032-of-00107.safetensors +3 -0
  14. model-00033-of-00107.safetensors +3 -0
  15. model-00034-of-00107.safetensors +3 -0
  16. model-00036-of-00107.safetensors +3 -0
  17. model-00042-of-00107.safetensors +3 -0
  18. model-00044-of-00107.safetensors +3 -0
  19. model-00045-of-00107.safetensors +3 -0
  20. model-00047-of-00107.safetensors +3 -0
  21. model-00050-of-00107.safetensors +3 -0
  22. model-00051-of-00107.safetensors +3 -0
  23. model-00052-of-00107.safetensors +3 -0
  24. model-00053-of-00107.safetensors +3 -0
  25. model-00056-of-00107.safetensors +3 -0
  26. model-00058-of-00107.safetensors +3 -0
  27. model-00060-of-00107.safetensors +3 -0
  28. model-00061-of-00107.safetensors +3 -0
  29. model-00063-of-00107.safetensors +3 -0
  30. model-00067-of-00107.safetensors +3 -0
  31. model-00068-of-00107.safetensors +3 -0
  32. model-00069-of-00107.safetensors +3 -0
  33. model-00072-of-00107.safetensors +3 -0
  34. model-00073-of-00107.safetensors +3 -0
  35. model-00076-of-00107.safetensors +3 -0
  36. model-00077-of-00107.safetensors +3 -0
  37. model-00078-of-00107.safetensors +3 -0
  38. model-00083-of-00107.safetensors +3 -0
  39. model-00087-of-00107.safetensors +3 -0
  40. model-00090-of-00107.safetensors +3 -0
  41. model-00097-of-00107.safetensors +3 -0
  42. model-00098-of-00107.safetensors +3 -0
  43. special_tokens_map.json +40 -0
  44. tokenizer_config.json +11 -59
README.md CHANGED
@@ -1,8 +1,34 @@
1
  ---
 
 
 
 
2
  license: other
3
  license_name: modified-mit
4
  library_name: transformers
5
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6
  <div align="center">
7
  <picture>
8
  <img src="figures/kimi-logo.png" width="30%" alt="Kimi K2: Open Agentic Intellignece">
@@ -28,15 +54,9 @@ library_name: transformers
28
  </div>
29
 
30
  <p align="center">
31
- <b>📰&nbsp;&nbsp;<a href="https://moonshotai.github.io/Kimi-K2/">Tech Blog</a></b> &nbsp;&nbsp;&nbsp; | &nbsp;&nbsp;&nbsp; <b>📄&nbsp;&nbsp;Paper Link (coming soon)</b>
32
  </p>
33
 
34
- ## 0. Changelog
35
-
36
- ### 2025.7.15
37
- - We have updated our tokenizer implementation. Now special tokens like `[EOS]` can be encoded to their token ids.
38
- - We fixed a bug in the chat template that was breaking multi-turn tool calls.
39
-
40
  ## 1. Model Introduction
41
 
42
  Kimi K2 is a state-of-the-art mixture-of-experts (MoE) language model with 32 billion activated parameters and 1 trillion total parameters. Trained with the Muon optimizer, Kimi K2 achieves exceptional performance across frontier knowledge, reasoning, and coding tasks while being meticulously optimized for agentic capabilities.
 
1
  ---
2
+ tags:
3
+ - unsloth
4
+ base_model:
5
+ - moonshotai/Kimi-K2-Instruct
6
  license: other
7
  license_name: modified-mit
8
  library_name: transformers
9
  ---
10
+
11
+ > [!NOTE]
12
+ > Includes our **chat template fixes**! <br> For `llama.cpp`, use `--jinja`
13
+ >
14
+
15
+ <div>
16
+ <p style="margin-top: 0;margin-bottom: 0;">
17
+ <em><a href="https://docs.unsloth.ai/basics/unsloth-dynamic-v2.0-gguf">Unsloth Dynamic 2.0</a> achieves superior accuracy & outperforms other leading quants.</em>
18
+ </p>
19
+ <div style="display: flex; gap: 5px; align-items: center; ">
20
+ <a href="https://github.com/unslothai/unsloth/">
21
+ <img src="https://github.com/unslothai/unsloth/raw/main/images/unsloth%20new%20logo.png" width="133">
22
+ </a>
23
+ <a href="https://discord.gg/unsloth">
24
+ <img src="https://github.com/unslothai/unsloth/raw/main/images/Discord%20button.png" width="173">
25
+ </a>
26
+ <a href="https://docs.unsloth.ai/">
27
+ <img src="https://raw.githubusercontent.com/unslothai/unsloth/refs/heads/main/images/documentation%20green%20button.png" width="143">
28
+ </a>
29
+ </div>
30
+ </div>
31
+
32
  <div align="center">
33
  <picture>
34
  <img src="figures/kimi-logo.png" width="30%" alt="Kimi K2: Open Agentic Intellignece">
 
54
  </div>
55
 
56
  <p align="center">
57
+ <b>📰&nbsp;&nbsp;<a href="https://moonshotai.github.io/Kimi-K2/">Tech Blog</a></b> &nbsp;&nbsp;&nbsp; | &nbsp;&nbsp;&nbsp; <b>📄&nbsp;&nbsp;Paper Link (comming soon)</b>
58
  </p>
59
 
 
 
 
 
 
 
60
  ## 1. Model Introduction
61
 
62
  Kimi K2 is a state-of-the-art mixture-of-experts (MoE) language model with 32 billion activated parameters and 1 trillion total parameters. Trained with the Muon optimizer, Kimi K2 achieves exceptional performance across frontier knowledge, reasoning, and coding tasks while being meticulously optimized for agentic capabilities.
chat_template.jinja ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if tools -%}
2
+ <|im_system|>tool_declare<|im_middle|>{{ tools | tojson }}<|im_end|>
3
+ {%- endif -%}
4
+ {%- for message in messages -%}
5
+ {%- if loop.first and messages[0]['role'] != 'system' -%}
6
+ <|im_system|>system<|im_middle|>You are Kimi, an AI assistant created by Moonshot AI.<|im_end|>
7
+ {%- endif -%}
8
+ {%- if message['role'] == 'system' -%}
9
+ <|im_system|>system<|im_middle|>
10
+ {%- elif message['role'] == 'user' -%}
11
+ <|im_user|>user<|im_middle|>
12
+ {%- elif message['role'] == 'assistant' -%}
13
+ <|im_assistant|>assistant<|im_middle|>
14
+ {%- elif message['role'] == 'tool' -%}
15
+ <|im_system|>tool<|im_middle|>
16
+ {%- endif -%}
17
+ {%- if message['role'] == 'assistant' and message.get('tool_calls') -%}
18
+ {%- if message['content'] -%}{{ message['content'] }}{%- endif -%}
19
+ <|tool_calls_section_begin|>
20
+ {%- for tool_call in message['tool_calls'] -%}
21
+ {%- set formatted_id = tool_call['id'] -%}
22
+ <|tool_call_begin|>{{ formatted_id }}<|tool_call_argument_begin|>{% if tool_call['function']['arguments'] is string %}{{ tool_call['function']['arguments'] }}{% else %}{{ tool_call['function']['arguments'] | tojson }}{% endif %}<|tool_call_end|>
23
+ {%- endfor -%}
24
+ <|tool_calls_section_end|>
25
+ {%- elif message['role'] == 'tool' -%}
26
+ ## Return of {{ message.tool_call_id }}
27
+ {{ message['content'] }}
28
+ {%- elif message['content'] is string -%}
29
+ {{ message['content'] }}
30
+ {%- elif message['content'] is not none -%}
31
+ {% for content in message['content'] -%}
32
+ {% if content['type'] == 'image' or 'image' in content or 'image_url' in content -%}
33
+ <|media_start|>image<|media_content|><|media_pad|><|media_end|>
34
+ {% else -%}
35
+ {{ content['text'] }}
36
+ {%- endif -%}
37
+ {%- endfor -%}
38
+ {%- endif -%}
39
+ <|im_end|>
40
+ {%- endfor -%}
41
+ {%- if add_generation_prompt -%}
42
+ <|im_assistant|>assistant<|im_middle|>
43
+ {%- endif -%}
config.json CHANGED
@@ -11,7 +11,7 @@
11
  },
12
  "aux_loss_alpha": 0.001,
13
  "bos_token_id": 163584,
14
- "eos_token_id": 163585,
15
  "ep_size": 1,
16
  "first_k_dense_replace": 1,
17
  "hidden_act": "silu",
@@ -32,6 +32,7 @@
32
  "num_hidden_layers": 61,
33
  "num_key_value_heads": 64,
34
  "num_nextn_predict_layers": 0,
 
35
  "pretraining_tp": 1,
36
  "q_lora_rank": 1536,
37
  "qk_nope_head_dim": 128,
@@ -39,7 +40,7 @@
39
  "quantization_config": {
40
  "_load_in_4bit": true,
41
  "_load_in_8bit": false,
42
- "bnb_4bit_compute_dtype": "float16",
43
  "bnb_4bit_quant_storage": "uint8",
44
  "bnb_4bit_quant_type": "nf4",
45
  "bnb_4bit_use_double_quant": true,
@@ -68,8 +69,9 @@
68
  "tie_word_embeddings": false,
69
  "topk_group": 1,
70
  "topk_method": "noaux_tc",
71
- "torch_dtype": "float16",
72
- "transformers_version": "4.53.2",
 
73
  "use_cache": true,
74
  "v_head_dim": 128,
75
  "vocab_size": 163840
 
11
  },
12
  "aux_loss_alpha": 0.001,
13
  "bos_token_id": 163584,
14
+ "eos_token_id": 163586,
15
  "ep_size": 1,
16
  "first_k_dense_replace": 1,
17
  "hidden_act": "silu",
 
32
  "num_hidden_layers": 61,
33
  "num_key_value_heads": 64,
34
  "num_nextn_predict_layers": 0,
35
+ "pad_token_id": 163839,
36
  "pretraining_tp": 1,
37
  "q_lora_rank": 1536,
38
  "qk_nope_head_dim": 128,
 
40
  "quantization_config": {
41
  "_load_in_4bit": true,
42
  "_load_in_8bit": false,
43
+ "bnb_4bit_compute_dtype": "bfloat16",
44
  "bnb_4bit_quant_storage": "uint8",
45
  "bnb_4bit_quant_type": "nf4",
46
  "bnb_4bit_use_double_quant": true,
 
69
  "tie_word_embeddings": false,
70
  "topk_group": 1,
71
  "topk_method": "noaux_tc",
72
+ "torch_dtype": "bfloat16",
73
+ "transformers_version": "4.56.0.dev0",
74
+ "unsloth_fixed": true,
75
  "use_cache": true,
76
  "v_head_dim": 128,
77
  "vocab_size": 163840
generation_config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
  "eos_token_id": 163586,
3
  "max_length": 131072,
4
- "transformers_version": "4.53.2"
5
  }
 
1
  {
2
  "eos_token_id": 163586,
3
  "max_length": 131072,
4
+ "transformers_version": "4.56.0.dev0"
5
  }
model-00001-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97ef31c9fdd3fb0b3b124a408532ee9fc6c7b10c830fbd8f31ac69c8bc2de543
3
+ size 4998356350
model-00003-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95127ca90eb2304cef4dfac752f608b999cafc7e6309548bf5d30bdd9669eb36
3
+ size 4996628390
model-00007-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:713d4a8ed063483005e2e5b314848c846e6954ee804b299d129e8e8a93c4cc99
3
+ size 4999514397
model-00008-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea748a060ed14eb9a72d28e4c29b6e2fb4e5204b0b5ed1b9e7180f515e9e1312
3
+ size 4996628486
model-00009-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64d0b2bddf001e463e8c5a1592fb2cc132d54bfce0e9747205402d5d162d90dc
3
+ size 4999515346
model-00011-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c72b100829345d077d6ac595140ef13231033a7f69cd225ad363d77cb62c4df
3
+ size 4999515457
model-00015-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5eab419792d677454f61700591198e2f94906c8a917c1da8a7d660feee33c2d3
3
+ size 4996628676
model-00030-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6670e90b4e28e420fb00feff6c8c9de40ad2ab9a1f099843f4bbf9c15f24e96c
3
+ size 4999518733
model-00032-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:601d78a1e79ca1598be661804c499e71c13599bba7ff23cb8f89e87e2007348a
3
+ size 4999519429
model-00033-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:418b35f6d71b123c0e8900e1af92fc3878e23d356ed4eda42b514ac5e928534f
3
+ size 4996632296
model-00034-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:822c7c2b347cc6f51313b5d75d2dc25b43966255b8b00f5bfcbaf3dd660df16c
3
+ size 4997450464
model-00036-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a87df2eb9a2aba10c9ea2953c0fe7234cb5e3f9beb90bcc8a5d6ea9d235f44c
3
+ size 4996633205
model-00042-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a80d1f3f41c40dc4d34d8bba7c8c36118011733674b5054dbc792a54b5366d86
3
+ size 4996632272
model-00044-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d06bab1c417250a7cb2199f3010354383dc8412839232edd6c22992c1afaa1d
3
+ size 4999518335
model-00045-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:415ce38326fae2dfdce4f35294c046ed82fe714408e10527c074c5d4a8b20410
3
+ size 4996632463
model-00047-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:159e571424bc6d0dafff79ac67e3c3f733eb4eb40debadf302e02141ae3b5edf
3
+ size 4996632289
model-00050-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36ddb29cfb2f01757ce4c4aa714fcd2c9c23da9ed763f1cbe64dc228b8206f24
3
+ size 4996633587
model-00051-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:79859c8287e53ec505cb10bdc1283ae2c391759cb71dac89c14df72fa5e35287
3
+ size 4999518160
model-00052-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf5fc3ca21d008f4efbe8926975cf037b06433633367520ec6e5ef9fd988eaa1
3
+ size 4996632622
model-00053-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c25b7c501b380eecee8ee898db5b312266c746f4985143c9a2032f1a2e8b5d34
3
+ size 4999519066
model-00056-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9a6b14f7f9a6b9e7262e6966bc725ee17aab4e6c258733d00e0bf31584a38b6
3
+ size 4996632310
model-00058-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:81086f3824bf028cb91c5e57ea938e92f0e0dc49ca749506ea3c07f9e4ac7eb9
3
+ size 4999517950
model-00060-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3205833cbd007e28bc4301e00ba5bb9100739cda38a2db8554e7053ce6aa3810
3
+ size 4999518882
model-00061-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f512233be2241251e37c52075a14025223d47f904add8a364a513ec52ec9d40
3
+ size 4996632277
model-00063-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1de12142c0a73e928a6e6ceaba21cfc94111b4c45aebf2b02f49066708ce6c35
3
+ size 4996632292
model-00067-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f070e321702270958ae0979ccd3a1ff73f0114774ccd3eb6ea20d7ec62e60f9
3
+ size 4999518700
model-00068-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:637cbfcf1bdb1c6910e201b41016a3bb2a743a28382d979b38a0875c32788d02
3
+ size 4996632300
model-00069-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9fe8b6ef84aeddfd4f81cee763d274213293bd31df261111672126f17d688de
3
+ size 4999519433
model-00072-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed52726d211289e635360e3fb230401d93d6cddac9bd67d44de444a2a6e2edd0
3
+ size 4996632313
model-00073-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b24f25a65a7a57ac2db611edd2aeac76378883ed087092c715d7f3d2055dff8
3
+ size 4996633218
model-00076-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:414169dff92d709b4207ba2ba23ade02258581947212f81fb672625d4dfe5e57
3
+ size 4999519428
model-00077-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92ca44f0ddefe997539172d80a91edf77ffbda8570919ea0330e0524b021c7c3
3
+ size 4996632302
model-00078-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e9447b947555418f3f05c092dda22d7e0c089455fbd0f5980bb9345b53a0e57
3
+ size 4999519442
model-00083-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a63e00a1d40845a4620234002a7a4d58dbc18f7a2a746c8bec4a99ed399b90ea
3
+ size 4999519258
model-00087-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:037b61611954714480c78a68cb702aac75dcc0fabf8d2e861a1b9a11df622341
3
+ size 4996633612
model-00090-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a1a42ed00cb1c8381a2142aed043c234b297a2eccd1c23a3e36ba88acdb919b
3
+ size 4999519071
model-00097-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:251d3a5650aa0237a3a4ded14bad19b150eafc0987b81b760168c6cd2977b2f0
3
+ size 4999518889
model-00098-of-00107.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c9691cb82ca142a84ef9ee2965bd38cf8e3ea05936c50d779d560c2d269dd4a
3
+ size 4996632316
special_tokens_map.json ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_end|>",
4
+ "<|im_user|>",
5
+ "<|im_assistant|>",
6
+ "<|start_header_id|>",
7
+ "<|end_header_id|>",
8
+ "[EOT]",
9
+ "<|im_system|>",
10
+ "<|im_middle|>"
11
+ ],
12
+ "bos_token": {
13
+ "content": "[BOS]",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false
18
+ },
19
+ "eos_token": {
20
+ "content": "<|im_end|>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false
25
+ },
26
+ "pad_token": {
27
+ "content": "[PAD]",
28
+ "lstrip": false,
29
+ "normalized": false,
30
+ "rstrip": false,
31
+ "single_word": false
32
+ },
33
+ "unk_token": {
34
+ "content": "[UNK]",
35
+ "lstrip": false,
36
+ "normalized": false,
37
+ "rstrip": false,
38
+ "single_word": false
39
+ }
40
+ }
tokenizer_config.json CHANGED
@@ -8,14 +8,6 @@
8
  "single_word": false,
9
  "special": true
10
  },
11
- "163585": {
12
- "content": "[EOS]",
13
- "lstrip": false,
14
- "normalized": false,
15
- "rstrip": false,
16
- "single_word": false,
17
- "special": true
18
- },
19
  "163586": {
20
  "content": "<|im_end|>",
21
  "lstrip": false,
@@ -72,46 +64,6 @@
72
  "single_word": false,
73
  "special": true
74
  },
75
- "163595": {
76
- "content": "<|tool_calls_section_begin|>",
77
- "lstrip": false,
78
- "normalized": false,
79
- "rstrip": false,
80
- "single_word": false,
81
- "special": false
82
- },
83
- "163596": {
84
- "content": "<|tool_calls_section_end|>",
85
- "lstrip": false,
86
- "normalized": false,
87
- "rstrip": false,
88
- "single_word": false,
89
- "special": false
90
- },
91
- "163597": {
92
- "content": "<|tool_call_begin|>",
93
- "lstrip": false,
94
- "normalized": false,
95
- "rstrip": false,
96
- "single_word": false,
97
- "special": false
98
- },
99
- "163598": {
100
- "content": "<|tool_call_argument_begin|>",
101
- "lstrip": false,
102
- "normalized": false,
103
- "rstrip": false,
104
- "single_word": false,
105
- "special": false
106
- },
107
- "163599": {
108
- "content": "<|tool_call_end|>",
109
- "lstrip": false,
110
- "normalized": false,
111
- "rstrip": false,
112
- "single_word": false,
113
- "special": false
114
- },
115
  "163601": {
116
  "content": "<|im_middle|>",
117
  "lstrip": false,
@@ -147,19 +99,19 @@
147
  "<|im_system|>",
148
  "<|im_middle|>"
149
  ],
150
- "bos_token": "[BOS]",
151
- "clean_up_tokenization_spaces": false,
152
- "eos_token": "[EOS]",
153
- "extra_special_tokens": {},
154
- "chat_template": "{%- if tools -%}\n <|im_system|>tool_declare<|im_middle|>{{ tools | tojson }}<|im_end|>\n{%- endif -%}\n{%- for message in messages -%}\n {%- if loop.first and messages[0]['role'] != 'system' -%}\n <|im_system|>system<|im_middle|>You are a helpful assistant<|im_end|>\n {%- endif -%}\n {%- if message['role'] == 'system' -%}\n <|im_system|>system<|im_middle|>\n {%- elif message['role'] == 'user' -%}\n <|im_user|>user<|im_middle|>\n {%- elif message['role'] == 'assistant' -%}\n <|im_assistant|>assistant<|im_middle|>\n {%- elif message['role'] == 'tool' -%}\n <|im_system|>tool<|im_middle|>\n {%- endif -%}\n {%- if message['role'] == 'assistant' and message.get('tool_calls') -%}\n {%- if message['content'] -%}{{ message['content'] }}{%- endif -%}\n <|tool_calls_section_begin|>\n {%- for tool_call in message['tool_calls'] -%}\n {%- set func_name = tool_call['function']['name'] -%}\n {%- set formatted_id = 'functions.' + func_name + ':' + loop.index0|string -%}\n <|tool_call_begin|>{{ formatted_id }}<|tool_call_argument_begin|>{{ tool_call['function']['arguments'] | tojson}}<|tool_call_end|>\n {%- endfor -%}\n <|tool_calls_section_end|>\n {%- elif message['role'] == 'tool' -%}\n ## Return of {{ message.tool_call_id }}\\n{{ message['content'] }}\n {%- elif message['content'] is string -%}\n {{ message['content'] }}\n {%- elif message['content'] is not none -%}\n {% for content in message['content'] -%}\n {% if content['type'] == 'image' or 'image' in content or 'image_url' in content -%}\n <|media_start|>image<|media_content|><|media_pad|><|media_end|>\n {% else -%}\n {{ content['text'] }}\n {%- endif -%}\n {%- endfor -%}\n {%- endif -%}\n <|im_end|>\n{%- endfor -%}\n{%- if add_generation_prompt -%}\n <|im_assistant|>assistant<|im_middle|>\n{%- endif -%}",
155
- "model_max_length": 1000000000000000019884624838656,
156
- "pad_token": "[PAD]",
157
- "tokenizer_class": "TikTokenTokenizer",
158
- "unk_token": "[UNK]",
159
  "auto_map": {
160
  "AutoTokenizer": [
161
  "tokenization_kimi.TikTokenTokenizer",
162
  null
163
  ]
164
- }
165
- }
 
 
 
 
 
 
 
 
 
 
8
  "single_word": false,
9
  "special": true
10
  },
 
 
 
 
 
 
 
 
11
  "163586": {
12
  "content": "<|im_end|>",
13
  "lstrip": false,
 
64
  "single_word": false,
65
  "special": true
66
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
67
  "163601": {
68
  "content": "<|im_middle|>",
69
  "lstrip": false,
 
99
  "<|im_system|>",
100
  "<|im_middle|>"
101
  ],
 
 
 
 
 
 
 
 
 
102
  "auto_map": {
103
  "AutoTokenizer": [
104
  "tokenization_kimi.TikTokenTokenizer",
105
  null
106
  ]
107
+ },
108
+ "bos_token": "[BOS]",
109
+ "clean_up_tokenization_spaces": false,
110
+ "eos_token": "<|im_end|>",
111
+ "extra_special_tokens": {},
112
+ "model_max_length": 131072,
113
+ "pad_token": "[PAD]",
114
+ "padding_side": "left",
115
+ "tokenizer_class": "TikTokenTokenizer",
116
+ "unk_token": "[UNK]"
117
+ }