danielhanchen commited on
Commit
aad5bbd
·
verified ·
1 Parent(s): 57d94f7

Add files using upload-large-folder tool

Browse files
README.md CHANGED
@@ -1,16 +1,30 @@
1
  ---
2
- base_model: deepseek-ai/DeepSeek-Prover-V2-671B
3
- language:
4
- - en
5
- library_name: transformers
6
  tags:
7
- - deepseek
8
  - unsloth
9
- - transformers
10
- license: mit
 
11
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
12
 
13
- # deepseek-ai/DeepSeek-Prover-V2-671B
 
 
14
 
15
  <div align="center">
16
  <img src="https://github.com/deepseek-ai/DeepSeek-V2/blob/main/figures/logo.svg?raw=true" width="60%" alt="DeepSeek-V3" />
 
1
  ---
 
 
 
 
2
  tags:
 
3
  - unsloth
4
+ base_model:
5
+ - deepseek-ai/DeepSeek-Prover-V2-671B
6
+ library_name: transformers
7
  ---
8
+ <div>
9
+ <p style="margin-top: 0;margin-bottom: 0;">
10
+ <em><a href="https://docs.unsloth.ai/basics/unsloth-dynamic-v2.0-gguf">Unsloth Dynamic 2.0</a> achieves superior accuracy & outperforms other leading quants.</em>
11
+ </p>
12
+ <div style="display: flex; gap: 5px; align-items: center; ">
13
+ <a href="https://github.com/unslothai/unsloth/">
14
+ <img src="https://github.com/unslothai/unsloth/raw/main/images/unsloth%20new%20logo.png" width="133">
15
+ </a>
16
+ <a href="https://discord.gg/unsloth">
17
+ <img src="https://github.com/unslothai/unsloth/raw/main/images/Discord%20button.png" width="173">
18
+ </a>
19
+ <a href="https://docs.unsloth.ai/basics/qwen3-how-to-run-and-fine-tune">
20
+ <img src="https://raw.githubusercontent.com/unslothai/unsloth/refs/heads/main/images/documentation%20green%20button.png" width="143">
21
+ </a>
22
+ </div>
23
+ </div>
24
 
25
+ <!-- markdownlint-disable first-line-h1 -->
26
+ <!-- markdownlint-disable html -->
27
+ <!-- markdownlint-disable no-duplicate-header -->
28
 
29
  <div align="center">
30
  <img src="https://github.com/deepseek-ai/DeepSeek-V2/blob/main/figures/logo.svg?raw=true" width="60%" alt="DeepSeek-V3" />
chat_template.jinja ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% set ns = namespace(is_first=false, is_tool=false, is_output_first=true, system_prompt='', is_first_sp=true, is_last_user=false) %}{%- for message in messages %}{%- if message['role'] == 'system' %}{%- if ns.is_first_sp %}{% set ns.system_prompt = ns.system_prompt + message['content'] %}{% set ns.is_first_sp = false %}{%- else %}{% set ns.system_prompt = ns.system_prompt + '
2
+
3
+ ' + message['content'] %}{%- endif %}{%- endif %}{%- endfor %}{{ bos_token }}{{ ns.system_prompt }}{%- for message in messages %}{%- if message['role'] == 'user' %}{%- set ns.is_tool = false -%}{%- set ns.is_first = false -%}{%- set ns.is_last_user = true -%}{{'<|User|>' + message['content'] + '<|Assistant|>'}}{%- endif %}{%- if message['role'] == 'assistant' and message['tool_calls'] is defined and message['tool_calls'] is not none %}{%- set ns.is_last_user = false -%}{%- if ns.is_tool %}{{'<|tool▁outputs▁end|>'}}{%- endif %}{%- set ns.is_first = false %}{%- set ns.is_tool = false -%}{%- set ns.is_output_first = true %}{%- for tool in message['tool_calls'] %}{%- if not ns.is_first %}{%- if message['content'] is none %}{{'<|tool▁calls▁begin|><|tool▁call▁begin|>' + tool['type'] + '<|tool▁sep|>' + tool['function']['name'] + '
4
+ ' + '```json' + '
5
+ ' + tool['function']['arguments'] + '
6
+ ' + '```' + '<|tool▁call▁end|>'}}{%- else %}{{message['content'] + '<|tool▁calls▁begin|><|tool▁call▁begin|>' + tool['type'] + '<|tool▁sep|>' + tool['function']['name'] + '
7
+ ' + '```json' + '
8
+ ' + tool['function']['arguments'] + '
9
+ ' + '```' + '<|tool▁call▁end|>'}}{%- endif %}{%- set ns.is_first = true -%}{%- else %}{{'
10
+ ' + '<|tool▁call▁begin|>' + tool['type'] + '<|tool▁sep|>' + tool['function']['name'] + '
11
+ ' + '```json' + '
12
+ ' + tool['function']['arguments'] + '
13
+ ' + '```' + '<|tool▁call▁end|>'}}{%- endif %}{%- endfor %}{{'<|tool▁calls▁end|><|end▁of▁sentence|>'}}{%- endif %}{%- if message['role'] == 'assistant' and (message['tool_calls'] is not defined or message['tool_calls'] is none)%}{%- set ns.is_last_user = false -%}{%- if ns.is_tool %}{{'<|tool▁outputs▁end|>' + message['content'] + '<|end▁of▁sentence|>'}}{%- set ns.is_tool = false -%}{%- else %}{% set content = message['content'] %}{{content + '<|end▁of▁sentence|>'}}{%- endif %}{%- endif %}{%- if message['role'] == 'tool' %}{%- set ns.is_last_user = false -%}{%- set ns.is_tool = true -%}{%- if ns.is_output_first %}{{'<|tool▁outputs▁begin|><|tool▁output▁begin|>' + message['content'] + '<|tool▁output▁end|>'}}{%- set ns.is_output_first = false %}{%- else %}{{'
14
+ <|tool▁output▁begin|>' + message['content'] + '<|tool▁output▁end|>'}}{%- endif %}{%- endif %}{%- endfor -%}{% if ns.is_tool %}{{'<|tool▁outputs▁end|>'}}{% endif %}{% if add_generation_prompt and not ns.is_last_user and not ns.is_tool %}{{'<|Assistant|>'}}{% endif %}
config.json CHANGED
@@ -13,6 +13,7 @@
13
  "eos_token_id": 1,
14
  "ep_size": 1,
15
  "first_k_dense_replace": 3,
 
16
  "hidden_act": "silu",
17
  "hidden_size": 7168,
18
  "initializer_range": 0.02,
@@ -31,7 +32,10 @@
31
  "num_hidden_layers": 61,
32
  "num_key_value_heads": 128,
33
  "num_nextn_predict_layers": 1,
 
 
34
  "q_lora_rank": 1536,
 
35
  "qk_nope_head_dim": 128,
36
  "qk_rope_head_dim": 64,
37
  "quantization_config": {
@@ -44,13 +48,15 @@
44
  ]
45
  },
46
  "rms_norm_eps": 1e-06,
 
47
  "rope_scaling": {
48
- "beta_fast": 32,
49
- "beta_slow": 1,
50
- "factor": 40,
51
  "mscale": 1.0,
52
  "mscale_all_dim": 1.0,
53
  "original_max_position_embeddings": 4096,
 
54
  "type": "yarn"
55
  },
56
  "rope_theta": 10000,
@@ -60,7 +66,8 @@
60
  "topk_group": 4,
61
  "topk_method": "noaux_tc",
62
  "torch_dtype": "bfloat16",
63
- "transformers_version": "4.46.3",
 
64
  "use_cache": true,
65
  "v_head_dim": 128,
66
  "vocab_size": 129280
 
13
  "eos_token_id": 1,
14
  "ep_size": 1,
15
  "first_k_dense_replace": 3,
16
+ "head_dim": 64,
17
  "hidden_act": "silu",
18
  "hidden_size": 7168,
19
  "initializer_range": 0.02,
 
32
  "num_hidden_layers": 61,
33
  "num_key_value_heads": 128,
34
  "num_nextn_predict_layers": 1,
35
+ "pad_token_id": 2,
36
+ "pretraining_tp": 1,
37
  "q_lora_rank": 1536,
38
+ "qk_head_dim": 192,
39
  "qk_nope_head_dim": 128,
40
  "qk_rope_head_dim": 64,
41
  "quantization_config": {
 
48
  ]
49
  },
50
  "rms_norm_eps": 1e-06,
51
+ "rope_interleave": true,
52
  "rope_scaling": {
53
+ "beta_fast": 32.0,
54
+ "beta_slow": 1.0,
55
+ "factor": 40.0,
56
  "mscale": 1.0,
57
  "mscale_all_dim": 1.0,
58
  "original_max_position_embeddings": 4096,
59
+ "rope_type": "yarn",
60
  "type": "yarn"
61
  },
62
  "rope_theta": 10000,
 
66
  "topk_group": 4,
67
  "topk_method": "noaux_tc",
68
  "torch_dtype": "bfloat16",
69
+ "transformers_version": "4.52.3",
70
+ "unsloth_fixed": true,
71
  "use_cache": true,
72
  "v_head_dim": 128,
73
  "vocab_size": 129280
special_tokens_map.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|begin▁of▁sentence|>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|end▁of▁sentence|>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<|▁pad▁|>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ }
23
+ }
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
The diff for this file is too large to render. See raw diff