Night-Quiet commited on
Commit
d042a27
·
verified ·
1 Parent(s): 6a670e2

Add files using upload-large-folder tool

Browse files
.ipynb_checkpoints/README-checkpoint.md ADDED
@@ -0,0 +1,79 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - 中医大模型
5
+
6
+ #model-type:
7
+ ##如 gpt、phi、llama、chatglm、baichuan 等
8
+ #- gpt
9
+
10
+ #domain:
11
+ ##如 nlp、cv、audio、multi-modal
12
+ #- nlp
13
+
14
+ #language:
15
+ ##语言代码列表 https://help.aliyun.com/document_detail/215387.html?spm=a2c4g.11186623.0.0.9f8d7467kni6Aa
16
+ #- cn
17
+
18
+ #metrics:
19
+ ##如 CIDEr、Blue、ROUGE 等
20
+ #- CIDEr
21
+
22
+ #tags:
23
+ ##各种自定义,包括 pretrained、fine-tuned、instruction-tuned、RL-tuned 等训练方法和其他
24
+ #- pretrained
25
+
26
+ #tools:
27
+ ##如 vllm、fastchat、llamacpp、AdaSeq 等
28
+ #- vllm
29
+ - 心语心言
30
+ - 医疗
31
+ - 医疗大模型
32
+ language:
33
+ - zh
34
+ frameworks: PyTorch
35
+ tasks:
36
+ - text-generation
37
+ base_model:
38
+ - Qwen/Qwen3-Next-80B-A3B-Instruct
39
+ base_model_relation: finetune
40
+ metrics:
41
+ - accuracy
42
+ ---
43
+ # DeepPulse-80B TCM Large Model Series
44
+
45
+ **DeepPulse (深度把脉)** is the core achievement of 心语心言's open-source Traditional Chinese Medicine (TCM) large model series. This series of models uses Qwen3-Next-80B as the base model and has undergone deep fine-tuning using a self-built high-quality TCM clinical medical dataset. This release includes two versions:
46
+
47
+ * **DeepPulse-80B-Thinking-V0.1**: Focuses on complex clinical reasoning and assisted diagnosis, achieving first place in total score in public evaluations, demonstrating top-tier logical reasoning capabilities in the TCM domain.
48
+ * **DeepPulse-80B-Instruct-V0.1**: Possesses excellent TCM instruction-following capabilities, suitable for a wide range of TCM Q&A and interactive scenarios, with a comprehensive ranking of sixth.
49
+
50
+ # Public TCM Benchmark Metrics Comparison (MedBench - TCM-5CEval)
51
+
52
+ TCM-5CEval is an authoritative evaluation benchmark for TCM large models, comprising the following five subtasks that comprehensively assess the model's TCM capabilities:
53
+
54
+ * **TCM-Exam (中医考试)**: Evaluates the mastery and application of fundamental TCM theories (Yin-Yang, Zang-Fu organs, etc.) and diagnostics knowledge.
55
+ * **TCM-LitQA (典籍问答)**: Tests deep understanding and reasoning of classic TCM texts such as "Huangdi Neijing" and "Shanghan Lun".
56
+ * **TCM-MRCD (临床诊疗)**: Simulates real clinical scenarios, evaluating the model's ability to analyze medical cases, perform pattern differentiation, and make prescription decisions.
57
+ * **TCM-CMM (中药方剂)**: Measures the model's knowledge of Chinese materia medica properties, effects, compatibility contraindications, and formula applications.
58
+ * **TCM-ClinNPT (非药物疗法)**: Assesses ability in acupoint selection for acupuncture, Tuina massage techniques, and pattern-based treatment for specific clinical scenarios.
59
+
60
+ | No. | Model Name | Organization/Team Name | Release Date | Type | Parameters | Total Score | TCM-Exam | TCM-LitQA | TCM-MRCD | TCM-CMM | TCM-ClinNPT |
61
+ | --- | --- | --- | --- | --- | --- | --- | --- | --- | --- | --- | --- |
62
+ | 1 | <font color="red">DeepPulse-80B-Thinking-V0.1</font> | <font color="red">心语心言</font> | <font color="red">2025/12/23</font> | <font color="red">开源</font> | <font color="red">80B</font> | <font color="red">71.3</font> | <font color="red">83.0</font> | <font color="red">45.5</font> | <font color="red">75.4</font> | <font color="red">84.9</font> | <font color="red">67.6</font> |
63
+ | 2 | HKR_TCM_HW_v1 | 港仔机器人主动健管团队 | 2025/12/12 | 闭源 | 671B | 70.8 | 85.4 | 44.2 | 73.1 | 83.8 | 67.5 |
64
+ | 3 | Gemini-2.5-Pro-nothinking | Google | 2025/03/25 | 闭源 | N/A | 69.2 | 77.9 | 62.0 | 72.4 | 72.6 | 61.2 |
65
+ | 4 | DeepSeek-V3.2 | DeepSeek | 2025/12/01 | 开源 | 671B | 66.8 | 74.5 | 44.4 | 66.8 | 80.0 | 68.3 |
66
+ | 5 | Grok-4 | xAI | 2025/07/09 | 闭源 | N/A | 66.6 | 73.0 | 59.3 | 68.4 | 68.0 | 64.2 |
67
+ | 6 | <font color="red">DeepPulse-80B-Instruct-V0.1</font> | <font color="red">心语心言</font> | <font color="red">2025/12/23</font> | <font color="red">开源</font> | <font color="red">80B</font> | <font color="red">66.2</font> | <font color="red">74.4</font> | <font color="red">40.7</font> | <font color="red">70.6</font> | <font color="red">79.7</font> | <font color="red">65.6</font> |
68
+ | 7 | Qwen3-235B-A22B-Thinking-2507 | Alibaba | 2025/08/17 | 开源 | 235B | 64.8 | 75.5 | 40.3 | 68.5 | 78.2 | 61.5 |
69
+ | 8 | Claude-Sonnet-4.5 | Anthropic | 2025/09/29 | 闭源 | N/A | 64.8 | 69.8 | 59.3 | 67.2 | 71.7 | 56.0 |
70
+ | 9 | GPT-5 | OpenAI | 2025/08/07 | 闭源 | N/A | 63.6 | 75.0 | 51.9 | 64.1 | 66.6 | 60.6 |
71
+ | 10 | Qwen3-Next-80B-A3B-Thinking | Alibaba | 2025/09/15 | 开源 | 80B | 63.5 | 76.0 | 38.2 | 66.2 | 77.9 | 59.4 |
72
+ | 11 | Llama-4-maverick | Meta | 2025/04/06 | 开源 | 400B | 57.2 | 72.1 | 51.3 | 63.8 | 54.4 | 44.3 |
73
+ | 12 | GPT-4o | OpenAI | 2025/05/13 | 闭源 | 200B | 55.9 | 66.5 | 46.9 | 60.9 | 57.1 | 47.9 |
74
+
75
+ > Note: "N/A" in the Parameters column indicates that the model's parameter count has not been publicly disclosed.
76
+ >
77
+ > Except for `DeepSeek-V3.2`, `Qwen3-235B-A22B-Thinking-2507`, `Qwen3-Next-80B-A3B-Thinking` which are self-tested deployment data, other models reference publicly available leaderboard data.
78
+ >
79
+ > TCM-5CEval: https://medbench.opencompass.org.cn/track-detail/tcmeval
README.md ADDED
@@ -0,0 +1,79 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - 中医大模型
5
+
6
+ #model-type:
7
+ ##如 gpt、phi、llama、chatglm、baichuan 等
8
+ #- gpt
9
+
10
+ #domain:
11
+ ##如 nlp、cv、audio、multi-modal
12
+ #- nlp
13
+
14
+ #language:
15
+ ##语言代码列表 https://help.aliyun.com/document_detail/215387.html?spm=a2c4g.11186623.0.0.9f8d7467kni6Aa
16
+ #- cn
17
+
18
+ #metrics:
19
+ ##如 CIDEr、Blue、ROUGE 等
20
+ #- CIDEr
21
+
22
+ #tags:
23
+ ##各种自定义,包括 pretrained、fine-tuned、instruction-tuned、RL-tuned 等训练方法和其他
24
+ #- pretrained
25
+
26
+ #tools:
27
+ ##如 vllm、fastchat、llamacpp、AdaSeq 等
28
+ #- vllm
29
+ - 心语心言
30
+ - 医疗
31
+ - 医疗大模型
32
+ language:
33
+ - zh
34
+ frameworks: PyTorch
35
+ tasks:
36
+ - text-generation
37
+ base_model:
38
+ - Qwen/Qwen3-Next-80B-A3B-Instruct
39
+ base_model_relation: finetune
40
+ metrics:
41
+ - accuracy
42
+ ---
43
+ # DeepPulse-80B TCM Large Model Series
44
+
45
+ **DeepPulse (深度把脉)** is the core achievement of 心语心言's open-source Traditional Chinese Medicine (TCM) large model series. This series of models uses Qwen3-Next-80B as the base model and has undergone deep fine-tuning using a self-built high-quality TCM clinical medical dataset. This release includes two versions:
46
+
47
+ * **DeepPulse-80B-Thinking-V0.1**: Focuses on complex clinical reasoning and assisted diagnosis, achieving first place in total score in public evaluations, demonstrating top-tier logical reasoning capabilities in the TCM domain.
48
+ * **DeepPulse-80B-Instruct-V0.1**: Possesses excellent TCM instruction-following capabilities, suitable for a wide range of TCM Q&A and interactive scenarios, with a comprehensive ranking of sixth.
49
+
50
+ # Public TCM Benchmark Metrics Comparison (MedBench - TCM-5CEval)
51
+
52
+ TCM-5CEval is an authoritative evaluation benchmark for TCM large models, comprising the following five subtasks that comprehensively assess the model's TCM capabilities:
53
+
54
+ * **TCM-Exam (中医考试)**: Evaluates the mastery and application of fundamental TCM theories (Yin-Yang, Zang-Fu organs, etc.) and diagnostics knowledge.
55
+ * **TCM-LitQA (典籍问答)**: Tests deep understanding and reasoning of classic TCM texts such as "Huangdi Neijing" and "Shanghan Lun".
56
+ * **TCM-MRCD (临床诊疗)**: Simulates real clinical scenarios, evaluating the model's ability to analyze medical cases, perform pattern differentiation, and make prescription decisions.
57
+ * **TCM-CMM (中药方剂)**: Measures the model's knowledge of Chinese materia medica properties, effects, compatibility contraindications, and formula applications.
58
+ * **TCM-ClinNPT (非药物疗法)**: Assesses ability in acupoint selection for acupuncture, Tuina massage techniques, and pattern-based treatment for specific clinical scenarios.
59
+
60
+ | No. | Model Name | Organization/Team Name | Release Date | Type | Parameters | Total Score | TCM-Exam | TCM-LitQA | TCM-MRCD | TCM-CMM | TCM-ClinNPT |
61
+ | --- | --- | --- | --- | --- | --- | --- | --- | --- | --- | --- | --- |
62
+ | 1 | <font color="red">DeepPulse-80B-Thinking-V0.1</font> | <font color="red">心语心言</font> | <font color="red">2025/12/23</font> | <font color="red">开源</font> | <font color="red">80B</font> | <font color="red">71.3</font> | <font color="red">83.0</font> | <font color="red">45.5</font> | <font color="red">75.4</font> | <font color="red">84.9</font> | <font color="red">67.6</font> |
63
+ | 2 | HKR_TCM_HW_v1 | 港仔机器人主动健管团队 | 2025/12/12 | 闭源 | 671B | 70.8 | 85.4 | 44.2 | 73.1 | 83.8 | 67.5 |
64
+ | 3 | Gemini-2.5-Pro-nothinking | Google | 2025/03/25 | 闭源 | N/A | 69.2 | 77.9 | 62.0 | 72.4 | 72.6 | 61.2 |
65
+ | 4 | DeepSeek-V3.2 | DeepSeek | 2025/12/01 | 开源 | 671B | 66.8 | 74.5 | 44.4 | 66.8 | 80.0 | 68.3 |
66
+ | 5 | Grok-4 | xAI | 2025/07/09 | 闭源 | N/A | 66.6 | 73.0 | 59.3 | 68.4 | 68.0 | 64.2 |
67
+ | 6 | <font color="red">DeepPulse-80B-Instruct-V0.1</font> | <font color="red">心语心言</font> | <font color="red">2025/12/23</font> | <font color="red">开源</font> | <font color="red">80B</font> | <font color="red">66.2</font> | <font color="red">74.4</font> | <font color="red">40.7</font> | <font color="red">70.6</font> | <font color="red">79.7</font> | <font color="red">65.6</font> |
68
+ | 7 | Qwen3-235B-A22B-Thinking-2507 | Alibaba | 2025/08/17 | 开源 | 235B | 64.8 | 75.5 | 40.3 | 68.5 | 78.2 | 61.5 |
69
+ | 8 | Claude-Sonnet-4.5 | Anthropic | 2025/09/29 | 闭源 | N/A | 64.8 | 69.8 | 59.3 | 67.2 | 71.7 | 56.0 |
70
+ | 9 | GPT-5 | OpenAI | 2025/08/07 | 闭源 | N/A | 63.6 | 75.0 | 51.9 | 64.1 | 66.6 | 60.6 |
71
+ | 10 | Qwen3-Next-80B-A3B-Thinking | Alibaba | 2025/09/15 | 开源 | 80B | 63.5 | 76.0 | 38.2 | 66.2 | 77.9 | 59.4 |
72
+ | 11 | Llama-4-maverick | Meta | 2025/04/06 | 开源 | 400B | 57.2 | 72.1 | 51.3 | 63.8 | 54.4 | 44.3 |
73
+ | 12 | GPT-4o | OpenAI | 2025/05/13 | 闭源 | 200B | 55.9 | 66.5 | 46.9 | 60.9 | 57.1 | 47.9 |
74
+
75
+ > Note: "N/A" in the Parameters column indicates that the model's parameter count has not been publicly disclosed.
76
+ >
77
+ > Except for `DeepSeek-V3.2`, `Qwen3-235B-A22B-Thinking-2507`, `Qwen3-Next-80B-A3B-Thinking` which are self-tested deployment data, other models reference publicly available leaderboard data.
78
+ >
79
+ > TCM-5CEval: https://medbench.opencompass.org.cn/track-detail/tcmeval
added_tokens.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</think>": 151668,
3
+ "</tool_call>": 151658,
4
+ "</tool_response>": 151666,
5
+ "<think>": 151667,
6
+ "<tool_call>": 151657,
7
+ "<tool_response>": 151665,
8
+ "<|box_end|>": 151649,
9
+ "<|box_start|>": 151648,
10
+ "<|endoftext|>": 151643,
11
+ "<|file_sep|>": 151664,
12
+ "<|fim_middle|>": 151660,
13
+ "<|fim_pad|>": 151662,
14
+ "<|fim_prefix|>": 151659,
15
+ "<|fim_suffix|>": 151661,
16
+ "<|im_end|>": 151645,
17
+ "<|im_start|>": 151644,
18
+ "<|image_pad|>": 151655,
19
+ "<|object_ref_end|>": 151647,
20
+ "<|object_ref_start|>": 151646,
21
+ "<|quad_end|>": 151651,
22
+ "<|quad_start|>": 151650,
23
+ "<|repo_name|>": 151663,
24
+ "<|video_pad|>": 151656,
25
+ "<|vision_end|>": 151653,
26
+ "<|vision_pad|>": 151654,
27
+ "<|vision_start|>": 151652
28
+ }
chat_template.jinja ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if tools %}
2
+ {{- '<|im_start|>system\n' }}
3
+ {%- if messages[0].role == 'system' %}
4
+ {{- messages[0].content + '\n\n' }}
5
+ {%- endif %}
6
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
7
+ {%- for tool in tools %}
8
+ {{- "\n" }}
9
+ {{- tool | tojson }}
10
+ {%- endfor %}
11
+ {{- "\n</tools>\n\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call><|im_end|>\n" }}
12
+ {%- else %}
13
+ {%- if messages[0].role == 'system' %}
14
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
15
+ {%- endif %}
16
+ {%- endif %}
17
+ {%- for message in messages %}
18
+ {%- if message.content is string %}
19
+ {%- set content = message.content %}
20
+ {%- else %}
21
+ {%- set content = '' %}
22
+ {%- endif %}
23
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
24
+ {{- '<|im_start|>' + message.role + '\n' + content + '<|im_end|>' + '\n' }}
25
+ {%- elif message.role == "assistant" %}
26
+ {{- '<|im_start|>' + message.role + '\n' + content }}
27
+ {%- if message.tool_calls %}
28
+ {%- for tool_call in message.tool_calls %}
29
+ {%- if (loop.first and content) or (not loop.first) %}
30
+ {{- '\n' }}
31
+ {%- endif %}
32
+ {%- if tool_call.function %}
33
+ {%- set tool_call = tool_call.function %}
34
+ {%- endif %}
35
+ {{- '<tool_call>\n{"name": "' }}
36
+ {{- tool_call.name }}
37
+ {{- '", "arguments": ' }}
38
+ {%- if tool_call.arguments is string %}
39
+ {{- tool_call.arguments }}
40
+ {%- else %}
41
+ {{- tool_call.arguments | tojson }}
42
+ {%- endif %}
43
+ {{- '}\n</tool_call>' }}
44
+ {%- endfor %}
45
+ {%- endif %}
46
+ {{- '<|im_end|>\n' }}
47
+ {%- elif message.role == "tool" %}
48
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
49
+ {{- '<|im_start|>user' }}
50
+ {%- endif %}
51
+ {{- '\n<tool_response>\n' }}
52
+ {{- content }}
53
+ {{- '\n</tool_response>' }}
54
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
55
+ {{- '<|im_end|>\n' }}
56
+ {%- endif %}
57
+ {%- endif %}
58
+ {%- endfor %}
59
+ {%- if add_generation_prompt %}
60
+ {{- '<|im_start|>assistant\n' }}
61
+ {%- endif %}
config.json ADDED
@@ -0,0 +1,94 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen3NextForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 151643,
8
+ "decoder_sparse_step": 1,
9
+ "dtype": "bfloat16",
10
+ "eos_token_id": 151645,
11
+ "full_attention_interval": 4,
12
+ "head_dim": 256,
13
+ "hidden_act": "silu",
14
+ "hidden_size": 2048,
15
+ "initializer_range": 0.02,
16
+ "intermediate_size": 5120,
17
+ "layer_types": [
18
+ "linear_attention",
19
+ "linear_attention",
20
+ "linear_attention",
21
+ "full_attention",
22
+ "linear_attention",
23
+ "linear_attention",
24
+ "linear_attention",
25
+ "full_attention",
26
+ "linear_attention",
27
+ "linear_attention",
28
+ "linear_attention",
29
+ "full_attention",
30
+ "linear_attention",
31
+ "linear_attention",
32
+ "linear_attention",
33
+ "full_attention",
34
+ "linear_attention",
35
+ "linear_attention",
36
+ "linear_attention",
37
+ "full_attention",
38
+ "linear_attention",
39
+ "linear_attention",
40
+ "linear_attention",
41
+ "full_attention",
42
+ "linear_attention",
43
+ "linear_attention",
44
+ "linear_attention",
45
+ "full_attention",
46
+ "linear_attention",
47
+ "linear_attention",
48
+ "linear_attention",
49
+ "full_attention",
50
+ "linear_attention",
51
+ "linear_attention",
52
+ "linear_attention",
53
+ "full_attention",
54
+ "linear_attention",
55
+ "linear_attention",
56
+ "linear_attention",
57
+ "full_attention",
58
+ "linear_attention",
59
+ "linear_attention",
60
+ "linear_attention",
61
+ "full_attention",
62
+ "linear_attention",
63
+ "linear_attention",
64
+ "linear_attention",
65
+ "full_attention"
66
+ ],
67
+ "linear_conv_kernel_dim": 4,
68
+ "linear_key_head_dim": 128,
69
+ "linear_num_key_heads": 16,
70
+ "linear_num_value_heads": 32,
71
+ "linear_value_head_dim": 128,
72
+ "max_position_embeddings": 262144,
73
+ "mlp_only_layers": [],
74
+ "model_type": "qwen3_next",
75
+ "moe_intermediate_size": 512,
76
+ "norm_topk_prob": true,
77
+ "num_attention_heads": 16,
78
+ "num_experts": 512,
79
+ "num_experts_per_tok": 10,
80
+ "num_hidden_layers": 48,
81
+ "num_key_value_heads": 2,
82
+ "output_router_logits": false,
83
+ "partial_rotary_factor": 0.25,
84
+ "rms_norm_eps": 1e-06,
85
+ "rope_scaling": null,
86
+ "rope_theta": 10000000,
87
+ "router_aux_loss_coef": 0.001,
88
+ "shared_expert_intermediate_size": 512,
89
+ "tie_word_embeddings": false,
90
+ "transformers_version": "4.57.1",
91
+ "use_cache": true,
92
+ "use_sliding_window": false,
93
+ "vocab_size": 151936
94
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
model17_0.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b224fe5b651e0b0da8b1208776bb86d18690f4ff6a2d1d85a671a89beaca3f6
3
+ size 4832125960
model18_0.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8fe7a9d44b35f45564a134b8888a3170ad830cddc08e9c1e7fced650f5e1ee76
3
+ size 4832125960
model19_0.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d4facec77c41938e366e7ff290cfe9934362d3dfb5688b5ec167351988b45a7
3
+ size 4832125960
model1_0.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13dedd057368d95a8a03539b3bbd14ad013aa91309e468577cbc572227a9a864
3
+ size 4832124808
model7_0.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9866a670e13ef5324617f069ed411b666488de7cd6b6e8287b6c490ed9a0cfff
3
+ size 4832124808
special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|im_end|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|endoftext|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }