Upload folder using huggingface_hub

#1
.gitattributes CHANGED
@@ -1,35 +1,15 @@
1
- *.7z filter=lfs diff=lfs merge=lfs -text
2
- *.arrow filter=lfs diff=lfs merge=lfs -text
 
 
3
  *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bz2 filter=lfs diff=lfs merge=lfs -text
5
- *.ckpt filter=lfs diff=lfs merge=lfs -text
6
- *.ftz filter=lfs diff=lfs merge=lfs -text
7
- *.gz filter=lfs diff=lfs merge=lfs -text
8
- *.h5 filter=lfs diff=lfs merge=lfs -text
9
- *.joblib filter=lfs diff=lfs merge=lfs -text
10
- *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
- *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
  *.model filter=lfs diff=lfs merge=lfs -text
13
- *.msgpack filter=lfs diff=lfs merge=lfs -text
14
- *.npy filter=lfs diff=lfs merge=lfs -text
15
- *.npz filter=lfs diff=lfs merge=lfs -text
16
- *.onnx filter=lfs diff=lfs merge=lfs -text
17
- *.ot filter=lfs diff=lfs merge=lfs -text
18
- *.parquet filter=lfs diff=lfs merge=lfs -text
19
- *.pb filter=lfs diff=lfs merge=lfs -text
20
- *.pickle filter=lfs diff=lfs merge=lfs -text
21
- *.pkl filter=lfs diff=lfs merge=lfs -text
22
- *.pt filter=lfs diff=lfs merge=lfs -text
23
- *.pth filter=lfs diff=lfs merge=lfs -text
24
- *.rar filter=lfs diff=lfs merge=lfs -text
25
- *.safetensors filter=lfs diff=lfs merge=lfs -text
26
- saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
- *.tar.* filter=lfs diff=lfs merge=lfs -text
28
- *.tar filter=lfs diff=lfs merge=lfs -text
29
- *.tflite filter=lfs diff=lfs merge=lfs -text
30
- *.tgz filter=lfs diff=lfs merge=lfs -text
31
- *.wasm filter=lfs diff=lfs merge=lfs -text
32
- *.xz filter=lfs diff=lfs merge=lfs -text
33
- *.zip filter=lfs diff=lfs merge=lfs -text
34
- *.zst filter=lfs diff=lfs merge=lfs -text
35
- *tfevents* filter=lfs diff=lfs merge=lfs -text
 
1
+ model-00001-of-00004.safetensors filter=lfs diff=lfs merge=lfs -text
2
+ model-00002-of-00004.safetensors filter=lfs diff=lfs merge=lfs -text
3
+ model-00003-of-00004.safetensors filter=lfs diff=lfs merge=lfs -text
4
+ model-00004-of-00004.safetensors filter=lfs diff=lfs merge=lfs -text
5
  *.bin filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
6
  *.model filter=lfs diff=lfs merge=lfs -text
7
+ generation_config.json filter=lfs diff=lfs merge=lfs -text
8
+ vocab.json filter=lfs diff=lfs merge=lfs -text
9
+ merges.txt filter=lfs diff=lfs merge=lfs -text
10
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
11
+ added_tokens.json filter=lfs diff=lfs merge=lfs -text
12
+ config.json filter=lfs diff=lfs merge=lfs -text
13
+ model.safetensors.index.json filter=lfs diff=lfs merge=lfs -text
14
+ special_tokens_map.json filter=lfs diff=lfs merge=lfs -text
15
+ tokenizer_config.json filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
README.md CHANGED
@@ -1,3 +1,129 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # ⚛️ Atom-Searcher: Enhancing Agentic Deep Research via Fine-Grained Atomic Thought Reward
2
+
3
+ <p align="center">
4
+ <a href="https://arxiv.org/abs/2508.12800" target="_blank">
5
+ <img src="https://img.shields.io/badge/arXiv-2508.12800-b31b1b.svg?style=for-the-badge" alt="ArXiv">
6
+ </a>
7
+ <a href="https://huggingface.co/collections/ant-group/atom-searcher" target="_blank">
8
+ <img src="https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face-Models-yellow?style=for-the-badge" alt="Hugging Face">
9
+ </a>
10
+ <a href="https://github.com/antgroup/Research-Venus" target="_blank">
11
+ <img src="https://img.shields.io/badge/GitHub-Repo-blue?style=for-the-badge&logo=github" alt="GitHub">
12
+ </a>
13
+ </p>
14
+
15
+ [简体中文](README_CN.md)
16
+
17
+
18
+
19
+ ## 📖 Introduction
20
+
21
+ Atom-Searcher is a novel framework designed to enhance the deep research capabilities of Large Language Models (LLMs). While LLMs show great promise, their static internal knowledge limits their ability to handle complex, multi-step tasksExisting methods like Retrieval-Augmented Generation (RAG) and outcome-based reinforcement learning (RL) often fall short due to rigid workflows, reward sparsity, and conflicting gradients during training.
22
+
23
+ To overcome these challenges, we introduce **Atom-Searcher**, a new reinforcement learning framework built on the concept of **Atomic Thought**. This paradigm decomposes complex reasoning into fine-grained, functional units. Each "atomic thought" is evaluated by a Reasoning Reward Model (RRM), providing a fine-grained **Atomic Thought Reward (ATR)** that guides the agent's learning process.
24
+
25
+ The framework uses a curriculum-inspired reward schedule that initially prioritizes high-quality reasoning processes before shifting focus to final outcomes, which accelerates the discovery of effective problem-solving strategies.
26
+
27
+ Key advantages of Atom-Searcher include:
28
+ * **State-of-the-Art Performance**: Achieves consistent improvements over existing models on seven different benchmarks.
29
+ * **Enhanced Interpretability**: Exhibits more human-like and understandable reasoning patterns by breaking down its thought process.
30
+ * **Efficient Training**: Mitigates issues of reward sparsity and gradient conflicts, leading to more efficient policy optimization.
31
+ * **Scalable Computation**: Effectively scales its computational efforts during test-time to tackle more complex queries.
32
+
33
+ <p align="center">
34
+ <img src="png/sota_results.png" alt="Atom-Searcher SOTA Performance"/>
35
+ </p>
36
+ -----
37
+
38
+ # Overview
39
+
40
+ * [Key Highlights](https://www.google.com/search?q=%23key-highlights)
41
+ * [Evaluation](https://www.google.com/search?q=%23evaluation)
42
+ * [Citation](https://www.google.com/search?q=%23citation)
43
+
44
+ -----
45
+
46
+ # ✨ Key Highlights
47
+
48
+ We introduce **Atom-Searcher**, an agentic deep research framework that significantly improves LLM problem-solving by refining the reasoning process itself, not just the final outcome.
49
+
50
+ -----
51
+
52
+ ### 💡 Introducing the "Atomic Thought" Paradigm
53
+
54
+ We propose **Atomic Thought**, a novel thinking paradigm that decomposes complex reasoning into fine-grained, interpretable functional units. Instead of a single monolithic block of thought, the agent generates a sequence of atomic thoughts like `<OBSERVATION>`, `<HYPOTHESIS_TESTING>`, and `<RISK_ANALYSIS>` This structured approach leads to:
55
+
56
+ - ✅ More human-like, interpretable, and in-depth reasoning patterns
57
+ - ✅ Scales computation at test-time
58
+ - ✅ Provides supervision anchors for RRMs, bridging deep research tasks and RRMs.
59
+
60
+
61
+ -----
62
+
63
+ ### 🎯 Process-Supervised Reinforcement Learning with Fine-Grained Rewards
64
+
65
+ Current agents rely on outcome-based reinforcement learning (RL), which suffers from **reward sparsity** and **gradient conflicts**—penalizing an entire reasoning chain for one wrong final answer. Atom-Searcher addresses this with:
66
+
67
+ - 🔹 **Reasoning Reward Models (RRMs):** An RRM scores each individual Atomic Thought, providing dense, fine-grained process-level rewards called Atomic Thought Rewards (ATR).
68
+ - 🔹 **Curriculum-Inspired Reward Schedule:** The framework dynamically balances the weight of process-level ATR and final outcome rewards. Early in training, it prioritizes good reasoning (ATR), and as the agent improves, it shifts focus to correct answers.
69
+ - 🔹 **Efficient Optimization:** This hybrid reward structure alleviates reward sparsity and guides the agent to discover effective reasoning paths much faster.
70
+
71
+ -----
72
+
73
+ ### 🚀 SOTA Performance and Scalable Reasoning
74
+
75
+ We demonstrate through extensive experiments that Atom-Searcher sets a new state-of-the-art in agentic deep research.
76
+
77
+ - 📈 It achieves significant performance gains over strong baselines like **DeepResearcher** and **R1-Searcher** on seven distinct benchmarks.
78
+ - 🧠 At test time, Atom-Searcher **scales its computation effectively**, generating 3.2x more tokens and making 1.24x more tool calls on average than the SOTA baseline, indicating deeper exploration and reasoning without explicit incentives.
79
+
80
+ 👉[Hugging Face Model](https://www.google.com/search?q=https://huggingface.co/collections/ant-group/atom-searcher)
81
+
82
+ -----
83
+
84
+
85
+
86
+
87
+ ## Evaluation
88
+
89
+ Atom-Searcher's effectiveness is validated across a diverse set of seven open-domain QA benchmarks
90
+
91
+ ### Main Results on In-Domain and Out-of-Domain Benchmarks
92
+
93
+ Atom-Searcher consistently outperforms both training-based and prompt-based methods. All scores are F1 scores.
94
+
95
+ | **Type** | **Method** | **NQ** | **TQ** | **HotpotQA** | **2Wiki** | **Musique** | **Bamboogle** | **PopQA** |
96
+ | :--- | :--- | :---: | :---: | :---: | :---: | :---: | :---: | :---: |
97
+ | Prompt Based | Search-al-Web | 32.4 | 58.9 | 33.0 | 30.9 | 14.7 | 46.6 | 38.3 |
98
+ | Training Based | Search-R1-Instruct | 33.1 | 44.7 | 45.7 | 43.4 | 26.5 | 45.0 | 43.0 |
99
+ | | R1-Searcher | 35.4 | 73.1 | 44.8 | 59.4 | 22.8 | 64.8 | 42.7 |
100
+ | | DeepResearcher | 39.6 | 78.4 | 52.8 | 59.7 | 27.1 | **71.0** | 48.5 |
101
+ | | **Atom-Searcher (Ours)** | **44.0** | **81.8** | **57.3** | **66.9** | **27.6** | 70.7 | **50.3** |
102
+
103
+ > 🔝 **Experimental results show that Atom-Searcher achieves new state-of-the-art performance on 6 out of 7 benchmarks, with an average improvement of 8.5% on in-domain tasks and 2.5% on out-of-domain tasks over the previous SOTA, DeepResearcher.**
104
+
105
+ ### Ablation Study
106
+
107
+ The ablation study confirms that both **Atomic Thought** and the **Reasoning Reward Model (RRM)** are critical for performance. Adding RRM rewards without the structured Atomic Thoughts provides minimal benefit.
108
+
109
+ | **Method** | **NQ** | **TQ** | **HotpotQA** | **2Wiki** | **Musique** | **Bamboogle** | **PopQA** |
110
+ | :--- | :---: | :---: | :---: | :---: | :---: | :---: | :---: |
111
+ | Base (DeepResearcher) | 39.6 | 78.4 | 52.8 | 59.7 | 27.1 | 71.0 | 48.5 |
112
+ | + RRM | 40.1 | 78.2 | 53.5 | 60.0 | 25.7 | 70.5 | 48.8 |
113
+ | **Atom-Searcher (Base + RRM + Atomic Thought)** | **44.0** | **81.8** | **57.3** | **66.9** | **27.6** | **70.7** | **50.3** |
114
+
115
+ # Citation
116
+
117
+ Please consider citing if you find our work useful:
118
+
119
+ ```plain
120
+ @misc{deng2025atomsearcherenhancingagenticdeep,
121
+ title={Atom-Searcher: Enhancing Agentic Deep Research via Fine-Grained Atomic Thought Reward},
122
+ author={Yong Deng and Guoqing Wang and Zhenzhe Ying and Xiaofeng Wu and Jinzhen Lin and Wenwen Xiong and Yuqin Dai and Shuo Yang and Zhanwei Zhang and Qiwen Wang and Yang Qin and Changhua Meng},
123
+ year={2025},
124
+ eprint={2508.12800},
125
+ archivePrefix={arXiv},
126
+ primaryClass={cs.CL},
127
+ url={https://arxiv.org/abs/2508.12800},
128
+ }
129
+ ```
README_CN.md ADDED
@@ -0,0 +1,124 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ # ⚛️ Atom-Searcher: 通过细粒度原子化思想奖励增强智能体的深度研究能力
3
+
4
+ <p align="center">
5
+ <a href="https://arxiv.org/abs/2508.12800" target="_blank">
6
+ <img src="https://img.shields.io/badge/arXiv-2508.12800-b31b1b.svg?style=for-the-badge" alt="ArXiv">
7
+ </a>
8
+ <a href="https://huggingface.co/collections/ant-group/atom-searcher" target="_blank">
9
+ <img src="https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face-Models-yellow?style=for-the-badge" alt="Hugging Face">
10
+ </a>
11
+ <a href="https://github.com/antgroup/Research-Venus" target="_blank">
12
+ <img src="https://img.shields.io/badge/GitHub-Repo-blue?style=for-the-badge&logo=github" alt="GitHub">
13
+ </a>
14
+ </p>
15
+
16
+
17
+
18
+ ## 📖 引言
19
+
20
+ Atom-Searcher 是一个旨在增强大语言模型(LLMs)深度研究能力的新颖框架。尽管 LLMs 展现出巨大潜力,但其静态的内部知识限制了它们处理复杂、多步骤任务的能力。现有的方法,如检索增强生成(RAG)和基于结果的强化学习(RL),常常因其固化的工作流程、奖励稀疏性以及训练过程中的梯度冲突而表现不佳。
21
+
22
+ 为克服这些挑战,我们引入了 **Atom-Searcher**,这是一个建立在**原子化思想(Atomic Thought)概念之上的全新强化学习框架。该范式将复杂的推理过程分解为细粒度的功能单元。每一个“原子化思想”都由一个推理奖励模型(Reasoning Reward Model, RRM)进行评估,从而提供细粒度的原子化思想奖励(Atomic Thought Reward, ATR)**,用以指导智能体的学习过程。
23
+
24
+ 该框架采用了一种受课程学习启发的奖励机制,在初期优先奖励高质量的推理过程,随后再将重点转移到最终结果上,从而加速发现有效的问题解决策略。
25
+
26
+ Atom-Searcher 的主要优势包括:
27
+
28
+ * **最先进的性能**:在七个不同的基准测试中,相较于现有模型均取得了一致的提升。
29
+ * **增强的可解释性**:通过分解其思考过程,展现出更类似人类且易于理解的推理模式。
30
+ * **高效的训练**:缓解了奖励稀疏性和梯度冲突问题,使策略优化更为高效。
31
+ * **可扩展的计算能力**:在测试时能有效扩展其计算投入,以解决更复杂的查询。
32
+
33
+ <p align="center">
34
+ <img src="png/sota_results.png" alt="Atom-Searcher SOTA Performance"/>
35
+ </p>
36
+
37
+ # 概览
38
+
39
+ * [主要亮点](https://www.google.com/search?q=%23key-highlights)
40
+ * [评估](https://www.google.com/search?q=%23evaluation)
41
+ * [引用](https://www.google.com/search?q=%23citation)
42
+
43
+ -----
44
+
45
+ # ✨ 主要亮点
46
+
47
+ 我们推出了 **Atom-Searcher**,一个旨在提升智能体深度研究能力的框架。它通过优化推理过程本身,而不仅仅是最终结果,显著提高了 LLM 解决问题的能力。
48
+
49
+ -----
50
+
51
+ ### 💡 引入“原子化思想”范式
52
+
53
+ 我们提出了**原子化思想(Atomic Thought)**,这是一种新颖的思维范式,它将复杂的推理过程分解为细粒度的、可解释的功能单元。智能体不再生成一个单一、庞大的思想块,而是生成一系列原子化的思想,如 `<OBSERVATION>`(观察)、`<HYPOTHESIS_TESTING>`(假设检验)和 `<RISK_ANALYSIS>`(风险分析)。这种结构化的方法带来了:
54
+
55
+ - ✅ 更类似人类、可解释且更深入的推理模式
56
+ - ✅ 在测试时可扩展计算资源
57
+ - ✅ 为推理奖励模型(RRM)提供监督锚点,将深度研究任务与 RRM 联系起来。
58
+
59
+ -----
60
+
61
+ ### 🎯 结合细粒度奖励的过程监督强化学习
62
+
63
+ 当前的智能体依赖于基于结果的强化学习(RL),但这种方法存在**奖励稀疏性**和**梯度冲突**的问题——即因为一个最终答案的错误而惩罚整个推理链。Atom-Searcher 通过以下方式解决此问题:
64
+
65
+ - 🔹 **推理奖励模型(RRM):** RRM 为每一个独立的原子化思想打分,提供密集的、细粒度的过程级奖励,我们称之为原子化思想奖励(ATR)。
66
+ - 🔹 **课程学习式奖励策略:** 该框架动态平衡过程级 ATR 和最终结果奖励的权重。在训练初期,它优先鼓励良好的推理过程(ATR),随着智能体能力的提升,逐渐将重点转移到产出正确答案上。
67
+ - 🔹 **高效优化:** 这种混合奖励结构缓解了奖励稀疏性问题,并引导智能体更快地发现有效的推理路径。
68
+
69
+ -----
70
+
71
+ ### 🚀 SOTA 性能与可扩展的推理能力
72
+
73
+ 我们通过大量实验证明,Atom-Searcher 在智能体深度研究领域树立了新的技术标杆(SOTA)。
74
+
75
+ - 📈 在七个不同的基准测试中,它相较于 **DeepResearcher** 和 **R1-Searcher** 等强大的基线模型取得了显著的性能提升。
76
+ - 🧠 在测试时,Atom-Searcher 能**有效地扩展其计算资源**,与 SOTA 基线模型相比,平均多生成 3.2 倍的 token 并多进行 1.24 倍的工具调用,这表明在没有明确激励的情况下,它也能进行更深度的探索和推理。
77
+
78
+ 👉[Hugging Face 模���](https://www.google.com/search?q=https://huggingface.co/collections/ant-group/atom-searcher)
79
+
80
+ -----
81
+
82
+ ## 评估
83
+
84
+ Atom-Searcher 的有效性在一系列多样化的开放域问答(QA)基准测试中得到了验证,共涵盖七个数据集。
85
+
86
+ ### 在域内和域外基准测试上的主要结果
87
+
88
+ Atom-Searcher 的表现始终优于基于训练和基于提示的方法。所有分数均为 F1 分数。
89
+
90
+ | **类型** | **方法** | **NQ** | **TQ** | **HotpotQA** | **2Wiki** | **Musique** | **Bamboogle** | **PopQA** |
91
+ | :--- | :--- | :---: | :---: | :---: | :---: | :---: | :---: | :---: |
92
+ | 基于提示 | Search-al-Web | 32.4 | 58.9 | 33.0 | 30.9 | 14.7 | 46.6 | 38.3 |
93
+ | 基于训练 | Search-R1-Instruct | 33.1 | 44.7 | 45.7 | 43.4 | 26.5 | 45.0 | 43.0 |
94
+ | | R1-Searcher | 35.4 | 73.1 | 44.8 | 59.4 | 22.8 | 64.8 | 42.7 |
95
+ | | DeepResearcher | 39.6 | 78.4 | 52.8 | 59.7 | 27.1 | **71.0** | 48.5 |
96
+ | | **Atom-Searcher (我们的模型)** | **44.0** | **81.8** | **57.3** | **66.9** | **27.6** | 70.7 | **50.3** |
97
+
98
+ > 🔝 **实验结果表明,Atom-Searcher 在 7 个基准测试中的 6 个上取得了新的 SOTA 性能,与之前的 SOTA 模型 DeepResearcher 相比,在域内任务上平均提升了 8.5%,在域外任务上平均提升了 2.5%。**
99
+
100
+ ### 消融实验
101
+
102
+ 消融实验证实,**原子化思想**和\*\*推理奖励模型(RRM)\*\*对性能都至关重要。在没有结构化原子思想的情况下,仅添加 RRM 奖励所带来的收益微乎其微。
103
+
104
+ | **方法** | **NQ** | **TQ** | **HotpotQA** | **2Wiki** | **Musique** | **Bamboogle** | **PopQA** |
105
+ | :--- | :---: | :---: | :---: | :---: | :---: | :---: | :---: |
106
+ | 基线 (DeepResearcher) | 39.6 | 78.4 | 52.8 | 59.7 | 27.1 | 71.0 | 48.5 |
107
+ | + RRM | 40.1 | 78.2 | 53.5 | 60.0 | 25.7 | 70.5 | 48.8 |
108
+ | **Atom-Searcher (基线 + RRM + 原子化思想)** | **44.0** | **81.8** | **57.3** | **66.9** | **27.6** | **70.7** | **50.3** |
109
+
110
+ # 引用
111
+
112
+ 如果您觉得我们的工作对您有用,请考虑引用:
113
+
114
+ ```plain
115
+ @misc{deng2025atomsearcherenhancingagenticdeep,
116
+ title={Atom-Searcher: Enhancing Agentic Deep Research via Fine-Grained Atomic Thought Reward},
117
+ author={Yong Deng and Guoqing Wang and Zhenzhe Ying and Xiaofeng Wu and Jinzhen Lin and Wenwen Xiong and Yuqin Dai and Shuo Yang and Zhanwei Zhang and Qiwen Wang and Yang Qin and Changhua Meng},
118
+ year={2025},
119
+ eprint={2508.12800},
120
+ archivePrefix={arXiv},
121
+ primaryClass={cs.CL},
122
+ url={https://arxiv.org/abs/2508.12800},
123
+ }
124
+ ```
added_tokens.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</tool_call>": 151658,
3
+ "<tool_call>": 151657,
4
+ "<|box_end|>": 151649,
5
+ "<|box_start|>": 151648,
6
+ "<|endoftext|>": 151643,
7
+ "<|file_sep|>": 151664,
8
+ "<|fim_middle|>": 151660,
9
+ "<|fim_pad|>": 151662,
10
+ "<|fim_prefix|>": 151659,
11
+ "<|fim_suffix|>": 151661,
12
+ "<|im_end|>": 151645,
13
+ "<|im_start|>": 151644,
14
+ "<|image_pad|>": 151655,
15
+ "<|object_ref_end|>": 151647,
16
+ "<|object_ref_start|>": 151646,
17
+ "<|quad_end|>": 151651,
18
+ "<|quad_start|>": 151650,
19
+ "<|repo_name|>": 151663,
20
+ "<|video_pad|>": 151656,
21
+ "<|vision_end|>": 151653,
22
+ "<|vision_pad|>": 151654,
23
+ "<|vision_start|>": 151652
24
+ }
config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen2ForCausalLM"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token_id": 151643,
7
+ "eos_token_id": 151645,
8
+ "hidden_act": "silu",
9
+ "hidden_size": 3584,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 18944,
12
+ "max_position_embeddings": 32768,
13
+ "max_window_layers": 28,
14
+ "model_type": "qwen2",
15
+ "num_attention_heads": 28,
16
+ "num_hidden_layers": 28,
17
+ "num_key_value_heads": 4,
18
+ "rms_norm_eps": 1e-06,
19
+ "rope_scaling": null,
20
+ "rope_theta": 1000000.0,
21
+ "sliding_window": 131072,
22
+ "tie_word_embeddings": false,
23
+ "torch_dtype": "bfloat16",
24
+ "transformers_version": "4.51.3",
25
+ "use_cache": true,
26
+ "use_sliding_window": false,
27
+ "vocab_size": 152064
28
+ }
generation_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 151643,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 151645,
6
+ 151643
7
+ ],
8
+ "pad_token_id": 151643,
9
+ "repetition_penalty": 1.05,
10
+ "temperature": 0.7,
11
+ "top_k": 20,
12
+ "top_p": 0.8,
13
+ "transformers_version": "4.51.3"
14
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model-00001-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7df92c87f85b706ae6d14f3956d6fd38c2e28f7edcc2a3677acbaa48b6e880ec
3
+ size 4954735584
model-00002-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c897ca3c4f45624a03cb7a41814e0993a2b07952228a51fbf6524d9315189870
3
+ size 4026214376
model-00003-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32a3ca4173aea1d32616fa880ccb57b301b9e3b0e9c618efa46c5b725d2d9be7
3
+ size 4995166408
model-00004-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ad07daba8737cfe902364a2eca6be31667d34e272236e8edce6da8d1a1db2bd8
3
+ size 1255155504
model.safetensors.index.json ADDED
@@ -0,0 +1,346 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 15231233024
4
+ },
5
+ "weight_map": {
6
+ "lm_head.weight": "model-00004-of-00004.safetensors",
7
+ "model.embed_tokens.weight": "model-00003-of-00004.safetensors",
8
+ "model.layers.0.input_layernorm.weight": "model-00001-of-00004.safetensors",
9
+ "model.layers.0.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
10
+ "model.layers.0.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
11
+ "model.layers.0.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
12
+ "model.layers.0.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
13
+ "model.layers.0.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
14
+ "model.layers.0.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
15
+ "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
16
+ "model.layers.0.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
17
+ "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
18
+ "model.layers.0.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
19
+ "model.layers.0.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
20
+ "model.layers.1.input_layernorm.weight": "model-00002-of-00004.safetensors",
21
+ "model.layers.1.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
22
+ "model.layers.1.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
23
+ "model.layers.1.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
24
+ "model.layers.1.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
25
+ "model.layers.1.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
26
+ "model.layers.1.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
27
+ "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
28
+ "model.layers.1.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
29
+ "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
30
+ "model.layers.1.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
31
+ "model.layers.1.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
32
+ "model.layers.10.input_layernorm.weight": "model-00002-of-00004.safetensors",
33
+ "model.layers.10.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
34
+ "model.layers.10.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
35
+ "model.layers.10.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
36
+ "model.layers.10.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
37
+ "model.layers.10.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
38
+ "model.layers.10.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
39
+ "model.layers.10.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
40
+ "model.layers.10.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
41
+ "model.layers.10.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
42
+ "model.layers.10.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
43
+ "model.layers.10.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
44
+ "model.layers.11.input_layernorm.weight": "model-00003-of-00004.safetensors",
45
+ "model.layers.11.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
46
+ "model.layers.11.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
47
+ "model.layers.11.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
48
+ "model.layers.11.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
49
+ "model.layers.11.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
50
+ "model.layers.11.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
51
+ "model.layers.11.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
52
+ "model.layers.11.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
53
+ "model.layers.11.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
54
+ "model.layers.11.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
55
+ "model.layers.11.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
56
+ "model.layers.12.input_layernorm.weight": "model-00001-of-00004.safetensors",
57
+ "model.layers.12.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
58
+ "model.layers.12.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
59
+ "model.layers.12.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
60
+ "model.layers.12.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
61
+ "model.layers.12.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
62
+ "model.layers.12.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
63
+ "model.layers.12.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
64
+ "model.layers.12.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
65
+ "model.layers.12.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
66
+ "model.layers.12.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
67
+ "model.layers.12.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
68
+ "model.layers.13.input_layernorm.weight": "model-00001-of-00004.safetensors",
69
+ "model.layers.13.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
70
+ "model.layers.13.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
71
+ "model.layers.13.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
72
+ "model.layers.13.post_attention_layernorm.weight": "model-00004-of-00004.safetensors",
73
+ "model.layers.13.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
74
+ "model.layers.13.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
75
+ "model.layers.13.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
76
+ "model.layers.13.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
77
+ "model.layers.13.self_attn.q_proj.weight": "model-00004-of-00004.safetensors",
78
+ "model.layers.13.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
79
+ "model.layers.13.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
80
+ "model.layers.14.input_layernorm.weight": "model-00003-of-00004.safetensors",
81
+ "model.layers.14.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
82
+ "model.layers.14.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
83
+ "model.layers.14.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
84
+ "model.layers.14.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
85
+ "model.layers.14.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
86
+ "model.layers.14.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
87
+ "model.layers.14.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
88
+ "model.layers.14.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
89
+ "model.layers.14.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
90
+ "model.layers.14.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
91
+ "model.layers.14.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
92
+ "model.layers.15.input_layernorm.weight": "model-00003-of-00004.safetensors",
93
+ "model.layers.15.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
94
+ "model.layers.15.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
95
+ "model.layers.15.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
96
+ "model.layers.15.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
97
+ "model.layers.15.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
98
+ "model.layers.15.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
99
+ "model.layers.15.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
100
+ "model.layers.15.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
101
+ "model.layers.15.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
102
+ "model.layers.15.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
103
+ "model.layers.15.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
104
+ "model.layers.16.input_layernorm.weight": "model-00001-of-00004.safetensors",
105
+ "model.layers.16.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
106
+ "model.layers.16.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
107
+ "model.layers.16.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
108
+ "model.layers.16.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
109
+ "model.layers.16.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
110
+ "model.layers.16.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
111
+ "model.layers.16.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
112
+ "model.layers.16.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
113
+ "model.layers.16.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
114
+ "model.layers.16.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
115
+ "model.layers.16.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
116
+ "model.layers.17.input_layernorm.weight": "model-00003-of-00004.safetensors",
117
+ "model.layers.17.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
118
+ "model.layers.17.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
119
+ "model.layers.17.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
120
+ "model.layers.17.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
121
+ "model.layers.17.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
122
+ "model.layers.17.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
123
+ "model.layers.17.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
124
+ "model.layers.17.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
125
+ "model.layers.17.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
126
+ "model.layers.17.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
127
+ "model.layers.17.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
128
+ "model.layers.18.input_layernorm.weight": "model-00003-of-00004.safetensors",
129
+ "model.layers.18.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
130
+ "model.layers.18.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
131
+ "model.layers.18.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
132
+ "model.layers.18.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
133
+ "model.layers.18.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
134
+ "model.layers.18.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
135
+ "model.layers.18.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
136
+ "model.layers.18.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
137
+ "model.layers.18.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
138
+ "model.layers.18.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
139
+ "model.layers.18.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
140
+ "model.layers.19.input_layernorm.weight": "model-00003-of-00004.safetensors",
141
+ "model.layers.19.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
142
+ "model.layers.19.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
143
+ "model.layers.19.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
144
+ "model.layers.19.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
145
+ "model.layers.19.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
146
+ "model.layers.19.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
147
+ "model.layers.19.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
148
+ "model.layers.19.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
149
+ "model.layers.19.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
150
+ "model.layers.19.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
151
+ "model.layers.19.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
152
+ "model.layers.2.input_layernorm.weight": "model-00003-of-00004.safetensors",
153
+ "model.layers.2.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
154
+ "model.layers.2.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
155
+ "model.layers.2.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
156
+ "model.layers.2.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
157
+ "model.layers.2.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
158
+ "model.layers.2.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
159
+ "model.layers.2.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
160
+ "model.layers.2.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
161
+ "model.layers.2.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
162
+ "model.layers.2.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
163
+ "model.layers.2.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
164
+ "model.layers.20.input_layernorm.weight": "model-00001-of-00004.safetensors",
165
+ "model.layers.20.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
166
+ "model.layers.20.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
167
+ "model.layers.20.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
168
+ "model.layers.20.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
169
+ "model.layers.20.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
170
+ "model.layers.20.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
171
+ "model.layers.20.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
172
+ "model.layers.20.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
173
+ "model.layers.20.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
174
+ "model.layers.20.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
175
+ "model.layers.20.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
176
+ "model.layers.21.input_layernorm.weight": "model-00001-of-00004.safetensors",
177
+ "model.layers.21.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
178
+ "model.layers.21.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
179
+ "model.layers.21.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
180
+ "model.layers.21.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
181
+ "model.layers.21.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
182
+ "model.layers.21.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
183
+ "model.layers.21.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
184
+ "model.layers.21.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
185
+ "model.layers.21.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
186
+ "model.layers.21.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
187
+ "model.layers.21.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
188
+ "model.layers.22.input_layernorm.weight": "model-00003-of-00004.safetensors",
189
+ "model.layers.22.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
190
+ "model.layers.22.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
191
+ "model.layers.22.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
192
+ "model.layers.22.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
193
+ "model.layers.22.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
194
+ "model.layers.22.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
195
+ "model.layers.22.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
196
+ "model.layers.22.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
197
+ "model.layers.22.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
198
+ "model.layers.22.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
199
+ "model.layers.22.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
200
+ "model.layers.23.input_layernorm.weight": "model-00001-of-00004.safetensors",
201
+ "model.layers.23.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
202
+ "model.layers.23.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
203
+ "model.layers.23.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
204
+ "model.layers.23.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
205
+ "model.layers.23.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
206
+ "model.layers.23.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
207
+ "model.layers.23.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
208
+ "model.layers.23.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
209
+ "model.layers.23.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
210
+ "model.layers.23.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
211
+ "model.layers.23.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
212
+ "model.layers.24.input_layernorm.weight": "model-00001-of-00004.safetensors",
213
+ "model.layers.24.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
214
+ "model.layers.24.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
215
+ "model.layers.24.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
216
+ "model.layers.24.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
217
+ "model.layers.24.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
218
+ "model.layers.24.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
219
+ "model.layers.24.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
220
+ "model.layers.24.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
221
+ "model.layers.24.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
222
+ "model.layers.24.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
223
+ "model.layers.24.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
224
+ "model.layers.25.input_layernorm.weight": "model-00003-of-00004.safetensors",
225
+ "model.layers.25.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
226
+ "model.layers.25.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
227
+ "model.layers.25.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
228
+ "model.layers.25.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
229
+ "model.layers.25.self_attn.k_proj.bias": "model-00004-of-00004.safetensors",
230
+ "model.layers.25.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
231
+ "model.layers.25.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
232
+ "model.layers.25.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
233
+ "model.layers.25.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
234
+ "model.layers.25.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
235
+ "model.layers.25.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
236
+ "model.layers.26.input_layernorm.weight": "model-00002-of-00004.safetensors",
237
+ "model.layers.26.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
238
+ "model.layers.26.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
239
+ "model.layers.26.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
240
+ "model.layers.26.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
241
+ "model.layers.26.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
242
+ "model.layers.26.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
243
+ "model.layers.26.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
244
+ "model.layers.26.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
245
+ "model.layers.26.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
246
+ "model.layers.26.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
247
+ "model.layers.26.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
248
+ "model.layers.27.input_layernorm.weight": "model-00001-of-00004.safetensors",
249
+ "model.layers.27.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
250
+ "model.layers.27.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
251
+ "model.layers.27.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
252
+ "model.layers.27.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
253
+ "model.layers.27.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
254
+ "model.layers.27.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
255
+ "model.layers.27.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
256
+ "model.layers.27.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
257
+ "model.layers.27.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
258
+ "model.layers.27.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
259
+ "model.layers.27.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
260
+ "model.layers.3.input_layernorm.weight": "model-00001-of-00004.safetensors",
261
+ "model.layers.3.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
262
+ "model.layers.3.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
263
+ "model.layers.3.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
264
+ "model.layers.3.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
265
+ "model.layers.3.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
266
+ "model.layers.3.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
267
+ "model.layers.3.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
268
+ "model.layers.3.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
269
+ "model.layers.3.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
270
+ "model.layers.3.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
271
+ "model.layers.3.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
272
+ "model.layers.4.input_layernorm.weight": "model-00002-of-00004.safetensors",
273
+ "model.layers.4.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
274
+ "model.layers.4.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
275
+ "model.layers.4.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
276
+ "model.layers.4.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
277
+ "model.layers.4.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
278
+ "model.layers.4.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
279
+ "model.layers.4.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
280
+ "model.layers.4.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
281
+ "model.layers.4.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
282
+ "model.layers.4.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
283
+ "model.layers.4.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
284
+ "model.layers.5.input_layernorm.weight": "model-00002-of-00004.safetensors",
285
+ "model.layers.5.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
286
+ "model.layers.5.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
287
+ "model.layers.5.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
288
+ "model.layers.5.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
289
+ "model.layers.5.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
290
+ "model.layers.5.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
291
+ "model.layers.5.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
292
+ "model.layers.5.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
293
+ "model.layers.5.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
294
+ "model.layers.5.self_attn.v_proj.bias": "model-00004-of-00004.safetensors",
295
+ "model.layers.5.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
296
+ "model.layers.6.input_layernorm.weight": "model-00003-of-00004.safetensors",
297
+ "model.layers.6.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
298
+ "model.layers.6.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
299
+ "model.layers.6.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
300
+ "model.layers.6.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
301
+ "model.layers.6.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
302
+ "model.layers.6.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
303
+ "model.layers.6.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
304
+ "model.layers.6.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
305
+ "model.layers.6.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
306
+ "model.layers.6.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
307
+ "model.layers.6.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
308
+ "model.layers.7.input_layernorm.weight": "model-00003-of-00004.safetensors",
309
+ "model.layers.7.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
310
+ "model.layers.7.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
311
+ "model.layers.7.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
312
+ "model.layers.7.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
313
+ "model.layers.7.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
314
+ "model.layers.7.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
315
+ "model.layers.7.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
316
+ "model.layers.7.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
317
+ "model.layers.7.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
318
+ "model.layers.7.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
319
+ "model.layers.7.self_attn.v_proj.weight": "model-00004-of-00004.safetensors",
320
+ "model.layers.8.input_layernorm.weight": "model-00003-of-00004.safetensors",
321
+ "model.layers.8.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
322
+ "model.layers.8.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
323
+ "model.layers.8.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
324
+ "model.layers.8.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
325
+ "model.layers.8.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
326
+ "model.layers.8.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
327
+ "model.layers.8.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
328
+ "model.layers.8.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
329
+ "model.layers.8.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
330
+ "model.layers.8.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
331
+ "model.layers.8.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
332
+ "model.layers.9.input_layernorm.weight": "model-00002-of-00004.safetensors",
333
+ "model.layers.9.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
334
+ "model.layers.9.mlp.gate_proj.weight": "model-00004-of-00004.safetensors",
335
+ "model.layers.9.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
336
+ "model.layers.9.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
337
+ "model.layers.9.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
338
+ "model.layers.9.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
339
+ "model.layers.9.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
340
+ "model.layers.9.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
341
+ "model.layers.9.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
342
+ "model.layers.9.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
343
+ "model.layers.9.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
344
+ "model.norm.weight": "model-00003-of-00004.safetensors"
345
+ }
346
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|im_end|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|endoftext|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c5ae00e602b8860cbd784ba82a8aa14e8feecec692e7076590d014d7b7fdafa
3
+ size 11421896
tokenizer_config.json ADDED
@@ -0,0 +1,208 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ }
181
+ },
182
+ "additional_special_tokens": [
183
+ "<|im_start|>",
184
+ "<|im_end|>",
185
+ "<|object_ref_start|>",
186
+ "<|object_ref_end|>",
187
+ "<|box_start|>",
188
+ "<|box_end|>",
189
+ "<|quad_start|>",
190
+ "<|quad_end|>",
191
+ "<|vision_start|>",
192
+ "<|vision_end|>",
193
+ "<|vision_pad|>",
194
+ "<|image_pad|>",
195
+ "<|video_pad|>"
196
+ ],
197
+ "bos_token": null,
198
+ "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are Qwen, created by Alibaba Cloud. You are a helpful assistant.' }}\n {%- endif %}\n {{- \"\\n\\n# Tools\\n\\nYou may call one or more functions to assist with the user query.\\n\\nYou are provided with function signatures within <tools></tools> XML tags:\\n<tools>\" }}\n {%- for tool in tools %}\n {{- \"\\n\" }}\n {{- tool | tojson }}\n {%- endfor %}\n {{- \"\\n</tools>\\n\\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\\n<tool_call>\\n{{\\\"name\\\": <function-name>, \\\"arguments\\\": <args-json-object>}}\\n</tool_call><|im_end|>\\n\" }}\n{%- else %}\n {%- if messages[0]['role'] == 'system' %}\n {{- '<|im_start|>system\\n' + messages[0]['content'] + '<|im_end|>\\n' }}\n {%- else %}\n {{- '<|im_start|>system\\nYou are Qwen, created by Alibaba Cloud. You are a helpful assistant.<|im_end|>\\n' }}\n {%- endif %}\n{%- endif %}\n{%- for message in messages %}\n {%- if (message.role == \"user\") or (message.role == \"system\" and not loop.first) or (message.role == \"assistant\" and not message.tool_calls) %}\n {{- '<|im_start|>' + message.role + '\\n' + message.content + '<|im_end|>' + '\\n' }}\n {%- elif message.role == \"assistant\" %}\n {{- '<|im_start|>' + message.role }}\n {%- if message.content %}\n {{- '\\n' + message.content }}\n {%- endif %}\n {%- for tool_call in message.tool_calls %}\n {%- if tool_call.function is defined %}\n {%- set tool_call = tool_call.function %}\n {%- endif %}\n {{- '\\n<tool_call>\\n{\"name\": \"' }}\n {{- tool_call.name }}\n {{- '\", \"arguments\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- '}\\n</tool_call>' }}\n {%- endfor %}\n {{- '<|im_end|>\\n' }}\n {%- elif message.role == \"tool\" %}\n {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != \"tool\") %}\n {{- '<|im_start|>user' }}\n {%- endif %}\n {{- '\\n<tool_response>\\n' }}\n {{- message.content }}\n {{- '\\n</tool_response>' }}\n {%- if loop.last or (messages[loop.index0 + 1].role != \"tool\") %}\n {{- '<|im_end|>\\n' }}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|im_start|>assistant\\n' }}\n{%- endif %}\n",
199
+ "clean_up_tokenization_spaces": false,
200
+ "eos_token": "<|im_end|>",
201
+ "errors": "replace",
202
+ "extra_special_tokens": {},
203
+ "model_max_length": 131072,
204
+ "pad_token": "<|endoftext|>",
205
+ "split_special_tokens": false,
206
+ "tokenizer_class": "Qwen2Tokenizer",
207
+ "unk_token": null
208
+ }
vocab.json ADDED
The diff for this file is too large to render. See raw diff