zwpride-iquestlab commited on
Commit
5eed073
·
verified ·
1 Parent(s): 1fa15f7

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +18 -10
README.md CHANGED
@@ -12,14 +12,16 @@ library_name: transformers
12
  ![Evaluation Results](./papers/iquest-coder-v1-logo.png)
13
 
14
  <p align="center">
15
- 📘 <a href="https://iquestlab.github.io">Blog</a >
 
 
16
  &nbsp;•&nbsp;
17
  📄 <a href="https://github.com/IQuestLab/IQuest-Coder-V1/blob/main/papers/IQuest_Coder_Technical_Report.pdf">Technical Report</a >
18
  </p >
19
 
20
- # IQuest-Coder-V1 Model Family
21
 
22
- 🚀 **[[IQuest-Coder-V1 Update](https://iquestlab.github.io/release-1.0-2602/index.html)]**: Released 7B & 14B Family Models and 40B-Thinking, specially optimized for tool use, CLI agents (Like Claude Code and OpenCode) & HTML/SVG generation, all with 128K context, now on Hugging Face! 
23
 
24
  ## 7B Models
25
 
@@ -48,10 +50,14 @@ library_name: transformers
48
  | IQuest-Coder-V1-40B-Instruct | [🤗 Hugging Face](https://huggingface.co/IQuestLab/IQuest-Coder-V1-40B-Instruct) |
49
  | IQuest-Coder-V1-40B-Loop-Instruct | [🤗 Hugging Face](https://huggingface.co/IQuestLab/IQuest-Coder-V1-40B-Loop-Instruct) |
50
  | IQuest-Coder-V1-40B-Thinking | [🤗 Hugging Face](https://huggingface.co/IQuestLab/IQuest-Coder-V1-40B-Thinking) |
 
51
 
52
  ## Sampling Parameters:
53
  For the IQuest-Coder-V1-Instruct: We suggest using Temperature=0.6, TopP=0.85, TopK=20.
54
 
 
 
 
55
  ## IQuest-Coder-V1 Highlights
56
 
57
  IQuest-Coder-V1 is a new family of code large language models (LLMs) designed to advance autonomous software engineering and code intelligence. Built on the innovative code-flow multi-stage training paradigm, IQuest-Coder-V1 captures the dynamic evolution of software logic, delivering state-of-the-art performance across critical dimensions:
@@ -63,6 +69,8 @@ IQuest-Coder-V1 is a new family of code large language models (LLMs) designed to
63
  - **Native Long Context**: All models natively support up to 128K tokens without requiring additional scaling techniques.
64
  - **CLI Agent Integration**: Demonstrates initial deployment capabilities on ClaudeCode and OpenCode platforms, with the ability to integrate into CLI-based agent workflows.
65
  - **HTML and SVG Generation**: Features preliminary support for HTML and SVG code generation.
 
 
66
 
67
  ## Model Overview
68
 
@@ -156,13 +164,13 @@ For Thinking models with reasoning support:
156
  vllm serve IQuestLab/IQuest-Coder-V1-40B-Thinking --reasoning-parser qwen3 --tensor-parallel-size 8
157
  ```
158
 
159
- When using tool, `IQuest-Coder-V1-40B-Instruct` and `IQuest-Coder-V1-40B-Loop-Instruct` should use `--tool-parser qwen3`, while `IQuest-Coder-V1-7B-Instruct`, `IQuest-Coder-V1-7B-Thinking`, `IQuest-Coder-V1-14B-Instruct`, `IQuest-Coder-V1-14B-Thinking` and `IQuest-Coder-V1-40B-Thinking` should use `--tool-parser qwen3_coder`.
160
 
161
  ### CLI-Like Agents and Tools Usage
162
 
163
- CLI-like agent capabilities are available for the following models: `IQuest-Coder-V1-7B-Instruct`, `IQuest-Coder-V1-7B-Thinking`, `IQuest-Coder-V1-14B-Instruct`, `IQuest-Coder-V1-14B-Thinking` and `IQuest-Coder-V1-40B-Thinking`.
164
 
165
- **Step 1:**: Deploy the model with vLLM and set tool parser (**Attention: Do not set reasoning parser for Instruct LLMs, otherwise it will cause unexpected errors**):
166
 
167
  ```bash
168
  vllm serve IQuestLab/IQuest-Coder-V1-7B-Instruct --tool-parser qwen3_coder
@@ -174,7 +182,7 @@ or
174
  vllm serve IQuestLab/IQuest-Coder-V1-7B-Thinking --tool-parser qwen3_coder --reasoning-parser qwen3
175
  ```
176
 
177
- **Step 2:**: Use Claude Code to enjoy it:
178
 
179
  ```bash
180
  export ANTHROPIC_BASE_URL="http://iquestcoder.link"
@@ -183,10 +191,10 @@ claude --model IQuestCoder-V1-7B-Instruct
183
  ```
184
 
185
 
 
186
 
 
187
 
188
-
189
- ## Evaluation Results
190
  ![Evaluation Results](./papers/results.png)
191
 
192
  ### Benchmark Parameters
@@ -198,7 +206,7 @@ claude --model IQuestCoder-V1-7B-Instruct
198
  | **BigCodeBench** | 0.0 | - |
199
  | **FullStackBench** | 0.0 | - |
200
  | **CruxEval** | 0.0 | - |
201
- | **LiveCodeBench** | 0.6 | 0.95 |
202
  | **Aider-Polyglot** | 0.95 | 0.85 |
203
  | **Mercury** | 0.2 | 0.85 |
204
  | **Bird** | 0.2 | 0.95 |
 
12
  ![Evaluation Results](./papers/iquest-coder-v1-logo.png)
13
 
14
  <p align="center">
15
+ 📘 <a href="https://iquestlab.github.io">Blog (2026-01-01)</a >
16
+ &nbsp;•&nbsp;
17
+ 📘 <a href="https://iquestlab.github.io">Blog (2026-03-02)</a >
18
  &nbsp;•&nbsp;
19
  📄 <a href="https://github.com/IQuestLab/IQuest-Coder-V1/blob/main/papers/IQuest_Coder_Technical_Report.pdf">Technical Report</a >
20
  </p >
21
 
22
+ # IQuest-Coder-V1 Model Family Update
23
 
24
+ 🚀🚀🚀 [IQuest-Coder-V1 Model Family Update](https://iquestlab.github.io/release-1.0-2602/index.html): Released 7B & 14B Family Models, 40B-Thinking and 40B-Loop-Thinking, specially optimized for tool use, CLI agents (Like `Claude Code` and `OpenCode`) & HTML/SVG generation, all with 128K context, now on Hugging Face!
25
 
26
  ## 7B Models
27
 
 
50
  | IQuest-Coder-V1-40B-Instruct | [🤗 Hugging Face](https://huggingface.co/IQuestLab/IQuest-Coder-V1-40B-Instruct) |
51
  | IQuest-Coder-V1-40B-Loop-Instruct | [🤗 Hugging Face](https://huggingface.co/IQuestLab/IQuest-Coder-V1-40B-Loop-Instruct) |
52
  | IQuest-Coder-V1-40B-Thinking | [🤗 Hugging Face](https://huggingface.co/IQuestLab/IQuest-Coder-V1-40B-Thinking) |
53
+ | IQuest-Coder-V1-40B-Loop-Thinking | [🤗 Hugging Face](https://huggingface.co/IQuestLab/IQuest-Coder-V1-40B-Loop-Thinking) |
54
 
55
  ## Sampling Parameters:
56
  For the IQuest-Coder-V1-Instruct: We suggest using Temperature=0.6, TopP=0.85, TopK=20.
57
 
58
+ For the IQuest-Coder-V1-Thinking: We suggest using Temperature=1.0, TopP=0.95, TopK=20.
59
+
60
+
61
  ## IQuest-Coder-V1 Highlights
62
 
63
  IQuest-Coder-V1 is a new family of code large language models (LLMs) designed to advance autonomous software engineering and code intelligence. Built on the innovative code-flow multi-stage training paradigm, IQuest-Coder-V1 captures the dynamic evolution of software logic, delivering state-of-the-art performance across critical dimensions:
 
69
  - **Native Long Context**: All models natively support up to 128K tokens without requiring additional scaling techniques.
70
  - **CLI Agent Integration**: Demonstrates initial deployment capabilities on ClaudeCode and OpenCode platforms, with the ability to integrate into CLI-based agent workflows.
71
  - **HTML and SVG Generation**: Features preliminary support for HTML and SVG code generation.
72
+ - **Architectural Chain-of-Thought via Recurrent Depth**: 40B-Loop-Thinking is a research-oriented, experimental model prototype designed to explore how structural chains of thought and procedural chains of thought can be combined within a single system. The model uniquely integrates structural chains of thought—realized through loop-based computation enabled by the dual-iteration LoopCoder architecture—with procedural chains of thought derived from explicit reasoning trajectories trained via reinforcement learning. Unlike standard reasoning models that rely solely on token-level chain-of-thought expansion, Loop-Thinking introduces implicit multi-step computation at the architectural level through a looped Transformer design. In this design, the second iteration refines the hidden states produced by the first iteration using a global–local attention gating mechanism. This results in a nested reasoning mechanism: the loop structure supports iterative representation refinement, while the reasoning-oriented training paradigm injects explicit problem decomposition behavior. It is important to note that this model is not intended to achieve state-of-the-art performance across benchmarks, but rather to validate the complementary roles of loop-based computation and reasoning-oriented training in shaping reasoning structures, and to provide experimental evidence for future model design.
73
+
74
 
75
  ## Model Overview
76
 
 
164
  vllm serve IQuestLab/IQuest-Coder-V1-40B-Thinking --reasoning-parser qwen3 --tensor-parallel-size 8
165
  ```
166
 
167
+ When using tool, `IQuest-Coder-V1-40B-Instruct` and `IQuest-Coder-V1-40B-Loop-Instruct` should use `--tool-parser qwen3`, while `IQuest-Coder-V1-7B-Instruct`, `IQuest-Coder-V1-7B-Thinking`, `IQuest-Coder-V1-14B-Instruct`, `IQuest-Coder-V1-14B-Thinking`, `IQuest-Coder-V1-40B-Thinking` and `IQuest-Coder-V1-40B-Loop-Thinking` should use `--tool-parser qwen3_coder`.
168
 
169
  ### CLI-Like Agents and Tools Usage
170
 
171
+ CLI-like agent capabilities are available for the following models: `IQuest-Coder-V1-7B-Instruct`, `IQuest-Coder-V1-7B-Thinking`, `IQuest-Coder-V1-14B-Instruct`, `IQuest-Coder-V1-14B-Thinking`, `IQuest-Coder-V1-40B-Thinking` and `IQuest-Coder-V1-40B-Loop-Thinking`.
172
 
173
+ **Step 1:** Deploy the model with vLLM and set tool parser (**Attention: Do not set reasoning parser for Instruct LLMs, otherwise it will cause unexpected errors**):
174
 
175
  ```bash
176
  vllm serve IQuestLab/IQuest-Coder-V1-7B-Instruct --tool-parser qwen3_coder
 
182
  vllm serve IQuestLab/IQuest-Coder-V1-7B-Thinking --tool-parser qwen3_coder --reasoning-parser qwen3
183
  ```
184
 
185
+ **Step 2:** Use Claude Code to enjoy it:
186
 
187
  ```bash
188
  export ANTHROPIC_BASE_URL="http://iquestcoder.link"
 
191
  ```
192
 
193
 
194
+ ## Evaluation Results
195
 
196
+ ![Evaluation Results](./papers/results-20260302.png)
197
 
 
 
198
  ![Evaluation Results](./papers/results.png)
199
 
200
  ### Benchmark Parameters
 
206
  | **BigCodeBench** | 0.0 | - |
207
  | **FullStackBench** | 0.0 | - |
208
  | **CruxEval** | 0.0 | - |
209
+ | **LiveCodeBench** | 1.0 | 1.0 |
210
  | **Aider-Polyglot** | 0.95 | 0.85 |
211
  | **Mercury** | 0.2 | 0.85 |
212
  | **Bird** | 0.2 | 0.95 |