zwpride-iquestlab commited on
Commit
7b6ee5a
·
verified ·
1 Parent(s): f8013f1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +18 -10
README.md CHANGED
@@ -11,14 +11,16 @@ library_name: transformers
11
  ![Evaluation Results](./papers/iquest-coder-v1-logo.png)
12
 
13
  <p align="center">
14
- 📘 <a href="https://iquestlab.github.io">Blog</a >
 
 
15
  &nbsp;•&nbsp;
16
  📄 <a href="https://github.com/IQuestLab/IQuest-Coder-V1/blob/main/papers/IQuest_Coder_Technical_Report.pdf">Technical Report</a >
17
  </p >
18
 
19
- # IQuest-Coder-V1 Model Family
20
 
21
- 🚀 **[[IQuest-Coder-V1 Update](https://iquestlab.github.io/release-1.0-2602/index.html)]**: Released 7B & 14B Family Models and 40B-Thinking, specially optimized for tool use, CLI agents (Like Claude Code and OpenCode) & HTML/SVG generation, all with 128K context, now on Hugging Face! 
22
 
23
  ## 7B Models
24
 
@@ -47,10 +49,14 @@ library_name: transformers
47
  | IQuest-Coder-V1-40B-Instruct | [🤗 Hugging Face](https://huggingface.co/IQuestLab/IQuest-Coder-V1-40B-Instruct) |
48
  | IQuest-Coder-V1-40B-Loop-Instruct | [🤗 Hugging Face](https://huggingface.co/IQuestLab/IQuest-Coder-V1-40B-Loop-Instruct) |
49
  | IQuest-Coder-V1-40B-Thinking | [🤗 Hugging Face](https://huggingface.co/IQuestLab/IQuest-Coder-V1-40B-Thinking) |
 
50
 
51
  ## Sampling Parameters:
52
  For the IQuest-Coder-V1-Instruct: We suggest using Temperature=0.6, TopP=0.85, TopK=20.
53
 
 
 
 
54
  ## IQuest-Coder-V1 Highlights
55
 
56
  IQuest-Coder-V1 is a new family of code large language models (LLMs) designed to advance autonomous software engineering and code intelligence. Built on the innovative code-flow multi-stage training paradigm, IQuest-Coder-V1 captures the dynamic evolution of software logic, delivering state-of-the-art performance across critical dimensions:
@@ -62,6 +68,8 @@ IQuest-Coder-V1 is a new family of code large language models (LLMs) designed to
62
  - **Native Long Context**: All models natively support up to 128K tokens without requiring additional scaling techniques.
63
  - **CLI Agent Integration**: Demonstrates initial deployment capabilities on ClaudeCode and OpenCode platforms, with the ability to integrate into CLI-based agent workflows.
64
  - **HTML and SVG Generation**: Features preliminary support for HTML and SVG code generation.
 
 
65
 
66
  ## Model Overview
67
 
@@ -155,13 +163,13 @@ For Thinking models with reasoning support:
155
  vllm serve IQuestLab/IQuest-Coder-V1-40B-Thinking --reasoning-parser qwen3 --tensor-parallel-size 8
156
  ```
157
 
158
- When using tool, `IQuest-Coder-V1-40B-Instruct` and `IQuest-Coder-V1-40B-Loop-Instruct` should use `--tool-parser qwen3`, while `IQuest-Coder-V1-7B-Instruct`, `IQuest-Coder-V1-7B-Thinking`, `IQuest-Coder-V1-14B-Instruct`, `IQuest-Coder-V1-14B-Thinking` and `IQuest-Coder-V1-40B-Thinking` should use `--tool-parser qwen3_coder`.
159
 
160
  ### CLI-Like Agents and Tools Usage
161
 
162
- CLI-like agent capabilities are available for the following models: `IQuest-Coder-V1-7B-Instruct`, `IQuest-Coder-V1-7B-Thinking`, `IQuest-Coder-V1-14B-Instruct`, `IQuest-Coder-V1-14B-Thinking` and `IQuest-Coder-V1-40B-Thinking`.
163
 
164
- **Step 1:**: Deploy the model with vLLM and set tool parser (**Attention: Do not set reasoning parser for Instruct LLMs, otherwise it will cause unexpected errors**):
165
 
166
  ```bash
167
  vllm serve IQuestLab/IQuest-Coder-V1-7B-Instruct --tool-parser qwen3_coder
@@ -173,7 +181,7 @@ or
173
  vllm serve IQuestLab/IQuest-Coder-V1-7B-Thinking --tool-parser qwen3_coder --reasoning-parser qwen3
174
  ```
175
 
176
- **Step 2:**: Use Claude Code to enjoy it:
177
 
178
  ```bash
179
  export ANTHROPIC_BASE_URL="http://iquestcoder.link"
@@ -182,10 +190,10 @@ claude --model IQuestCoder-V1-7B-Instruct
182
  ```
183
 
184
 
 
185
 
 
186
 
187
-
188
- ## Evaluation Results
189
  ![Evaluation Results](./papers/results.png)
190
 
191
  ### Benchmark Parameters
@@ -197,7 +205,7 @@ claude --model IQuestCoder-V1-7B-Instruct
197
  | **BigCodeBench** | 0.0 | - |
198
  | **FullStackBench** | 0.0 | - |
199
  | **CruxEval** | 0.0 | - |
200
- | **LiveCodeBench** | 0.6 | 0.95 |
201
  | **Aider-Polyglot** | 0.95 | 0.85 |
202
  | **Mercury** | 0.2 | 0.85 |
203
  | **Bird** | 0.2 | 0.95 |
 
11
  ![Evaluation Results](./papers/iquest-coder-v1-logo.png)
12
 
13
  <p align="center">
14
+ 📘 <a href="https://iquestlab.github.io">Blog (2026-01-01)</a >
15
+ &nbsp;•&nbsp;
16
+ 📘 <a href="https://iquestlab.github.io">Blog (2026-03-02)</a >
17
  &nbsp;•&nbsp;
18
  📄 <a href="https://github.com/IQuestLab/IQuest-Coder-V1/blob/main/papers/IQuest_Coder_Technical_Report.pdf">Technical Report</a >
19
  </p >
20
 
21
+ # IQuest-Coder-V1 Model Family Update
22
 
23
+ 🚀🚀🚀 [IQuest-Coder-V1 Model Family Update](https://iquestlab.github.io/release-1.0-2602/index.html): Released 7B & 14B Family Models, 40B-Thinking and 40B-Loop-Thinking, specially optimized for tool use, CLI agents (Like `Claude Code` and `OpenCode`) & HTML/SVG generation, all with 128K context, now on Hugging Face!
24
 
25
  ## 7B Models
26
 
 
49
  | IQuest-Coder-V1-40B-Instruct | [🤗 Hugging Face](https://huggingface.co/IQuestLab/IQuest-Coder-V1-40B-Instruct) |
50
  | IQuest-Coder-V1-40B-Loop-Instruct | [🤗 Hugging Face](https://huggingface.co/IQuestLab/IQuest-Coder-V1-40B-Loop-Instruct) |
51
  | IQuest-Coder-V1-40B-Thinking | [🤗 Hugging Face](https://huggingface.co/IQuestLab/IQuest-Coder-V1-40B-Thinking) |
52
+ | IQuest-Coder-V1-40B-Loop-Thinking | [🤗 Hugging Face](https://huggingface.co/IQuestLab/IQuest-Coder-V1-40B-Loop-Thinking) |
53
 
54
  ## Sampling Parameters:
55
  For the IQuest-Coder-V1-Instruct: We suggest using Temperature=0.6, TopP=0.85, TopK=20.
56
 
57
+ For the IQuest-Coder-V1-Thinking: We suggest using Temperature=1.0, TopP=0.95, TopK=20.
58
+
59
+
60
  ## IQuest-Coder-V1 Highlights
61
 
62
  IQuest-Coder-V1 is a new family of code large language models (LLMs) designed to advance autonomous software engineering and code intelligence. Built on the innovative code-flow multi-stage training paradigm, IQuest-Coder-V1 captures the dynamic evolution of software logic, delivering state-of-the-art performance across critical dimensions:
 
68
  - **Native Long Context**: All models natively support up to 128K tokens without requiring additional scaling techniques.
69
  - **CLI Agent Integration**: Demonstrates initial deployment capabilities on ClaudeCode and OpenCode platforms, with the ability to integrate into CLI-based agent workflows.
70
  - **HTML and SVG Generation**: Features preliminary support for HTML and SVG code generation.
71
+ - **Architectural Chain-of-Thought via Recurrent Depth**: 40B-Loop-Thinking is a research-oriented, experimental model prototype designed to explore how structural chains of thought and procedural chains of thought can be combined within a single system. The model uniquely integrates structural chains of thought—realized through loop-based computation enabled by the dual-iteration LoopCoder architecture—with procedural chains of thought derived from explicit reasoning trajectories trained via reinforcement learning. Unlike standard reasoning models that rely solely on token-level chain-of-thought expansion, Loop-Thinking introduces implicit multi-step computation at the architectural level through a looped Transformer design. In this design, the second iteration refines the hidden states produced by the first iteration using a global–local attention gating mechanism. This results in a nested reasoning mechanism: the loop structure supports iterative representation refinement, while the reasoning-oriented training paradigm injects explicit problem decomposition behavior. It is important to note that this model is not intended to achieve state-of-the-art performance across benchmarks, but rather to validate the complementary roles of loop-based computation and reasoning-oriented training in shaping reasoning structures, and to provide experimental evidence for future model design.
72
+
73
 
74
  ## Model Overview
75
 
 
163
  vllm serve IQuestLab/IQuest-Coder-V1-40B-Thinking --reasoning-parser qwen3 --tensor-parallel-size 8
164
  ```
165
 
166
+ When using tool, `IQuest-Coder-V1-40B-Instruct` and `IQuest-Coder-V1-40B-Loop-Instruct` should use `--tool-parser qwen3`, while `IQuest-Coder-V1-7B-Instruct`, `IQuest-Coder-V1-7B-Thinking`, `IQuest-Coder-V1-14B-Instruct`, `IQuest-Coder-V1-14B-Thinking`, `IQuest-Coder-V1-40B-Thinking` and `IQuest-Coder-V1-40B-Loop-Thinking` should use `--tool-parser qwen3_coder`.
167
 
168
  ### CLI-Like Agents and Tools Usage
169
 
170
+ CLI-like agent capabilities are available for the following models: `IQuest-Coder-V1-7B-Instruct`, `IQuest-Coder-V1-7B-Thinking`, `IQuest-Coder-V1-14B-Instruct`, `IQuest-Coder-V1-14B-Thinking`, `IQuest-Coder-V1-40B-Thinking` and `IQuest-Coder-V1-40B-Loop-Thinking`.
171
 
172
+ **Step 1:** Deploy the model with vLLM and set tool parser (**Attention: Do not set reasoning parser for Instruct LLMs, otherwise it will cause unexpected errors**):
173
 
174
  ```bash
175
  vllm serve IQuestLab/IQuest-Coder-V1-7B-Instruct --tool-parser qwen3_coder
 
181
  vllm serve IQuestLab/IQuest-Coder-V1-7B-Thinking --tool-parser qwen3_coder --reasoning-parser qwen3
182
  ```
183
 
184
+ **Step 2:** Use Claude Code to enjoy it:
185
 
186
  ```bash
187
  export ANTHROPIC_BASE_URL="http://iquestcoder.link"
 
190
  ```
191
 
192
 
193
+ ## Evaluation Results
194
 
195
+ ![Evaluation Results](./papers/results-20260302.png)
196
 
 
 
197
  ![Evaluation Results](./papers/results.png)
198
 
199
  ### Benchmark Parameters
 
205
  | **BigCodeBench** | 0.0 | - |
206
  | **FullStackBench** | 0.0 | - |
207
  | **CruxEval** | 0.0 | - |
208
+ | **LiveCodeBench** | 1.0 | 1.0 |
209
  | **Aider-Polyglot** | 0.95 | 0.85 |
210
  | **Mercury** | 0.2 | 0.85 |
211
  | **Bird** | 0.2 | 0.95 |