File size: 4,765 Bytes
bc24ac8
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0a0023a
bc24ac8
 
 
 
 
 
 
 
 
 
0144546
bc24ac8
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
---
license: apache-2.0
language:
- en
tags:
- code
- coding
- programming
- reasoning
- small-model
- efficient
- local
- qwen
- qwen3
- qwen3.5
- 4b
- small
- developer
- coding-assistant
- python
- debugging
- daily-use
- localai
- ai
- gpt
- dqnlabs
- dqngpt
- gguf
- lmstudio
- ollama
pipeline_tag: text-generation
---

# dqnCode-v1

dqnCode-v1 is a 4B-parameter language model designed for fast, clear, and practical coding assistance.

It focuses on writing, fixing, and explaining code efficiently, with minimal verbosity and strong real-world usefulness. It is optimized for everyday programming tasks with low latency and concise outputs.

![dqnCode Banner](dqnCode.png)
---

## Benchmark

dqnCode-v1 is positioned as a high-performance compact coding model, with strong results on standard code generation benchmarks. It is trained with simple prompts in mind, so you don't need to be a developer to use it!

### HumanEval

- **pass@1:** 63.4%  

This score places dqnCode-v1 among the strongest models in the 4B parameter class for coding tasks (only beaten by one other model in the 4B or below models class!)

| Model                     | Provider        | HumanEval (pass@1) |
|--------------------------|-----------------|--------------------|
| GPT-3.5 Turbo            | OpenAI          | 68%                |
| GPT-4                    | OpenAI          | 67%                |
| dqnCode v1 (4B)          | DQN Labs        | 63.4%              |
| Phi-3.5-mini-instruct    | Microsoft       | 62.8%              |
| DeepSeek Coder 33B       | DeepSeek        | 52.4%              |
| Gemma 2 27B              | Google          | 51.8%              |
| Nous Hermes 3 405B       | Nous Research   | 51.4%              |
---

## Benchmark Context

- Evaluated on HumanEval (Python code generation benchmark)  
- Focused on functional correctness of generated code  
- Designed to reflect real-world coding performance in a compact model  

---

## Positioning

dqnCode-v1 is designed to deliver coding performance comparable to significantly larger models while maintaining a small footprint.

- Approaches GPT-3.5 / GPT-4-level performance on many practical coding tasks  
- Achieves strong performance with significantly fewer parameters  
- Optimized for fast local inference and real-world development workflows
- Excellent at all languages, especially HTML-CSS-JS, Python, Java, and more!


## Model Description

- **Model type:** Causal Language Model  
- **Parameters:** 4B  
- **Primary use:** Coding and programming assistance  
- **Style:** Direct code generation with minimal explanation  

dqnCode-v1 4B is optimized for practical coding workflows rather than long-form development or benchmark performance.

---

## Intended Uses

### Direct Use

- Writing code snippets  
- Debugging errors  
- Explaining code logic  
- Converting code between languages  
- Assisting with small to medium programming tasks  
- Competitive programming support  

---

## Key Characteristics

- Produces clean and functional code  
- Prioritizes correctness and simplicity  
- Minimizes unnecessary explanation  
- Strong at debugging and fixing issues  
- Designed for consistent outputs across common coding tasks  
- Excellent instruction following and stability.
- Trained with simple prompts in mind, so you don't need to be a developer to use it!

---

## Example

**Input**
```text
Write a Python function to check if a number is prime
```

**Output**
```python
def is_prime(n):
    if n < 2:
        return False
    for i in range(2, int(n**0.5) + 1):
        if n % i == 0:
            return False
    return True
```

---

**Input**
```text
Fix this code:
for i in range(5)
    print(i)
```

**Output**
```python
for i in range(5):
    print(i)
```

---

## Usage

This model is available on many platforms and is compatible with many formats.

The GGUF format is compatible with llama.cpp and LM Studio.  
Other formats include MLX (LM Studio, optimized for Apple devices), and HF (universal compatibility).

---

## Training Details

dqnCode-v1 is fine-tuned for practical coding tasks and efficient problem solving.

The training process emphasizes:

- Functional correctness  
- Minimal and clean outputs  
- Real-world coding scenarios  
- Debugging and code repair  

---

## Limitations

- Limited performance on very large or complex codebases  
- Not optimized for long-form software architecture design  
- May simplify explanations rather than provide deep theoretical detail  

---

## Efficiency

dqnCode-v1 is designed to run efficiently on consumer hardware, with support for quantized formats.

---

## License

Apache 2.0

---

## Author

Developed by DQN Labs.  
This model card was generated with the help of dqnGPT v0.2!