Transformers
GGUF
English
qwen3
text-generation-inference
unsloth
conversational
File size: 5,743 Bytes
4c5b555
 
 
 
 
 
 
 
 
 
 
576b487
 
 
4c5b555
cd54f7a
98a58f1
7e56751
98a58f1
7e56751
 
98a58f1
ab5c531
98a58f1
ab5c531
4c5b555
98a58f1
 
45f88b7
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4c5b555
576b487
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
---
base_model: unsloth/Qwen3-14B-unsloth-bnb-4bit
tags:
- text-generation-inference
- transformers
- unsloth
- qwen3
- gguf
license: apache-2.0
language:
- en
datasets:
- mugivara1/convex-reasoning-new-train
- mugivara1/convex-reasoning
---
# Qwisine 14B
<!-- Smaller mascot image -->
<div align="center">
  <img src="https://i.imgur.com/JbfjKSy.png" alt="Qwisine Mascot" style="width:140px; margin-bottom:20px;" />
</div>

<!-- Banner with logo and evaluation chart -->
<div align="center">
  <img src="https://i.imgur.com/AxRWCK3.png" alt="Qwisine Banner" style="width:100%; height:350px; object-fit: cover; margin-bottom:20px;" />
</div>



---

## Model details

| Field             | Description                                                                                                     |
| ----------------- | --------------------------------------------------------------------------------------------------------------- |
| **Base model**    | [Qwen‑3-14B](https://huggingface.co/Qwen/Qwen3-14B) (pre‑trained)                                                                  |
| **Fine‑tuned by** | *Mugi*                                                                                         |
| **Task**          | Question‑Answering & Code Generation for the [Convex](https://convex.dev) TypeScript backend/database framework |
| **Language(s)**   | English (developer‑oriented)                                                                                    |
| **License**       | *NAH just use it.*                                                                                              |
| **Model name**    | **Qwisine**                                                                                                     |

Qwisine is a specialised version of Qwen‑3 fine‑tuned on curated Convex documentation & synthethic code and community Q\&A. The model understands Convex‐specific concepts (data modelling, mutations, actions, idioms, client usage, etc.) and can generate code snippets or explain behaviour in plain English.

---

## Intended use & limitations

**Primary use‑case**

* Conversational assistant for developers building on Convex.
* Drafting / Helping with convex orientated questions & tasks.
* Documentation chatbots or support assistants.

**Out‑of‑scope**

* Production‑critical decision making without human review.

---

## Dataset

* **Size**  : 938 Q\&A pairs
* **Source**: Convex official docs, example apps, public issues, community Discord, and synthetic edge‑cases.
* **Question types** (distilled)

  * `what_is`   – factual look‑ups (no reasoning)
  * `why`       – causal explanations (no reasoning)
  * `task`      – recipe‑style how‑to (with reasoning)
  * `edge_case` – tricky or undocumented scenarios (with reasoning)
  * `v‑task`    – verbose multi‑step tasks (with reasoning)

Reasoning‑bearing examples represent \~85 % of the dataset.

---

## Training procedure -- will add later since i ran & experimented MANY RUNS 😭😭😭😭

* **Epochs**  : ** 
* **Batch**   : **
* **LR / schedule** : **
* **Optimizer** : **

Fine‑tuning followed standard QLORA with unsloth. No additional RLHF was applied.

---

## Evaluation results

| Category       | **Think** mode              | Fully **Non‑Think** mode |
| -------------- | --------------------------- | ------------------------ |
| Fundamentals   | **75.05 %**                 | 73.44 %                  |
| Data modelling | **82.82 %**                 | **87.36 %**              |
| Queries        | 74.38 %                     | 74.19 %                  |
| Mutations      | 71.04 %                     | 73.59 %                  |
| Actions        | 63.05 %                     | **49.27 %**              |
| Idioms         | 75.06 %                     | 75.06 %                  |
| Clients        | 69.84 %                     | 69.84 %                  |
| **Average**    | **73.03 %**                 | 71.82 %                  |

---

### Think Mode

| Parameter     | Value | Notes                           |
| ------------- | ----- | ------------------------------- |
| `temperature` | 0.6   | Reasoned answers with structure |
| `top_p`       | 0.95  | Wider beam of sampling          |
| `top_k`       | 20    |                                 |
| `min_p`       | 0     |                                 |

### Non-Think Mode

| Parameter     | Value | Notes                             |
| ------------- | ----- | --------------------------------- |
| `temperature` | 0.7   | More diversity for simple prompts |
| `top_p`       | 0.8   | Slightly tighter sampling         |
| `top_k`       | 20    |                                   |
| `min_p`       | 0     |                                   |

<sub>*Adjust as needed for your deployment; these were used in LM Studio during evaluation.*</sub>

---

## How to run locally

```bash
# LM Studio
search "Qwisine" in models menu.

# Ollama
il add soon.
# Llama‑cpp
il add soon.
```

---

## Limitations & biases

* Training data is entirely Convex‑centred; the model may hallucinate.
* The dataset size is modest (938 samples); edge‑case coverage is still incomplete and so is more complex prompts like create project from scratch with multiple steps and instructions.

---

## Future work

*not sure yet*

---

## Citation

```bibtex
@misc{qwisine2025,
  title        = {Qwisine: A Qwen‑3 model fine‑tuned for Convex},
  author       = {mugi},
  year         = {2025},
  url          = {https://huggingface.co/mugivara1/Qwisine},
}
```

---

## Acknowledgements

*(To be completed)*

Convex • Qwen‑3 • ...