File size: 5,743 Bytes
4c5b555 576b487 4c5b555 cd54f7a 98a58f1 7e56751 98a58f1 7e56751 98a58f1 ab5c531 98a58f1 ab5c531 4c5b555 98a58f1 45f88b7 4c5b555 576b487 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 | ---
base_model: unsloth/Qwen3-14B-unsloth-bnb-4bit
tags:
- text-generation-inference
- transformers
- unsloth
- qwen3
- gguf
license: apache-2.0
language:
- en
datasets:
- mugivara1/convex-reasoning-new-train
- mugivara1/convex-reasoning
---
# Qwisine 14B
<!-- Smaller mascot image -->
<div align="center">
<img src="https://i.imgur.com/JbfjKSy.png" alt="Qwisine Mascot" style="width:140px; margin-bottom:20px;" />
</div>
<!-- Banner with logo and evaluation chart -->
<div align="center">
<img src="https://i.imgur.com/AxRWCK3.png" alt="Qwisine Banner" style="width:100%; height:350px; object-fit: cover; margin-bottom:20px;" />
</div>
---
## Model details
| Field | Description |
| ----------------- | --------------------------------------------------------------------------------------------------------------- |
| **Base model** | [Qwen‑3-14B](https://huggingface.co/Qwen/Qwen3-14B) (pre‑trained) |
| **Fine‑tuned by** | *Mugi* |
| **Task** | Question‑Answering & Code Generation for the [Convex](https://convex.dev) TypeScript backend/database framework |
| **Language(s)** | English (developer‑oriented) |
| **License** | *NAH just use it.* |
| **Model name** | **Qwisine** |
Qwisine is a specialised version of Qwen‑3 fine‑tuned on curated Convex documentation & synthethic code and community Q\&A. The model understands Convex‐specific concepts (data modelling, mutations, actions, idioms, client usage, etc.) and can generate code snippets or explain behaviour in plain English.
---
## Intended use & limitations
**Primary use‑case**
* Conversational assistant for developers building on Convex.
* Drafting / Helping with convex orientated questions & tasks.
* Documentation chatbots or support assistants.
**Out‑of‑scope**
* Production‑critical decision making without human review.
---
## Dataset
* **Size** : 938 Q\&A pairs
* **Source**: Convex official docs, example apps, public issues, community Discord, and synthetic edge‑cases.
* **Question types** (distilled)
* `what_is` – factual look‑ups (no reasoning)
* `why` – causal explanations (no reasoning)
* `task` – recipe‑style how‑to (with reasoning)
* `edge_case` – tricky or undocumented scenarios (with reasoning)
* `v‑task` – verbose multi‑step tasks (with reasoning)
Reasoning‑bearing examples represent \~85 % of the dataset.
---
## Training procedure -- will add later since i ran & experimented MANY RUNS 😭😭😭😭
* **Epochs** : **
* **Batch** : **
* **LR / schedule** : **
* **Optimizer** : **
Fine‑tuning followed standard QLORA with unsloth. No additional RLHF was applied.
---
## Evaluation results
| Category | **Think** mode | Fully **Non‑Think** mode |
| -------------- | --------------------------- | ------------------------ |
| Fundamentals | **75.05 %** | 73.44 % |
| Data modelling | **82.82 %** | **87.36 %** |
| Queries | 74.38 % | 74.19 % |
| Mutations | 71.04 % | 73.59 % |
| Actions | 63.05 % | **49.27 %** |
| Idioms | 75.06 % | 75.06 % |
| Clients | 69.84 % | 69.84 % |
| **Average** | **73.03 %** | 71.82 % |
---
### Think Mode
| Parameter | Value | Notes |
| ------------- | ----- | ------------------------------- |
| `temperature` | 0.6 | Reasoned answers with structure |
| `top_p` | 0.95 | Wider beam of sampling |
| `top_k` | 20 | |
| `min_p` | 0 | |
### Non-Think Mode
| Parameter | Value | Notes |
| ------------- | ----- | --------------------------------- |
| `temperature` | 0.7 | More diversity for simple prompts |
| `top_p` | 0.8 | Slightly tighter sampling |
| `top_k` | 20 | |
| `min_p` | 0 | |
<sub>*Adjust as needed for your deployment; these were used in LM Studio during evaluation.*</sub>
---
## How to run locally
```bash
# LM Studio
search "Qwisine" in models menu.
# Ollama
il add soon.
# Llama‑cpp
il add soon.
```
---
## Limitations & biases
* Training data is entirely Convex‑centred; the model may hallucinate.
* The dataset size is modest (938 samples); edge‑case coverage is still incomplete and so is more complex prompts like create project from scratch with multiple steps and instructions.
---
## Future work
*not sure yet*
---
## Citation
```bibtex
@misc{qwisine2025,
title = {Qwisine: A Qwen‑3 model fine‑tuned for Convex},
author = {mugi},
year = {2025},
url = {https://huggingface.co/mugivara1/Qwisine},
}
```
---
## Acknowledgements
*(To be completed)*
Convex • Qwen‑3 • ... |