Transformers
GGUF
English
qwen3
text-generation-inference
unsloth
conversational

Qwisine 14B

Qwisine Mascot
Qwisine Banner

Model details

Field Description
Base model Qwenโ€‘3-14B (preโ€‘trained)
Fineโ€‘tuned by Mugi
Task Questionโ€‘Answering & Code Generation for the Convex TypeScript backend/database framework
Language(s) English (developerโ€‘oriented)
License NAH just use it.
Model name Qwisine

Qwisine is a specialised version of Qwenโ€‘3 fineโ€‘tuned on curated Convex documentation & synthethic code and community Q&A. The model understands Convexโ€specific concepts (data modelling, mutations, actions, idioms, client usage, etc.) and can generate code snippets or explain behaviour in plain English.


Intended use & limitations

Primary useโ€‘case

  • Conversational assistant for developers building on Convex.
  • Drafting / Helping with convex orientated questions & tasks.
  • Documentation chatbots or support assistants.

Outโ€‘ofโ€‘scope

  • Productionโ€‘critical decision making without human review.

Dataset

  • Size  : 938 Q&A pairs

  • Source: Convex official docs, example apps, public issues, community Discord, and synthetic edgeโ€‘cases.

  • Question types (distilled)

    • what_is   โ€“ factual lookโ€‘ups (no reasoning)
    • why       โ€“ causal explanations (no reasoning)
    • task      โ€“ recipeโ€‘style howโ€‘to (with reasoning)
    • edge_case โ€“ tricky or undocumented scenarios (with reasoning)
    • vโ€‘task    โ€“ verbose multiโ€‘step tasks (with reasoning)

Reasoningโ€‘bearing examples represent ~85โ€ฏ% of the dataset.


Training procedure -- will add later since i ran & experimented MANY RUNS ๐Ÿ˜ญ๐Ÿ˜ญ๐Ÿ˜ญ๐Ÿ˜ญ

  • Epochs  : **
  • Batch   : **
  • LR / schedule : **
  • Optimizer : **

Fineโ€‘tuning followed standard QLORA with unsloth. No additional RLHF was applied.


Evaluation results

Category Think mode Fully Nonโ€‘Think mode
Fundamentals 75.05โ€ฏ% 73.44โ€ฏ%
Data modelling 82.82โ€ฏ% 87.36โ€ฏ%
Queries 74.38โ€ฏ% 74.19โ€ฏ%
Mutations 71.04โ€ฏ% 73.59โ€ฏ%
Actions 63.05โ€ฏ% 49.27โ€ฏ%
Idioms 75.06โ€ฏ% 75.06โ€ฏ%
Clients 69.84โ€ฏ% 69.84โ€ฏ%
Average 73.03โ€ฏ% 71.82โ€ฏ%

Think Mode

Parameter Value Notes
temperature 0.6 Reasoned answers with structure
top_p 0.95 Wider beam of sampling
top_k 20
min_p 0

Non-Think Mode

Parameter Value Notes
temperature 0.7 More diversity for simple prompts
top_p 0.8 Slightly tighter sampling
top_k 20
min_p 0

Adjust as needed for your deployment; these were used in LM Studio during evaluation.


How to run locally

# LM Studio
search "Qwisine" in models menu.

# Ollama
il add soon.
# Llamaโ€‘cpp
il add soon.

Limitations & biases

  • Training data is entirely Convexโ€‘centred; the model may hallucinate.
  • The dataset size is modest (938 samples); edgeโ€‘case coverage is still incomplete and so is more complex prompts like create project from scratch with multiple steps and instructions.

Future work

not sure yet


Citation

@misc{qwisine2025,
  title        = {Qwisine: A Qwenโ€‘3 model fineโ€‘tuned for Convex},
  author       = {mugi},
  year         = {2025},
  url          = {https://huggingface.co/mugivara1/Qwisine},
}

Acknowledgements

(To be completed)

Convex โ€ข Qwenโ€‘3 โ€ขโ€ฏ...

Downloads last month
7
GGUF
Model size
15B params
Architecture
qwen3
Hardware compatibility
Log In to add your hardware

5-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for moogin/Qwisine

Finetuned
Qwen/Qwen3-14B
Quantized
(24)
this model