o-duffy commited on
Commit
cd14514
·
verified ·
1 Parent(s): ee2cfa3

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +61 -0
README.md ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: Qwen/Qwen2.5-Coder-14B-Instruct
4
+ tags:
5
+ - unreal-engine
6
+ - ue5
7
+ - game-development
8
+ - gguf
9
+ - qwen2
10
+ language:
11
+ - en
12
+ pipeline_tag: text-generation
13
+ ---
14
+
15
+ # ue-expert-v2 (Qwen2.5-Coder-14B-Instruct + SFT)
16
+
17
+ Fine-tuned Unreal Engine 5 expert model, specialized in C++ and Blueprint development.
18
+
19
+ ## Model Details
20
+
21
+ - **Base model:** [Qwen2.5-Coder-14B-Instruct](https://huggingface.co/Qwen/Qwen2.5-Coder-14B-Instruct)
22
+ - **Fine-tuning:** QLoRA (rank 32, alpha 64) SFT on 27,738 curated UE5 Q&A pairs
23
+ - **Negative examples:** 9.3% of training data teaches the model to say "I don't know" for hallucinated/non-existent APIs
24
+ - **Quantization:** Q4_K_M (4.87 bits per weight)
25
+ - **Size:** ~8.4 GB GGUF
26
+
27
+ ## Training Data
28
+
29
+ - 27,738 training pairs covering UE5 C++ APIs, Blueprint patterns, architecture, and best practices
30
+ - 7 template categories for positive examples (hierarchy, API lookup, code patterns, etc.)
31
+ - Negative examples include fabricated class names, non-existent functions, and plausible-but-wrong API claims
32
+ - Quality-gated: all pairs scored >= 0.4 by automated quality pipeline
33
+
34
+ ## Usage with Ollama
35
+
36
+ ```bash
37
+ # Download the GGUF and Modelfile, then:
38
+ ollama create ue-expert -f Modelfile
39
+ ollama run ue-expert "What is the parent class of ACharacter?"
40
+ ```
41
+
42
+ ## Files
43
+
44
+ - `model-q4_k_m.gguf` — Quantized model (Q4_K_M, 8.4 GB)
45
+ - `sft_adapter/` — LoRA adapter weights (for further fine-tuning)
46
+ - `Modelfile` — Ollama model definition with ChatML template
47
+
48
+ ## Training Metrics
49
+
50
+ | Metric | Value |
51
+ |--------|-------|
52
+ | Steps | 1299 (3 epochs) |
53
+ | Final train loss | 0.668 |
54
+ | Final eval loss | 0.677 |
55
+ | Hardware | RunPod A100 SXM 80GB |
56
+ | Training time | ~3h 46m |
57
+ | VRAM usage | 15.8 GB / 80 GB |
58
+
59
+ ## Part of game-dev-docs
60
+
61
+ This model is the synthesis layer for a RAG + fine-tuned model + MCP server pipeline that provides deep Unreal Engine awareness to Claude Code. The RAG pipeline provides retrieval over 302K indexed documentation chunks; this model provides internalized domain knowledge for synthesis and judgment calls.