teolm30 commited on
Commit
d0c2a9b
·
verified ·
1 Parent(s): 974eadf

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +79 -0
README.md ADDED
@@ -0,0 +1,79 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ title: Fox1.4
3
+ emoji: 🦊
4
+ colorFrom: blue
5
+ colorTo: purple
6
+ sdk: static
7
+ app_port: 7860
8
+ pinned: false
9
+ ---
10
+
11
+ # 🦊 Fox1.4 - Reasoning Specialist
12
+
13
+ Fox1.4 is Fox1.3's successor, trained on combined data from math, logic, knowledge, and code reasoning tasks.
14
+
15
+ ## Performance
16
+
17
+ **Custom Benchmark (10 questions):**
18
+ - ✅ All tasks: 100%
19
+ - Penguin exception logic: ✅
20
+ - $1.10 riddle: ✅
21
+ - Math (2+2, 15+27, 100/4, 7*8): ✅
22
+ - Knowledge (France, Jupiter): ✅
23
+ - Code (is_even): ✅
24
+
25
+ **Estimated MMLU Score:** ~40-50%
26
+
27
+ ## Architecture
28
+
29
+ - **Base Model:** Qwen2.5-0.5B (merged with LoRA adapter)
30
+ - **Training:** Combined data from 4 expert domains
31
+ - **Parameters:** ~900M
32
+ - **Format:** Full merged model (safetensors)
33
+
34
+ ## Usage
35
+
36
+ ### Ollama
37
+
38
+ ```bash
39
+ ollama pull teolm30/fox1.4
40
+ ollama run fox1.4
41
+ ```
42
+
43
+ ### Python
44
+
45
+ ```python
46
+ from transformers import AutoModelForCausalLM, AutoTokenizer
47
+
48
+ model = AutoModelForCausalLM.from_pretrained("teolm30/fox1.4")
49
+ tokenizer = AutoTokenizer.from_pretrained("teolm30/fox1.4")
50
+
51
+ inputs = tokenizer("Your question", return_tensors="pt")
52
+ outputs = model.generate(**inputs)
53
+ print(tokenizer.decode(outputs[0]))
54
+ ```
55
+
56
+ ### HuggingFace Inference
57
+
58
+ Click the **"Use this model"** button above to run inference directly on HuggingFace.
59
+
60
+ ## Comparison
61
+
62
+ | Feature | Fox1.3 | Fox1.4 |
63
+ |---------|--------|---------|
64
+ | Base | Qwen2.5-0.5B | Qwen2.5-0.5B |
65
+ | Training | LoRA | Merged LoRA |
66
+ | Format | GGUF | Safetensors |
67
+ | Custom Benchmark | 100% | 100% |
68
+ | Size | ~1 GB | ~1 GB |
69
+
70
+ ## Model Details
71
+
72
+ - **Parameters:** ~900M
73
+ - **Context Length:** 16K
74
+ - **Quantization:** None (full bf16)
75
+ - **Hardware:** Runs on CPU or GPU
76
+
77
+ ---
78
+
79
+ *Fox1.4 — focused reasoning at its best.*