| --- |
| title: Fox1.4 |
| emoji: π¦ |
| colorFrom: blue |
| colorTo: purple |
| sdk: static |
| app_port: 7860 |
| pinned: false |
| license: apache-2.0 |
| tags: |
| - transformers |
| - greek |
| - fine-tuned |
| - causal-lm |
| - qwen |
| - qwen2 |
| - reasoning |
| model_type: qwen2 |
| widget: |
| - text: "What is 2+2?" |
| - text: "Solve this riddle: I have hands but cannot clap" |
| - text: "Write python code to check if a number is prime" |
| inference: |
| minutes: 10 |
| --- |
| |
| # π¦ Fox1.4 - Reasoning Specialist |
|
|
| Fox1.4 is Fox1.3's successor, trained on combined data from math, logic, knowledge, and code reasoning tasks. |
|
|
| ## Performance |
|
|
| **Custom Benchmark (10 questions):** |
| - β
All tasks: 100% |
| - Penguin exception logic: β
|
| - $1.10 riddle: β
|
| - Math (2+2, 15+27, 100/4, 7*8): β
|
| - Knowledge (France, Jupiter): β
|
| - Code (is_even): β
|
| |
| **Estimated MMLU Score:** ~40-50% |
| |
| ## Architecture |
| |
| - **Base Model:** Qwen2.5-0.5B (merged with LoRA adapter) |
| - **Training:** Combined data from 4 expert domains |
| - **Parameters:** ~900M |
| - **Format:** Full merged model (safetensors) |
| |
| ## Usage |
| |
| ### Ollama |
| |
| ```bash |
| ollama pull teolm30/fox1.4 |
| ollama run fox1.4 |
| ``` |
| |
| ### Python |
| |
| ```python |
| from transformers import AutoModelForCausalLM, AutoTokenizer |
| |
| model = AutoModelForCausalLM.from_pretrained("teolm30/fox1.4") |
| tokenizer = AutoTokenizer.from_pretrained("teolm30/fox1.4") |
| |
| inputs = tokenizer("What is 2+2?", return_tensors="pt") |
| output = model.generate(**inputs, max_new_tokens=50) |
| print(tokenizer.decode(output[0])) |
| ``` |
| |