| # Qwen3-8B-Elizabeth-Simple | |
| A fine-tuned version of Qwen3-8B specifically optimized for tool use capabilities, trained on the Elizabeth tool use minipack. | |
| ## Model Details | |
| ### Base Model | |
| - **Model:** Qwen/Qwen3-8B | |
| - **Architecture:** Transformer decoder-only | |
| - **Parameters:** 8 billion | |
| - **Context Length:** 4096 tokens | |
| ### Training Details | |
| - **Training Method:** Full fine-tuning (no LoRA/adapters) | |
| - **Precision:** bfloat16 | |
| - **Training Data:** Elizabeth tool use minipack (198 high-quality examples) | |
| - **Training Time:** 2 minutes 36 seconds | |
| - **Final Loss:** 0.436 (from 3.27 → 0.16) | |
| - **Hardware:** 2x NVIDIA H200 (283GB total VRAM) | |
| ### Performance | |
| - **Training Speed:** 3.8 samples/second | |
| - **Convergence:** Excellent (3.27 → 0.16 loss) | |
| - **Tool Use Accuracy:** Optimized for reliable tool calling | |
| ## Usage | |
| ```python | |
| from transformers import AutoModelForCausalLM, AutoTokenizer | |
| model = AutoModelForCausalLM.from_pretrained( | |
| "LevelUp2x/qwen3-8b-elizabeth-simple", | |
| torch_dtype=torch.bfloat16, | |
| device_map="auto" | |
| ) | |
| tokenizer = AutoTokenizer.from_pretrained("LevelUp2x/qwen3-8b-elizabeth-simple") | |
| # Tool use example | |
| prompt = "Please help me calculate the square root of 144 using the calculator tool." | |
| inputs = tokenizer(prompt, return_tensors="pt").to(model.device) | |
| outputs = model.generate(**inputs, max_length=512) | |
| print(tokenizer.decode(outputs[0])) | |
| ``` | |
| ## Training Methodology | |
| ### Pure Weight Evolution | |
| This model was trained using pure weight evolution methodology - no external adapters, LoRA, or quantization were used. The entire base model weights were updated to bake Elizabeth's identity and tool use capabilities directly into the model parameters. | |
| ### Data Quality | |
| - **Dataset Size:** 198 carefully curated examples | |
| - **Quality:** High-quality tool use demonstrations | |
| - **Diversity:** Multiple tool types and usage patterns | |
| - **Consistency:** Uniform formatting and instruction following | |
| ### Optimization | |
| - **Gradient Accumulation:** 16 steps | |
| - **Effective Batch Size:** 64 | |
| - **Learning Rate:** 2e-5 | |
| - **Optimizer:** AdamW with cosine scheduler | |
| - **Epochs:** 3.0 | |
| ## Deployment | |
| ### Hardware Requirements | |
| - **GPU Memory:** Minimum 80GB VRAM (recommended 120GB+) | |
| - **Precision:** bfloat16 recommended | |
| - **Batch Size:** Optimal batch size of 4 | |
| ### Serving | |
| Recommended serving with vLLM for optimal performance: | |
| ```bash | |
| python -m vllm.entrypoints.api_server \ | |
| --model LevelUp2x/qwen3-8b-elizabeth-simple \ | |
| --dtype bfloat16 \ | |
| --gpu-memory-utilization 0.9 | |
| ``` | |
| ## License | |
| Apache 2.0 | |
| ## Citation | |
| ```bibtex | |
| @software{qwen3_8b_elizabeth_simple, | |
| title = {Qwen3-8B-Elizabeth-Simple: Tool Use Fine-Tuned Model}, | |
| author = {ADAPT-Chase and Nova Prime}, | |
| year = {2025}, | |
| url = {https://huggingface.co/LevelUp2x/qwen3-8b-elizabeth-simple}, | |
| publisher = {Hugging Face}, | |
| version = {1.0.0} | |
| } | |
| ``` | |
| ## Contact | |
| For questions about this model, please open an issue on the Hugging Face repository or contact the maintainers. |