|
|
--- |
|
|
language: |
|
|
- en |
|
|
license: apache-2.0 |
|
|
library_name: transformers |
|
|
tags: |
|
|
- tool-use |
|
|
- fine-tuned |
|
|
- qwen3 |
|
|
- 8b |
|
|
- elizabeth |
|
|
pipeline_tag: text-generation |
|
|
--- |
|
|
|
|
|
# Model Card for Qwen3-8B-Elizabeth-Simple |
|
|
|
|
|
## Model Details |
|
|
|
|
|
### Model Description |
|
|
- **Developed by:** ADAPT-Chase |
|
|
- **Model type:** Transformer-based language model |
|
|
- **Language(s):** English |
|
|
- **License:** Apache 2.0 |
|
|
- **Finetuned from:** Qwen/Qwen3-8B |
|
|
|
|
|
### Model Sources |
|
|
- **Repository:** https://huggingface.co/LevelUp2x/qwen3-8b-elizabeth-simple |
|
|
- **Paper:** N/A |
|
|
- **Demo:** N/A |
|
|
|
|
|
## Uses |
|
|
|
|
|
### Direct Use |
|
|
This model is designed for tool use and function calling tasks. It can be used for: |
|
|
- Automated tool invocation |
|
|
- API calling |
|
|
- Function execution |
|
|
- Task automation |
|
|
- Agent systems |
|
|
|
|
|
### Out-of-Scope Use |
|
|
- Medical advice |
|
|
- Legal decisions |
|
|
- Financial recommendations |
|
|
- Harmful content generation |
|
|
|
|
|
## Bias, Risks, and Limitations |
|
|
|
|
|
This model inherits biases from its base model Qwen3-8B and may exhibit: |
|
|
- Social biases present in training data |
|
|
- Limitations in tool use accuracy |
|
|
- Potential hallucination of tool responses |
|
|
|
|
|
### Recommendations |
|
|
Users should: |
|
|
- Validate tool outputs |
|
|
- Implement safety checks |
|
|
- Monitor for unexpected behavior |
|
|
- Use in controlled environments |
|
|
|
|
|
## Training Details |
|
|
|
|
|
### Training Data |
|
|
- **Dataset:** Elizabeth tool use minipack |
|
|
- **Samples:** 198 high-quality examples |
|
|
- **Format:** Instruction-response pairs with tool calls |
|
|
|
|
|
### Training Procedure |
|
|
- **Training regime:** Full fine-tuning |
|
|
- **Precision:** bfloat16 |
|
|
- **Hardware:** 2x NVIDIA H200 |
|
|
- **Training time:** 2 minutes 36 seconds |
|
|
|
|
|
#### Training Hyperparameters |
|
|
- **Learning rate:** 2e-5 |
|
|
- **Batch size:** 4 (effective 64 with accumulation) |
|
|
- **Epochs:** 3.0 |
|
|
- **Optimizer:** AdamW |
|
|
- **Scheduler:** Cosine |
|
|
|
|
|
## Evaluation |
|
|
|
|
|
### Testing Data |
|
|
- **Factors:** Tool use accuracy, response quality |
|
|
- **Metrics:** Loss, perplexity, tool call success rate |
|
|
|
|
|
### Results |
|
|
- **Final loss:** 0.436 |
|
|
- **Training speed:** 3.8 samples/second |
|
|
- **Convergence:** Excellent (3.27 → 0.16) |
|
|
|
|
|
## Environmental Impact |
|
|
|
|
|
- **Hardware Type:** NVIDIA H200 GPUs |
|
|
- **Hours used:** 0.043 hours |
|
|
- **Cloud Provider:** Private infrastructure |
|
|
- **Carbon Emitted:** Minimal (estimated < 0.1 kgCO2eq) |
|
|
|
|
|
## Technical Specifications |
|
|
|
|
|
### Model Architecture and Objective |
|
|
- **Architecture:** Transformer decoder |
|
|
- **Objective:** Causal language modeling |
|
|
- **Params:** 8 billion |
|
|
- **Context length:** 4096 |
|
|
|
|
|
### Compute Infrastructure |
|
|
- **Hardware:** 2x NVIDIA H200 |
|
|
- **VRAM used:** ~120GB during training |
|
|
|
|
|
## Citation |
|
|
|
|
|
**BibTeX:** |
|
|
```bibtex |
|
|
@software{qwen3_8b_elizabeth_simple_2025, |
|
|
title = {Qwen3-8B-Elizabeth-Simple}, |
|
|
author = {ADAPT-Chase and Nova Prime}, |
|
|
year = {2025}, |
|
|
url = {https://huggingface.co/LevelUp2x/qwen3-8b-elizabeth-simple}, |
|
|
publisher = {Hugging Face} |
|
|
} |
|
|
``` |
|
|
|
|
|
## Glossary |
|
|
|
|
|
- **Pure Weight Evolution:** Full fine-tuning without adapters |
|
|
- **Tool Use:** Ability to call external functions/APIs |
|
|
- **bfloat16:** Brain floating point format |
|
|
|
|
|
## Model Card Authors |
|
|
|
|
|
ADAPT-Chase and Nova Prime |
|
|
|
|
|
## How to Get Help |
|
|
|
|
|
Open an issue on the Hugging Face repository or contact the maintainers. |