Model Card for GEAR-2-500m-Identity
Model Details
Model Description
GEAR-2-500m-Identity is a lightweight Transformer LLM with approximately 0.5 billion parameters, fine-tuned on the Qwen2.5 architecture using Unsloth. It is designed to run extremely fast on local machines (CPU/Edge) with minimal memory usage. The model embodies the persona of Gear, an intelligent assistant created by HeavensHack.
It is capable of code generation (Python) and general chat. While efficient, it is a small model and may struggle with complex reasoning compared to larger parameters.
- Developed by: HeavensHack
- Model type: Qwen2 For Causal LM
- Language(s) (NLP): English, Python (Code) , (New) Russian
- License: Apache 2.0
- Finetuned from model: Qwen/Qwen2.5-0.5B-Instruct
Uses
Direct Use
- Fast local chat assistant
- Python code generation and debugging
Out-of-Scope Use
- Complex mathematical reasoning
- High-stakes decision making
- Long-context analysis requiring high accuracy
Bias, Risks, and Limitations
- Hallucinations: Due to the 0.5B parameter size, it may generate plausible but incorrect information.
- Identity: The model is strictly fine-tuned to identify itself as "Gear" by HeavensHack.
- Inconsistency: Behavior might be variable in long conversations.
Recommendations
- Use for educational purposes, hobby projects, or low-resource environments.
- Verify any code generated before running it in production.
How to Get Started
- Load the model using Unsloth or standard Hugging Face transformers.
- Optimized for local inference.
Training Details
- Training Data: Custom identity dataset (HeavensHack), Alpaca (English), and Python Code instructions.
- Training Procedure: Fine-tuned using Unsloth (LoRA) for efficiency.
- Training Regime: Mixed precision (BF16/FP16).
Evaluation
- Validated for identity retention and basic coding tasks.
- Not benchmarked for enterprise production use.
Environmental Impact
- Extremely low compute cost during training due to Unsloth optimization.
Model Card Contact
But...
- At first, only the GGUF file will be available. The rest will follow later.