--- license: apache-2.0 base_model: - WeiboAI/VibeThinker-1.5B datasets: - OpceanAI/Yuuki-Personality-v2 language: - en - es library_name: transformers tags: - reasoning - unsloth - pytorch - bilingual - opceanai - yuuki - rxg - fine-tuned - chat - deepseek - qwen2 pipeline_tag: text-generation ---
|
**Architecture**
| Property | Value |
|:---------|:------|
| Base Model | VibeThinker-1.5B |
| Base Architecture | Qwen2.5-Math-1.5B |
| Parameters | 1.5B |
| Fine-tuning Method | QLoRA SFT |
| Trainable Parameters | 18.4M (1.18%) |
| Context Length | 4,096 tokens |
| Chat Template | ChatML |
| Thinking Protocol | Native ` |
**Release** | Property | Value | |:---------|:------| | Organization | OpceanAI | | Release Date | April 2026 | | Version | v1.0 | | Languages | English, Spanish | | License | Apache 2.0 | | Evaluation | lm-evaluation-harness | | Training Cost | < $15 USD | | Training Time | ~90 minutes | |
| **Hardware** | Component | Specification | |:----------|:-------------| | GPU | NVIDIA A100 40GB | | Precision | BF16 native | | Framework | Unsloth 2026.4 + TRL | | Flash Attention | Xformers fallback | | Cloud Compute | Google Colab Pro | | Training Time | ~90 minutes | | Total Cost | < $15 USD | | **LoRA Configuration** | Parameter | Value | |:----------|:-----:| | Rank (r) | 16 | | Alpha | 32 | | Dropout | 0.0 | | Target Modules | q, k, v, o, gate, up, down | | Trainable Parameters | 18.4M (1.18%) | | Gradient Checkpointing | Unsloth smart offload | | Quantization | 4-bit NF4 (QLoRA) | |