| Yuuki - Mobile-Trained Code Language Model | |
| Copyright 2026 OpceanAI | |
| This product includes a language model trained entirely on a mobile device | |
| (Qualcomm Snapdragon 685) over 42 days with zero GPU budget. | |
| Training Details: | |
| - Base model: DistilGPT-2 (82M parameters) | |
| - Training period: January-March 2026 | |
| - Hardware: Android device (Snapdragon 685, 6GB RAM) | |
| - Dataset: The Stack (75,000 examples for v0.1) | |
| - Total cost: $0 in cloud/GPU compute | |
| Third-party Components: | |
| - Transformers library by HuggingFace (Apache 2.0) | |
| - PyTorch (BSD-3-Clause) | |
| - The Stack dataset by BigCode (BigCode OpenRAIL-M) | |
| - DistilGPT-2 base model (Apache 2.0) | |
| Special Thanks: | |
| - My snapdragon 685 | |
| - HuggingFace for infrastructure | |
| - The ML community | |