--- title: Opus Research emoji: 🧠 colorFrom: purple colorTo: blue sdk: static pinned: false ---

🧠 Opus Research

Training AI models from scratch, one parameter at a time.

"We stand at the right place at the right time."

## 👋 About Us We're two teenage AI enthusiasts (ages 15 & 17) passionate about understanding AI from the ground up. Instead of just using pre-trained models, we build them ourselves. ## 🚀 Our Models | Model | Parameters | Architecture | Status | |-------|-----------|--------------|--------| | [Opus 1.5](https://huggingface.co/opus-research/opus-1.5) | 0.88B | LLaMA-style | ✅ Released | | Opus 2.0 | 3B+ | LLaMA + Reasoning | 🔜 Coming Soon | ## 🔬 Research Focus - **Training from scratch** - No pre-trained weights, 100% original - **Chain-of-thought reasoning** - Teaching models to think before answering - **Efficient architectures** - Sub-3B models that run on consumer GPUs ## 📊 Opus 1.5 Highlights - **0.88 billion parameters** - **~2 GB VRAM** for inference - **42 hours** training on 2x RTX 4090 - **LLaMA architecture** with RoPE, SwiGLU, GQA, FlashAttention-2 ## 🔗 Links - 📦 [Opus 1.5 on HuggingFace](https://huggingface.co/opus-research/opus-1.5) ## ⚡ Powered By - ❤️ Two teenagers who should probably be doing homework - 🎮 Two RTX 4090s that haven't known rest since December 2025 - 🔥 Questionable GPU thermals - 📊 A concerning amount of time staring at loss curves - 💡 48GB of VRAM - ⚡ ~847 kWh of electricity we'll never get back - 💸 Our parents' electricity bill - 🤔 The firm belief that we could've just used ChatGPT

But where's the fun in that?