spoodddddd commited on
Commit
323cc3b
Β·
verified Β·
1 Parent(s): 180b996

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +12 -2
README.md CHANGED
@@ -6,31 +6,41 @@ colorTo: blue
6
  sdk: static
7
  pinned: false
8
  ---
 
9
  <div align="center">
10
  <h1>🧠 Opus Research</h1>
11
  <p><strong>Training AI models from scratch, one parameter at a time.</strong></p>
12
  <p><em>"We stand at the right place at the right time."</em></p>
13
  </div>
14
- ---
15
  ## πŸ‘‹ About Us
 
16
  We're two teenage AI enthusiasts (ages 15 & 17) passionate about understanding AI from the ground up. Instead of just using pre-trained models, we build them ourselves.
 
17
  ## πŸš€ Our Models
 
18
  | Model | Parameters | Architecture | Status |
19
  |-------|-----------|--------------|--------|
20
  | [Opus 1.5](https://huggingface.co/opus-research/opus-1.5) | 0.88B | LLaMA-style | βœ… Released |
21
  | Opus 2.0 | 3B+ | LLaMA + Reasoning | πŸ”œ Coming Soon |
 
22
  ## πŸ”¬ Research Focus
 
23
  - **Training from scratch** - No pre-trained weights, 100% original
24
  - **Chain-of-thought reasoning** - Teaching models to think before answering
25
  - **Efficient architectures** - Sub-3B models that run on consumer GPUs
 
26
  ## πŸ“Š Opus 1.5 Highlights
 
27
  - **0.88 billion parameters**
28
  - **~2 GB VRAM** for inference
29
  - **42 hours** training on 2x RTX 4090
30
  - **LLaMA architecture** with RoPE, SwiGLU, GQA, FlashAttention-2
 
31
  ## πŸ”— Links
 
32
  - πŸ“¦ [Opus 1.5 on HuggingFace](https://huggingface.co/opus-research/opus-1.5)
33
- ---
34
  <div align="center">
35
  <p>Made with ❀️ and way too much GPU power</p>
36
  </div>
 
6
  sdk: static
7
  pinned: false
8
  ---
9
+
10
  <div align="center">
11
  <h1>🧠 Opus Research</h1>
12
  <p><strong>Training AI models from scratch, one parameter at a time.</strong></p>
13
  <p><em>"We stand at the right place at the right time."</em></p>
14
  </div>
15
+
16
  ## πŸ‘‹ About Us
17
+
18
  We're two teenage AI enthusiasts (ages 15 & 17) passionate about understanding AI from the ground up. Instead of just using pre-trained models, we build them ourselves.
19
+
20
  ## πŸš€ Our Models
21
+
22
  | Model | Parameters | Architecture | Status |
23
  |-------|-----------|--------------|--------|
24
  | [Opus 1.5](https://huggingface.co/opus-research/opus-1.5) | 0.88B | LLaMA-style | βœ… Released |
25
  | Opus 2.0 | 3B+ | LLaMA + Reasoning | πŸ”œ Coming Soon |
26
+
27
  ## πŸ”¬ Research Focus
28
+
29
  - **Training from scratch** - No pre-trained weights, 100% original
30
  - **Chain-of-thought reasoning** - Teaching models to think before answering
31
  - **Efficient architectures** - Sub-3B models that run on consumer GPUs
32
+
33
  ## πŸ“Š Opus 1.5 Highlights
34
+
35
  - **0.88 billion parameters**
36
  - **~2 GB VRAM** for inference
37
  - **42 hours** training on 2x RTX 4090
38
  - **LLaMA architecture** with RoPE, SwiGLU, GQA, FlashAttention-2
39
+
40
  ## πŸ”— Links
41
+
42
  - πŸ“¦ [Opus 1.5 on HuggingFace](https://huggingface.co/opus-research/opus-1.5)
43
+
44
  <div align="center">
45
  <p>Made with ❀️ and way too much GPU power</p>
46
  </div>