| # Stack 2.9 Model Registry |
|
|
| > Version tracking for all Stack 2.9 model variants. |
|
|
| --- |
|
|
| ## Model Versions |
|
|
| | Version | Status | Date | Base Model | Parameters | Dataset | Performance | Use Case | |
| |---------|--------|------|------------|------------|---------|-------------|----------| |
| | `stack-2.9-1.5B` | 🟡 In Training | 2026-04-06 | Llama 3.2-1B | 1.5B | Stack 2.9 dedup | TBD | Research, fine-tuning base | |
| | `stack-2.9-7B` | 🔴 Planned | TBD | Llama 3.1-8B | 7B | Stack 2.9 dedup | TBD | General-purpose inference | |
| | `stack-2.9-7B-QLoRA` | 🔴 Planned | TBD | Llama 3.1-8B | 7B (quantized) | Stack 2.9 dedup | TBD | Edge deployment, low-memory | |
|
|
| --- |
|
|
| ## Version Details |
|
|
| ### stack-2.9-1.5B (Current) |
|
|
| - **Status:** In Training |
| - **Architecture:** Transformer (pretrained) |
| - **Base Model:** Llama 3.2-1B |
| - **Parameters:** 1.5B |
| - **Training Data:** Stack 2.9 deduplicated |
| - **Context Length:** 128k tokens |
| - **Vocabulary Size:** ~128K |
| - **Precision:** BF16 |
| - **Training Hardware:** 8x H100 (TBD确认) |
| - **Expected Completion:** TBD |
| - **Notes:** First iteration of Stack 2.9, used as baseline for larger variants |
|
|
| ### stack-2.9-7B (Planned) |
|
|
| - **Status:** Planned |
| - **Architecture:** Transformer (pretrained) |
| - **Base Model:** Llama 3.1-8B |
| - **Parameters:** 7B |
| - **Training Data:** Stack 2.9 deduplicated |
| - **Context Length:** 128k tokens |
| - **Vocabulary Size:** ~128K |
| - **Precision:** BF16 |
| - **Training Hardware:** TBD |
| - **Expected Start:** TBD |
| - **Notes:** Scale-up from 1.5B, targeting general-purpose use |
|
|
| ### stack-2.9-7B-QLoRA (Planned) |
|
|
| - **Status:** Planned |
| - **Architecture:** Transformer + QLoRA |
| - **Base Model:** Llama 3.1-8B |
| - **Parameters:** 7B (4-bit quantized) |
| - **Training Data:** Stack 2.9 deduplicated |
| - **Context Length:** 128k tokens |
| - **Vocabulary Size:** ~128K |
| - **Quantization:** 4-bit NF4 |
| - **LoRA Rank:** TBD |
| - **LoRA Alpha:** TBD |
| - **LoRA Dropout:** TBD |
| - **Target Modules:** TBD |
| - **Notes:** Quantized for consumer GPU deployment (e.g., 24GB VRAM) |
|
|
| --- |
|
|
| ## Changelog |
|
|
| | Date | Version | Change | |
| |------|---------|--------| |
| | 2026-04-06 | stack-2.9-1.5B | Initial entry — training started | |
|
|