Stack-2-9-finetuned / docs /MODEL_REGISTRY.md
walidsobhie-code
reorganize: consolidate root level to 20 folders
b8e3e42

Stack 2.9 Model Registry

Version tracking for all Stack 2.9 model variants.


Model Versions

Version Status Date Base Model Parameters Dataset Performance Use Case
stack-2.9-1.5B 🟡 In Training 2026-04-06 Llama 3.2-1B 1.5B Stack 2.9 dedup TBD Research, fine-tuning base
stack-2.9-7B 🔴 Planned TBD Llama 3.1-8B 7B Stack 2.9 dedup TBD General-purpose inference
stack-2.9-7B-QLoRA 🔴 Planned TBD Llama 3.1-8B 7B (quantized) Stack 2.9 dedup TBD Edge deployment, low-memory

Version Details

stack-2.9-1.5B (Current)

  • Status: In Training
  • Architecture: Transformer (pretrained)
  • Base Model: Llama 3.2-1B
  • Parameters: 1.5B
  • Training Data: Stack 2.9 deduplicated
  • Context Length: 128k tokens
  • Vocabulary Size: ~128K
  • Precision: BF16
  • Training Hardware: 8x H100 (TBD确认)
  • Expected Completion: TBD
  • Notes: First iteration of Stack 2.9, used as baseline for larger variants

stack-2.9-7B (Planned)

  • Status: Planned
  • Architecture: Transformer (pretrained)
  • Base Model: Llama 3.1-8B
  • Parameters: 7B
  • Training Data: Stack 2.9 deduplicated
  • Context Length: 128k tokens
  • Vocabulary Size: ~128K
  • Precision: BF16
  • Training Hardware: TBD
  • Expected Start: TBD
  • Notes: Scale-up from 1.5B, targeting general-purpose use

stack-2.9-7B-QLoRA (Planned)

  • Status: Planned
  • Architecture: Transformer + QLoRA
  • Base Model: Llama 3.1-8B
  • Parameters: 7B (4-bit quantized)
  • Training Data: Stack 2.9 deduplicated
  • Context Length: 128k tokens
  • Vocabulary Size: ~128K
  • Quantization: 4-bit NF4
  • LoRA Rank: TBD
  • LoRA Alpha: TBD
  • LoRA Dropout: TBD
  • Target Modules: TBD
  • Notes: Quantized for consumer GPU deployment (e.g., 24GB VRAM)

Changelog

Date Version Change
2026-04-06 stack-2.9-1.5B Initial entry — training started