gary-boon Claude Opus 4.5 commited on
Commit
65c6e2e
·
1 Parent(s): 688efad

docs: Mark GPU HF Space Devstral deployment complete

Browse files

- A100 (80GB VRAM) configured
- DEFAULT_MODEL=devstral-small set
- Phase 2c fully complete

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com>

Files changed (1) hide show
  1. docs/devstral-spark-plan-phased.md +2 -2
docs/devstral-spark-plan-phased.md CHANGED
@@ -1829,12 +1829,12 @@ Before marking each phase complete, verify:
1829
  - Dynamic vocab size from modelInfo
1830
  - Dynamic head_dim derived from actual matrix data
1831
  - Removed hardcoded "64 dimensions" in tutorial
1832
- - [x] **Phase 2c**: API route conversion ✅ COMPLETE (partial)
1833
  - All 8 API routes converted to use backendFetch helper
1834
  - Server-side auth with HF token for private Spaces
1835
  - Per-user backend routing working
 
1836
  - ⏸️ Spark toggle deferred (no benefit until PyTorch supports GB10)
1837
- - ⏸️ GPU HF Space Devstral deployment pending (requires VRAM upgrade)
1838
  - [ ] **Phase 3**: Deploy Devstral to DGX Spark ⏸️ BLOCKED (PyTorch sm_121 support)
1839
  - [ ] **Phase 4**: Future enhancements (optional)
1840
 
 
1829
  - Dynamic vocab size from modelInfo
1830
  - Dynamic head_dim derived from actual matrix data
1831
  - Removed hardcoded "64 dimensions" in tutorial
1832
+ - [x] **Phase 2c**: API route conversion + GPU HF Space ✅ COMPLETE
1833
  - All 8 API routes converted to use backendFetch helper
1834
  - Server-side auth with HF token for private Spaces
1835
  - Per-user backend routing working
1836
+ - GPU HF Space configured: A100 (80GB), DEFAULT_MODEL=devstral-small
1837
  - ⏸️ Spark toggle deferred (no benefit until PyTorch supports GB10)
 
1838
  - [ ] **Phase 3**: Deploy Devstral to DGX Spark ⏸️ BLOCKED (PyTorch sm_121 support)
1839
  - [ ] **Phase 4**: Future enhancements (optional)
1840