| | --- |
| | title: Math Conjecture Trainer |
| | sdk: gradio |
| | sdk_version: "6.6.0" |
| | python_version: "3.10" |
| | app_file: app.py |
| | pinned: false |
| | emoji: 🧮 |
| |
|
| | --- |
| | |
| | # Math Conjecture Trainer Space |
| |
|
| | An autonomous training operations console for DeepSeek-Math that runs multi-stage curriculum fine-tuning on Space GPU, executes post-training quality evaluation, and publishes only qualified adapters, checkpoints, and run reports to your Hugging Face model repository. |
| |
|
| | This Space is the tactical operations console for `maths-conjuncture-solutions` and is wired to: |
| |
|
| | - dataset: `NorthernTribe-Research/math-conjecture-training-corpus` |
| | - model repo: `NorthernTribe-Research/math-conjecture-model` |
| |
|
| | ## End-to-end flow |
| |
|
| | 1. Download released parquet splits (`train/validation/test`). |
| | 2. Build runtime config from `configs/deepseek_math_sota.yaml`. |
| | 3. Run 4-stage curriculum LoRA fine-tuning with `scripts/train_sota.py`. |
| | 4. Run post-train evaluation (`pass@1`, `pass@k`, exact/boxed, family metrics). |
| | 5. Apply quality gate thresholds before hub push. |
| | 6. Emit `training_summary.json` + `post_eval_report.json` and stream live telemetry in UI. |
| |
|
| | ## Access posture |
| |
|
| | Credentials and publish permissions are handled by deployment runtime settings. |
| |
|
| | ## Runtime controls |
| |
|
| | - `Autonomous Mode`: enabled by default; applies full-stage training/eval/gate/publish profile automatically. |
| | - `Run Evaluation After Training`: toggles post-train eval in runtime config. |
| | - `Enforce Quality Gate`: enables/disables promotion gate checks. |
| | - `Gate Min pass@1`, `Gate Min pass@k`, `Gate Min Rows`: runtime gate thresholds. |
| | - `Live Tactical Telemetry`: real-time stage progression, runtime posture, and training-loss graph (sparkline) with gate/push state. |
| | - `Ops Console (Live Log + Mission JSON)`: unified panel for line-by-line runtime stream, heartbeats, and structured mission summary. |
| | - `Validation Mode (No Training)`: validates pipeline with `--dry-run`. |
| | - `Force Dataset Redownload`: bypasses cached parquet files. |
| | - `Abort Active Run`: cancels active subprocess tree. |
| |
|
| | ## Artifacts |
| |
|
| | - runtime config: `workspace/runtime/deepseek_math_sota.runtime.yaml` |
| | - run output root: `workspace/runs/math-conjecture-sota` |
| | - final adapter: `workspace/runs/math-conjecture-sota/final_adapter` |
| | - training summary: `workspace/runs/math-conjecture-sota/training_summary.json` |
| | - post-eval report: `workspace/runs/math-conjecture-sota/post_eval_report.json` |
| |
|
| | ## Notes |
| |
|
| | - Full training runs on GPU when available and automatically falls back to CPU mode when CUDA is unavailable. |
| | - App handles Gradio copy-button compatibility across versions automatically. |
| |
|