Spaces:
Runtime error
Runtime error
metadata
title: OrbGen Training
emoji: 🚀
colorFrom: blue
colorTo: purple
sdk: docker
pinned: false
license: apache-2.0
OrbGen Training
Training code for OrbGen - a model that generates valid Orbital schemas from natural language.
Quick Start
Local Training
# Install dependencies
pip install -r requirements.txt
# Full training (with local GPU)
python train.py --config config.yaml
# Debug mode (1 epoch, no W&B)
python train.py --config config.yaml --debug
HuggingFace Cloud Training
# Use the HuggingFace-optimized config
python train.py --config config-huggingface.yaml
Configuration Files
| Config | GPU | VRAM | Use Case |
|---|---|---|---|
config.yaml |
Local (RTX 3000) | 6GB | Local testing with QLoRA |
config-huggingface.yaml |
A10G | 24GB | HuggingFace Spaces training |
Evaluation
# Basic evaluation
python evaluate.py --checkpoint ./orbgen-1.5b/final
# With Orbital validator
python evaluate.py --checkpoint ./orbgen-1.5b/final --use_validator
Generate
# Single generation
python generate.py --prompt "Create a task management app with projects and due dates"
# Interactive mode
python generate.py --interactive
# Save to file
python generate.py --prompt "..." --output schema.orb --validate
Files
| File | Description |
|---|---|
train.py |
Main training script with SFT |
evaluate.py |
Evaluation with Orbital validation |
generate.py |
Inference and generation |
config.yaml |
Config for local 6GB GPU (QLoRA) |
config-huggingface.yaml |
Config for HuggingFace A10G (24GB) |
Dockerfile |
Container for HuggingFace Spaces |
requirements.txt |
Python dependencies |
Training on HuggingFace Spaces
Prerequisites
Upload dataset to HuggingFace:
python scripts/upload_dataset.pySet Space secrets:
HUGGINGFACE_TOKEN- HF token with write accessWANDB_API_KEY- Weights & Biases API key
Push training code:
cd orbgen-training huggingface-cli upload orbital-ai/orbgen-training . --repo-type spaceConfigure Space with A10G GPU in settings
Training will start automatically
Hardware Requirements
| Phase | GPU | VRAM | Notes |
|---|---|---|---|
| Training (local) | RTX 3000 | 6GB | Uses QLoRA (4-bit) |
| Training (cloud) | A10G | 24GB | Full bf16 training |
| Inference | T4 | 16GB | Production inference |
Model Output
After training, the model is saved to ./orbgen-1.5b/final/:
adapter_config.json- LoRA configurationadapter_model.safetensors- LoRA weightstokenizer.json- Tokenizerconfig.json- Model config
Model is automatically pushed to orbital-ai/orbgen-1.5b when using HuggingFace config.