orbgen-training / README.md
javasop's picture
Upload folder using huggingface_hub
9791706 verified
metadata
title: OrbGen Training
emoji: 🚀
colorFrom: blue
colorTo: purple
sdk: docker
pinned: false
license: apache-2.0

OrbGen Training

Training code for OrbGen - a model that generates valid Orbital schemas from natural language.

Quick Start

Local Training

# Install dependencies
pip install -r requirements.txt

# Full training (with local GPU)
python train.py --config config.yaml

# Debug mode (1 epoch, no W&B)
python train.py --config config.yaml --debug

HuggingFace Cloud Training

# Use the HuggingFace-optimized config
python train.py --config config-huggingface.yaml

Configuration Files

Config GPU VRAM Use Case
config.yaml Local (RTX 3000) 6GB Local testing with QLoRA
config-huggingface.yaml A10G 24GB HuggingFace Spaces training

Evaluation

# Basic evaluation
python evaluate.py --checkpoint ./orbgen-1.5b/final

# With Orbital validator
python evaluate.py --checkpoint ./orbgen-1.5b/final --use_validator

Generate

# Single generation
python generate.py --prompt "Create a task management app with projects and due dates"

# Interactive mode
python generate.py --interactive

# Save to file
python generate.py --prompt "..." --output schema.orb --validate

Files

File Description
train.py Main training script with SFT
evaluate.py Evaluation with Orbital validation
generate.py Inference and generation
config.yaml Config for local 6GB GPU (QLoRA)
config-huggingface.yaml Config for HuggingFace A10G (24GB)
Dockerfile Container for HuggingFace Spaces
requirements.txt Python dependencies

Training on HuggingFace Spaces

Prerequisites

  1. Upload dataset to HuggingFace:

    python scripts/upload_dataset.py
    
  2. Set Space secrets:

    • HUGGINGFACE_TOKEN - HF token with write access
    • WANDB_API_KEY - Weights & Biases API key
  3. Push training code:

    cd orbgen-training
    huggingface-cli upload orbital-ai/orbgen-training . --repo-type space
    
  4. Configure Space with A10G GPU in settings

  5. Training will start automatically

Hardware Requirements

Phase GPU VRAM Notes
Training (local) RTX 3000 6GB Uses QLoRA (4-bit)
Training (cloud) A10G 24GB Full bf16 training
Inference T4 16GB Production inference

Model Output

After training, the model is saved to ./orbgen-1.5b/final/:

  • adapter_config.json - LoRA configuration
  • adapter_model.safetensors - LoRA weights
  • tokenizer.json - Tokenizer
  • config.json - Model config

Model is automatically pushed to orbital-ai/orbgen-1.5b when using HuggingFace config.