Mer / README.md
POISONX's picture
Duplicate from unsloth/jobs
c8f983f
---
viewer: false
tags:
- uv-script
- unsloth
- training
- hf-jobs
- vlm
- fine-tuning
---
# 🦥 Unsloth Training Scripts for HF Jobs
UV scripts for fine-tuning LLMs and VLMs using [Unsloth](https://github.com/unslothai/unsloth) on [HF Jobs](https://huggingface.co/docs/hub/jobs) (on-demand cloud GPUs). UV handles dependency installation automatically, so you can run these scripts directly without any local setup.
These scripts can also be used or adapted by agents to train models for you.
## Prerequisites
- A Hugging Face account
- The [HF CLI](https://huggingface.co/docs/huggingface_hub/main/en/guides/cli) installed and authenticated (`hf auth login`)
- A dataset on the Hub in the appropriate format (see format requirements below). A strong LLM agent can often convert your data into the right format if needed.
## Data Formats
### LLM Fine-tuning (SFT)
Requires conversation data in ShareGPT or similar format:
```python
{
"messages": [
{"from": "human", "value": "What is the capital of France?"},
{"from": "gpt", "value": "The capital of France is Paris."}
]
}
```
The script auto-converts common formats (ShareGPT, Alpaca, etc.) via `standardize_data_formats`. See [mlabonne/FineTome-100k](https://huggingface.co/datasets/mlabonne/FineTome-100k) for a working dataset example.
### VLM Fine-tuning
Requires `images` and `messages` columns:
```python
{
"images": [<PIL.Image>], # List of images
"messages": [
{
"role": "user",
"content": [
{"type": "image"},
{"type": "text", "text": "What's in this image?"}
]
},
{
"role": "assistant",
"content": [
{"type": "text", "text": "A golden retriever playing fetch in a park."}
]
}
]
}
```
See [davanstrien/iconclass-vlm-sft](https://huggingface.co/datasets/davanstrien/iconclass-vlm-sft) for a working dataset example, and [davanstrien/iconclass-vlm-qwen3-best](https://huggingface.co/davanstrien/iconclass-vlm-qwen3-best) for a model trained with these scripts.
### Continued Pretraining
Any dataset with a text column:
```python
{"text": "Your domain-specific text here..."}
```
Use `--text-column` if your column has a different name.
## Usage
View available options for any script:
```bash
uv run https://huggingface.co/datasets/unsloth/jobs/raw/main/sft-lfm2.5.py --help
```
### LLM fine-tuning
Fine-tune [LFM2.5-1.2B-Instruct](https://huggingface.co/LiquidAI/LFM2.5-1.2B-Instruct), a compact and efficient text model from Liquid AI:
```bash
hf jobs uv run \
https://huggingface.co/datasets/unsloth/jobs/raw/main/sft-lfm2.5.py \
--flavor a10g-small --secrets HF_TOKEN --timeout 4h \
-- --dataset mlabonne/FineTome-100k \
--num-epochs 1 \
--eval-split 0.2 \
--output-repo your-username/lfm-finetuned
```
### VLM fine-tuning
```bash
hf jobs uv run \
https://huggingface.co/datasets/unsloth/jobs/raw/main/sft-qwen3-vl.py \
--flavor a100-large --secrets HF_TOKEN \
-- --dataset your-username/dataset \
--trackio-space your-username/trackio \
--output-repo your-username/my-model
```
### Continued pretraining
```bash
hf jobs uv run \
https://huggingface.co/datasets/unsloth/jobs/raw/main/continued-pretraining.py \
--flavor a100-large --secrets HF_TOKEN \
-- --dataset your-username/domain-corpus \
--text-column content \
--max-steps 1000 \
--output-repo your-username/domain-llm
```
### With Trackio monitoring
```bash
hf jobs uv run \
https://huggingface.co/datasets/unsloth/jobs/raw/main/sft-lfm2.5.py \
--flavor a10g-small --secrets HF_TOKEN \
-- --dataset mlabonne/FineTome-100k \
--trackio-space your-username/trackio \
--output-repo your-username/lfm-finetuned
```
## Scripts
| Script | Base Model | Task |
| ------------------------------------------------------ | -------------------- | ----------------------------- |
| [`sft-lfm2.5.py`](sft-lfm2.5.py) | LFM2.5-1.2B-Instruct | LLM fine-tuning (recommended) |
| [`sft-qwen3-vl.py`](sft-qwen3-vl.py) | Qwen3-VL-8B | VLM fine-tuning |
| [`sft-gemma3-vlm.py`](sft-gemma3-vlm.py) | Gemma 3 4B | VLM fine-tuning (smaller) |
| [`continued-pretraining.py`](continued-pretraining.py) | Qwen3-0.6B | Domain adaptation |
## Common Options
| Option | Description | Default |
| ------------------------- | -------------------------------------- | ------------ |
| `--dataset` | HF dataset ID | _required_ |
| `--output-repo` | Where to save trained model | _required_ |
| `--max-steps` | Number of training steps | 500 |
| `--num-epochs` | Train for N epochs instead of steps | - |
| `--eval-split` | Fraction for evaluation (e.g., 0.2) | 0 (disabled) |
| `--batch-size` | Per-device batch size | 2 |
| `--gradient-accumulation` | Gradient accumulation steps | 4 |
| `--lora-r` | LoRA rank | 16 |
| `--learning-rate` | Learning rate | 2e-4 |
| `--merge-model` | Upload merged model (not just adapter) | false |
| `--trackio-space` | HF Space for live monitoring | - |
| `--run-name` | Custom name for Trackio run | auto |
## Tips
- Use `--max-steps 10` to verify everything works before a full run
- `--eval-split 0.1` helps detect overfitting
- Run `hf jobs hardware` to see GPU pricing (A100-large ~$2.50/hr, L40S ~$1.80/hr)
- Add `--streaming` for very large datasets
- First training step may take a few minutes (CUDA kernel compilation)
## Links
- [HF Jobs Quickstart](https://huggingface.co/docs/hub/jobs-quickstart)
- [Unsloth Documentation](https://docs.unsloth.ai/)
- [UV Scripts Guide](https://docs.astral.sh/uv/guides/scripts/)