zen
zenlm
zeekay commited on
Commit
44abfbe
·
verified ·
1 Parent(s): 3fa5ab1

Update README: proper training infrastructure model card, remove Gradio Space metadata

Browse files
Files changed (1) hide show
  1. README.md +55 -122
README.md CHANGED
@@ -1,148 +1,81 @@
1
  ---
2
- title: Zen Training
3
- emoji: 🧘
4
- colorFrom: blue
5
- colorTo: purple
6
- sdk: gradio
7
- sdk_version: 4.0.0
8
- app_file: app.py
9
- pinned: true
10
  license: apache-2.0
11
- hardware: a10g-large
 
 
 
 
12
  ---
13
 
14
- # 🧘 Zen Training Space
15
 
16
- **Unified Training Platform for All Zen Models**
17
 
18
- Train any Zen model with any dataset combination from HuggingFace. Everything runs directly from HF datasets - no local storage needed!
19
 
20
- ## 🎯 Features
21
 
22
- ### Supported Models
23
 
24
- **Language Models:**
25
- - `zen-nano` (0.6B) - Edge deployment
26
- - `zen-eco` (4B) - Balanced performance
27
- - `zen-omni` (7B) - Multi-task
28
- - `zen-coder` (14B) - Code generation
29
- - `zen-next` (32B) - Frontier performance
30
 
31
- **Vision-Language Models:**
32
- - `zen-vl-4b` - Efficient VL with function calling
33
- - `zen-vl-8b` - Enhanced VL capabilities
34
- - `zen-vl-30b` - Maximum VL performance
 
 
35
 
36
- ### Supported Datasets
37
 
38
- **Agent Training (ADP):**
39
- - AgentTuning OS/KG/DB (~15k samples)
40
- - Synatra (99k agent trajectories)
41
- - Code Feedback (66k samples)
42
- - Go Browse (27k web interactions)
 
 
43
 
44
- **Function Calling:**
45
- - xLAM 60k (Salesforce high-quality function calling)
46
 
47
- **Instruction Tuning:**
48
- - Alpaca (52k instruction samples)
 
 
 
 
49
 
50
- ## 🚀 How to Use
51
 
52
- 1. **Select Model**: Choose from language or vision-language models
53
- 2. **Select Datasets**: Check multiple datasets to combine them
54
- 3. **Configure Training**: Set epochs, batch size, learning rate, max samples
55
- 4. **Set Output Repo**: Specify HuggingFace repo for trained model
56
- 5. **Start Training**: Click the button and monitor logs
57
 
58
- ## ⚙️ Training Configuration
 
 
 
 
59
 
60
- ### Recommended Settings
61
 
62
- **4B Models (A10G - 24GB):**
63
- - Batch Size: 1-2
64
- - Max Samples: 10,000-30,000
65
- - Time: 4-8 hours
66
- - Cost: ~$3-5
67
 
68
- **8B Models (A100 - 40GB):**
69
- - Batch Size: 2-4
70
- - Max Samples: 30,000-50,000
71
- - Time: 8-12 hours
72
- - Cost: ~$15-20
73
-
74
- **32B Models (A100 - 80GB):**
75
- - Batch Size: 1-2
76
- - Max Samples: 50,000-100,000
77
- - Time: 20-30 hours
78
- - Cost: ~$50-80
79
-
80
- ## 📊 Dataset Combinations
81
-
82
- ### For Agent Training:
83
- ```
84
- ADP Synatra (80%) + xLAM (20%)
85
- = Strong agent + quality function calling
86
- ```
87
-
88
- ### For Code Models:
89
- ```
90
- Code Feedback (70%) + Alpaca (30%)
91
- = Code expertise + general instruction following
92
- ```
93
-
94
- ### For VL Models:
95
- ```
96
- ADP (all configs) + xLAM
97
- = Complete vision-language agent training
98
  ```
99
 
100
- ## 🔒 Requirements
101
-
102
- - HuggingFace Pro account (for GPU access)
103
- - Write access to output repository
104
- - HF_TOKEN secret set in Space settings
105
-
106
- ## 💡 Tips
107
 
108
- 1. **Start Small**: Test with 1,000 samples first
109
- 2. **Mix Datasets**: Combine complementary datasets for best results
110
- 3. **Monitor Logs**: Watch for OOM errors and adjust batch size
111
- 4. **Save Often**: Lower save_steps for longer training runs
 
 
112
 
113
- ## 📚 Resources
114
-
115
- - **Website**: https://zenlm.org
116
- - **GitHub**: https://github.com/zenlm
117
- - **Models**: https://huggingface.co/zenlm
118
- - **Datasets**:
119
- - [ADP](https://huggingface.co/datasets/neulab/agent-data-collection)
120
- - [xLAM](https://huggingface.co/datasets/Salesforce/xlam-function-calling-60k)
121
-
122
- ## 📄 License
123
 
124
  Apache 2.0
125
-
126
- ## 🙏 Citations
127
-
128
- ```bibtex
129
- @software{zen-training-2025,
130
- title={Zen Training: Unified Training Platform for Zen Models},
131
- author={Zen AI Team},
132
- year={2025},
133
- url={https://huggingface.co/spaces/zenlm/zen-training}
134
- }
135
-
136
- @article{adp2024,
137
- title={Agent Data Protocol},
138
- author={NeuLab},
139
- journal={arXiv preprint arXiv:2510.24702},
140
- year={2024}
141
- }
142
-
143
- @dataset{xlam2024,
144
- title={xLAM Function Calling Dataset},
145
- author={Salesforce Research},
146
- year={2024}
147
- }
148
- ```
 
1
  ---
2
+ language: en
 
 
 
 
 
 
 
3
  license: apache-2.0
4
+ tags:
5
+ - training
6
+ - zen
7
+ - zenlm
8
+ - hanzo
9
  ---
10
 
11
+ # Zen Training
12
 
13
+ Training infrastructure and recipes for the Zen model family.
14
 
15
+ **Zen LM by Hanzo AI** Open training configurations for all Zen models.
16
 
17
+ ## Overview
18
 
19
+ This repository contains the training configurations, scripts, and recipes used to train Zen models using the Zen MoDE (Mixture of Distilled Experts) architecture. All training runs use mixed-precision distributed training with full support for LoRA/QLoRA fine-tuning and alignment techniques.
20
 
21
+ ## Training Recipes
 
 
 
 
 
22
 
23
+ | Model | Type | Parameters | Context | Hardware |
24
+ |-------|------|-----------|---------|----------|
25
+ | Zen Nano | Dense | 0.6B | 32K | 1x H100 |
26
+ | Zen Eco | Dense | 4B | 64K | 4x H100 |
27
+ | Zen Pro | Dense | 8B | 128K | 8x H100 |
28
+ | Zen MAX | MoE | 235B (22B active) | 128K | 64x H100 |
29
 
30
+ ## Features
31
 
32
+ - Mixed precision training (BF16)
33
+ - Gradient checkpointing
34
+ - Distributed training with FSDP / DeepSpeed ZeRO-3
35
+ - LoRA / QLoRA fine-tuning support
36
+ - RLHF and DPO alignment pipelines
37
+ - Dataset mixing and curriculum scheduling
38
+ - Evaluation harness integration
39
 
40
+ ## Supported Training Tasks
 
41
 
42
+ - Instruction tuning
43
+ - Function calling
44
+ - Agent trajectory training
45
+ - Vision-language alignment
46
+ - Code generation fine-tuning
47
+ - Reasoning / chain-of-thought distillation
48
 
49
+ ## Dataset Support
50
 
51
+ Training recipes support direct streaming from HuggingFace datasets:
 
 
 
 
52
 
53
+ - Instruction tuning corpora
54
+ - Agent behavior datasets
55
+ - Function calling datasets
56
+ - Code and math reasoning sets
57
+ - Multilingual alignment data
58
 
59
+ ## Quick Start
60
 
61
+ See [github.com/zenlm/zen-family](https://github.com/zenlm/zen-family) for full documentation, training scripts, and configuration files.
 
 
 
 
62
 
63
+ ```bash
64
+ git clone https://github.com/zenlm/zen-family
65
+ cd zen-family/training
66
+ pip install -r requirements.txt
67
+ python train.py --config configs/zen-pro-8b.yaml
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
68
  ```
69
 
70
+ ## Related Repositories
 
 
 
 
 
 
71
 
72
+ | Repo | Description |
73
+ |------|-------------|
74
+ | [zenlm/zen-family](https://huggingface.co/zenlm/zen-family) | Model family overview |
75
+ | [zenlm/zen-nano-600m-instruct](https://huggingface.co/zenlm/zen-nano-600m-instruct) | Zen Nano 0.6B |
76
+ | [zenlm/zen-pro-8b-instruct](https://huggingface.co/zenlm/zen-pro-8b-instruct) | Zen Pro — 8B |
77
+ | [zenlm/zen-max-235b-a22b-instruct](https://huggingface.co/zenlm/zen-max-235b-a22b-instruct) | Zen MAX — 235B MoE |
78
 
79
+ ## License
 
 
 
 
 
 
 
 
 
80
 
81
  Apache 2.0