zeekay commited on
Commit
8530b27
·
verified ·
1 Parent(s): cb7abf4

Update model files

Browse files
Files changed (1) hide show
  1. README.md +20 -128
README.md CHANGED
@@ -1,154 +1,46 @@
1
  ---
2
  license: apache-2.0
3
- tags:
4
- - zen-research
5
- - zen-ai
6
- - hypermodal
7
- - language-model
8
  language:
9
  - en
10
- library_name: transformers
11
  pipeline_tag: text-generation
 
 
 
 
 
12
  ---
13
 
14
- # zen-agent-4b
15
 
16
- 4B parameter tool-calling agent with Model Context Protocol (MCP) support
17
 
18
  ## Model Details
19
 
20
- - **Developed by**: Zen Research Authors
21
- - **Organization**: Zen Research DAO under [Zoo Labs Inc](https://github.com/zenlm) (501(c)(3) Non-Profit)
22
- - **Location**: San Francisco, California, USA
23
- - **Model type**: language-model
24
- - **Architecture**: Qwen3-4B with MCP
25
  - **Parameters**: 4B
26
- - **License**: Apache 2.0
27
- - **Training**: Trained with [Zen Gym](https://github.com/zenlm/zen-gym)
28
- - **Inference**: Optimized for [Zen Engine](https://github.com/zenlm/zen-engine)
29
-
30
- ## 🌟 Zen AI Ecosystem
31
-
32
- This model is part of the **Zen Research** hypermodal AI family - the world's most comprehensive open-source AI ecosystem.
33
-
34
- ### Complete Model Family
35
-
36
- **Language Models:**
37
- - [zen-nano-0.6b](https://huggingface.co/zenlm/zen-nano-0.6b) - 0.6B edge model (44K tokens/sec)
38
- - [zen-eco-4b-instruct](https://huggingface.co/zenlm/zen-eco-4b-instruct) - 4B instruction model
39
- - [zen-eco-4b-thinking](https://huggingface.co/zenlm/zen-eco-4b-thinking) - 4B reasoning model
40
- - [zen-agent-4b](https://huggingface.co/zenlm/zen-agent-4b) - 4B tool-calling agent
41
-
42
- **3D & World Generation:**
43
- - [zen-3d](https://huggingface.co/zenlm/zen-3d) - Controllable 3D asset generation
44
- - [zen-voyager](https://huggingface.co/zenlm/zen-voyager) - Camera-controlled world exploration
45
- - [zen-world](https://huggingface.co/zenlm/zen-world) - Large-scale world simulation
46
-
47
- **Video Generation:**
48
- - [zen-director](https://huggingface.co/zenlm/zen-director) - Text/image-to-video (5B)
49
- - [zen-video](https://huggingface.co/zenlm/zen-video) - Professional video synthesis
50
- - [zen-video-i2v](https://huggingface.co/zenlm/zen-video-i2v) - Image-to-video animation
51
-
52
- **Audio Generation:**
53
- - [zen-musician](https://huggingface.co/zenlm/zen-musician) - Music generation (7B)
54
- - [zen-foley](https://huggingface.co/zenlm/zen-foley) - Video-to-audio Foley effects
55
-
56
- **Infrastructure:**
57
- - [Zen Gym](https://github.com/zenlm/zen-gym) - Unified training platform
58
- - [Zen Engine](https://github.com/zenlm/zen-engine) - High-performance inference
59
 
60
  ## Usage
61
 
62
- ### Quick Start
63
-
64
  ```python
65
  from transformers import AutoModelForCausalLM, AutoTokenizer
66
 
67
- model = AutoModelForCausalLM.from_pretrained("zenlm/zen-agent-4b")
68
- tokenizer = AutoTokenizer.from_pretrained("zenlm/zen-agent-4b")
69
-
70
- inputs = tokenizer("Hello!", return_tensors="pt")\noutputs = model.generate(**inputs)\nprint(tokenizer.decode(outputs[0]))
71
- ```
72
-
73
- ### With Zen Engine
74
-
75
- ```bash
76
- # High-performance inference (44K tokens/sec on M3 Max)
77
- zen-engine serve --model zenlm/zen-agent-4b --port 3690
78
- ```
79
-
80
- ```python
81
- # OpenAI-compatible API
82
- from openai import OpenAI
83
 
84
- client = OpenAI(base_url="http://localhost:3690/v1")
85
- response = client.chat.completions.create(
86
- model="zenlm/zen-agent-4b",
87
- messages=[{"role": "user", "content": "Hello!"}]
88
- )
89
  ```
90
 
91
  ## Training
92
 
93
- Fine-tune with [Zen Gym](https://github.com/zenlm/zen-gym):
94
-
95
- ```bash
96
- git clone https://github.com/zenlm/zen-gym
97
- cd zen-gym
98
-
99
- # LoRA fine-tuning
100
- llamafactory-cli train --config configs/zen_lora.yaml \
101
- --model_name_or_path zenlm/zen-agent-4b
102
-
103
- # GRPO reinforcement learning (40-60% memory reduction)
104
- llamafactory-cli train --config configs/zen_grpo.yaml \
105
- --model_name_or_path zenlm/zen-agent-4b
106
- ```
107
-
108
- Supported methods: LoRA, QLoRA, DoRA, GRPO, GSPO, DPO, PPO, KTO, ORPO, SimPO, Unsloth
109
-
110
- ## Performance
111
-
112
- - **Speed**: 28K tokens/sec (RTX 4090)
113
- - **Memory**: 2.5GB (Q4_K_M) to 8GB (F16)
114
- - **MCP**: Full Model Context Protocol support
115
- - **Tools**: 100+ function calling accuracy
116
-
117
- ## Ethical Considerations
118
-
119
- - **Open Research**: Released under Apache 2.0 for maximum accessibility
120
- - **Environmental Impact**: Optimized for eco-friendly deployment
121
- - **Transparency**: Full training details and model architecture disclosed
122
- - **Safety**: Comprehensive testing and evaluation
123
- - **Non-Profit**: Developed by Zoo Labs Inc (501(c)(3)) for public benefit
124
-
125
- ## Citation
126
-
127
- ```bibtex
128
- @misc{zenzenagent4b2025,
129
- title={zen-agent-4b: 4B parameter tool-calling agent with Model Context Protocol (MCP) support},
130
- author={Zen Research Authors},
131
- year={2025},
132
- publisher={Zoo Labs Inc},
133
- organization={Zen Research DAO},
134
- url={https://huggingface.co/zenlm/zen-agent-4b}
135
- }
136
- ```
137
-
138
- ## Links
139
-
140
- - **Organization**: [github.com/zenlm](https://github.com/zenlm) • [huggingface.co/zenlm](https://huggingface.co/zenlm)
141
- - **Training Platform**: [Zen Gym](https://github.com/zenlm/zen-gym)
142
- - **Inference Engine**: [Zen Engine](https://github.com/zenlm/zen-engine)
143
- - **Parent Org**: [Zoo Labs Inc](https://github.com/zenlm) (501(c)(3) Non-Profit, San Francisco)
144
- - **Contact**: dev@hanzo.ai • +1 (913) 777-4443
145
 
146
  ## License
147
 
148
- Apache License 2.0
149
-
150
- Copyright 2025 Zen Research Authors
151
-
152
- ---
153
-
154
- **Zen Research** - Building open, eco-friendly AI for everyone 🌱
 
1
  ---
2
  license: apache-2.0
 
 
 
 
 
3
  language:
4
  - en
 
5
  pipeline_tag: text-generation
6
+ tags:
7
+ - zen
8
+ - hanzo-ai
9
+ - qwen3
10
+ - agent
11
  ---
12
 
13
+ # zenlm/zen-eco-4b-agent
14
 
15
+ Zen Eco 4B Agent - Tool-calling agent model
16
 
17
  ## Model Details
18
 
19
+ - **Architecture**: Qwen3 base
 
 
 
 
20
  - **Parameters**: 4B
21
+ - **Training**: Fine-tuned with Zen identity
22
+ - **Developer**: Hanzo AI
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
23
 
24
  ## Usage
25
 
 
 
26
  ```python
27
  from transformers import AutoModelForCausalLM, AutoTokenizer
28
 
29
+ model = AutoModelForCausalLM.from_pretrained("zenlm/zen-eco-4b-agent")
30
+ tokenizer = AutoTokenizer.from_pretrained("zenlm/zen-eco-4b-agent")
 
 
 
 
 
 
 
 
 
 
 
 
 
 
31
 
32
+ prompt = "Hello, who are you?"
33
+ inputs = tokenizer(prompt, return_tensors="pt")
34
+ outputs = model.generate(**inputs, max_new_tokens=50)
35
+ response = tokenizer.decode(outputs[0], skip_special_tokens=True)
36
+ print(response)
37
  ```
38
 
39
  ## Training
40
 
41
+ Trained with fixed seed (42) for reproducibility.
42
+ Base model: Qwen3-4B
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
43
 
44
  ## License
45
 
46
+ Apache 2.0