Trouter-Library commited on
Commit
21434ad
·
verified ·
1 Parent(s): d6f46cf

Create model_variants.json

Browse files
Files changed (1) hide show
  1. model_variants.json +240 -0
model_variants.json ADDED
@@ -0,0 +1,240 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_family": "Helion",
3
+ "version": "2.5",
4
+ "release_type": "research_and_development",
5
+ "variants": {
6
+ "base": {
7
+ "name": "Helion-2.5-Rnd",
8
+ "full_name": "DeepXR/Helion-2.5-Rnd",
9
+ "description": "Base research model with full precision (BF16)",
10
+ "parameters": "70B",
11
+ "precision": "bfloat16",
12
+ "context_length": 131072,
13
+ "status": "active",
14
+ "recommended_use": [
15
+ "Research",
16
+ "Development",
17
+ "High-accuracy inference"
18
+ ],
19
+ "hardware_requirements": {
20
+ "min_vram_gb": 160,
21
+ "min_gpus": 2,
22
+ "recommended_gpu": "A100 80GB"
23
+ }
24
+ },
25
+ "instruct": {
26
+ "name": "Helion-2.5-Rnd-Instruct",
27
+ "full_name": "DeepXR/Helion-2.5-Rnd-Instruct",
28
+ "description": "Instruction-tuned variant optimized for following instructions",
29
+ "parameters": "70B",
30
+ "precision": "bfloat16",
31
+ "context_length": 131072,
32
+ "status": "planned",
33
+ "recommended_use": [
34
+ "Instruction following",
35
+ "Task completion",
36
+ "Structured outputs"
37
+ ],
38
+ "fine_tuning": {
39
+ "type": "supervised",
40
+ "data_focus": "instruction_pairs"
41
+ }
42
+ },
43
+ "chat": {
44
+ "name": "Helion-2.5-Rnd-Chat",
45
+ "full_name": "DeepXR/Helion-2.5-Rnd-Chat",
46
+ "description": "Conversational variant optimized for multi-turn dialogue",
47
+ "parameters": "70B",
48
+ "precision": "bfloat16",
49
+ "context_length": 131072,
50
+ "status": "planned",
51
+ "recommended_use": [
52
+ "Conversational AI",
53
+ "Customer service",
54
+ "Interactive applications"
55
+ ],
56
+ "fine_tuning": {
57
+ "type": "rlhf",
58
+ "data_focus": "conversational_data"
59
+ }
60
+ },
61
+ "code": {
62
+ "name": "Helion-2.5-Rnd-Code",
63
+ "full_name": "DeepXR/Helion-2.5-Rnd-Code",
64
+ "description": "Code-specialized variant with enhanced programming capabilities",
65
+ "parameters": "70B",
66
+ "precision": "bfloat16",
67
+ "context_length": 131072,
68
+ "status": "planned",
69
+ "recommended_use": [
70
+ "Code generation",
71
+ "Code review",
72
+ "Bug fixing",
73
+ "Documentation"
74
+ ],
75
+ "fine_tuning": {
76
+ "type": "supervised",
77
+ "data_focus": "code_repositories"
78
+ },
79
+ "enhanced_languages": [
80
+ "Python",
81
+ "JavaScript",
82
+ "TypeScript",
83
+ "Rust",
84
+ "Go",
85
+ "Java"
86
+ ]
87
+ },
88
+ "math": {
89
+ "name": "Helion-2.5-Rnd-Math",
90
+ "full_name": "DeepXR/Helion-2.5-Rnd-Math",
91
+ "description": "Mathematics-specialized variant for advanced problem solving",
92
+ "parameters": "70B",
93
+ "precision": "bfloat16",
94
+ "context_length": 131072,
95
+ "status": "planned",
96
+ "recommended_use": [
97
+ "Mathematical reasoning",
98
+ "Proof generation",
99
+ "Problem solving",
100
+ "Educational applications"
101
+ ],
102
+ "fine_tuning": {
103
+ "type": "supervised",
104
+ "data_focus": "mathematical_proofs"
105
+ }
106
+ }
107
+ },
108
+ "deployment_configurations": {
109
+ "production": {
110
+ "description": "Production-ready configuration with optimizations",
111
+ "settings": {
112
+ "tensor_parallel_size": 4,
113
+ "gpu_memory_utilization": 0.95,
114
+ "max_batch_size": 32,
115
+ "enable_prefix_caching": true,
116
+ "enable_chunked_prefill": true
117
+ }
118
+ },
119
+ "development": {
120
+ "description": "Development configuration for testing",
121
+ "settings": {
122
+ "tensor_parallel_size": 2,
123
+ "gpu_memory_utilization": 0.85,
124
+ "max_batch_size": 8,
125
+ "enable_prefix_caching": false,
126
+ "enable_chunked_prefill": false
127
+ }
128
+ },
129
+ "research": {
130
+ "description": "Research configuration for experimentation",
131
+ "settings": {
132
+ "tensor_parallel_size": 2,
133
+ "gpu_memory_utilization": 0.90,
134
+ "max_batch_size": 4,
135
+ "enable_prefix_caching": false,
136
+ "enable_chunked_prefill": false,
137
+ "enable_logging": true
138
+ }
139
+ }
140
+ },
141
+ "comparison_matrix": {
142
+ "base_vs_instruct": {
143
+ "base_advantages": [
144
+ "More flexible for fine-tuning",
145
+ "Better for creative tasks",
146
+ "Less constrained outputs"
147
+ ],
148
+ "instruct_advantages": [
149
+ "Better instruction following",
150
+ "More structured outputs",
151
+ "Improved task completion"
152
+ ]
153
+ },
154
+ "base_vs_chat": {
155
+ "base_advantages": [
156
+ "Better for single-turn tasks",
157
+ "More diverse outputs",
158
+ "Flexible formatting"
159
+ ],
160
+ "chat_advantages": [
161
+ "Better conversation coherence",
162
+ "Improved context awareness",
163
+ "Natural dialogue flow"
164
+ ]
165
+ }
166
+ },
167
+ "migration_guide": {
168
+ "from_base_to_instruct": {
169
+ "steps": [
170
+ "Update prompt format to instruction style",
171
+ "Adjust temperature (typically lower)",
172
+ "Add explicit task descriptions",
173
+ "Use structured output formats"
174
+ ],
175
+ "example_prompt_change": {
176
+ "base": "Write a function to sort a list",
177
+ "instruct": "### Instruction:\nWrite a Python function that sorts a list in ascending order.\n\n### Response:"
178
+ }
179
+ },
180
+ "from_base_to_chat": {
181
+ "steps": [
182
+ "Convert to chat message format",
183
+ "Add system prompts",
184
+ "Maintain conversation history",
185
+ "Use appropriate message roles"
186
+ ],
187
+ "example_format_change": {
188
+ "base": "Hello, how are you?",
189
+ "chat": [
190
+ {
191
+ "role": "system",
192
+ "content": "You are a helpful assistant."
193
+ },
194
+ {
195
+ "role": "user",
196
+ "content": "Hello, how are you?"
197
+ }
198
+ ]
199
+ }
200
+ }
201
+ },
202
+ "version_history": {
203
+ "2.5.0-rnd": {
204
+ "release_date": "2025-01-30",
205
+ "status": "current",
206
+ "changes": [
207
+ "Initial research release",
208
+ "70B parameter model",
209
+ "131K context with YARN",
210
+ "SafeTensors format (96 shards)",
211
+ "Full precision (BF16)"
212
+ ]
213
+ }
214
+ },
215
+ "roadmap": {
216
+ "upcoming_variants": [
217
+ {
218
+ "name": "Helion-2.5-Rnd-Instruct",
219
+ "expected": "Q2 2025",
220
+ "status": "in_development"
221
+ },
222
+ {
223
+ "name": "Helion-2.5-Rnd-Chat",
224
+ "expected": "Q2 2025",
225
+ "status": "planned"
226
+ },
227
+ {
228
+ "name": "Helion-2.5-Rnd-Code",
229
+ "expected": "Q3 2025",
230
+ "status": "planned"
231
+ }
232
+ ],
233
+ "future_features": [
234
+ "Multi-modal capabilities",
235
+ "Extended context to 256K",
236
+ "Improved multilingual support",
237
+ "Domain-specific variants"
238
+ ]
239
+ }
240
+ }