Trouter-Library commited on
Commit
1a02a20
·
verified ·
1 Parent(s): 5b6df24

Create model-card.yaml

Browse files
Files changed (1) hide show
  1. model-card.yaml +209 -0
model-card.yaml ADDED
@@ -0,0 +1,209 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Model Card for Helion-V2
2
+
3
+ model_details:
4
+ name: Helion-V2
5
+ version: 2.0
6
+ description: A 7.2B parameter large language model optimized for daily use, featuring strong performance across reasoning, coding, and conversational tasks.
7
+ organization: DeepXR
8
+ license: Apache-2.0
9
+ release_date: 2024-11-15
10
+ model_type: Causal Language Model
11
+ architecture: Decoder-only Transformer
12
+ parameters: 7200000000
13
+ precision: bfloat16
14
+
15
+ intended_use:
16
+ primary_use: General-purpose language model for conversational AI, code generation, and text completion
17
+ intended_users: Developers, researchers, businesses, and individuals
18
+ use_cases:
19
+ - Conversational AI assistants
20
+ - Code generation and debugging
21
+ - Content creation and writing assistance
22
+ - Question answering systems
23
+ - Educational tutoring
24
+ - Data analysis and summarization
25
+ out_of_scope:
26
+ - Medical diagnosis or treatment recommendations
27
+ - Legal advice or contractual interpretation
28
+ - Financial investment decisions
29
+ - Safety-critical systems
30
+ - Autonomous decision-making without human oversight
31
+ - Generating harmful or illegal content
32
+
33
+ factors:
34
+ languages:
35
+ - English (primary)
36
+ - Spanish
37
+ - French
38
+ - German
39
+ - Italian
40
+ - Portuguese
41
+ - Dutch
42
+ - Russian
43
+ - Chinese
44
+ - Japanese
45
+ - Korean
46
+ - Arabic
47
+ - Hindi
48
+ demographic_considerations:
49
+ - Model trained on diverse global data sources
50
+ - Performance may vary across languages and cultural contexts
51
+ - English language performance is strongest
52
+ technical_limitations:
53
+ - Context length limited to 8,192 tokens
54
+ - Knowledge cutoff at October 2024
55
+ - May generate plausible but incorrect information
56
+ - Performance degrades with highly specialized technical content
57
+
58
+ metrics:
59
+ evaluation_benchmarks:
60
+ MMLU:
61
+ score: 64.2
62
+ type: accuracy
63
+ description: Massive Multitask Language Understanding (5-shot)
64
+ HumanEval:
65
+ score: 48.2
66
+ type: pass@1
67
+ description: Code generation benchmark
68
+ HellaSwag:
69
+ score: 80.5
70
+ type: accuracy
71
+ description: Commonsense reasoning (10-shot)
72
+ TruthfulQA:
73
+ score: 52.1
74
+ type: mc2_accuracy
75
+ description: Truthfulness and factual accuracy
76
+ GSM8K:
77
+ score: 68.7
78
+ type: accuracy
79
+ description: Grade school math (8-shot, chain-of-thought)
80
+ ARC_Challenge:
81
+ score: 58.3
82
+ type: accuracy
83
+ description: AI2 Reasoning Challenge (25-shot)
84
+ MT_Bench:
85
+ score: 7.85
86
+ type: rating
87
+ description: Multi-turn conversation quality
88
+ ToxiGen:
89
+ score: 0.08
90
+ type: toxicity
91
+ description: Toxicity detection (lower is better)
92
+
93
+ training_data:
94
+ description: Diverse corpus of approximately 2.5 trillion tokens
95
+ sources:
96
+ - Web documents and articles (45%)
97
+ - Code repositories (20%)
98
+ - Books and educational materials (15%)
99
+ - Scientific papers (10%)
100
+ - Instruction-following datasets (10%)
101
+ preprocessing:
102
+ - Quality filtering with perplexity thresholds
103
+ - Deduplication using MinHash LSH
104
+ - Toxicity filtering with Perspective API
105
+ - PII removal and scrubbing
106
+ - License compliance verification
107
+ knowledge_cutoff: 2024-10-31
108
+
109
+ training_procedure:
110
+ optimizer: AdamW
111
+ learning_rate: 0.0003
112
+ lr_schedule: Cosine with warmup
113
+ warmup_steps: 2000
114
+ batch_size: 4194304 tokens
115
+ sequence_length: 8192
116
+ training_steps: 600000
117
+ epochs: 3
118
+ precision: bfloat16
119
+ hardware: 128x NVIDIA H100 80GB GPUs
120
+ framework: PyTorch 2.1.2
121
+ distributed_strategy: DeepSpeed ZeRO-3
122
+ training_time: 21 days
123
+ compute_hours: 64512 GPU-hours
124
+
125
+ evaluation_data:
126
+ datasets:
127
+ - MMLU (Massive Multitask Language Understanding)
128
+ - HumanEval (Code generation)
129
+ - MBPP (Mostly Basic Python Problems)
130
+ - HellaSwag (Commonsense reasoning)
131
+ - PIQA (Physical commonsense)
132
+ - WinoGrande (Coreference resolution)
133
+ - ARC (AI2 Reasoning Challenge)
134
+ - TruthfulQA (Truthfulness)
135
+ - GSM8K (Math reasoning)
136
+ - MATH (Advanced mathematics)
137
+ - BBH (Big-Bench Hard)
138
+ - MT-Bench (Multi-turn conversation)
139
+ - AlpacaEval (Instruction following)
140
+ - ToxiGen (Toxicity detection)
141
+ - CrowS-Pairs (Bias detection)
142
+
143
+ ethical_considerations:
144
+ bias_analysis:
145
+ - Training data may contain societal biases
146
+ - Gender, racial, cultural, and geographic biases possible
147
+ - Users should validate outputs for fairness
148
+ - Ongoing monitoring and evaluation recommended
149
+ risks_and_limitations:
150
+ - May generate plausible but incorrect information
151
+ - Can be misused for generating harmful content
152
+ - Should not replace professional advice
153
+ - Requires appropriate safeguards for production use
154
+ mitigation_strategies:
155
+ - Content filtering during training
156
+ - Safety fine-tuning with human feedback
157
+ - Built-in refusal mechanisms
158
+ - Comprehensive safety documentation
159
+ - Rate limiting and monitoring recommended
160
+ safety_features:
161
+ - Low toxicity score (0.08 on ToxiGen)
162
+ - High truthfulness (52.1% on TruthfulQA)
163
+ - Content moderation capabilities
164
+ - PII detection and redaction
165
+ - Crisis detection and resource routing
166
+
167
+ environmental_impact:
168
+ carbon_emissions: 8500 kg CO2eq
169
+ compute_region: United States
170
+ hardware: 128x NVIDIA H100 80GB GPUs
171
+ training_hours: 64512 GPU-hours
172
+ estimated_cost: 450000 USD
173
+ mitigation:
174
+ - Support for quantization (4-bit, 8-bit)
175
+ - Efficient inference optimization
176
+ - Carbon offset programs recommended
177
+
178
+ citations:
179
+ bibtex: |
180
+ @misc{helion-v2-2024,
181
+ title={Helion-V2: An Efficient and Truthful Large Language Model for Daily Use},
182
+ author={DeepXR Team},
183
+ year={2024},
184
+ month={November},
185
+ publisher={HuggingFace},
186
+ url={https://huggingface.co/DeepXR/Helion-V2},
187
+ note={7.2B parameter decoder-only transformer with grouped query attention}
188
+ }
189
+
190
+ contact:
191
+ email: contact@deepxr.ai
192
+ github: https://github.com/DeepXR/Helion-V2
193
+ issues: https://github.com/DeepXR/Helion-V2/issues
194
+ twitter: "@DeepXR_AI"
195
+ discord: https://discord.gg/deepxr
196
+
197
+ license:
198
+ name: Apache License 2.0
199
+ url: https://www.apache.org/licenses/LICENSE-2.0
200
+ commercial_use: true
201
+ modifications_allowed: true
202
+ distribution_allowed: true
203
+ patent_use: true
204
+
205
+ model_card_version: 1.0
206
+ model_card_authors:
207
+ - DeepXR Team
208
+ model_card_contact: contact@deepxr.ai
209
+ model_card_date: 2024-11-15