File size: 3,406 Bytes
72b94b9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
{
  "model_name": "Helion-V2",
  "model_version": "2.0",
  "model_type": "causal-lm",
  "architecture": "helion",
  "organization": "DeepXR",
  "release_date": "2024-11-15",
  "license": "Apache-2.0",
  "homepage": "https://huggingface.co/DeepXR/Helion-V2",
  "repository": "https://github.com/DeepXR/Helion-V2",
  "documentation": "https://docs.deepxr.ai/helion-v2",
  "parameters": {
    "total": 7200000000,
    "trainable": 7200000000,
    "non_trainable": 0
  },
  "model_size": {
    "fp32": "28.8 GB",
    "fp16": "14.4 GB",
    "bfloat16": "14.4 GB",
    "int8": "7.2 GB",
    "int4": "3.6 GB"
  },
  "context_length": 8192,
  "vocabulary_size": 32768,
  "languages": [
    "en",
    "es",
    "fr",
    "de",
    "it",
    "pt",
    "nl",
    "ru",
    "zh",
    "ja",
    "ko",
    "ar",
    "hi"
  ],
  "primary_language": "en",
  "tasks": [
    "text-generation",
    "conversational",
    "code-generation",
    "question-answering",
    "summarization",
    "translation",
    "creative-writing"
  ],
  "framework": {
    "name": "transformers",
    "version": "4.40.0+",
    "backend": "pytorch",
    "backend_version": "2.1.0+"
  },
  "training": {
    "dataset_tokens": 2500000000000,
    "training_steps": 600000,
    "epochs": 3,
    "batch_size": 4194304,
    "learning_rate": 0.0003,
    "optimizer": "AdamW",
    "hardware": "128x NVIDIA H100 80GB",
    "training_time_days": 21,
    "knowledge_cutoff": "2024-10-31"
  },
  "benchmarks": {
    "mmlu": {
      "score": 64.2,
      "metric": "accuracy",
      "shots": 5
    },
    "humaneval": {
      "score": 48.2,
      "metric": "pass@1",
      "shots": 0
    },
    "hellaswag": {
      "score": 80.5,
      "metric": "accuracy",
      "shots": 10
    },
    "truthfulqa": {
      "score": 52.1,
      "metric": "mc2_accuracy",
      "shots": 0
    },
    "gsm8k": {
      "score": 68.7,
      "metric": "accuracy",
      "shots": 8
    },
    "arc_challenge": {
      "score": 58.3,
      "metric": "accuracy",
      "shots": 25
    },
    "mt_bench": {
      "score": 7.85,
      "metric": "rating",
      "shots": 0
    }
  },
  "safety": {
    "toxigen_score": 0.08,
    "crows_pairs_bias": 54.2,
    "content_filtering": true,
    "pii_detection": true,
    "crisis_detection": true,
    "safety_classifiers": [
      "hate_speech",
      "violence",
      "sexual_content",
      "self_harm",
      "illegal_activity",
      "child_safety"
    ]
  },
  "deployment": {
    "recommended_gpu": [
      "NVIDIA A100",
      "NVIDIA H100",
      "NVIDIA RTX 4090",
      "NVIDIA A6000"
    ],
    "minimum_vram_gb": 16,
    "recommended_vram_gb": 24,
    "supports_quantization": true,
    "quantization_formats": [
      "int8",
      "int4",
      "GPTQ",
      "AWQ",
      "GGUF"
    ],
    "inference_frameworks": [
      "transformers",
      "vllm",
      "llama.cpp",
      "text-generation-inference"
    ]
  },
  "contact": {
    "email": "contact@deepxr.ai",
    "github": "https://github.com/DeepXR/Helion-V2",
    "twitter": "@DeepXR_AI",
    "discord": "https://discord.gg/deepxr"
  },
  "tags": [
    "text-generation",
    "conversational",
    "code",
    "instruction-following",
    "pytorch",
    "causal-lm",
    "llm",
    "reasoning",
    "multilingual",
    "apache-2.0",
    "production-ready",
    "safe-ai"
  ],
  "checksum": {
    "algorithm": "SHA256",
    "model_files": "computed_on_upload"
  }
}