Trouter-Library commited on
Commit
df07e20
·
verified ·
1 Parent(s): f7e8b30

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +65 -108
README.md CHANGED
@@ -4,15 +4,11 @@ base_model: meta-llama/Llama-2-7b-hf
4
  tags:
5
  - text-generation
6
  - conversational
7
- - assistant
8
- - safety
9
  - llama-2
10
- - autotrain
11
  - autotrain_compatible
 
12
  language:
13
  - en
14
- datasets:
15
- - custom
16
  pipeline_tag: text-generation
17
  library_name: transformers
18
  model-index:
@@ -40,105 +36,84 @@ model-index:
40
  name: Win Rate %
41
  - task:
42
  type: text-generation
43
- name: Coding
44
  dataset:
45
  name: HumanEval
46
  type: humaneval
47
  metrics:
48
- - type: pass_at_1
49
  value: 42.3
50
- name: Pass@1 Score
51
  widget:
52
- - text: "How do I learn Python programming?"
53
- example_title: "Programming Help"
54
- - text: "Explain quantum computing in simple terms"
55
  example_title: "Technical Explanation"
56
- - text: "Write a short story about a robot"
57
- example_title: "Creative Writing"
58
  ---
59
 
60
  <div align="center">
 
61
  <img src="https://imgur.com/aUIJXf7.png" alt="Helion-V1 Logo" width="100%"/>
 
62
  </div>
63
 
64
  ---
65
 
66
  # Helion-V1.5
67
 
68
- Helion-V1.5 is an improved conversational AI assistant fine-tuned with HuggingFace AutoTrain. Built on Llama-2-7B, it combines helpfulness, safety, and performance with enhanced training techniques.
 
 
 
 
69
 
70
  ## Model Details
71
 
72
- ### Model Description
 
 
 
 
73
 
74
- - **Developed by:** DeepXR
75
- - **Model type:** Causal Language Model (Decoder-only Transformer)
76
- - **Base model:** meta-llama/Llama-2-7b-hf
77
- - **Language(s):** English
78
- - **License:** Apache 2.0
79
- - **Finetuned from:** Llama-2-7B using LoRA/QLoRA
80
- - **Training method:** HuggingFace AutoTrain
81
- - **Parameters:** 7 billion
82
- - **Context length:** 4096 tokens
83
 
84
- ### Model Architecture
 
 
 
 
 
 
 
85
 
86
- | Component | Specification |
87
- |-----------|--------------|
88
- | Architecture | Llama-2 (Transformer Decoder) |
89
- | Layers | 32 |
90
  | Hidden Size | 4096 |
 
91
  | Attention Heads | 32 |
92
- | Head Dimension | 128 |
93
  | Intermediate Size | 11008 |
94
- | Vocabulary Size | 32000 |
95
- | Position Embeddings | Rotary (RoPE) |
96
- | Normalization | RMSNorm |
97
- | Activation | SwiGLU |
 
 
 
 
 
98
 
99
- ### Training Configuration
100
 
101
- **LoRA Parameters:**
102
- - Rank (r): 64
103
- - Alpha: 128
104
- - Dropout: 0.05
105
- - Target Modules: q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj
106
-
107
- **Training Hyperparameters:**
108
- - Learning Rate: 2e-5
109
- - Batch Size: 4 per device
110
- - Gradient Accumulation: 8 steps
111
- - Epochs: 3
112
- - Warmup Steps: 100
113
- - Max Sequence Length: 4096
114
- - Optimizer: AdamW
115
- - Scheduler: Cosine with warmup
116
- - Mixed Precision: bfloat16
117
-
118
- **Hardware:**
119
- - Training: 1x NVIDIA A100 (40GB)
120
- - Training Time: ~6 hours
121
- - Total Steps: ~5,000
122
-
123
- ## Intended Use
124
-
125
- ### Primary Use Cases
126
-
127
- ✅ **General Conversation** - Natural, helpful dialogue
128
- ✅ **Question Answering** - Accurate information retrieval
129
- ✅ **Code Assistance** - Programming help and debugging
130
- ✅ **Writing Support** - Content creation and editing
131
- ✅ **Education** - Explanations and tutoring
132
- ✅ **Problem Solving** - Logical reasoning and analysis
133
-
134
- ### Out-of-Scope Use
135
-
136
- ❌ **Medical Advice** - Not qualified for medical diagnosis/treatment
137
- ❌ **Legal Advice** - Not a substitute for legal counsel
138
- ❌ **Financial Advice** - Not for investment decisions
139
- ❌ **Harmful Content** - Will refuse to generate dangerous content
140
- ❌ **Impersonation** - Not for pretending to be real people
141
- ❌ **Misinformation** - Not for spreading false information
142
 
143
  ## How to Use
144
 
@@ -289,51 +264,33 @@ The model may exhibit biases present in the training data. We've implemented:
289
  - User feedback integration
290
  - Ongoing bias mitigation efforts
291
 
292
- ## Ethical Considerations
293
-
294
- ### Responsible Use
295
 
296
  Users should:
297
- - Verify important information from authoritative sources
298
- - Monitor outputs for accuracy in production
299
- - Provide proper attribution for AI-generated content
300
- - Implement appropriate safeguards for your use case
301
- - Follow applicable laws and regulations
302
-
303
- ### Environmental Impact
304
-
305
- - **Training CO2 Emissions:** ~15 kg CO2eq (estimated)
306
- - **Training Energy:** ~30 kWh
307
- - **Compute Used:** 1x A100 GPU for 6 hours
308
 
309
  ## Citation
310
 
311
  ```bibtex
312
- @misc{helion-v1.5,
313
  author = {DeepXR},
314
- title = {Helion-V1.5: An Enhanced Conversational AI Assistant},
315
  year = {2024},
316
  publisher = {HuggingFace},
317
- howpublished = {\url{https://huggingface.co/DeepXR/Helion-V1.5}},
318
- note = {Trained with HuggingFace AutoTrain}
319
  }
320
  ```
321
 
322
- ## Model Card Authors
323
-
324
- DeepXR Team
325
-
326
-
327
- ## Acknowledgments
328
 
329
- - Built on Meta's Llama-2 foundation
330
- - Trained using HuggingFace AutoTrain
331
- - Community feedback and testing
332
- - Open-source ecosystem support
333
 
334
  ---
335
 
336
- **Version:** 1.5.0
337
- **Release Date:** November 2024
338
- **Status:** Production Ready
339
- **AutoTrain Compatible:** Yes
 
4
  tags:
5
  - text-generation
6
  - conversational
 
 
7
  - llama-2
 
8
  - autotrain_compatible
9
+ - function-calling
10
  language:
11
  - en
 
 
12
  pipeline_tag: text-generation
13
  library_name: transformers
14
  model-index:
 
36
  name: Win Rate %
37
  - task:
38
  type: text-generation
39
+ name: Code Generation
40
  dataset:
41
  name: HumanEval
42
  type: humaneval
43
  metrics:
44
+ - type: pass@1
45
  value: 42.3
46
+ name: Pass@1
47
  widget:
48
+ - text: "Explain the difference between machine learning and deep learning"
 
 
49
  example_title: "Technical Explanation"
50
+ - text: "Write a Python function to calculate fibonacci numbers"
51
+ example_title: "Code Generation"
52
  ---
53
 
54
  <div align="center">
55
+
56
  <img src="https://imgur.com/aUIJXf7.png" alt="Helion-V1 Logo" width="100%"/>
57
+
58
  </div>
59
 
60
  ---
61
 
62
  # Helion-V1.5
63
 
64
+ <div align="center">
65
+ <img src="https://huggingface.co/datasets/huggingface/badges/resolve/main/powered-by-autotrain.svg" alt="Powered by AutoTrain"/>
66
+ </div>
67
+
68
+ **Helion-V1.5** is a 7B parameter conversational AI model fine-tuned from Llama-2 using QLoRA. It delivers improved performance over Helion-V1 with enhanced instruction following, code generation, and multi-turn dialogue capabilities.
69
 
70
  ## Model Details
71
 
72
+ **Architecture:** Llama-2-7B with LoRA adapters
73
+ **Parameters:** 7 billion (base) + 67M (LoRA)
74
+ **Context Length:** 4096 tokens
75
+ **Training:** QLoRA (4-bit) fine-tuning on high-quality instruction data
76
+ **License:** Apache 2.0
77
 
78
+ ### Key Improvements over Helion-V1
 
 
 
 
 
 
 
 
79
 
80
+ | Feature | Helion-V1 | Helion-V1.5 | Improvement |
81
+ |---------|-----------|-------------|-------------|
82
+ | **MT-Bench Score** | 6.8 | 7.2 | +5.9% |
83
+ | **AlpacaEval Win Rate** | 72.3% | 78.5% | +8.6% |
84
+ | **HumanEval Pass@1** | 38.1% | 42.3% | +11.0% |
85
+ | **Avg Response Time** | 2.3s | 1.8s | -21.7% |
86
+ | **Function Calling** | ❌ | ✅ | New |
87
+ | **Streaming Support** | Basic | Full | Enhanced |
88
 
89
+ ### Technical Specifications
90
+
91
+ | Component | Value |
92
+ |-----------|-------|
93
  | Hidden Size | 4096 |
94
+ | Layers | 32 |
95
  | Attention Heads | 32 |
 
96
  | Intermediate Size | 11008 |
97
+ | Vocabulary | 32000 tokens |
98
+ | Position Encoding | RoPE |
99
+ | Precision | bfloat16 |
100
+
101
+ **LoRA Configuration:**
102
+ - Rank: 64
103
+ - Alpha: 128
104
+ - Target Modules: All linear layers (q,k,v,o,gate,up,down)
105
+ - Dropout: 0.05
106
 
107
+ ## Performance Benchmarks
108
 
109
+ | Benchmark | Score | Category |
110
+ |-----------|-------|----------|
111
+ | MT-Bench | 7.2/10 | Multi-turn conversation |
112
+ | AlpacaEval | 78.5% | Instruction following |
113
+ | HumanEval | 42.3% | Code generation |
114
+ | GSM8K | 35.7% | Mathematical reasoning |
115
+ | TruthfulQA | 51.2% | Factual accuracy |
116
+ | MMLU | 48.9% | Knowledge |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
117
 
118
  ## How to Use
119
 
 
264
  - User feedback integration
265
  - Ongoing bias mitigation efforts
266
 
267
+ ## Responsible Use
 
 
268
 
269
  Users should:
270
+ - Verify critical information from authoritative sources
271
+ - Implement appropriate safeguards for production use
272
+ - Monitor outputs for accuracy and appropriateness
273
+ - Comply with applicable laws and regulations
274
+ - Provide proper attribution for AI-generated content
 
 
 
 
 
 
275
 
276
  ## Citation
277
 
278
  ```bibtex
279
+ @misc{helion-v1.5-2024,
280
  author = {DeepXR},
281
+ title = {Helion-V1.5: Enhanced Conversational AI},
282
  year = {2024},
283
  publisher = {HuggingFace},
284
+ url = {https://huggingface.co/DeepXR/Helion-V1.5}
 
285
  }
286
  ```
287
 
288
+ ## Contact
 
 
 
 
 
289
 
290
+ - **Repository:** https://huggingface.co/DeepXR/Helion-V1.5
291
+ - **Issues:** https://huggingface.co/DeepXR/Helion-V1.5/discussions
292
+ - **Email:** contact@deepxr.ai
 
293
 
294
  ---
295
 
296
+ **Model Version:** 1.5.0 | **Release:** November 2024 | **Status:** Production Ready