Update README.md
Browse files
README.md
CHANGED
|
@@ -18,7 +18,7 @@ pipeline_tag: text-generation
|
|
| 18 |
|
| 19 |
Welcome to **GenZ**, an advanced Large Language Model (LLM) fine-tuned on the foundation of Meta's open-source Llama V2 70B parameter model. At Bud Ecosystem, we believe in the power of open-source collaboration to drive the advancement of technology at an accelerated pace. Our vision is to democratize access to fine-tuned LLMs, and to that end, we will be releasing a series of models across different parameter counts (7B, 13B, and 70B) and quantizations (32-bit and 4-bit) for the open-source community to use, enhance, and build upon.
|
| 20 |
|
| 21 |
-
<p align="center"><img src="https://raw.githubusercontent.com/BudEcosystem/GenZ/main/assets/
|
| 22 |
|
| 23 |
The smaller quantization version of our models makes them more accessible, enabling their use even on personal computers. This opens up a world of possibilities for developers, researchers, and enthusiasts to experiment with these models and contribute to the collective advancement of language model technology.
|
| 24 |
|
|
@@ -47,6 +47,28 @@ And this isn't the end. It's just the beginning of a journey towards creating mo
|
|
| 47 |
|
| 48 |
---
|
| 49 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 50 |
|
| 51 |
<h2>Getting Started on Hugging Face 🤗</h2>
|
| 52 |
|
|
@@ -157,29 +179,6 @@ Here are the hyperparameters we used for fine-tuning:
|
|
| 157 |
| Precision | FP16 |
|
| 158 |
| Optimizer | AdamW |
|
| 159 |
|
| 160 |
-
---
|
| 161 |
-
|
| 162 |
-
<h2>Evaluations 🎯</h2>
|
| 163 |
-
|
| 164 |
-
Evaluating our model is a key part of our fine-tuning process. It helps us understand how our model is performing and how it stacks up against other models. Here's a look at some of the key evaluations for GenZ 70B:
|
| 165 |
-
|
| 166 |
-
<h3>Benchmark Comparison</h3>
|
| 167 |
-
|
| 168 |
-
We've compared GenZ models to understand the improvements our fine-tuning has achieved.
|
| 169 |
-
|
| 170 |
-
| Model Name | MT Bench | MMLU | Human Eval | Hellaswag | BBH |
|
| 171 |
-
|:----------:|:--------:|:----:|:----------:|:---------:|:----:|
|
| 172 |
-
| Genz 13B | 6.12 | 53.62| 17.68 | 77.38 | 37.76|
|
| 173 |
-
| Genz 13B v2| 6.79 | 53.68| 21.95 | 77.48 | 38.1 |
|
| 174 |
-
| Genz 70B | 7.33 | 70.32| | | |
|
| 175 |
-
|
| 176 |
-
<h3>MT Bench Score</h3>
|
| 177 |
-
|
| 178 |
-
A key evaluation metric we use is the MT Bench score. This score provides a comprehensive assessment of our model's performance across a range of tasks.
|
| 179 |
-
|
| 180 |
-
<p align="center"><img src="https://raw.githubusercontent.com/BudEcosystem/GenZ/main/assets/mt_bench_score.png" width="500"></p>
|
| 181 |
-
|
| 182 |
-
|
| 183 |
---
|
| 184 |
|
| 185 |
<h2>Looking Ahead 👀</h2>
|
|
|
|
| 18 |
|
| 19 |
Welcome to **GenZ**, an advanced Large Language Model (LLM) fine-tuned on the foundation of Meta's open-source Llama V2 70B parameter model. At Bud Ecosystem, we believe in the power of open-source collaboration to drive the advancement of technology at an accelerated pace. Our vision is to democratize access to fine-tuned LLMs, and to that end, we will be releasing a series of models across different parameter counts (7B, 13B, and 70B) and quantizations (32-bit and 4-bit) for the open-source community to use, enhance, and build upon.
|
| 20 |
|
| 21 |
+
<p align="center"><img src="https://raw.githubusercontent.com/BudEcosystem/GenZ/main/assets/mt_bench_compare" width="500"></p>
|
| 22 |
|
| 23 |
The smaller quantization version of our models makes them more accessible, enabling their use even on personal computers. This opens up a world of possibilities for developers, researchers, and enthusiasts to experiment with these models and contribute to the collective advancement of language model technology.
|
| 24 |
|
|
|
|
| 47 |
|
| 48 |
---
|
| 49 |
|
| 50 |
+
<h2>Evaluations 🎯</h2>
|
| 51 |
+
|
| 52 |
+
Evaluating our model is a key part of our fine-tuning process. It helps us understand how our model is performing and how it stacks up against other models. Here's a look at some of the key evaluations for GenZ 70B:
|
| 53 |
+
|
| 54 |
+
<h3>Benchmark Comparison</h3>
|
| 55 |
+
|
| 56 |
+
We've compared GenZ models to understand the improvements our fine-tuning has achieved.
|
| 57 |
+
|
| 58 |
+
| Model Name | MT Bench | MMLU | Human Eval | BBH |
|
| 59 |
+
|:----------:|:--------:|:----:|:----------:|:----:|
|
| 60 |
+
| Genz 13B | 6.12 | 53.62| 17.68 | 37.76|
|
| 61 |
+
| Genz 13B v2| 6.79 | 53.68| 21.95 | 38.1 |
|
| 62 |
+
| Genz 70B | 7.33 | 70.32| 37.8 |54.69 |
|
| 63 |
+
|
| 64 |
+
<h3>MT Bench Score</h3>
|
| 65 |
+
|
| 66 |
+
A key evaluation metric we use is the MT Bench score. This score provides a comprehensive assessment of our model's performance across a range of tasks.
|
| 67 |
+
|
| 68 |
+
<p align="center"><img src="https://raw.githubusercontent.com/BudEcosystem/GenZ/main/assets/mt_bench_score.png" width="500"></p>
|
| 69 |
+
|
| 70 |
+
|
| 71 |
+
---
|
| 72 |
|
| 73 |
<h2>Getting Started on Hugging Face 🤗</h2>
|
| 74 |
|
|
|
|
| 179 |
| Precision | FP16 |
|
| 180 |
| Optimizer | AdamW |
|
| 181 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 182 |
---
|
| 183 |
|
| 184 |
<h2>Looking Ahead 👀</h2>
|