Updating model files
Browse files
README.md
CHANGED
|
@@ -1,6 +1,17 @@
|
|
| 1 |
---
|
| 2 |
license: other
|
| 3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 4 |
|
| 5 |
# WizardLM: An Instruction-following LLM Using Evol-Instruct
|
| 6 |
|
|
@@ -23,6 +34,17 @@ This repo contains the full unquantised model files in HF format for GPU inferen
|
|
| 23 |
* [4bit GGML models for CPU inference](https://huggingface.co/TheBloke/wizardLM-7B-GGML)
|
| 24 |
* [4bit GPTQ models for GPU inference](https://huggingface.co/TheBloke/wizardLM-7B-GPTQ)
|
| 25 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 26 |
# Original model info
|
| 27 |
|
| 28 |
## Full details in the model's Github page
|
|
@@ -36,4 +58,4 @@ Evol-Instruct is a novel method using LLMs instead of humans to automatically ma
|
|
| 36 |
Although on our complexity-balanced test set, WizardLM-7B outperforms ChatGPT in the high-complexity instructions, it still lag behind ChatGPT on the entire test set, and we also consider WizardLM to still be in a baby state. This repository will continue to improve WizardLM, train on larger scales, add more training data, and innovate more advanced large-model training methods.
|
| 37 |
|
| 38 |

|
| 39 |
-

|
|
|
|
| 1 |
---
|
| 2 |
license: other
|
| 3 |
---
|
| 4 |
+
<div style="width: 100%;">
|
| 5 |
+
<img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;">
|
| 6 |
+
</div>
|
| 7 |
+
<div style="display: flex; justify-content: space-between; width: 100%;">
|
| 8 |
+
<div style="display: flex; flex-direction: column; align-items: flex-start;">
|
| 9 |
+
<p><a href="https://discord.gg/UBgz4VXf">Chat & support: my new Discord server</a></p>
|
| 10 |
+
</div>
|
| 11 |
+
<div style="display: flex; flex-direction: column; align-items: flex-end;">
|
| 12 |
+
<p><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? Patreon coming soon!</a></p>
|
| 13 |
+
</div>
|
| 14 |
+
</div>
|
| 15 |
|
| 16 |
# WizardLM: An Instruction-following LLM Using Evol-Instruct
|
| 17 |
|
|
|
|
| 34 |
* [4bit GGML models for CPU inference](https://huggingface.co/TheBloke/wizardLM-7B-GGML)
|
| 35 |
* [4bit GPTQ models for GPU inference](https://huggingface.co/TheBloke/wizardLM-7B-GPTQ)
|
| 36 |
|
| 37 |
+
## Want to support my work?
|
| 38 |
+
|
| 39 |
+
I've had a lot of people ask if they can contribute. I love providing models and helping people, but it is starting to rack up pretty big cloud computing bills.
|
| 40 |
+
|
| 41 |
+
So if you're able and willing to contribute, it'd be most gratefully received and will help me to keep providing models, and work on various AI projects.
|
| 42 |
+
|
| 43 |
+
Donaters will get priority support on any and all AI/LLM/model questions, and I'll gladly quantise any model you'd like to try.
|
| 44 |
+
|
| 45 |
+
* Patreon: coming soon! (just awaiting approval)
|
| 46 |
+
* Ko-Fi: https://ko-fi.com/TheBlokeAI
|
| 47 |
+
* Discord: https://discord.gg/UBgz4VXf
|
| 48 |
# Original model info
|
| 49 |
|
| 50 |
## Full details in the model's Github page
|
|
|
|
| 58 |
Although on our complexity-balanced test set, WizardLM-7B outperforms ChatGPT in the high-complexity instructions, it still lag behind ChatGPT on the entire test set, and we also consider WizardLM to still be in a baby state. This repository will continue to improve WizardLM, train on larger scales, add more training data, and innovate more advanced large-model training methods.
|
| 59 |
|
| 60 |

|
| 61 |
+

|