Commit
·
d82a6a4
1
Parent(s):
96379e3
Update README.md
Browse files
README.md
CHANGED
|
@@ -34,13 +34,13 @@ tags:
|
|
| 34 |
|
| 35 |
*Image drawn by GPT-4 DALL·E 3* TL;DR: Perhaps this 7B model, better than all existing models <= 33B, in most quantitative evaluations...
|
| 36 |
|
| 37 |
-
**Some problems with llama.cpp on GPT2Tokenizer, gotta fix soon...**
|
| 38 |
-
|
| 39 |
# Please Stop Using WRONG unofficial quant models unless you know what you're doing
|
| 40 |
|
| 41 |
-
GPTQ quants require a good dataset for calibration, and the default C4 dataset is not capable
|
|
|
|
|
|
|
|
|
|
| 42 |
|
| 43 |
-
**Some problems with llama.cpp on GPT2Tokenizer, gotta fix soon...**
|
| 44 |
|
| 45 |
## Read Me:
|
| 46 |
|
|
@@ -91,7 +91,8 @@ Hard acc:48.03
|
|
| 91 |
**Zero-shot ACC 0.5921152388172858** (Outperforms WizardMath-7B and Qwen-7B)
|
| 92 |
|
| 93 |
|
| 94 |
-
**
|
|
|
|
| 95 |
|
| 96 |
## 请读我:
|
| 97 |
|
|
|
|
| 34 |
|
| 35 |
*Image drawn by GPT-4 DALL·E 3* TL;DR: Perhaps this 7B model, better than all existing models <= 33B, in most quantitative evaluations...
|
| 36 |
|
|
|
|
|
|
|
| 37 |
# Please Stop Using WRONG unofficial quant models unless you know what you're doing
|
| 38 |
|
| 39 |
+
GPTQ quants require a good dataset for calibration, and the default C4 dataset is not capable.
|
| 40 |
+
|
| 41 |
+
**llama.cpp GGUF models**
|
| 42 |
+
GPT2Tokenizer fixed by [Kerfuffle](https://github.com/KerfuffleV2) on [https://github.com/ggerganov/llama.cpp/pull/3743](https://github.com/ggerganov/llama.cpp/pull/3743), new models to be reuploaded.
|
| 43 |
|
|
|
|
| 44 |
|
| 45 |
## Read Me:
|
| 46 |
|
|
|
|
| 91 |
**Zero-shot ACC 0.5921152388172858** (Outperforms WizardMath-7B and Qwen-7B)
|
| 92 |
|
| 93 |
|
| 94 |
+
**llama.cpp GGUF models**
|
| 95 |
+
GPT2Tokenizer 支持由 [Kerfuffle](https://github.com/KerfuffleV2) 修复于 [https://github.com/ggerganov/llama.cpp/pull/3743](https://github.com/ggerganov/llama.cpp/pull/3743),新模型稍后上传。
|
| 96 |
|
| 97 |
## 请读我:
|
| 98 |
|