woodchen7 commited on
Commit
fc2ec0b
·
verified ·
1 Parent(s): 3254bd5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -17,7 +17,7 @@ Dedicated to building a more intuitive, comprehensive, and efficient LLMs compre
17
  </h3>
18
 
19
  <p align="center">
20
- 📖 <a href="https://angelslim.readthedocs.io/">Documentation</a>&nbsp&nbsp | &nbsp&nbsp🤗 <a href="https://huggingface.co/AngelSlim">Hugging Face</a>&nbsp&nbsp | &nbsp&nbsp🤖 <a href="https://modelscope.cn/organization/AngelSlim">ModelScope</a>&nbsp&nbsp | &nbsp&nbsp💬 <a href="./docs/source/assets/angel_slim_wechat.png">WeChat</a>
21
  <br>
22
  </p>
23
 
@@ -42,7 +42,7 @@ For more detailed information, please refer to[[AngelSlim]](https://github.com/T
42
 
43
  Benchmark results for HY-1.8B-2Bit equivalent weights on vLLM across **cmmlu**,**ceval**,**arc**,**bbh**,**gsm8k**,**humaneval**,**livecodebench** and **gpqa_diamond**.
44
 
45
- xxx
46
 
47
  | Model | cmmlu | ceval | arc | bbh | gsm8k | humaneval<br/>(pass@3) | livecodebench | gpqa_diamond<br/>(pass@3) |
48
  |------------------|--------|--------|--------|--------|--------|-------------------|---------------|----------------------|
 
17
  </h3>
18
 
19
  <p align="center">
20
+ 📣 <a href="https://huggingface.co/AngelSlim/HY-1.8B-2Bit-GGUF">GGUF</a>&nbsp&nbsp | &nbsp&nbsp 📖 <a href="https://angelslim.readthedocs.io/">Documentation</a>&nbsp&nbsp | &nbsp&nbsp🤗 <a href="https://huggingface.co/AngelSlim">Hugging Face</a>&nbsp&nbsp | &nbsp&nbsp🤖 <a href="https://modelscope.cn/organization/AngelSlim">ModelScope</a>&nbsp&nbsp | &nbsp&nbsp💬 <a href="./docs/source/assets/angel_slim_wechat.png">WeChat</a>
21
  <br>
22
  </p>
23
 
 
42
 
43
  Benchmark results for HY-1.8B-2Bit equivalent weights on vLLM across **cmmlu**,**ceval**,**arc**,**bbh**,**gsm8k**,**humaneval**,**livecodebench** and **gpqa_diamond**.
44
 
45
+ The empirical results reveal that HY-1.8B-2Bit maintains high-tier performance despite the extreme reduction in bit-width, incurring a marginal average degradation of only 3.97\% compared to its full-precision 1.8B teacher. Remarkably, HY-1.8B-2Bit performs nearly on par with the INT4 variant,with a negligible accuracy gap of only 0.13\%, while utilizing only half the weight precision. When compared to the dense HY-0.5B model, which occupies a comparable model size, the superiority of the 2-bit QAT approach becomes evident.  While the 0.5B dense model suffers a catastrophic 21.87\% drop in average accuracy, HY-1.8B-2Bit remains robust, outperforming the smaller dense counterpart by 22.29\% in GSM8K and 20.62\% in LiveCodeBench.
46
 
47
  | Model | cmmlu | ceval | arc | bbh | gsm8k | humaneval<br/>(pass@3) | livecodebench | gpqa_diamond<br/>(pass@3) |
48
  |------------------|--------|--------|--------|--------|--------|-------------------|---------------|----------------------|