yhyu13
Update readme
955bb3b
---
license: llama2
---
# Edit
PS : https://github.com/Cornell-RelaxML/quip-sharp/issues/13
As mentioned in the above issue thread,
- for accurate hessian generation, use a larger devset (e.g., 4096) and consider accumulating hessians in fp32 if consumer GPUs with fast fp64 are not available.
- changing the Hessian dataset from a natural language dataset to a mathematical dataset, as the task is a math model.
---
Experiment QUIP 2-bit E8P12 version that works in textgen-webui with quip mode loader
Generated by using scripts from https://gitee.com/yhyu13/llama_-tools
Original weight : https://huggingface.co/Xwin-LM/Xwin-Math-7B-V1.0
GPTQ 4bit : https://huggingface.co/Yhyu13/Xwin-Math-7B-V1.0-GPTQ-4bit
---
This repo used `hessian_offline_llama.py` provided by QUIP repo to generate hessian specifically for the orignal model before applying Quip quantization.
It took quite a long time for hessian for all 31 layers, about 6 hours for 7B models on a single RTX3090. I am not sure if I made any error.
QUIP byproducts are also uploaded.
Perplexity calcaultead using `eval_ppl.py` provided by QUIP repo
QUIP PPL:
- wikitext2 perplexity: 11.247852325439453
- c4 perplexity: 16.275997161865234
Original model PPL:
- wikitext2 perplexity: 6.042122840881348
- c4 perplexity: 8.430611610412598
Looks like something is wrong, the quantized model is a disaster.
---
Here is some testing done in textgen-webui, I was using Q&A from this dataset https://huggingface.co/datasets/TIGER-Lab/MathInstruct
It seems the 2 bit could hardly answer any question correctly, in compare to GPTQ 4bit version. But the https://huggingface.co/relaxml/Llama-2-13b-E8P-2Bit model made by the author of QUIP seems to work fine, just as good as GPTQ.
So in conclusion, this is a very experimental model that I made just to testify QUIP, I may made some error. But I think it is a good start.
QUIP 2 bit version:
![Alt text](img/XMath_Quip2.png)
GPTQ 4 bit version:
![Alt text](img/textgen-xinmath.png)