Update README.md
Browse files
README.md
CHANGED
|
@@ -29,58 +29,39 @@ Join me at: https://discord.gg/UBgz4VXf
|
|
| 29 |
|
| 30 |
## EXPERIMENTAL
|
| 31 |
|
| 32 |
-
Please note this is an experimental
|
|
|
|
|
|
|
| 33 |
|
| 34 |
To use it you will require:
|
| 35 |
|
| 36 |
1. AutoGPTQ, from the latest `main` branch and compiled with `pip install .`
|
| 37 |
2. `pip install einops`
|
| 38 |
|
| 39 |
-
You can then use it immediately from Python code - see example code below
|
| 40 |
-
|
| 41 |
-
## text-generation-webui
|
| 42 |
-
|
| 43 |
-
There is also provisional AutoGPTQ support in text-generation-webui.
|
| 44 |
-
|
| 45 |
-
However at the time I'm writing this, a commit is needed to text-generation-webui to enable it to load this model.
|
| 46 |
|
| 47 |
-
|
| 48 |
|
| 49 |
-
To
|
| 50 |
-
1. Edit `text-generation-webui/modules/AutoGPTQ_loader.py`
|
| 51 |
-
2. Make the following change:
|
| 52 |
-
|
| 53 |
-
Find the line that says:
|
| 54 |
-
```
|
| 55 |
-
'use_safetensors': use_safetensors,
|
| 56 |
-
```
|
| 57 |
-
|
| 58 |
-
And after it, add:
|
| 59 |
```
|
| 60 |
-
|
|
|
|
|
|
|
| 61 |
```
|
| 62 |
|
| 63 |
-
|
| 64 |
|
| 65 |
-
|
| 66 |
-
|
| 67 |
-
```
|
| 68 |
-
pip install einops
|
| 69 |
-
```
|
| 70 |
|
| 71 |
-
|
| 72 |
|
| 73 |
-
|
| 74 |
-
git clone https://github.com/PanQiWei/AutoGPTQ
|
| 75 |
-
cd AutoGPTQ
|
| 76 |
-
pip install . # This step requires CUDA toolkit installed
|
| 77 |
-
```
|
| 78 |
|
| 79 |
-
|
| 80 |
|
| 81 |
## How to download and use this model in text-generation-webui
|
| 82 |
|
| 83 |
-
1. Launch text-generation-webui with the following command-line arguments: `--autogptq --
|
| 84 |
2. Click the **Model tab**.
|
| 85 |
3. Under **Download custom model or LoRA**, enter `TheBloke/falcon-7B-instruct-GPTQ`.
|
| 86 |
4. Click **Download**.
|
|
|
|
| 29 |
|
| 30 |
## EXPERIMENTAL
|
| 31 |
|
| 32 |
+
Please note this is an experimental GPTQ model. Support for it is currently quite limited.
|
| 33 |
+
|
| 34 |
+
It is also expected to be **VERY SLOW**. This is unavoidable at the moment, but is being looked at.
|
| 35 |
|
| 36 |
To use it you will require:
|
| 37 |
|
| 38 |
1. AutoGPTQ, from the latest `main` branch and compiled with `pip install .`
|
| 39 |
2. `pip install einops`
|
| 40 |
|
| 41 |
+
You can then use it immediately from Python code - see example code below - or from text-generation-webui.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 42 |
|
| 43 |
+
## AutoGPTQ
|
| 44 |
|
| 45 |
+
To install AutoGPTQ please follow these instructions:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 46 |
```
|
| 47 |
+
git clone https://github.com/PanQiWei/AutoGPTQ
|
| 48 |
+
cd AutoGPTQ
|
| 49 |
+
pip install .
|
| 50 |
```
|
| 51 |
|
| 52 |
+
These steps will require that you have the [Nvidia CUDA toolkit](https://developer.nvidia.com/cuda-12-0-1-download-archive) installed.
|
| 53 |
|
| 54 |
+
## text-generation-webui
|
|
|
|
|
|
|
|
|
|
|
|
|
| 55 |
|
| 56 |
+
There is also provisional AutoGPTQ support in text-generation-webui.
|
| 57 |
|
| 58 |
+
This requires text-generation-webui as of commit 204731952ae59d79ea3805a425c73dd171d943c3.
|
|
|
|
|
|
|
|
|
|
|
|
|
| 59 |
|
| 60 |
+
So please first update text-genration-webui to the latest version.
|
| 61 |
|
| 62 |
## How to download and use this model in text-generation-webui
|
| 63 |
|
| 64 |
+
1. Launch text-generation-webui with the following command-line arguments: `--autogptq --trust-remote-code`
|
| 65 |
2. Click the **Model tab**.
|
| 66 |
3. Under **Download custom model or LoRA**, enter `TheBloke/falcon-7B-instruct-GPTQ`.
|
| 67 |
4. Click **Download**.
|