Add link to llama-imatrix
Browse files
README.md
CHANGED
|
@@ -144,6 +144,8 @@ is unfortunately very frequent).
|
|
| 144 |
|
| 145 |
## What do I need to do to compute imatrix files for large models?
|
| 146 |
|
|
|
|
|
|
|
| 147 |
### Hardware
|
| 148 |
|
| 149 |
* RAM: A lot of RAM is required to compute imatrix files. Example: 512 GB is just enough to compute 405B imatrix quants in Q8.
|
|
|
|
| 144 |
|
| 145 |
## What do I need to do to compute imatrix files for large models?
|
| 146 |
|
| 147 |
+
Use [`llama-imatrix`](https://github.com/ggml-org/llama.cpp/blob/master/examples/imatrix/README.md) to compute imatrix files.
|
| 148 |
+
|
| 149 |
### Hardware
|
| 150 |
|
| 151 |
* RAM: A lot of RAM is required to compute imatrix files. Example: 512 GB is just enough to compute 405B imatrix quants in Q8.
|