Add llama.cpp to the examples
#3
by
b-a-s-e-d
- opened
README.md
CHANGED
|
@@ -107,6 +107,7 @@ The model can also be deployed with the following libraries:
|
|
| 107 |
- [`mistral-inference`](https://github.com/mistralai/mistral-inference): See [here](#mistral-inference)
|
| 108 |
- [`transformers`](https://github.com/huggingface/transformers): See [here](#transformers)
|
| 109 |
- [`LMStudio`](https://lmstudio.ai/): See [here](#lmstudio)
|
|
|
|
| 110 |
- [`ollama`](https://github.com/ollama/ollama): See [here](#ollama)
|
| 111 |
|
| 112 |
|
|
@@ -394,6 +395,23 @@ docker run -it --rm --pull=always \
|
|
| 394 |
Click “see advanced setting” on the second line.
|
| 395 |
In the new tab, toggle advanced to on. Set the custom model to be mistral/devstralq4_k_m and Base URL the api address we get from the last step in LM Studio. Set API Key to dummy. Click save changes.
|
| 396 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 397 |
|
| 398 |
### Ollama
|
| 399 |
|
|
|
|
| 107 |
- [`mistral-inference`](https://github.com/mistralai/mistral-inference): See [here](#mistral-inference)
|
| 108 |
- [`transformers`](https://github.com/huggingface/transformers): See [here](#transformers)
|
| 109 |
- [`LMStudio`](https://lmstudio.ai/): See [here](#lmstudio)
|
| 110 |
+
- [`llama.cpp`](https://github.com/ggml-org/llama.cpp): See [here](#llama.cpp)
|
| 111 |
- [`ollama`](https://github.com/ollama/ollama): See [here](#ollama)
|
| 112 |
|
| 113 |
|
|
|
|
| 395 |
Click “see advanced setting” on the second line.
|
| 396 |
In the new tab, toggle advanced to on. Set the custom model to be mistral/devstralq4_k_m and Base URL the api address we get from the last step in LM Studio. Set API Key to dummy. Click save changes.
|
| 397 |
|
| 398 |
+
### llama.cpp
|
| 399 |
+
|
| 400 |
+
Download the weights from huggingface:
|
| 401 |
+
|
| 402 |
+
```
|
| 403 |
+
pip install -U "huggingface_hub[cli]"
|
| 404 |
+
huggingface-cli download \
|
| 405 |
+
"mistralai/Devstral-Small-2505_gguf" \
|
| 406 |
+
--include "devstralQ4_K_M.gguf" \
|
| 407 |
+
--local-dir "mistralai/Devstral-Small-2505_gguf/"
|
| 408 |
+
```
|
| 409 |
+
|
| 410 |
+
Then run Devstral using the llama.cpp CLI.
|
| 411 |
+
|
| 412 |
+
```bash
|
| 413 |
+
./llama-cli -m Devstral-Small-2505_gguf/devstralQ4_K_M.gguf -cnv
|
| 414 |
+
```
|
| 415 |
|
| 416 |
### Ollama
|
| 417 |
|