Update README.md
Browse files
README.md
CHANGED
|
@@ -21,6 +21,10 @@ license: other
|
|
| 21 |
|
| 22 |
These files are GGML format model files for [Gorilla LLM's Gorilla 7B](https://huggingface.co/gorilla-llm/gorilla-7b-hf-delta-v0).
|
| 23 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 24 |
GGML files are for CPU + GPU inference using [llama.cpp](https://github.com/ggerganov/llama.cpp) and libraries and UIs which support this format, such as:
|
| 25 |
* [text-generation-webui](https://github.com/oobabooga/text-generation-webui)
|
| 26 |
* [KoboldCpp](https://github.com/LostRuins/koboldcpp)
|
|
@@ -34,6 +38,13 @@ GGML files are for CPU + GPU inference using [llama.cpp](https://github.com/gger
|
|
| 34 |
* [2, 3, 4, 5, 6 and 8-bit GGML models for CPU+GPU inference](https://huggingface.co/TheBloke/gorilla-7B-GGML)
|
| 35 |
* [Unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/TheBloke/gorilla-7B-fp16)
|
| 36 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 37 |
<!-- compatibility_ggml start -->
|
| 38 |
## Compatibility
|
| 39 |
|
|
|
|
| 21 |
|
| 22 |
These files are GGML format model files for [Gorilla LLM's Gorilla 7B](https://huggingface.co/gorilla-llm/gorilla-7b-hf-delta-v0).
|
| 23 |
|
| 24 |
+
**NOTE**: This is not a regular LLM. It is designed to allow LLMs to use tools by invoking APIs.
|
| 25 |
+
|
| 26 |
+
"Gorilla enables LLMs to use tools by invoking APIs. Given a natural language query, Gorilla can write a semantically- and syntactically- correct API to invoke. With Gorilla, we are the first to demonstrate how to use LLMs to invoke 1,600+ (and growing) API calls accurately while reducing hallucination. "
|
| 27 |
+
|
| 28 |
GGML files are for CPU + GPU inference using [llama.cpp](https://github.com/ggerganov/llama.cpp) and libraries and UIs which support this format, such as:
|
| 29 |
* [text-generation-webui](https://github.com/oobabooga/text-generation-webui)
|
| 30 |
* [KoboldCpp](https://github.com/LostRuins/koboldcpp)
|
|
|
|
| 38 |
* [2, 3, 4, 5, 6 and 8-bit GGML models for CPU+GPU inference](https://huggingface.co/TheBloke/gorilla-7B-GGML)
|
| 39 |
* [Unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/TheBloke/gorilla-7B-fp16)
|
| 40 |
|
| 41 |
+
## Prompt template
|
| 42 |
+
|
| 43 |
+
```
|
| 44 |
+
###USER: find me an API to generate cute cat images
|
| 45 |
+
###ASSISTANT:
|
| 46 |
+
```
|
| 47 |
+
|
| 48 |
<!-- compatibility_ggml start -->
|
| 49 |
## Compatibility
|
| 50 |
|