Add exllamav2 quant link
#5
by
bartowski
- opened
README.md
CHANGED
|
@@ -71,4 +71,8 @@ Hercules-3.0-Mistral-7B is fine-tuned from the following sources:
|
|
| 71 |
- No model parameters were frozen.
|
| 72 |
- This model was trained on OpenAI's ChatML prompt format. Because this model has function calling capabilities, the prompt format is slightly different, here's what it would look like: ```<|im_start|>system\n{message}<|im_end|>\n<|im_start|>user\n{user message}<|im_end|>\n<|im_start|>call\n{function call message}<|im_end|>\n<|im_start|>function\n{function response message}<|im_end|>\n<|im_start|>assistant\n{assistant message}</s>```
|
| 73 |
|
| 74 |
-
This model was fine-tuned using the TPU-Alignment repository. https://github.com/Locutusque/TPU-Alignment
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 71 |
- No model parameters were frozen.
|
| 72 |
- This model was trained on OpenAI's ChatML prompt format. Because this model has function calling capabilities, the prompt format is slightly different, here's what it would look like: ```<|im_start|>system\n{message}<|im_end|>\n<|im_start|>user\n{user message}<|im_end|>\n<|im_start|>call\n{function call message}<|im_end|>\n<|im_start|>function\n{function response message}<|im_end|>\n<|im_start|>assistant\n{assistant message}</s>```
|
| 73 |
|
| 74 |
+
This model was fine-tuned using the TPU-Alignment repository. https://github.com/Locutusque/TPU-Alignment
|
| 75 |
+
|
| 76 |
+
# Quants
|
| 77 |
+
|
| 78 |
+
ExLlamaV2 by bartowski https://huggingface.co/bartowski/Hercules-3.0-Mistral-7B-exl2
|