| | --- |
| | license: apache-2.0 |
| | tags: |
| | - gguf |
| | --- |
| | GGUF [llama.cpp](https://github.com/ggerganov/llama.cpp) quantized version of: |
| | - Original model: [Mistral-Nemo-Instruct-2407](https://huggingface.co/mistralai/Mistral-Nemo-Instruct-2407) |
| | - Model creator: [Mistral AI](https://huggingface.co/mistralai) |
| | - [License](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/apache-2.0.md) |
| |
|
| | ## Recommended Prompt Format (Mistral) |
| | ``` |
| | <s>[INST]Provide some context and/or instructions to the model.[/INST] |
| | AI message goes here</s> |
| | [INST] The user’s message goes here [/INST] |
| | ``` |
| |
|
| | Quant Version: [b3437](https://github.com/ggerganov/llama.cpp/releases/tag/b3437) with [imatrix](https://github.com/ggerganov/llama.cpp/discussions/5263#discussioncomment-8395384) |