| library_name: transformers | |
| pipeline_tag: text-generation | |
| tags: | |
| - code | |
| - llama | |
| - llama-2 | |
| - text-generation-inference | |
| base_model: codellama/CodeLlama-7b-Instruct-hf | |
| inference: false | |
| # Mistral-7B-v0.1-GGUF | |
| - Quantized version of [CodeLlama-7b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-7b-Instruct-hf) | |
| - Created using llama.cpp | |
| ## Available Quants | |
| * Q2_K | |
| * Q3_K_L | |
| * Q3_K_M | |
| * Q3_K_S | |
| * Q4_0 | |
| * Q4_K_M | |
| * Q4_K_S | |
| * Q5_0 | |
| * Q5_K_M | |
| * Q5_K_S | |
| * Q6_K | |
| * Q8_0 |