--- license: llama2 pipeline_tag: text-generation library_name: gguf --- **NOTE**: You will need a recent build of llama.cpp to run these quants (i.e. at least commit `494c870`). GGUF importance matrix (imatrix) quants for https://huggingface.co/m-a-p/OpenCodeInterpreter-CL-70B * The importance matrix was trained for ~50K tokens (105 batches of 512 tokens) using a [general purpose imatrix calibration dataset](https://github.com/ggerganov/llama.cpp/discussions/5263#discussioncomment-8395384). * The [imatrix is being used on the K-quants](https://github.com/ggerganov/llama.cpp/pull/4930) as well. > OpenCodeInterpreter is a family of open-source code generation systems designed to bridge the gap between large language models and advanced proprietary systems like the GPT-4 Code Interpreter. It significantly advances code generation capabilities by integrating execution and iterative refinement functionalities. This model is based on [CodeLlama-70b-Python-hf](https://huggingface.co/codellama/CodeLlama-70b-Python-hf). | Layers | Context | Template | | --- | --- | --- | |
80
|
16384
|
\[INST] {prompt} [/INST]
|