| | --- |
| | license: apache-2.0 |
| | base_model: |
| | - allenai/Olmo-3.1-32B-Instruct |
| | tags: |
| | - llmcompressor |
| | --- |
| | This is [allenai/Olmo-3.1-32B-Instruct](https://huggingface.co/allenai/Olmo-3.1-32B-Instruct) quantized with [LLM Compressor](https://github.com/vllm-project/llm-compressor) with the recipe in the "recipe.yaml" file. The model is compatible with vLLM (tested: v0.12.0). Tested with an RTX Pro 6000. |
| |
|
| | How the models perform (token efficiency, accuracy per domain, ...) and how to use them: |
| | [Quantizing Olmo 3: Most Efficient and Accurate Formats](https://kaitchup.substack.com/p/quantizing-olmo-3-most-efficient) |
| |
|
| |  |
| |
|
| | - **Developed by:** [The Kaitchup](https://kaitchup.substack.com/) |
| | - **License:** Apache 2.0 license |
| |
|
| | ## How to Support My Work |
| | Subscribe to [The Kaitchup](https://kaitchup.substack.com/subscribe). Or you can "[buy me a kofi](https://ko-fi.com/bnjmn_marie)". |
| |
|