--- license: apache-2.0 base_model: - Nanbeige/Nanbeige4.1-3B tags: - llm-compressor ---
The Kaitchup -- AI on a Budget
Subscribe and Support
This is [Nanbeige/Nanbeige4.1-3B](https://huggingface.co/Nanbeige/Nanbeige4.1-3B) quantized with [llm-compressor](https://github.com/vllm-project/llm-compressor) to W8A8 (FP8) . The model is compatible with vLLM (tested: v0.15.1). Tested with an L4 (Google Colab). - **Developed by:** [The Kaitchup](https://kaitchup.substack.com/)