GGUF quantization for the InternLM-20b-chat model.
Requires a recent version of llama.cpp: https://github.com/ggerganov/llama.cpp/pull/4283
Link to the original model: https://huggingface.co/internlm/internlm-chat-20b
Github repository: https://github.com/InternLM/InternLM
Prompt format:
<|User|>: Your input.<eoh>
<|Bot|>:
License:
Model weights are fully open for academic research and also allow free commercial usage. To apply for a commercial license, please fill in the application form (English)/申请表(中文). For other questions or collaborations, please contact internlm@pjlab.org.cn.
- Downloads last month
- 1
Hardware compatibility
Log In to add your hardware
2-bit
3-bit
4-bit
5-bit
6-bit
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support