GGUF

GGUF quantization for the InternLM-20b-chat model.

Requires a recent version of llama.cpp: https://github.com/ggerganov/llama.cpp/pull/4283

Link to the original model: https://huggingface.co/internlm/internlm-chat-20b

Github repository: https://github.com/InternLM/InternLM

Prompt format:

<|User|>: Your input.<eoh>
<|Bot|>:

License:

Model weights are fully open for academic research and also allow free commercial usage. To apply for a commercial license, please fill in the application form (English)/申请表(中文). For other questions or collaborations, please contact internlm@pjlab.org.cn.

Downloads last month
1
GGUF
Model size
20B params
Architecture
llama
Hardware compatibility
Log In to add your hardware

2-bit

3-bit

4-bit

5-bit

6-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support