Instructions to use ModelCloud/glm-4-9b-gptq-4bit with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use ModelCloud/glm-4-9b-gptq-4bit with Transformers:
# Load model directly from transformers import AutoModel model = AutoModel.from_pretrained("ModelCloud/glm-4-9b-gptq-4bit", trust_remote_code=True, dtype="auto") - Notebooks
- Google Colab
- Kaggle
YAML Metadata Warning:empty or missing yaml metadata in repo card
Check out the documentation for more information.
World's first gptq 4bit quant of glm-4-9b model.
Autogptq PR: https://github.com/AutoGPTQ/AutoGPTQ/pull/683
Please note ChatGLM has tendency to switch from English to Chinese in mid-reply or in direct reply to English prompt. This issue happens in both native and quantized model and needs further investigation.
- Downloads last month
- 9
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support