Instructions to use zai-org/chatglm3-6b with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use zai-org/chatglm3-6b with Transformers:
# Load model directly from transformers import AutoModel model = AutoModel.from_pretrained("zai-org/chatglm3-6b", trust_remote_code=True, dtype="auto") - Notebooks
- Google Colab
- Kaggle
本地模型加载太慢
#46
by pxkkkkk - opened
请问量化后加载过程还是很慢是怎么回事???Loading checkpoint shards: 100%|██████████| 7/7 [08:51<00:00, 75.92s/it]
这个模型没有做量化上的优化,你如果在CPU加载肯定很慢
pxkkkkk changed discussion status to closed