Instructions to use mlx-community/Mistral-Nemo-Instruct-2407-8bit with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- MLX
How to use mlx-community/Mistral-Nemo-Instruct-2407-8bit with MLX:
# Download the model from the Hub pip install huggingface_hub[hf_xet] huggingface-cli download --local-dir Mistral-Nemo-Instruct-2407-8bit mlx-community/Mistral-Nemo-Instruct-2407-8bit
- Notebooks
- Google Colab
- Kaggle
- Local Apps
- LM Studio
ValueError: Expected shape (5120, 1280) but received shape (4096, 1280)
#2
by silverslash - opened
Getting this error when running the example:
ValueError: Expected shape (5120, 1280) but received shape (4096, 1280) for parameter model.layers.0.self_attn.q_proj.weight
How was this even possible? Was the quantization passed a wrong config? I wasted a 12 GB download...
It seems it need a mlx-lm version that's unreleased:
https://huggingface.co/mlx-community/Mistral-Nemo-Instruct-2407-4bit/discussions/2#669a7c9a49d1c43c4a42f702