Instructions to use mlx-community/CodeLlama-70b-Instruct-hf-4bit-MLX with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- MLX
How to use mlx-community/CodeLlama-70b-Instruct-hf-4bit-MLX with MLX:
# Make sure mlx-lm is installed # pip install --upgrade mlx-lm # Generate text with mlx-lm from mlx_lm import load, generate model, tokenizer = load("mlx-community/CodeLlama-70b-Instruct-hf-4bit-MLX") prompt = "Write a story about Einstein" messages = [{"role": "user", "content": prompt}] prompt = tokenizer.apply_chat_template( messages, add_generation_prompt=True ) text = generate(model, tokenizer, prompt=prompt, verbose=True) - Notebooks
- Google Colab
- Kaggle
- Local Apps
- LM Studio
- MLX LM
How to use mlx-community/CodeLlama-70b-Instruct-hf-4bit-MLX with MLX LM:
Generate or start a chat session
# Install MLX LM uv tool install mlx-lm # Interactive chat REPL mlx_lm.chat --model "mlx-community/CodeLlama-70b-Instruct-hf-4bit-MLX"
Run an OpenAI-compatible server
# Install MLX LM uv tool install mlx-lm # Start the server mlx_lm.server --model "mlx-community/CodeLlama-70b-Instruct-hf-4bit-MLX" # Calling the OpenAI-compatible server with curl curl -X POST "http://localhost:8000/v1/chat/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "mlx-community/CodeLlama-70b-Instruct-hf-4bit-MLX", "messages": [ {"role": "user", "content": "Hello"} ] }'
`std::runtime_error: [Matmul::eval_cpu] Currently only supports float32`
#2
by adhishthite - opened
Hello,
I am getting this output on my Mac M1 Pro on Sonoma 14.3. Python version used is 3.11. Used latest PyTorch.
Please visit this link for full output: https://app.warp.dev/block/NMbYuCAkwfcxcQ7zjhZv8n
In [5]: response = generate(model, tokenizer, prompt="<step>Source: user Fibonacci series in Python<step> Source: assistant Destination: user", verbose=True)
...:
...:
==========
Prompt: <step>Source: user Fibonacci series in Python<step> Source: assistant Destination: user
libc++abi: terminating due to uncaught exception of type std::runtime_error: [Matmul::eval_cpu] Currently only supports float32.
[1] 9782 abort ipython
How much RAM in your M1 Pro? This model requires a machine with at least 64GB to run with q4
@ivanfioravanti I have the M1 Pro 16GB, but still got this error with a 7B Q4 quantized model.
Please see my github issue if you know how to fix it. Many thanks
https://github.com/ml-explore/mlx/issues/753