| | --- |
| | license: apache-2.0 |
| | license_name: jamba-open-model-license |
| | license_link: https://www.ai21.com/licenses/jamba-open-model-license |
| | pipeline_tag: text-generation |
| | library_name: mlx |
| | tags: |
| | - mlx |
| | - safetensors |
| | - jamba |
| | - text-generation |
| | - reasoning |
| | base_model: ai21labs/AI21-Jamba-Reasoning-3B |
| | --- |
| | |
| | # AI21-Jamba-Reasoning-3B MLX |
| |
|
| | This repository contains a public MLX `safetensors` export of |
| | [`ai21labs/AI21-Jamba-Reasoning-3B`](https://huggingface.co/ai21labs/AI21-Jamba-Reasoning-3B) |
| | for Apple Silicon workflows with `mlx-lm`. |
| |
|
| | ## Model Details |
| |
|
| | - Base model: `ai21labs/AI21-Jamba-Reasoning-3B` |
| | - Format: MLX `safetensors` |
| | - Quantization: none |
| | - Intended use: local text generation and chat on MLX-compatible Apple devices |
| |
|
| | ## Quick Start |
| |
|
| | Install the runtime: |
| |
|
| | ```bash |
| | pip install -U mlx-lm |
| | ``` |
| |
|
| | Run a one-shot generation: |
| |
|
| | ```bash |
| | mlx_lm.generate --model ssdataanalysis/AI21-Jamba-Reasoning-3B-mlx-fp16 --prompt "Explain why the sky is blue." |
| | ``` |
| |
|
| | Start an interactive chat: |
| |
|
| | ```bash |
| | mlx_lm.chat --model ssdataanalysis/AI21-Jamba-Reasoning-3B-mlx-fp16 |
| | ``` |
| |
|
| | Run the HTTP server: |
| |
|
| | ```bash |
| | mlx_lm.server --model ssdataanalysis/AI21-Jamba-Reasoning-3B-mlx-fp16 --host 127.0.0.1 --port 8080 |
| | ``` |
| |
|
| | You can replace the model ID above with a local path if you have already |
| | downloaded the repository. |
| |
|
| | ## Notes |
| |
|
| | - This is an MLX export intended for `mlx-lm`. |
| | - The upstream model license remains Apache-2.0 with the Jamba open model |
| | license details linked above. |
| | - For the original source checkpoint and upstream documentation, see |
| | [`ai21labs/AI21-Jamba-Reasoning-3B`](https://huggingface.co/ai21labs/AI21-Jamba-Reasoning-3B). |
| |
|