| | --- |
| | license: apache-2.0 |
| | datasets: |
| | - common-pile/comma_v0.1_training_dataset |
| | language: |
| | - en |
| | base_model: common-pile/comma-v0.1-2t |
| | pipeline_tag: text-generation |
| | tags: |
| | - mlx |
| | library_name: mlx |
| | --- |
| | # common-pile/comma-v0.1-2t ported to MLX |
| |
|
| | See [common-pile/comma-v0.1-2t](https://huggingface.co/common-pile/comma-v0.1-2t) for the original model card. |
| |
|
| | Try this model out using [uv](https://github.com/astral-sh/uv) like this, which will download a 15GB model the first time you run it: |
| | ``` |
| | uv run --python 3.12 \ |
| | --with mlx-lm \ |
| | mlx_lm.generate \ |
| | --model simonw/comma-v0.1-2t-mlx \ |
| | --prompt 'Facts about pelicans:' |
| | ``` |
| |
|
| | More notes [on my blog](https://simonwillison.net/2025/Jun/7/comma/). I created the MLX port using this command: |
| |
|
| | ```bash |
| | uv run --python 3.12 \ |
| | --with mlx-lm \ |
| | python -m mlx_lm convert \ |
| | --hf-path common-pile/comma-v0.1-2t \ |
| | --mlx-path ./comma-v0.1-2t-mlx |
| | ``` |
| |
|