| # Gemma 3 vision |
|
|
| > [!IMPORTANT] |
| > |
| > This is very experimental, only used for demo purpose. |
|
|
| ## Quick started |
|
|
| You can use pre-quantized model from [ggml-org](https://huggingface.co/ggml-org)'s Hugging Face account |
|
|
| ```bash |
| # build |
| cmake -B build |
| cmake --build build --target llama-mtmd-cli |
| |
| # alternatively, install from brew (MacOS) |
| brew install llama.cpp |
| |
| # run it |
| llama-mtmd-cli -hf ggml-org/gemma-3-4b-it-GGUF |
| llama-mtmd-cli -hf ggml-org/gemma-3-12b-it-GGUF |
| llama-mtmd-cli -hf ggml-org/gemma-3-27b-it-GGUF |
| |
| # note: 1B model does not support vision |
| ``` |
|
|
| ## How to get mmproj.gguf? |
|
|
| Simply to add `--mmproj` in when converting model via `convert_hf_to_gguf.py`: |
|
|
| ```bash |
| cd gemma-3-4b-it |
| python ../llama.cpp/convert_hf_to_gguf.py --outfile model.gguf --outtype f16 --mmproj . |
| # output file: mmproj-model.gguf |
| ``` |
|
|
| ## How to run it? |
|
|
| What you need: |
| - The text model GGUF, can be converted using `convert_hf_to_gguf.py` |
| - The mmproj file from step above |
| - An image file |
|
|
| ```bash |
| # build |
| cmake -B build |
| cmake --build build --target llama-mtmd-cli |
| |
| # run it |
| ./build/bin/llama-mtmd-cli -m {text_model}.gguf --mmproj mmproj.gguf --image your_image.jpg |
| ``` |
|
|