Cosmos-Reason2-2B-GGUF

GGUF conversion of nvidia/Cosmos-Reason2-2B.

Files

  • Cosmos-Reason2-2B-BF16-split-00001-of-00002.gguf to ...-00002-of-00002.gguf: Main model (BF16, split into 2 parts)
  • mmproj-Cosmos-Reason2-2B-BF16.gguf: Vision encoder (multimodal projector)

Download

huggingface-cli download robertzty/Cosmos-Reason2-2B-GGUF --local-dir ./Cosmos-Reason2-2B-GGUF

Usage with llama.cpp

No need to merge split files - llama.cpp loads them automatically:

llama-mtmd-cli -m ./Cosmos-Reason2-2B-GGUF/Cosmos-Reason2-2B-BF16-split-00001-of-00002.gguf \
    --mmproj ./Cosmos-Reason2-2B-GGUF/mmproj-Cosmos-Reason2-2B-BF16.gguf -cnv

Optional: merge into single file:

llama-gguf-split --merge Cosmos-Reason2-2B-BF16-split-00001-of-00002.gguf Cosmos-Reason2-2B-BF16.gguf

Notes

  • Requires llama.cpp b7480+ for qwen3vl architecture support
  • Ollama and LM Studio may have compatibility issues with qwen3vl models
Downloads last month
167
GGUF
Model size
2B params
Architecture
qwen3vl
Hardware compatibility
Log In to view the estimation

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for robertzty/Cosmos-Reason2-2B-GGUF

Quantized
(1)
this model