--- license: apache-2.0 tags: - diffusion - masked-diffusion - dream - qwen2 - gguf - diffuse-cpp base_model: Dream-org/Dream-v0-Instruct-7B pipeline_tag: text-generation --- # Dream-v0-Instruct-7B-GGUF GGUF quantizations of [Dream-org/Dream-v0-Instruct-7B](https://huggingface.co/Dream-org/Dream-v0-Instruct-7B) for use with [diffuse-cpp](https://github.com/iafiscal1212/diffuse-cpp), a CPU inference engine for Diffusion Language Models. Dream is a masked diffusion language model based on the Qwen2.5-7B backbone with bidirectional attention and Grouped Query Attention (GQA, 28 query heads / 4 KV heads). ## Available Quantizations | File | Type | Size | Description | |------|------|------|-------------| | `dream-7b-f16.gguf` | F16 | ~15 GB | Full precision, best quality | | `dream-7b-q8_0.gguf` | Q8_0 | ~8.2 GB | 8-bit quantization, near-lossless | | `dream-7b-q4km.gguf` | Q4_K_M | ~5.0 GB | 4-bit mixed quantization, best quality/size ratio | **Recommended:** Q4_K_M for most users. Q8_0 if you have enough RAM and want minimal quality loss. ## Performance Benchmarked on diffuse-cpp with entropy_exit + inter-step KV cache, 12 threads, seed=42: | Prompt | tok/s | Steps | vs llama.cpp | |--------|-------|-------|-------------| | Capital of France? | 21.6 | 2 | 2.5x | | Translate to French | 14.3 | 6 | 1.7x | | 15 x 23? | 21.6 | 2 | 2.5x | | Translate to Spanish | 13.2 | 10 | 1.6x | | Python is_prime() | 8.2 | 7 | 1.0x | | Why sky blue? | 4.9 | 16 | 0.6x | | List planets | 4.9 | 16 | 0.6x | | Poem about ocean | 4.5 | 16 | 0.5x | | **Average** | **11.6** | | **1.4x** | - Easy prompts (factual, math): **14-22 tok/s** (1.6-2.5x faster than llama.cpp) - Hard prompts (creative, long-form): **4.5-4.9 tok/s** - llama.cpp baseline: 8.51 tok/s (Qwen2.5-7B-Instruct, Q4_K_M, same hardware) ## Usage ```bash # Download huggingface-cli download diffuse-cpp/Dream-v0-Instruct-7B-GGUF dream-7b-q4km.gguf # Run (requires diffuse-cpp v0.2.0+) ./diffuse-cli -m dream-7b-q4km.gguf -p "What is the capital of France?" -n 64 -s 16 ``` ## Model Details - **Architecture:** Qwen2.5-7B backbone with bidirectional attention - **Parameters:** 7.62B - **Layers:** 28 - **Hidden size:** 3584 - **Attention:** GQA (28 query heads, 4 KV heads, head dim 128) - **FFN:** SwiGLU, intermediate size 18944 - **Vocabulary:** 152,064 tokens - **RoPE theta:** 1,000,000 - **Mask token ID:** 151666 - **Training:** Masked diffusion on text, with autoregressive logit shift ## Conversion Details Converted from SafeTensors using `convert-dream.py` from diffuse-cpp: - 339 tensors total (255 weights + 84 QKV biases) - QKV biases kept at F32 in all quantizations - Edge layers (first/last) quantized to Q6_K in Q4_K_M scheme ## Citation ```bibtex @misc{dream2025, title={Dream 7B - Scalable Discrete Denoising Diffusion Models for Text Generation}, author={Ye, Jiacheng and others}, year={2025} } ``` ## License Apache 2.0, following the original Dream model license.