| --- |
| library_name: gguf |
| tags: |
| - moe |
| - expert-prefetch |
| - madvise |
| - llama-cpp |
| - apple-silicon |
| - cuda |
| - on-device |
| license: mit |
| pipeline_tag: text-generation |
| --- |
| |
| # llama.cpp Expert Sniper β madvise prefetch for MoE inference |
|
|
| ~65 lines of C++ that enables MoE models larger than RAM on llama.cpp. |
|
|
| Stock llama.cpp thrashes indefinitely. This build generates tokens. |
|
|
| ## Results |
|
|
| | Hardware | RAM | Model | Stock llama.cpp | madvise build | |
| |----------|-----|-------|-----------------|---------------| |
| | M2 MacBook Air | 8 GB | Qwen3.5-35B-A3B IQ2_M (10.6 GB) | 0 tok/s (thrash) | **0.57 tok/s** | |
| | M2 MacBook Air | 8 GB | Same model, no-op callback only | 0 tok/s (thrash) | 0.46 tok/s | |
| |
| On GPU machines with abundant RAM (A100 251GB, RTX 3090 31GB), stock llama.cpp is faster β the OS page cache handles it. madvise helps specifically when **system RAM < model size** and **layers are on CPU (ngl 0)**. |
| |
| ## How it works |
| |
| MoE models activate 8 of 128+ experts per token. Consecutive tokens share ~87% of active experts. Stock llama.cpp uses mmap but the OS has no idea which expert pages are hot β it evicts them randomly, causing a page fault storm. |
| |
| Our patch hooks llama.cpp's eval callback, intercepts every `ggml_mul_mat_id` operation, reads the router's top-k expert selection from `t->src[2]`, and calls `madvise(MADV_WILLNEED)` on each active expert's memory range. This tells the kernel which pages to prefetch before the compute needs them. |
| |
| Zero allocation. Zero memcpy. Zero mutex. One syscall per expert slice. |
| |
| ## What we learned |
| |
| **1. madvise beats LRU cache everywhere.** We first built a 460-line LRU cache. It was 2.4x slower than 15 lines of madvise (0.24 vs 0.57 tok/s on 8GB MacBook). The cache stole 5GB from the OS page cache for duplicate data. Don't fight the OS page cache β coach it. |
| |
| **2. Even a no-op callback prevents thrashing.** Just hooking the eval callback and inspecting tensor pointers (without any madvise) produces 0.46 tok/s where stock produces zero. The callback inadvertently warms mmap pages through pointer inspection. |
| |
| **3. Device pointer bug.** All prior GPU benchmarks were silently invalid β the callback dereferenced `t->src[2]->data` without checking `ggml_backend_buffer_is_host()`. On GPU layers this was a CUDA device pointer. Fixed. |
| |
| **4. On abundant RAM, do nothing.** The OS page cache is remarkably good when it has enough room. Any intervention is pure overhead when RAM exceeds model size. |
| |
| ## Build |
| |
| ```bash |
| git clone https://github.com/walter-grace/mac-code |
| cd mac-code/research/expert-sniper/llama-cpp |
| |
| # macOS (Metal) |
| cmake -B build -DGGML_METAL=ON -DCMAKE_BUILD_TYPE=Release |
| cmake --build build -j$(nproc) --target llama-server |
|
|
| # NVIDIA GPU (CUDA) |
| cmake -B build -DGGML_CUDA=ON -DCMAKE_BUILD_TYPE=Release |
| cmake --build build -j$(nproc) --target llama-server |
| ``` |
| |
| ## Usage |
| |
| ```bash |
| # Enable madvise prefetch |
| ./build/bin/llama-server \ |
| -m Qwen3.5-35B-A3B-UD-IQ2_M.gguf \ |
| -ngl 0 \ |
| --expert-cache-size 1 \ |
| --port 8201 |
|
|
| # Test no-op mode (isolate callback overhead) |
| EXPERT_CACHE_NOOP=1 ./build/bin/llama-server \ |
| -m model.gguf \ |
| -ngl 0 \ |
| --expert-cache-size 1 \ |
| --port 8201 |
| ``` |
| |
| ## Files changed vs stock llama.cpp |
| |
| **New files (~430 lines):** |
| |
| | File | Purpose | |
| |------|---------| |
| | `src/llama-expert-cache-ctx.cpp` | Eval callback, madvise prefetch, tensor identification | |
| | `src/llama-expert-cache-ctx.h` | Context struct and declarations | |
| | `src/llama-expert-cache.cpp` | LRU cache (deprecated, retained for reference) | |
| | `src/llama-expert-cache.h` | Cache class definition | |
| |
| **Patched files (~30 lines across 5 files):** |
| |
| | File | Change | |
| |------|--------| |
| | `src/CMakeLists.txt` | Added new source files to build | |
| | `src/llama-context.h` | Added expert cache context member | |
| | `common/common.h` | Added `expert_cache_size` parameter | |
| | `common/common.cpp` | Cache init + eval callback registration | |
| | `common/arg.cpp` | `--expert-cache-size` CLI flag | |
| |
| ## The research journey |
| |
| ``` |
| 460-line LRU cache β 0.24 tok/s (stole RAM from OS page cache) |
| 15-line madvise β 0.57 tok/s (coached the OS page cache) |
| no-op callback β 0.46 tok/s (accidental page warming) |
|
|
| The cache was the experiment. madvise was the answer. |
| ``` |
| |
| ## Full three-way benchmark (8 GB MacBook Air) |
| |
| | Config | tok/s | Mechanism | |
| |--------|-------|-----------| |
| | Stock llama.cpp | 0 (thrash) | OS blind LRU, no domain knowledge | |
| | No-op callback | 0.46 | Accidental page warming from tensor inspection | |
| | madvise prefetch | 0.57 | Explicit kernel prefetch hints | |
| | LRU cache (5 GB) | 0.24 | Duplicate data in user-space heap | |
| |
| ## Gemma 4-26B-A4B β MoE Sparsity Benchmark |
| |
| Google Gemma 4 has 128 experts with top-8 routing (4B active of 26B total). Tested at multiple quantization levels on Apple Silicon: |
| |
| | Hardware | Quant | Model size | RAM | Speed | Notes | |
| |----------|-------|-----------|-----|-------|-------| |
| | M2 MacBook Air | IQ2_M | 9.3 GB | 8 GB | **1.37 tok/s** | Model exceeds RAM, MoE sparsity prevents thrash | |
| | M4 Mac Mini | IQ2_M | 9.3 GB | 16 GB | **36.5 tok/s** | Fits in RAM, full GPU speed | |
| | M4 Mac Mini | Q4_K_M | 16.9 GB | 16 GB | **5.18 tok/s** | Exceeds RAM, still runs smoothly | |
| | M4 Mac Mini | Q8_0 | 26.9 GB | 16 GB | **0 tok/s (thrash)** | CPU_REPACK doubles memory to 51 GB, can't load | |
| |
| All results: stock llama.cpp with mmap, no madvise. Canberra verified on all configs. |
| |
| **Finding:** Gemma 4's low activation ratio (15.4%) lets the OS page cache handle memory pressure without explicit madvise. The madvise sniper is most valuable for denser MoE models (Qwen 35B) where the per-token working set overwhelms the page cache. |
| |
| ## Related |
| |
| - **MLX Expert Sniper** (Apple Silicon, 5.4 tok/s on 35B): [huggingface.co/waltgrace/mlx-expert-sniper](https://huggingface.co/waltgrace/mlx-expert-sniper) |
| - **Full research + code**: [github.com/walter-grace/mac-code/tree/main/research/expert-sniper](https://github.com/walter-grace/mac-code/tree/main/research/expert-sniper) |
| |