File size: 2,401 Bytes
c7db2ae 4e1cfd7 9da1b7d c7db2ae af8f393 8a29dab 7a6a274 34b84c9 8ad9e6c 9da1b7d af8f393 073ce48 af8f393 57a0195 af8f393 9da1b7d 305b851 9da1b7d 305b851 9da1b7d af8f393 073ce48 57a0195 af8f393 073ce48 af8f393 c7db2ae 073ce48 c7db2ae 799d2d3 528aea5 799d2d3 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 | ---
license: apache-2.0
tags:
- Image-Text-to-Text Models
- Audio-Text-to-Text
- text-to-Text
- llama_cpp
- any to any
- Multimodal AI
---
nzgnzg73
llama_cpp_WebUI
Github
https://github.com/nzgnzg73/llama_cpp_WebUI
## Want to talk or ask something?
Just click the YouTube link below! You'll find my 📧 email there and can message me easily. 👇
🎥 YouTube Channel: @nzg73
🔗 https://youtube.com/@NZG73
## Contact Email 👇👇👀
E-mail:-
nzgnzg73@gmail.com

## Image-Text-to-Text Models
## Gemma-3
CPU. RAM 20GB
OR
GPU. 4 VRAM

1. gemma-3-12b-it-Q4_K_S.gguf
2. mmproj-model-f16-12B.gguf
## -Text-to-Text Models
## GPT OSS 20


## Qwen3
CPU. RAM 25GB
OR
GPU. 4 VRAM
1. Qwen3-VL-2B-Instruct-Q8_0.gguf
2. mmproj-Qwen3-VL-2B-Instruct-Q8_0.gguf

## Qwen2.5-Omni
CPU. RAM 40GB
OR
GPU. 8 VRAM
1. Qwen2.5-Omni-7B-BF16.gguf
2. mmproj-F16.gguf 2GB

## Audio-Text-to-Text
## Llama-3.2
CPU. RAM 10GB

1. Llama-3.2-1B-Instruct-Q4_K_M.gguf
2. Llama-3.2-1B-Instruct-Q8_0.gguf
3. mmproj-ultravox-v0_5-llama-3_2-1b-f16.gguf
## run.bat
Local Server
llama-server.exe --n-gpu-layers 2 --ctx-size 111192 -m ".\models\mistralai\mistralai_Voxtral-Mini-3B-2507-Q8_0.gguf" --mmproj ".\models\mistralai\mmproj-mistralai_Voxtral-Mini-3B-2507-bf16.gguf" --host 0.0.0.0 --port 8005
public URL
llama-server --n-gpu-layers 15 --ctx-size 8192 -m models/ollma/Llama-3.2-1B-Instruct-Q8_0.gguf --mmproj models/ollma/mmproj-ultravox-v0_5-llama-3_2-1b-f16.gguf --host 127.0.0.1 --port 8083 |