File size: 2,936 Bytes
c7db2ae 4e1cfd7 9da1b7d 7fedc95 5a1a4cd c7db2ae af8f393 8a29dab 7a6a274 34b84c9 8ad9e6c b4cb3bc 8ad9e6c 9da1b7d 5a1a4cd 9da1b7d 66c32c8 9da1b7d 66c32c8 5a1a4cd af8f393 073ce48 af8f393 57a0195 af8f393 9da1b7d 305b851 9da1b7d 305b851 9da1b7d af8f393 073ce48 57a0195 af8f393 073ce48 af8f393 c7db2ae 073ce48 c7db2ae 799d2d3 528aea5 799d2d3 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 |
---
license: apache-2.0
tags:
- Image-Text-to-Text Models
- Audio-Text-to-Text
- text-to-Text
- llama_cpp
- any to any
- Multimodal AI
- video-Text-to-Text
- Llama Model Switcher
---
nzgnzg73
llama_cpp_WebUI
Github
https://github.com/nzgnzg73/llama_cpp_WebUI
## Want to talk or ask something?
Just click the YouTube link below! You'll find my π§ email there and can message me easily. π
π₯ YouTube Channel: @nzg73
π https://youtube.com/@NZG73
## Contact Email πππ
E-mail:-
nzgnzg73@gmail.com
llama Cpp (cp\gpu)
Old Version llama cpp ππππ
llama-b7200-bin-win-cpu-x64.zip
NEW Update Version llama cpp ππππ
llama-b7541-bin-win-cpu-x64

## Llama Model Switcher
CMD
model_switcher.py
pip install flask
pip install flask psutil GPUtil


## Image-Text-to-Text Models
## Gemma-3
CPU. RAM 20GB
OR
GPU. 4 VRAM

1. gemma-3-12b-it-Q4_K_S.gguf
2. mmproj-model-f16-12B.gguf
## -Text-to-Text Models
## GPT OSS 20


## Qwen3
CPU. RAM 25GB
OR
GPU. 4 VRAM
1. Qwen3-VL-2B-Instruct-Q8_0.gguf
2. mmproj-Qwen3-VL-2B-Instruct-Q8_0.gguf

## Qwen2.5-Omni
CPU. RAM 40GB
OR
GPU. 8 VRAM
1. Qwen2.5-Omni-7B-BF16.gguf
2. mmproj-F16.gguf 2GB

## Audio-Text-to-Text
## Llama-3.2
CPU. RAM 10GB

1. Llama-3.2-1B-Instruct-Q4_K_M.gguf
2. Llama-3.2-1B-Instruct-Q8_0.gguf
3. mmproj-ultravox-v0_5-llama-3_2-1b-f16.gguf
## run.bat
Local Server
llama-server.exe --n-gpu-layers 2 --ctx-size 111192 -m ".\models\mistralai\mistralai_Voxtral-Mini-3B-2507-Q8_0.gguf" --mmproj ".\models\mistralai\mmproj-mistralai_Voxtral-Mini-3B-2507-bf16.gguf" --host 0.0.0.0Β --portΒ 8005
public URL
llama-server --n-gpu-layers 15 --ctx-size 8192 -m models/ollma/Llama-3.2-1B-Instruct-Q8_0.gguf --mmproj models/ollma/mmproj-ultravox-v0_5-llama-3_2-1b-f16.gguf --host 127.0.0.1Β --portΒ 8083 |