parameters guide
samplers guide
model generation
role play settings
quant selection
arm quants
iq quants vs q quants
optimal model setting
gibberish fixes
coherence
instructing following
quality generation
chat settings
quality settings
llamacpp server
llamacpp
lmstudio
sillytavern
koboldcpp
backyard
ollama
model generation steering
steering
model generation fixes
text generation webui
ggufs
exl2
full precision
quants
imatrix
neo imatrix
llama
llama-3
gemma
gemma2
gemma3
llama-2
llama-3.1
llama-3.2
mistral
Mixture of Experts
mixture of experts
mixtral
Update README.md
Browse files
README.md
CHANGED
|
@@ -12,6 +12,20 @@ tags:
|
|
| 12 |
- quality generation
|
| 13 |
- chat settings
|
| 14 |
- quality settings
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 15 |
---
|
| 16 |
|
| 17 |
<h3>Maximizing Model Performance for All Quants Types And Full-Precision using Samplers, Advance Samplers and Parameters Guide</h3>
|
|
@@ -170,6 +184,8 @@ It appears "DRY" / "XTC" samplers has been added to LLAMACPP and SILLYTAVERN.
|
|
| 170 |
|
| 171 |
It is available via "llama-server.exe". Likely this sampler will also become available "downstream" in applications that use LLAMACPP in due time.
|
| 172 |
|
|
|
|
|
|
|
| 173 |
---
|
| 174 |
|
| 175 |
DETAILED NOTES ON PARAMETERS, SAMPLERS and ADVANCED SAMPLERS:
|
|
@@ -199,6 +215,10 @@ Benchmarking-and-Guiding-Adaptive-Sampling-Decoding https://github.com/ZhouYuxua
|
|
| 199 |
|
| 200 |
Depending on the AI/LLM "apps" you are using, additional reference material for parameters / samplers may also exist.
|
| 201 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 202 |
---
|
| 203 |
|
| 204 |
CRITICAL NOTES:
|
|
@@ -341,7 +361,7 @@ https://huggingface.co/EnragedAntelope
|
|
| 341 |
|
| 342 |
https://github.com/EnragedAntelope
|
| 343 |
|
| 344 |
-
Please see sections below this for advanced usage, more details, settings notes etc etc.
|
| 345 |
|
| 346 |
<small>
|
| 347 |
# LLM Parameters Reference Table
|
|
|
|
| 12 |
- quality generation
|
| 13 |
- chat settings
|
| 14 |
- quality settings
|
| 15 |
+
- llamacpp server
|
| 16 |
+
- llamacpp
|
| 17 |
+
- lmstudio
|
| 18 |
+
- sillytavern
|
| 19 |
+
- koboldcpp
|
| 20 |
+
- backyard
|
| 21 |
+
- ollama
|
| 22 |
+
- text generation webui
|
| 23 |
+
- ggufs
|
| 24 |
+
- exl2
|
| 25 |
+
- full precision
|
| 26 |
+
- quants
|
| 27 |
+
- imatrix
|
| 28 |
+
- neo imatrix
|
| 29 |
---
|
| 30 |
|
| 31 |
<h3>Maximizing Model Performance for All Quants Types And Full-Precision using Samplers, Advance Samplers and Parameters Guide</h3>
|
|
|
|
| 184 |
|
| 185 |
It is available via "llama-server.exe". Likely this sampler will also become available "downstream" in applications that use LLAMACPP in due time.
|
| 186 |
|
| 187 |
+
[ https://github.com/ggerganov/llama.cpp/blob/master/examples/server/README.md ]
|
| 188 |
+
|
| 189 |
---
|
| 190 |
|
| 191 |
DETAILED NOTES ON PARAMETERS, SAMPLERS and ADVANCED SAMPLERS:
|
|
|
|
| 215 |
|
| 216 |
Depending on the AI/LLM "apps" you are using, additional reference material for parameters / samplers may also exist.
|
| 217 |
|
| 218 |
+
LLAMACPP-SERVER EXE:
|
| 219 |
+
|
| 220 |
+
https://github.com/ggerganov/llama.cpp/blob/master/examples/server/README.md
|
| 221 |
+
|
| 222 |
---
|
| 223 |
|
| 224 |
CRITICAL NOTES:
|
|
|
|
| 361 |
|
| 362 |
https://github.com/EnragedAntelope
|
| 363 |
|
| 364 |
+
Please see sections below this for advanced usage, more details, settings, notes etc etc.
|
| 365 |
|
| 366 |
<small>
|
| 367 |
# LLM Parameters Reference Table
|