parameters guide
samplers guide
model generation
role play settings
quant selection
arm quants
iq quants vs q quants
optimal model setting
gibberish fixes
coherence
instructing following
quality generation
chat settings
quality settings
llamacpp server
llamacpp
lmstudio
sillytavern
koboldcpp
backyard
ollama
model generation steering
steering
model generation fixes
text generation webui
ggufs
exl2
full precision
quants
imatrix
neo imatrix
llama
llama-3
gemma
gemma2
gemma3
llama-2
llama-3.1
llama-3.2
mistral
Mixture of Experts
mixture of experts
mixtral
Update README.md
Browse files
README.md
CHANGED
|
@@ -197,13 +197,15 @@ Special note:
|
|
| 197 |
|
| 198 |
It appears "DRY" / "XTC" samplers has been added to LLAMACPP and SILLYTAVERN.
|
| 199 |
|
| 200 |
-
It is available (Llamacpp) via "llama-server.exe". Likely this sampler will also become available "downstream" in applications that use LLAMACPP in due time.
|
| 201 |
|
| 202 |
[ https://github.com/ggerganov/llama.cpp/blob/master/examples/server/README.md ]
|
| 203 |
|
| 204 |
Operating Systems:
|
| 205 |
|
| 206 |
-
Most AI/LLM apps operate on Windows, Mac, and Linux.
|
|
|
|
|
|
|
| 207 |
|
| 208 |
---
|
| 209 |
|
|
@@ -574,6 +576,22 @@ A word change, rephrasing, punctation , even a comma, or semi-colon can drastica
|
|
| 574 |
|
| 575 |
Likewise the size, and complexity of your prompt impacts generation too ; especially clarity and direction.
|
| 576 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 577 |
<B>temp / temperature</B>
|
| 578 |
|
| 579 |
temperature (default: 0.8)
|
|
|
|
| 197 |
|
| 198 |
It appears "DRY" / "XTC" samplers has been added to LLAMACPP and SILLYTAVERN.
|
| 199 |
|
| 200 |
+
It is available (Llamacpp) via "server.exe / llama-server.exe". Likely this sampler will also become available "downstream" in applications that use LLAMACPP in due time.
|
| 201 |
|
| 202 |
[ https://github.com/ggerganov/llama.cpp/blob/master/examples/server/README.md ]
|
| 203 |
|
| 204 |
Operating Systems:
|
| 205 |
|
| 206 |
+
Most AI/LLM apps operate on Windows, Mac, and Linux.
|
| 207 |
+
|
| 208 |
+
Mobile devices (and O/S) are in many cases also supported.
|
| 209 |
|
| 210 |
---
|
| 211 |
|
|
|
|
| 576 |
|
| 577 |
Likewise the size, and complexity of your prompt impacts generation too ; especially clarity and direction.
|
| 578 |
|
| 579 |
+
Special note:
|
| 580 |
+
|
| 581 |
+
Pre-prompts / system role are not discussed here. Many of the model repo cards (at my repo) have an optional pre-prompt you can use to aid generation (and can impact instruction following too).
|
| 582 |
+
|
| 583 |
+
Some of my newer models repo cards use a limited form of this called a "prose control" (discussed and shown by example).
|
| 584 |
+
|
| 585 |
+
Roughly a pre-prompt / system role is embedded during each prompt and can act as a guide and/or set of directives for processing the prompt and/or containing generation instructions.
|
| 586 |
+
|
| 587 |
+
A prose control is a simplifed version of this, which preceeds the main prompt(s) - but the idea / effect is relatively the same.
|
| 588 |
+
|
| 589 |
+
I strongly suggest you research these online, as they are a powerful addition to your generation toolbox.
|
| 590 |
+
|
| 591 |
+
They are especially potent with newer model archs due to newer model types having stronger instruction following ablities AND increase context too.
|
| 592 |
+
|
| 593 |
+
PRIMARY SAMPLERS:
|
| 594 |
+
|
| 595 |
<B>temp / temperature</B>
|
| 596 |
|
| 597 |
temperature (default: 0.8)
|