generation
chat settings
sampling
sampler_config
sampling-strategies
parameters guide
samplers guide
decoding
nucleus-sampling
optimized
optimization
experimentation
role play settings
generation-features
optimal model setting
coherence
steering
high_quality
top-k
top-p
temperature
repetition-penalty
sillytavern
koboldcpp
mistral
gemma
llama
gpt2
Major improvements for V1, V2 and related parameters
Browse filesAlso merged V3 with V2, and added ASSISTANT MODE for V2
README.md
CHANGED
|
@@ -138,7 +138,7 @@ Avoid with strict settings.
|
|
| 138 |
Values below 0.96 are not recommended.
|
| 139 |
</details>
|
| 140 |
<details><summary>Adaptive-P (new):</summary>
|
| 141 |
-
|
| 142 |
</details>
|
| 143 |
<details><summary>DRY:</summary>
|
| 144 |
Avoid in any case.
|
|
@@ -157,16 +157,16 @@ Probability - the chance for Threshold to cutoff the desired most likely tokens.
|
|
| 157 |
<details><summary>V1 **-CREATIVE-BALANCED-**</summary>
|
| 158 |
<img src="https://gitlab.com/Azuro721/trueperfect-ai/-/raw/main/PERF1.png" style="float:right; width:200px; height:300px; padding:10px;">
|
| 159 |
|
| 160 |
-
Very fine with very good creativity, level of detail, emotional connections, "surprising" outcomes and descriptions.
|
| 161 |
|
| 162 |
-
|
|
|
|
| 163 |
|
| 164 |
-
|
| 165 |
-
|
| 166 |
-
In this case, Top-K 206 with Repetition Penalty 1.02612 will improve attention to finer details with less attention to creative things and slower progression. **Top-K 134 (with Repetition Penalty 1.02612) is generally more preferred due to being more subtle, with faster progression (good in most cases) and better creative outcomes**.
|
| 167 |
|
| 168 |
In some specific models (most likely), TFS 0.9551 might improve things even further, with more attention, creativity, and performance overall; do not use if you notice overly long descriptions (extremely long after each section).
|
| 169 |
</details>
|
|
|
|
| 170 |
<details><summary>V2 **-INSANE-DETAIL/ATTENTION-**</summary>
|
| 171 |
<img src="https://gitlab.com/Azuro721/trueperfect-ai/-/raw/main/PERF2.png" style="float:right; width:200px; height:300px; padding:10px;">
|
| 172 |
|
|
@@ -176,22 +176,40 @@ OVERKILL.
|
|
| 176 |
|
| 177 |
Maintains insane amount of details, attention, accuracy, and length: focused outputs, "surprising" outcomes and descriptions (noticeably (in some models) less compared to Top-A 0.07 but still generally good).
|
| 178 |
|
| 179 |
-
|
| 180 |
|
| 181 |
-
|
|
|
|
|
|
|
| 182 |
|
| 183 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 184 |
|
| 185 |
-
|
|
|
|
|
|
|
| 186 |
</details>
|
| 187 |
-
<details><summary>V2-Alt **-INSANE-DETAIL/ATTENTION-ASSISTANT-**</summary>
|
| 188 |
-
<img src="https://gitlab.com/Azuro721/trueperfect-ai/-/raw/main/PERF3.png" style="float:right; width:200px; height:300px; padding:10px;">
|
| 189 |
|
| 190 |
-
|
| 191 |
|
| 192 |
-
|
| 193 |
|
| 194 |
-
|
| 195 |
</details>
|
| 196 |
|
| 197 |
To preserve the versatility, I would like to describe complete and specific sampler values below **Additional fine-tuning:**, to aim perfection for any case.
|
|
@@ -215,31 +233,26 @@ Lower temperature will be output less "exciting" and creative results (like less
|
|
| 215 |
|
| 216 |
Higher Top-K will expand the attention to smaller details, and preserve attention to multiple simultaneous events, and also can fix smaller text-related issues (like with quotation marks, asterisks, hyphens and etc.)
|
| 217 |
|
| 218 |
-
Top-K 278, as described earlier, might cause overly descriptive results, which will most likely
|
| 219 |
-
|
| 220 |
-
Top-K 206 is the more attentive one, which fits more with assistant tasks, as it will take away some of creativity, so might not be suitable for Temperature 4.8.
|
| 221 |
-
|
| 222 |
-
Top-K 134 is the middle-balanced one, with better creativity, good level of detail and fine transitions. Recommeded one for in-character actions and strong roleplay scenarios.
|
| 223 |
|
| 224 |
-
|
| 225 |
|
| 226 |
-
|
| 227 |
-
|
| 228 |
-
Temperature 2.4 with Top-K 206 and TFS 0.8413 will output more attentive results, with less variety, emotions and "surprising" moments.
|
| 229 |
-
|
| 230 |
-
Temperature 1.2 with Top-K 206 and TFS 0.8413 will output slightly more attentive results, with even less variety, emotions and "surprising" moments.
|
| 231 |
|
| 232 |
Further experimentation with Top-K might not be possible, mostly due to logical limit for all settings combined.
|
| 233 |
</details>
|
| 234 |
<details><summary>Repetition Penalty:</summary>
|
| 235 |
Base value: 1.12082, which will output more creative, emotional, varied, smart and "exciting" results. But tends to have issues with asterisks and quotation marks; similar to 1.02612, but with more creativity, less descriptions, faster pace, but prone to issues if input has logical inconsistencies or lots of typos.
|
| 236 |
|
| 237 |
-
1.105 (
|
| 238 |
|
| 239 |
-
1.05 (
|
| 240 |
|
| 241 |
1.02612: very specific one; will output very descriptive, attentive and expanded results. Will try to pay attention to noticeably more things compared to other variants. Will preserve character details and much more things as events go by. Great as an **assistant.**. Great for very complex instructions, very complex character cards and complex scenes. Great attention to multiple characters.
|
| 242 |
|
|
|
|
|
|
|
|
|
|
| 243 |
<details><summary>Other values (might output unstable results):</summary>
|
| 244 |
|
| 245 |
Feel free to experiment with these variants, and show any good results (if stable enough to be used for at least ~6K tokens).
|
|
@@ -263,13 +276,14 @@ Base value: 0.915, which will output very attentive, consistent and stable resul
|
|
| 263 |
|
| 264 |
0.905: pays more attention to specific details, slightly less emotions, and very close to being repetitive.
|
| 265 |
|
| 266 |
-
0.95/0.97: very creative and unpredictable; might be used for better models, but generally less attentive.
|
| 267 |
</details>
|
| 268 |
<details><summary>Top-A:</summary>
|
| 269 |
-
Base value: 0.07, which will output very consistent,
|
|
|
|
| 270 |
0.0001: will output insane amount of details, attention, accuracy and other things described in **V2**.
|
| 271 |
<details><summary>Other values (experimental):</summary>
|
| 272 |
-
0.043725: more attention to anatomy, but unstable and tends to be unpredictable. Works better with Repetition Penalty 1.02612 or **V2** but
|
| 273 |
|
| 274 |
0.2025: more creative, descriptive, "exciting" and emotional, but tends to skip some details. Less accurate and have lots of issues with high Temperature; not suitable to be used generally, only to get initial creative inputs.
|
| 275 |
</details>
|
|
|
|
| 138 |
Values below 0.96 are not recommended.
|
| 139 |
</details>
|
| 140 |
<details><summary>Adaptive-P (new):</summary>
|
| 141 |
+
Can't work with high Temperature; Avoid at any case.
|
| 142 |
</details>
|
| 143 |
<details><summary>DRY:</summary>
|
| 144 |
Avoid in any case.
|
|
|
|
| 157 |
<details><summary>V1 **-CREATIVE-BALANCED-**</summary>
|
| 158 |
<img src="https://gitlab.com/Azuro721/trueperfect-ai/-/raw/main/PERF1.png" style="float:right; width:200px; height:300px; padding:10px;">
|
| 159 |
|
| 160 |
+
Very fine with very good creativity, level of detail (might skip some due to higher Top-A compared to V2), emotional connections, "surprising" outcomes and descriptions.
|
| 161 |
|
| 162 |
+
<details><summary>Optional Adjustments</summary>
|
| 163 |
+
If Repetition Penalty 1.12082 outputs overly descriptive results, try improving the descriptions for character cards or altering the instructions, it will fix most of such issues; otherwise (if nothing helps), use 1.121.
|
| 164 |
|
| 165 |
+
In some cases, tends to confuse things like character names, species (mostly from trained data), pronouns, misuse (confuse) of User's actions, improper pronouns and etc (mostly due to low-probability token picks). **In such cases, lower Repetition Penalty to 1.02612; this will noticeably reduce the frequency of such occurrences and alter the outputs in a more focused and less creative way, with attention to finer details and more subtle transitions**.
|
|
|
|
|
|
|
| 166 |
|
| 167 |
In some specific models (most likely), TFS 0.9551 might improve things even further, with more attention, creativity, and performance overall; do not use if you notice overly long descriptions (extremely long after each section).
|
| 168 |
</details>
|
| 169 |
+
</details>
|
| 170 |
<details><summary>V2 **-INSANE-DETAIL/ATTENTION-**</summary>
|
| 171 |
<img src="https://gitlab.com/Azuro721/trueperfect-ai/-/raw/main/PERF2.png" style="float:right; width:200px; height:300px; padding:10px;">
|
| 172 |
|
|
|
|
| 176 |
|
| 177 |
Maintains insane amount of details, attention, accuracy, and length: focused outputs, "surprising" outcomes and descriptions (noticeably (in some models) less compared to Top-A 0.07 but still generally good).
|
| 178 |
|
| 179 |
+
V2 will be more bland and less descriptive compared to V1 in favor of maximum stability and fixes related to general incoherence with character names (mostly from trained data) and improper pronouns (complete removal of very-low-probability token picks that produce rare misspellings and related issues).
|
| 180 |
|
| 181 |
+
<details><summary>ASSISTANT MODE</summary>
|
| 182 |
+
TFS: 0.9551
|
| 183 |
+
Repetition Penalty: 1.02612
|
| 184 |
|
| 185 |
+
Insane for maximum accuracy for ASSISTANT-related tasks (personal assistant); Will be less creative in favor of attention.
|
| 186 |
+
<details><summary>If any issues occur (too detailed with incoherence)</summary>
|
| 187 |
+
Decrease TFS to 0.8413 at cost of lower accuracy and level of detail, but noticeably better stability.
|
| 188 |
+
</details>
|
| 189 |
+
<details><summary>Optional Adjustments (might degrade stability and accuracy)</summary>
|
| 190 |
+
Might provide very good results in specific models, but generally unstable.
|
| 191 |
+
<details><summary>Repetition Penalty 1.12082 / 1.121</summary>
|
| 192 |
+
Will improve creativity with slight reduction of accuracy.
|
| 193 |
+
</details>
|
| 194 |
+
<details><summary>Temperature 4.8 with Top-K 134</summary>
|
| 195 |
+
Will make outputs more lively, creative and "surprising", but might also increase chances of instabilities, such as overly-high description with sudden incoherence and overall degradation over higher amount of tokens.
|
| 196 |
+
</details>
|
| 197 |
+
<details><summary>Temperature 4.8 with Top-K 278 (284 / 296)</summary>
|
| 198 |
+
Will make outputs more lively, creative, "surprising", descriptive and attentive, but tends to have higher chances of instability compared to Top-K134, with even faster degradation.
|
| 199 |
+
</details>
|
| 200 |
+
<details><summary>TFS 0.9551 (Special)</summary>
|
| 201 |
+
Will improve attention to details and overall performance in many aspects.
|
| 202 |
|
| 203 |
+
Special because tends to have more chances to work correctly across different models.
|
| 204 |
+
</details>
|
| 205 |
+
</details>
|
| 206 |
</details>
|
|
|
|
|
|
|
| 207 |
|
| 208 |
+
Lower Repetition Penalty to get even more insane attention and level of detail, but sacrifice a bit of creativity and "surprising" outcomes (especially with 1.02612).
|
| 209 |
|
| 210 |
+
Repetition Penalty 1.02612 is preferred as the lowest point; will output very attentive and detailed descriptions, with other things described earlier.
|
| 211 |
|
| 212 |
+
Use Top-K 134 for faster transitions and attention to more "surprising" moments (works (mostly) only with Top-A 0.07).
|
| 213 |
</details>
|
| 214 |
|
| 215 |
To preserve the versatility, I would like to describe complete and specific sampler values below **Additional fine-tuning:**, to aim perfection for any case.
|
|
|
|
| 233 |
|
| 234 |
Higher Top-K will expand the attention to smaller details, and preserve attention to multiple simultaneous events, and also can fix smaller text-related issues (like with quotation marks, asterisks, hyphens and etc.)
|
| 235 |
|
| 236 |
+
Top-K 278, as described earlier, might cause overly descriptive results, which will most likely lead to incoherent results.
|
|
|
|
|
|
|
|
|
|
|
|
|
| 237 |
|
| 238 |
+
Top-K 206 is the more attentive one, which fits more with assistant tasks, as it will take away some of creativity, but tends to be more repetitive and might lead to incoherence.
|
| 239 |
|
| 240 |
+
Top-K 134 is the middle-balanced one, with better creativity, good level of detail and fine transitions. Recommeded one for in-character actions and strong roleplay scenarios **NOT suitable for Top-A 0.0001**.
|
|
|
|
|
|
|
|
|
|
|
|
|
| 241 |
|
| 242 |
Further experimentation with Top-K might not be possible, mostly due to logical limit for all settings combined.
|
| 243 |
</details>
|
| 244 |
<details><summary>Repetition Penalty:</summary>
|
| 245 |
Base value: 1.12082, which will output more creative, emotional, varied, smart and "exciting" results. But tends to have issues with asterisks and quotation marks; similar to 1.02612, but with more creativity, less descriptions, faster pace, but prone to issues if input has logical inconsistencies or lots of typos.
|
| 246 |
|
| 247 |
+
1.105 (not fine-tuned): specific value I found out during experimentation. Will output less "exciting" results, but fairly better compared to 1.05.
|
| 248 |
|
| 249 |
+
1.05 (not fine-tuned, not recommended): base value, which is widely used in various LLMs. Might output focused results with average creativity (better than 1.02612), but prone to be less stable compared to 1.105.
|
| 250 |
|
| 251 |
1.02612: very specific one; will output very descriptive, attentive and expanded results. Will try to pay attention to noticeably more things compared to other variants. Will preserve character details and much more things as events go by. Great as an **assistant.**. Great for very complex instructions, very complex character cards and complex scenes. Great attention to multiple characters.
|
| 252 |
|
| 253 |
+
1.15 (not fine-tuned): tends to be more creative with shorter descriptions; might be better with Top-A 0.0001 and might be incoherent (might perform well on specific models).
|
| 254 |
+
|
| 255 |
+
1.23 (not fine-tuned): tends to be even more creative with slight shorter descriptions; might be better with Top-A 0.0001 and tends to have less chances to be coherent, but might perform well in rare cases (with specific models).
|
| 256 |
<details><summary>Other values (might output unstable results):</summary>
|
| 257 |
|
| 258 |
Feel free to experiment with these variants, and show any good results (if stable enough to be used for at least ~6K tokens).
|
|
|
|
| 276 |
|
| 277 |
0.905: pays more attention to specific details, slightly less emotions, and very close to being repetitive.
|
| 278 |
|
| 279 |
+
0.95/0.97: very creative and unpredictable; might be used for better models, but generally less attentive (might perform well on higher-quality models (12B+)).
|
| 280 |
</details>
|
| 281 |
<details><summary>Top-A:</summary>
|
| 282 |
+
Base value: 0.07, which will output very consistent, generally stable results, with smooth transitions and relatively good attention to most details. Recommeded for all cases.
|
| 283 |
+
|
| 284 |
0.0001: will output insane amount of details, attention, accuracy and other things described in **V2**.
|
| 285 |
<details><summary>Other values (experimental):</summary>
|
| 286 |
+
0.043725: more attention to anatomy, but unstable and tends to be unpredictable. Works better with Repetition Penalty 1.02612 or **V2** but degrades.
|
| 287 |
|
| 288 |
0.2025: more creative, descriptive, "exciting" and emotional, but tends to skip some details. Less accurate and have lots of issues with high Temperature; not suitable to be used generally, only to get initial creative inputs.
|
| 289 |
</details>
|