generation
chat settings
sampling
sampler_config
sampling-strategies
parameters guide
samplers guide
decoding
nucleus-sampling
optimized
optimization
experimentation
role play settings
generation-features
optimal model setting
coherence
steering
high_quality
top-k
top-p
temperature
repetition-penalty
sillytavern
koboldcpp
mistral
gemma
llama
gpt2
Added more important info for README.md
Browse filesMore info related to main settings, Temperature+Top-K
README.md
CHANGED
|
@@ -166,14 +166,16 @@ Might confuse things like character names, species and etc (due to high Temperat
|
|
| 166 |
|
| 167 |
OVERKILL.
|
| 168 |
|
| 169 |
-
|
|
|
|
|
|
|
| 170 |
|
| 171 |
Lower Repetition Penalty to get even more insane attention and accuracy, but sacrifice a bit of creativity.
|
| 172 |
</details>
|
| 173 |
<details><summary>V2-Alt **-INSANE-DETAIL/ATTENTION-ASSISTANT-**</summary>
|
| 174 |
<img src="https://gitlab.com/Azuro721/trueperfect-ai/-/raw/main/PERF3.png" style="float:right; width:200px; height:300px; padding:10px;">
|
| 175 |
|
| 176 |
-
Insane for maximum accuracy for ASSISTANT tasks; can be used with overly complexive RP scenarios.
|
| 177 |
</details>
|
| 178 |
|
| 179 |
To preserve the versatility, I would like to describe complete and specific sampler values below **Additional fine-tuning:**, to aim perfection for any case.
|
|
@@ -183,23 +185,31 @@ Here I will describe additional values for special cases, as well as dependencie
|
|
| 183 |
<details><summary>Temperature+Top-K (only higher values):</summary>
|
| 184 |
Temperature is related to Top-K, and in order to achieve perfect part for this specific parameters, both of these needs to be adjusted.
|
| 185 |
|
| 186 |
-
For example Temperature 2.4 **needs** to have Top-K 134, increased by 72.
|
| 187 |
|
| 188 |
-
**Acceptable values of Top-K for Temperature 2.4**: 134, 206
|
| 189 |
|
| 190 |
Different values will cause inconsistency, instability and other issues.
|
| 191 |
|
| 192 |
-
**Acceptable values of Temperature for Top-K**:
|
|
|
|
|
|
|
| 193 |
|
| 194 |
Lower temperature will be output less "exciting" and creative results (like less emotions, variety and predictability by one output), and might trigger repetitions, which can be mostly fixed by raising Top-K.
|
| 195 |
|
| 196 |
Higher Top-K will expand the attention to smaller details, and preserve attention to multiple simultaneous events, and also can fix smaller text-related issues (like with quotation marks, asterisks, hyphens and etc.)
|
| 197 |
|
| 198 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 199 |
|
| 200 |
-
Temperature 2.4 with Top-K 206
|
| 201 |
|
| 202 |
-
Temperature 1.2 with Top-K 206
|
| 203 |
</details>
|
| 204 |
<details><summary>Repetition penalty:</summary>
|
| 205 |
Base value: 1.12082, which will output more creative, emotional, varied, smart and "exciting" results. But tends to have issues with asterisks and quotation marks. Good as an **assistant**. Similar to 1.02612, but with more creativity, less descriptions, faster pace, more interesting responses across multiple characters simultaneously.
|
|
@@ -235,7 +245,7 @@ Base value: 0.915, which will output very attentive, consistent and stable resul
|
|
| 235 |
0.95/0.97: very creative and unpredictable; might be used for better models, but generally less attentive.
|
| 236 |
</details>
|
| 237 |
<details><summary>Top-A:</summary>
|
| 238 |
-
Base value: 0.07, which will output extremely consistent, stable results, with smooth
|
| 239 |
</details>
|
| 240 |
<details><summary>Other values (experimental):</summary>
|
| 241 |
0.043725: more attention to anatomy, but unstable and tends to be unpredictable. Works well with Repetition Penalty 1.02612
|
|
|
|
| 166 |
|
| 167 |
OVERKILL.
|
| 168 |
|
| 169 |
+
Maintains insane amount of details, attention, accuracy, and length.
|
| 170 |
+
|
| 171 |
+
Top-K 206 might improve attention to details, but not generally recommended for Temperature 4.8, mostly due to more focused output on certain things, followed by transitions, which lowers the cases with better creativity and some supposed variations in one output.
|
| 172 |
|
| 173 |
Lower Repetition Penalty to get even more insane attention and accuracy, but sacrifice a bit of creativity.
|
| 174 |
</details>
|
| 175 |
<details><summary>V2-Alt **-INSANE-DETAIL/ATTENTION-ASSISTANT-**</summary>
|
| 176 |
<img src="https://gitlab.com/Azuro721/trueperfect-ai/-/raw/main/PERF3.png" style="float:right; width:200px; height:300px; padding:10px;">
|
| 177 |
|
| 178 |
+
Insane for maximum accuracy for ASSISTANT tasks; can be used with overly complexive RP scenarios (with favor of attention). Will be less creative in favor of attention.
|
| 179 |
</details>
|
| 180 |
|
| 181 |
To preserve the versatility, I would like to describe complete and specific sampler values below **Additional fine-tuning:**, to aim perfection for any case.
|
|
|
|
| 185 |
<details><summary>Temperature+Top-K (only higher values):</summary>
|
| 186 |
Temperature is related to Top-K, and in order to achieve perfect part for this specific parameters, both of these needs to be adjusted.
|
| 187 |
|
| 188 |
+
For example Temperature 2.4 **needs** to have at least Top-K 134, increased by 72.
|
| 189 |
|
| 190 |
+
**Acceptable values of Top-K for Temperature 2.4/4.8**: 134, 206.
|
| 191 |
|
| 192 |
Different values will cause inconsistency, instability and other issues.
|
| 193 |
|
| 194 |
+
**Acceptable values of Temperature for Top-K**: 2.4, 4.8.
|
| 195 |
+
|
| 196 |
+
(1.2) is not prioritized mostly because in most cases it will output unsatisfactory results (e.g. bland and boring).
|
| 197 |
|
| 198 |
Lower temperature will be output less "exciting" and creative results (like less emotions, variety and predictability by one output), and might trigger repetitions, which can be mostly fixed by raising Top-K.
|
| 199 |
|
| 200 |
Higher Top-K will expand the attention to smaller details, and preserve attention to multiple simultaneous events, and also can fix smaller text-related issues (like with quotation marks, asterisks, hyphens and etc.)
|
| 201 |
|
| 202 |
+
Top-K 206 is the more attentive one, which fits more with assistant tasks, as it will take away some of creativity, so might not be suitable for Temperature 4.8.
|
| 203 |
+
|
| 204 |
+
Top-K 134 is the middle-balanced one, with better creativity, good level of detail and fine transitions. Recommeded one for in-character actions and strong roleplay scenarios.
|
| 205 |
+
|
| 206 |
+
Temperature 4.8 with Top-K 206 and TFS 0.8413 will output more creative, emotional and "surprising" results compared to lower ones.
|
| 207 |
+
|
| 208 |
+
**Temperature 2.4 with Top-K 206 and TFS 0.9551 will output the best possible results as an assistant.**
|
| 209 |
|
| 210 |
+
Temperature 2.4 with Top-K 206 and TFS 0.8413 will output more attentive results, with less variety, emotions and "surprising" moments.
|
| 211 |
|
| 212 |
+
Temperature 1.2 with Top-K 206 and TFS 0.8413 will output slightly more attentive results, with even less variety, emotions and "surprising" moments.
|
| 213 |
</details>
|
| 214 |
<details><summary>Repetition penalty:</summary>
|
| 215 |
Base value: 1.12082, which will output more creative, emotional, varied, smart and "exciting" results. But tends to have issues with asterisks and quotation marks. Good as an **assistant**. Similar to 1.02612, but with more creativity, less descriptions, faster pace, more interesting responses across multiple characters simultaneously.
|
|
|
|
| 245 |
0.95/0.97: very creative and unpredictable; might be used for better models, but generally less attentive.
|
| 246 |
</details>
|
| 247 |
<details><summary>Top-A:</summary>
|
| 248 |
+
Base value: 0.07, which will output extremely consistent, stable results, with smooth transitions and very good attention to most things. Recommeded for all cases.
|
| 249 |
</details>
|
| 250 |
<details><summary>Other values (experimental):</summary>
|
| 251 |
0.043725: more attention to anatomy, but unstable and tends to be unpredictable. Works well with Repetition Penalty 1.02612
|