Commit
·
3caf86c
1
Parent(s):
c15682b
Update locon/notes.md
Browse files- locon/notes.md +21 -0
locon/notes.md
CHANGED
|
@@ -0,0 +1,21 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# LoCon Testing
|
| 2 |
+
https://github.com/KohakuBlueleaf/LoCon
|
| 3 |
+
|
| 4 |
+
## Dim/Alpha
|
| 5 |
+
Testing with
|
| 6 |
+
conv: 1dim 1alpha (due to coding error the examples are using 4alpha, has since been fixed in the code base)
|
| 7 |
+
linear(network): 4dim 4alpha
|
| 8 |
+
We will refer to dim and alpha in the format of dim/alpha, ie 1/1, 4/4
|
| 9 |
+
|
| 10 |
+
## Training Time
|
| 11 |
+
Character appears to gain significant accuracy sooner than using Lora (600 steps vs 800). The time per step is a little bit slower but a better analysis of this is required.
|
| 12 |
+
We suspect that we can train characters faster in terms of real time.
|
| 13 |
+
|
| 14 |
+
## Mixing
|
| 15 |
+
Applying style on character appears to work well. The example wlop style applied on Amber is 1/0.25 and 4/1. The character loses some stability but the style LoCon can easily be lowered and still retain great style retention while significantly reducing the impact to character correctness.
|
| 16 |
+
Overall it appears that mixing is more stable with LoCon than lora but this could be due to the dim/alpha and training settings between these two locon being closer compared to LoRA land where popular dim sizes range from 8 to 128.
|
| 17 |
+
|
| 18 |
+
## Finetune model extractions
|
| 19 |
+
To be tested. We suspect the performance should be much better than LoRA extraction since all the model will be extracted from.
|
| 20 |
+
https://github.com/KohakuBlueleaf/LoCon/blob/main/extract_locon.py
|
| 21 |
+
|