quuMerge branch 'main' of https://huggingface.co/alea31415/LyCORIS-experiments into main
Browse files
.gitattributes
CHANGED
|
@@ -110,3 +110,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 110 |
0325_aniscreen_fanart_styles/0325exp-cs1-tte1e-4-step-12000.safetensors filter=lfs diff=lfs merge=lfs -text
|
| 111 |
0325_aniscreen_fanart_styles/0325exp-cs2-step-6000.safetensors filter=lfs diff=lfs merge=lfs -text
|
| 112 |
0324_all_aniscreen_tags/0324exp-cs1.safetensors filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
| 110 |
0325_aniscreen_fanart_styles/0325exp-cs1-tte1e-4-step-12000.safetensors filter=lfs diff=lfs merge=lfs -text
|
| 111 |
0325_aniscreen_fanart_styles/0325exp-cs2-step-6000.safetensors filter=lfs diff=lfs merge=lfs -text
|
| 112 |
0324_all_aniscreen_tags/0324exp-cs1.safetensors filter=lfs diff=lfs merge=lfs -text
|
| 113 |
+
generated_samples/00165-20230327023658.jpg filter=lfs diff=lfs merge=lfs -text
|
| 114 |
+
generated_samples/00165-20230327023658.png filter=lfs diff=lfs merge=lfs -text
|
README.md
CHANGED
|
@@ -2,7 +2,7 @@
|
|
| 2 |
license: creativeml-openrail-m
|
| 3 |
---
|
| 4 |
|
| 5 |
-
|
| 6 |
|
| 7 |
```
|
| 8 |
Anisphia, Euphyllia, Tilty, OyamaMahiro, OyamaMihari
|
|
@@ -14,7 +14,7 @@ For `0324_all_aniscreen_tags`, I accidentally tag all the character images with
|
|
| 14 |
For `0325_aniscreen_fanart_styles`, things are done correctly (anime screenshots tagged as `aniscreen`, fanart tagged as `fanart`).
|
| 15 |
|
| 16 |
|
| 17 |
-
|
| 18 |
|
| 19 |
Default settings are
|
| 20 |
- loha net dim 8, conv dim 4, alpha 1
|
|
@@ -28,7 +28,28 @@ The configuration json files can otherwsie be found in the `config` subdirectori
|
|
| 28 |
However, some experiments concern the effect of tags for which I regenerate the txt file and the difference can not be seen from the configuration file in this case.
|
| 29 |
For now this concerns `05tag` for which tags are only used with probability 0.5.
|
| 30 |
|
| 31 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 32 |
|
| 33 |
Here is the composition of the datasets
|
| 34 |
```
|
|
@@ -51,6 +72,4 @@ Here is the composition of the datasets
|
|
| 51 |
5_artists~onono_imoko: 373
|
| 52 |
7_characters~screenshots~Tilty: 95
|
| 53 |
9_characters~fanart~Anisphia+Euphyllia: 97
|
| 54 |
-
```
|
| 55 |
-
|
| 56 |
-
Tks
|
|
|
|
| 2 |
license: creativeml-openrail-m
|
| 3 |
---
|
| 4 |
|
| 5 |
+
### Trigger words
|
| 6 |
|
| 7 |
```
|
| 8 |
Anisphia, Euphyllia, Tilty, OyamaMahiro, OyamaMihari
|
|
|
|
| 14 |
For `0325_aniscreen_fanart_styles`, things are done correctly (anime screenshots tagged as `aniscreen`, fanart tagged as `fanart`).
|
| 15 |
|
| 16 |
|
| 17 |
+
### Settings
|
| 18 |
|
| 19 |
Default settings are
|
| 20 |
- loha net dim 8, conv dim 4, alpha 1
|
|
|
|
| 28 |
However, some experiments concern the effect of tags for which I regenerate the txt file and the difference can not be seen from the configuration file in this case.
|
| 29 |
For now this concerns `05tag` for which tags are only used with probability 0.5.
|
| 30 |
|
| 31 |
+
### Some observations
|
| 32 |
+
|
| 33 |
+
For a thorough comparaison please refer to the `generated_samples` folder.
|
| 34 |
+
|
| 35 |
+
#### Captioning
|
| 36 |
+
|
| 37 |
+
Dataset, in general, is the most important out of all.
|
| 38 |
+
The common wisdom that we should prune anything that we want to be attach to the trigger word is exactly the way to go for.
|
| 39 |
+
No tags at all is terrible, especially for style training.
|
| 40 |
+
Having all the tags remove the traits from subjects if these tags are not used during sampling (not completely true but more or less the case).
|
| 41 |
+
|
| 42 |
+

|
| 43 |
+
|
| 44 |
+
|
| 45 |
+
#### Others
|
| 46 |
+
|
| 47 |
+
1. I barely see any difference for training at clip skip 1 and 2.
|
| 48 |
+
2. Setting text encoder learning rate to be half of that of unet makes training two times slower while I cannot see how it helps.
|
| 49 |
+
3. The difference between lora, locon, and loha are very subtle.
|
| 50 |
+
4. Training at higher resolution helps generating more complex backgrounds etc, but it is very time-consuming and most of the time it isn't worth it (simpler to just switch base model) unless this is exactly the goal of the lora you're training.
|
| 51 |
+
|
| 52 |
+
### Datasets
|
| 53 |
|
| 54 |
Here is the composition of the datasets
|
| 55 |
```
|
|
|
|
| 72 |
5_artists~onono_imoko: 373
|
| 73 |
7_characters~screenshots~Tilty: 95
|
| 74 |
9_characters~fanart~Anisphia+Euphyllia: 97
|
| 75 |
+
```
|
|
|
|
|
|
generated_samples/00165-20230327023658.jpg
ADDED
|
Git LFS Details
|
generated_samples/00165-20230327023658.png
ADDED
|
Git LFS Details
|
generated_samples/00165-20230327023658.txt
ADDED
|
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
OyamaMahiro; fanart; (w-sitting, wariza:1.1), 1girl, solo, skirt, eating lollipop
|
| 2 |
+
looking at viewer, detailed background, outdoors, in village, cloud
|
| 3 |
+
<lora:0325exp-cs1:1>
|
| 4 |
+
Negative prompt: aniscreen, far from viewer, realistic, bad hands, lowres, bad anatomy, text, error, missing fingers, extra digit, fewer digits, cropped, worst quality, low quality, normal quality, jpeg artifacts, signature, watermark, username, blurry
|
| 5 |
+
Steps: 25, Sampler: Euler a, CFG scale: 8, Seed: 1234, Size: 520x680, Model hash: 89d59c3dde, Model: base_nai-full-pruned, Script: X/Y/Z plot, X Type: Prompt S/R, X Values: "<lora:0325exp-cs1:1>,<lora:0325exp-cs1-tte1e-4:1>,<lora:0326exp-cs1-lora:1>,
|
| 6 |
+
<lora:0325exp-cs1-768-step-28000:1>, <lora:0326exp-cs1-5e-4:1>,<lora:0326exp-cs1-dim32-16:1>, <lora:0326exp-cs1-dim32-16-alpha-half:1>, <lora:0326exp-cs1-dim32-16-5e-4:1>, <lora:0326exp-cs1-character-only:1>", Y Type: Seed, Y Values: "1234,555,98,444,343242", Fixed Y Values: "1234, 555, 98, 444, 343242"
|