checkpoint54144sd Bluvoll commited on
Commit
31d4a23
·
verified ·
0 Parent(s):

Duplicate from CabalResearch/NoobAI-RectifiedFlow-Experimental

Browse files

Co-authored-by: Bluvoll <Bluvoll@users.noreply.huggingface.co>

.gitattributes ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
NoobAI-RF-0.2-Slight-Aesthetic.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b4f6cab4e3d583e2eea35b84679db933662f29aff19137147f216a61709710ba
3
+ size 6938043248
NoobAI-RF-v0.2-Base.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:391c760519331221049e47c79fe255b537b0c7e198b7308341a531456e39031d
3
+ size 6938043248
Noobai Flow Basic Workflow.json ADDED
@@ -0,0 +1,529 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "id": "a4a1f988-221c-41e4-aaf1-674cec119bc1",
3
+ "revision": 0,
4
+ "last_node_id": 13,
5
+ "last_link_id": 13,
6
+ "nodes": [
7
+ {
8
+ "id": 8,
9
+ "type": "AddReflectionToVAE",
10
+ "pos": [
11
+ 430,
12
+ 340
13
+ ],
14
+ "size": [
15
+ 210,
16
+ 30
17
+ ],
18
+ "flags": {},
19
+ "order": 6,
20
+ "mode": 4,
21
+ "inputs": [
22
+ {
23
+ "name": "vae",
24
+ "type": "VAE",
25
+ "link": 4
26
+ }
27
+ ],
28
+ "outputs": [
29
+ {
30
+ "name": "VAE",
31
+ "type": "VAE",
32
+ "links": [
33
+ 7
34
+ ]
35
+ }
36
+ ],
37
+ "properties": {
38
+ "cnr_id": "comfyui-vae-reflection",
39
+ "ver": "31903567117dccff29df914f2ff7b93fd1255e6e",
40
+ "Node name for S&R": "AddReflectionToVAE"
41
+ },
42
+ "widgets_values": [],
43
+ "shape": 1
44
+ },
45
+ {
46
+ "id": 12,
47
+ "type": "Note",
48
+ "pos": [
49
+ 430,
50
+ 410
51
+ ],
52
+ "size": [
53
+ 210,
54
+ 88
55
+ ],
56
+ "flags": {},
57
+ "order": 0,
58
+ "mode": 0,
59
+ "inputs": [],
60
+ "outputs": [],
61
+ "properties": {},
62
+ "widgets_values": [
63
+ "Padding is optional, but required for correct image edges. You can get the node at: https://github.com/Jelosus2/comfyui-vae-reflection"
64
+ ],
65
+ "color": "#432",
66
+ "bgcolor": "#653",
67
+ "shape": 1
68
+ },
69
+ {
70
+ "id": 6,
71
+ "type": "VAEDecodeTiled",
72
+ "pos": [
73
+ 900,
74
+ 240
75
+ ],
76
+ "size": [
77
+ 220,
78
+ 160
79
+ ],
80
+ "flags": {},
81
+ "order": 8,
82
+ "mode": 0,
83
+ "inputs": [
84
+ {
85
+ "name": "samples",
86
+ "type": "LATENT",
87
+ "link": 2
88
+ },
89
+ {
90
+ "name": "vae",
91
+ "type": "VAE",
92
+ "link": 7
93
+ }
94
+ ],
95
+ "outputs": [
96
+ {
97
+ "name": "IMAGE",
98
+ "type": "IMAGE",
99
+ "slot_index": 0,
100
+ "links": [
101
+ 8
102
+ ]
103
+ }
104
+ ],
105
+ "properties": {
106
+ "cnr_id": "comfy-core",
107
+ "ver": "0.3.26",
108
+ "Node name for S&R": "VAEDecodeTiled"
109
+ },
110
+ "widgets_values": [
111
+ 512,
112
+ 64,
113
+ 64,
114
+ 8
115
+ ],
116
+ "shape": 1
117
+ },
118
+ {
119
+ "id": 11,
120
+ "type": "PreviewImage",
121
+ "pos": [
122
+ 1130,
123
+ 240
124
+ ],
125
+ "size": [
126
+ 460,
127
+ 700
128
+ ],
129
+ "flags": {},
130
+ "order": 9,
131
+ "mode": 0,
132
+ "inputs": [
133
+ {
134
+ "name": "images",
135
+ "type": "IMAGE",
136
+ "link": 8
137
+ }
138
+ ],
139
+ "outputs": [],
140
+ "title": "Image Preview",
141
+ "properties": {
142
+ "cnr_id": "comfy-core",
143
+ "ver": "0.3.26",
144
+ "Node name for S&R": "Tuned"
145
+ },
146
+ "widgets_values": [],
147
+ "color": "#223",
148
+ "bgcolor": "#335",
149
+ "shape": 1
150
+ },
151
+ {
152
+ "id": 3,
153
+ "type": "KSampler",
154
+ "pos": [
155
+ 650,
156
+ 240
157
+ ],
158
+ "size": [
159
+ 240,
160
+ 474
161
+ ],
162
+ "flags": {},
163
+ "order": 7,
164
+ "mode": 0,
165
+ "inputs": [
166
+ {
167
+ "name": "model",
168
+ "type": "MODEL",
169
+ "link": 13
170
+ },
171
+ {
172
+ "name": "positive",
173
+ "type": "CONDITIONING",
174
+ "link": 10
175
+ },
176
+ {
177
+ "name": "negative",
178
+ "type": "CONDITIONING",
179
+ "link": 11
180
+ },
181
+ {
182
+ "name": "latent_image",
183
+ "type": "LATENT",
184
+ "link": 9
185
+ }
186
+ ],
187
+ "outputs": [
188
+ {
189
+ "name": "LATENT",
190
+ "type": "LATENT",
191
+ "slot_index": 0,
192
+ "links": [
193
+ 2
194
+ ]
195
+ }
196
+ ],
197
+ "properties": {
198
+ "cnr_id": "comfy-core",
199
+ "ver": "0.3.26",
200
+ "Node name for S&R": "KSampler"
201
+ },
202
+ "widgets_values": [
203
+ 236456494,
204
+ "fixed",
205
+ 24,
206
+ 7,
207
+ "euler",
208
+ "normal",
209
+ 1
210
+ ],
211
+ "shape": 1
212
+ },
213
+ {
214
+ "id": 13,
215
+ "type": "EmptyLatentImage",
216
+ "pos": [
217
+ 430,
218
+ 540
219
+ ],
220
+ "size": [
221
+ 210,
222
+ 110
223
+ ],
224
+ "flags": {},
225
+ "order": 1,
226
+ "mode": 0,
227
+ "inputs": [],
228
+ "outputs": [
229
+ {
230
+ "name": "LATENT",
231
+ "type": "LATENT",
232
+ "slot_index": 0,
233
+ "links": [
234
+ 9
235
+ ]
236
+ }
237
+ ],
238
+ "properties": {
239
+ "cnr_id": "comfy-core",
240
+ "ver": "0.3.26",
241
+ "Node name for S&R": "EmptyLatentImage"
242
+ },
243
+ "widgets_values": [
244
+ 832,
245
+ 1216,
246
+ 1
247
+ ],
248
+ "shape": 1
249
+ },
250
+ {
251
+ "id": 9,
252
+ "type": "CLIPTextEncode",
253
+ "pos": [
254
+ 10,
255
+ 540
256
+ ],
257
+ "size": [
258
+ 410,
259
+ 110
260
+ ],
261
+ "flags": {},
262
+ "order": 5,
263
+ "mode": 0,
264
+ "inputs": [
265
+ {
266
+ "name": "clip",
267
+ "type": "CLIP",
268
+ "link": 6
269
+ }
270
+ ],
271
+ "outputs": [
272
+ {
273
+ "name": "CONDITIONING",
274
+ "type": "CONDITIONING",
275
+ "slot_index": 0,
276
+ "links": [
277
+ 11
278
+ ]
279
+ }
280
+ ],
281
+ "properties": {
282
+ "cnr_id": "comfy-core",
283
+ "ver": "0.3.26",
284
+ "Node name for S&R": "CLIPTextEncode"
285
+ },
286
+ "widgets_values": [
287
+ "worst quality, normal quality, bad anatomy"
288
+ ],
289
+ "shape": 1
290
+ },
291
+ {
292
+ "id": 10,
293
+ "type": "CLIPTextEncode",
294
+ "pos": [
295
+ 10,
296
+ 380
297
+ ],
298
+ "size": [
299
+ 410,
300
+ 120
301
+ ],
302
+ "flags": {},
303
+ "order": 4,
304
+ "mode": 0,
305
+ "inputs": [
306
+ {
307
+ "name": "clip",
308
+ "type": "CLIP",
309
+ "link": 5
310
+ }
311
+ ],
312
+ "outputs": [
313
+ {
314
+ "name": "CONDITIONING",
315
+ "type": "CONDITIONING",
316
+ "slot_index": 0,
317
+ "links": [
318
+ 10
319
+ ]
320
+ }
321
+ ],
322
+ "properties": {
323
+ "cnr_id": "comfy-core",
324
+ "ver": "0.3.26",
325
+ "Node name for S&R": "CLIPTextEncode"
326
+ },
327
+ "widgets_values": [
328
+ "masterpiece, best quality, 1girl, upper body"
329
+ ],
330
+ "shape": 1
331
+ },
332
+ {
333
+ "id": 1,
334
+ "type": "CheckpointLoaderSimple",
335
+ "pos": [
336
+ 10,
337
+ 240
338
+ ],
339
+ "size": [
340
+ 410,
341
+ 98
342
+ ],
343
+ "flags": {},
344
+ "order": 2,
345
+ "mode": 0,
346
+ "inputs": [],
347
+ "outputs": [
348
+ {
349
+ "name": "MODEL",
350
+ "type": "MODEL",
351
+ "links": [
352
+ 12
353
+ ]
354
+ },
355
+ {
356
+ "name": "CLIP",
357
+ "type": "CLIP",
358
+ "links": [
359
+ 5,
360
+ 6
361
+ ]
362
+ },
363
+ {
364
+ "name": "VAE",
365
+ "type": "VAE",
366
+ "links": [
367
+ 4
368
+ ]
369
+ }
370
+ ],
371
+ "properties": {
372
+ "cnr_id": "comfy-core",
373
+ "ver": "0.3.51",
374
+ "Node name for S&R": "CheckpointLoaderSimple"
375
+ },
376
+ "widgets_values": [
377
+ "NoobAI-RF-0.5TEST-AT-000014.safetensors"
378
+ ],
379
+ "shape": 1
380
+ },
381
+ {
382
+ "id": 2,
383
+ "type": "ModelSamplingSD3",
384
+ "pos": [
385
+ 430,
386
+ 240
387
+ ],
388
+ "size": [
389
+ 210,
390
+ 58
391
+ ],
392
+ "flags": {},
393
+ "order": 3,
394
+ "mode": 0,
395
+ "inputs": [
396
+ {
397
+ "name": "model",
398
+ "type": "MODEL",
399
+ "link": 12
400
+ }
401
+ ],
402
+ "outputs": [
403
+ {
404
+ "name": "MODEL",
405
+ "type": "MODEL",
406
+ "links": [
407
+ 13
408
+ ]
409
+ }
410
+ ],
411
+ "properties": {
412
+ "cnr_id": "comfy-core",
413
+ "ver": "0.3.51",
414
+ "Node name for S&R": "ModelSamplingSD3"
415
+ },
416
+ "widgets_values": [
417
+ 2.5000000000000004
418
+ ],
419
+ "shape": 1
420
+ }
421
+ ],
422
+ "links": [
423
+ [
424
+ 2,
425
+ 3,
426
+ 0,
427
+ 6,
428
+ 0,
429
+ "LATENT"
430
+ ],
431
+ [
432
+ 4,
433
+ 1,
434
+ 2,
435
+ 8,
436
+ 0,
437
+ "VAE"
438
+ ],
439
+ [
440
+ 5,
441
+ 1,
442
+ 1,
443
+ 10,
444
+ 0,
445
+ "CLIP"
446
+ ],
447
+ [
448
+ 6,
449
+ 1,
450
+ 1,
451
+ 9,
452
+ 0,
453
+ "CLIP"
454
+ ],
455
+ [
456
+ 7,
457
+ 8,
458
+ 0,
459
+ 6,
460
+ 1,
461
+ "VAE"
462
+ ],
463
+ [
464
+ 8,
465
+ 6,
466
+ 0,
467
+ 11,
468
+ 0,
469
+ "IMAGE"
470
+ ],
471
+ [
472
+ 9,
473
+ 13,
474
+ 0,
475
+ 3,
476
+ 3,
477
+ "LATENT"
478
+ ],
479
+ [
480
+ 10,
481
+ 10,
482
+ 0,
483
+ 3,
484
+ 1,
485
+ "CONDITIONING"
486
+ ],
487
+ [
488
+ 11,
489
+ 9,
490
+ 0,
491
+ 3,
492
+ 2,
493
+ "CONDITIONING"
494
+ ],
495
+ [
496
+ 12,
497
+ 1,
498
+ 0,
499
+ 2,
500
+ 0,
501
+ "MODEL"
502
+ ],
503
+ [
504
+ 13,
505
+ 2,
506
+ 0,
507
+ 3,
508
+ 0,
509
+ "MODEL"
510
+ ]
511
+ ],
512
+ "groups": [],
513
+ "config": {},
514
+ "extra": {
515
+ "ds": {
516
+ "scale": 1.1167815779424906,
517
+ "offset": [
518
+ 102.55456858602435,
519
+ -4.436191422064923
520
+ ]
521
+ },
522
+ "frontendVersion": "1.23.4",
523
+ "VHS_latentpreview": false,
524
+ "VHS_latentpreviewrate": 0,
525
+ "VHS_MetadataImage": false,
526
+ "VHS_KeepIntermediate": true
527
+ },
528
+ "version": 0.4
529
+ }
README.md ADDED
@@ -0,0 +1,229 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ license_name: fair-ai-public-license-1.0-sd
4
+ license_link: https://freedevproject.org/faipl-1.0-sd/
5
+ base_model:
6
+ - Bluvoll/Experimental_EQ-VAE_NoobAI_tests
7
+ - Laxhar/noobai-XL-Vpred-1.0
8
+ library_name: diffusers
9
+ ---
10
+
11
+ ![Без имени-1 копия](https://cdn-uploads.huggingface.co/production/uploads/633b43d29fe04b13f46c8988/6Ac5cvhD-KBTe9zyktmh8.png)
12
+
13
+ ## Model Details
14
+
15
+ This is an Experimental Conversion of Noobai v-pred to Rectified Flow target, using EQ-VAE.
16
+
17
+
18
+ ![rf vs vpred](https://cdn-uploads.huggingface.co/production/uploads/633b43d29fe04b13f46c8988/e8Qi7d5DWtQQvOg3vvb2n.png)
19
+
20
+ ### Model Description
21
+
22
+ Model is a continuation of Noobai training on same dataset, with new diffusion target and few improvements to existing tag approach*. Given the scope of this undertaking, this is only an experimental version, utilizing only subset of full original data.
23
+
24
+ Current state of model is acceptable for general and research purposes, like Image Generation, Finetuning, LoRA Training, and others. We will provide example settings for common style training approach below.
25
+
26
+ Generally, model is fairly stable, but can suffer certain drawbacks coming from lack of training, like malformed understanding of certain tags and colors in our tests, but are not, or rarely observed, in normal prompts in practice.
27
+
28
+
29
+
30
+ - **Developed by:** Cabal Research (Bluvoll, Anzhc)
31
+ - **Funded by:** Community, Bluvoll
32
+ - **License:** [fair-ai-public-license-1.0-sd](https://freedevproject.org/faipl-1.0-sd/)
33
+ - **Finetuned from model:** [Noobai V-pred 1.0](https://huggingface.co/Laxhar/noobai-XL-Vpred-1.0)
34
+
35
+ *Removed massive(in some cases over 6 tags) keep token`, introduced "protected tags", which allows for indiscriminate shuffling, while keeping tokens undroppable.
36
+
37
+ ## Bias and Limitations
38
+
39
+ Due to low budget(~150$ total), we have not been successful in fully stabilizing the model, so you can and will encounter some issues that we were not able to find in our tests, or were not able to address. That wouldn't be too different from the performance of other base models, but your mileage will vary.
40
+
41
+ Most biases of official dataset will apply(Blue Archive, etc.).
42
+
43
+ Some color biases were not reduced, or became more apparent due to some of the quirks in convergence of rectified flow from Noobai v-pred. We did our best to mitigate it by training a bit further, but you will encounter them in certain strong color prompts. Some colors are in unstable state and are hard to achieve due to unfortunate state of their convergence at current step (Black and dark in particular, for example, `dark` will not generate dark image, you need to prompt `dark theme` for that.)
44
+
45
+
46
+ ## Model Output Examples
47
+
48
+ ![01290-2943123450](https://cdn-uploads.huggingface.co/production/uploads/633b43d29fe04b13f46c8988/D9PLWrawp_QEMZyk5dpbl.png)
49
+
50
+ ![01292-1874776530](https://cdn-uploads.huggingface.co/production/uploads/633b43d29fe04b13f46c8988/cN1UUxoywBcA7L3m22bQb.png)
51
+
52
+ ![01291-2943123455](https://cdn-uploads.huggingface.co/production/uploads/633b43d29fe04b13f46c8988/wXp2AtQA6ifv_50_2o7WE.png)
53
+
54
+ ![01294-1874776532](https://cdn-uploads.huggingface.co/production/uploads/633b43d29fe04b13f46c8988/ZVRCxcN-vjUUa6EPDkRsO.png)
55
+
56
+ ![01296-1021002911](https://cdn-uploads.huggingface.co/production/uploads/633b43d29fe04b13f46c8988/y8G-lW4AE7JcRpKu5G3Os.png)
57
+
58
+ ![01298-2775208673](https://cdn-uploads.huggingface.co/production/uploads/633b43d29fe04b13f46c8988/QUNU14SHnw0ILS1rCVHm1.png)
59
+
60
+ ![01293-1874776531](https://cdn-uploads.huggingface.co/production/uploads/633b43d29fe04b13f46c8988/NlrQ1ce66fGqet1GCQTTY.png)
61
+
62
+ ![01295-1874776534](https://cdn-uploads.huggingface.co/production/uploads/633b43d29fe04b13f46c8988/LwEQ3DvhL9GFDkK2zL12V.png)
63
+
64
+ ![01297-2775208672](https://cdn-uploads.huggingface.co/production/uploads/633b43d29fe04b13f46c8988/_jSMO7b7niZs_1stOOo-u.png)
65
+
66
+ ![01287-37110566](https://cdn-uploads.huggingface.co/production/uploads/633b43d29fe04b13f46c8988/pSgA6H158rxIQoe7DYsWQ.png)
67
+
68
+ ![01274-3354982185](https://cdn-uploads.huggingface.co/production/uploads/633b43d29fe04b13f46c8988/sR-JTq5GB573kmAA8Nisa.png)
69
+
70
+ ## Recommendations
71
+
72
+ ### Inference
73
+
74
+ #### Comfy
75
+
76
+ ![изображение](https://cdn-uploads.huggingface.co/production/uploads/633b43d29fe04b13f46c8988/IQFZizmP_NSbEMYE5LC7T.png)
77
+ (Workflow is available alongside model in repo)
78
+
79
+ Same as your normal inference, but with addition of SD3 sampling node, and optional conv padding node, which is required for correct edges(VAE and model has been trained with padded convs in vae, to allow for easier edge content learning.)
80
+
81
+ Recommended Parameters:
82
+ **Sampler**: Euler, Euler A, DPM++ SDE, etc.
83
+ **Steps**: 20-28
84
+ **CFG**: 5-7
85
+ **Schedule**: Normal/Simple
86
+ **Positive Quality Tags**: `masterpiece, best quality`
87
+ **Negative Tags**: `worst quality, normal quality, bad anatomy`
88
+
89
+
90
+ #### A1111 WebUI
91
+
92
+
93
+ Recommended WebUI: [ReForge](https://github.com/Panchovix/stable-diffusion-webui-reForge) - has native support for both RF, and conv padding.
94
+ Possible WebUIs:
95
+ [ErsatzForge](https://github.com/DenOfEquity/ersatzForge) - Has native support for RF, but written in a hardcoded name-checking way, so will not work out of the box. I'm also not able to verify if approach is correct, but it worked after adding the model name to list of checked.
96
+
97
+ **How to use in ReForge**:
98
+
99
+ ![изображение](https://cdn-uploads.huggingface.co/production/uploads/633b43d29fe04b13f46c8988/UV5Yp66H7YlccdQqborPf.png)
100
+ (ignore Sigma max field at the top, this is not used in RF)
101
+
102
+ Support for RF in ReForge is being implemented through a built-in extension:
103
+
104
+ ![изображение](https://cdn-uploads.huggingface.co/production/uploads/633b43d29fe04b13f46c8988/LpMF0lmC96X001Au9fFU_.png)
105
+
106
+ Set parameters to that, and you're good to go.
107
+
108
+ **How to turn on padding**:
109
+
110
+ ![изображение](https://cdn-uploads.huggingface.co/production/uploads/633b43d29fe04b13f46c8988/GmieYDa5l1C9sUiN363xt.png)
111
+
112
+ Turn this on, save, FULLY RELOAD the UI, by closing console and launching it again. This is required. Setting does not change until UI is fully reloaded.
113
+ Recommended Parameters:
114
+ **Sampler**: Euler A Comfy RF, Euler, DPM++ SDE Comfy, etc. **ALL VARIANTS MUST BE RF OR COMFY, IF AVAILABLE. In ComfyUI routing is automatic, but not in the case of WebUI.**
115
+ **Steps**: 20-28
116
+ **CFG**: 5-7
117
+ **Schedule**: Normal/Simple
118
+ **Positive Quality Tags**: `masterpiece, best quality`
119
+ **Negative Tags**: `worst quality, normal quality, bad anatomy`
120
+
121
+ **ADETAILER FIX FOR RF**:
122
+ By default, Adetailer discards Advanced Model Sampling extension, which breaks RF. You need to add AMS to this part of settings:
123
+
124
+ ![изображение](https://cdn-uploads.huggingface.co/production/uploads/633b43d29fe04b13f46c8988/RQMtfm5Xi3V7oNsqXoZJN.png)
125
+
126
+ Add: `advanced_model_sampling_script,advanced_model_sampling_script_backported` to there.
127
+
128
+ If that does not work, go into adetailer extension, find args.py, open it, replace _builtin_scripts like this:
129
+
130
+ ![изображение](https://cdn-uploads.huggingface.co/production/uploads/633b43d29fe04b13f46c8988/rmnS-i_kciJzTZmeR-mGP.png)
131
+
132
+ Here is a copypaste for easy copy:
133
+ ```
134
+ _builtin_script = (
135
+ "advanced_model_sampling_script",
136
+ "advanced_model_sampling_script_backported",
137
+ "hypertile_script",
138
+ "soft_inpainting",
139
+ )
140
+ ```
141
+ ## Training
142
+
143
+ ### Model Composition
144
+ (Relative to base it's trained from)
145
+
146
+ Unet: Same
147
+ CLIP L: Same, Frozen
148
+ CLIP G: Same, Frozen
149
+ VAE: Changed, new VAE - [EQB7](https://huggingface.co/Anzhc/MS-LC-EQ-D-VR_VAE) w/conv padding.
150
+
151
+
152
+ ### Training Details
153
+ (Base / quality-tuned)
154
+
155
+ **Samples seen**(unbatched steps): ~2kk / ~400k
156
+ **Learning Rate**: 2e-5 / 2e-5
157
+ **Effective Batch size**: 1280 (40 real * 4 accum * 8 devices) / 1280 (40 * 4 * 8)
158
+ **Precision**: Full BF16
159
+ **Optimizer**: AdamW8bit with Kahan Summation
160
+ **Weight Decay**: 0.01
161
+ **Schedule**: Constant with warmup
162
+ **Timestep Sampling Strategy**: Logit-Normal (sometimes referred to as Lognorm), Shift 2.5
163
+ **Text Encoders**: Frozen
164
+ **Keep Token**: False (Used "Protected Tags" instead), all tags are shuffled.
165
+ **Tag Dropout**: 10%
166
+ **Uncond Dropout**: 10%
167
+ **Optimal Transport**: True
168
+
169
+
170
+
171
+ **VAE Conv Padding**: True
172
+ **VAE Shift**: 0.1726
173
+ **VAE Scale**: 0.1280
174
+
175
+ (Computed against ~80k of anime images prior to training. Scale is +- same as in base SDXL VAE (negligible difference), but drastically different Shift, 0.1726 vs ~1.60)
176
+
177
+ #### Training Data
178
+
179
+ "Original" Noobai data subset of ~2 million samples, then WAF* subset of ~20 thousand for quality tuning of this intermediate checkpoint. Tags were not changed, data was taken "as-is", as per the wishes of community.
180
+
181
+ *WAF - Weighted Aesthetic Filter, our recent solution for filtering data based on input of multiple scoring models at the same time(at varied weight, adapted for their specific prediction classes/range), including specialized models for specific content. High general threshold was used, resulting in top ~5% of data being selected for quality tuning.
182
+
183
+
184
+ ### LoRA Trainig
185
+
186
+ Current base is highly trainable. We are mostly style trainers and finetuners, so we would give you current recommendation for that, from which you can derive settings you find reasonable based on your experience with other model types.
187
+
188
+ My current style training settings (Anzhc):
189
+
190
+ **Learning Rate**: tested up to **7.5e-4**, LoRA is still stable at that. Somehow. Prolonged training(300+ images for 50 epochs) at that LR did not result in degradation, likely can be pushed even further, likely up to 1e-3, at least at the batch im using.
191
+ **Batch Size**: 144 (6 real * 24 accum), using SGA(Stochastic Gradient Accumulation) - without SGA I probably would lower accum to 4-8.
192
+ **Optimizer**: Adamw8bit with Kahan summation
193
+ **Schedule**: ReREX (Use REX for simplicity)
194
+ **Precision**: Full BF16
195
+ **Weight Decay**: 0.02
196
+ **Timestep Sampling Strategy**: Logit-Normal, Shift 2.5 (Closest to what i use result-wise)
197
+
198
+ **Dim/Alpha/Conv/Alpha**: 24/24/24/24 (Lycoris/Locon)
199
+
200
+ **Text Encoders**: Frozen
201
+
202
+ **Optimal Transport**: True
203
+
204
+ **Expected Dataset Size**: 100 images (Can be even 10, but balance with repeats to roughly this target.)
205
+ **Epochs**: 50 (Yes, even with 10 repeats. 500 effective epochs works just fine and doesn't break from my tests.)
206
+
207
+
208
+
209
+
210
+ ### Hardware
211
+
212
+ Model was trained on cloud 8xA100 node.
213
+
214
+ ### Software
215
+
216
+ Custom fork of [SD-Scripts](https://github.com/bluvoll/sd-scripts)(maintained by Bluvoll)
217
+
218
+ ## Acknowledgements
219
+
220
+ ### Special Thanks
221
+
222
+ **To supporting individuals of the community, who have donated funds to kickstart this training.**
223
+
224
+ - Itterative
225
+ - Sab
226
+ - Puzll
227
+ - Kyonisus
228
+
229
+ It wouldn't have happened without you at this scale.