shreenithi20 commited on
Commit
d653910
·
verified ·
1 Parent(s): 2ca62dc

Upload 3 files

Browse files
lumina_2.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b703b020a8ff07994e06b3dfa1b28625192bdbfa39aa086a33ea6d6fd206b186
3
+ size 10620240765
lumina_consistent_3panels_workflow.json ADDED
@@ -0,0 +1,1109 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "id": "15b7f508-0607-4b30-9b55-626079e783cb",
3
+ "revision": 0,
4
+ "last_node_id": 26,
5
+ "last_link_id": 35,
6
+ "nodes": [
7
+ {
8
+ "id": 14,
9
+ "type": "Note",
10
+ "pos": [
11
+ -438.65673828125,
12
+ -189.08949279785156
13
+ ],
14
+ "size": [
15
+ 310,
16
+ 180
17
+ ],
18
+ "flags": {},
19
+ "order": 0,
20
+ "mode": 0,
21
+ "inputs": [],
22
+ "outputs": [],
23
+ "properties": {},
24
+ "widgets_values": [
25
+ "The official way to sample this model is: shift 6 with 36 steps\n\nSampling it with lower steps works but you might have to lower the shift value to reduce the amount of artifacts.\n\nEx: 20 steps with shift 3 seems to not produce artifacts"
26
+ ],
27
+ "color": "#432",
28
+ "bgcolor": "#653"
29
+ },
30
+ {
31
+ "id": 15,
32
+ "type": "Note",
33
+ "pos": [
34
+ -84.70723724365234,
35
+ -196.49012756347656
36
+ ],
37
+ "size": [
38
+ 319.26513671875,
39
+ 197.89625549316406
40
+ ],
41
+ "flags": {},
42
+ "order": 1,
43
+ "mode": 0,
44
+ "inputs": [],
45
+ "outputs": [],
46
+ "properties": {},
47
+ "widgets_values": [
48
+ "The \"You are an assistant... <Prompt Start> \" text before the actual prompt is the one used in the official example.\n\nThe reason it is exposed to the user like this is because the model still works if you modify or remove it."
49
+ ],
50
+ "color": "#432",
51
+ "bgcolor": "#653"
52
+ },
53
+ {
54
+ "id": 16,
55
+ "type": "LoraLoaderModelOnly",
56
+ "pos": [
57
+ 16.060720443725586,
58
+ 638.0568237304688
59
+ ],
60
+ "size": [
61
+ 315,
62
+ 82
63
+ ],
64
+ "flags": {},
65
+ "order": 6,
66
+ "mode": 0,
67
+ "inputs": [
68
+ {
69
+ "name": "model",
70
+ "type": "MODEL",
71
+ "link": 18
72
+ }
73
+ ],
74
+ "outputs": [
75
+ {
76
+ "name": "MODEL",
77
+ "type": "MODEL",
78
+ "links": [
79
+ 19
80
+ ]
81
+ }
82
+ ],
83
+ "properties": {
84
+ "cnr_id": "comfy-core",
85
+ "ver": "0.3.27",
86
+ "Node name for S&R": "LoraLoaderModelOnly"
87
+ },
88
+ "widgets_values": [
89
+ "my_first_lumina_lora_v1.safetensors",
90
+ 1
91
+ ]
92
+ },
93
+ {
94
+ "id": 6,
95
+ "type": "CLIPTextEncode",
96
+ "pos": [
97
+ 466.8511657714844,
98
+ 261.74090576171875
99
+ ],
100
+ "size": [
101
+ 423.83001708984375,
102
+ 177.11770629882812
103
+ ],
104
+ "flags": {},
105
+ "order": 7,
106
+ "mode": 0,
107
+ "inputs": [
108
+ {
109
+ "name": "clip",
110
+ "type": "CLIP",
111
+ "link": 3
112
+ }
113
+ ],
114
+ "outputs": [
115
+ {
116
+ "name": "CONDITIONING",
117
+ "type": "CONDITIONING",
118
+ "slot_index": 0,
119
+ "links": [
120
+ 4
121
+ ]
122
+ }
123
+ ],
124
+ "title": "CLIP Text Encode (Positive Prompt)",
125
+ "properties": {
126
+ "cnr_id": "comfy-core",
127
+ "ver": "0.3.27",
128
+ "Node name for S&R": "CLIPTextEncode"
129
+ },
130
+ "widgets_values": [
131
+ "You are an assistant designed to generate superior images with the superior degree of image-text alignment based on textual prompts or user prompts. <Prompt Start> A cheerful barista with long brown hair in a vibrant red apron, pouring coffee in a colorful café comic style"
132
+ ],
133
+ "color": "#232",
134
+ "bgcolor": "#353"
135
+ },
136
+ {
137
+ "id": 13,
138
+ "type": "EmptySD3LatentImage",
139
+ "pos": [
140
+ 529.3159790039062,
141
+ 714.1939086914062
142
+ ],
143
+ "size": [
144
+ 315,
145
+ 106
146
+ ],
147
+ "flags": {},
148
+ "order": 2,
149
+ "mode": 0,
150
+ "inputs": [],
151
+ "outputs": [
152
+ {
153
+ "name": "LATENT",
154
+ "type": "LATENT",
155
+ "slot_index": 0,
156
+ "links": [
157
+ 17
158
+ ]
159
+ }
160
+ ],
161
+ "properties": {
162
+ "cnr_id": "comfy-core",
163
+ "ver": "0.3.27",
164
+ "Node name for S&R": "EmptySD3LatentImage"
165
+ },
166
+ "widgets_values": [
167
+ 1024,
168
+ 1024,
169
+ 1
170
+ ]
171
+ },
172
+ {
173
+ "id": 8,
174
+ "type": "VAEDecode",
175
+ "pos": [
176
+ 1411.045654296875,
177
+ 130.90013122558594
178
+ ],
179
+ "size": [
180
+ 210,
181
+ 46
182
+ ],
183
+ "flags": {},
184
+ "order": 15,
185
+ "mode": 0,
186
+ "inputs": [
187
+ {
188
+ "name": "samples",
189
+ "type": "LATENT",
190
+ "link": 14
191
+ },
192
+ {
193
+ "name": "vae",
194
+ "type": "VAE",
195
+ "link": 8
196
+ }
197
+ ],
198
+ "outputs": [
199
+ {
200
+ "name": "IMAGE",
201
+ "type": "IMAGE",
202
+ "slot_index": 0,
203
+ "links": [
204
+ 16
205
+ ]
206
+ }
207
+ ],
208
+ "properties": {
209
+ "cnr_id": "comfy-core",
210
+ "ver": "0.3.27",
211
+ "Node name for S&R": "VAEDecode"
212
+ },
213
+ "widgets_values": []
214
+ },
215
+ {
216
+ "id": 9,
217
+ "type": "SaveImage",
218
+ "pos": [
219
+ 1826.6982421875,
220
+ -84.63604736328125
221
+ ],
222
+ "size": [
223
+ 566.2537841796875,
224
+ 595.7423095703125
225
+ ],
226
+ "flags": {},
227
+ "order": 18,
228
+ "mode": 0,
229
+ "inputs": [
230
+ {
231
+ "name": "images",
232
+ "type": "IMAGE",
233
+ "link": 16
234
+ }
235
+ ],
236
+ "outputs": [],
237
+ "properties": {
238
+ "cnr_id": "comfy-core",
239
+ "ver": "0.3.27"
240
+ },
241
+ "widgets_values": [
242
+ "ComfyUI",
243
+ ""
244
+ ]
245
+ },
246
+ {
247
+ "id": 18,
248
+ "type": "EmptySD3LatentImage",
249
+ "pos": [
250
+ 524.0302734375,
251
+ 1452.2642822265625
252
+ ],
253
+ "size": [
254
+ 315,
255
+ 106
256
+ ],
257
+ "flags": {},
258
+ "order": 3,
259
+ "mode": 0,
260
+ "inputs": [],
261
+ "outputs": [
262
+ {
263
+ "name": "LATENT",
264
+ "type": "LATENT",
265
+ "slot_index": 0,
266
+ "links": [
267
+ 21
268
+ ]
269
+ }
270
+ ],
271
+ "properties": {
272
+ "cnr_id": "comfy-core",
273
+ "ver": "0.3.27",
274
+ "Node name for S&R": "EmptySD3LatentImage"
275
+ },
276
+ "widgets_values": [
277
+ 1024,
278
+ 1024,
279
+ 1
280
+ ]
281
+ },
282
+ {
283
+ "id": 20,
284
+ "type": "VAEDecode",
285
+ "pos": [
286
+ 1405.7601318359375,
287
+ 868.970703125
288
+ ],
289
+ "size": [
290
+ 210,
291
+ 46
292
+ ],
293
+ "flags": {},
294
+ "order": 16,
295
+ "mode": 0,
296
+ "inputs": [
297
+ {
298
+ "name": "samples",
299
+ "type": "LATENT",
300
+ "link": 22
301
+ },
302
+ {
303
+ "name": "vae",
304
+ "type": "VAE",
305
+ "link": 30
306
+ }
307
+ ],
308
+ "outputs": [
309
+ {
310
+ "name": "IMAGE",
311
+ "type": "IMAGE",
312
+ "slot_index": 0,
313
+ "links": [
314
+ 23
315
+ ]
316
+ }
317
+ ],
318
+ "properties": {
319
+ "cnr_id": "comfy-core",
320
+ "ver": "0.3.27",
321
+ "Node name for S&R": "VAEDecode"
322
+ },
323
+ "widgets_values": []
324
+ },
325
+ {
326
+ "id": 21,
327
+ "type": "SaveImage",
328
+ "pos": [
329
+ 1821.4127197265625,
330
+ 653.4346313476562
331
+ ],
332
+ "size": [
333
+ 566.2537841796875,
334
+ 595.7423095703125
335
+ ],
336
+ "flags": {},
337
+ "order": 19,
338
+ "mode": 0,
339
+ "inputs": [
340
+ {
341
+ "name": "images",
342
+ "type": "IMAGE",
343
+ "link": 23
344
+ }
345
+ ],
346
+ "outputs": [],
347
+ "properties": {
348
+ "cnr_id": "comfy-core",
349
+ "ver": "0.3.27"
350
+ },
351
+ "widgets_values": [
352
+ "ComfyUI",
353
+ ""
354
+ ]
355
+ },
356
+ {
357
+ "id": 23,
358
+ "type": "EmptySD3LatentImage",
359
+ "pos": [
360
+ 525.4942626953125,
361
+ 2197.490478515625
362
+ ],
363
+ "size": [
364
+ 315,
365
+ 106
366
+ ],
367
+ "flags": {},
368
+ "order": 4,
369
+ "mode": 0,
370
+ "inputs": [],
371
+ "outputs": [
372
+ {
373
+ "name": "LATENT",
374
+ "type": "LATENT",
375
+ "slot_index": 0,
376
+ "links": [
377
+ 25
378
+ ]
379
+ }
380
+ ],
381
+ "properties": {
382
+ "cnr_id": "comfy-core",
383
+ "ver": "0.3.27",
384
+ "Node name for S&R": "EmptySD3LatentImage"
385
+ },
386
+ "widgets_values": [
387
+ 1024,
388
+ 1024,
389
+ 1
390
+ ]
391
+ },
392
+ {
393
+ "id": 25,
394
+ "type": "VAEDecode",
395
+ "pos": [
396
+ 1407.2235107421875,
397
+ 1614.197265625
398
+ ],
399
+ "size": [
400
+ 210,
401
+ 46
402
+ ],
403
+ "flags": {},
404
+ "order": 17,
405
+ "mode": 0,
406
+ "inputs": [
407
+ {
408
+ "name": "samples",
409
+ "type": "LATENT",
410
+ "link": 26
411
+ },
412
+ {
413
+ "name": "vae",
414
+ "type": "VAE",
415
+ "link": 31
416
+ }
417
+ ],
418
+ "outputs": [
419
+ {
420
+ "name": "IMAGE",
421
+ "type": "IMAGE",
422
+ "slot_index": 0,
423
+ "links": [
424
+ 27
425
+ ]
426
+ }
427
+ ],
428
+ "properties": {
429
+ "cnr_id": "comfy-core",
430
+ "ver": "0.3.27",
431
+ "Node name for S&R": "VAEDecode"
432
+ },
433
+ "widgets_values": []
434
+ },
435
+ {
436
+ "id": 26,
437
+ "type": "SaveImage",
438
+ "pos": [
439
+ 1822.8760986328125,
440
+ 1398.6610107421875
441
+ ],
442
+ "size": [
443
+ 566.2537841796875,
444
+ 595.7423095703125
445
+ ],
446
+ "flags": {},
447
+ "order": 20,
448
+ "mode": 0,
449
+ "inputs": [
450
+ {
451
+ "name": "images",
452
+ "type": "IMAGE",
453
+ "link": 27
454
+ }
455
+ ],
456
+ "outputs": [],
457
+ "properties": {
458
+ "cnr_id": "comfy-core",
459
+ "ver": "0.3.27"
460
+ },
461
+ "widgets_values": [
462
+ "ComfyUI",
463
+ ""
464
+ ]
465
+ },
466
+ {
467
+ "id": 4,
468
+ "type": "CheckpointLoaderSimple",
469
+ "pos": [
470
+ -412.145263671875,
471
+ 465.4569396972656
472
+ ],
473
+ "size": [
474
+ 315,
475
+ 98
476
+ ],
477
+ "flags": {},
478
+ "order": 5,
479
+ "mode": 0,
480
+ "inputs": [],
481
+ "outputs": [
482
+ {
483
+ "name": "MODEL",
484
+ "type": "MODEL",
485
+ "slot_index": 0,
486
+ "links": [
487
+ 18
488
+ ]
489
+ },
490
+ {
491
+ "name": "CLIP",
492
+ "type": "CLIP",
493
+ "slot_index": 1,
494
+ "links": [
495
+ 3,
496
+ 5,
497
+ 28,
498
+ 29
499
+ ]
500
+ },
501
+ {
502
+ "name": "VAE",
503
+ "type": "VAE",
504
+ "slot_index": 2,
505
+ "links": [
506
+ 8,
507
+ 30,
508
+ 31
509
+ ]
510
+ }
511
+ ],
512
+ "properties": {
513
+ "cnr_id": "comfy-core",
514
+ "ver": "0.3.27",
515
+ "Node name for S&R": "CheckpointLoaderSimple"
516
+ },
517
+ "widgets_values": [
518
+ "lumina_2.safetensors"
519
+ ]
520
+ },
521
+ {
522
+ "id": 11,
523
+ "type": "ModelSamplingAuraFlow",
524
+ "pos": [
525
+ 489.6622314453125,
526
+ 82.98014831542969
527
+ ],
528
+ "size": [
529
+ 315,
530
+ 58
531
+ ],
532
+ "flags": {},
533
+ "order": 11,
534
+ "mode": 0,
535
+ "inputs": [
536
+ {
537
+ "name": "model",
538
+ "type": "MODEL",
539
+ "link": 19
540
+ }
541
+ ],
542
+ "outputs": [
543
+ {
544
+ "name": "MODEL",
545
+ "type": "MODEL",
546
+ "slot_index": 0,
547
+ "links": [
548
+ 13,
549
+ 32,
550
+ 33
551
+ ]
552
+ }
553
+ ],
554
+ "properties": {
555
+ "cnr_id": "comfy-core",
556
+ "ver": "0.3.27",
557
+ "Node name for S&R": "ModelSamplingAuraFlow"
558
+ },
559
+ "widgets_values": [
560
+ 6
561
+ ]
562
+ },
563
+ {
564
+ "id": 7,
565
+ "type": "CLIPTextEncode",
566
+ "pos": [
567
+ 458.0665283203125,
568
+ -162.50588989257812
569
+ ],
570
+ "size": [
571
+ 425.27801513671875,
572
+ 180.6060791015625
573
+ ],
574
+ "flags": {},
575
+ "order": 8,
576
+ "mode": 0,
577
+ "inputs": [
578
+ {
579
+ "name": "clip",
580
+ "type": "CLIP",
581
+ "link": 5
582
+ }
583
+ ],
584
+ "outputs": [
585
+ {
586
+ "name": "CONDITIONING",
587
+ "type": "CONDITIONING",
588
+ "slot_index": 0,
589
+ "links": [
590
+ 6,
591
+ 34,
592
+ 35
593
+ ]
594
+ }
595
+ ],
596
+ "title": "CLIP Text Encode (Negative Prompt)",
597
+ "properties": {
598
+ "cnr_id": "comfy-core",
599
+ "ver": "0.3.27",
600
+ "Node name for S&R": "CLIPTextEncode"
601
+ },
602
+ "widgets_values": [
603
+ "flaws in the eyes, flaws in the face, flaws, lowres, non-HDRi, low quality, worst quality,artifacts noise, text, watermark, glitch, deformed, mutated, ugly, disfigured, hands, low resolution, partially rendered objects, deformed or partially rendered eyes, deformed, deformed eyeballs, cross-eyed, blurry, duplicate person"
604
+ ],
605
+ "color": "#322",
606
+ "bgcolor": "#533"
607
+ },
608
+ {
609
+ "id": 17,
610
+ "type": "CLIPTextEncode",
611
+ "pos": [
612
+ 461.5656433105469,
613
+ 999.811279296875
614
+ ],
615
+ "size": [
616
+ 423.83001708984375,
617
+ 177.11770629882812
618
+ ],
619
+ "flags": {},
620
+ "order": 9,
621
+ "mode": 0,
622
+ "inputs": [
623
+ {
624
+ "name": "clip",
625
+ "type": "CLIP",
626
+ "link": 28
627
+ }
628
+ ],
629
+ "outputs": [
630
+ {
631
+ "name": "CONDITIONING",
632
+ "type": "CONDITIONING",
633
+ "slot_index": 0,
634
+ "links": [
635
+ 20
636
+ ]
637
+ }
638
+ ],
639
+ "title": "CLIP Text Encode (Positive Prompt)",
640
+ "properties": {
641
+ "cnr_id": "comfy-core",
642
+ "ver": "0.3.27",
643
+ "Node name for S&R": "CLIPTextEncode"
644
+ },
645
+ "widgets_values": [
646
+ "You are an assistant designed to generate superior images with the superior degree of image-text alignment based on textual prompts or user prompts. <Prompt Start> A cheerful barista with long brown hair in a vibrant red apron, serving pastries in a colorful café comic style"
647
+ ],
648
+ "color": "#232",
649
+ "bgcolor": "#353"
650
+ },
651
+ {
652
+ "id": 22,
653
+ "type": "CLIPTextEncode",
654
+ "pos": [
655
+ 463.02935791015625,
656
+ 1745.038330078125
657
+ ],
658
+ "size": [
659
+ 423.83001708984375,
660
+ 177.11770629882812
661
+ ],
662
+ "flags": {},
663
+ "order": 10,
664
+ "mode": 0,
665
+ "inputs": [
666
+ {
667
+ "name": "clip",
668
+ "type": "CLIP",
669
+ "link": 29
670
+ }
671
+ ],
672
+ "outputs": [
673
+ {
674
+ "name": "CONDITIONING",
675
+ "type": "CONDITIONING",
676
+ "slot_index": 0,
677
+ "links": [
678
+ 24
679
+ ]
680
+ }
681
+ ],
682
+ "title": "CLIP Text Encode (Positive Prompt)",
683
+ "properties": {
684
+ "cnr_id": "comfy-core",
685
+ "ver": "0.3.27",
686
+ "Node name for S&R": "CLIPTextEncode"
687
+ },
688
+ "widgets_values": [
689
+ "You are an assistant designed to generate superior images with the superior degree of image-text alignment based on textual prompts or user prompts. <Prompt Start> A cheerful barista with long brown hair in a vibrant red apron, chatting with customers in a colorful café comic style"
690
+ ],
691
+ "color": "#232",
692
+ "bgcolor": "#353"
693
+ },
694
+ {
695
+ "id": 3,
696
+ "type": "KSampler",
697
+ "pos": [
698
+ 1130.93017578125,
699
+ 449.5379943847656
700
+ ],
701
+ "size": [
702
+ 315,
703
+ 262
704
+ ],
705
+ "flags": {},
706
+ "order": 12,
707
+ "mode": 0,
708
+ "inputs": [
709
+ {
710
+ "name": "model",
711
+ "type": "MODEL",
712
+ "link": 13
713
+ },
714
+ {
715
+ "name": "positive",
716
+ "type": "CONDITIONING",
717
+ "link": 4
718
+ },
719
+ {
720
+ "name": "negative",
721
+ "type": "CONDITIONING",
722
+ "link": 6
723
+ },
724
+ {
725
+ "name": "latent_image",
726
+ "type": "LATENT",
727
+ "link": 17
728
+ }
729
+ ],
730
+ "outputs": [
731
+ {
732
+ "name": "LATENT",
733
+ "type": "LATENT",
734
+ "slot_index": 0,
735
+ "links": [
736
+ 14
737
+ ]
738
+ }
739
+ ],
740
+ "properties": {
741
+ "cnr_id": "comfy-core",
742
+ "ver": "0.3.27",
743
+ "Node name for S&R": "KSampler"
744
+ },
745
+ "widgets_values": [
746
+ 440509408231901,
747
+ "fixed",
748
+ 25,
749
+ 4,
750
+ "res_multistep",
751
+ "simple",
752
+ 1
753
+ ]
754
+ },
755
+ {
756
+ "id": 24,
757
+ "type": "KSampler",
758
+ "pos": [
759
+ 1127.1080322265625,
760
+ 1932.8349609375
761
+ ],
762
+ "size": [
763
+ 315,
764
+ 262
765
+ ],
766
+ "flags": {},
767
+ "order": 14,
768
+ "mode": 0,
769
+ "inputs": [
770
+ {
771
+ "name": "model",
772
+ "type": "MODEL",
773
+ "link": 33
774
+ },
775
+ {
776
+ "name": "positive",
777
+ "type": "CONDITIONING",
778
+ "link": 24
779
+ },
780
+ {
781
+ "name": "negative",
782
+ "type": "CONDITIONING",
783
+ "link": 35
784
+ },
785
+ {
786
+ "name": "latent_image",
787
+ "type": "LATENT",
788
+ "link": 25
789
+ }
790
+ ],
791
+ "outputs": [
792
+ {
793
+ "name": "LATENT",
794
+ "type": "LATENT",
795
+ "slot_index": 0,
796
+ "links": [
797
+ 26
798
+ ]
799
+ }
800
+ ],
801
+ "properties": {
802
+ "cnr_id": "comfy-core",
803
+ "ver": "0.3.27",
804
+ "Node name for S&R": "KSampler"
805
+ },
806
+ "widgets_values": [
807
+ 440509408231901,
808
+ "fixed",
809
+ 25,
810
+ 4,
811
+ "res_multistep",
812
+ "simple",
813
+ 1
814
+ ]
815
+ },
816
+ {
817
+ "id": 19,
818
+ "type": "KSampler",
819
+ "pos": [
820
+ 1125.6446533203125,
821
+ 1187.6083984375
822
+ ],
823
+ "size": [
824
+ 315,
825
+ 262
826
+ ],
827
+ "flags": {},
828
+ "order": 13,
829
+ "mode": 0,
830
+ "inputs": [
831
+ {
832
+ "name": "model",
833
+ "type": "MODEL",
834
+ "link": 32
835
+ },
836
+ {
837
+ "name": "positive",
838
+ "type": "CONDITIONING",
839
+ "link": 20
840
+ },
841
+ {
842
+ "name": "negative",
843
+ "type": "CONDITIONING",
844
+ "link": 34
845
+ },
846
+ {
847
+ "name": "latent_image",
848
+ "type": "LATENT",
849
+ "link": 21
850
+ }
851
+ ],
852
+ "outputs": [
853
+ {
854
+ "name": "LATENT",
855
+ "type": "LATENT",
856
+ "slot_index": 0,
857
+ "links": [
858
+ 22
859
+ ]
860
+ }
861
+ ],
862
+ "properties": {
863
+ "cnr_id": "comfy-core",
864
+ "ver": "0.3.27",
865
+ "Node name for S&R": "KSampler"
866
+ },
867
+ "widgets_values": [
868
+ 440509408231901,
869
+ "fixed",
870
+ 25,
871
+ 4,
872
+ "res_multistep",
873
+ "simple",
874
+ 1
875
+ ]
876
+ }
877
+ ],
878
+ "links": [
879
+ [
880
+ 3,
881
+ 4,
882
+ 1,
883
+ 6,
884
+ 0,
885
+ "CLIP"
886
+ ],
887
+ [
888
+ 4,
889
+ 6,
890
+ 0,
891
+ 3,
892
+ 1,
893
+ "CONDITIONING"
894
+ ],
895
+ [
896
+ 5,
897
+ 4,
898
+ 1,
899
+ 7,
900
+ 0,
901
+ "CLIP"
902
+ ],
903
+ [
904
+ 6,
905
+ 7,
906
+ 0,
907
+ 3,
908
+ 2,
909
+ "CONDITIONING"
910
+ ],
911
+ [
912
+ 8,
913
+ 4,
914
+ 2,
915
+ 8,
916
+ 1,
917
+ "VAE"
918
+ ],
919
+ [
920
+ 13,
921
+ 11,
922
+ 0,
923
+ 3,
924
+ 0,
925
+ "MODEL"
926
+ ],
927
+ [
928
+ 14,
929
+ 3,
930
+ 0,
931
+ 8,
932
+ 0,
933
+ "LATENT"
934
+ ],
935
+ [
936
+ 16,
937
+ 8,
938
+ 0,
939
+ 9,
940
+ 0,
941
+ "IMAGE"
942
+ ],
943
+ [
944
+ 17,
945
+ 13,
946
+ 0,
947
+ 3,
948
+ 3,
949
+ "LATENT"
950
+ ],
951
+ [
952
+ 18,
953
+ 4,
954
+ 0,
955
+ 16,
956
+ 0,
957
+ "MODEL"
958
+ ],
959
+ [
960
+ 19,
961
+ 16,
962
+ 0,
963
+ 11,
964
+ 0,
965
+ "MODEL"
966
+ ],
967
+ [
968
+ 20,
969
+ 17,
970
+ 0,
971
+ 19,
972
+ 1,
973
+ "CONDITIONING"
974
+ ],
975
+ [
976
+ 21,
977
+ 18,
978
+ 0,
979
+ 19,
980
+ 3,
981
+ "LATENT"
982
+ ],
983
+ [
984
+ 22,
985
+ 19,
986
+ 0,
987
+ 20,
988
+ 0,
989
+ "LATENT"
990
+ ],
991
+ [
992
+ 23,
993
+ 20,
994
+ 0,
995
+ 21,
996
+ 0,
997
+ "IMAGE"
998
+ ],
999
+ [
1000
+ 24,
1001
+ 22,
1002
+ 0,
1003
+ 24,
1004
+ 1,
1005
+ "CONDITIONING"
1006
+ ],
1007
+ [
1008
+ 25,
1009
+ 23,
1010
+ 0,
1011
+ 24,
1012
+ 3,
1013
+ "LATENT"
1014
+ ],
1015
+ [
1016
+ 26,
1017
+ 24,
1018
+ 0,
1019
+ 25,
1020
+ 0,
1021
+ "LATENT"
1022
+ ],
1023
+ [
1024
+ 27,
1025
+ 25,
1026
+ 0,
1027
+ 26,
1028
+ 0,
1029
+ "IMAGE"
1030
+ ],
1031
+ [
1032
+ 28,
1033
+ 4,
1034
+ 1,
1035
+ 17,
1036
+ 0,
1037
+ "CLIP"
1038
+ ],
1039
+ [
1040
+ 29,
1041
+ 4,
1042
+ 1,
1043
+ 22,
1044
+ 0,
1045
+ "CLIP"
1046
+ ],
1047
+ [
1048
+ 30,
1049
+ 4,
1050
+ 2,
1051
+ 20,
1052
+ 1,
1053
+ "VAE"
1054
+ ],
1055
+ [
1056
+ 31,
1057
+ 4,
1058
+ 2,
1059
+ 25,
1060
+ 1,
1061
+ "VAE"
1062
+ ],
1063
+ [
1064
+ 32,
1065
+ 11,
1066
+ 0,
1067
+ 19,
1068
+ 0,
1069
+ "MODEL"
1070
+ ],
1071
+ [
1072
+ 33,
1073
+ 11,
1074
+ 0,
1075
+ 24,
1076
+ 0,
1077
+ "MODEL"
1078
+ ],
1079
+ [
1080
+ 34,
1081
+ 7,
1082
+ 0,
1083
+ 19,
1084
+ 2,
1085
+ "CONDITIONING"
1086
+ ],
1087
+ [
1088
+ 35,
1089
+ 7,
1090
+ 0,
1091
+ 24,
1092
+ 2,
1093
+ "CONDITIONING"
1094
+ ]
1095
+ ],
1096
+ "groups": [],
1097
+ "config": {},
1098
+ "extra": {
1099
+ "ds": {
1100
+ "scale": 0.46650738020973376,
1101
+ "offset": [
1102
+ 3405.894589828991,
1103
+ -643.7852115386304
1104
+ ]
1105
+ },
1106
+ "ue_links": []
1107
+ },
1108
+ "version": 0.4
1109
+ }
my_first_lumina_lora_v1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95a76cccd5e73c56d21c9217364e971a08adeeaf3355131865f68ecfcf180c51
3
+ size 57157416