NicRaf commited on
Commit
003b478
·
verified ·
1 Parent(s): 7d75543

Upload 17 files

Browse files
.gitattributes CHANGED
@@ -33,3 +33,6 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ text_encoders/Qwen2.5-VL-7B-Instruct-UD-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
37
+ text_encoders/Qwen3-4B-UD-Q6_K_XL.gguf filter=lfs diff=lfs merge=lfs -text
38
+ text_encoders/Qwen3-8B-abliterated-q8_0.gguf filter=lfs diff=lfs merge=lfs -text
loras/A2R_Klein_Standard.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1586036a63f0e68ed2c84b71aaf538115f6549977009eb4b062421624ac2caa
3
+ size 82866696
loras/KLEIN-Unchained-V2.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bcf84b6b72ccef01f6e36565bd8b69ce8513907e80204476cb3e42098b453341
3
+ size 331378480
loras/Qwen-Image-Lightning-4steps-V1.0.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9526e90d71c4290392feeccf3c2172cb77ab3a489f1faeb956637f97acb4c8b1
3
+ size 1698951104
loras/bfs_head_v1_flux-klein_9b_step3500_rank128.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70d8aaf332d710b905d5085afaa87c3ef577edffd54ffcfadeb8c47a854f9044
3
+ size 662729912
loras/bfs_head_v1_flux-klein_9b_step3500_rank128.safetensors.rgthree-info.json ADDED
@@ -0,0 +1,349 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "file": "flux-klein\\bfs_head_v1_flux-klein_9b_step3500_rank128.safetensors",
3
+ "path": "E:\\AI_Softwares\\ImageGenerator\\ComfyUI_windows_portable\\ComfyUI\\models\\loras\\flux-klein\\bfs_head_v1_flux-klein_9b_step3500_rank128.safetensors",
4
+ "modified": 1772192604735.2175,
5
+ "images": [
6
+ {
7
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/97116a13-5c80-4054-8127-eafe3576b28a/original=true/117946449.jpeg",
8
+ "civitaiUrl": "https://civitai.com/images/117946449",
9
+ "width": 4056,
10
+ "height": 2556,
11
+ "type": "image",
12
+ "nsfwLevel": 1,
13
+ "seed": null,
14
+ "positive": null,
15
+ "negative": null,
16
+ "steps": null,
17
+ "sampler": null,
18
+ "cfg": null,
19
+ "model": null,
20
+ "resources": null
21
+ },
22
+ {
23
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/4e569bc7-947a-49ad-bc0d-6ed89a5d14c1/original=true/117946480.jpeg",
24
+ "civitaiUrl": "https://civitai.com/images/117946480",
25
+ "width": 4107,
26
+ "height": 2556,
27
+ "type": "image",
28
+ "nsfwLevel": 4,
29
+ "seed": 432262096973493,
30
+ "positive": "head_swap: start with Picture 1 as the base image, keeping its lighting, environment, and background. remove the head from Picture 1 completely and replace it with the head from Picture 2, strictly preserving the hair, eye color, nose structure of Picture 2. copy the direction of the eye, head rotation, micro expressions from Picture 1, high quality, sharp details, 4k\n\nThe woman has a focused, intense, and slightly defiant facial expression.\n\nKey features:\n- His eyes are wide open and staring directly at the camera with an unblinking gaze.\n- His eyebrows are raised slightly, contributing to an alert or challenging look.\n- His mouth is set in a firm line; it's not smiling but rather held tight, suggesting concentration or determination.\n- There\u2019s visible tension around his jawline and cheekbones, indicating he's exerting effort while skateboarding.\n- A cigarette rests between his lips, adding a casual yet rebellious element to his demeanor.\n\nOverall, this isn't a relaxed or happy expression\u2014it conveys energy, focus, and perhaps a bit of edgy confidence as he rides down the street. The direct eye contact creates a sense of connection and intensity with the viewer.",
31
+ "negative": "bad quality, noise, blurry, worst quality, low resolution, blur, distortion, unnatural blending, cartoon, illustration, painting",
32
+ "steps": 4,
33
+ "sampler": "LCM",
34
+ "cfg": 1,
35
+ "model": null,
36
+ "resources": null
37
+ },
38
+ {
39
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/b9f85a65-8920-4897-97ea-d2bf0e50ff7b/original=true/117946477.jpeg",
40
+ "civitaiUrl": "https://civitai.com/images/117946477",
41
+ "width": 4056,
42
+ "height": 2556,
43
+ "type": "image",
44
+ "nsfwLevel": 1,
45
+ "seed": null,
46
+ "positive": null,
47
+ "negative": null,
48
+ "steps": null,
49
+ "sampler": null,
50
+ "cfg": null,
51
+ "model": null,
52
+ "resources": null
53
+ },
54
+ {
55
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/78ab6400-cfa5-4145-9990-d983c53aac8e/original=true/117947768.jpeg",
56
+ "civitaiUrl": "https://civitai.com/images/117947768",
57
+ "width": 3228,
58
+ "height": 1912,
59
+ "type": "image",
60
+ "nsfwLevel": 1,
61
+ "seed": 432262096973493,
62
+ "positive": "head_swap: start with Picture 1 as the base image, keeping its lighting, environment, and background. remove the head from Picture 1 completely and replace it with the head from Picture 2, strictly preserving the hair, eye color, nose structure of Picture 2. copy the direction of the eye, head rotation, micro expressions from Picture 1, high quality, sharp details, 4k.\n\nDescribe the expression in Picture 1 and copy it to the new image.",
63
+ "negative": "bad quality, noise, blurry, worst quality, low resolution, blur, distortion, unnatural blending, cartoon, illustration, painting",
64
+ "steps": 4,
65
+ "sampler": "LCM",
66
+ "cfg": 1,
67
+ "model": null,
68
+ "resources": null
69
+ }
70
+ ],
71
+ "raw": {
72
+ "metadata": {
73
+ "format": "pt",
74
+ "sshs_legacy_hash": "a00c7d57",
75
+ "ss_output_name": "head_swap_flux-klein_9b",
76
+ "sshs_model_hash": "aa1383e02a3da9765499c0e63e532a1c3d0694346578f86077b1079ed92539ef",
77
+ "training_info": {
78
+ "step": 3500,
79
+ "epoch": 3
80
+ },
81
+ "version": "1.0",
82
+ "name": "head_swap_flux-klein_9b",
83
+ "ss_base_model_version": "flux2_klein_9b",
84
+ "software": {
85
+ "name": "ai-toolkit",
86
+ "repo": "https://github.com/ostris/ai-toolkit",
87
+ "version": "0.7.20"
88
+ },
89
+ "_sha256": "70d8aaf332d710b905d5085afaa87c3ef577edffd54ffcfadeb8c47a854f9044"
90
+ },
91
+ "civitai": {
92
+ "id": 2610018,
93
+ "modelId": 2027766,
94
+ "name": "F.K. 9b - V1 Focus Head",
95
+ "nsfwLevel": 60,
96
+ "createdAt": "2026-01-18T22:54:55.856Z",
97
+ "updatedAt": "2026-03-03T02:06:42.902Z",
98
+ "status": "Published",
99
+ "publishedAt": "2026-01-18T22:59:47.176Z",
100
+ "trainedWords": [
101
+ "head_swap: start with Picture 1 as the base image, keeping its lighting, environment, and background. remove the head from Picture 1 completely and replace it with the head from Picture 2, strictly preserving the hair, eye color, nose structure of Picture 2. copy the direction of the eye, head rotation, micro expressions from Picture 1, high quality, sharp details, 4k. Describe the expression in Picture 1 and copy it to the new image.",
102
+ "head_swap: start with Picture 1 as the base image, keeping its lighting, environment, and background. remove the head from Picture 1 completely and replace it with the head from Picture 2, strictly preserving the hair, eye color, nose structure of Picture 2. copy the direction of the eye, head rotation, micro expressions from Picture 1, high quality, sharp details, 4k."
103
+ ],
104
+ "trainingStatus": null,
105
+ "trainingDetails": null,
106
+ "baseModel": "Flux.2 Klein 9B-base",
107
+ "baseModelType": "Standard",
108
+ "earlyAccessEndsAt": null,
109
+ "earlyAccessConfig": null,
110
+ "description": "<p>One really cool thing about this model is that you can now add information like expressions and other details bellow the original prompt, and it will add them without any problem.<br /></p><p>[Note]</p><p>The training data file is a zip file containing a rank 64 version of LoRa, unlike the attached LoRa file which is rank 128. I advise you to test both and see which one you prefer. Rank 128 is step 3500 of the training, and rank 64 is step 3750.<br /><br />If you want to take a test: <a target=\"_blank\" rel=\"ugc\" href=\"https://www.seaart.ai/workFlowAppDetail/d5n22s5e878c739fpaf0\">SeaArt AI | BFS - Best Face Swap - Flux.2 Klein</a><br /></p>",
111
+ "uploadType": "Created",
112
+ "usageControl": "Download",
113
+ "air": "urn:air:flux2:lora:civitai:2027766@2610018",
114
+ "stats": {
115
+ "downloadCount": 4456,
116
+ "thumbsUpCount": 297
117
+ },
118
+ "model": {
119
+ "name": "BFS - Best Face Swap",
120
+ "type": "LORA",
121
+ "nsfw": true,
122
+ "poi": false
123
+ },
124
+ "files": [
125
+ {
126
+ "id": 2497437,
127
+ "sizeKB": 323613.078125,
128
+ "name": "head_swap_flux-klein_9b_000003750_rank64.zip",
129
+ "type": "Training Data",
130
+ "pickleScanResult": "Success",
131
+ "pickleScanMessage": "No Pickle imports",
132
+ "virusScanResult": "Success",
133
+ "virusScanMessage": null,
134
+ "scannedAt": "2026-01-18T23:40:35.957Z",
135
+ "metadata": {
136
+ "format": "Other",
137
+ "size": null,
138
+ "fp": null
139
+ },
140
+ "hashes": {
141
+ "AutoV1": "670D3DAB",
142
+ "AutoV2": "ACB5EBF462",
143
+ "SHA256": "ACB5EBF462ED835F0530880E31A94DD99C4D88E6D790C526CD5750A0883BC352",
144
+ "CRC32": "6165E35D",
145
+ "BLAKE3": "30B0BB73558F7D2F65361E600342059BF6027DC867E25E282C4CC4B021BDB30B",
146
+ "AutoV3": "C1CD5C0E0783"
147
+ },
148
+ "primary": false,
149
+ "downloadUrl": "https://civitai.com/api/download/models/2610018?type=Training%20Data"
150
+ },
151
+ {
152
+ "id": 2497344,
153
+ "sizeKB": 647197.1796875,
154
+ "name": "head_swap_flux-klein_9b_000003500.safetensors",
155
+ "type": "Model",
156
+ "pickleScanResult": "Success",
157
+ "pickleScanMessage": "No Pickle imports",
158
+ "virusScanResult": "Success",
159
+ "virusScanMessage": null,
160
+ "scannedAt": "2026-01-18T23:05:26.424Z",
161
+ "metadata": {
162
+ "format": "SafeTensor",
163
+ "size": null,
164
+ "fp": null
165
+ },
166
+ "hashes": {
167
+ "AutoV1": "E578C295",
168
+ "AutoV2": "70D8AAF332",
169
+ "SHA256": "70D8AAF332D710B905D5085AFAA87C3EF577EDFFD54FFCFADEB8C47A854F9044",
170
+ "CRC32": "8116A009",
171
+ "BLAKE3": "4AB698E17CF74CD601BB3683C1BF9D72AACE6C256E102BA7292F4DD98562D51B",
172
+ "AutoV3": "AA1383E02A3D"
173
+ },
174
+ "primary": true,
175
+ "downloadUrl": "https://civitai.com/api/download/models/2610018"
176
+ }
177
+ ],
178
+ "images": [
179
+ {
180
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/97116a13-5c80-4054-8127-eafe3576b28a/original=true/117946449.jpeg",
181
+ "nsfwLevel": 1,
182
+ "width": 4056,
183
+ "height": 2556,
184
+ "hash": "U3GtgS{HK~$]005[v=EA0f_LrpoF0JDOXSw]",
185
+ "type": "image",
186
+ "metadata": {
187
+ "hash": "U3GtgS{HK~$]005[v=EA0f_LrpoF0JDOXSw]",
188
+ "size": 9700043,
189
+ "width": 4056,
190
+ "height": 2556
191
+ },
192
+ "minor": false,
193
+ "poi": false,
194
+ "meta": null,
195
+ "availability": "Public",
196
+ "hasMeta": false,
197
+ "hasPositivePrompt": false,
198
+ "onSite": false,
199
+ "remixOfId": null
200
+ },
201
+ {
202
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/4e569bc7-947a-49ad-bc0d-6ed89a5d14c1/original=true/117946480.jpeg",
203
+ "nsfwLevel": 4,
204
+ "width": 4107,
205
+ "height": 2556,
206
+ "hash": "UCHn?6Mdi{RP~qJAJ8OE0f?G?G%LROjGM{jG",
207
+ "type": "image",
208
+ "metadata": {
209
+ "hash": "UCHn?6Mdi{RP~qJAJ8OE0f?G?G%LROjGM{jG",
210
+ "size": 11302983,
211
+ "width": 4107,
212
+ "height": 2556
213
+ },
214
+ "minor": false,
215
+ "poi": false,
216
+ "meta": {
217
+ "seed": 432262096973493,
218
+ "steps": 4,
219
+ "prompt": "head_swap: start with Picture 1 as the base image, keeping its lighting, environment, and background. remove the head from Picture 1 completely and replace it with the head from Picture 2, strictly preserving the hair, eye color, nose structure of Picture 2. copy the direction of the eye, head rotation, micro expressions from Picture 1, high quality, sharp details, 4k\n\nThe woman has a focused, intense, and slightly defiant facial expression.\n\nKey features:\n- His eyes are wide open and staring directly at the camera with an unblinking gaze.\n- His eyebrows are raised slightly, contributing to an alert or challenging look.\n- His mouth is set in a firm line; it's not smiling but rather held tight, suggesting concentration or determination.\n- There\u2019s visible tension around his jawline and cheekbones, indicating he's exerting effort while skateboarding.\n- A cigarette rests between his lips, adding a casual yet rebellious element to his demeanor.\n\nOverall, this isn't a relaxed or happy expression\u2014it conveys energy, focus, and perhaps a bit of edgy confidence as he rides down the street. The direct eye contact creates a sense of connection and intensity with the viewer.",
220
+ "sampler": "LCM",
221
+ "cfgScale": 1,
222
+ "negativePrompt": "bad quality, noise, blurry, worst quality, low resolution, blur, distortion, unnatural blending, cartoon, illustration, painting"
223
+ },
224
+ "availability": "Public",
225
+ "hasMeta": true,
226
+ "hasPositivePrompt": true,
227
+ "onSite": false,
228
+ "remixOfId": null
229
+ },
230
+ {
231
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/b9f85a65-8920-4897-97ea-d2bf0e50ff7b/original=true/117946477.jpeg",
232
+ "nsfwLevel": 1,
233
+ "width": 4056,
234
+ "height": 2556,
235
+ "hash": "U8Klm-5*E}tS:{9@$%,=PAXB-Cn-0Ns7RQjD",
236
+ "type": "image",
237
+ "metadata": {
238
+ "hash": "U8Klm-5*E}tS:{9@$%,=PAXB-Cn-0Ns7RQjD",
239
+ "size": 13323724,
240
+ "width": 4056,
241
+ "height": 2556
242
+ },
243
+ "minor": false,
244
+ "poi": false,
245
+ "meta": null,
246
+ "availability": "Public",
247
+ "hasMeta": false,
248
+ "hasPositivePrompt": false,
249
+ "onSite": false,
250
+ "remixOfId": null
251
+ },
252
+ {
253
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/78ab6400-cfa5-4145-9990-d983c53aac8e/original=true/117947768.jpeg",
254
+ "nsfwLevel": 1,
255
+ "width": 3228,
256
+ "height": 1912,
257
+ "hash": "U7H-[O~X-qtS-qELIUX80g^5^k$%0#.8N^NG",
258
+ "type": "image",
259
+ "metadata": {
260
+ "hash": "U7H-[O~X-qtS-qELIUX80g^5^k$%0#.8N^NG",
261
+ "size": 9057822,
262
+ "width": 3228,
263
+ "height": 1912
264
+ },
265
+ "minor": false,
266
+ "poi": false,
267
+ "meta": {
268
+ "seed": 432262096973493,
269
+ "steps": 4,
270
+ "prompt": "head_swap: start with Picture 1 as the base image, keeping its lighting, environment, and background. remove the head from Picture 1 completely and replace it with the head from Picture 2, strictly preserving the hair, eye color, nose structure of Picture 2. copy the direction of the eye, head rotation, micro expressions from Picture 1, high quality, sharp details, 4k.\n\nDescribe the expression in Picture 1 and copy it to the new image.",
271
+ "sampler": "LCM",
272
+ "cfgScale": 1,
273
+ "negativePrompt": "bad quality, noise, blurry, worst quality, low resolution, blur, distortion, unnatural blending, cartoon, illustration, painting"
274
+ },
275
+ "availability": "Public",
276
+ "hasMeta": true,
277
+ "hasPositivePrompt": true,
278
+ "onSite": false,
279
+ "remixOfId": null
280
+ }
281
+ ],
282
+ "downloadUrl": "https://civitai.com/api/download/models/2610018",
283
+ "_sha256": "70d8aaf332d710b905d5085afaa87c3ef577edffd54ffcfadeb8c47a854f9044",
284
+ "_civitai_api": "https://civitai.com/api/v1/model-versions/by-hash/70d8aaf332d710b905d5085afaa87c3ef577edffd54ffcfadeb8c47a854f9044"
285
+ }
286
+ },
287
+ "trainedWords": [
288
+ {
289
+ "word": "head_swap: start with Picture 1 as the base image",
290
+ "civitai": true
291
+ },
292
+ {
293
+ "word": "keeping its lighting",
294
+ "civitai": true
295
+ },
296
+ {
297
+ "word": "environment",
298
+ "civitai": true
299
+ },
300
+ {
301
+ "word": "and background. remove the head from Picture 1 completely and replace it with the head from Picture 2",
302
+ "civitai": true
303
+ },
304
+ {
305
+ "word": "strictly preserving the hair",
306
+ "civitai": true
307
+ },
308
+ {
309
+ "word": "eye color",
310
+ "civitai": true
311
+ },
312
+ {
313
+ "word": "nose structure of Picture 2. copy the direction of the eye",
314
+ "civitai": true
315
+ },
316
+ {
317
+ "word": "head rotation",
318
+ "civitai": true
319
+ },
320
+ {
321
+ "word": "micro expressions from Picture 1",
322
+ "civitai": true
323
+ },
324
+ {
325
+ "word": "high quality",
326
+ "civitai": true
327
+ },
328
+ {
329
+ "word": "sharp details",
330
+ "civitai": true
331
+ },
332
+ {
333
+ "word": "4k. Describe the expression in Picture 1 and copy it to the new image.",
334
+ "civitai": true
335
+ },
336
+ {
337
+ "word": "4k.",
338
+ "civitai": true
339
+ }
340
+ ],
341
+ "sha256": "70d8aaf332d710b905d5085afaa87c3ef577edffd54ffcfadeb8c47a854f9044",
342
+ "name": "BFS - Best Face Swap - F.K. 9b - V1 Focus Head",
343
+ "type": "LORA",
344
+ "baseModel": "Flux.2 Klein 9B-base",
345
+ "links": [
346
+ "https://civitai.com/models/2027766?modelVersionId=2610018",
347
+ "https://civitai.com/api/v1/model-versions/by-hash/70d8aaf332d710b905d5085afaa87c3ef577edffd54ffcfadeb8c47a854f9044"
348
+ ]
349
+ }
loras/bfs_head_v1_flux-klein_9b_step3750_rank64.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49907a240602488d2846861d7570dd95d80609f44be7e3347965735712b1be16
3
+ size 331379568
loras/f2k_consist_20260225.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b76181b7c07093e287a959ae1cfcd5e2e666dc567bfb98c3178674d9d7e2d79f
3
+ size 570458992
loras/flux20klein2020NSFW.1M27.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0209eddba9755ba209145ab800fd028164f74cfdc1739f96f288d22e851de5f7
3
+ size 165704392
loras/klein_slider_comic.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b790414cac216841cfcafccace9d4361da991267f0fdb62834facc98fb4b29b7
3
+ size 20738136
loras/klein_slider_comic.safetensors.rgthree-info.json ADDED
@@ -0,0 +1,325 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "file": "flux-klein\\klein_slider_comic.safetensors",
3
+ "path": "E:\\AI_Softwares\\ImageGenerator\\ComfyUI_windows_portable\\ComfyUI\\models\\loras\\flux-klein\\klein_slider_comic.safetensors",
4
+ "modified": 1772208422393.8274,
5
+ "images": [
6
+ {
7
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/477d7c78-3012-4f6e-af0c-26633cbb7d34/original=true/118500300.mp4",
8
+ "civitaiUrl": "https://civitai.com/images/118500300",
9
+ "width": 768,
10
+ "height": 768,
11
+ "type": "video",
12
+ "nsfwLevel": 1,
13
+ "seed": null,
14
+ "positive": null,
15
+ "negative": null,
16
+ "steps": null,
17
+ "sampler": null,
18
+ "cfg": null,
19
+ "model": null,
20
+ "resources": null
21
+ },
22
+ {
23
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/07a95be1-c5a7-41fd-830a-26af1dc92514/original=true/118500356.jpeg",
24
+ "civitaiUrl": "https://civitai.com/images/118500356",
25
+ "width": 2048,
26
+ "height": 2048,
27
+ "type": "image",
28
+ "nsfwLevel": 1,
29
+ "seed": 9,
30
+ "positive": null,
31
+ "negative": null,
32
+ "steps": null,
33
+ "sampler": "Euler",
34
+ "cfg": null,
35
+ "model": null,
36
+ "resources": null
37
+ },
38
+ {
39
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/d0a613be-ba68-4ec0-9f54-8bca72556aa5/original=true/118500355.jpeg",
40
+ "civitaiUrl": "https://civitai.com/images/118500355",
41
+ "width": 2048,
42
+ "height": 2048,
43
+ "type": "image",
44
+ "nsfwLevel": 1,
45
+ "seed": 9,
46
+ "positive": null,
47
+ "negative": null,
48
+ "steps": null,
49
+ "sampler": "Euler",
50
+ "cfg": null,
51
+ "model": null,
52
+ "resources": null
53
+ },
54
+ {
55
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/e35ed27e-3e12-4bcd-932e-ef711b5718e5/original=true/118500357.jpeg",
56
+ "civitaiUrl": "https://civitai.com/images/118500357",
57
+ "width": 2048,
58
+ "height": 2048,
59
+ "type": "image",
60
+ "nsfwLevel": 1,
61
+ "seed": 9,
62
+ "positive": null,
63
+ "negative": null,
64
+ "steps": null,
65
+ "sampler": "Euler",
66
+ "cfg": null,
67
+ "model": null,
68
+ "resources": null
69
+ },
70
+ {
71
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/378f32a7-dded-43cf-9457-b88a222087c7/original=true/118502437.jpeg",
72
+ "civitaiUrl": "https://civitai.com/images/118502437",
73
+ "width": 2048,
74
+ "height": 1024,
75
+ "type": "image",
76
+ "nsfwLevel": 1,
77
+ "seed": 9,
78
+ "positive": null,
79
+ "negative": null,
80
+ "steps": null,
81
+ "sampler": "Euler",
82
+ "cfg": null,
83
+ "model": null,
84
+ "resources": null
85
+ }
86
+ ],
87
+ "raw": {
88
+ "metadata": {
89
+ "sshs_model_hash": "35a3e50f517d5f7b9d1021a8ec919a1251294ac5cfd319664a839b079d694736",
90
+ "sshs_legacy_hash": "3c0892df",
91
+ "training_info": {
92
+ "step": 50,
93
+ "epoch": 6
94
+ },
95
+ "format": "pt",
96
+ "name": "klein_slider_comic",
97
+ "software": {
98
+ "name": "ai-toolkit",
99
+ "repo": "https://github.com/ostris/ai-toolkit",
100
+ "version": "0.7.20"
101
+ },
102
+ "version": "1.0",
103
+ "ss_base_model_version": "flux2_klein_9b",
104
+ "ss_output_name": "klein_slider_comic",
105
+ "_sha256": "b790414cac216841cfcafccace9d4361da991267f0fdb62834facc98fb4b29b7"
106
+ },
107
+ "civitai": {
108
+ "id": 2623808,
109
+ "modelId": 2332534,
110
+ "name": "Klein 9B",
111
+ "nsfwLevel": 1,
112
+ "createdAt": "2026-01-23T20:57:02.192Z",
113
+ "updatedAt": "2026-01-23T20:58:23.010Z",
114
+ "status": "Published",
115
+ "publishedAt": "2026-01-23T20:58:22.979Z",
116
+ "trainedWords": [],
117
+ "trainingStatus": null,
118
+ "trainingDetails": null,
119
+ "baseModel": "Flux.2 Klein 9B",
120
+ "baseModelType": "Standard",
121
+ "earlyAccessEndsAt": null,
122
+ "earlyAccessConfig": null,
123
+ "description": null,
124
+ "uploadType": "Created",
125
+ "usageControl": "Download",
126
+ "air": "urn:air:flux2:lora:civitai:2332534@2623808",
127
+ "stats": {
128
+ "downloadCount": 965,
129
+ "thumbsUpCount": 67
130
+ },
131
+ "model": {
132
+ "name": "Klein Realistic / Comic Slider",
133
+ "type": "LORA",
134
+ "nsfw": false,
135
+ "poi": false
136
+ },
137
+ "files": [
138
+ {
139
+ "id": 2511385,
140
+ "sizeKB": 20252.0859375,
141
+ "name": "klein_slider_comic.safetensors",
142
+ "type": "Model",
143
+ "pickleScanResult": "Success",
144
+ "pickleScanMessage": "No Pickle imports",
145
+ "virusScanResult": "Success",
146
+ "virusScanMessage": null,
147
+ "scannedAt": "2026-01-23T21:00:25.205Z",
148
+ "metadata": {
149
+ "format": "SafeTensor",
150
+ "size": null,
151
+ "fp": null
152
+ },
153
+ "hashes": {
154
+ "AutoV1": "B908CE28",
155
+ "AutoV2": "B790414CAC",
156
+ "SHA256": "B790414CAC216841CFCAFCCACE9D4361DA991267F0FDB62834FACC98FB4B29B7",
157
+ "CRC32": "C53DC187",
158
+ "BLAKE3": "57CFB94784EF72440D9F8EA2DDF1A0AAF9B76962AD87E513D84FF0D3A141FF13",
159
+ "AutoV3": "35A3E50F517D"
160
+ },
161
+ "primary": true,
162
+ "downloadUrl": "https://civitai.com/api/download/models/2623808"
163
+ }
164
+ ],
165
+ "images": [
166
+ {
167
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/477d7c78-3012-4f6e-af0c-26633cbb7d34/original=true/118500300.mp4",
168
+ "nsfwLevel": 1,
169
+ "width": 768,
170
+ "height": 768,
171
+ "hash": "UHIf*DpF0h~SWY%JNb%0obag-nM}E4WVSzaf",
172
+ "type": "video",
173
+ "metadata": {
174
+ "hash": "U00000fQfQfQfQfQfQfQfQfQfQfQfQfQfQfQ",
175
+ "size": 466666,
176
+ "audio": false,
177
+ "width": 768,
178
+ "height": 768,
179
+ "duration": 1.563
180
+ },
181
+ "minor": false,
182
+ "poi": false,
183
+ "meta": null,
184
+ "availability": "Public",
185
+ "hasMeta": false,
186
+ "hasPositivePrompt": false,
187
+ "onSite": false,
188
+ "remixOfId": null
189
+ },
190
+ {
191
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/07a95be1-c5a7-41fd-830a-26af1dc92514/original=true/118500356.jpeg",
192
+ "nsfwLevel": 1,
193
+ "width": 2048,
194
+ "height": 2048,
195
+ "hash": "U1I3s.3?01wY0u9_8|9[1H$fD*D+*NXkD+-i",
196
+ "type": "image",
197
+ "metadata": {
198
+ "hash": "U1I3s.3?01wY0u9_8|9[1H$fD*D+*NXkD+-i",
199
+ "size": 7049699,
200
+ "width": 2048,
201
+ "height": 2048,
202
+ "nsfwLevelReason": "Knights Vote"
203
+ },
204
+ "minor": false,
205
+ "poi": false,
206
+ "meta": {
207
+ "seed": 9,
208
+ "vaes": [
209
+ "flux2-vae.safetensors"
210
+ ],
211
+ "comfy": "{\"prompt\": {\"9\": {\"inputs\": {\"filename_prefix\": \"Flux2-Klein\", \"images\": [\"102\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"100\": {\"inputs\": {\"sampler_name\": \"euler\"}, \"class_type\": \"KSamplerSelect\", \"_meta\": {\"title\": \"KSamplerSelect\"}}, \"101\": {\"inputs\": {\"noise\": [\"106\", 0], \"guider\": [\"105\", 0], \"sampler\": [\"100\", 0], \"sigmas\": [\"109\", 0], \"latent_image\": [\"113\", 0]}, \"class_type\": \"SamplerCustomAdvanced\", \"_meta\": {\"title\": \"SamplerCustomAdvanced\"}}, \"102\": {\"inputs\": {\"samples\": [\"101\", 0], \"vae\": [\"104\", 0]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"103\": {\"inputs\": {\"clip_name\": \"qwen_3_8b_fp8mixed.safetensors\", \"type\": \"flux2\", \"device\": \"default\"}, \"class_type\": \"CLIPLoader\", \"_meta\": {\"title\": \"Load CLIP\"}}, \"104\": {\"inputs\": {\"vae_name\": \"flux2-vae.safetensors\"}, \"class_type\": \"VAELoader\", \"_meta\": {\"title\": \"Load VAE\"}}, \"105\": {\"inputs\": {\"cfg\": 1.0, \"model\": [\"116\", 0], \"positive\": [\"110\", 0], \"negative\": [\"111\", 0]}, \"class_type\": \"CFGGuider\", \"_meta\": {\"title\": \"CFGGuider\"}}, \"106\": {\"inputs\": {\"noise_seed\": 9}, \"class_type\": \"RandomNoise\", \"_meta\": {\"title\": \"RandomNoise\"}}, \"107\": {\"inputs\": {\"unet_name\": \"flux-2-klein-9b.safetensors\", \"weight_dtype\": \"default\"}, \"class_type\": \"UNETLoader\", \"_meta\": {\"title\": \"Load Diffusion Model\"}}, \"109\": {\"inputs\": {\"steps\": 4, \"width\": [\"114\", 0], \"height\": [\"115\", 0]}, \"class_type\": \"Flux2Scheduler\", \"_meta\": {\"title\": \"Flux2Scheduler\"}}, \"110\": {\"inputs\": {\"text\": \"A close-up portrait of a young woman with messy auburn hair and green eyes, wearing a knitted wool sweater. Soft cinematic lighting, shallow depth of field, detailed eyelashes, natural expression.\", \"clip\": [\"103\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Positive Prompt)\"}}, \"111\": {\"inputs\": {\"conditioning\": [\"110\", 0]}, \"class_type\": \"ConditioningZeroOut\", \"_meta\": {\"title\": \"ConditioningZeroOut\"}}, \"113\": {\"inputs\": {\"width\": [\"114\", 0], \"height\": [\"115\", 0], \"batch_size\": 1}, \"class_type\": \"EmptyFlux2LatentImage\", \"_meta\": {\"title\": \"Empty Flux 2 Latent\"}}, \"114\": {\"inputs\": {\"value\": 1024}, \"class_type\": \"PrimitiveInt\", \"_meta\": {\"title\": \"Width\"}}, \"115\": {\"inputs\": {\"value\": 1024}, \"class_type\": \"PrimitiveInt\", \"_meta\": {\"title\": \"Height\"}}, \"116\": {\"inputs\": {\"lora_name\": \"klein_slider_comic_000000050.safetensors\", \"strength_model\": 0.0, \"model\": [\"107\", 0]}, \"class_type\": \"LoraLoaderModelOnly\", \"_meta\": {\"title\": \"LoraLoaderModelOnly\"}}}, \"workflow\": {\"id\": \"92112d97-bb64-4b44-86f2-ea5691ef8f6e\", \"revision\": 0, \"last_node_id\": 120, \"last_link_id\": 208, \"nodes\": [{\"id\": 100, \"type\": \"KSamplerSelect\", \"pos\": [433.0836518325727, 250.7621681845402], \"size\": [270, 58], \"flags\": {}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"SAMPLER\", \"type\": \"SAMPLER\", \"links\": [184]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"KSamplerSelect\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [\"euler\"]}, {\"id\": 101, \"type\": \"SamplerCustomAdvanced\", \"pos\": [1020.278813335458, 101.36819395749974], \"size\": [213.275, 106], \"flags\": {}, \"order\": 15, \"mode\": 0, \"inputs\": [{\"name\": \"noise\", \"type\": \"NOISE\", \"link\": 182}, {\"name\": \"guider\", \"type\": \"GUIDER\", \"link\": 183}, {\"name\": \"sampler\", \"type\": \"SAMPLER\", \"link\": 184}, {\"name\": \"sigmas\", \"type\": \"SIGMAS\", \"link\": 185}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 198}], \"outputs\": [{\"name\": \"output\", \"type\": \"LATENT\", \"links\": [186]}, {\"name\": \"denoised_output\", \"type\": \"LATENT\", \"links\": []}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"SamplerCustomAdvanced\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": []}, {\"id\": 102, \"type\": \"VAEDecode\", \"pos\": [1274.2787087340078, 31.368183492318206], \"size\": [220, 46], \"flags\": {}, \"order\": 16, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 186}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 187}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"slot_index\": 0, \"links\": [189]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"VAEDecode\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": []}, {\"id\": 103, \"type\": \"CLIPLoader\", \"pos\": [-205.72122391625783, 231.2015448252655], \"size\": [364.42708333333337, 106], \"flags\": {}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [188]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"CLIPLoader\", \"models\": [{\"name\": \"qwen_3_8b_fp8mixed.safetensors\", \"url\": \"https://huggingface.co/Comfy-Org/flux2-klein-9B/resolve/main/split_files/text_encoders/qwen_3_8b_fp8mixed.safetensors\", \"directory\": \"text_encoders\"}], \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [\"qwen_3_8b_fp8mixed.safetensors\", \"flux2\", \"default\"]}, {\"id\": 104, \"type\": \"VAELoader\", \"pos\": [-205.72122391625783, 401.201573126844], \"size\": [364.42708333333337, 58], \"flags\": {}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [187]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"VAELoader\", \"models\": [{\"name\": \"flux2-vae.safetensors\", \"url\": \"https://huggingface.co/Comfy-Org/flux2-dev/resolve/main/split_files/vae/flux2-vae.safetensors\", \"directory\": \"vae\"}], \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [\"flux2-vae.safetensors\"]}, {\"id\": 111, \"type\": \"ConditioningZeroOut\", \"pos\": [308.27845799443645, 510.418847617246], \"size\": [204.1750030517578, 26], \"flags\": {}, \"order\": 12, \"mode\": 0, \"inputs\": [{\"name\": \"conditioning\", \"type\": \"CONDITIONING\", \"link\": 191}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [196]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.9.1\", \"Node name for S&R\": \"ConditioningZeroOut\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": []}, {\"id\": 97, \"type\": \"MarkdownNote\", \"pos\": [-600, 810], \"size\": [480, 610], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [], \"outputs\": [], \"properties\": {}, \"widgets_values\": [\"Guide: [Subgraph](https://docs.comfy.org/interface/features/subgraph)\\n\\n## Model links (for local users)\\n\\n\\n**diffusion_models**\\n\\n- [flux-2-klein-9b-fp8.safetensors](https://huggingface.co/black-forest-labs/FLUX.2-klein-9b-fp8)\\n\\n> Please visit BFL's repo, accept the agreement in the repo, and then download the models.\\n\\n**text_encoders**\\n\\n- [qwen_3_8b_fp8mixed.safetensors](https://huggingface.co/Comfy-Org/flux2-klein-9B/resolve/main/split_files/text_encoders/qwen_3_8b_fp8mixed.safetensors)\\n\\n**vae**\\n\\n- [flux2-vae.safetensors](https://huggingface.co/Comfy-Org/flux2-dev/resolve/main/split_files/vae/flux2-vae.safetensors)\\n\\n\\nModel Storage Location\\n\\n```\\n\\ud83d\\udcc2 ComfyUI/\\n\\u251c\\u2500\\u2500 \\ud83d\\udcc2 models/\\n\\u2502 \\u251c\\u2500\\u2500 \\ud83d\\udcc2 diffusion_models/\\n\\u2502 \\u2502 \\u2514\\u2500\\u2500 flux-2-klein-9b-fp8.safetensors\\n\\u2502 \\u251c\\u2500\\u2500 \\ud83d\\udcc2 text_encoders/\\n\\u2502 \\u2502 \\u2514\\u2500\\u2500 qwen_3_8b_fp8mixed.safetensors\\n\\u2502 \\u2514\\u2500\\u2500 \\ud83d\\udcc2 vae/\\n\\u2502 \\u2514\\u2500\\u2500 flux2-vae.safetensors\\n```\\n## Report issue\\n\\nNote: please update ComfyUI first ([guide](https://docs.comfy.org/zh-CN/installation/update_comfyui)) and prepare required models. Desktop/Cloud ship stable builds; nightly-supported models may not be included yet, please wait for the next stable release.\\n\\n- Cannot run / runtime errors: [ComfyUI/issues](https://github.com/comfyanonymous/ComfyUI/issues)\\n- UI / frontend issues: [ComfyUI_frontend/issues](https://github.com/Comfy-Org/ComfyUI_frontend/issues)\\n- Workflow issues: [workflow_templates/issues](https://github.com/Comfy-Org/workflow_templates/issues)\\n\"], \"color\": \"#222\", \"bgcolor\": \"#000\"}, {\"id\": 109, \"type\": \"Flux2Scheduler\", \"pos\": [444.57531947043185, 356.1096083006014], \"size\": [270, 106], \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"width\", \"type\": \"INT\", \"widget\": {\"name\": \"width\"}, \"link\": 201}, {\"name\": \"height\", \"type\": \"INT\", \"widget\": {\"name\": \"height\"}, \"link\": 202}], \"outputs\": [{\"name\": \"SIGMAS\", \"type\": \"SIGMAS\", \"links\": [185]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"Flux2Scheduler\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [4, 1024, 1536]}, {\"id\": 113, \"type\": \"EmptyFlux2LatentImage\", \"pos\": [1040.036060147257, 402.4920705620544], \"size\": [270, 106], \"flags\": {}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"width\", \"type\": \"INT\", \"widget\": {\"name\": \"width\"}, \"link\": 199}, {\"name\": \"height\", \"type\": \"INT\", \"widget\": {\"name\": \"height\"}, \"link\": 200}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [198]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"EmptyFlux2LatentImage\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [1024, 1024, 1]}, {\"id\": 114, \"type\": \"PrimitiveInt\", \"pos\": [494.59000806697486, -508.7012168025424], \"size\": [270, 82], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"INT\", \"type\": \"INT\", \"links\": [199, 201]}], \"title\": \"Width\", \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"PrimitiveInt\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [1024, \"fixed\"]}, {\"id\": 105, \"type\": \"CFGGuider\", \"pos\": [433.0836518325727, 110.76212705122477], \"size\": [270, 98], \"flags\": {}, \"order\": 14, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 204}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 195}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 196}], \"outputs\": [{\"name\": \"GUIDER\", \"type\": \"GUIDER\", \"links\": [183]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"CFGGuider\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [1]}, {\"id\": 107, \"type\": \"UNETLoader\", \"pos\": [-348.6323733305521, -232.84532892939507], \"size\": [364.42708333333337, 82], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [192]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"UNETLoader\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [\"flux-2-klein-9b.safetensors\", \"default\"]}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [834.9079794814677, -446.2581642417122], \"size\": [375, 405], \"flags\": {}, \"order\": 17, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 189}], \"outputs\": [], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"SaveImage\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [\"Flux2-Klein\"]}, {\"id\": 115, \"type\": \"PrimitiveInt\", \"pos\": [494.156336121289, -366.941480099023], \"size\": [270, 82], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"INT\", \"type\": \"INT\", \"links\": [200, 202]}], \"title\": \"Height\", \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"PrimitiveInt\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [1024, \"fixed\"]}, {\"id\": 108, \"type\": \"LoraLoaderModelOnly\", \"pos\": [71.00668339294343, -247.979305695621], \"size\": [270, 82], \"flags\": {}, \"order\": 9, \"mode\": 4, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 192}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [203]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.9.2\", \"Node name for S&R\": \"LoraLoaderModelOnly\"}, \"widgets_values\": [\"klein_m4n02_3K.safetensors\", 1]}, {\"id\": 106, \"type\": \"RandomNoise\", \"pos\": [444.53333469869284, -104.13110719389196], \"size\": [270, 82], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"NOISE\", \"type\": \"NOISE\", \"links\": [182]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"RandomNoise\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [9, \"fixed\"]}, {\"id\": 110, \"type\": \"CLIPTextEncode\", \"pos\": [257.3010312783703, 70.9723493087957], \"size\": [450, 380], \"flags\": {}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 188}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"slot_index\": 0, \"links\": [191, 195]}], \"title\": \"CLIP Text Encode (Positive Prompt)\", \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"CLIPTextEncode\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [\"A close-up portrait of a young woman with messy auburn hair and green eyes, wearing a knitted wool sweater. Soft cinematic lighting, shallow depth of field, detailed eyelashes, natural expression.\"], \"color\": \"#232\", \"bgcolor\": \"#353\"}, {\"id\": 116, \"type\": \"LoraLoaderModelOnly\", \"pos\": [399.17761636912036, -234.25210187034656], \"size\": [270, 82], \"flags\": {}, \"order\": 13, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 203}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [204]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.9.2\", \"Node name for S&R\": \"LoraLoaderModelOnly\"}, \"widgets_values\": [\"klein_slider_comic_000000050.safetensors\", 0]}], \"links\": [[182, 106, 0, 101, 0, \"NOISE\"], [183, 105, 0, 101, 1, \"GUIDER\"], [184, 100, 0, 101, 2, \"SAMPLER\"], [185, 109, 0, 101, 3, \"SIGMAS\"], [186, 101, 0, 102, 0, \"LATENT\"], [187, 104, 0, 102, 1, \"VAE\"], [188, 103, 0, 110, 0, \"CLIP\"], [189, 102, 0, 9, 0, \"IMAGE\"], [191, 110, 0, 111, 0, \"CONDITIONING\"], [192, 107, 0, 108, 0, \"MODEL\"], [195, 110, 0, 105, 1, \"CONDITIONING\"], [196, 111, 0, 105, 2, \"CONDITIONING\"], [198, 113, 0, 101, 4, \"LATENT\"], [199, 114, 0, 113, 0, \"INT\"], [200, 115, 0, 113, 1, \"INT\"], [201, 114, 0, 109, 0, \"INT\"], [202, 115, 0, 109, 1, \"INT\"], [203, 108, 0, 116, 0, \"MODEL\"], [204, 116, 0, 105, 0, \"MODEL\"]], \"groups\": [{\"id\": 1, \"title\": \"Models\", \"bounding\": [-209.3878920261178, -2.6595892865098936, 380, 550], \"color\": \"#3f789e\", \"font_size\": 24, \"flags\": {}}, {\"id\": 2, \"title\": \"Prompt\", \"bounding\": [227.30951520925015, 11.175419318941024, 470, 550], \"color\": \"#3f789e\", \"font_size\": 24, \"flags\": {}}, {\"id\": 3, \"title\": \"Sampler\", \"bounding\": [720.6121079738838, -2.6595892865098936, 532.3638671875, 550], \"color\": \"#3f789e\", \"font_size\": 24, \"flags\": {}}], \"config\": {}, \"extra\": {\"frontendVersion\": \"1.36.14\", \"workflowRendererVersion\": \"LG\", \"VHS_latentpreview\": false, \"VHS_latentpreviewrate\": 0, \"VHS_MetadataImage\": true, \"VHS_KeepIntermediate\": true, \"ds\": {\"scale\": 1.3974240692428268, \"offset\": [359.37949133306165, 364.7837773367647]}}, \"version\": 0.4}}",
212
+ "sampler": "Euler"
213
+ },
214
+ "availability": "Public",
215
+ "hasMeta": true,
216
+ "hasPositivePrompt": false,
217
+ "onSite": false,
218
+ "remixOfId": null
219
+ },
220
+ {
221
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/d0a613be-ba68-4ec0-9f54-8bca72556aa5/original=true/118500355.jpeg",
222
+ "nsfwLevel": 1,
223
+ "width": 2048,
224
+ "height": 2048,
225
+ "hash": "UIJ72AtlRQ?FtmkDM{j@o}oLIoRk~AoyM|od",
226
+ "type": "image",
227
+ "metadata": {
228
+ "hash": "UIJ72AtlRQ?FtmkDM{j@o}oLIoRk~AoyM|od",
229
+ "size": 6982925,
230
+ "width": 2048,
231
+ "height": 2048,
232
+ "nsfwLevelReason": "Knights Vote"
233
+ },
234
+ "minor": false,
235
+ "poi": false,
236
+ "meta": {
237
+ "seed": 9,
238
+ "vaes": [
239
+ "flux2-vae.safetensors"
240
+ ],
241
+ "comfy": "{\"prompt\": {\"9\": {\"inputs\": {\"filename_prefix\": \"Flux2-Klein\", \"images\": [\"102\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"100\": {\"inputs\": {\"sampler_name\": \"euler\"}, \"class_type\": \"KSamplerSelect\", \"_meta\": {\"title\": \"KSamplerSelect\"}}, \"101\": {\"inputs\": {\"noise\": [\"106\", 0], \"guider\": [\"105\", 0], \"sampler\": [\"100\", 0], \"sigmas\": [\"109\", 0], \"latent_image\": [\"113\", 0]}, \"class_type\": \"SamplerCustomAdvanced\", \"_meta\": {\"title\": \"SamplerCustomAdvanced\"}}, \"102\": {\"inputs\": {\"samples\": [\"101\", 0], \"vae\": [\"104\", 0]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"103\": {\"inputs\": {\"clip_name\": \"qwen_3_8b_fp8mixed.safetensors\", \"type\": \"flux2\", \"device\": \"default\"}, \"class_type\": \"CLIPLoader\", \"_meta\": {\"title\": \"Load CLIP\"}}, \"104\": {\"inputs\": {\"vae_name\": \"flux2-vae.safetensors\"}, \"class_type\": \"VAELoader\", \"_meta\": {\"title\": \"Load VAE\"}}, \"105\": {\"inputs\": {\"cfg\": 1.0, \"model\": [\"116\", 0], \"positive\": [\"110\", 0], \"negative\": [\"111\", 0]}, \"class_type\": \"CFGGuider\", \"_meta\": {\"title\": \"CFGGuider\"}}, \"106\": {\"inputs\": {\"noise_seed\": 9}, \"class_type\": \"RandomNoise\", \"_meta\": {\"title\": \"RandomNoise\"}}, \"107\": {\"inputs\": {\"unet_name\": \"flux-2-klein-9b.safetensors\", \"weight_dtype\": \"default\"}, \"class_type\": \"UNETLoader\", \"_meta\": {\"title\": \"Load Diffusion Model\"}}, \"109\": {\"inputs\": {\"steps\": 4, \"width\": [\"114\", 0], \"height\": [\"115\", 0]}, \"class_type\": \"Flux2Scheduler\", \"_meta\": {\"title\": \"Flux2Scheduler\"}}, \"110\": {\"inputs\": {\"text\": \"A young woman with her hair in a bun working behind a busy coffee shop counter. She is pouring milk into a ceramic cup. Warm interior lighting, steam from the espresso machine, shelves of coffee beans in the background, realistic fabric texture on her green apron.\", \"clip\": [\"103\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Positive Prompt)\"}}, \"111\": {\"inputs\": {\"conditioning\": [\"110\", 0]}, \"class_type\": \"ConditioningZeroOut\", \"_meta\": {\"title\": \"ConditioningZeroOut\"}}, \"113\": {\"inputs\": {\"width\": [\"114\", 0], \"height\": [\"115\", 0], \"batch_size\": 1}, \"class_type\": \"EmptyFlux2LatentImage\", \"_meta\": {\"title\": \"Empty Flux 2 Latent\"}}, \"114\": {\"inputs\": {\"value\": 1024}, \"class_type\": \"PrimitiveInt\", \"_meta\": {\"title\": \"Width\"}}, \"115\": {\"inputs\": {\"value\": 1024}, \"class_type\": \"PrimitiveInt\", \"_meta\": {\"title\": \"Height\"}}, \"116\": {\"inputs\": {\"lora_name\": \"klein_slider_comic_000000050.safetensors\", \"strength_model\": 0.0, \"model\": [\"107\", 0]}, \"class_type\": \"LoraLoaderModelOnly\", \"_meta\": {\"title\": \"LoraLoaderModelOnly\"}}}, \"workflow\": {\"id\": \"92112d97-bb64-4b44-86f2-ea5691ef8f6e\", \"revision\": 0, \"last_node_id\": 120, \"last_link_id\": 208, \"nodes\": [{\"id\": 100, \"type\": \"KSamplerSelect\", \"pos\": [433.0836518325727, 250.7621681845402], \"size\": [270, 58], \"flags\": {}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"SAMPLER\", \"type\": \"SAMPLER\", \"links\": [184]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"KSamplerSelect\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [\"euler\"]}, {\"id\": 101, \"type\": \"SamplerCustomAdvanced\", \"pos\": [1020.278813335458, 101.36819395749974], \"size\": [213.275, 106], \"flags\": {}, \"order\": 15, \"mode\": 0, \"inputs\": [{\"name\": \"noise\", \"type\": \"NOISE\", \"link\": 182}, {\"name\": \"guider\", \"type\": \"GUIDER\", \"link\": 183}, {\"name\": \"sampler\", \"type\": \"SAMPLER\", \"link\": 184}, {\"name\": \"sigmas\", \"type\": \"SIGMAS\", \"link\": 185}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 198}], \"outputs\": [{\"name\": \"output\", \"type\": \"LATENT\", \"links\": [186]}, {\"name\": \"denoised_output\", \"type\": \"LATENT\", \"links\": []}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"SamplerCustomAdvanced\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": []}, {\"id\": 102, \"type\": \"VAEDecode\", \"pos\": [1274.2787087340078, 31.368183492318206], \"size\": [220, 46], \"flags\": {}, \"order\": 16, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 186}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 187}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"slot_index\": 0, \"links\": [189]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"VAEDecode\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": []}, {\"id\": 103, \"type\": \"CLIPLoader\", \"pos\": [-205.72122391625783, 231.2015448252655], \"size\": [364.42708333333337, 106], \"flags\": {}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [188]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"CLIPLoader\", \"models\": [{\"name\": \"qwen_3_8b_fp8mixed.safetensors\", \"url\": \"https://huggingface.co/Comfy-Org/flux2-klein-9B/resolve/main/split_files/text_encoders/qwen_3_8b_fp8mixed.safetensors\", \"directory\": \"text_encoders\"}], \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [\"qwen_3_8b_fp8mixed.safetensors\", \"flux2\", \"default\"]}, {\"id\": 104, \"type\": \"VAELoader\", \"pos\": [-205.72122391625783, 401.201573126844], \"size\": [364.42708333333337, 58], \"flags\": {}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [187]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"VAELoader\", \"models\": [{\"name\": \"flux2-vae.safetensors\", \"url\": \"https://huggingface.co/Comfy-Org/flux2-dev/resolve/main/split_files/vae/flux2-vae.safetensors\", \"directory\": \"vae\"}], \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [\"flux2-vae.safetensors\"]}, {\"id\": 111, \"type\": \"ConditioningZeroOut\", \"pos\": [308.27845799443645, 510.418847617246], \"size\": [204.1750030517578, 26], \"flags\": {}, \"order\": 12, \"mode\": 0, \"inputs\": [{\"name\": \"conditioning\", \"type\": \"CONDITIONING\", \"link\": 191}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [196]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.9.1\", \"Node name for S&R\": \"ConditioningZeroOut\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": []}, {\"id\": 97, \"type\": \"MarkdownNote\", \"pos\": [-600, 810], \"size\": [480, 610], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [], \"outputs\": [], \"properties\": {}, \"widgets_values\": [\"Guide: [Subgraph](https://docs.comfy.org/interface/features/subgraph)\\n\\n## Model links (for local users)\\n\\n\\n**diffusion_models**\\n\\n- [flux-2-klein-9b-fp8.safetensors](https://huggingface.co/black-forest-labs/FLUX.2-klein-9b-fp8)\\n\\n> Please visit BFL's repo, accept the agreement in the repo, and then download the models.\\n\\n**text_encoders**\\n\\n- [qwen_3_8b_fp8mixed.safetensors](https://huggingface.co/Comfy-Org/flux2-klein-9B/resolve/main/split_files/text_encoders/qwen_3_8b_fp8mixed.safetensors)\\n\\n**vae**\\n\\n- [flux2-vae.safetensors](https://huggingface.co/Comfy-Org/flux2-dev/resolve/main/split_files/vae/flux2-vae.safetensors)\\n\\n\\nModel Storage Location\\n\\n```\\n\\ud83d\\udcc2 ComfyUI/\\n\\u251c\\u2500\\u2500 \\ud83d\\udcc2 models/\\n\\u2502 \\u251c\\u2500\\u2500 \\ud83d\\udcc2 diffusion_models/\\n\\u2502 \\u2502 \\u2514\\u2500\\u2500 flux-2-klein-9b-fp8.safetensors\\n\\u2502 \\u251c\\u2500\\u2500 \\ud83d\\udcc2 text_encoders/\\n\\u2502 \\u2502 \\u2514\\u2500\\u2500 qwen_3_8b_fp8mixed.safetensors\\n\\u2502 \\u2514\\u2500\\u2500 \\ud83d\\udcc2 vae/\\n\\u2502 \\u2514\\u2500\\u2500 flux2-vae.safetensors\\n```\\n## Report issue\\n\\nNote: please update ComfyUI first ([guide](https://docs.comfy.org/zh-CN/installation/update_comfyui)) and prepare required models. Desktop/Cloud ship stable builds; nightly-supported models may not be included yet, please wait for the next stable release.\\n\\n- Cannot run / runtime errors: [ComfyUI/issues](https://github.com/comfyanonymous/ComfyUI/issues)\\n- UI / frontend issues: [ComfyUI_frontend/issues](https://github.com/Comfy-Org/ComfyUI_frontend/issues)\\n- Workflow issues: [workflow_templates/issues](https://github.com/Comfy-Org/workflow_templates/issues)\\n\"], \"color\": \"#222\", \"bgcolor\": \"#000\"}, {\"id\": 109, \"type\": \"Flux2Scheduler\", \"pos\": [444.57531947043185, 356.1096083006014], \"size\": [270, 106], \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"width\", \"type\": \"INT\", \"widget\": {\"name\": \"width\"}, \"link\": 201}, {\"name\": \"height\", \"type\": \"INT\", \"widget\": {\"name\": \"height\"}, \"link\": 202}], \"outputs\": [{\"name\": \"SIGMAS\", \"type\": \"SIGMAS\", \"links\": [185]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"Flux2Scheduler\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [4, 1024, 1536]}, {\"id\": 113, \"type\": \"EmptyFlux2LatentImage\", \"pos\": [1040.036060147257, 402.4920705620544], \"size\": [270, 106], \"flags\": {}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"width\", \"type\": \"INT\", \"widget\": {\"name\": \"width\"}, \"link\": 199}, {\"name\": \"height\", \"type\": \"INT\", \"widget\": {\"name\": \"height\"}, \"link\": 200}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [198]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"EmptyFlux2LatentImage\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [1024, 1024, 1]}, {\"id\": 114, \"type\": \"PrimitiveInt\", \"pos\": [494.59000806697486, -508.7012168025424], \"size\": [270, 82], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"INT\", \"type\": \"INT\", \"links\": [199, 201]}], \"title\": \"Width\", \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"PrimitiveInt\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [1024, \"fixed\"]}, {\"id\": 105, \"type\": \"CFGGuider\", \"pos\": [433.0836518325727, 110.76212705122477], \"size\": [270, 98], \"flags\": {}, \"order\": 14, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 204}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 195}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 196}], \"outputs\": [{\"name\": \"GUIDER\", \"type\": \"GUIDER\", \"links\": [183]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"CFGGuider\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [1]}, {\"id\": 107, \"type\": \"UNETLoader\", \"pos\": [-348.6323733305521, -232.84532892939507], \"size\": [364.42708333333337, 82], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [192]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"UNETLoader\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [\"flux-2-klein-9b.safetensors\", \"default\"]}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [834.9079794814677, -446.2581642417122], \"size\": [375, 405], \"flags\": {}, \"order\": 17, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 189}], \"outputs\": [], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"SaveImage\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [\"Flux2-Klein\"]}, {\"id\": 115, \"type\": \"PrimitiveInt\", \"pos\": [494.156336121289, -366.941480099023], \"size\": [270, 82], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"INT\", \"type\": \"INT\", \"links\": [200, 202]}], \"title\": \"Height\", \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"PrimitiveInt\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [1024, \"fixed\"]}, {\"id\": 108, \"type\": \"LoraLoaderModelOnly\", \"pos\": [71.00668339294343, -247.979305695621], \"size\": [270, 82], \"flags\": {}, \"order\": 9, \"mode\": 4, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 192}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [203]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.9.2\", \"Node name for S&R\": \"LoraLoaderModelOnly\"}, \"widgets_values\": [\"klein_m4n02_3K.safetensors\", 1]}, {\"id\": 106, \"type\": \"RandomNoise\", \"pos\": [444.53333469869284, -104.13110719389196], \"size\": [270, 82], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"NOISE\", \"type\": \"NOISE\", \"links\": [182]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"RandomNoise\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [9, \"fixed\"]}, {\"id\": 110, \"type\": \"CLIPTextEncode\", \"pos\": [257.3010312783703, 70.9723493087957], \"size\": [450, 380], \"flags\": {}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 188}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"slot_index\": 0, \"links\": [191, 195]}], \"title\": \"CLIP Text Encode (Positive Prompt)\", \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"CLIPTextEncode\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [\"A young woman with her hair in a bun working behind a busy coffee shop counter. She is pouring milk into a ceramic cup. Warm interior lighting, steam from the espresso machine, shelves of coffee beans in the background, realistic fabric texture on her green apron.\"], \"color\": \"#232\", \"bgcolor\": \"#353\"}, {\"id\": 116, \"type\": \"LoraLoaderModelOnly\", \"pos\": [399.17761636912036, -234.25210187034656], \"size\": [270, 82], \"flags\": {}, \"order\": 13, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 203}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [204]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.9.2\", \"Node name for S&R\": \"LoraLoaderModelOnly\"}, \"widgets_values\": [\"klein_slider_comic_000000050.safetensors\", 0]}], \"links\": [[182, 106, 0, 101, 0, \"NOISE\"], [183, 105, 0, 101, 1, \"GUIDER\"], [184, 100, 0, 101, 2, \"SAMPLER\"], [185, 109, 0, 101, 3, \"SIGMAS\"], [186, 101, 0, 102, 0, \"LATENT\"], [187, 104, 0, 102, 1, \"VAE\"], [188, 103, 0, 110, 0, \"CLIP\"], [189, 102, 0, 9, 0, \"IMAGE\"], [191, 110, 0, 111, 0, \"CONDITIONING\"], [192, 107, 0, 108, 0, \"MODEL\"], [195, 110, 0, 105, 1, \"CONDITIONING\"], [196, 111, 0, 105, 2, \"CONDITIONING\"], [198, 113, 0, 101, 4, \"LATENT\"], [199, 114, 0, 113, 0, \"INT\"], [200, 115, 0, 113, 1, \"INT\"], [201, 114, 0, 109, 0, \"INT\"], [202, 115, 0, 109, 1, \"INT\"], [203, 108, 0, 116, 0, \"MODEL\"], [204, 116, 0, 105, 0, \"MODEL\"]], \"groups\": [{\"id\": 1, \"title\": \"Models\", \"bounding\": [-209.3878920261178, -2.6595892865098936, 380, 550], \"color\": \"#3f789e\", \"font_size\": 24, \"flags\": {}}, {\"id\": 2, \"title\": \"Prompt\", \"bounding\": [227.30951520925015, 11.175419318941024, 470, 550], \"color\": \"#3f789e\", \"font_size\": 24, \"flags\": {}}, {\"id\": 3, \"title\": \"Sampler\", \"bounding\": [720.6121079738838, -2.6595892865098936, 532.3638671875, 550], \"color\": \"#3f789e\", \"font_size\": 24, \"flags\": {}}], \"config\": {}, \"extra\": {\"frontendVersion\": \"1.36.14\", \"workflowRendererVersion\": \"LG\", \"VHS_latentpreview\": false, \"VHS_latentpreviewrate\": 0, \"VHS_MetadataImage\": true, \"VHS_KeepIntermediate\": true, \"ds\": {\"scale\": 1.3974240692428272, \"offset\": [294.46602400850605, 460.54522199035114]}}, \"version\": 0.4}}",
242
+ "sampler": "Euler"
243
+ },
244
+ "availability": "Public",
245
+ "hasMeta": true,
246
+ "hasPositivePrompt": false,
247
+ "onSite": false,
248
+ "remixOfId": null
249
+ },
250
+ {
251
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/e35ed27e-3e12-4bcd-932e-ef711b5718e5/original=true/118500357.jpeg",
252
+ "nsfwLevel": 1,
253
+ "width": 2048,
254
+ "height": 2048,
255
+ "hash": "UYJPrTt7oe-6%foyt6jYtRW;s:j?~AayjZWA",
256
+ "type": "image",
257
+ "metadata": {
258
+ "hash": "UYJPrTt7oe-6%foyt6jYtRW;s:j?~AayjZWA",
259
+ "size": 6439871,
260
+ "width": 2048,
261
+ "height": 2048,
262
+ "nsfwLevelReason": "Knights Vote"
263
+ },
264
+ "minor": false,
265
+ "poi": false,
266
+ "meta": {
267
+ "seed": 9,
268
+ "vaes": [
269
+ "flux2-vae.safetensors"
270
+ ],
271
+ "comfy": "{\"prompt\": {\"9\": {\"inputs\": {\"filename_prefix\": \"Flux2-Klein\", \"images\": [\"102\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"100\": {\"inputs\": {\"sampler_name\": \"euler\"}, \"class_type\": \"KSamplerSelect\", \"_meta\": {\"title\": \"KSamplerSelect\"}}, \"101\": {\"inputs\": {\"noise\": [\"106\", 0], \"guider\": [\"105\", 0], \"sampler\": [\"100\", 0], \"sigmas\": [\"109\", 0], \"latent_image\": [\"113\", 0]}, \"class_type\": \"SamplerCustomAdvanced\", \"_meta\": {\"title\": \"SamplerCustomAdvanced\"}}, \"102\": {\"inputs\": {\"samples\": [\"101\", 0], \"vae\": [\"104\", 0]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"103\": {\"inputs\": {\"clip_name\": \"qwen_3_8b_fp8mixed.safetensors\", \"type\": \"flux2\", \"device\": \"default\"}, \"class_type\": \"CLIPLoader\", \"_meta\": {\"title\": \"Load CLIP\"}}, \"104\": {\"inputs\": {\"vae_name\": \"flux2-vae.safetensors\"}, \"class_type\": \"VAELoader\", \"_meta\": {\"title\": \"Load VAE\"}}, \"105\": {\"inputs\": {\"cfg\": 1.0, \"model\": [\"116\", 0], \"positive\": [\"110\", 0], \"negative\": [\"111\", 0]}, \"class_type\": \"CFGGuider\", \"_meta\": {\"title\": \"CFGGuider\"}}, \"106\": {\"inputs\": {\"noise_seed\": 9}, \"class_type\": \"RandomNoise\", \"_meta\": {\"title\": \"RandomNoise\"}}, \"107\": {\"inputs\": {\"unet_name\": \"flux-2-klein-9b.safetensors\", \"weight_dtype\": \"default\"}, \"class_type\": \"UNETLoader\", \"_meta\": {\"title\": \"Load Diffusion Model\"}}, \"109\": {\"inputs\": {\"steps\": 4, \"width\": [\"114\", 0], \"height\": [\"115\", 0]}, \"class_type\": \"Flux2Scheduler\", \"_meta\": {\"title\": \"Flux2Scheduler\"}}, \"110\": {\"inputs\": {\"text\": \"a man cutting onions in the kitchen, sunset, city street is visible through the window\", \"clip\": [\"103\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Positive Prompt)\"}}, \"111\": {\"inputs\": {\"conditioning\": [\"110\", 0]}, \"class_type\": \"ConditioningZeroOut\", \"_meta\": {\"title\": \"ConditioningZeroOut\"}}, \"113\": {\"inputs\": {\"width\": [\"114\", 0], \"height\": [\"115\", 0], \"batch_size\": 1}, \"class_type\": \"EmptyFlux2LatentImage\", \"_meta\": {\"title\": \"Empty Flux 2 Latent\"}}, \"114\": {\"inputs\": {\"value\": 1024}, \"class_type\": \"PrimitiveInt\", \"_meta\": {\"title\": \"Width\"}}, \"115\": {\"inputs\": {\"value\": 1024}, \"class_type\": \"PrimitiveInt\", \"_meta\": {\"title\": \"Height\"}}, \"116\": {\"inputs\": {\"lora_name\": \"klein_slider_comic_000000050.safetensors\", \"strength_model\": 0.0, \"model\": [\"107\", 0]}, \"class_type\": \"LoraLoaderModelOnly\", \"_meta\": {\"title\": \"LoraLoaderModelOnly\"}}}, \"workflow\": {\"id\": \"92112d97-bb64-4b44-86f2-ea5691ef8f6e\", \"revision\": 0, \"last_node_id\": 120, \"last_link_id\": 208, \"nodes\": [{\"id\": 100, \"type\": \"KSamplerSelect\", \"pos\": [433.0836518325727, 250.7621681845402], \"size\": [270, 58], \"flags\": {}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"SAMPLER\", \"type\": \"SAMPLER\", \"links\": [184]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"KSamplerSelect\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [\"euler\"]}, {\"id\": 101, \"type\": \"SamplerCustomAdvanced\", \"pos\": [1020.278813335458, 101.36819395749974], \"size\": [213.275, 106], \"flags\": {}, \"order\": 15, \"mode\": 0, \"inputs\": [{\"name\": \"noise\", \"type\": \"NOISE\", \"link\": 182}, {\"name\": \"guider\", \"type\": \"GUIDER\", \"link\": 183}, {\"name\": \"sampler\", \"type\": \"SAMPLER\", \"link\": 184}, {\"name\": \"sigmas\", \"type\": \"SIGMAS\", \"link\": 185}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 198}], \"outputs\": [{\"name\": \"output\", \"type\": \"LATENT\", \"links\": [186]}, {\"name\": \"denoised_output\", \"type\": \"LATENT\", \"links\": []}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"SamplerCustomAdvanced\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": []}, {\"id\": 102, \"type\": \"VAEDecode\", \"pos\": [1274.2787087340078, 31.368183492318206], \"size\": [220, 46], \"flags\": {}, \"order\": 16, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 186}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 187}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"slot_index\": 0, \"links\": [189]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"VAEDecode\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": []}, {\"id\": 103, \"type\": \"CLIPLoader\", \"pos\": [-205.72122391625783, 231.2015448252655], \"size\": [364.42708333333337, 106], \"flags\": {}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [188]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"CLIPLoader\", \"models\": [{\"name\": \"qwen_3_8b_fp8mixed.safetensors\", \"url\": \"https://huggingface.co/Comfy-Org/flux2-klein-9B/resolve/main/split_files/text_encoders/qwen_3_8b_fp8mixed.safetensors\", \"directory\": \"text_encoders\"}], \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [\"qwen_3_8b_fp8mixed.safetensors\", \"flux2\", \"default\"]}, {\"id\": 104, \"type\": \"VAELoader\", \"pos\": [-205.72122391625783, 401.201573126844], \"size\": [364.42708333333337, 58], \"flags\": {}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [187]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"VAELoader\", \"models\": [{\"name\": \"flux2-vae.safetensors\", \"url\": \"https://huggingface.co/Comfy-Org/flux2-dev/resolve/main/split_files/vae/flux2-vae.safetensors\", \"directory\": \"vae\"}], \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [\"flux2-vae.safetensors\"]}, {\"id\": 111, \"type\": \"ConditioningZeroOut\", \"pos\": [308.27845799443645, 510.418847617246], \"size\": [204.1750030517578, 26], \"flags\": {}, \"order\": 12, \"mode\": 0, \"inputs\": [{\"name\": \"conditioning\", \"type\": \"CONDITIONING\", \"link\": 191}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [196]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.9.1\", \"Node name for S&R\": \"ConditioningZeroOut\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": []}, {\"id\": 97, \"type\": \"MarkdownNote\", \"pos\": [-600, 810], \"size\": [480, 610], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [], \"outputs\": [], \"properties\": {}, \"widgets_values\": [\"Guide: [Subgraph](https://docs.comfy.org/interface/features/subgraph)\\n\\n## Model links (for local users)\\n\\n\\n**diffusion_models**\\n\\n- [flux-2-klein-9b-fp8.safetensors](https://huggingface.co/black-forest-labs/FLUX.2-klein-9b-fp8)\\n\\n> Please visit BFL's repo, accept the agreement in the repo, and then download the models.\\n\\n**text_encoders**\\n\\n- [qwen_3_8b_fp8mixed.safetensors](https://huggingface.co/Comfy-Org/flux2-klein-9B/resolve/main/split_files/text_encoders/qwen_3_8b_fp8mixed.safetensors)\\n\\n**vae**\\n\\n- [flux2-vae.safetensors](https://huggingface.co/Comfy-Org/flux2-dev/resolve/main/split_files/vae/flux2-vae.safetensors)\\n\\n\\nModel Storage Location\\n\\n```\\n\\ud83d\\udcc2 ComfyUI/\\n\\u251c\\u2500\\u2500 \\ud83d\\udcc2 models/\\n\\u2502 \\u251c\\u2500\\u2500 \\ud83d\\udcc2 diffusion_models/\\n\\u2502 \\u2502 \\u2514\\u2500\\u2500 flux-2-klein-9b-fp8.safetensors\\n\\u2502 \\u251c\\u2500\\u2500 \\ud83d\\udcc2 text_encoders/\\n\\u2502 \\u2502 \\u2514\\u2500\\u2500 qwen_3_8b_fp8mixed.safetensors\\n\\u2502 \\u2514\\u2500\\u2500 \\ud83d\\udcc2 vae/\\n\\u2502 \\u2514\\u2500\\u2500 flux2-vae.safetensors\\n```\\n## Report issue\\n\\nNote: please update ComfyUI first ([guide](https://docs.comfy.org/zh-CN/installation/update_comfyui)) and prepare required models. Desktop/Cloud ship stable builds; nightly-supported models may not be included yet, please wait for the next stable release.\\n\\n- Cannot run / runtime errors: [ComfyUI/issues](https://github.com/comfyanonymous/ComfyUI/issues)\\n- UI / frontend issues: [ComfyUI_frontend/issues](https://github.com/Comfy-Org/ComfyUI_frontend/issues)\\n- Workflow issues: [workflow_templates/issues](https://github.com/Comfy-Org/workflow_templates/issues)\\n\"], \"color\": \"#222\", \"bgcolor\": \"#000\"}, {\"id\": 109, \"type\": \"Flux2Scheduler\", \"pos\": [444.57531947043185, 356.1096083006014], \"size\": [270, 106], \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"width\", \"type\": \"INT\", \"widget\": {\"name\": \"width\"}, \"link\": 201}, {\"name\": \"height\", \"type\": \"INT\", \"widget\": {\"name\": \"height\"}, \"link\": 202}], \"outputs\": [{\"name\": \"SIGMAS\", \"type\": \"SIGMAS\", \"links\": [185]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"Flux2Scheduler\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [4, 1024, 1536]}, {\"id\": 113, \"type\": \"EmptyFlux2LatentImage\", \"pos\": [1040.036060147257, 402.4920705620544], \"size\": [270, 106], \"flags\": {}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"width\", \"type\": \"INT\", \"widget\": {\"name\": \"width\"}, \"link\": 199}, {\"name\": \"height\", \"type\": \"INT\", \"widget\": {\"name\": \"height\"}, \"link\": 200}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [198]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"EmptyFlux2LatentImage\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [1024, 1024, 1]}, {\"id\": 114, \"type\": \"PrimitiveInt\", \"pos\": [494.59000806697486, -508.7012168025424], \"size\": [270, 82], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"INT\", \"type\": \"INT\", \"links\": [199, 201]}], \"title\": \"Width\", \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"PrimitiveInt\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [1024, \"fixed\"]}, {\"id\": 105, \"type\": \"CFGGuider\", \"pos\": [433.0836518325727, 110.76212705122477], \"size\": [270, 98], \"flags\": {}, \"order\": 14, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 204}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 195}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 196}], \"outputs\": [{\"name\": \"GUIDER\", \"type\": \"GUIDER\", \"links\": [183]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"CFGGuider\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [1]}, {\"id\": 107, \"type\": \"UNETLoader\", \"pos\": [-348.6323733305521, -232.84532892939507], \"size\": [364.42708333333337, 82], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [192]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"UNETLoader\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [\"flux-2-klein-9b.safetensors\", \"default\"]}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [834.9079794814677, -446.2581642417122], \"size\": [375, 405], \"flags\": {}, \"order\": 17, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 189}], \"outputs\": [], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"SaveImage\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [\"Flux2-Klein\"]}, {\"id\": 115, \"type\": \"PrimitiveInt\", \"pos\": [494.156336121289, -366.941480099023], \"size\": [270, 82], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"INT\", \"type\": \"INT\", \"links\": [200, 202]}], \"title\": \"Height\", \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"PrimitiveInt\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [1024, \"fixed\"]}, {\"id\": 108, \"type\": \"LoraLoaderModelOnly\", \"pos\": [71.00668339294343, -247.979305695621], \"size\": [270, 82], \"flags\": {}, \"order\": 9, \"mode\": 4, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 192}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [203]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.9.2\", \"Node name for S&R\": \"LoraLoaderModelOnly\"}, \"widgets_values\": [\"klein_m4n02_3K.safetensors\", 1]}, {\"id\": 106, \"type\": \"RandomNoise\", \"pos\": [444.53333469869284, -104.13110719389196], \"size\": [270, 82], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"NOISE\", \"type\": \"NOISE\", \"links\": [182]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"RandomNoise\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [9, \"fixed\"]}, {\"id\": 110, \"type\": \"CLIPTextEncode\", \"pos\": [257.3010312783703, 70.9723493087957], \"size\": [450, 380], \"flags\": {}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 188}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"slot_index\": 0, \"links\": [191, 195]}], \"title\": \"CLIP Text Encode (Positive Prompt)\", \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"CLIPTextEncode\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [\"a man cutting onions in the kitchen, sunset, city street is visible through the window\"], \"color\": \"#232\", \"bgcolor\": \"#353\"}, {\"id\": 116, \"type\": \"LoraLoaderModelOnly\", \"pos\": [399.17761636912036, -234.25210187034656], \"size\": [270, 82], \"flags\": {}, \"order\": 13, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 203}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [204]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.9.2\", \"Node name for S&R\": \"LoraLoaderModelOnly\"}, \"widgets_values\": [\"klein_slider_comic_000000050.safetensors\", 0]}], \"links\": [[182, 106, 0, 101, 0, \"NOISE\"], [183, 105, 0, 101, 1, \"GUIDER\"], [184, 100, 0, 101, 2, \"SAMPLER\"], [185, 109, 0, 101, 3, \"SIGMAS\"], [186, 101, 0, 102, 0, \"LATENT\"], [187, 104, 0, 102, 1, \"VAE\"], [188, 103, 0, 110, 0, \"CLIP\"], [189, 102, 0, 9, 0, \"IMAGE\"], [191, 110, 0, 111, 0, \"CONDITIONING\"], [192, 107, 0, 108, 0, \"MODEL\"], [195, 110, 0, 105, 1, \"CONDITIONING\"], [196, 111, 0, 105, 2, \"CONDITIONING\"], [198, 113, 0, 101, 4, \"LATENT\"], [199, 114, 0, 113, 0, \"INT\"], [200, 115, 0, 113, 1, \"INT\"], [201, 114, 0, 109, 0, \"INT\"], [202, 115, 0, 109, 1, \"INT\"], [203, 108, 0, 116, 0, \"MODEL\"], [204, 116, 0, 105, 0, \"MODEL\"]], \"groups\": [{\"id\": 1, \"title\": \"Models\", \"bounding\": [-209.3878920261178, -2.6595892865098936, 380, 550], \"color\": \"#3f789e\", \"font_size\": 24, \"flags\": {}}, {\"id\": 2, \"title\": \"Prompt\", \"bounding\": [227.30951520925015, 11.175419318941024, 470, 550], \"color\": \"#3f789e\", \"font_size\": 24, \"flags\": {}}, {\"id\": 3, \"title\": \"Sampler\", \"bounding\": [720.6121079738838, -2.6595892865098936, 532.3638671875, 550], \"color\": \"#3f789e\", \"font_size\": 24, \"flags\": {}}], \"config\": {}, \"extra\": {\"frontendVersion\": \"1.36.14\", \"workflowRendererVersion\": \"LG\", \"VHS_latentpreview\": false, \"VHS_latentpreviewrate\": 0, \"VHS_MetadataImage\": true, \"VHS_KeepIntermediate\": true, \"ds\": {\"scale\": 1.3974240692428281, \"offset\": [330.90624035389277, 618.046264074153]}}, \"version\": 0.4}}",
272
+ "sampler": "Euler"
273
+ },
274
+ "availability": "Public",
275
+ "hasMeta": true,
276
+ "hasPositivePrompt": false,
277
+ "onSite": false,
278
+ "remixOfId": null
279
+ },
280
+ {
281
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/378f32a7-dded-43cf-9457-b88a222087c7/original=true/118502437.jpeg",
282
+ "nsfwLevel": 1,
283
+ "width": 2048,
284
+ "height": 1024,
285
+ "hash": "UHKw,]XmkWxt~Bj[R*sS_NIVRjxC?baxWVog",
286
+ "type": "image",
287
+ "metadata": {
288
+ "hash": "UHKw,]XmkWxt~Bj[R*sS_NIVRjxC?baxWVog",
289
+ "size": 3328111,
290
+ "width": 2048,
291
+ "height": 1024,
292
+ "nsfwLevelReason": "Knights Vote"
293
+ },
294
+ "minor": false,
295
+ "poi": false,
296
+ "meta": {
297
+ "seed": 9,
298
+ "vaes": [
299
+ "flux2-vae.safetensors"
300
+ ],
301
+ "comfy": "{\"prompt\": {\"9\": {\"inputs\": {\"filename_prefix\": \"Flux2-Klein\", \"images\": [\"102\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"100\": {\"inputs\": {\"sampler_name\": \"euler\"}, \"class_type\": \"KSamplerSelect\", \"_meta\": {\"title\": \"KSamplerSelect\"}}, \"101\": {\"inputs\": {\"noise\": [\"106\", 0], \"guider\": [\"105\", 0], \"sampler\": [\"100\", 0], \"sigmas\": [\"109\", 0], \"latent_image\": [\"113\", 0]}, \"class_type\": \"SamplerCustomAdvanced\", \"_meta\": {\"title\": \"SamplerCustomAdvanced\"}}, \"102\": {\"inputs\": {\"samples\": [\"101\", 0], \"vae\": [\"104\", 0]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"103\": {\"inputs\": {\"clip_name\": \"qwen_3_8b_fp8mixed.safetensors\", \"type\": \"flux2\", \"device\": \"default\"}, \"class_type\": \"CLIPLoader\", \"_meta\": {\"title\": \"Load CLIP\"}}, \"104\": {\"inputs\": {\"vae_name\": \"flux2-vae.safetensors\"}, \"class_type\": \"VAELoader\", \"_meta\": {\"title\": \"Load VAE\"}}, \"105\": {\"inputs\": {\"cfg\": 1.0, \"model\": [\"107\", 0], \"positive\": [\"110\", 0], \"negative\": [\"111\", 0]}, \"class_type\": \"CFGGuider\", \"_meta\": {\"title\": \"CFGGuider\"}}, \"106\": {\"inputs\": {\"noise_seed\": 9}, \"class_type\": \"RandomNoise\", \"_meta\": {\"title\": \"RandomNoise\"}}, \"107\": {\"inputs\": {\"unet_name\": \"flux-2-klein-9b.safetensors\", \"weight_dtype\": \"default\"}, \"class_type\": \"UNETLoader\", \"_meta\": {\"title\": \"Load Diffusion Model\"}}, \"109\": {\"inputs\": {\"steps\": 4, \"width\": [\"114\", 0], \"height\": [\"115\", 0]}, \"class_type\": \"Flux2Scheduler\", \"_meta\": {\"title\": \"Flux2Scheduler\"}}, \"110\": {\"inputs\": {\"text\": \"an imaginary superhero woman wearing a white costume walking on a street, comic style\", \"clip\": [\"103\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Positive Prompt)\"}}, \"111\": {\"inputs\": {\"conditioning\": [\"110\", 0]}, \"class_type\": \"ConditioningZeroOut\", \"_meta\": {\"title\": \"ConditioningZeroOut\"}}, \"113\": {\"inputs\": {\"width\": [\"114\", 0], \"height\": [\"115\", 0], \"batch_size\": 1}, \"class_type\": \"EmptyFlux2LatentImage\", \"_meta\": {\"title\": \"Empty Flux 2 Latent\"}}, \"114\": {\"inputs\": {\"value\": 1024}, \"class_type\": \"PrimitiveInt\", \"_meta\": {\"title\": \"Width\"}}, \"115\": {\"inputs\": {\"value\": 1024}, \"class_type\": \"PrimitiveInt\", \"_meta\": {\"title\": \"Height\"}}}, \"workflow\": {\"id\": \"92112d97-bb64-4b44-86f2-ea5691ef8f6e\", \"revision\": 0, \"last_node_id\": 120, \"last_link_id\": 208, \"nodes\": [{\"id\": 100, \"type\": \"KSamplerSelect\", \"pos\": [433.0836518325727, 250.7621681845402], \"size\": [270, 58], \"flags\": {}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"SAMPLER\", \"type\": \"SAMPLER\", \"links\": [184]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"KSamplerSelect\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [\"euler\"]}, {\"id\": 101, \"type\": \"SamplerCustomAdvanced\", \"pos\": [1020.278813335458, 101.36819395749974], \"size\": [213.275, 106], \"flags\": {}, \"order\": 15, \"mode\": 0, \"inputs\": [{\"name\": \"noise\", \"type\": \"NOISE\", \"link\": 182}, {\"name\": \"guider\", \"type\": \"GUIDER\", \"link\": 183}, {\"name\": \"sampler\", \"type\": \"SAMPLER\", \"link\": 184}, {\"name\": \"sigmas\", \"type\": \"SIGMAS\", \"link\": 185}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 198}], \"outputs\": [{\"name\": \"output\", \"type\": \"LATENT\", \"links\": [186]}, {\"name\": \"denoised_output\", \"type\": \"LATENT\", \"links\": []}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"SamplerCustomAdvanced\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": []}, {\"id\": 102, \"type\": \"VAEDecode\", \"pos\": [1274.2787087340078, 31.368183492318206], \"size\": [220, 46], \"flags\": {}, \"order\": 16, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 186}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 187}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"slot_index\": 0, \"links\": [189]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"VAEDecode\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": []}, {\"id\": 103, \"type\": \"CLIPLoader\", \"pos\": [-205.72122391625783, 231.2015448252655], \"size\": [364.42708333333337, 106], \"flags\": {}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [188]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"CLIPLoader\", \"models\": [{\"name\": \"qwen_3_8b_fp8mixed.safetensors\", \"url\": \"https://huggingface.co/Comfy-Org/flux2-klein-9B/resolve/main/split_files/text_encoders/qwen_3_8b_fp8mixed.safetensors\", \"directory\": \"text_encoders\"}], \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [\"qwen_3_8b_fp8mixed.safetensors\", \"flux2\", \"default\"]}, {\"id\": 104, \"type\": \"VAELoader\", \"pos\": [-205.72122391625783, 401.201573126844], \"size\": [364.42708333333337, 58], \"flags\": {}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [187]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"VAELoader\", \"models\": [{\"name\": \"flux2-vae.safetensors\", \"url\": \"https://huggingface.co/Comfy-Org/flux2-dev/resolve/main/split_files/vae/flux2-vae.safetensors\", \"directory\": \"vae\"}], \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [\"flux2-vae.safetensors\"]}, {\"id\": 111, \"type\": \"ConditioningZeroOut\", \"pos\": [308.27845799443645, 510.418847617246], \"size\": [204.1750030517578, 26], \"flags\": {}, \"order\": 12, \"mode\": 0, \"inputs\": [{\"name\": \"conditioning\", \"type\": \"CONDITIONING\", \"link\": 191}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [196]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.9.1\", \"Node name for S&R\": \"ConditioningZeroOut\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": []}, {\"id\": 97, \"type\": \"MarkdownNote\", \"pos\": [-600, 810], \"size\": [480, 610], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [], \"outputs\": [], \"properties\": {}, \"widgets_values\": [\"Guide: [Subgraph](https://docs.comfy.org/interface/features/subgraph)\\n\\n## Model links (for local users)\\n\\n\\n**diffusion_models**\\n\\n- [flux-2-klein-9b-fp8.safetensors](https://huggingface.co/black-forest-labs/FLUX.2-klein-9b-fp8)\\n\\n> Please visit BFL's repo, accept the agreement in the repo, and then download the models.\\n\\n**text_encoders**\\n\\n- [qwen_3_8b_fp8mixed.safetensors](https://huggingface.co/Comfy-Org/flux2-klein-9B/resolve/main/split_files/text_encoders/qwen_3_8b_fp8mixed.safetensors)\\n\\n**vae**\\n\\n- [flux2-vae.safetensors](https://huggingface.co/Comfy-Org/flux2-dev/resolve/main/split_files/vae/flux2-vae.safetensors)\\n\\n\\nModel Storage Location\\n\\n```\\n\\ud83d\\udcc2 ComfyUI/\\n\\u251c\\u2500\\u2500 \\ud83d\\udcc2 models/\\n\\u2502 \\u251c\\u2500\\u2500 \\ud83d\\udcc2 diffusion_models/\\n\\u2502 \\u2502 \\u2514\\u2500\\u2500 flux-2-klein-9b-fp8.safetensors\\n\\u2502 \\u251c\\u2500\\u2500 \\ud83d\\udcc2 text_encoders/\\n\\u2502 \\u2502 \\u2514\\u2500\\u2500 qwen_3_8b_fp8mixed.safetensors\\n\\u2502 \\u2514\\u2500\\u2500 \\ud83d\\udcc2 vae/\\n\\u2502 \\u2514\\u2500\\u2500 flux2-vae.safetensors\\n```\\n## Report issue\\n\\nNote: please update ComfyUI first ([guide](https://docs.comfy.org/zh-CN/installation/update_comfyui)) and prepare required models. Desktop/Cloud ship stable builds; nightly-supported models may not be included yet, please wait for the next stable release.\\n\\n- Cannot run / runtime errors: [ComfyUI/issues](https://github.com/comfyanonymous/ComfyUI/issues)\\n- UI / frontend issues: [ComfyUI_frontend/issues](https://github.com/Comfy-Org/ComfyUI_frontend/issues)\\n- Workflow issues: [workflow_templates/issues](https://github.com/Comfy-Org/workflow_templates/issues)\\n\"], \"color\": \"#222\", \"bgcolor\": \"#000\"}, {\"id\": 109, \"type\": \"Flux2Scheduler\", \"pos\": [444.57531947043185, 356.1096083006014], \"size\": [270, 106], \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"width\", \"type\": \"INT\", \"widget\": {\"name\": \"width\"}, \"link\": 201}, {\"name\": \"height\", \"type\": \"INT\", \"widget\": {\"name\": \"height\"}, \"link\": 202}], \"outputs\": [{\"name\": \"SIGMAS\", \"type\": \"SIGMAS\", \"links\": [185]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"Flux2Scheduler\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [4, 1024, 1536]}, {\"id\": 113, \"type\": \"EmptyFlux2LatentImage\", \"pos\": [1040.036060147257, 402.4920705620544], \"size\": [270, 106], \"flags\": {}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"width\", \"type\": \"INT\", \"widget\": {\"name\": \"width\"}, \"link\": 199}, {\"name\": \"height\", \"type\": \"INT\", \"widget\": {\"name\": \"height\"}, \"link\": 200}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [198]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"EmptyFlux2LatentImage\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [1024, 1024, 1]}, {\"id\": 114, \"type\": \"PrimitiveInt\", \"pos\": [494.59000806697486, -508.7012168025424], \"size\": [270, 82], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"INT\", \"type\": \"INT\", \"links\": [199, 201]}], \"title\": \"Width\", \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"PrimitiveInt\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [1024, \"fixed\"]}, {\"id\": 105, \"type\": \"CFGGuider\", \"pos\": [433.0836518325727, 110.76212705122477], \"size\": [270, 98], \"flags\": {}, \"order\": 14, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 204}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 195}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 196}], \"outputs\": [{\"name\": \"GUIDER\", \"type\": \"GUIDER\", \"links\": [183]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"CFGGuider\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [1]}, {\"id\": 107, \"type\": \"UNETLoader\", \"pos\": [-348.6323733305521, -232.84532892939507], \"size\": [364.42708333333337, 82], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [192]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"UNETLoader\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [\"flux-2-klein-9b.safetensors\", \"default\"]}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [834.9079794814677, -446.2581642417122], \"size\": [375, 405], \"flags\": {}, \"order\": 17, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 189}], \"outputs\": [], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"SaveImage\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [\"Flux2-Klein\"]}, {\"id\": 115, \"type\": \"PrimitiveInt\", \"pos\": [494.156336121289, -366.941480099023], \"size\": [270, 82], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"INT\", \"type\": \"INT\", \"links\": [200, 202]}], \"title\": \"Height\", \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"PrimitiveInt\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [1024, \"fixed\"]}, {\"id\": 108, \"type\": \"LoraLoaderModelOnly\", \"pos\": [71.00668339294343, -247.979305695621], \"size\": [270, 82], \"flags\": {}, \"order\": 9, \"mode\": 4, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 192}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [203]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.9.2\", \"Node name for S&R\": \"LoraLoaderModelOnly\"}, \"widgets_values\": [\"klein_m4n02_3K.safetensors\", 1]}, {\"id\": 116, \"type\": \"LoraLoaderModelOnly\", \"pos\": [399.17761636912036, -234.25210187034656], \"size\": [270, 82], \"flags\": {}, \"order\": 13, \"mode\": 4, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 203}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [204]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.9.2\", \"Node name for S&R\": \"LoraLoaderModelOnly\"}, \"widgets_values\": [\"klein_slider_comic_000000050.safetensors\", 0]}, {\"id\": 110, \"type\": \"CLIPTextEncode\", \"pos\": [257.3010312783703, 70.9723493087957], \"size\": [450, 380], \"flags\": {}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 188}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"slot_index\": 0, \"links\": [191, 195]}], \"title\": \"CLIP Text Encode (Positive Prompt)\", \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"CLIPTextEncode\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [\"an imaginary superhero woman wearing a white costume walking on a street, comic style\"], \"color\": \"#232\", \"bgcolor\": \"#353\"}, {\"id\": 106, \"type\": \"RandomNoise\", \"pos\": [444.53333469869284, -104.13110719389196], \"size\": [270, 82], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"NOISE\", \"type\": \"NOISE\", \"links\": [182]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.8.2\", \"Node name for S&R\": \"RandomNoise\", \"enableTabs\": false, \"tabWidth\": 65, \"tabXOffset\": 10, \"hasSecondTab\": false, \"secondTabText\": \"Send Back\", \"secondTabOffset\": 80, \"secondTabWidth\": 65}, \"widgets_values\": [9, \"fixed\"]}], \"links\": [[182, 106, 0, 101, 0, \"NOISE\"], [183, 105, 0, 101, 1, \"GUIDER\"], [184, 100, 0, 101, 2, \"SAMPLER\"], [185, 109, 0, 101, 3, \"SIGMAS\"], [186, 101, 0, 102, 0, \"LATENT\"], [187, 104, 0, 102, 1, \"VAE\"], [188, 103, 0, 110, 0, \"CLIP\"], [189, 102, 0, 9, 0, \"IMAGE\"], [191, 110, 0, 111, 0, \"CONDITIONING\"], [192, 107, 0, 108, 0, \"MODEL\"], [195, 110, 0, 105, 1, \"CONDITIONING\"], [196, 111, 0, 105, 2, \"CONDITIONING\"], [198, 113, 0, 101, 4, \"LATENT\"], [199, 114, 0, 113, 0, \"INT\"], [200, 115, 0, 113, 1, \"INT\"], [201, 114, 0, 109, 0, \"INT\"], [202, 115, 0, 109, 1, \"INT\"], [203, 108, 0, 116, 0, \"MODEL\"], [204, 116, 0, 105, 0, \"MODEL\"]], \"groups\": [{\"id\": 1, \"title\": \"Models\", \"bounding\": [-209.3878920261178, -2.6595892865098936, 380, 550], \"color\": \"#3f789e\", \"font_size\": 24, \"flags\": {}}, {\"id\": 2, \"title\": \"Prompt\", \"bounding\": [227.30951520925015, 11.175419318941024, 470, 550], \"color\": \"#3f789e\", \"font_size\": 24, \"flags\": {}}, {\"id\": 3, \"title\": \"Sampler\", \"bounding\": [720.6121079738838, -2.6595892865098936, 532.3638671875, 550], \"color\": \"#3f789e\", \"font_size\": 24, \"flags\": {}}], \"config\": {}, \"extra\": {\"frontendVersion\": \"1.36.14\", \"workflowRendererVersion\": \"LG\", \"VHS_latentpreview\": false, \"VHS_latentpreviewrate\": 0, \"VHS_MetadataImage\": true, \"VHS_KeepIntermediate\": true, \"ds\": {\"scale\": 1.3974240692428308, \"offset\": [9.652408525107944, 584.169183778874]}}, \"version\": 0.4}}",
302
+ "sampler": "Euler"
303
+ },
304
+ "availability": "Public",
305
+ "hasMeta": true,
306
+ "hasPositivePrompt": false,
307
+ "onSite": false,
308
+ "remixOfId": null
309
+ }
310
+ ],
311
+ "downloadUrl": "https://civitai.com/api/download/models/2623808",
312
+ "_sha256": "b790414cac216841cfcafccace9d4361da991267f0fdb62834facc98fb4b29b7",
313
+ "_civitai_api": "https://civitai.com/api/v1/model-versions/by-hash/b790414cac216841cfcafccace9d4361da991267f0fdb62834facc98fb4b29b7"
314
+ }
315
+ },
316
+ "trainedWords": [],
317
+ "sha256": "b790414cac216841cfcafccace9d4361da991267f0fdb62834facc98fb4b29b7",
318
+ "name": "Klein Realistic / Comic Slider - Klein 9B",
319
+ "type": "LORA",
320
+ "baseModel": "Flux.2 Klein 9B",
321
+ "links": [
322
+ "https://civitai.com/models/2332534?modelVersionId=2623808",
323
+ "https://civitai.com/api/v1/model-versions/by-hash/b790414cac216841cfcafccace9d4361da991267f0fdb62834facc98fb4b29b7"
324
+ ]
325
+ }
loras/klein_snofs_v1_2.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae6b2f510212741a25c946500e349aca47063199f107413b80ead017af5b7d84
3
+ size 1090563760
loras/uncrop_F2K9B.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3fe90629390454f4b18a6cbea8918e8e36223781996a78d35b4c18db8731e9db
3
+ size 165704384
loras/uncrop_F2K9B.safetensors.rgthree-info.json ADDED
@@ -0,0 +1,450 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "file": "flux-klein\\uncrop_F2K9B.safetensors",
3
+ "path": "E:\\AI_Softwares\\ImageGenerator\\ComfyUI_windows_portable\\ComfyUI\\models\\loras\\flux-klein\\uncrop_F2K9B.safetensors",
4
+ "modified": 1771930633536.0254,
5
+ "images": [
6
+ {
7
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/ac9009e9-cfcf-4443-af2c-0ae36d5762c7/original=true/118388616.mp4",
8
+ "civitaiUrl": "https://civitai.com/images/118388616",
9
+ "width": 832,
10
+ "height": 1248,
11
+ "type": "video",
12
+ "nsfwLevel": 1,
13
+ "seed": 43,
14
+ "positive": "remove the white parts, use the image for context. add a cowboy hat",
15
+ "negative": null,
16
+ "steps": 6,
17
+ "sampler": null,
18
+ "cfg": null,
19
+ "model": null,
20
+ "resources": null
21
+ },
22
+ {
23
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/a0fc4429-2466-429a-a46c-397d60dd6916/original=true/118381644.jpeg",
24
+ "civitaiUrl": "https://civitai.com/images/118381644",
25
+ "width": 853,
26
+ "height": 1280,
27
+ "type": "image",
28
+ "nsfwLevel": 4,
29
+ "seed": 44,
30
+ "positive": "Remove the white parts. Use the image for context. She is looking to the left",
31
+ "negative": null,
32
+ "steps": 6,
33
+ "sampler": null,
34
+ "cfg": null,
35
+ "model": null,
36
+ "resources": null
37
+ },
38
+ {
39
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/76cc2dee-4702-48c9-b52f-7a1edc7f6950/original=true/118382203.jpeg",
40
+ "civitaiUrl": "https://civitai.com/images/118382203",
41
+ "width": 853,
42
+ "height": 1280,
43
+ "type": "image",
44
+ "nsfwLevel": 4,
45
+ "seed": 42,
46
+ "positive": "Remove the white parts, use the butt for context with matching skin color",
47
+ "negative": null,
48
+ "steps": 6,
49
+ "sampler": null,
50
+ "cfg": null,
51
+ "model": null,
52
+ "resources": null
53
+ },
54
+ {
55
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/8814fbb4-6f52-4e1d-93dc-7622fedd5b17/original=true/118377591.jpeg",
56
+ "civitaiUrl": "https://civitai.com/images/118377591",
57
+ "width": 853,
58
+ "height": 1280,
59
+ "type": "image",
60
+ "nsfwLevel": 4,
61
+ "seed": 43,
62
+ "positive": "Remove the white parts, use the image for context",
63
+ "negative": null,
64
+ "steps": 4,
65
+ "sampler": null,
66
+ "cfg": null,
67
+ "model": null,
68
+ "resources": null
69
+ },
70
+ {
71
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/403915be-05e9-4544-9398-0661be289af3/original=true/118378048.jpeg",
72
+ "civitaiUrl": "https://civitai.com/images/118378048",
73
+ "width": 853,
74
+ "height": 1280,
75
+ "type": "image",
76
+ "nsfwLevel": 4,
77
+ "seed": 43,
78
+ "positive": "Remove the white parts. Use the image for context.",
79
+ "negative": null,
80
+ "steps": 4,
81
+ "sampler": null,
82
+ "cfg": null,
83
+ "model": null,
84
+ "resources": null
85
+ },
86
+ {
87
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/74c28b74-fa85-48fb-bbea-06bb298cb420/original=true/118381990.jpeg",
88
+ "civitaiUrl": "https://civitai.com/images/118381990",
89
+ "width": 853,
90
+ "height": 1280,
91
+ "type": "image",
92
+ "nsfwLevel": 4,
93
+ "seed": 44,
94
+ "positive": "Remove the white parts. Use the image for context.",
95
+ "negative": null,
96
+ "steps": 4,
97
+ "sampler": null,
98
+ "cfg": null,
99
+ "model": null,
100
+ "resources": null
101
+ },
102
+ {
103
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/ea6b9b04-aaa2-42d1-9421-8d7cc9e059da/original=true/118382581.jpeg",
104
+ "civitaiUrl": "https://civitai.com/images/118382581",
105
+ "width": 853,
106
+ "height": 1280,
107
+ "type": "image",
108
+ "nsfwLevel": 4,
109
+ "seed": 42,
110
+ "positive": "Remove the bikini in image 1, use the dress in image 2 for context. preserve her body shape",
111
+ "negative": null,
112
+ "steps": 5,
113
+ "sampler": null,
114
+ "cfg": null,
115
+ "model": null,
116
+ "resources": null
117
+ },
118
+ {
119
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/cb93c16f-0d58-4aa7-b1d5-03cea43dd6de/original=true/118383163.jpeg",
120
+ "civitaiUrl": "https://civitai.com/images/118383163",
121
+ "width": 853,
122
+ "height": 1280,
123
+ "type": "image",
124
+ "nsfwLevel": 4,
125
+ "seed": 42,
126
+ "positive": "remove the white parts, use the bikini in image 2 for context",
127
+ "negative": null,
128
+ "steps": 6,
129
+ "sampler": null,
130
+ "cfg": null,
131
+ "model": null,
132
+ "resources": null
133
+ }
134
+ ],
135
+ "raw": {
136
+ "metadata": {
137
+ "version": "1.0",
138
+ "ss_output_name": "uncrop_F2K9B",
139
+ "format": "pt",
140
+ "training_info": {
141
+ "step": 1500,
142
+ "epoch": 16
143
+ },
144
+ "name": "uncrop_F2K9B",
145
+ "software": {
146
+ "name": "ai-toolkit",
147
+ "repo": "https://github.com/ostris/ai-toolkit",
148
+ "version": "0.7.20"
149
+ },
150
+ "sshs_model_hash": "a8e56a5afddbd755ec397989c3d2ec4081da7dae4a510485b91c0d1e59a5f511",
151
+ "sshs_legacy_hash": "6fdedec8",
152
+ "ss_base_model_version": "flux2_klein_9b",
153
+ "_sha256": "3fe90629390454f4b18a6cbea8918e8e36223781996a78d35b4c18db8731e9db"
154
+ },
155
+ "civitai": {
156
+ "id": 2620761,
157
+ "modelId": 2106308,
158
+ "name": "F.2 Klein 9B",
159
+ "nsfwLevel": 5,
160
+ "createdAt": "2026-01-22T18:02:58.623Z",
161
+ "updatedAt": "2026-01-22T19:22:23.133Z",
162
+ "status": "Published",
163
+ "publishedAt": "2026-01-22T19:18:42.804Z",
164
+ "trainedWords": [
165
+ "remove the white parts, use the image for context"
166
+ ],
167
+ "trainingStatus": null,
168
+ "trainingDetails": null,
169
+ "baseModel": "Flux.2 Klein 9B-base",
170
+ "baseModelType": "Standard",
171
+ "earlyAccessEndsAt": null,
172
+ "earlyAccessConfig": null,
173
+ "description": "<p>Suggested tags: <code>remove the white parts, use the image for context</code></p><p>Suggested strength: 1.0</p><p>Flux.2 Klein 9B is already very good at doing this without this LoRA. I was quite amazed when testing on/off how many results looked fine without the LoRA. This LoRA still has its use cases in my dataset creation workflow and I spent time and money on <a target=\"_blank\" rel=\"ugc\" href=\"https://runpod.io/?ref=tbivgil2\">Runpod</a> making it so I'll still release it, but feel free to try it with/without. Also tested with colors besides white and it worked fine. My workflow is embedded in all the images again. </p><p>Examples generated using <a target=\"_blank\" rel=\"ugc\" href=\"https://huggingface.co/black-forest-labs/FLUX.2-klein-9b-fp8/blob/main/flux-2-klein-9b-fp8.safetensors\">flux-2-klein-9b-fp8.safetensors</a> and corresponding <a target=\"_blank\" rel=\"ugc\" href=\"https://huggingface.co/Comfy-Org/vae-text-encorder-for-flux-klein-9b/resolve/main/split_files/text_encoders/qwen_3_8b_fp8mixed.safetensors\">fp8 text encoder</a> at around 4-6 steps. Also Flux 2.Klein i2i <a target=\"_blank\" rel=\"ugc\" href=\"https://docs.bfl.ai/guides/prompting_guide_kontext_i2i\">prompting guide</a> is helpful. </p><p>If you are using 9b-base, in my workflow change sampler to euler and increase number of steps.</p>",
174
+ "uploadType": "Created",
175
+ "usageControl": "Download",
176
+ "air": "urn:air:flux2:lora:civitai:2106308@2620761",
177
+ "stats": {
178
+ "downloadCount": 1331,
179
+ "thumbsUpCount": 105
180
+ },
181
+ "model": {
182
+ "name": "UNCROP/INPAINT/OUTPAINT WITH CONTEXT IMAGE for F.2 Klein 9B + QWEN IMAGE EDIT LORA",
183
+ "type": "LORA",
184
+ "nsfw": false,
185
+ "poi": false
186
+ },
187
+ "files": [
188
+ {
189
+ "id": 2508257,
190
+ "sizeKB": 161820.6875,
191
+ "name": "uncrop_F2K9B.safetensors",
192
+ "type": "Model",
193
+ "pickleScanResult": "Success",
194
+ "pickleScanMessage": "No Pickle imports",
195
+ "virusScanResult": "Success",
196
+ "virusScanMessage": null,
197
+ "scannedAt": "2026-01-22T18:06:05.233Z",
198
+ "metadata": {
199
+ "format": "SafeTensor",
200
+ "size": null,
201
+ "fp": null
202
+ },
203
+ "hashes": {
204
+ "AutoV1": "8D64A47F",
205
+ "AutoV2": "3FE9062939",
206
+ "SHA256": "3FE90629390454F4B18A6CBEA8918E8E36223781996A78D35B4C18DB8731E9DB",
207
+ "CRC32": "D603D94B",
208
+ "BLAKE3": "B918399B285D51202BCFF9F06BC101E6233A20EC87EAD61B49FC9131D322522A",
209
+ "AutoV3": "A8E56A5AFDDB"
210
+ },
211
+ "primary": true,
212
+ "downloadUrl": "https://civitai.com/api/download/models/2620761"
213
+ }
214
+ ],
215
+ "images": [
216
+ {
217
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/ac9009e9-cfcf-4443-af2c-0ae36d5762c7/original=true/118388616.mp4",
218
+ "nsfwLevel": 1,
219
+ "width": 832,
220
+ "height": 1248,
221
+ "hash": "U00000fQfQfQfQfQfQfQfQfQfQfQfQfQfQfQ",
222
+ "type": "video",
223
+ "metadata": {
224
+ "hash": "U00000fQfQfQfQfQfQfQfQfQfQfQfQfQfQfQ",
225
+ "size": 590994,
226
+ "audio": true,
227
+ "width": 832,
228
+ "height": 1248,
229
+ "duration": 4.68
230
+ },
231
+ "minor": false,
232
+ "poi": false,
233
+ "meta": {
234
+ "seed": 43,
235
+ "steps": 6,
236
+ "prompt": "remove the white parts, use the image for context. add a cowboy hat"
237
+ },
238
+ "availability": "Public",
239
+ "hasMeta": true,
240
+ "hasPositivePrompt": true,
241
+ "onSite": false,
242
+ "remixOfId": null
243
+ },
244
+ {
245
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/a0fc4429-2466-429a-a46c-397d60dd6916/original=true/118381644.jpeg",
246
+ "nsfwLevel": 4,
247
+ "width": 853,
248
+ "height": 1280,
249
+ "hash": "U6GHby}n9aE23E=xIoxY^,OrxuDi00%#-prW",
250
+ "type": "image",
251
+ "metadata": {
252
+ "hash": "U6GHby}n9aE23E=xIoxY^,OrxuDi00%#-prW",
253
+ "size": 1458325,
254
+ "width": 853,
255
+ "height": 1280
256
+ },
257
+ "minor": false,
258
+ "poi": false,
259
+ "meta": {
260
+ "seed": 44,
261
+ "steps": 6,
262
+ "prompt": "Remove the white parts. Use the image for context. She is looking to the left"
263
+ },
264
+ "availability": "Public",
265
+ "hasMeta": true,
266
+ "hasPositivePrompt": true,
267
+ "onSite": false,
268
+ "remixOfId": null
269
+ },
270
+ {
271
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/76cc2dee-4702-48c9-b52f-7a1edc7f6950/original=true/118382203.jpeg",
272
+ "nsfwLevel": 4,
273
+ "width": 853,
274
+ "height": 1280,
275
+ "hash": "U9INmo.R~V=YbxIp-oI;00ibVsS$00#QRjtR",
276
+ "type": "image",
277
+ "metadata": {
278
+ "hash": "U9INmo.R~V=YbxIp-oI;00ibVsS$00#QRjtR",
279
+ "size": 1344661,
280
+ "width": 853,
281
+ "height": 1280
282
+ },
283
+ "minor": false,
284
+ "poi": false,
285
+ "meta": {
286
+ "seed": 42,
287
+ "steps": 6,
288
+ "prompt": "Remove the white parts, use the butt for context with matching skin color"
289
+ },
290
+ "availability": "Public",
291
+ "hasMeta": true,
292
+ "hasPositivePrompt": true,
293
+ "onSite": false,
294
+ "remixOfId": null
295
+ },
296
+ {
297
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/8814fbb4-6f52-4e1d-93dc-7622fedd5b17/original=true/118377591.jpeg",
298
+ "nsfwLevel": 4,
299
+ "width": 853,
300
+ "height": 1280,
301
+ "hash": "UNKd3%9Ef+I;.mIAkCM{IA%MozNG00x]t7NG",
302
+ "type": "image",
303
+ "metadata": {
304
+ "hash": "UNKd3%9Ef+I;.mIAkCM{IA%MozNG00x]t7NG",
305
+ "size": 1279333,
306
+ "width": 853,
307
+ "height": 1280
308
+ },
309
+ "minor": false,
310
+ "poi": false,
311
+ "meta": {
312
+ "seed": 43,
313
+ "steps": 4,
314
+ "prompt": "Remove the white parts, use the image for context"
315
+ },
316
+ "availability": "Public",
317
+ "hasMeta": true,
318
+ "hasPositivePrompt": true,
319
+ "onSite": false,
320
+ "remixOfId": null
321
+ },
322
+ {
323
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/403915be-05e9-4544-9398-0661be289af3/original=true/118378048.jpeg",
324
+ "nsfwLevel": 4,
325
+ "width": 853,
326
+ "height": 1280,
327
+ "hash": "UALfdch~r_ZQqFm+nnEy00,-j^XT00-UW:Ng",
328
+ "type": "image",
329
+ "metadata": {
330
+ "hash": "UALfdch~r_ZQqFm+nnEy00,-j^XT00-UW:Ng",
331
+ "size": 1389306,
332
+ "width": 853,
333
+ "height": 1280
334
+ },
335
+ "minor": false,
336
+ "poi": false,
337
+ "meta": {
338
+ "seed": 43,
339
+ "steps": 4,
340
+ "prompt": "Remove the white parts. Use the image for context."
341
+ },
342
+ "availability": "Public",
343
+ "hasMeta": true,
344
+ "hasPositivePrompt": true,
345
+ "onSite": false,
346
+ "remixOfId": null
347
+ },
348
+ {
349
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/74c28b74-fa85-48fb-bbea-06bb298cb420/original=true/118381990.jpeg",
350
+ "nsfwLevel": 4,
351
+ "width": 853,
352
+ "height": 1280,
353
+ "hash": "UQM7P*~q%Mr=Io4n%MR-IUxZM_D%ITt5WAoz",
354
+ "type": "image",
355
+ "metadata": {
356
+ "hash": "UQM7P*~q%Mr=Io4n%MR-IUxZM_D%ITt5WAoz",
357
+ "size": 862092,
358
+ "width": 853,
359
+ "height": 1280
360
+ },
361
+ "minor": false,
362
+ "poi": false,
363
+ "meta": {
364
+ "seed": 44,
365
+ "steps": 4,
366
+ "prompt": "Remove the white parts. Use the image for context."
367
+ },
368
+ "availability": "Public",
369
+ "hasMeta": true,
370
+ "hasPositivePrompt": true,
371
+ "onSite": false,
372
+ "remixOfId": null
373
+ },
374
+ {
375
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/ea6b9b04-aaa2-42d1-9421-8d7cc9e059da/original=true/118382581.jpeg",
376
+ "nsfwLevel": 4,
377
+ "width": 853,
378
+ "height": 1280,
379
+ "hash": "UHKmbO}^xukXof$,oz5PVt%MIo%2tTTJInEL",
380
+ "type": "image",
381
+ "metadata": {
382
+ "hash": "UHKmbO}^xukXof$,oz5PVt%MIo%2tTTJInEL",
383
+ "size": 1674338,
384
+ "width": 853,
385
+ "height": 1280
386
+ },
387
+ "minor": false,
388
+ "poi": false,
389
+ "meta": {
390
+ "seed": 42,
391
+ "steps": 5,
392
+ "prompt": "Remove the bikini in image 1, use the dress in image 2 for context. preserve her body shape"
393
+ },
394
+ "availability": "Public",
395
+ "hasMeta": true,
396
+ "hasPositivePrompt": true,
397
+ "onSite": false,
398
+ "remixOfId": null
399
+ },
400
+ {
401
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/cb93c16f-0d58-4aa7-b1d5-03cea43dd6de/original=true/118383163.jpeg",
402
+ "nsfwLevel": 4,
403
+ "width": 853,
404
+ "height": 1280,
405
+ "hash": "ULHnT#Mwrq=|_4IUi]kDE2xDaJX9E1%2fkt7",
406
+ "type": "image",
407
+ "metadata": {
408
+ "hash": "ULHnT#Mwrq=|_4IUi]kDE2xDaJX9E1%2fkt7",
409
+ "size": 1438893,
410
+ "width": 853,
411
+ "height": 1280
412
+ },
413
+ "minor": false,
414
+ "poi": false,
415
+ "meta": {
416
+ "seed": 42,
417
+ "steps": 6,
418
+ "prompt": "remove the white parts, use the bikini in image 2 for context"
419
+ },
420
+ "availability": "Public",
421
+ "hasMeta": true,
422
+ "hasPositivePrompt": true,
423
+ "onSite": false,
424
+ "remixOfId": null
425
+ }
426
+ ],
427
+ "downloadUrl": "https://civitai.com/api/download/models/2620761",
428
+ "_sha256": "3fe90629390454f4b18a6cbea8918e8e36223781996a78d35b4c18db8731e9db",
429
+ "_civitai_api": "https://civitai.com/api/v1/model-versions/by-hash/3fe90629390454f4b18a6cbea8918e8e36223781996a78d35b4c18db8731e9db"
430
+ }
431
+ },
432
+ "trainedWords": [
433
+ {
434
+ "word": "remove the white parts",
435
+ "civitai": true
436
+ },
437
+ {
438
+ "word": "use the image for context",
439
+ "civitai": true
440
+ }
441
+ ],
442
+ "sha256": "3fe90629390454f4b18a6cbea8918e8e36223781996a78d35b4c18db8731e9db",
443
+ "name": "UNCROP/INPAINT/OUTPAINT WITH CONTEXT IMAGE for F.2 Klein 9B + QWEN IMAGE EDIT LORA - F.2 Klein 9B",
444
+ "type": "LORA",
445
+ "baseModel": "Flux.2 Klein 9B-base",
446
+ "links": [
447
+ "https://civitai.com/models/2106308?modelVersionId=2620761",
448
+ "https://civitai.com/api/v1/model-versions/by-hash/3fe90629390454f4b18a6cbea8918e8e36223781996a78d35b4c18db8731e9db"
449
+ ]
450
+ }
text_encoders/Corresponding text encoders.txt ADDED
File without changes
text_encoders/Qwen2.5-VL-7B-Instruct-UD-Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8edd10e3ae170ab8b8d114e70763ff63a220f16deec9a56acf9330034efc0bb5
3
+ size 4785159040
text_encoders/Qwen3-4B-UD-Q6_K_XL.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac0767b5e9c9f16efe57ce422253a33747970c166c3131c4d4d59d20511f07e1
3
+ size 3658223392
text_encoders/Qwen3-8B-abliterated-q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:efb23366a9cfd3e631e1f9f500c7e95f0a308d6560d21a9f5809a0808466dc15
3
+ size 8709518304