msrcam commited on
Commit
e7bbaad
·
verified ·
1 Parent(s): 77d7c5d

Upload 12 files

Browse files
FLUX2_Klein_4LoRA.json ADDED
@@ -0,0 +1,333 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "last_node_id": 15,
3
+ "last_link_id": 18,
4
+ "nodes": [
5
+ {
6
+ "id": 1,
7
+ "type": "UNETLoader",
8
+ "pos": [100, 100],
9
+ "size": [315, 58],
10
+ "flags": {},
11
+ "order": 0,
12
+ "mode": 0,
13
+ "outputs": [
14
+ {
15
+ "name": "MODEL",
16
+ "type": "MODEL",
17
+ "links": [1]
18
+ }
19
+ ],
20
+ "properties": {},
21
+ "widgets_values": ["flux-2-klein-base-9b-fp8.safetensors", "default"]
22
+ },
23
+ {
24
+ "id": 2,
25
+ "type": "DualCLIPLoader",
26
+ "pos": [100, 200],
27
+ "size": [315, 106],
28
+ "flags": {},
29
+ "order": 1,
30
+ "mode": 0,
31
+ "outputs": [
32
+ {
33
+ "name": "CLIP",
34
+ "type": "CLIP",
35
+ "links": [2]
36
+ }
37
+ ],
38
+ "properties": {},
39
+ "widgets_values": ["clip_l.safetensors", "qwen_3_8b_fp8mixed.safetensors", "flux", "default"]
40
+ },
41
+ {
42
+ "id": 3,
43
+ "type": "VAELoader",
44
+ "pos": [100, 350],
45
+ "size": [315, 58],
46
+ "flags": {},
47
+ "order": 2,
48
+ "mode": 0,
49
+ "outputs": [
50
+ {
51
+ "name": "VAE",
52
+ "type": "VAE",
53
+ "links": [3]
54
+ }
55
+ ],
56
+ "properties": {},
57
+ "widgets_values": ["flux2-vae.safetensors"]
58
+ },
59
+ {
60
+ "id": 4,
61
+ "type": "LoraLoaderModelOnly",
62
+ "pos": [500, 100],
63
+ "size": [315, 82],
64
+ "flags": {},
65
+ "order": 3,
66
+ "mode": 0,
67
+ "inputs": [
68
+ {
69
+ "name": "model",
70
+ "type": "MODEL",
71
+ "link": 1
72
+ }
73
+ ],
74
+ "outputs": [
75
+ {
76
+ "name": "MODEL",
77
+ "type": "MODEL",
78
+ "links": [4]
79
+ }
80
+ ],
81
+ "properties": {},
82
+ "widgets_values": ["clothes_remover_v0.safetensors", 1.0]
83
+ },
84
+ {
85
+ "id": 5,
86
+ "type": "LoraLoaderModelOnly",
87
+ "pos": [500, 220],
88
+ "size": [315, 82],
89
+ "flags": {},
90
+ "order": 4,
91
+ "mode": 0,
92
+ "inputs": [
93
+ {
94
+ "name": "model",
95
+ "type": "MODEL",
96
+ "link": 4
97
+ }
98
+ ],
99
+ "outputs": [
100
+ {
101
+ "name": "MODEL",
102
+ "type": "MODEL",
103
+ "links": [5]
104
+ }
105
+ ],
106
+ "properties": {},
107
+ "widgets_values": ["mytryon4_comfy.safetensors", 1.0]
108
+ },
109
+ {
110
+ "id": 6,
111
+ "type": "LoraLoaderModelOnly",
112
+ "pos": [500, 340],
113
+ "size": [315, 82],
114
+ "flags": {},
115
+ "order": 5,
116
+ "mode": 0,
117
+ "inputs": [
118
+ {
119
+ "name": "model",
120
+ "type": "MODEL",
121
+ "link": 5
122
+ }
123
+ ],
124
+ "outputs": [
125
+ {
126
+ "name": "MODEL",
127
+ "type": "MODEL",
128
+ "links": [6]
129
+ }
130
+ ],
131
+ "properties": {},
132
+ "widgets_values": ["Zoeeee_Body_Adjuster_kontext.safetensors", 1.0]
133
+ },
134
+ {
135
+ "id": 7,
136
+ "type": "LoraLoaderModelOnly",
137
+ "pos": [500, 460],
138
+ "size": [315, 82],
139
+ "flags": {},
140
+ "order": 6,
141
+ "mode": 0,
142
+ "inputs": [
143
+ {
144
+ "name": "model",
145
+ "type": "MODEL",
146
+ "link": 6
147
+ }
148
+ ],
149
+ "outputs": [
150
+ {
151
+ "name": "MODEL",
152
+ "type": "MODEL",
153
+ "links": [7]
154
+ }
155
+ ],
156
+ "properties": {},
157
+ "widgets_values": ["JD3s_Nudify_Kontext.safetensors", 1.0]
158
+ },
159
+ {
160
+ "id": 8,
161
+ "type": "CLIPTextEncode",
162
+ "pos": [500, 580],
163
+ "size": [400, 200],
164
+ "flags": {},
165
+ "order": 7,
166
+ "mode": 0,
167
+ "inputs": [
168
+ {
169
+ "name": "clip",
170
+ "type": "CLIP",
171
+ "link": 2
172
+ }
173
+ ],
174
+ "outputs": [
175
+ {
176
+ "name": "CONDITIONING",
177
+ "type": "CONDITIONING",
178
+ "links": [8]
179
+ }
180
+ ],
181
+ "properties": {},
182
+ "widgets_values": ["a beautiful woman in elegant dress, professional photo, studio lighting, high quality"]
183
+ },
184
+ {
185
+ "id": 9,
186
+ "type": "CLIPTextEncode",
187
+ "pos": [500, 820],
188
+ "size": [400, 130],
189
+ "flags": {},
190
+ "order": 8,
191
+ "mode": 0,
192
+ "inputs": [
193
+ {
194
+ "name": "clip",
195
+ "type": "CLIP",
196
+ "link": 2
197
+ }
198
+ ],
199
+ "outputs": [
200
+ {
201
+ "name": "CONDITIONING",
202
+ "type": "CONDITIONING",
203
+ "links": [9]
204
+ }
205
+ ],
206
+ "properties": {},
207
+ "widgets_values": ["ugly, blurry, low quality, deformed"]
208
+ },
209
+ {
210
+ "id": 10,
211
+ "type": "EmptyLatentImage",
212
+ "pos": [900, 700],
213
+ "size": [315, 106],
214
+ "flags": {},
215
+ "order": 9,
216
+ "mode": 0,
217
+ "outputs": [
218
+ {
219
+ "name": "LATENT",
220
+ "type": "LATENT",
221
+ "links": [10]
222
+ }
223
+ ],
224
+ "properties": {},
225
+ "widgets_values": [1024, 1024, 1]
226
+ },
227
+ {
228
+ "id": 11,
229
+ "type": "KSampler",
230
+ "pos": [950, 100],
231
+ "size": [315, 262],
232
+ "flags": {},
233
+ "order": 10,
234
+ "mode": 0,
235
+ "inputs": [
236
+ {
237
+ "name": "model",
238
+ "type": "MODEL",
239
+ "link": 7
240
+ },
241
+ {
242
+ "name": "positive",
243
+ "type": "CONDITIONING",
244
+ "link": 8
245
+ },
246
+ {
247
+ "name": "negative",
248
+ "type": "CONDITIONING",
249
+ "link": 9
250
+ },
251
+ {
252
+ "name": "latent_image",
253
+ "type": "LATENT",
254
+ "link": 10
255
+ }
256
+ ],
257
+ "outputs": [
258
+ {
259
+ "name": "LATENT",
260
+ "type": "LATENT",
261
+ "links": [11]
262
+ }
263
+ ],
264
+ "properties": {},
265
+ "widgets_values": [123456789, "randomize", 20, 3.5, "euler", "simple", 1.0]
266
+ },
267
+ {
268
+ "id": 12,
269
+ "type": "VAEDecode",
270
+ "pos": [1300, 100],
271
+ "size": [210, 46],
272
+ "flags": {},
273
+ "order": 11,
274
+ "mode": 0,
275
+ "inputs": [
276
+ {
277
+ "name": "samples",
278
+ "type": "LATENT",
279
+ "link": 11
280
+ },
281
+ {
282
+ "name": "vae",
283
+ "type": "VAE",
284
+ "link": 3
285
+ }
286
+ ],
287
+ "outputs": [
288
+ {
289
+ "name": "IMAGE",
290
+ "type": "IMAGE",
291
+ "links": [12]
292
+ }
293
+ ],
294
+ "properties": {}
295
+ },
296
+ {
297
+ "id": 13,
298
+ "type": "SaveImage",
299
+ "pos": [1300, 200],
300
+ "size": [315, 270],
301
+ "flags": {},
302
+ "order": 12,
303
+ "mode": 0,
304
+ "inputs": [
305
+ {
306
+ "name": "images",
307
+ "type": "IMAGE",
308
+ "link": 12
309
+ }
310
+ ],
311
+ "properties": {},
312
+ "widgets_values": ["Klein_output"]
313
+ }
314
+ ],
315
+ "links": [
316
+ [1, 1, 0, 4, 0, "MODEL"],
317
+ [2, 2, 0, 8, 0, "CLIP"],
318
+ [3, 3, 0, 12, 1, "VAE"],
319
+ [4, 4, 0, 5, 0, "MODEL"],
320
+ [5, 5, 0, 6, 0, "MODEL"],
321
+ [6, 6, 0, 7, 0, "MODEL"],
322
+ [7, 7, 0, 11, 0, "MODEL"],
323
+ [8, 8, 0, 11, 1, "CONDITIONING"],
324
+ [9, 9, 0, 11, 2, "CONDITIONING"],
325
+ [10, 10, 0, 11, 3, "LATENT"],
326
+ [11, 11, 0, 12, 0, "LATENT"],
327
+ [12, 12, 0, 13, 0, "IMAGE"]
328
+ ],
329
+ "groups": [],
330
+ "config": {},
331
+ "extra": {},
332
+ "version": 0.4
333
+ }
FLUX2_Klein_img2img_4LoRA.json ADDED
@@ -0,0 +1,369 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "last_node_id": 16,
3
+ "last_link_id": 20,
4
+ "nodes": [
5
+ {
6
+ "id": 1,
7
+ "type": "UNETLoader",
8
+ "pos": [100, 100],
9
+ "size": [315, 58],
10
+ "flags": {},
11
+ "order": 0,
12
+ "mode": 0,
13
+ "outputs": [
14
+ {
15
+ "name": "MODEL",
16
+ "type": "MODEL",
17
+ "links": [1]
18
+ }
19
+ ],
20
+ "properties": {},
21
+ "widgets_values": ["flux-2-klein-base-9b-fp8.safetensors", "default"]
22
+ },
23
+ {
24
+ "id": 2,
25
+ "type": "DualCLIPLoader",
26
+ "pos": [100, 200],
27
+ "size": [315, 106],
28
+ "flags": {},
29
+ "order": 1,
30
+ "mode": 0,
31
+ "outputs": [
32
+ {
33
+ "name": "CLIP",
34
+ "type": "CLIP",
35
+ "links": [2]
36
+ }
37
+ ],
38
+ "properties": {},
39
+ "widgets_values": ["clip_l.safetensors", "qwen_3_8b_fp8mixed.safetensors", "flux", "default"]
40
+ },
41
+ {
42
+ "id": 3,
43
+ "type": "VAELoader",
44
+ "pos": [100, 350],
45
+ "size": [315, 58],
46
+ "flags": {},
47
+ "order": 2,
48
+ "mode": 0,
49
+ "outputs": [
50
+ {
51
+ "name": "VAE",
52
+ "type": "VAE",
53
+ "links": [3, 13]
54
+ }
55
+ ],
56
+ "properties": {},
57
+ "widgets_values": ["flux2-vae.safetensors"]
58
+ },
59
+ {
60
+ "id": 14,
61
+ "type": "LoadImage",
62
+ "pos": [100, 450],
63
+ "size": [315, 314],
64
+ "flags": {},
65
+ "order": 3,
66
+ "mode": 0,
67
+ "outputs": [
68
+ {
69
+ "name": "IMAGE",
70
+ "type": "IMAGE",
71
+ "links": [14]
72
+ },
73
+ {
74
+ "name": "MASK",
75
+ "type": "MASK",
76
+ "links": null
77
+ }
78
+ ],
79
+ "properties": {},
80
+ "widgets_values": ["example.png", "image"]
81
+ },
82
+ {
83
+ "id": 15,
84
+ "type": "VAEEncode",
85
+ "pos": [500, 450],
86
+ "size": [210, 46],
87
+ "flags": {},
88
+ "order": 4,
89
+ "mode": 0,
90
+ "inputs": [
91
+ {
92
+ "name": "pixels",
93
+ "type": "IMAGE",
94
+ "link": 14
95
+ },
96
+ {
97
+ "name": "vae",
98
+ "type": "VAE",
99
+ "link": 13
100
+ }
101
+ ],
102
+ "outputs": [
103
+ {
104
+ "name": "LATENT",
105
+ "type": "LATENT",
106
+ "links": [15]
107
+ }
108
+ ],
109
+ "properties": {}
110
+ },
111
+ {
112
+ "id": 4,
113
+ "type": "LoraLoaderModelOnly",
114
+ "pos": [500, 100],
115
+ "size": [315, 82],
116
+ "flags": {},
117
+ "order": 5,
118
+ "mode": 0,
119
+ "inputs": [
120
+ {
121
+ "name": "model",
122
+ "type": "MODEL",
123
+ "link": 1
124
+ }
125
+ ],
126
+ "outputs": [
127
+ {
128
+ "name": "MODEL",
129
+ "type": "MODEL",
130
+ "links": [4]
131
+ }
132
+ ],
133
+ "properties": {},
134
+ "widgets_values": ["clothes_remover_v0.safetensors", 0.8]
135
+ },
136
+ {
137
+ "id": 5,
138
+ "type": "LoraLoaderModelOnly",
139
+ "pos": [850, 100],
140
+ "size": [315, 82],
141
+ "flags": {},
142
+ "order": 6,
143
+ "mode": 0,
144
+ "inputs": [
145
+ {
146
+ "name": "model",
147
+ "type": "MODEL",
148
+ "link": 4
149
+ }
150
+ ],
151
+ "outputs": [
152
+ {
153
+ "name": "MODEL",
154
+ "type": "MODEL",
155
+ "links": [5]
156
+ }
157
+ ],
158
+ "properties": {},
159
+ "widgets_values": ["mytryon4_comfy.safetensors", 0.8]
160
+ },
161
+ {
162
+ "id": 6,
163
+ "type": "LoraLoaderModelOnly",
164
+ "pos": [1200, 100],
165
+ "size": [315, 82],
166
+ "flags": {},
167
+ "order": 7,
168
+ "mode": 0,
169
+ "inputs": [
170
+ {
171
+ "name": "model",
172
+ "type": "MODEL",
173
+ "link": 5
174
+ }
175
+ ],
176
+ "outputs": [
177
+ {
178
+ "name": "MODEL",
179
+ "type": "MODEL",
180
+ "links": [6]
181
+ }
182
+ ],
183
+ "properties": {},
184
+ "widgets_values": ["Zoeeee_Body_Adjuster_kontext.safetensors", 0.6]
185
+ },
186
+ {
187
+ "id": 7,
188
+ "type": "LoraLoaderModelOnly",
189
+ "pos": [1550, 100],
190
+ "size": [315, 82],
191
+ "flags": {},
192
+ "order": 8,
193
+ "mode": 0,
194
+ "inputs": [
195
+ {
196
+ "name": "model",
197
+ "type": "MODEL",
198
+ "link": 6
199
+ }
200
+ ],
201
+ "outputs": [
202
+ {
203
+ "name": "MODEL",
204
+ "type": "MODEL",
205
+ "links": [7]
206
+ }
207
+ ],
208
+ "properties": {},
209
+ "widgets_values": ["JD3s_Nudify_Kontext.safetensors", 0.7]
210
+ },
211
+ {
212
+ "id": 8,
213
+ "type": "CLIPTextEncode",
214
+ "pos": [500, 550],
215
+ "size": [400, 150],
216
+ "flags": {},
217
+ "order": 9,
218
+ "mode": 0,
219
+ "inputs": [
220
+ {
221
+ "name": "clip",
222
+ "type": "CLIP",
223
+ "link": 2
224
+ }
225
+ ],
226
+ "outputs": [
227
+ {
228
+ "name": "CONDITIONING",
229
+ "type": "CONDITIONING",
230
+ "links": [8]
231
+ }
232
+ ],
233
+ "properties": {},
234
+ "widgets_values": ["same person, change clothes to red dress, keep face identical, keep hair identical, keep body shape, professional photo"]
235
+ },
236
+ {
237
+ "id": 9,
238
+ "type": "CLIPTextEncode",
239
+ "pos": [500, 740],
240
+ "size": [400, 100],
241
+ "flags": {},
242
+ "order": 10,
243
+ "mode": 0,
244
+ "inputs": [
245
+ {
246
+ "name": "clip",
247
+ "type": "CLIP",
248
+ "link": 2
249
+ }
250
+ ],
251
+ "outputs": [
252
+ {
253
+ "name": "CONDITIONING",
254
+ "type": "CONDITIONING",
255
+ "links": [9]
256
+ }
257
+ ],
258
+ "properties": {},
259
+ "widgets_values": ["different person, changed face, ugly, blurry, deformed"]
260
+ },
261
+ {
262
+ "id": 11,
263
+ "type": "KSampler",
264
+ "pos": [950, 250],
265
+ "size": [315, 262],
266
+ "flags": {},
267
+ "order": 11,
268
+ "mode": 0,
269
+ "inputs": [
270
+ {
271
+ "name": "model",
272
+ "type": "MODEL",
273
+ "link": 7
274
+ },
275
+ {
276
+ "name": "positive",
277
+ "type": "CONDITIONING",
278
+ "link": 8
279
+ },
280
+ {
281
+ "name": "negative",
282
+ "type": "CONDITIONING",
283
+ "link": 9
284
+ },
285
+ {
286
+ "name": "latent_image",
287
+ "type": "LATENT",
288
+ "link": 15
289
+ }
290
+ ],
291
+ "outputs": [
292
+ {
293
+ "name": "LATENT",
294
+ "type": "LATENT",
295
+ "links": [11]
296
+ }
297
+ ],
298
+ "properties": {},
299
+ "widgets_values": [123456789, "randomize", 25, 3.5, "euler", "simple", 0.45]
300
+ },
301
+ {
302
+ "id": 12,
303
+ "type": "VAEDecode",
304
+ "pos": [1300, 250],
305
+ "size": [210, 46],
306
+ "flags": {},
307
+ "order": 12,
308
+ "mode": 0,
309
+ "inputs": [
310
+ {
311
+ "name": "samples",
312
+ "type": "LATENT",
313
+ "link": 11
314
+ },
315
+ {
316
+ "name": "vae",
317
+ "type": "VAE",
318
+ "link": 3
319
+ }
320
+ ],
321
+ "outputs": [
322
+ {
323
+ "name": "IMAGE",
324
+ "type": "IMAGE",
325
+ "links": [12]
326
+ }
327
+ ],
328
+ "properties": {}
329
+ },
330
+ {
331
+ "id": 13,
332
+ "type": "SaveImage",
333
+ "pos": [1300, 350],
334
+ "size": [315, 270],
335
+ "flags": {},
336
+ "order": 13,
337
+ "mode": 0,
338
+ "inputs": [
339
+ {
340
+ "name": "images",
341
+ "type": "IMAGE",
342
+ "link": 12
343
+ }
344
+ ],
345
+ "properties": {},
346
+ "widgets_values": ["Klein_img2img"]
347
+ }
348
+ ],
349
+ "links": [
350
+ [1, 1, 0, 4, 0, "MODEL"],
351
+ [2, 2, 0, 8, 0, "CLIP"],
352
+ [3, 3, 0, 12, 1, "VAE"],
353
+ [4, 4, 0, 5, 0, "MODEL"],
354
+ [5, 5, 0, 6, 0, "MODEL"],
355
+ [6, 6, 0, 7, 0, "MODEL"],
356
+ [7, 7, 0, 11, 0, "MODEL"],
357
+ [8, 8, 0, 11, 1, "CONDITIONING"],
358
+ [9, 9, 0, 11, 2, "CONDITIONING"],
359
+ [11, 11, 0, 12, 0, "LATENT"],
360
+ [12, 12, 0, 13, 0, "IMAGE"],
361
+ [13, 3, 0, 15, 1, "VAE"],
362
+ [14, 14, 0, 15, 0, "IMAGE"],
363
+ [15, 15, 0, 11, 3, "LATENT"]
364
+ ],
365
+ "groups": [],
366
+ "config": {},
367
+ "extra": {},
368
+ "version": 0.4
369
+ }
Flux.2 D - 2000s style.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca463db1bf7d0193dcee5e2a74dfc2f47ecbeeb3bdc11f4d27545721fc20bc5a
3
+ size 390112168
I2I_flux2_klein_and_flux1_kontext_Nudifier.json ADDED
The diff for this file is too large to render. See raw diff
 
NSFW-klein.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:911db07558661f9dd2b57271b6e16b6b0016fea09d34f61575f7ec1aea0a949b
3
+ size 165704392
The_Body_Version_A_Flux2.k.9B_r16_AdamW8Bit_Weighted_768_woman_000005000.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8cec9dce5775d9d1134dc7bf3399c569e03267b9454d078a8303bb500b7a5155
3
+ size 82866856
anus_fem_klein.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd8b088d067a5d4aa9f5643d8658297122468bc95d7244806291461b3c49ac79
3
+ size 165704384
bj_20260120_22-22-29epoch15_comfy.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6cb0f5363ed04d6d6c369f9a2753c1398f7175ab4c543422bdefc89c7be9efe1
3
+ size 99644032
cameltoe_klein.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c91c87da4b85bf6c076e2069842c449b06d83798c10a8a385f08c00271101f0b
3
+ size 82866712
nipplediffusion-f2-klein-9b.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f297b81fbe8d7c6ddfd49803f0c4871407ece773c3a21f32f5de1976552334c
3
+ size 165721560
pawg_klein.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:973d611fa277920fe64ada89ec820958f0ceb9c6c335063371672c8899629804
3
+ size 165704376
pussydiffusion-f2-klein-9b.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca41a1c408e233d8ac972b0d131685a5136c947aea42005a4b7307d0e9dcdd84
3
+ size 165720472