Sarim-Hash commited on
Commit
e167e99
·
verified ·
1 Parent(s): 69ad353

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +146 -0
  2. vlcs_upscaled/model/README.md +132 -0
  3. vlcs_upscaled/model/all_image_files_pacs.json +0 -0
  4. vlcs_upscaled/model/all_text_cache_files_text-embeds.json +1 -0
  5. vlcs_upscaled/model/all_vae_cache_files_pacs.json +1 -0
  6. vlcs_upscaled/model/assets/image_0_0.png +3 -0
  7. vlcs_upscaled/model/assets/image_1_0.png +3 -0
  8. vlcs_upscaled/model/benchmarks/base_model/unconditional_512x512.png +3 -0
  9. vlcs_upscaled/model/benchmarks/base_model/validation_512x512.png +3 -0
  10. vlcs_upscaled/model/checkpoint-1000/README.md +132 -0
  11. vlcs_upscaled/model/checkpoint-1000/assets/image_0_0.png +3 -0
  12. vlcs_upscaled/model/checkpoint-1000/assets/image_1_0.png +3 -0
  13. vlcs_upscaled/model/checkpoint-1000/optimizer.bin +3 -0
  14. vlcs_upscaled/model/checkpoint-1000/pytorch_lora_weights.safetensors +3 -0
  15. vlcs_upscaled/model/checkpoint-1000/random_states_0.pkl +3 -0
  16. vlcs_upscaled/model/checkpoint-1000/scheduler.bin +3 -0
  17. vlcs_upscaled/model/checkpoint-1000/training_state-pacs.json +0 -0
  18. vlcs_upscaled/model/checkpoint-1000/training_state.json +1 -0
  19. vlcs_upscaled/model/checkpoint-1250/README.md +132 -0
  20. vlcs_upscaled/model/checkpoint-1250/assets/image_0_0.png +3 -0
  21. vlcs_upscaled/model/checkpoint-1250/assets/image_1_0.png +3 -0
  22. vlcs_upscaled/model/checkpoint-1250/optimizer.bin +3 -0
  23. vlcs_upscaled/model/checkpoint-1250/pytorch_lora_weights.safetensors +3 -0
  24. vlcs_upscaled/model/checkpoint-1250/random_states_0.pkl +3 -0
  25. vlcs_upscaled/model/checkpoint-1250/scheduler.bin +3 -0
  26. vlcs_upscaled/model/checkpoint-1250/training_state-pacs.json +0 -0
  27. vlcs_upscaled/model/checkpoint-1250/training_state.json +1 -0
  28. vlcs_upscaled/model/checkpoint-1500/README.md +132 -0
  29. vlcs_upscaled/model/checkpoint-1500/assets/image_0_0.png +3 -0
  30. vlcs_upscaled/model/checkpoint-1500/assets/image_1_0.png +3 -0
  31. vlcs_upscaled/model/checkpoint-1500/optimizer.bin +3 -0
  32. vlcs_upscaled/model/checkpoint-1500/pytorch_lora_weights.safetensors +3 -0
  33. vlcs_upscaled/model/checkpoint-1500/random_states_0.pkl +3 -0
  34. vlcs_upscaled/model/checkpoint-1500/scheduler.bin +3 -0
  35. vlcs_upscaled/model/checkpoint-1500/training_state-pacs.json +0 -0
  36. vlcs_upscaled/model/checkpoint-1500/training_state.json +1 -0
  37. vlcs_upscaled/model/checkpoint-1750/README.md +132 -0
  38. vlcs_upscaled/model/checkpoint-1750/assets/image_0_0.png +3 -0
  39. vlcs_upscaled/model/checkpoint-1750/assets/image_1_0.png +3 -0
  40. vlcs_upscaled/model/checkpoint-1750/optimizer.bin +3 -0
  41. vlcs_upscaled/model/checkpoint-1750/pytorch_lora_weights.safetensors +3 -0
  42. vlcs_upscaled/model/checkpoint-1750/random_states_0.pkl +3 -0
  43. vlcs_upscaled/model/checkpoint-1750/scheduler.bin +3 -0
  44. vlcs_upscaled/model/checkpoint-1750/training_state-pacs.json +0 -0
  45. vlcs_upscaled/model/checkpoint-1750/training_state.json +1 -0
  46. vlcs_upscaled/model/checkpoint-2000/README.md +132 -0
  47. vlcs_upscaled/model/checkpoint-2000/assets/image_0_0.png +3 -0
  48. vlcs_upscaled/model/checkpoint-2000/assets/image_1_0.png +3 -0
  49. vlcs_upscaled/model/checkpoint-2000/optimizer.bin +3 -0
  50. vlcs_upscaled/model/checkpoint-2000/pytorch_lora_weights.safetensors +3 -0
.gitattributes CHANGED
@@ -179,3 +179,149 @@ digit_upscaled/validation_images/step_900_unconditional_512x512.png filter=lfs d
179
  digit_upscaled/validation_images/step_900_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
180
  digit_upscaled/validation_images/step_950_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
181
  digit_upscaled/validation_images/step_950_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
179
  digit_upscaled/validation_images/step_900_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
180
  digit_upscaled/validation_images/step_950_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
181
  digit_upscaled/validation_images/step_950_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
182
+ vlcs_upscaled/model/assets/image_0_0.png filter=lfs diff=lfs merge=lfs -text
183
+ vlcs_upscaled/model/assets/image_1_0.png filter=lfs diff=lfs merge=lfs -text
184
+ vlcs_upscaled/model/benchmarks/base_model/unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
185
+ vlcs_upscaled/model/benchmarks/base_model/validation_512x512.png filter=lfs diff=lfs merge=lfs -text
186
+ vlcs_upscaled/model/checkpoint-1000/assets/image_0_0.png filter=lfs diff=lfs merge=lfs -text
187
+ vlcs_upscaled/model/checkpoint-1000/assets/image_1_0.png filter=lfs diff=lfs merge=lfs -text
188
+ vlcs_upscaled/model/checkpoint-1250/assets/image_0_0.png filter=lfs diff=lfs merge=lfs -text
189
+ vlcs_upscaled/model/checkpoint-1250/assets/image_1_0.png filter=lfs diff=lfs merge=lfs -text
190
+ vlcs_upscaled/model/checkpoint-1500/assets/image_0_0.png filter=lfs diff=lfs merge=lfs -text
191
+ vlcs_upscaled/model/checkpoint-1500/assets/image_1_0.png filter=lfs diff=lfs merge=lfs -text
192
+ vlcs_upscaled/model/checkpoint-1750/assets/image_0_0.png filter=lfs diff=lfs merge=lfs -text
193
+ vlcs_upscaled/model/checkpoint-1750/assets/image_1_0.png filter=lfs diff=lfs merge=lfs -text
194
+ vlcs_upscaled/model/checkpoint-2000/assets/image_0_0.png filter=lfs diff=lfs merge=lfs -text
195
+ vlcs_upscaled/model/checkpoint-2000/assets/image_1_0.png filter=lfs diff=lfs merge=lfs -text
196
+ vlcs_upscaled/model/checkpoint-2250/assets/image_0_0.png filter=lfs diff=lfs merge=lfs -text
197
+ vlcs_upscaled/model/checkpoint-2250/assets/image_1_0.png filter=lfs diff=lfs merge=lfs -text
198
+ vlcs_upscaled/model/checkpoint-250/assets/image_0_0.png filter=lfs diff=lfs merge=lfs -text
199
+ vlcs_upscaled/model/checkpoint-250/assets/image_1_0.png filter=lfs diff=lfs merge=lfs -text
200
+ vlcs_upscaled/model/checkpoint-2500/assets/image_0_0.png filter=lfs diff=lfs merge=lfs -text
201
+ vlcs_upscaled/model/checkpoint-2500/assets/image_1_0.png filter=lfs diff=lfs merge=lfs -text
202
+ vlcs_upscaled/model/checkpoint-2750/assets/image_0_0.png filter=lfs diff=lfs merge=lfs -text
203
+ vlcs_upscaled/model/checkpoint-2750/assets/image_1_0.png filter=lfs diff=lfs merge=lfs -text
204
+ vlcs_upscaled/model/checkpoint-500/assets/image_0_0.png filter=lfs diff=lfs merge=lfs -text
205
+ vlcs_upscaled/model/checkpoint-500/assets/image_1_0.png filter=lfs diff=lfs merge=lfs -text
206
+ vlcs_upscaled/model/checkpoint-750/assets/image_0_0.png filter=lfs diff=lfs merge=lfs -text
207
+ vlcs_upscaled/model/checkpoint-750/assets/image_1_0.png filter=lfs diff=lfs merge=lfs -text
208
+ vlcs_upscaled/model/validation_images/step_0_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
209
+ vlcs_upscaled/model/validation_images/step_0_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
210
+ vlcs_upscaled/model/validation_images/step_1000_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
211
+ vlcs_upscaled/model/validation_images/step_1000_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
212
+ vlcs_upscaled/model/validation_images/step_100_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
213
+ vlcs_upscaled/model/validation_images/step_100_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
214
+ vlcs_upscaled/model/validation_images/step_1050_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
215
+ vlcs_upscaled/model/validation_images/step_1050_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
216
+ vlcs_upscaled/model/validation_images/step_1100_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
217
+ vlcs_upscaled/model/validation_images/step_1100_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
218
+ vlcs_upscaled/model/validation_images/step_1150_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
219
+ vlcs_upscaled/model/validation_images/step_1150_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
220
+ vlcs_upscaled/model/validation_images/step_1200_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
221
+ vlcs_upscaled/model/validation_images/step_1200_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
222
+ vlcs_upscaled/model/validation_images/step_1250_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
223
+ vlcs_upscaled/model/validation_images/step_1250_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
224
+ vlcs_upscaled/model/validation_images/step_1300_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
225
+ vlcs_upscaled/model/validation_images/step_1300_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
226
+ vlcs_upscaled/model/validation_images/step_1350_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
227
+ vlcs_upscaled/model/validation_images/step_1350_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
228
+ vlcs_upscaled/model/validation_images/step_1400_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
229
+ vlcs_upscaled/model/validation_images/step_1400_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
230
+ vlcs_upscaled/model/validation_images/step_1450_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
231
+ vlcs_upscaled/model/validation_images/step_1450_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
232
+ vlcs_upscaled/model/validation_images/step_1500_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
233
+ vlcs_upscaled/model/validation_images/step_1500_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
234
+ vlcs_upscaled/model/validation_images/step_150_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
235
+ vlcs_upscaled/model/validation_images/step_150_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
236
+ vlcs_upscaled/model/validation_images/step_1550_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
237
+ vlcs_upscaled/model/validation_images/step_1550_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
238
+ vlcs_upscaled/model/validation_images/step_1600_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
239
+ vlcs_upscaled/model/validation_images/step_1600_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
240
+ vlcs_upscaled/model/validation_images/step_1650_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
241
+ vlcs_upscaled/model/validation_images/step_1650_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
242
+ vlcs_upscaled/model/validation_images/step_1700_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
243
+ vlcs_upscaled/model/validation_images/step_1700_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
244
+ vlcs_upscaled/model/validation_images/step_1750_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
245
+ vlcs_upscaled/model/validation_images/step_1750_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
246
+ vlcs_upscaled/model/validation_images/step_1800_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
247
+ vlcs_upscaled/model/validation_images/step_1800_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
248
+ vlcs_upscaled/model/validation_images/step_1850_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
249
+ vlcs_upscaled/model/validation_images/step_1850_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
250
+ vlcs_upscaled/model/validation_images/step_1900_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
251
+ vlcs_upscaled/model/validation_images/step_1900_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
252
+ vlcs_upscaled/model/validation_images/step_1950_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
253
+ vlcs_upscaled/model/validation_images/step_1950_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
254
+ vlcs_upscaled/model/validation_images/step_2000_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
255
+ vlcs_upscaled/model/validation_images/step_2000_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
256
+ vlcs_upscaled/model/validation_images/step_200_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
257
+ vlcs_upscaled/model/validation_images/step_200_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
258
+ vlcs_upscaled/model/validation_images/step_2050_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
259
+ vlcs_upscaled/model/validation_images/step_2050_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
260
+ vlcs_upscaled/model/validation_images/step_2100_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
261
+ vlcs_upscaled/model/validation_images/step_2100_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
262
+ vlcs_upscaled/model/validation_images/step_2150_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
263
+ vlcs_upscaled/model/validation_images/step_2150_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
264
+ vlcs_upscaled/model/validation_images/step_2200_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
265
+ vlcs_upscaled/model/validation_images/step_2200_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
266
+ vlcs_upscaled/model/validation_images/step_2250_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
267
+ vlcs_upscaled/model/validation_images/step_2250_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
268
+ vlcs_upscaled/model/validation_images/step_2300_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
269
+ vlcs_upscaled/model/validation_images/step_2300_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
270
+ vlcs_upscaled/model/validation_images/step_2350_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
271
+ vlcs_upscaled/model/validation_images/step_2350_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
272
+ vlcs_upscaled/model/validation_images/step_2400_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
273
+ vlcs_upscaled/model/validation_images/step_2400_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
274
+ vlcs_upscaled/model/validation_images/step_2450_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
275
+ vlcs_upscaled/model/validation_images/step_2450_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
276
+ vlcs_upscaled/model/validation_images/step_2500_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
277
+ vlcs_upscaled/model/validation_images/step_2500_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
278
+ vlcs_upscaled/model/validation_images/step_250_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
279
+ vlcs_upscaled/model/validation_images/step_250_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
280
+ vlcs_upscaled/model/validation_images/step_2550_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
281
+ vlcs_upscaled/model/validation_images/step_2550_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
282
+ vlcs_upscaled/model/validation_images/step_2600_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
283
+ vlcs_upscaled/model/validation_images/step_2600_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
284
+ vlcs_upscaled/model/validation_images/step_2650_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
285
+ vlcs_upscaled/model/validation_images/step_2650_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
286
+ vlcs_upscaled/model/validation_images/step_2700_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
287
+ vlcs_upscaled/model/validation_images/step_2700_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
288
+ vlcs_upscaled/model/validation_images/step_2750_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
289
+ vlcs_upscaled/model/validation_images/step_2750_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
290
+ vlcs_upscaled/model/validation_images/step_2800_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
291
+ vlcs_upscaled/model/validation_images/step_2800_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
292
+ vlcs_upscaled/model/validation_images/step_2850_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
293
+ vlcs_upscaled/model/validation_images/step_2850_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
294
+ vlcs_upscaled/model/validation_images/step_2900_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
295
+ vlcs_upscaled/model/validation_images/step_2900_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
296
+ vlcs_upscaled/model/validation_images/step_2950_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
297
+ vlcs_upscaled/model/validation_images/step_2950_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
298
+ vlcs_upscaled/model/validation_images/step_300_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
299
+ vlcs_upscaled/model/validation_images/step_300_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
300
+ vlcs_upscaled/model/validation_images/step_350_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
301
+ vlcs_upscaled/model/validation_images/step_350_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
302
+ vlcs_upscaled/model/validation_images/step_400_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
303
+ vlcs_upscaled/model/validation_images/step_400_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
304
+ vlcs_upscaled/model/validation_images/step_450_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
305
+ vlcs_upscaled/model/validation_images/step_450_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
306
+ vlcs_upscaled/model/validation_images/step_500_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
307
+ vlcs_upscaled/model/validation_images/step_500_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
308
+ vlcs_upscaled/model/validation_images/step_50_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
309
+ vlcs_upscaled/model/validation_images/step_50_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
310
+ vlcs_upscaled/model/validation_images/step_550_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
311
+ vlcs_upscaled/model/validation_images/step_550_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
312
+ vlcs_upscaled/model/validation_images/step_600_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
313
+ vlcs_upscaled/model/validation_images/step_600_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
314
+ vlcs_upscaled/model/validation_images/step_650_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
315
+ vlcs_upscaled/model/validation_images/step_650_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
316
+ vlcs_upscaled/model/validation_images/step_700_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
317
+ vlcs_upscaled/model/validation_images/step_700_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
318
+ vlcs_upscaled/model/validation_images/step_750_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
319
+ vlcs_upscaled/model/validation_images/step_750_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
320
+ vlcs_upscaled/model/validation_images/step_800_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
321
+ vlcs_upscaled/model/validation_images/step_800_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
322
+ vlcs_upscaled/model/validation_images/step_850_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
323
+ vlcs_upscaled/model/validation_images/step_850_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
324
+ vlcs_upscaled/model/validation_images/step_900_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
325
+ vlcs_upscaled/model/validation_images/step_900_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
326
+ vlcs_upscaled/model/validation_images/step_950_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
327
+ vlcs_upscaled/model/validation_images/step_950_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
vlcs_upscaled/model/README.md ADDED
@@ -0,0 +1,132 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ base_model: "sd3/unknown-model"
4
+ tags:
5
+ - sd3
6
+ - sd3-diffusers
7
+ - text-to-image
8
+ - diffusers
9
+ - simpletuner
10
+ - not-for-all-audiences
11
+ - lora
12
+ - template:sd-lora
13
+ - standard
14
+ inference: true
15
+ widget:
16
+ - text: 'unconditional (blank prompt)'
17
+ parameters:
18
+ negative_prompt: 'blurry, cropped, ugly'
19
+ output:
20
+ url: ./assets/image_0_0.png
21
+ - text: 'A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant''s posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant.'
22
+ parameters:
23
+ negative_prompt: 'blurry, cropped, ugly'
24
+ output:
25
+ url: ./assets/image_1_0.png
26
+ ---
27
+
28
+ # simpletuner-lora
29
+
30
+ This is a standard PEFT LoRA derived from [sd3/unknown-model](https://huggingface.co/sd3/unknown-model).
31
+
32
+
33
+ The main validation prompt used during training was:
34
+ ```
35
+ A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant's posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant.
36
+ ```
37
+
38
+
39
+ ## Validation settings
40
+ - CFG: `7.5`
41
+ - CFG Rescale: `0.0`
42
+ - Steps: `35`
43
+ - Sampler: `FlowMatchEulerDiscreteScheduler`
44
+ - Seed: `42`
45
+ - Resolution: `512x512`
46
+ - Skip-layer guidance:
47
+
48
+ Note: The validation settings are not necessarily the same as the [training settings](#training-settings).
49
+
50
+ You can find some example images in the following gallery:
51
+
52
+
53
+ <Gallery />
54
+
55
+ The text encoder **was not** trained.
56
+ You may reuse the base model text encoder for inference.
57
+
58
+
59
+ ## Training settings
60
+
61
+ - Training epochs: 6
62
+ - Training steps: 3000
63
+ - Learning rate: 0.0001
64
+ - Learning rate schedule: cosine
65
+ - Warmup steps: 100
66
+ - Max grad norm: 2.0
67
+ - Effective batch size: 16
68
+ - Micro-batch size: 4
69
+ - Gradient accumulation steps: 4
70
+ - Number of GPUs: 1
71
+ - Gradient checkpointing: True
72
+ - Prediction type: flow-matching (extra parameters=['shift=3'])
73
+ - Optimizer: adamw_bf16
74
+ - Trainable parameter precision: Pure BF16
75
+ - Caption dropout probability: 10.0%
76
+
77
+
78
+ - LoRA Rank: 128
79
+ - LoRA Alpha: None
80
+ - LoRA Dropout: 0.1
81
+ - LoRA initialisation style: default
82
+
83
+
84
+ ## Datasets
85
+
86
+ ### pacs
87
+ - Repeats: 0
88
+ - Total number of images: 7680
89
+ - Total number of aspect buckets: 1
90
+ - Resolution: 1.0 megapixels
91
+ - Cropped: False
92
+ - Crop style: None
93
+ - Crop aspect: None
94
+ - Used for regularisation data: No
95
+
96
+
97
+ ## Inference
98
+
99
+
100
+ ```python
101
+ import torch
102
+ from diffusers import DiffusionPipeline
103
+
104
+ model_id = '/ephemeral/shashmi/llava_lets_go/chimaa_finetuner/stable-diffusion-3.5-medium'
105
+ adapter_id = 'Sarim-Hash/simpletuner-lora'
106
+ pipeline = DiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.bfloat16) # loading directly in bf16
107
+ pipeline.load_lora_weights(adapter_id)
108
+
109
+ prompt = "A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant's posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant."
110
+ negative_prompt = 'blurry, cropped, ugly'
111
+
112
+ ## Optional: quantise the model to save on vram.
113
+ ## Note: The model was quantised during training, and so it is recommended to do the same during inference time.
114
+ from optimum.quanto import quantize, freeze, qint8
115
+ quantize(pipeline.transformer, weights=qint8)
116
+ freeze(pipeline.transformer)
117
+
118
+ pipeline.to('cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu') # the pipeline is already in its target precision level
119
+ image = pipeline(
120
+ prompt=prompt,
121
+ negative_prompt=negative_prompt,
122
+ num_inference_steps=35,
123
+ generator=torch.Generator(device='cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu').manual_seed(42),
124
+ width=512,
125
+ height=512,
126
+ guidance_scale=7.5,
127
+ ).images[0]
128
+ image.save("output.png", format="PNG")
129
+ ```
130
+
131
+
132
+
vlcs_upscaled/model/all_image_files_pacs.json ADDED
The diff for this file is too large to render. See raw diff
 
vlcs_upscaled/model/all_text_cache_files_text-embeds.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {}
vlcs_upscaled/model/all_vae_cache_files_pacs.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {}
vlcs_upscaled/model/assets/image_0_0.png ADDED

Git LFS Details

  • SHA256: a81c244726e2fe2fe56d7360bfaa5aafb3023d44eec7e0f64b1cdf1af28d40f0
  • Pointer size: 131 Bytes
  • Size of remote file: 903 kB
vlcs_upscaled/model/assets/image_1_0.png ADDED

Git LFS Details

  • SHA256: 0f529f723fa92b805fc2635c35a2639aab6ad1cf65e17931bf759888611a1724
  • Pointer size: 131 Bytes
  • Size of remote file: 389 kB
vlcs_upscaled/model/benchmarks/base_model/unconditional_512x512.png ADDED

Git LFS Details

  • SHA256: fabd8007d625e53537564a8d3c4121fb79b8117fa8a837f033441b4c6ea69148
  • Pointer size: 131 Bytes
  • Size of remote file: 441 kB
vlcs_upscaled/model/benchmarks/base_model/validation_512x512.png ADDED

Git LFS Details

  • SHA256: 2c6c00eafc749ace816a52e01bc84dde4f4918c4071fba465c1c9283a3b7ff65
  • Pointer size: 131 Bytes
  • Size of remote file: 208 kB
vlcs_upscaled/model/checkpoint-1000/README.md ADDED
@@ -0,0 +1,132 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ base_model: "sd3/unknown-model"
4
+ tags:
5
+ - sd3
6
+ - sd3-diffusers
7
+ - text-to-image
8
+ - diffusers
9
+ - simpletuner
10
+ - not-for-all-audiences
11
+ - lora
12
+ - template:sd-lora
13
+ - standard
14
+ inference: true
15
+ widget:
16
+ - text: 'unconditional (blank prompt)'
17
+ parameters:
18
+ negative_prompt: 'blurry, cropped, ugly'
19
+ output:
20
+ url: ./assets/image_0_0.png
21
+ - text: 'A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant''s posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant.'
22
+ parameters:
23
+ negative_prompt: 'blurry, cropped, ugly'
24
+ output:
25
+ url: ./assets/image_1_0.png
26
+ ---
27
+
28
+ # simpletuner-lora
29
+
30
+ This is a standard PEFT LoRA derived from [sd3/unknown-model](https://huggingface.co/sd3/unknown-model).
31
+
32
+
33
+ The main validation prompt used during training was:
34
+ ```
35
+ A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant's posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant.
36
+ ```
37
+
38
+
39
+ ## Validation settings
40
+ - CFG: `7.5`
41
+ - CFG Rescale: `0.0`
42
+ - Steps: `35`
43
+ - Sampler: `FlowMatchEulerDiscreteScheduler`
44
+ - Seed: `42`
45
+ - Resolution: `512x512`
46
+ - Skip-layer guidance:
47
+
48
+ Note: The validation settings are not necessarily the same as the [training settings](#training-settings).
49
+
50
+ You can find some example images in the following gallery:
51
+
52
+
53
+ <Gallery />
54
+
55
+ The text encoder **was not** trained.
56
+ You may reuse the base model text encoder for inference.
57
+
58
+
59
+ ## Training settings
60
+
61
+ - Training epochs: 2
62
+ - Training steps: 1000
63
+ - Learning rate: 0.0001
64
+ - Learning rate schedule: cosine
65
+ - Warmup steps: 100
66
+ - Max grad norm: 2.0
67
+ - Effective batch size: 16
68
+ - Micro-batch size: 4
69
+ - Gradient accumulation steps: 4
70
+ - Number of GPUs: 1
71
+ - Gradient checkpointing: True
72
+ - Prediction type: flow-matching (extra parameters=['shift=3'])
73
+ - Optimizer: adamw_bf16
74
+ - Trainable parameter precision: Pure BF16
75
+ - Caption dropout probability: 10.0%
76
+
77
+
78
+ - LoRA Rank: 128
79
+ - LoRA Alpha: None
80
+ - LoRA Dropout: 0.1
81
+ - LoRA initialisation style: default
82
+
83
+
84
+ ## Datasets
85
+
86
+ ### pacs
87
+ - Repeats: 0
88
+ - Total number of images: 7680
89
+ - Total number of aspect buckets: 1
90
+ - Resolution: 1.0 megapixels
91
+ - Cropped: False
92
+ - Crop style: None
93
+ - Crop aspect: None
94
+ - Used for regularisation data: No
95
+
96
+
97
+ ## Inference
98
+
99
+
100
+ ```python
101
+ import torch
102
+ from diffusers import DiffusionPipeline
103
+
104
+ model_id = '/ephemeral/shashmi/llava_lets_go/chimaa_finetuner/stable-diffusion-3.5-medium'
105
+ adapter_id = 'Sarim-Hash/simpletuner-lora'
106
+ pipeline = DiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.bfloat16) # loading directly in bf16
107
+ pipeline.load_lora_weights(adapter_id)
108
+
109
+ prompt = "A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant's posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant."
110
+ negative_prompt = 'blurry, cropped, ugly'
111
+
112
+ ## Optional: quantise the model to save on vram.
113
+ ## Note: The model was quantised during training, and so it is recommended to do the same during inference time.
114
+ from optimum.quanto import quantize, freeze, qint8
115
+ quantize(pipeline.transformer, weights=qint8)
116
+ freeze(pipeline.transformer)
117
+
118
+ pipeline.to('cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu') # the pipeline is already in its target precision level
119
+ image = pipeline(
120
+ prompt=prompt,
121
+ negative_prompt=negative_prompt,
122
+ num_inference_steps=35,
123
+ generator=torch.Generator(device='cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu').manual_seed(42),
124
+ width=512,
125
+ height=512,
126
+ guidance_scale=7.5,
127
+ ).images[0]
128
+ image.save("output.png", format="PNG")
129
+ ```
130
+
131
+
132
+
vlcs_upscaled/model/checkpoint-1000/assets/image_0_0.png ADDED

Git LFS Details

  • SHA256: cc3a75a87e9281b697f42605b646855e6ba941cce2777aac681af431258dec4f
  • Pointer size: 131 Bytes
  • Size of remote file: 912 kB
vlcs_upscaled/model/checkpoint-1000/assets/image_1_0.png ADDED

Git LFS Details

  • SHA256: 866d6eb39080cc6d8685f8dfeba8dc7afc94430c5ee444c548f78419780c7c51
  • Pointer size: 131 Bytes
  • Size of remote file: 398 kB
vlcs_upscaled/model/checkpoint-1000/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:889d68b8772e7897888a4b3ce89f269e6b5318c534500638a1238b6088889582
3
+ size 349442426
vlcs_upscaled/model/checkpoint-1000/pytorch_lora_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0292ac0c32e7aec02b5fad98be33f643652f992f4d5b0c42d3307ccfafb79b3d
3
+ size 116431016
vlcs_upscaled/model/checkpoint-1000/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:865bbff9ce0789ab10f17862fc6b85da55fb8fd0bc263602232a38ee64766d5c
3
+ size 14408
vlcs_upscaled/model/checkpoint-1000/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b40c7683c794eb2d2c5f51a0d043e67c0aac67ed037d02fc10ed51860ad3226
3
+ size 1128
vlcs_upscaled/model/checkpoint-1000/training_state-pacs.json ADDED
The diff for this file is too large to render. See raw diff
 
vlcs_upscaled/model/checkpoint-1000/training_state.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"global_step": 1000, "epoch_step": 1000, "epoch": 3, "exhausted_backends": [], "repeats": {"pacs": 0}}
vlcs_upscaled/model/checkpoint-1250/README.md ADDED
@@ -0,0 +1,132 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ base_model: "sd3/unknown-model"
4
+ tags:
5
+ - sd3
6
+ - sd3-diffusers
7
+ - text-to-image
8
+ - diffusers
9
+ - simpletuner
10
+ - not-for-all-audiences
11
+ - lora
12
+ - template:sd-lora
13
+ - standard
14
+ inference: true
15
+ widget:
16
+ - text: 'unconditional (blank prompt)'
17
+ parameters:
18
+ negative_prompt: 'blurry, cropped, ugly'
19
+ output:
20
+ url: ./assets/image_0_0.png
21
+ - text: 'A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant''s posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant.'
22
+ parameters:
23
+ negative_prompt: 'blurry, cropped, ugly'
24
+ output:
25
+ url: ./assets/image_1_0.png
26
+ ---
27
+
28
+ # simpletuner-lora
29
+
30
+ This is a standard PEFT LoRA derived from [sd3/unknown-model](https://huggingface.co/sd3/unknown-model).
31
+
32
+
33
+ The main validation prompt used during training was:
34
+ ```
35
+ A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant's posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant.
36
+ ```
37
+
38
+
39
+ ## Validation settings
40
+ - CFG: `7.5`
41
+ - CFG Rescale: `0.0`
42
+ - Steps: `35`
43
+ - Sampler: `FlowMatchEulerDiscreteScheduler`
44
+ - Seed: `42`
45
+ - Resolution: `512x512`
46
+ - Skip-layer guidance:
47
+
48
+ Note: The validation settings are not necessarily the same as the [training settings](#training-settings).
49
+
50
+ You can find some example images in the following gallery:
51
+
52
+
53
+ <Gallery />
54
+
55
+ The text encoder **was not** trained.
56
+ You may reuse the base model text encoder for inference.
57
+
58
+
59
+ ## Training settings
60
+
61
+ - Training epochs: 2
62
+ - Training steps: 1250
63
+ - Learning rate: 0.0001
64
+ - Learning rate schedule: cosine
65
+ - Warmup steps: 100
66
+ - Max grad norm: 2.0
67
+ - Effective batch size: 16
68
+ - Micro-batch size: 4
69
+ - Gradient accumulation steps: 4
70
+ - Number of GPUs: 1
71
+ - Gradient checkpointing: True
72
+ - Prediction type: flow-matching (extra parameters=['shift=3'])
73
+ - Optimizer: adamw_bf16
74
+ - Trainable parameter precision: Pure BF16
75
+ - Caption dropout probability: 10.0%
76
+
77
+
78
+ - LoRA Rank: 128
79
+ - LoRA Alpha: None
80
+ - LoRA Dropout: 0.1
81
+ - LoRA initialisation style: default
82
+
83
+
84
+ ## Datasets
85
+
86
+ ### pacs
87
+ - Repeats: 0
88
+ - Total number of images: 7680
89
+ - Total number of aspect buckets: 1
90
+ - Resolution: 1.0 megapixels
91
+ - Cropped: False
92
+ - Crop style: None
93
+ - Crop aspect: None
94
+ - Used for regularisation data: No
95
+
96
+
97
+ ## Inference
98
+
99
+
100
+ ```python
101
+ import torch
102
+ from diffusers import DiffusionPipeline
103
+
104
+ model_id = '/ephemeral/shashmi/llava_lets_go/chimaa_finetuner/stable-diffusion-3.5-medium'
105
+ adapter_id = 'Sarim-Hash/simpletuner-lora'
106
+ pipeline = DiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.bfloat16) # loading directly in bf16
107
+ pipeline.load_lora_weights(adapter_id)
108
+
109
+ prompt = "A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant's posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant."
110
+ negative_prompt = 'blurry, cropped, ugly'
111
+
112
+ ## Optional: quantise the model to save on vram.
113
+ ## Note: The model was quantised during training, and so it is recommended to do the same during inference time.
114
+ from optimum.quanto import quantize, freeze, qint8
115
+ quantize(pipeline.transformer, weights=qint8)
116
+ freeze(pipeline.transformer)
117
+
118
+ pipeline.to('cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu') # the pipeline is already in its target precision level
119
+ image = pipeline(
120
+ prompt=prompt,
121
+ negative_prompt=negative_prompt,
122
+ num_inference_steps=35,
123
+ generator=torch.Generator(device='cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu').manual_seed(42),
124
+ width=512,
125
+ height=512,
126
+ guidance_scale=7.5,
127
+ ).images[0]
128
+ image.save("output.png", format="PNG")
129
+ ```
130
+
131
+
132
+
vlcs_upscaled/model/checkpoint-1250/assets/image_0_0.png ADDED

Git LFS Details

  • SHA256: fe7240ffad28f0135df2ed454d60d2aa9cfca43b13b09e492a1ca62f2e487ba7
  • Pointer size: 131 Bytes
  • Size of remote file: 908 kB
vlcs_upscaled/model/checkpoint-1250/assets/image_1_0.png ADDED

Git LFS Details

  • SHA256: 6905810598cbb2baee2f3a3bb4e93251bf5945a6e4b8cde5ae61144a1d21a0c4
  • Pointer size: 131 Bytes
  • Size of remote file: 383 kB
vlcs_upscaled/model/checkpoint-1250/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:baefc5dcceeceb30e54adb2a5b86184251e23a68386a6d13a1cfc4ef6bc82470
3
+ size 349442426
vlcs_upscaled/model/checkpoint-1250/pytorch_lora_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d48056e7a5d32122b81672db8c1777b8d9491ca88348e98072f25b5a1f5170dc
3
+ size 116431016
vlcs_upscaled/model/checkpoint-1250/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b133853f0107ce4e5268fcb5ae0d264f4db8d4bd1931e6f9ae6a6471ce7e5513
3
+ size 14344
vlcs_upscaled/model/checkpoint-1250/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f0396e8354db804fd5f3b80b54b57a56511fa235d6226728e6557cfe331c1e6
3
+ size 1128
vlcs_upscaled/model/checkpoint-1250/training_state-pacs.json ADDED
The diff for this file is too large to render. See raw diff
 
vlcs_upscaled/model/checkpoint-1250/training_state.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"global_step": 1250, "epoch_step": 1250, "epoch": 3, "exhausted_backends": [], "repeats": {"pacs": 0}}
vlcs_upscaled/model/checkpoint-1500/README.md ADDED
@@ -0,0 +1,132 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ base_model: "sd3/unknown-model"
4
+ tags:
5
+ - sd3
6
+ - sd3-diffusers
7
+ - text-to-image
8
+ - diffusers
9
+ - simpletuner
10
+ - not-for-all-audiences
11
+ - lora
12
+ - template:sd-lora
13
+ - standard
14
+ inference: true
15
+ widget:
16
+ - text: 'unconditional (blank prompt)'
17
+ parameters:
18
+ negative_prompt: 'blurry, cropped, ugly'
19
+ output:
20
+ url: ./assets/image_0_0.png
21
+ - text: 'A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant''s posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant.'
22
+ parameters:
23
+ negative_prompt: 'blurry, cropped, ugly'
24
+ output:
25
+ url: ./assets/image_1_0.png
26
+ ---
27
+
28
+ # simpletuner-lora
29
+
30
+ This is a standard PEFT LoRA derived from [sd3/unknown-model](https://huggingface.co/sd3/unknown-model).
31
+
32
+
33
+ The main validation prompt used during training was:
34
+ ```
35
+ A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant's posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant.
36
+ ```
37
+
38
+
39
+ ## Validation settings
40
+ - CFG: `7.5`
41
+ - CFG Rescale: `0.0`
42
+ - Steps: `35`
43
+ - Sampler: `FlowMatchEulerDiscreteScheduler`
44
+ - Seed: `42`
45
+ - Resolution: `512x512`
46
+ - Skip-layer guidance:
47
+
48
+ Note: The validation settings are not necessarily the same as the [training settings](#training-settings).
49
+
50
+ You can find some example images in the following gallery:
51
+
52
+
53
+ <Gallery />
54
+
55
+ The text encoder **was not** trained.
56
+ You may reuse the base model text encoder for inference.
57
+
58
+
59
+ ## Training settings
60
+
61
+ - Training epochs: 3
62
+ - Training steps: 1500
63
+ - Learning rate: 0.0001
64
+ - Learning rate schedule: cosine
65
+ - Warmup steps: 100
66
+ - Max grad norm: 2.0
67
+ - Effective batch size: 16
68
+ - Micro-batch size: 4
69
+ - Gradient accumulation steps: 4
70
+ - Number of GPUs: 1
71
+ - Gradient checkpointing: True
72
+ - Prediction type: flow-matching (extra parameters=['shift=3'])
73
+ - Optimizer: adamw_bf16
74
+ - Trainable parameter precision: Pure BF16
75
+ - Caption dropout probability: 10.0%
76
+
77
+
78
+ - LoRA Rank: 128
79
+ - LoRA Alpha: None
80
+ - LoRA Dropout: 0.1
81
+ - LoRA initialisation style: default
82
+
83
+
84
+ ## Datasets
85
+
86
+ ### pacs
87
+ - Repeats: 0
88
+ - Total number of images: 7680
89
+ - Total number of aspect buckets: 1
90
+ - Resolution: 1.0 megapixels
91
+ - Cropped: False
92
+ - Crop style: None
93
+ - Crop aspect: None
94
+ - Used for regularisation data: No
95
+
96
+
97
+ ## Inference
98
+
99
+
100
+ ```python
101
+ import torch
102
+ from diffusers import DiffusionPipeline
103
+
104
+ model_id = '/ephemeral/shashmi/llava_lets_go/chimaa_finetuner/stable-diffusion-3.5-medium'
105
+ adapter_id = 'Sarim-Hash/simpletuner-lora'
106
+ pipeline = DiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.bfloat16) # loading directly in bf16
107
+ pipeline.load_lora_weights(adapter_id)
108
+
109
+ prompt = "A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant's posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant."
110
+ negative_prompt = 'blurry, cropped, ugly'
111
+
112
+ ## Optional: quantise the model to save on vram.
113
+ ## Note: The model was quantised during training, and so it is recommended to do the same during inference time.
114
+ from optimum.quanto import quantize, freeze, qint8
115
+ quantize(pipeline.transformer, weights=qint8)
116
+ freeze(pipeline.transformer)
117
+
118
+ pipeline.to('cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu') # the pipeline is already in its target precision level
119
+ image = pipeline(
120
+ prompt=prompt,
121
+ negative_prompt=negative_prompt,
122
+ num_inference_steps=35,
123
+ generator=torch.Generator(device='cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu').manual_seed(42),
124
+ width=512,
125
+ height=512,
126
+ guidance_scale=7.5,
127
+ ).images[0]
128
+ image.save("output.png", format="PNG")
129
+ ```
130
+
131
+
132
+
vlcs_upscaled/model/checkpoint-1500/assets/image_0_0.png ADDED

Git LFS Details

  • SHA256: d6e330a5181a2672f120515f17c90347ffde104e06b46f59062a652bc47a4461
  • Pointer size: 131 Bytes
  • Size of remote file: 907 kB
vlcs_upscaled/model/checkpoint-1500/assets/image_1_0.png ADDED

Git LFS Details

  • SHA256: bd6294f0e24896ced254c781c80a4fa2faf6bc44fead8e62e50d9acf2e010d3d
  • Pointer size: 131 Bytes
  • Size of remote file: 389 kB
vlcs_upscaled/model/checkpoint-1500/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a5203ef3b99bc48123d15f33d2a0757d980b0815b62c80d06398ae2a140894f
3
+ size 349442426
vlcs_upscaled/model/checkpoint-1500/pytorch_lora_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1d9b454f598ceb4258ad2216ec64424f3d12ea8ba908c35775435edabd7f320
3
+ size 116431016
vlcs_upscaled/model/checkpoint-1500/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ee22891055186d041691f54aa4b7126c693794c535050976985a128ba2678e2
3
+ size 14344
vlcs_upscaled/model/checkpoint-1500/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92c21aba5105638209ef13ea982d9ffda2517815d25498abf24fedadcdeec846
3
+ size 1128
vlcs_upscaled/model/checkpoint-1500/training_state-pacs.json ADDED
The diff for this file is too large to render. See raw diff
 
vlcs_upscaled/model/checkpoint-1500/training_state.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"global_step": 1500, "epoch_step": 1500, "epoch": 4, "exhausted_backends": [], "repeats": {"pacs": 0}}
vlcs_upscaled/model/checkpoint-1750/README.md ADDED
@@ -0,0 +1,132 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ base_model: "sd3/unknown-model"
4
+ tags:
5
+ - sd3
6
+ - sd3-diffusers
7
+ - text-to-image
8
+ - diffusers
9
+ - simpletuner
10
+ - not-for-all-audiences
11
+ - lora
12
+ - template:sd-lora
13
+ - standard
14
+ inference: true
15
+ widget:
16
+ - text: 'unconditional (blank prompt)'
17
+ parameters:
18
+ negative_prompt: 'blurry, cropped, ugly'
19
+ output:
20
+ url: ./assets/image_0_0.png
21
+ - text: 'A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant''s posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant.'
22
+ parameters:
23
+ negative_prompt: 'blurry, cropped, ugly'
24
+ output:
25
+ url: ./assets/image_1_0.png
26
+ ---
27
+
28
+ # simpletuner-lora
29
+
30
+ This is a standard PEFT LoRA derived from [sd3/unknown-model](https://huggingface.co/sd3/unknown-model).
31
+
32
+
33
+ The main validation prompt used during training was:
34
+ ```
35
+ A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant's posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant.
36
+ ```
37
+
38
+
39
+ ## Validation settings
40
+ - CFG: `7.5`
41
+ - CFG Rescale: `0.0`
42
+ - Steps: `35`
43
+ - Sampler: `FlowMatchEulerDiscreteScheduler`
44
+ - Seed: `42`
45
+ - Resolution: `512x512`
46
+ - Skip-layer guidance:
47
+
48
+ Note: The validation settings are not necessarily the same as the [training settings](#training-settings).
49
+
50
+ You can find some example images in the following gallery:
51
+
52
+
53
+ <Gallery />
54
+
55
+ The text encoder **was not** trained.
56
+ You may reuse the base model text encoder for inference.
57
+
58
+
59
+ ## Training settings
60
+
61
+ - Training epochs: 3
62
+ - Training steps: 1750
63
+ - Learning rate: 0.0001
64
+ - Learning rate schedule: cosine
65
+ - Warmup steps: 100
66
+ - Max grad norm: 2.0
67
+ - Effective batch size: 16
68
+ - Micro-batch size: 4
69
+ - Gradient accumulation steps: 4
70
+ - Number of GPUs: 1
71
+ - Gradient checkpointing: True
72
+ - Prediction type: flow-matching (extra parameters=['shift=3'])
73
+ - Optimizer: adamw_bf16
74
+ - Trainable parameter precision: Pure BF16
75
+ - Caption dropout probability: 10.0%
76
+
77
+
78
+ - LoRA Rank: 128
79
+ - LoRA Alpha: None
80
+ - LoRA Dropout: 0.1
81
+ - LoRA initialisation style: default
82
+
83
+
84
+ ## Datasets
85
+
86
+ ### pacs
87
+ - Repeats: 0
88
+ - Total number of images: 7680
89
+ - Total number of aspect buckets: 1
90
+ - Resolution: 1.0 megapixels
91
+ - Cropped: False
92
+ - Crop style: None
93
+ - Crop aspect: None
94
+ - Used for regularisation data: No
95
+
96
+
97
+ ## Inference
98
+
99
+
100
+ ```python
101
+ import torch
102
+ from diffusers import DiffusionPipeline
103
+
104
+ model_id = '/ephemeral/shashmi/llava_lets_go/chimaa_finetuner/stable-diffusion-3.5-medium'
105
+ adapter_id = 'Sarim-Hash/simpletuner-lora'
106
+ pipeline = DiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.bfloat16) # loading directly in bf16
107
+ pipeline.load_lora_weights(adapter_id)
108
+
109
+ prompt = "A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant's posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant."
110
+ negative_prompt = 'blurry, cropped, ugly'
111
+
112
+ ## Optional: quantise the model to save on vram.
113
+ ## Note: The model was quantised during training, and so it is recommended to do the same during inference time.
114
+ from optimum.quanto import quantize, freeze, qint8
115
+ quantize(pipeline.transformer, weights=qint8)
116
+ freeze(pipeline.transformer)
117
+
118
+ pipeline.to('cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu') # the pipeline is already in its target precision level
119
+ image = pipeline(
120
+ prompt=prompt,
121
+ negative_prompt=negative_prompt,
122
+ num_inference_steps=35,
123
+ generator=torch.Generator(device='cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu').manual_seed(42),
124
+ width=512,
125
+ height=512,
126
+ guidance_scale=7.5,
127
+ ).images[0]
128
+ image.save("output.png", format="PNG")
129
+ ```
130
+
131
+
132
+
vlcs_upscaled/model/checkpoint-1750/assets/image_0_0.png ADDED

Git LFS Details

  • SHA256: 09538a7a73837dbc6fa1a9f2c7722fb372cf55e203bb08d1eed7c84adae39de0
  • Pointer size: 131 Bytes
  • Size of remote file: 912 kB
vlcs_upscaled/model/checkpoint-1750/assets/image_1_0.png ADDED

Git LFS Details

  • SHA256: b193a1086bacbb927cc59b74776b96ed131f9016d5233be875e535617194ee1e
  • Pointer size: 131 Bytes
  • Size of remote file: 396 kB
vlcs_upscaled/model/checkpoint-1750/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:29680a012e5a8ed1a79a0ad8ae6b2a9e25131a337dc55b298c65bd308204d15d
3
+ size 349442426
vlcs_upscaled/model/checkpoint-1750/pytorch_lora_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a326e89559656ae8b8689c1f2d4cd0f76c892fe840bd6d8381776c28cf59aa78
3
+ size 116431016
vlcs_upscaled/model/checkpoint-1750/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:62cdc2b3120ca6f1f466356bccfb50819cb5137e0c2b90b1d4aa259593342f5d
3
+ size 14344
vlcs_upscaled/model/checkpoint-1750/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2922907515a7eda3d4adff43a1e668fe04b9659067a29ebea923dbbcf57043c
3
+ size 1128
vlcs_upscaled/model/checkpoint-1750/training_state-pacs.json ADDED
The diff for this file is too large to render. See raw diff
 
vlcs_upscaled/model/checkpoint-1750/training_state.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"global_step": 1750, "epoch_step": 1750, "epoch": 4, "exhausted_backends": [], "repeats": {"pacs": 0}}
vlcs_upscaled/model/checkpoint-2000/README.md ADDED
@@ -0,0 +1,132 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ base_model: "sd3/unknown-model"
4
+ tags:
5
+ - sd3
6
+ - sd3-diffusers
7
+ - text-to-image
8
+ - diffusers
9
+ - simpletuner
10
+ - not-for-all-audiences
11
+ - lora
12
+ - template:sd-lora
13
+ - standard
14
+ inference: true
15
+ widget:
16
+ - text: 'unconditional (blank prompt)'
17
+ parameters:
18
+ negative_prompt: 'blurry, cropped, ugly'
19
+ output:
20
+ url: ./assets/image_0_0.png
21
+ - text: 'A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant''s posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant.'
22
+ parameters:
23
+ negative_prompt: 'blurry, cropped, ugly'
24
+ output:
25
+ url: ./assets/image_1_0.png
26
+ ---
27
+
28
+ # simpletuner-lora
29
+
30
+ This is a standard PEFT LoRA derived from [sd3/unknown-model](https://huggingface.co/sd3/unknown-model).
31
+
32
+
33
+ The main validation prompt used during training was:
34
+ ```
35
+ A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant's posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant.
36
+ ```
37
+
38
+
39
+ ## Validation settings
40
+ - CFG: `7.5`
41
+ - CFG Rescale: `0.0`
42
+ - Steps: `35`
43
+ - Sampler: `FlowMatchEulerDiscreteScheduler`
44
+ - Seed: `42`
45
+ - Resolution: `512x512`
46
+ - Skip-layer guidance:
47
+
48
+ Note: The validation settings are not necessarily the same as the [training settings](#training-settings).
49
+
50
+ You can find some example images in the following gallery:
51
+
52
+
53
+ <Gallery />
54
+
55
+ The text encoder **was not** trained.
56
+ You may reuse the base model text encoder for inference.
57
+
58
+
59
+ ## Training settings
60
+
61
+ - Training epochs: 4
62
+ - Training steps: 2000
63
+ - Learning rate: 0.0001
64
+ - Learning rate schedule: cosine
65
+ - Warmup steps: 100
66
+ - Max grad norm: 2.0
67
+ - Effective batch size: 16
68
+ - Micro-batch size: 4
69
+ - Gradient accumulation steps: 4
70
+ - Number of GPUs: 1
71
+ - Gradient checkpointing: True
72
+ - Prediction type: flow-matching (extra parameters=['shift=3'])
73
+ - Optimizer: adamw_bf16
74
+ - Trainable parameter precision: Pure BF16
75
+ - Caption dropout probability: 10.0%
76
+
77
+
78
+ - LoRA Rank: 128
79
+ - LoRA Alpha: None
80
+ - LoRA Dropout: 0.1
81
+ - LoRA initialisation style: default
82
+
83
+
84
+ ## Datasets
85
+
86
+ ### pacs
87
+ - Repeats: 0
88
+ - Total number of images: 7680
89
+ - Total number of aspect buckets: 1
90
+ - Resolution: 1.0 megapixels
91
+ - Cropped: False
92
+ - Crop style: None
93
+ - Crop aspect: None
94
+ - Used for regularisation data: No
95
+
96
+
97
+ ## Inference
98
+
99
+
100
+ ```python
101
+ import torch
102
+ from diffusers import DiffusionPipeline
103
+
104
+ model_id = '/ephemeral/shashmi/llava_lets_go/chimaa_finetuner/stable-diffusion-3.5-medium'
105
+ adapter_id = 'Sarim-Hash/simpletuner-lora'
106
+ pipeline = DiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.bfloat16) # loading directly in bf16
107
+ pipeline.load_lora_weights(adapter_id)
108
+
109
+ prompt = "A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant's posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant."
110
+ negative_prompt = 'blurry, cropped, ugly'
111
+
112
+ ## Optional: quantise the model to save on vram.
113
+ ## Note: The model was quantised during training, and so it is recommended to do the same during inference time.
114
+ from optimum.quanto import quantize, freeze, qint8
115
+ quantize(pipeline.transformer, weights=qint8)
116
+ freeze(pipeline.transformer)
117
+
118
+ pipeline.to('cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu') # the pipeline is already in its target precision level
119
+ image = pipeline(
120
+ prompt=prompt,
121
+ negative_prompt=negative_prompt,
122
+ num_inference_steps=35,
123
+ generator=torch.Generator(device='cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu').manual_seed(42),
124
+ width=512,
125
+ height=512,
126
+ guidance_scale=7.5,
127
+ ).images[0]
128
+ image.save("output.png", format="PNG")
129
+ ```
130
+
131
+
132
+
vlcs_upscaled/model/checkpoint-2000/assets/image_0_0.png ADDED

Git LFS Details

  • SHA256: 3a27fdd988cc37a69a18f0c2929ab3dc3314cd16be124f935a7e6bb3ecd10b0f
  • Pointer size: 131 Bytes
  • Size of remote file: 911 kB
vlcs_upscaled/model/checkpoint-2000/assets/image_1_0.png ADDED

Git LFS Details

  • SHA256: 7e570feb5c76bcec897bce8cd6c3d4691ad5ee38307598ace2a46e5b77181183
  • Pointer size: 131 Bytes
  • Size of remote file: 392 kB
vlcs_upscaled/model/checkpoint-2000/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:752bd81543f096a119a6ce8b5aa79abc98ea4cc4270ff7fb733097a7880a4995
3
+ size 349442426
vlcs_upscaled/model/checkpoint-2000/pytorch_lora_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e853f34851f4a239f7f910f96fedaa04fb78e5ec02e9ef63ab8f61bc12ffe04
3
+ size 116431016