Sarim-Hash commited on
Commit
69ad353
·
verified ·
1 Parent(s): ac05801

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +146 -0
  2. .gitignore +1 -0
  3. digit_upscaled/README.md +132 -0
  4. digit_upscaled/all_image_files_pacs.json +0 -0
  5. digit_upscaled/all_text_cache_files_text-embeds.json +0 -0
  6. digit_upscaled/all_vae_cache_files_pacs.json +0 -0
  7. digit_upscaled/assets/image_0_0.png +3 -0
  8. digit_upscaled/assets/image_1_0.png +3 -0
  9. digit_upscaled/benchmarks/base_model/unconditional_512x512.png +3 -0
  10. digit_upscaled/benchmarks/base_model/validation_512x512.png +3 -0
  11. digit_upscaled/checkpoint-1000/README.md +132 -0
  12. digit_upscaled/checkpoint-1000/assets/image_0_0.png +3 -0
  13. digit_upscaled/checkpoint-1000/assets/image_1_0.png +3 -0
  14. digit_upscaled/checkpoint-1000/optimizer.bin +3 -0
  15. digit_upscaled/checkpoint-1000/pytorch_lora_weights.safetensors +3 -0
  16. digit_upscaled/checkpoint-1000/random_states_0.pkl +3 -0
  17. digit_upscaled/checkpoint-1000/scheduler.bin +3 -0
  18. digit_upscaled/checkpoint-1000/training_state-pacs.json +0 -0
  19. digit_upscaled/checkpoint-1000/training_state.json +1 -0
  20. digit_upscaled/checkpoint-1250/README.md +132 -0
  21. digit_upscaled/checkpoint-1250/assets/image_0_0.png +3 -0
  22. digit_upscaled/checkpoint-1250/assets/image_1_0.png +3 -0
  23. digit_upscaled/checkpoint-1250/optimizer.bin +3 -0
  24. digit_upscaled/checkpoint-1250/pytorch_lora_weights.safetensors +3 -0
  25. digit_upscaled/checkpoint-1250/random_states_0.pkl +3 -0
  26. digit_upscaled/checkpoint-1250/scheduler.bin +3 -0
  27. digit_upscaled/checkpoint-1250/training_state-pacs.json +0 -0
  28. digit_upscaled/checkpoint-1250/training_state.json +1 -0
  29. digit_upscaled/checkpoint-1500/README.md +132 -0
  30. digit_upscaled/checkpoint-1500/assets/image_0_0.png +3 -0
  31. digit_upscaled/checkpoint-1500/assets/image_1_0.png +3 -0
  32. digit_upscaled/checkpoint-1500/optimizer.bin +3 -0
  33. digit_upscaled/checkpoint-1500/pytorch_lora_weights.safetensors +3 -0
  34. digit_upscaled/checkpoint-1500/random_states_0.pkl +3 -0
  35. digit_upscaled/checkpoint-1500/scheduler.bin +3 -0
  36. digit_upscaled/checkpoint-1500/training_state-pacs.json +0 -0
  37. digit_upscaled/checkpoint-1500/training_state.json +1 -0
  38. digit_upscaled/checkpoint-1750/README.md +132 -0
  39. digit_upscaled/checkpoint-1750/assets/image_0_0.png +3 -0
  40. digit_upscaled/checkpoint-1750/assets/image_1_0.png +3 -0
  41. digit_upscaled/checkpoint-1750/optimizer.bin +3 -0
  42. digit_upscaled/checkpoint-1750/pytorch_lora_weights.safetensors +3 -0
  43. digit_upscaled/checkpoint-1750/random_states_0.pkl +3 -0
  44. digit_upscaled/checkpoint-1750/scheduler.bin +3 -0
  45. digit_upscaled/checkpoint-1750/training_state-pacs.json +0 -0
  46. digit_upscaled/checkpoint-1750/training_state.json +1 -0
  47. digit_upscaled/checkpoint-2000/README.md +132 -0
  48. digit_upscaled/checkpoint-2000/assets/image_0_0.png +3 -0
  49. digit_upscaled/checkpoint-2000/assets/image_1_0.png +3 -0
  50. digit_upscaled/checkpoint-2000/optimizer.bin +3 -0
.gitattributes CHANGED
@@ -33,3 +33,149 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ digit_upscaled/assets/image_0_0.png filter=lfs diff=lfs merge=lfs -text
37
+ digit_upscaled/assets/image_1_0.png filter=lfs diff=lfs merge=lfs -text
38
+ digit_upscaled/benchmarks/base_model/unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
39
+ digit_upscaled/benchmarks/base_model/validation_512x512.png filter=lfs diff=lfs merge=lfs -text
40
+ digit_upscaled/checkpoint-1000/assets/image_0_0.png filter=lfs diff=lfs merge=lfs -text
41
+ digit_upscaled/checkpoint-1000/assets/image_1_0.png filter=lfs diff=lfs merge=lfs -text
42
+ digit_upscaled/checkpoint-1250/assets/image_0_0.png filter=lfs diff=lfs merge=lfs -text
43
+ digit_upscaled/checkpoint-1250/assets/image_1_0.png filter=lfs diff=lfs merge=lfs -text
44
+ digit_upscaled/checkpoint-1500/assets/image_0_0.png filter=lfs diff=lfs merge=lfs -text
45
+ digit_upscaled/checkpoint-1500/assets/image_1_0.png filter=lfs diff=lfs merge=lfs -text
46
+ digit_upscaled/checkpoint-1750/assets/image_0_0.png filter=lfs diff=lfs merge=lfs -text
47
+ digit_upscaled/checkpoint-1750/assets/image_1_0.png filter=lfs diff=lfs merge=lfs -text
48
+ digit_upscaled/checkpoint-2000/assets/image_0_0.png filter=lfs diff=lfs merge=lfs -text
49
+ digit_upscaled/checkpoint-2000/assets/image_1_0.png filter=lfs diff=lfs merge=lfs -text
50
+ digit_upscaled/checkpoint-2250/assets/image_0_0.png filter=lfs diff=lfs merge=lfs -text
51
+ digit_upscaled/checkpoint-2250/assets/image_1_0.png filter=lfs diff=lfs merge=lfs -text
52
+ digit_upscaled/checkpoint-250/assets/image_0_0.png filter=lfs diff=lfs merge=lfs -text
53
+ digit_upscaled/checkpoint-250/assets/image_1_0.png filter=lfs diff=lfs merge=lfs -text
54
+ digit_upscaled/checkpoint-2500/assets/image_0_0.png filter=lfs diff=lfs merge=lfs -text
55
+ digit_upscaled/checkpoint-2500/assets/image_1_0.png filter=lfs diff=lfs merge=lfs -text
56
+ digit_upscaled/checkpoint-2750/assets/image_0_0.png filter=lfs diff=lfs merge=lfs -text
57
+ digit_upscaled/checkpoint-2750/assets/image_1_0.png filter=lfs diff=lfs merge=lfs -text
58
+ digit_upscaled/checkpoint-500/assets/image_0_0.png filter=lfs diff=lfs merge=lfs -text
59
+ digit_upscaled/checkpoint-500/assets/image_1_0.png filter=lfs diff=lfs merge=lfs -text
60
+ digit_upscaled/checkpoint-750/assets/image_0_0.png filter=lfs diff=lfs merge=lfs -text
61
+ digit_upscaled/checkpoint-750/assets/image_1_0.png filter=lfs diff=lfs merge=lfs -text
62
+ digit_upscaled/validation_images/step_0_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
63
+ digit_upscaled/validation_images/step_0_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
64
+ digit_upscaled/validation_images/step_1000_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
65
+ digit_upscaled/validation_images/step_1000_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
66
+ digit_upscaled/validation_images/step_100_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
67
+ digit_upscaled/validation_images/step_100_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
68
+ digit_upscaled/validation_images/step_1050_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
69
+ digit_upscaled/validation_images/step_1050_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
70
+ digit_upscaled/validation_images/step_1100_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
71
+ digit_upscaled/validation_images/step_1100_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
72
+ digit_upscaled/validation_images/step_1150_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
73
+ digit_upscaled/validation_images/step_1150_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
74
+ digit_upscaled/validation_images/step_1200_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
75
+ digit_upscaled/validation_images/step_1200_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
76
+ digit_upscaled/validation_images/step_1250_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
77
+ digit_upscaled/validation_images/step_1250_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
78
+ digit_upscaled/validation_images/step_1300_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
79
+ digit_upscaled/validation_images/step_1300_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
80
+ digit_upscaled/validation_images/step_1350_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
81
+ digit_upscaled/validation_images/step_1350_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
82
+ digit_upscaled/validation_images/step_1400_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
83
+ digit_upscaled/validation_images/step_1400_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
84
+ digit_upscaled/validation_images/step_1450_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
85
+ digit_upscaled/validation_images/step_1450_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
86
+ digit_upscaled/validation_images/step_1500_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
87
+ digit_upscaled/validation_images/step_1500_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
88
+ digit_upscaled/validation_images/step_150_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
89
+ digit_upscaled/validation_images/step_150_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
90
+ digit_upscaled/validation_images/step_1550_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
91
+ digit_upscaled/validation_images/step_1550_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
92
+ digit_upscaled/validation_images/step_1600_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
93
+ digit_upscaled/validation_images/step_1600_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
94
+ digit_upscaled/validation_images/step_1650_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
95
+ digit_upscaled/validation_images/step_1650_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
96
+ digit_upscaled/validation_images/step_1700_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
97
+ digit_upscaled/validation_images/step_1700_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
98
+ digit_upscaled/validation_images/step_1750_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
99
+ digit_upscaled/validation_images/step_1750_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
100
+ digit_upscaled/validation_images/step_1800_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
101
+ digit_upscaled/validation_images/step_1800_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
102
+ digit_upscaled/validation_images/step_1850_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
103
+ digit_upscaled/validation_images/step_1850_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
104
+ digit_upscaled/validation_images/step_1900_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
105
+ digit_upscaled/validation_images/step_1900_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
106
+ digit_upscaled/validation_images/step_1950_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
107
+ digit_upscaled/validation_images/step_1950_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
108
+ digit_upscaled/validation_images/step_2000_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
109
+ digit_upscaled/validation_images/step_2000_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
110
+ digit_upscaled/validation_images/step_200_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
111
+ digit_upscaled/validation_images/step_200_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
112
+ digit_upscaled/validation_images/step_2050_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
113
+ digit_upscaled/validation_images/step_2050_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
114
+ digit_upscaled/validation_images/step_2100_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
115
+ digit_upscaled/validation_images/step_2100_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
116
+ digit_upscaled/validation_images/step_2150_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
117
+ digit_upscaled/validation_images/step_2150_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
118
+ digit_upscaled/validation_images/step_2200_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
119
+ digit_upscaled/validation_images/step_2200_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
120
+ digit_upscaled/validation_images/step_2250_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
121
+ digit_upscaled/validation_images/step_2250_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
122
+ digit_upscaled/validation_images/step_2300_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
123
+ digit_upscaled/validation_images/step_2300_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
124
+ digit_upscaled/validation_images/step_2350_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
125
+ digit_upscaled/validation_images/step_2350_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
126
+ digit_upscaled/validation_images/step_2400_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
127
+ digit_upscaled/validation_images/step_2400_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
128
+ digit_upscaled/validation_images/step_2450_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
129
+ digit_upscaled/validation_images/step_2450_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
130
+ digit_upscaled/validation_images/step_2500_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
131
+ digit_upscaled/validation_images/step_2500_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
132
+ digit_upscaled/validation_images/step_250_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
133
+ digit_upscaled/validation_images/step_250_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
134
+ digit_upscaled/validation_images/step_2550_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
135
+ digit_upscaled/validation_images/step_2550_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
136
+ digit_upscaled/validation_images/step_2600_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
137
+ digit_upscaled/validation_images/step_2600_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
138
+ digit_upscaled/validation_images/step_2650_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
139
+ digit_upscaled/validation_images/step_2650_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
140
+ digit_upscaled/validation_images/step_2700_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
141
+ digit_upscaled/validation_images/step_2700_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
142
+ digit_upscaled/validation_images/step_2750_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
143
+ digit_upscaled/validation_images/step_2750_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
144
+ digit_upscaled/validation_images/step_2800_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
145
+ digit_upscaled/validation_images/step_2800_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
146
+ digit_upscaled/validation_images/step_2850_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
147
+ digit_upscaled/validation_images/step_2850_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
148
+ digit_upscaled/validation_images/step_2900_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
149
+ digit_upscaled/validation_images/step_2900_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
150
+ digit_upscaled/validation_images/step_2950_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
151
+ digit_upscaled/validation_images/step_2950_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
152
+ digit_upscaled/validation_images/step_300_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
153
+ digit_upscaled/validation_images/step_300_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
154
+ digit_upscaled/validation_images/step_350_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
155
+ digit_upscaled/validation_images/step_350_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
156
+ digit_upscaled/validation_images/step_400_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
157
+ digit_upscaled/validation_images/step_400_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
158
+ digit_upscaled/validation_images/step_450_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
159
+ digit_upscaled/validation_images/step_450_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
160
+ digit_upscaled/validation_images/step_500_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
161
+ digit_upscaled/validation_images/step_500_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
162
+ digit_upscaled/validation_images/step_50_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
163
+ digit_upscaled/validation_images/step_50_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
164
+ digit_upscaled/validation_images/step_550_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
165
+ digit_upscaled/validation_images/step_550_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
166
+ digit_upscaled/validation_images/step_600_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
167
+ digit_upscaled/validation_images/step_600_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
168
+ digit_upscaled/validation_images/step_650_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
169
+ digit_upscaled/validation_images/step_650_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
170
+ digit_upscaled/validation_images/step_700_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
171
+ digit_upscaled/validation_images/step_700_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
172
+ digit_upscaled/validation_images/step_750_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
173
+ digit_upscaled/validation_images/step_750_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
174
+ digit_upscaled/validation_images/step_800_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
175
+ digit_upscaled/validation_images/step_800_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
176
+ digit_upscaled/validation_images/step_850_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
177
+ digit_upscaled/validation_images/step_850_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
178
+ digit_upscaled/validation_images/step_900_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
179
+ digit_upscaled/validation_images/step_900_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
180
+ digit_upscaled/validation_images/step_950_unconditional_512x512.png filter=lfs diff=lfs merge=lfs -text
181
+ digit_upscaled/validation_images/step_950_validation_512x512.png filter=lfs diff=lfs merge=lfs -text
.gitignore ADDED
@@ -0,0 +1 @@
 
 
1
+ push.sh
digit_upscaled/README.md ADDED
@@ -0,0 +1,132 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ base_model: "sd3/unknown-model"
4
+ tags:
5
+ - sd3
6
+ - sd3-diffusers
7
+ - text-to-image
8
+ - diffusers
9
+ - simpletuner
10
+ - not-for-all-audiences
11
+ - lora
12
+ - template:sd-lora
13
+ - standard
14
+ inference: true
15
+ widget:
16
+ - text: 'unconditional (blank prompt)'
17
+ parameters:
18
+ negative_prompt: 'blurry, cropped, ugly'
19
+ output:
20
+ url: ./assets/image_0_0.png
21
+ - text: 'A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant''s posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant.'
22
+ parameters:
23
+ negative_prompt: 'blurry, cropped, ugly'
24
+ output:
25
+ url: ./assets/image_1_0.png
26
+ ---
27
+
28
+ # simpletuner-lora
29
+
30
+ This is a standard PEFT LoRA derived from [sd3/unknown-model](https://huggingface.co/sd3/unknown-model).
31
+
32
+
33
+ The main validation prompt used during training was:
34
+ ```
35
+ A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant's posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant.
36
+ ```
37
+
38
+
39
+ ## Validation settings
40
+ - CFG: `7.5`
41
+ - CFG Rescale: `0.0`
42
+ - Steps: `35`
43
+ - Sampler: `FlowMatchEulerDiscreteScheduler`
44
+ - Seed: `42`
45
+ - Resolution: `512x512`
46
+ - Skip-layer guidance:
47
+
48
+ Note: The validation settings are not necessarily the same as the [training settings](#training-settings).
49
+
50
+ You can find some example images in the following gallery:
51
+
52
+
53
+ <Gallery />
54
+
55
+ The text encoder **was not** trained.
56
+ You may reuse the base model text encoder for inference.
57
+
58
+
59
+ ## Training settings
60
+
61
+ - Training epochs: 1
62
+ - Training steps: 3000
63
+ - Learning rate: 0.0001
64
+ - Learning rate schedule: cosine
65
+ - Warmup steps: 100
66
+ - Max grad norm: 2.0
67
+ - Effective batch size: 16
68
+ - Micro-batch size: 4
69
+ - Gradient accumulation steps: 4
70
+ - Number of GPUs: 1
71
+ - Gradient checkpointing: True
72
+ - Prediction type: flow-matching (extra parameters=['shift=3'])
73
+ - Optimizer: adamw_bf16
74
+ - Trainable parameter precision: Pure BF16
75
+ - Caption dropout probability: 10.0%
76
+
77
+
78
+ - LoRA Rank: 128
79
+ - LoRA Alpha: None
80
+ - LoRA Dropout: 0.1
81
+ - LoRA initialisation style: default
82
+
83
+
84
+ ## Datasets
85
+
86
+ ### pacs
87
+ - Repeats: 0
88
+ - Total number of images: 24000
89
+ - Total number of aspect buckets: 1
90
+ - Resolution: 1.0 megapixels
91
+ - Cropped: False
92
+ - Crop style: None
93
+ - Crop aspect: None
94
+ - Used for regularisation data: No
95
+
96
+
97
+ ## Inference
98
+
99
+
100
+ ```python
101
+ import torch
102
+ from diffusers import DiffusionPipeline
103
+
104
+ model_id = '/ephemeral/shashmi/llava_lets_go/chimaa_finetuner/stable-diffusion-3.5-medium'
105
+ adapter_id = 'Sarim-Hash/simpletuner-lora'
106
+ pipeline = DiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.bfloat16) # loading directly in bf16
107
+ pipeline.load_lora_weights(adapter_id)
108
+
109
+ prompt = "A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant's posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant."
110
+ negative_prompt = 'blurry, cropped, ugly'
111
+
112
+ ## Optional: quantise the model to save on vram.
113
+ ## Note: The model was quantised during training, and so it is recommended to do the same during inference time.
114
+ from optimum.quanto import quantize, freeze, qint8
115
+ quantize(pipeline.transformer, weights=qint8)
116
+ freeze(pipeline.transformer)
117
+
118
+ pipeline.to('cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu') # the pipeline is already in its target precision level
119
+ image = pipeline(
120
+ prompt=prompt,
121
+ negative_prompt=negative_prompt,
122
+ num_inference_steps=35,
123
+ generator=torch.Generator(device='cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu').manual_seed(42),
124
+ width=512,
125
+ height=512,
126
+ guidance_scale=7.5,
127
+ ).images[0]
128
+ image.save("output.png", format="PNG")
129
+ ```
130
+
131
+
132
+
digit_upscaled/all_image_files_pacs.json ADDED
The diff for this file is too large to render. See raw diff
 
digit_upscaled/all_text_cache_files_text-embeds.json ADDED
The diff for this file is too large to render. See raw diff
 
digit_upscaled/all_vae_cache_files_pacs.json ADDED
The diff for this file is too large to render. See raw diff
 
digit_upscaled/assets/image_0_0.png ADDED

Git LFS Details

  • SHA256: aea394633dd6008300cc47c02d06fe9567ebf8c960d77478832203b064032fc7
  • Pointer size: 131 Bytes
  • Size of remote file: 877 kB
digit_upscaled/assets/image_1_0.png ADDED

Git LFS Details

  • SHA256: 60c7a09258f6383f1a4ee7eab8efc72daef8fa11530c5ccbc1e8271dbef943ef
  • Pointer size: 131 Bytes
  • Size of remote file: 408 kB
digit_upscaled/benchmarks/base_model/unconditional_512x512.png ADDED

Git LFS Details

  • SHA256: fabd8007d625e53537564a8d3c4121fb79b8117fa8a837f033441b4c6ea69148
  • Pointer size: 131 Bytes
  • Size of remote file: 441 kB
digit_upscaled/benchmarks/base_model/validation_512x512.png ADDED

Git LFS Details

  • SHA256: 2c6c00eafc749ace816a52e01bc84dde4f4918c4071fba465c1c9283a3b7ff65
  • Pointer size: 131 Bytes
  • Size of remote file: 208 kB
digit_upscaled/checkpoint-1000/README.md ADDED
@@ -0,0 +1,132 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ base_model: "sd3/unknown-model"
4
+ tags:
5
+ - sd3
6
+ - sd3-diffusers
7
+ - text-to-image
8
+ - diffusers
9
+ - simpletuner
10
+ - not-for-all-audiences
11
+ - lora
12
+ - template:sd-lora
13
+ - standard
14
+ inference: true
15
+ widget:
16
+ - text: 'unconditional (blank prompt)'
17
+ parameters:
18
+ negative_prompt: 'blurry, cropped, ugly'
19
+ output:
20
+ url: ./assets/image_0_0.png
21
+ - text: 'A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant''s posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant.'
22
+ parameters:
23
+ negative_prompt: 'blurry, cropped, ugly'
24
+ output:
25
+ url: ./assets/image_1_0.png
26
+ ---
27
+
28
+ # simpletuner-lora
29
+
30
+ This is a standard PEFT LoRA derived from [sd3/unknown-model](https://huggingface.co/sd3/unknown-model).
31
+
32
+
33
+ The main validation prompt used during training was:
34
+ ```
35
+ A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant's posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant.
36
+ ```
37
+
38
+
39
+ ## Validation settings
40
+ - CFG: `7.5`
41
+ - CFG Rescale: `0.0`
42
+ - Steps: `35`
43
+ - Sampler: `FlowMatchEulerDiscreteScheduler`
44
+ - Seed: `42`
45
+ - Resolution: `512x512`
46
+ - Skip-layer guidance:
47
+
48
+ Note: The validation settings are not necessarily the same as the [training settings](#training-settings).
49
+
50
+ You can find some example images in the following gallery:
51
+
52
+
53
+ <Gallery />
54
+
55
+ The text encoder **was not** trained.
56
+ You may reuse the base model text encoder for inference.
57
+
58
+
59
+ ## Training settings
60
+
61
+ - Training epochs: 0
62
+ - Training steps: 1000
63
+ - Learning rate: 0.0001
64
+ - Learning rate schedule: cosine
65
+ - Warmup steps: 100
66
+ - Max grad norm: 2.0
67
+ - Effective batch size: 16
68
+ - Micro-batch size: 4
69
+ - Gradient accumulation steps: 4
70
+ - Number of GPUs: 1
71
+ - Gradient checkpointing: True
72
+ - Prediction type: flow-matching (extra parameters=['shift=3'])
73
+ - Optimizer: adamw_bf16
74
+ - Trainable parameter precision: Pure BF16
75
+ - Caption dropout probability: 10.0%
76
+
77
+
78
+ - LoRA Rank: 128
79
+ - LoRA Alpha: None
80
+ - LoRA Dropout: 0.1
81
+ - LoRA initialisation style: default
82
+
83
+
84
+ ## Datasets
85
+
86
+ ### pacs
87
+ - Repeats: 0
88
+ - Total number of images: 24000
89
+ - Total number of aspect buckets: 1
90
+ - Resolution: 1.0 megapixels
91
+ - Cropped: False
92
+ - Crop style: None
93
+ - Crop aspect: None
94
+ - Used for regularisation data: No
95
+
96
+
97
+ ## Inference
98
+
99
+
100
+ ```python
101
+ import torch
102
+ from diffusers import DiffusionPipeline
103
+
104
+ model_id = '/ephemeral/shashmi/llava_lets_go/chimaa_finetuner/stable-diffusion-3.5-medium'
105
+ adapter_id = 'Sarim-Hash/simpletuner-lora'
106
+ pipeline = DiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.bfloat16) # loading directly in bf16
107
+ pipeline.load_lora_weights(adapter_id)
108
+
109
+ prompt = "A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant's posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant."
110
+ negative_prompt = 'blurry, cropped, ugly'
111
+
112
+ ## Optional: quantise the model to save on vram.
113
+ ## Note: The model was quantised during training, and so it is recommended to do the same during inference time.
114
+ from optimum.quanto import quantize, freeze, qint8
115
+ quantize(pipeline.transformer, weights=qint8)
116
+ freeze(pipeline.transformer)
117
+
118
+ pipeline.to('cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu') # the pipeline is already in its target precision level
119
+ image = pipeline(
120
+ prompt=prompt,
121
+ negative_prompt=negative_prompt,
122
+ num_inference_steps=35,
123
+ generator=torch.Generator(device='cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu').manual_seed(42),
124
+ width=512,
125
+ height=512,
126
+ guidance_scale=7.5,
127
+ ).images[0]
128
+ image.save("output.png", format="PNG")
129
+ ```
130
+
131
+
132
+
digit_upscaled/checkpoint-1000/assets/image_0_0.png ADDED

Git LFS Details

  • SHA256: 0bf6c013170e39a8f49bd06b657d9766915a18efabd15d8ce53446b93003934a
  • Pointer size: 131 Bytes
  • Size of remote file: 926 kB
digit_upscaled/checkpoint-1000/assets/image_1_0.png ADDED

Git LFS Details

  • SHA256: a84863d79d3e02c5ab61696901d1b2d9a670036892c943ce740a40bbb630e5e1
  • Pointer size: 131 Bytes
  • Size of remote file: 369 kB
digit_upscaled/checkpoint-1000/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e3e13b7623d8153911450bfe0465b597ac02d2aaf789bcab822f5fde5f2ecb1
3
+ size 349442426
digit_upscaled/checkpoint-1000/pytorch_lora_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf9d700ee5d758b2a906faa9bcd87d357616d9ddacb8aea5bff16a77bbfa5f82
3
+ size 116431016
digit_upscaled/checkpoint-1000/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e1fc83748a0eca9c703f587aca2569022da2d45c7f671e80708ccd4baa183d12
3
+ size 14408
digit_upscaled/checkpoint-1000/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b40c7683c794eb2d2c5f51a0d043e67c0aac67ed037d02fc10ed51860ad3226
3
+ size 1128
digit_upscaled/checkpoint-1000/training_state-pacs.json ADDED
The diff for this file is too large to render. See raw diff
 
digit_upscaled/checkpoint-1000/training_state.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"global_step": 1000, "epoch_step": 625, "epoch": 1, "exhausted_backends": [], "repeats": {}}
digit_upscaled/checkpoint-1250/README.md ADDED
@@ -0,0 +1,132 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ base_model: "sd3/unknown-model"
4
+ tags:
5
+ - sd3
6
+ - sd3-diffusers
7
+ - text-to-image
8
+ - diffusers
9
+ - simpletuner
10
+ - not-for-all-audiences
11
+ - lora
12
+ - template:sd-lora
13
+ - standard
14
+ inference: true
15
+ widget:
16
+ - text: 'unconditional (blank prompt)'
17
+ parameters:
18
+ negative_prompt: 'blurry, cropped, ugly'
19
+ output:
20
+ url: ./assets/image_0_0.png
21
+ - text: 'A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant''s posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant.'
22
+ parameters:
23
+ negative_prompt: 'blurry, cropped, ugly'
24
+ output:
25
+ url: ./assets/image_1_0.png
26
+ ---
27
+
28
+ # simpletuner-lora
29
+
30
+ This is a standard PEFT LoRA derived from [sd3/unknown-model](https://huggingface.co/sd3/unknown-model).
31
+
32
+
33
+ The main validation prompt used during training was:
34
+ ```
35
+ A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant's posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant.
36
+ ```
37
+
38
+
39
+ ## Validation settings
40
+ - CFG: `7.5`
41
+ - CFG Rescale: `0.0`
42
+ - Steps: `35`
43
+ - Sampler: `FlowMatchEulerDiscreteScheduler`
44
+ - Seed: `42`
45
+ - Resolution: `512x512`
46
+ - Skip-layer guidance:
47
+
48
+ Note: The validation settings are not necessarily the same as the [training settings](#training-settings).
49
+
50
+ You can find some example images in the following gallery:
51
+
52
+
53
+ <Gallery />
54
+
55
+ The text encoder **was not** trained.
56
+ You may reuse the base model text encoder for inference.
57
+
58
+
59
+ ## Training settings
60
+
61
+ - Training epochs: 0
62
+ - Training steps: 1250
63
+ - Learning rate: 0.0001
64
+ - Learning rate schedule: cosine
65
+ - Warmup steps: 100
66
+ - Max grad norm: 2.0
67
+ - Effective batch size: 16
68
+ - Micro-batch size: 4
69
+ - Gradient accumulation steps: 4
70
+ - Number of GPUs: 1
71
+ - Gradient checkpointing: True
72
+ - Prediction type: flow-matching (extra parameters=['shift=3'])
73
+ - Optimizer: adamw_bf16
74
+ - Trainable parameter precision: Pure BF16
75
+ - Caption dropout probability: 10.0%
76
+
77
+
78
+ - LoRA Rank: 128
79
+ - LoRA Alpha: None
80
+ - LoRA Dropout: 0.1
81
+ - LoRA initialisation style: default
82
+
83
+
84
+ ## Datasets
85
+
86
+ ### pacs
87
+ - Repeats: 0
88
+ - Total number of images: 24000
89
+ - Total number of aspect buckets: 1
90
+ - Resolution: 1.0 megapixels
91
+ - Cropped: False
92
+ - Crop style: None
93
+ - Crop aspect: None
94
+ - Used for regularisation data: No
95
+
96
+
97
+ ## Inference
98
+
99
+
100
+ ```python
101
+ import torch
102
+ from diffusers import DiffusionPipeline
103
+
104
+ model_id = '/ephemeral/shashmi/llava_lets_go/chimaa_finetuner/stable-diffusion-3.5-medium'
105
+ adapter_id = 'Sarim-Hash/simpletuner-lora'
106
+ pipeline = DiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.bfloat16) # loading directly in bf16
107
+ pipeline.load_lora_weights(adapter_id)
108
+
109
+ prompt = "A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant's posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant."
110
+ negative_prompt = 'blurry, cropped, ugly'
111
+
112
+ ## Optional: quantise the model to save on vram.
113
+ ## Note: The model was quantised during training, and so it is recommended to do the same during inference time.
114
+ from optimum.quanto import quantize, freeze, qint8
115
+ quantize(pipeline.transformer, weights=qint8)
116
+ freeze(pipeline.transformer)
117
+
118
+ pipeline.to('cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu') # the pipeline is already in its target precision level
119
+ image = pipeline(
120
+ prompt=prompt,
121
+ negative_prompt=negative_prompt,
122
+ num_inference_steps=35,
123
+ generator=torch.Generator(device='cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu').manual_seed(42),
124
+ width=512,
125
+ height=512,
126
+ guidance_scale=7.5,
127
+ ).images[0]
128
+ image.save("output.png", format="PNG")
129
+ ```
130
+
131
+
132
+
digit_upscaled/checkpoint-1250/assets/image_0_0.png ADDED

Git LFS Details

  • SHA256: a2426ed942e1c25b30f3f46f7279dd69f060a390666037926bcac4ae70e58ced
  • Pointer size: 131 Bytes
  • Size of remote file: 973 kB
digit_upscaled/checkpoint-1250/assets/image_1_0.png ADDED

Git LFS Details

  • SHA256: 315db9a8227d21bbbb2007a266a2543a2a46dfc236310f47895810e135ddd736
  • Pointer size: 131 Bytes
  • Size of remote file: 408 kB
digit_upscaled/checkpoint-1250/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:299a0b765cda0f80c1f2ed269d0423035de528ee821401a5a81d3055fc018f8a
3
+ size 349442426
digit_upscaled/checkpoint-1250/pytorch_lora_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5257eb32295ff5943f0f668636ab2c039a9c0a03b03472b2a0062feaf58fd93b
3
+ size 116431016
digit_upscaled/checkpoint-1250/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e380ad4a16f73c3a42ae21056855074ab404d6361ec2fd87ed6cb03d93233760
3
+ size 14344
digit_upscaled/checkpoint-1250/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f0396e8354db804fd5f3b80b54b57a56511fa235d6226728e6557cfe331c1e6
3
+ size 1128
digit_upscaled/checkpoint-1250/training_state-pacs.json ADDED
The diff for this file is too large to render. See raw diff
 
digit_upscaled/checkpoint-1250/training_state.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"global_step": 1250, "epoch_step": 875, "epoch": 1, "exhausted_backends": [], "repeats": {}}
digit_upscaled/checkpoint-1500/README.md ADDED
@@ -0,0 +1,132 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ base_model: "sd3/unknown-model"
4
+ tags:
5
+ - sd3
6
+ - sd3-diffusers
7
+ - text-to-image
8
+ - diffusers
9
+ - simpletuner
10
+ - not-for-all-audiences
11
+ - lora
12
+ - template:sd-lora
13
+ - standard
14
+ inference: true
15
+ widget:
16
+ - text: 'unconditional (blank prompt)'
17
+ parameters:
18
+ negative_prompt: 'blurry, cropped, ugly'
19
+ output:
20
+ url: ./assets/image_0_0.png
21
+ - text: 'A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant''s posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant.'
22
+ parameters:
23
+ negative_prompt: 'blurry, cropped, ugly'
24
+ output:
25
+ url: ./assets/image_1_0.png
26
+ ---
27
+
28
+ # simpletuner-lora
29
+
30
+ This is a standard PEFT LoRA derived from [sd3/unknown-model](https://huggingface.co/sd3/unknown-model).
31
+
32
+
33
+ The main validation prompt used during training was:
34
+ ```
35
+ A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant's posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant.
36
+ ```
37
+
38
+
39
+ ## Validation settings
40
+ - CFG: `7.5`
41
+ - CFG Rescale: `0.0`
42
+ - Steps: `35`
43
+ - Sampler: `FlowMatchEulerDiscreteScheduler`
44
+ - Seed: `42`
45
+ - Resolution: `512x512`
46
+ - Skip-layer guidance:
47
+
48
+ Note: The validation settings are not necessarily the same as the [training settings](#training-settings).
49
+
50
+ You can find some example images in the following gallery:
51
+
52
+
53
+ <Gallery />
54
+
55
+ The text encoder **was not** trained.
56
+ You may reuse the base model text encoder for inference.
57
+
58
+
59
+ ## Training settings
60
+
61
+ - Training epochs: 1
62
+ - Training steps: 1500
63
+ - Learning rate: 0.0001
64
+ - Learning rate schedule: cosine
65
+ - Warmup steps: 100
66
+ - Max grad norm: 2.0
67
+ - Effective batch size: 16
68
+ - Micro-batch size: 4
69
+ - Gradient accumulation steps: 4
70
+ - Number of GPUs: 1
71
+ - Gradient checkpointing: True
72
+ - Prediction type: flow-matching (extra parameters=['shift=3'])
73
+ - Optimizer: adamw_bf16
74
+ - Trainable parameter precision: Pure BF16
75
+ - Caption dropout probability: 10.0%
76
+
77
+
78
+ - LoRA Rank: 128
79
+ - LoRA Alpha: None
80
+ - LoRA Dropout: 0.1
81
+ - LoRA initialisation style: default
82
+
83
+
84
+ ## Datasets
85
+
86
+ ### pacs
87
+ - Repeats: 0
88
+ - Total number of images: 24000
89
+ - Total number of aspect buckets: 1
90
+ - Resolution: 1.0 megapixels
91
+ - Cropped: False
92
+ - Crop style: None
93
+ - Crop aspect: None
94
+ - Used for regularisation data: No
95
+
96
+
97
+ ## Inference
98
+
99
+
100
+ ```python
101
+ import torch
102
+ from diffusers import DiffusionPipeline
103
+
104
+ model_id = '/ephemeral/shashmi/llava_lets_go/chimaa_finetuner/stable-diffusion-3.5-medium'
105
+ adapter_id = 'Sarim-Hash/simpletuner-lora'
106
+ pipeline = DiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.bfloat16) # loading directly in bf16
107
+ pipeline.load_lora_weights(adapter_id)
108
+
109
+ prompt = "A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant's posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant."
110
+ negative_prompt = 'blurry, cropped, ugly'
111
+
112
+ ## Optional: quantise the model to save on vram.
113
+ ## Note: The model was quantised during training, and so it is recommended to do the same during inference time.
114
+ from optimum.quanto import quantize, freeze, qint8
115
+ quantize(pipeline.transformer, weights=qint8)
116
+ freeze(pipeline.transformer)
117
+
118
+ pipeline.to('cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu') # the pipeline is already in its target precision level
119
+ image = pipeline(
120
+ prompt=prompt,
121
+ negative_prompt=negative_prompt,
122
+ num_inference_steps=35,
123
+ generator=torch.Generator(device='cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu').manual_seed(42),
124
+ width=512,
125
+ height=512,
126
+ guidance_scale=7.5,
127
+ ).images[0]
128
+ image.save("output.png", format="PNG")
129
+ ```
130
+
131
+
132
+
digit_upscaled/checkpoint-1500/assets/image_0_0.png ADDED

Git LFS Details

  • SHA256: 9527d8d7ea02ea82f995a14744f9f9caa67938a179dfcc4bb16bd737fd485f5d
  • Pointer size: 131 Bytes
  • Size of remote file: 897 kB
digit_upscaled/checkpoint-1500/assets/image_1_0.png ADDED

Git LFS Details

  • SHA256: 5b20ca09ae4f41e5c8ab29e26cd8e60fee62629acf428f1698458dcb54442afe
  • Pointer size: 131 Bytes
  • Size of remote file: 387 kB
digit_upscaled/checkpoint-1500/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4409e3215e5e126426af6fce28ad77dc9d24b3177977f064c022e12377081941
3
+ size 349442426
digit_upscaled/checkpoint-1500/pytorch_lora_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46a06d90ccf831f94bc65b1f2c5515fbdcf3c490de95b76db2fcbc1e95613636
3
+ size 116431016
digit_upscaled/checkpoint-1500/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0618486845c2dda047253cbe4c5bf375fb865c3549ffd337c6941eb47da91f09
3
+ size 14344
digit_upscaled/checkpoint-1500/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92c21aba5105638209ef13ea982d9ffda2517815d25498abf24fedadcdeec846
3
+ size 1128
digit_upscaled/checkpoint-1500/training_state-pacs.json ADDED
The diff for this file is too large to render. See raw diff
 
digit_upscaled/checkpoint-1500/training_state.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"global_step": 1500, "epoch_step": 1125, "epoch": 1, "exhausted_backends": [], "repeats": {}}
digit_upscaled/checkpoint-1750/README.md ADDED
@@ -0,0 +1,132 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ base_model: "sd3/unknown-model"
4
+ tags:
5
+ - sd3
6
+ - sd3-diffusers
7
+ - text-to-image
8
+ - diffusers
9
+ - simpletuner
10
+ - not-for-all-audiences
11
+ - lora
12
+ - template:sd-lora
13
+ - standard
14
+ inference: true
15
+ widget:
16
+ - text: 'unconditional (blank prompt)'
17
+ parameters:
18
+ negative_prompt: 'blurry, cropped, ugly'
19
+ output:
20
+ url: ./assets/image_0_0.png
21
+ - text: 'A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant''s posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant.'
22
+ parameters:
23
+ negative_prompt: 'blurry, cropped, ugly'
24
+ output:
25
+ url: ./assets/image_1_0.png
26
+ ---
27
+
28
+ # simpletuner-lora
29
+
30
+ This is a standard PEFT LoRA derived from [sd3/unknown-model](https://huggingface.co/sd3/unknown-model).
31
+
32
+
33
+ The main validation prompt used during training was:
34
+ ```
35
+ A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant's posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant.
36
+ ```
37
+
38
+
39
+ ## Validation settings
40
+ - CFG: `7.5`
41
+ - CFG Rescale: `0.0`
42
+ - Steps: `35`
43
+ - Sampler: `FlowMatchEulerDiscreteScheduler`
44
+ - Seed: `42`
45
+ - Resolution: `512x512`
46
+ - Skip-layer guidance:
47
+
48
+ Note: The validation settings are not necessarily the same as the [training settings](#training-settings).
49
+
50
+ You can find some example images in the following gallery:
51
+
52
+
53
+ <Gallery />
54
+
55
+ The text encoder **was not** trained.
56
+ You may reuse the base model text encoder for inference.
57
+
58
+
59
+ ## Training settings
60
+
61
+ - Training epochs: 1
62
+ - Training steps: 1750
63
+ - Learning rate: 0.0001
64
+ - Learning rate schedule: cosine
65
+ - Warmup steps: 100
66
+ - Max grad norm: 2.0
67
+ - Effective batch size: 16
68
+ - Micro-batch size: 4
69
+ - Gradient accumulation steps: 4
70
+ - Number of GPUs: 1
71
+ - Gradient checkpointing: True
72
+ - Prediction type: flow-matching (extra parameters=['shift=3'])
73
+ - Optimizer: adamw_bf16
74
+ - Trainable parameter precision: Pure BF16
75
+ - Caption dropout probability: 10.0%
76
+
77
+
78
+ - LoRA Rank: 128
79
+ - LoRA Alpha: None
80
+ - LoRA Dropout: 0.1
81
+ - LoRA initialisation style: default
82
+
83
+
84
+ ## Datasets
85
+
86
+ ### pacs
87
+ - Repeats: 0
88
+ - Total number of images: 24000
89
+ - Total number of aspect buckets: 1
90
+ - Resolution: 1.0 megapixels
91
+ - Cropped: False
92
+ - Crop style: None
93
+ - Crop aspect: None
94
+ - Used for regularisation data: No
95
+
96
+
97
+ ## Inference
98
+
99
+
100
+ ```python
101
+ import torch
102
+ from diffusers import DiffusionPipeline
103
+
104
+ model_id = '/ephemeral/shashmi/llava_lets_go/chimaa_finetuner/stable-diffusion-3.5-medium'
105
+ adapter_id = 'Sarim-Hash/simpletuner-lora'
106
+ pipeline = DiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.bfloat16) # loading directly in bf16
107
+ pipeline.load_lora_weights(adapter_id)
108
+
109
+ prompt = "A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant's posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant."
110
+ negative_prompt = 'blurry, cropped, ugly'
111
+
112
+ ## Optional: quantise the model to save on vram.
113
+ ## Note: The model was quantised during training, and so it is recommended to do the same during inference time.
114
+ from optimum.quanto import quantize, freeze, qint8
115
+ quantize(pipeline.transformer, weights=qint8)
116
+ freeze(pipeline.transformer)
117
+
118
+ pipeline.to('cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu') # the pipeline is already in its target precision level
119
+ image = pipeline(
120
+ prompt=prompt,
121
+ negative_prompt=negative_prompt,
122
+ num_inference_steps=35,
123
+ generator=torch.Generator(device='cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu').manual_seed(42),
124
+ width=512,
125
+ height=512,
126
+ guidance_scale=7.5,
127
+ ).images[0]
128
+ image.save("output.png", format="PNG")
129
+ ```
130
+
131
+
132
+
digit_upscaled/checkpoint-1750/assets/image_0_0.png ADDED

Git LFS Details

  • SHA256: 67c0cd6134cef5cfd16e06fe4cef5dd57381d3bac110cf6495400ab07679570f
  • Pointer size: 131 Bytes
  • Size of remote file: 927 kB
digit_upscaled/checkpoint-1750/assets/image_1_0.png ADDED

Git LFS Details

  • SHA256: 8e418f11e7f261cd459542c7245856cc37e1a190af7381b1811931db2358b571
  • Pointer size: 131 Bytes
  • Size of remote file: 409 kB
digit_upscaled/checkpoint-1750/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8239de18d3e2c78c62444a03a09258ff4d98d96fa3407dc0e22005221750a211
3
+ size 349442426
digit_upscaled/checkpoint-1750/pytorch_lora_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1fbb3b4c6170665f40a5d2796b6926a3045eeb00cd6854ff8b3db03843a20b3
3
+ size 116431016
digit_upscaled/checkpoint-1750/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5451cc176407619b4710589bbb28334f84e86575a616512aee110f7c8f91ef17
3
+ size 14408
digit_upscaled/checkpoint-1750/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2922907515a7eda3d4adff43a1e668fe04b9659067a29ebea923dbbcf57043c
3
+ size 1128
digit_upscaled/checkpoint-1750/training_state-pacs.json ADDED
The diff for this file is too large to render. See raw diff
 
digit_upscaled/checkpoint-1750/training_state.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"global_step": 1750, "epoch_step": 1375, "epoch": 2, "exhausted_backends": [], "repeats": {"pacs": 0}}
digit_upscaled/checkpoint-2000/README.md ADDED
@@ -0,0 +1,132 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ base_model: "sd3/unknown-model"
4
+ tags:
5
+ - sd3
6
+ - sd3-diffusers
7
+ - text-to-image
8
+ - diffusers
9
+ - simpletuner
10
+ - not-for-all-audiences
11
+ - lora
12
+ - template:sd-lora
13
+ - standard
14
+ inference: true
15
+ widget:
16
+ - text: 'unconditional (blank prompt)'
17
+ parameters:
18
+ negative_prompt: 'blurry, cropped, ugly'
19
+ output:
20
+ url: ./assets/image_0_0.png
21
+ - text: 'A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant''s posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant.'
22
+ parameters:
23
+ negative_prompt: 'blurry, cropped, ugly'
24
+ output:
25
+ url: ./assets/image_1_0.png
26
+ ---
27
+
28
+ # simpletuner-lora
29
+
30
+ This is a standard PEFT LoRA derived from [sd3/unknown-model](https://huggingface.co/sd3/unknown-model).
31
+
32
+
33
+ The main validation prompt used during training was:
34
+ ```
35
+ A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant's posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant.
36
+ ```
37
+
38
+
39
+ ## Validation settings
40
+ - CFG: `7.5`
41
+ - CFG Rescale: `0.0`
42
+ - Steps: `35`
43
+ - Sampler: `FlowMatchEulerDiscreteScheduler`
44
+ - Seed: `42`
45
+ - Resolution: `512x512`
46
+ - Skip-layer guidance:
47
+
48
+ Note: The validation settings are not necessarily the same as the [training settings](#training-settings).
49
+
50
+ You can find some example images in the following gallery:
51
+
52
+
53
+ <Gallery />
54
+
55
+ The text encoder **was not** trained.
56
+ You may reuse the base model text encoder for inference.
57
+
58
+
59
+ ## Training settings
60
+
61
+ - Training epochs: 1
62
+ - Training steps: 2000
63
+ - Learning rate: 0.0001
64
+ - Learning rate schedule: cosine
65
+ - Warmup steps: 100
66
+ - Max grad norm: 2.0
67
+ - Effective batch size: 16
68
+ - Micro-batch size: 4
69
+ - Gradient accumulation steps: 4
70
+ - Number of GPUs: 1
71
+ - Gradient checkpointing: True
72
+ - Prediction type: flow-matching (extra parameters=['shift=3'])
73
+ - Optimizer: adamw_bf16
74
+ - Trainable parameter precision: Pure BF16
75
+ - Caption dropout probability: 10.0%
76
+
77
+
78
+ - LoRA Rank: 128
79
+ - LoRA Alpha: None
80
+ - LoRA Dropout: 0.1
81
+ - LoRA initialisation style: default
82
+
83
+
84
+ ## Datasets
85
+
86
+ ### pacs
87
+ - Repeats: 0
88
+ - Total number of images: 24000
89
+ - Total number of aspect buckets: 1
90
+ - Resolution: 1.0 megapixels
91
+ - Cropped: False
92
+ - Crop style: None
93
+ - Crop aspect: None
94
+ - Used for regularisation data: No
95
+
96
+
97
+ ## Inference
98
+
99
+
100
+ ```python
101
+ import torch
102
+ from diffusers import DiffusionPipeline
103
+
104
+ model_id = '/ephemeral/shashmi/llava_lets_go/chimaa_finetuner/stable-diffusion-3.5-medium'
105
+ adapter_id = 'Sarim-Hash/simpletuner-lora'
106
+ pipeline = DiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.bfloat16) # loading directly in bf16
107
+ pipeline.load_lora_weights(adapter_id)
108
+
109
+ prompt = "A simplistic, hand-drawn illustration of an elephant. the elephant is depicted in a walking pose, with its trunk raised slightly. the drawing is done in black ink on a white background. the elephant's posture and the positioning of its legs suggest movement. the style is minimalistic, with clean lines and a lack of intricate details. the lighting appears to be coming from the top left, casting a shadow on the right side of the elephant."
110
+ negative_prompt = 'blurry, cropped, ugly'
111
+
112
+ ## Optional: quantise the model to save on vram.
113
+ ## Note: The model was quantised during training, and so it is recommended to do the same during inference time.
114
+ from optimum.quanto import quantize, freeze, qint8
115
+ quantize(pipeline.transformer, weights=qint8)
116
+ freeze(pipeline.transformer)
117
+
118
+ pipeline.to('cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu') # the pipeline is already in its target precision level
119
+ image = pipeline(
120
+ prompt=prompt,
121
+ negative_prompt=negative_prompt,
122
+ num_inference_steps=35,
123
+ generator=torch.Generator(device='cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu').manual_seed(42),
124
+ width=512,
125
+ height=512,
126
+ guidance_scale=7.5,
127
+ ).images[0]
128
+ image.save("output.png", format="PNG")
129
+ ```
130
+
131
+
132
+
digit_upscaled/checkpoint-2000/assets/image_0_0.png ADDED

Git LFS Details

  • SHA256: 3c956b84a9273100527209d81d0f2085a8f6f0854781594d595d195d263b69e3
  • Pointer size: 131 Bytes
  • Size of remote file: 913 kB
digit_upscaled/checkpoint-2000/assets/image_1_0.png ADDED

Git LFS Details

  • SHA256: 1c25f14d9c7fbc46141f4e1eb87f2c61ab8b0471e8d812d1a767de9ed0d36590
  • Pointer size: 131 Bytes
  • Size of remote file: 389 kB
digit_upscaled/checkpoint-2000/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:547d2b19a017d3cad429dcad568765efe2b3557f3dbe5296a135105bfef755a3
3
+ size 349442426