Commit
·
c358db4
1
Parent(s):
4690128
ump success
Browse files
d-adaptation/notes.md
CHANGED
|
@@ -10,7 +10,8 @@ UMP ends up with image generations that look like a single brown square, still t
|
|
| 10 |
As noted in the same github issue, alpha/rank scaling modifies the gradient update to become smaller and thus d-adaptation to boost the learning rate. This could be the reason why it goes bad.
|
| 11 |
|
| 12 |
UMP redone at dim 8 alpha 8 showed recognizable character but still significantly degraded aesthetics and prompt coherence.
|
| 13 |
-
|
|
|
|
| 14 |
|
| 15 |
## Dim
|
| 16 |
128 dim shows some local noisy patterns. Reranking the model to a lower dim from 128 doesn't get rid of it. Converting the weights of the last up block in the unet does but also causes a noticable change in the generated character. Obviously you could reduce the last up block by a smaller amount.
|
|
|
|
| 10 |
As noted in the same github issue, alpha/rank scaling modifies the gradient update to become smaller and thus d-adaptation to boost the learning rate. This could be the reason why it goes bad.
|
| 11 |
|
| 12 |
UMP redone at dim 8 alpha 8 showed recognizable character but still significantly degraded aesthetics and prompt coherence.
|
| 13 |
+
After redoing UMP at dim 8 alpha 8 with less cosine restarts (16->9), the results are much better.
|
| 14 |
+
Consine restarts would likely affect how much time we spend at a high learning rate which could be the reason for blowing the model apart.
|
| 15 |
|
| 16 |
## Dim
|
| 17 |
128 dim shows some local noisy patterns. Reranking the model to a lower dim from 128 doesn't get rid of it. Converting the weights of the last up block in the unet does but also causes a noticable change in the generated character. Obviously you could reduce the last up block by a smaller amount.
|
d-adaptation/ump45_(girls'_frontline)/dim8_alpha8_less_restarts_success.jpg
ADDED
|
Git LFS Details
|