Upload folder using huggingface_hub
Browse files- samples/sample_0.jpg +2 -2
- samples/sample_1.jpg +2 -2
- samples/sample_2.jpg +2 -2
- samples/sample_decoded.jpg +2 -2
- samples/sample_real.jpg +2 -2
- train_vae.py +8 -7
- vae/diffusion_pytorch_model.safetensors +1 -1
samples/sample_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/sample_1.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/sample_2.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/sample_decoded.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/sample_real.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
train_vae.py
CHANGED
|
@@ -29,9 +29,9 @@ from collections import deque
|
|
| 29 |
# --------------------------- Параметры ---------------------------
|
| 30 |
ds_path = "/workspace/d23"
|
| 31 |
project = "vae"
|
| 32 |
-
batch_size =
|
| 33 |
base_learning_rate = 2e-6
|
| 34 |
-
min_learning_rate =
|
| 35 |
num_epochs = 25
|
| 36 |
sample_interval_share = 2
|
| 37 |
use_wandb = True
|
|
@@ -40,8 +40,8 @@ use_decay = True
|
|
| 40 |
optimizer_type = "adam8bit"
|
| 41 |
dtype = torch.float32
|
| 42 |
|
| 43 |
-
model_resolution =
|
| 44 |
-
high_resolution =
|
| 45 |
limit = 0
|
| 46 |
save_barrier = 1.3
|
| 47 |
warmup_percent = 0.001
|
|
@@ -50,7 +50,7 @@ beta2 = 0.997
|
|
| 50 |
eps = 1e-8
|
| 51 |
clip_grad_norm = 1.0
|
| 52 |
mixed_precision = "no"
|
| 53 |
-
gradient_accumulation_steps =
|
| 54 |
generated_folder = "samples"
|
| 55 |
save_as = "vae"
|
| 56 |
num_workers = 0
|
|
@@ -169,9 +169,10 @@ if full_training and not train_decoder_only:
|
|
| 169 |
else:
|
| 170 |
# учим только 0-й блок декодера + post_quant_conv
|
| 171 |
if hasattr(core, "decoder"):
|
| 172 |
-
# --- только 0-й up_block ---
|
| 173 |
if hasattr(core.decoder, "up_blocks") and len(core.decoder.up_blocks) > 0:
|
| 174 |
-
|
|
|
|
|
|
|
| 175 |
p.requires_grad = True
|
| 176 |
unfrozen_param_names.append(f"decoder.up_blocks.0.{name}")
|
| 177 |
else:
|
|
|
|
| 29 |
# --------------------------- Параметры ---------------------------
|
| 30 |
ds_path = "/workspace/d23"
|
| 31 |
project = "vae"
|
| 32 |
+
batch_size = 1
|
| 33 |
base_learning_rate = 2e-6
|
| 34 |
+
min_learning_rate = 8e-8
|
| 35 |
num_epochs = 25
|
| 36 |
sample_interval_share = 2
|
| 37 |
use_wandb = True
|
|
|
|
| 40 |
optimizer_type = "adam8bit"
|
| 41 |
dtype = torch.float32
|
| 42 |
|
| 43 |
+
model_resolution = 512
|
| 44 |
+
high_resolution = 1024
|
| 45 |
limit = 0
|
| 46 |
save_barrier = 1.3
|
| 47 |
warmup_percent = 0.001
|
|
|
|
| 50 |
eps = 1e-8
|
| 51 |
clip_grad_norm = 1.0
|
| 52 |
mixed_precision = "no"
|
| 53 |
+
gradient_accumulation_steps = 16
|
| 54 |
generated_folder = "samples"
|
| 55 |
save_as = "vae"
|
| 56 |
num_workers = 0
|
|
|
|
| 169 |
else:
|
| 170 |
# учим только 0-й блок декодера + post_quant_conv
|
| 171 |
if hasattr(core, "decoder"):
|
|
|
|
| 172 |
if hasattr(core.decoder, "up_blocks") and len(core.decoder.up_blocks) > 0:
|
| 173 |
+
# --- только 0-й up_block ---
|
| 174 |
+
# for name, p in core.decoder.up_blocks[0].named_parameters():
|
| 175 |
+
for name, p in core.decoder.up_blocks.named_parameters():
|
| 176 |
p.requires_grad = True
|
| 177 |
unfrozen_param_names.append(f"decoder.up_blocks.0.{name}")
|
| 178 |
else:
|
vae/diffusion_pytorch_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 444559412
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c917b96507d94e2c4954494c0479344574155ccbb479acbe0ba7a8c6f05e4af3
|
| 3 |
size 444559412
|