2907
Browse files- micro/config.json +2 -2
- micro/diffusion_pytorch_model.safetensors +2 -2
- samples/micro_192x384_0.jpg +2 -2
- samples/micro_256x384_0.jpg +2 -2
- samples/micro_320x384_0.jpg +2 -2
- samples/micro_384x192_0.jpg +2 -2
- samples/micro_384x256_0.jpg +2 -2
- samples/micro_384x320_0.jpg +2 -2
- samples/micro_384x384_0.jpg +2 -2
- samples/sdxl_192x384_0.jpg +3 -0
- samples/sdxl_256x384_0.jpg +3 -0
- samples/sdxl_320x384_0.jpg +3 -0
- samples/sdxl_384x192_0.jpg +3 -0
- samples/sdxl_384x256_0.jpg +3 -0
- samples/sdxl_384x320_0.jpg +3 -0
- src/micro_create.ipynb +2 -2
- train.py +6 -6
micro/config.json
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:18891ea8f81c705422b12d7575493f5abb949d60f2814a989a436fde65e84c82
|
| 3 |
+
size 1873
|
micro/diffusion_pytorch_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9fa21e51b39a8f8f44036b0e271cddc65914ea3591cce6e0a35506425026f183
|
| 3 |
+
size 3929714960
|
samples/micro_192x384_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/micro_256x384_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/micro_320x384_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/micro_384x192_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/micro_384x256_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/micro_384x320_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/micro_384x384_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/sdxl_192x384_0.jpg
ADDED
|
Git LFS Details
|
samples/sdxl_256x384_0.jpg
ADDED
|
Git LFS Details
|
samples/sdxl_320x384_0.jpg
ADDED
|
Git LFS Details
|
samples/sdxl_384x192_0.jpg
ADDED
|
Git LFS Details
|
samples/sdxl_384x256_0.jpg
ADDED
|
Git LFS Details
|
samples/sdxl_384x320_0.jpg
ADDED
|
Git LFS Details
|
src/micro_create.ipynb
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:856a34f139aad0b40359836b62caa7a005ee898990e416a7a04a9ded0881496a
|
| 3 |
+
size 33982
|
train.py
CHANGED
|
@@ -27,12 +27,12 @@ import torch.nn.functional as F
|
|
| 27 |
# --------------------------- Параметры ---------------------------
|
| 28 |
ds_path = "datasets/384"
|
| 29 |
project = "micro"
|
| 30 |
-
batch_size =
|
| 31 |
base_learning_rate = 1e-4
|
| 32 |
min_learning_rate = 5e-5
|
| 33 |
-
num_epochs =
|
| 34 |
# samples/save per epoch
|
| 35 |
-
sample_interval_share =
|
| 36 |
use_wandb = True
|
| 37 |
save_model = True
|
| 38 |
use_decay = True
|
|
@@ -54,12 +54,12 @@ dispersive_temperature=0.5
|
|
| 54 |
dispersive_weight= 0.05
|
| 55 |
percentile_clipping = 95 # 8bit optim
|
| 56 |
betta2 = 0.97
|
| 57 |
-
eps = 1e-
|
| 58 |
clip_grad_norm = 1.0
|
| 59 |
steps_offset = 0 # Scheduler
|
| 60 |
limit = 0
|
| 61 |
checkpoints_folder = ""
|
| 62 |
-
mixed_precision = "
|
| 63 |
gradient_accumulation_steps = 1
|
| 64 |
accelerator = Accelerator(
|
| 65 |
mixed_precision=mixed_precision,
|
|
@@ -807,7 +807,7 @@ for epoch in range(start_epoch, start_epoch + num_epochs):
|
|
| 807 |
avg_loss = np.mean(batch_losses[-sample_interval:])
|
| 808 |
avg_tloss = np.mean(batch_tlosses[-sample_interval:])
|
| 809 |
avg_grad = torch.mean(torch.stack(batch_grads[-sample_interval:])).cpu().item()
|
| 810 |
-
print(f"Эпоха {epoch}, шаг {global_step}, средний лосс: {avg_loss:.6f}")
|
| 811 |
|
| 812 |
if save_model:
|
| 813 |
print("saving:",avg_loss < min_loss*save_barrier)
|
|
|
|
| 27 |
# --------------------------- Параметры ---------------------------
|
| 28 |
ds_path = "datasets/384"
|
| 29 |
project = "micro"
|
| 30 |
+
batch_size = 64
|
| 31 |
base_learning_rate = 1e-4
|
| 32 |
min_learning_rate = 5e-5
|
| 33 |
+
num_epochs = 50
|
| 34 |
# samples/save per epoch
|
| 35 |
+
sample_interval_share = 10
|
| 36 |
use_wandb = True
|
| 37 |
save_model = True
|
| 38 |
use_decay = True
|
|
|
|
| 54 |
dispersive_weight= 0.05
|
| 55 |
percentile_clipping = 95 # 8bit optim
|
| 56 |
betta2 = 0.97
|
| 57 |
+
eps = 1e-6
|
| 58 |
clip_grad_norm = 1.0
|
| 59 |
steps_offset = 0 # Scheduler
|
| 60 |
limit = 0
|
| 61 |
checkpoints_folder = ""
|
| 62 |
+
mixed_precision = "no" #"fp16"
|
| 63 |
gradient_accumulation_steps = 1
|
| 64 |
accelerator = Accelerator(
|
| 65 |
mixed_precision=mixed_precision,
|
|
|
|
| 807 |
avg_loss = np.mean(batch_losses[-sample_interval:])
|
| 808 |
avg_tloss = np.mean(batch_tlosses[-sample_interval:])
|
| 809 |
avg_grad = torch.mean(torch.stack(batch_grads[-sample_interval:])).cpu().item()
|
| 810 |
+
print(f"Эпоха {epoch}, шаг {global_step}, средний лосс: {avg_loss:.6f}, grad: {avg_grad:.6f}")
|
| 811 |
|
| 812 |
if save_model:
|
| 813 |
print("saving:",avg_loss < min_loss*save_barrier)
|