2512
Browse files- samples/unet_384x768_0.jpg +2 -2
- samples/unet_416x768_0.jpg +2 -2
- samples/unet_448x768_0.jpg +2 -2
- samples/unet_480x768_0.jpg +2 -2
- samples/unet_512x768_0.jpg +2 -2
- samples/unet_544x768_0.jpg +2 -2
- samples/unet_576x768_0.jpg +2 -2
- samples/unet_608x768_0.jpg +2 -2
- samples/unet_640x768_0.jpg +2 -2
- samples/unet_672x768_0.jpg +2 -2
- samples/unet_704x768_0.jpg +2 -2
- samples/unet_736x768_0.jpg +2 -2
- samples/unet_768x384_0.jpg +2 -2
- samples/unet_768x416_0.jpg +2 -2
- samples/unet_768x448_0.jpg +2 -2
- samples/unet_768x480_0.jpg +2 -2
- samples/unet_768x512_0.jpg +2 -2
- samples/unet_768x544_0.jpg +2 -2
- samples/unet_768x576_0.jpg +2 -2
- samples/unet_768x608_0.jpg +2 -2
- samples/unet_768x640_0.jpg +2 -2
- samples/unet_768x672_0.jpg +2 -2
- samples/unet_768x704_0.jpg +2 -2
- samples/unet_768x736_0.jpg +2 -2
- samples/unet_768x768_0.jpg +2 -2
- src/sdxs_create.ipynb +2 -2
- src/sdxs_sdxxs_transfer.ipynb +2 -2
- test.ipynb +2 -2
- train.py +6 -6
- unet/config.json +2 -2
- unet/diffusion_pytorch_model.safetensors +2 -2
samples/unet_384x768_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/unet_416x768_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/unet_448x768_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/unet_480x768_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/unet_512x768_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/unet_544x768_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/unet_576x768_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/unet_608x768_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/unet_640x768_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/unet_672x768_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/unet_704x768_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/unet_736x768_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/unet_768x384_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/unet_768x416_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/unet_768x448_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/unet_768x480_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/unet_768x512_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/unet_768x544_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/unet_768x576_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/unet_768x608_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/unet_768x640_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/unet_768x672_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/unet_768x704_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/unet_768x736_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/unet_768x768_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
src/sdxs_create.ipynb
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8c2659b9cf3ffaea51ac60f748764234bb55403da46efc1e77ab6259ebd2bc44
|
| 3 |
+
size 8033
|
src/sdxs_sdxxs_transfer.ipynb
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bb62c419aa9b0daaf7da17486c3e2e55f7bdada8c9baa0092549114c8f7e45fd
|
| 3 |
+
size 27104
|
test.ipynb
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:487dd5242008c70fe41643c48bbfbf595908048b2faed4254cf3bbc613858299
|
| 3 |
+
size 10046639
|
train.py
CHANGED
|
@@ -29,11 +29,11 @@ from transformers import AutoTokenizer, AutoModel
|
|
| 29 |
# --------------------------- Параметры ---------------------------
|
| 30 |
ds_path = "/workspace/sdxs/datasets/768"
|
| 31 |
project = "unet"
|
| 32 |
-
batch_size =
|
| 33 |
-
base_learning_rate =
|
| 34 |
min_learning_rate = 2.7e-5
|
| 35 |
num_epochs = 100
|
| 36 |
-
sample_interval_share =
|
| 37 |
max_length = 192
|
| 38 |
use_wandb = True
|
| 39 |
use_comet_ml = False
|
|
@@ -173,12 +173,12 @@ def encode_texts(texts, max_length=max_length):
|
|
| 173 |
|
| 174 |
#return hidden, attention_mask
|
| 175 |
# --- НОВАЯ ЛОГИКА: ОБЪЕДИНЕНИЕ ДЛЯ КРОСС-ВНИМАНИЯ ---
|
| 176 |
-
# 1. Расширяем пулинг-вектор до последовательности [B, 1,
|
| 177 |
pooled_expanded = pooled.unsqueeze(1)
|
| 178 |
|
| 179 |
# 2. Объединяем последовательность токенов и пулинг-вектор
|
| 180 |
# !!! ИЗМЕНЕНИЕ ЗДЕСЬ !!!: Пулинг идет ПЕРВЫМ
|
| 181 |
-
# Теперь: [B, 1 + L,
|
| 182 |
new_encoder_hidden_states = torch.cat([pooled_expanded, hidden], dim=1)
|
| 183 |
|
| 184 |
# 3. Обновляем маску внимания для нового токена
|
|
@@ -476,7 +476,7 @@ fixed_samples = get_fixed_samples_by_resolution(dataset)
|
|
| 476 |
# --- [UPDATED] Функция для негативного эмбеддинга (возвращает 3 элемента) ---
|
| 477 |
def get_negative_embedding(neg_prompt="", batch_size=1):
|
| 478 |
if not neg_prompt:
|
| 479 |
-
hidden_dim =
|
| 480 |
seq_len = max_length
|
| 481 |
empty_emb = torch.zeros((batch_size, seq_len, hidden_dim), dtype=dtype, device=device)
|
| 482 |
empty_mask = torch.ones((batch_size, seq_len), dtype=torch.int64, device=device)
|
|
|
|
| 29 |
# --------------------------- Параметры ---------------------------
|
| 30 |
ds_path = "/workspace/sdxs/datasets/768"
|
| 31 |
project = "unet"
|
| 32 |
+
batch_size = 40
|
| 33 |
+
base_learning_rate = 4e-5
|
| 34 |
min_learning_rate = 2.7e-5
|
| 35 |
num_epochs = 100
|
| 36 |
+
sample_interval_share = 6
|
| 37 |
max_length = 192
|
| 38 |
use_wandb = True
|
| 39 |
use_comet_ml = False
|
|
|
|
| 173 |
|
| 174 |
#return hidden, attention_mask
|
| 175 |
# --- НОВАЯ ЛОГИКА: ОБЪЕДИНЕНИЕ ДЛЯ КРОСС-ВНИМАНИЯ ---
|
| 176 |
+
# 1. Расширяем пулинг-вектор до последовательности [B, 1, emb]
|
| 177 |
pooled_expanded = pooled.unsqueeze(1)
|
| 178 |
|
| 179 |
# 2. Объединяем последовательность токенов и пулинг-вектор
|
| 180 |
# !!! ИЗМЕНЕНИЕ ЗДЕСЬ !!!: Пулинг идет ПЕРВЫМ
|
| 181 |
+
# Теперь: [B, 1 + L, emb]. Пулинг стал токеном в НАЧАЛЕ.
|
| 182 |
new_encoder_hidden_states = torch.cat([pooled_expanded, hidden], dim=1)
|
| 183 |
|
| 184 |
# 3. Обновляем маску внимания для нового токена
|
|
|
|
| 476 |
# --- [UPDATED] Функция для негативного эмбеддинга (возвращает 3 элемента) ---
|
| 477 |
def get_negative_embedding(neg_prompt="", batch_size=1):
|
| 478 |
if not neg_prompt:
|
| 479 |
+
hidden_dim = 2048
|
| 480 |
seq_len = max_length
|
| 481 |
empty_emb = torch.zeros((batch_size, seq_len, hidden_dim), dtype=dtype, device=device)
|
| 482 |
empty_mask = torch.ones((batch_size, seq_len), dtype=torch.int64, device=device)
|
unet/config.json
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fb10c12478625b355068dbfce29e8db3c8a52d03e10508452e4feb5404545584
|
| 3 |
+
size 1876
|
unet/diffusion_pytorch_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2a548c55ca1e248fcbdd0b198192d0213bf0e20a8d66b63b01b722964e61af80
|
| 3 |
+
size 7444321360
|