Add files using upload-large-folder tool
Browse files- .gitattributes +2 -0
- .job_config.json +10 -16
- config.yaml +12 -16
- lana_lora_wan2.1_000000750.safetensors +3 -0
- lana_lora_wan2.1_000001000.safetensors +3 -0
- lana_lora_wan2.1_000001250.safetensors +3 -0
- lana_lora_wan2.1_000001500.safetensors +3 -0
- log.txt +0 -0
- loss_log.db +2 -2
- loss_log.db-shm +0 -0
- loss_log.db-wal +1 -1
- optimizer.pt +2 -2
- pid.txt +1 -1
- samples/1772099185860__000000000_0.jpg +0 -0
- samples/1772099511556__000000250_0.jpg +0 -0
- samples/1772099815995__000000500_0.jpg +3 -0
- samples/1772100143209__000000750_0.jpg +0 -0
- samples/1772100455705__000001000_0.jpg +0 -0
- samples/1772100773170__000001250_0.jpg +0 -0
- samples/1772101085673__000001500_0.jpg +0 -0
.gitattributes
CHANGED
|
@@ -52,3 +52,5 @@ samples/1771662989226__000003500_0.jpg filter=lfs diff=lfs merge=lfs -text
|
|
| 52 |
samples/1771664222598__000004000_0.jpg filter=lfs diff=lfs merge=lfs -text
|
| 53 |
samples/1771663598140__000003750_0.jpg filter=lfs diff=lfs merge=lfs -text
|
| 54 |
samples/1772010163437__000000750_0.jpg filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
| 52 |
samples/1771664222598__000004000_0.jpg filter=lfs diff=lfs merge=lfs -text
|
| 53 |
samples/1771663598140__000003750_0.jpg filter=lfs diff=lfs merge=lfs -text
|
| 54 |
samples/1772010163437__000000750_0.jpg filter=lfs diff=lfs merge=lfs -text
|
| 55 |
+
samples/1772099815995__000000500_0.jpg filter=lfs diff=lfs merge=lfs -text
|
| 56 |
+
log.txt filter=lfs diff=lfs merge=lfs -text
|
.job_config.json
CHANGED
|
@@ -1,7 +1,7 @@
|
|
| 1 |
{
|
| 2 |
"job": "extension",
|
| 3 |
"config": {
|
| 4 |
-
"name": "lana_lora_wan2.
|
| 5 |
"process": [
|
| 6 |
{
|
| 7 |
"type": "diffusion_trainer",
|
|
@@ -70,18 +70,18 @@
|
|
| 70 |
},
|
| 71 |
"unload_text_encoder": false,
|
| 72 |
"cache_text_embeddings": false,
|
| 73 |
-
"lr": 0.
|
| 74 |
"ema_config": {
|
| 75 |
"use_ema": false,
|
| 76 |
"ema_decay": 0.99
|
| 77 |
},
|
| 78 |
-
"skip_first_sample":
|
| 79 |
"force_first_sample": false,
|
| 80 |
"disable_sampling": false,
|
| 81 |
"dtype": "bf16",
|
| 82 |
-
"diff_output_preservation":
|
| 83 |
"diff_output_preservation_multiplier": 1,
|
| 84 |
-
"diff_output_preservation_class": "
|
| 85 |
"switch_boundary_every": 1,
|
| 86 |
"loss_type": "mse"
|
| 87 |
},
|
|
@@ -90,20 +90,14 @@
|
|
| 90 |
"use_ui_logger": true
|
| 91 |
},
|
| 92 |
"model": {
|
| 93 |
-
"name_or_path": "
|
| 94 |
"quantize": false,
|
| 95 |
"qtype": "qfloat8",
|
| 96 |
-
"quantize_te":
|
| 97 |
"qtype_te": "qfloat8",
|
| 98 |
-
"arch": "
|
| 99 |
"low_vram": false,
|
| 100 |
-
"model_kwargs": {
|
| 101 |
-
"train_high_noise": true,
|
| 102 |
-
"train_low_noise": true
|
| 103 |
-
},
|
| 104 |
-
"layer_offloading": false,
|
| 105 |
-
"layer_offloading_text_encoder_percent": 1,
|
| 106 |
-
"layer_offloading_transformer_percent": 1
|
| 107 |
},
|
| 108 |
"sample": {
|
| 109 |
"sampler": "flowmatch",
|
|
@@ -112,7 +106,7 @@
|
|
| 112 |
"height": 1024,
|
| 113 |
"samples": [
|
| 114 |
{
|
| 115 |
-
"prompt": "l4n4lux"
|
| 116 |
}
|
| 117 |
],
|
| 118 |
"neg": "",
|
|
|
|
| 1 |
{
|
| 2 |
"job": "extension",
|
| 3 |
"config": {
|
| 4 |
+
"name": "lana_lora_wan2.1",
|
| 5 |
"process": [
|
| 6 |
{
|
| 7 |
"type": "diffusion_trainer",
|
|
|
|
| 70 |
},
|
| 71 |
"unload_text_encoder": false,
|
| 72 |
"cache_text_embeddings": false,
|
| 73 |
+
"lr": 0.0001,
|
| 74 |
"ema_config": {
|
| 75 |
"use_ema": false,
|
| 76 |
"ema_decay": 0.99
|
| 77 |
},
|
| 78 |
+
"skip_first_sample": false,
|
| 79 |
"force_first_sample": false,
|
| 80 |
"disable_sampling": false,
|
| 81 |
"dtype": "bf16",
|
| 82 |
+
"diff_output_preservation": false,
|
| 83 |
"diff_output_preservation_multiplier": 1,
|
| 84 |
+
"diff_output_preservation_class": "person",
|
| 85 |
"switch_boundary_every": 1,
|
| 86 |
"loss_type": "mse"
|
| 87 |
},
|
|
|
|
| 90 |
"use_ui_logger": true
|
| 91 |
},
|
| 92 |
"model": {
|
| 93 |
+
"name_or_path": "Wan-AI/Wan2.1-T2V-14B-Diffusers",
|
| 94 |
"quantize": false,
|
| 95 |
"qtype": "qfloat8",
|
| 96 |
+
"quantize_te": false,
|
| 97 |
"qtype_te": "qfloat8",
|
| 98 |
+
"arch": "wan21:14b",
|
| 99 |
"low_vram": false,
|
| 100 |
+
"model_kwargs": {}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 101 |
},
|
| 102 |
"sample": {
|
| 103 |
"sampler": "flowmatch",
|
|
|
|
| 106 |
"height": 1024,
|
| 107 |
"samples": [
|
| 108 |
{
|
| 109 |
+
"prompt": "l4n4lux, woman with red hair, playing chess at the park, bomb going off in the background"
|
| 110 |
}
|
| 111 |
],
|
| 112 |
"neg": "",
|
config.yaml
CHANGED
|
@@ -1,6 +1,6 @@
|
|
| 1 |
job: extension
|
| 2 |
config:
|
| 3 |
-
name: lana_lora_wan2.
|
| 4 |
process:
|
| 5 |
- type: diffusion_trainer
|
| 6 |
training_folder: /root/lab/ai-toolkit/output
|
|
@@ -60,43 +60,39 @@ config:
|
|
| 60 |
weight_decay: 0.0001
|
| 61 |
unload_text_encoder: false
|
| 62 |
cache_text_embeddings: false
|
| 63 |
-
lr: 0.
|
| 64 |
ema_config:
|
| 65 |
use_ema: false
|
| 66 |
ema_decay: 0.99
|
| 67 |
-
skip_first_sample:
|
| 68 |
force_first_sample: false
|
| 69 |
disable_sampling: false
|
| 70 |
dtype: bf16
|
| 71 |
-
diff_output_preservation:
|
| 72 |
diff_output_preservation_multiplier: 1
|
| 73 |
-
diff_output_preservation_class:
|
| 74 |
switch_boundary_every: 1
|
| 75 |
loss_type: mse
|
| 76 |
logging:
|
| 77 |
log_every: 1
|
| 78 |
use_ui_logger: true
|
| 79 |
model:
|
| 80 |
-
name_or_path:
|
| 81 |
quantize: false
|
| 82 |
qtype: qfloat8
|
| 83 |
-
quantize_te:
|
| 84 |
qtype_te: qfloat8
|
| 85 |
-
arch:
|
| 86 |
low_vram: false
|
| 87 |
-
model_kwargs:
|
| 88 |
-
train_high_noise: true
|
| 89 |
-
train_low_noise: true
|
| 90 |
-
layer_offloading: false
|
| 91 |
-
layer_offloading_text_encoder_percent: 1
|
| 92 |
-
layer_offloading_transformer_percent: 1
|
| 93 |
sample:
|
| 94 |
sampler: flowmatch
|
| 95 |
sample_every: 250
|
| 96 |
width: 1024
|
| 97 |
height: 1024
|
| 98 |
samples:
|
| 99 |
-
- prompt: l4n4lux
|
|
|
|
| 100 |
neg: ''
|
| 101 |
seed: 42
|
| 102 |
walk_seed: true
|
|
@@ -105,5 +101,5 @@ config:
|
|
| 105 |
num_frames: 1
|
| 106 |
fps: 1
|
| 107 |
meta:
|
| 108 |
-
name: lana_lora_wan2.
|
| 109 |
version: '1.0'
|
|
|
|
| 1 |
job: extension
|
| 2 |
config:
|
| 3 |
+
name: lana_lora_wan2.1
|
| 4 |
process:
|
| 5 |
- type: diffusion_trainer
|
| 6 |
training_folder: /root/lab/ai-toolkit/output
|
|
|
|
| 60 |
weight_decay: 0.0001
|
| 61 |
unload_text_encoder: false
|
| 62 |
cache_text_embeddings: false
|
| 63 |
+
lr: 0.0001
|
| 64 |
ema_config:
|
| 65 |
use_ema: false
|
| 66 |
ema_decay: 0.99
|
| 67 |
+
skip_first_sample: false
|
| 68 |
force_first_sample: false
|
| 69 |
disable_sampling: false
|
| 70 |
dtype: bf16
|
| 71 |
+
diff_output_preservation: false
|
| 72 |
diff_output_preservation_multiplier: 1
|
| 73 |
+
diff_output_preservation_class: person
|
| 74 |
switch_boundary_every: 1
|
| 75 |
loss_type: mse
|
| 76 |
logging:
|
| 77 |
log_every: 1
|
| 78 |
use_ui_logger: true
|
| 79 |
model:
|
| 80 |
+
name_or_path: Wan-AI/Wan2.1-T2V-14B-Diffusers
|
| 81 |
quantize: false
|
| 82 |
qtype: qfloat8
|
| 83 |
+
quantize_te: false
|
| 84 |
qtype_te: qfloat8
|
| 85 |
+
arch: wan21:14b
|
| 86 |
low_vram: false
|
| 87 |
+
model_kwargs: {}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 88 |
sample:
|
| 89 |
sampler: flowmatch
|
| 90 |
sample_every: 250
|
| 91 |
width: 1024
|
| 92 |
height: 1024
|
| 93 |
samples:
|
| 94 |
+
- prompt: l4n4lux, woman with red hair, playing chess at the park, bomb going
|
| 95 |
+
off in the background
|
| 96 |
neg: ''
|
| 97 |
seed: 42
|
| 98 |
walk_seed: true
|
|
|
|
| 101 |
num_frames: 1
|
| 102 |
fps: 1
|
| 103 |
meta:
|
| 104 |
+
name: lana_lora_wan2.1
|
| 105 |
version: '1.0'
|
lana_lora_wan2.1_000000750.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f02581c4614bf8370b59bc667c57d3e73664427c609ec4c747648376f24c9307
|
| 3 |
+
size 613517200
|
lana_lora_wan2.1_000001000.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c21c609ee2ec3cc75de275fbcd94dae1c4f711c091b5ec4f44d5e36b66343811
|
| 3 |
+
size 613517200
|
lana_lora_wan2.1_000001250.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:91ac396b622212e94097d0df9b75bcd79dbea8ecbeb52d39062d115a5a27ed9b
|
| 3 |
+
size 613517200
|
lana_lora_wan2.1_000001500.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1553ceb84de86c1da882b523bd30572d73358616e93c8ec66bb50eabf1e60a7c
|
| 3 |
+
size 613517200
|
log.txt
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|
loss_log.db
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5b02e0c8850a39eef04ba54855b958fe435349e7327523f3f227c691f9fc3676
|
| 3 |
+
size 282624
|
loss_log.db-shm
CHANGED
|
Binary files a/loss_log.db-shm and b/loss_log.db-shm differ
|
|
|
loss_log.db-wal
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 4148872
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9435af1a76e7fda76ac9afcb8fc545785fc91230eb9d5877fcb8525bbaa45486
|
| 3 |
size 4148872
|
optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a784454ac27624675c6e6e7d50ea180b0751afc3905b4be478581cdf36e2628c
|
| 3 |
+
size 623951333
|
pid.txt
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
|
|
|
|
| 1 |
+
11146
|
samples/1772099185860__000000000_0.jpg
ADDED
|
samples/1772099511556__000000250_0.jpg
ADDED
|
samples/1772099815995__000000500_0.jpg
ADDED
|
Git LFS Details
|
samples/1772100143209__000000750_0.jpg
ADDED
|
samples/1772100455705__000001000_0.jpg
ADDED
|
samples/1772100773170__000001250_0.jpg
ADDED
|
samples/1772101085673__000001500_0.jpg
ADDED
|