Add files using upload-large-folder tool
Browse files- .job_config.json +6 -6
- config.yaml +7 -8
- lana_lora_wan2.2_000000250_high_noise.safetensors +3 -0
- lana_lora_wan2.2_000000250_low_noise.safetensors +3 -0
- log.txt +0 -0
- loss_log.db +2 -2
- loss_log.db-shm +0 -0
- loss_log.db-wal +2 -2
- optimizer.pt +2 -2
- pid.txt +1 -1
- samples/1772008869870__000000250_0.jpg +0 -0
.job_config.json
CHANGED
|
@@ -1,7 +1,7 @@
|
|
| 1 |
{
|
| 2 |
"job": "extension",
|
| 3 |
"config": {
|
| 4 |
-
"name": "
|
| 5 |
"process": [
|
| 6 |
{
|
| 7 |
"type": "diffusion_trainer",
|
|
@@ -70,7 +70,7 @@
|
|
| 70 |
},
|
| 71 |
"unload_text_encoder": false,
|
| 72 |
"cache_text_embeddings": false,
|
| 73 |
-
"lr": 0.
|
| 74 |
"ema_config": {
|
| 75 |
"use_ema": false,
|
| 76 |
"ema_decay": 0.99
|
|
@@ -93,12 +93,12 @@
|
|
| 93 |
"name_or_path": "ai-toolkit/Wan2.2-T2V-A14B-Diffusers-bf16",
|
| 94 |
"quantize": false,
|
| 95 |
"qtype": "qfloat8",
|
| 96 |
-
"quantize_te":
|
| 97 |
"qtype_te": "qfloat8",
|
| 98 |
"arch": "wan22_14b:t2v",
|
| 99 |
"low_vram": false,
|
| 100 |
"model_kwargs": {
|
| 101 |
-
"train_high_noise":
|
| 102 |
"train_low_noise": true
|
| 103 |
},
|
| 104 |
"layer_offloading": false,
|
|
@@ -112,14 +112,14 @@
|
|
| 112 |
"height": 1024,
|
| 113 |
"samples": [
|
| 114 |
{
|
| 115 |
-
"prompt": "l4n4lux
|
| 116 |
}
|
| 117 |
],
|
| 118 |
"neg": "",
|
| 119 |
"seed": 42,
|
| 120 |
"walk_seed": true,
|
| 121 |
"guidance_scale": 4,
|
| 122 |
-
"sample_steps":
|
| 123 |
"num_frames": 1,
|
| 124 |
"fps": 1
|
| 125 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"job": "extension",
|
| 3 |
"config": {
|
| 4 |
+
"name": "lana_lora_wan2.2",
|
| 5 |
"process": [
|
| 6 |
{
|
| 7 |
"type": "diffusion_trainer",
|
|
|
|
| 70 |
},
|
| 71 |
"unload_text_encoder": false,
|
| 72 |
"cache_text_embeddings": false,
|
| 73 |
+
"lr": 0.0002,
|
| 74 |
"ema_config": {
|
| 75 |
"use_ema": false,
|
| 76 |
"ema_decay": 0.99
|
|
|
|
| 93 |
"name_or_path": "ai-toolkit/Wan2.2-T2V-A14B-Diffusers-bf16",
|
| 94 |
"quantize": false,
|
| 95 |
"qtype": "qfloat8",
|
| 96 |
+
"quantize_te": true,
|
| 97 |
"qtype_te": "qfloat8",
|
| 98 |
"arch": "wan22_14b:t2v",
|
| 99 |
"low_vram": false,
|
| 100 |
"model_kwargs": {
|
| 101 |
+
"train_high_noise": true,
|
| 102 |
"train_low_noise": true
|
| 103 |
},
|
| 104 |
"layer_offloading": false,
|
|
|
|
| 112 |
"height": 1024,
|
| 113 |
"samples": [
|
| 114 |
{
|
| 115 |
+
"prompt": "l4n4lux"
|
| 116 |
}
|
| 117 |
],
|
| 118 |
"neg": "",
|
| 119 |
"seed": 42,
|
| 120 |
"walk_seed": true,
|
| 121 |
"guidance_scale": 4,
|
| 122 |
+
"sample_steps": 25,
|
| 123 |
"num_frames": 1,
|
| 124 |
"fps": 1
|
| 125 |
}
|
config.yaml
CHANGED
|
@@ -1,6 +1,6 @@
|
|
| 1 |
job: extension
|
| 2 |
config:
|
| 3 |
-
name:
|
| 4 |
process:
|
| 5 |
- type: diffusion_trainer
|
| 6 |
training_folder: /root/lab/ai-toolkit/output
|
|
@@ -60,7 +60,7 @@ config:
|
|
| 60 |
weight_decay: 0.0001
|
| 61 |
unload_text_encoder: false
|
| 62 |
cache_text_embeddings: false
|
| 63 |
-
lr: 0.
|
| 64 |
ema_config:
|
| 65 |
use_ema: false
|
| 66 |
ema_decay: 0.99
|
|
@@ -80,12 +80,12 @@ config:
|
|
| 80 |
name_or_path: ai-toolkit/Wan2.2-T2V-A14B-Diffusers-bf16
|
| 81 |
quantize: false
|
| 82 |
qtype: qfloat8
|
| 83 |
-
quantize_te:
|
| 84 |
qtype_te: qfloat8
|
| 85 |
arch: wan22_14b:t2v
|
| 86 |
low_vram: false
|
| 87 |
model_kwargs:
|
| 88 |
-
train_high_noise:
|
| 89 |
train_low_noise: true
|
| 90 |
layer_offloading: false
|
| 91 |
layer_offloading_text_encoder_percent: 1
|
|
@@ -96,15 +96,14 @@ config:
|
|
| 96 |
width: 1024
|
| 97 |
height: 1024
|
| 98 |
samples:
|
| 99 |
-
- prompt: l4n4lux
|
| 100 |
-
off in the background
|
| 101 |
neg: ''
|
| 102 |
seed: 42
|
| 103 |
walk_seed: true
|
| 104 |
guidance_scale: 4
|
| 105 |
-
sample_steps:
|
| 106 |
num_frames: 1
|
| 107 |
fps: 1
|
| 108 |
meta:
|
| 109 |
-
name:
|
| 110 |
version: '1.0'
|
|
|
|
| 1 |
job: extension
|
| 2 |
config:
|
| 3 |
+
name: lana_lora_wan2.2
|
| 4 |
process:
|
| 5 |
- type: diffusion_trainer
|
| 6 |
training_folder: /root/lab/ai-toolkit/output
|
|
|
|
| 60 |
weight_decay: 0.0001
|
| 61 |
unload_text_encoder: false
|
| 62 |
cache_text_embeddings: false
|
| 63 |
+
lr: 0.0002
|
| 64 |
ema_config:
|
| 65 |
use_ema: false
|
| 66 |
ema_decay: 0.99
|
|
|
|
| 80 |
name_or_path: ai-toolkit/Wan2.2-T2V-A14B-Diffusers-bf16
|
| 81 |
quantize: false
|
| 82 |
qtype: qfloat8
|
| 83 |
+
quantize_te: true
|
| 84 |
qtype_te: qfloat8
|
| 85 |
arch: wan22_14b:t2v
|
| 86 |
low_vram: false
|
| 87 |
model_kwargs:
|
| 88 |
+
train_high_noise: true
|
| 89 |
train_low_noise: true
|
| 90 |
layer_offloading: false
|
| 91 |
layer_offloading_text_encoder_percent: 1
|
|
|
|
| 96 |
width: 1024
|
| 97 |
height: 1024
|
| 98 |
samples:
|
| 99 |
+
- prompt: l4n4lux
|
|
|
|
| 100 |
neg: ''
|
| 101 |
seed: 42
|
| 102 |
walk_seed: true
|
| 103 |
guidance_scale: 4
|
| 104 |
+
sample_steps: 25
|
| 105 |
num_frames: 1
|
| 106 |
fps: 1
|
| 107 |
meta:
|
| 108 |
+
name: lana_lora_wan2.2
|
| 109 |
version: '1.0'
|
lana_lora_wan2.2_000000250_high_noise.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:489aa10eca8bb1b16a4e016b221d60650c10c6894d037f51edafff29f0d65513
|
| 3 |
+
size 613517200
|
lana_lora_wan2.2_000000250_low_noise.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7f75df0b0eb0598566910c9871e9b2c654b8fd0821dd7446ef3ed3c9185c2d48
|
| 3 |
+
size 613517200
|
log.txt
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|
loss_log.db
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9d5c6f0cd884ac1803325090a03697958f2b068c376281c1aabdbd3f0c0c4f78
|
| 3 |
+
size 57344
|
loss_log.db-shm
CHANGED
|
Binary files a/loss_log.db-shm and b/loss_log.db-shm differ
|
|
|
loss_log.db-wal
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1c3560114d7cefa3982f943229125dca4e853f0d1a2b82185d8bf3c379b9e501
|
| 3 |
+
size 4136512
|
optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4e6abdd9e39a2eb4503b2f6f6b21f8514a2e66f653a40a645871f657ea3ed3a0
|
| 3 |
+
size 1247900133
|
pid.txt
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
|
|
|
|
| 1 |
+
11172
|
samples/1772008869870__000000250_0.jpg
ADDED
|