Commit
·
8ff0ce1
0
Parent(s):
Duplicate from zatochu/EasyFluff
Browse filesCo-authored-by: zatochu <zatochu@users.noreply.huggingface.co>
This view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +35 -0
- EasyFluff.safetensors +3 -0
- EasyFluff.yaml +72 -0
- EasyFluffExperimental.safetensors +3 -0
- EasyFluffExperimental.yaml +72 -0
- EasyFluffV10-FunEdition.safetensors +3 -0
- EasyFluffV10-FunEdition.yaml +72 -0
- EasyFluffV10-FunEditionAlt.safetensors +3 -0
- EasyFluffV10-FunEditionAlt.yaml +72 -0
- EasyFluffV10-FunnerEdition.safetensors +3 -0
- EasyFluffV10-FunnerEdition.yaml +72 -0
- EasyFluffV10-PreRelease.safetensors +3 -0
- EasyFluffV10-PreRelease.yaml +72 -0
- EasyFluffV10-PreRelease2.safetensors +3 -0
- EasyFluffV10-PreRelease2.yaml +72 -0
- EasyFluffV10.1.safetensors +3 -0
- EasyFluffV10.1.yaml +72 -0
- EasyFluffV10.safetensors +3 -0
- EasyFluffV10.yaml +72 -0
- EasyFluffV11.1.safetensors +3 -0
- EasyFluffV11.1.yaml +72 -0
- EasyFluffV11.2.safetensors +3 -0
- EasyFluffV11.2.yaml +72 -0
- EasyFluffV11.safetensors +3 -0
- EasyFluffV11.yaml +72 -0
- EasyFluffV2.1.safetensors +3 -0
- EasyFluffV2.1.yaml +72 -0
- EasyFluffV2.safetensors +3 -0
- EasyFluffV2.yaml +72 -0
- EasyFluffV3-LessCreativeMoreAccurate.safetensors +3 -0
- EasyFluffV3-LessCreativeMoreAccurate.yaml +72 -0
- EasyFluffV3.safetensors +3 -0
- EasyFluffV3.yaml +72 -0
- EasyFluffV4-LessCreativeMoreAccurate.safetensors +3 -0
- EasyFluffV4-LessCreativeMoreAccurate.yaml +72 -0
- EasyFluffV4.safetensors +3 -0
- EasyFluffV4.yaml +72 -0
- EasyFluffV5.safetensors +3 -0
- EasyFluffV5.yaml +72 -0
- EasyFluffV6-OnlyFR-NoPF.safetensors +3 -0
- EasyFluffV6-OnlyFR-NoPF.yaml +72 -0
- EasyFluffV6.safetensors +3 -0
- EasyFluffV6.yaml +72 -0
- EasyFluffV7.safetensors +3 -0
- EasyFluffV7.yaml +72 -0
- EasyFluffV8.1.safetensors +3 -0
- EasyFluffV8.1.yaml +72 -0
- EasyFluffV8.safetensors +3 -0
- EasyFluffV8.yaml +72 -0
- EasyFluffV9.safetensors +3 -0
.gitattributes
ADDED
|
@@ -0,0 +1,35 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
*.7z filter=lfs diff=lfs merge=lfs -text
|
| 2 |
+
*.arrow filter=lfs diff=lfs merge=lfs -text
|
| 3 |
+
*.bin filter=lfs diff=lfs merge=lfs -text
|
| 4 |
+
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
| 5 |
+
*.ckpt filter=lfs diff=lfs merge=lfs -text
|
| 6 |
+
*.ftz filter=lfs diff=lfs merge=lfs -text
|
| 7 |
+
*.gz filter=lfs diff=lfs merge=lfs -text
|
| 8 |
+
*.h5 filter=lfs diff=lfs merge=lfs -text
|
| 9 |
+
*.joblib filter=lfs diff=lfs merge=lfs -text
|
| 10 |
+
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
| 11 |
+
*.mlmodel filter=lfs diff=lfs merge=lfs -text
|
| 12 |
+
*.model filter=lfs diff=lfs merge=lfs -text
|
| 13 |
+
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
| 14 |
+
*.npy filter=lfs diff=lfs merge=lfs -text
|
| 15 |
+
*.npz filter=lfs diff=lfs merge=lfs -text
|
| 16 |
+
*.onnx filter=lfs diff=lfs merge=lfs -text
|
| 17 |
+
*.ot filter=lfs diff=lfs merge=lfs -text
|
| 18 |
+
*.parquet filter=lfs diff=lfs merge=lfs -text
|
| 19 |
+
*.pb filter=lfs diff=lfs merge=lfs -text
|
| 20 |
+
*.pickle filter=lfs diff=lfs merge=lfs -text
|
| 21 |
+
*.pkl filter=lfs diff=lfs merge=lfs -text
|
| 22 |
+
*.pt filter=lfs diff=lfs merge=lfs -text
|
| 23 |
+
*.pth filter=lfs diff=lfs merge=lfs -text
|
| 24 |
+
*.rar filter=lfs diff=lfs merge=lfs -text
|
| 25 |
+
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
| 26 |
+
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
| 27 |
+
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
| 28 |
+
*.tar filter=lfs diff=lfs merge=lfs -text
|
| 29 |
+
*.tflite filter=lfs diff=lfs merge=lfs -text
|
| 30 |
+
*.tgz filter=lfs diff=lfs merge=lfs -text
|
| 31 |
+
*.wasm filter=lfs diff=lfs merge=lfs -text
|
| 32 |
+
*.xz filter=lfs diff=lfs merge=lfs -text
|
| 33 |
+
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
+
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
+
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
EasyFluff.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:322a67259a182ea3d27d62e789cd513dac85d42f68ff5d98df3a5582ad178a74
|
| 3 |
+
size 2132634592
|
EasyFluff.yaml
ADDED
|
@@ -0,0 +1,72 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
model:
|
| 2 |
+
base_learning_rate: 1.0e-04
|
| 3 |
+
target: ldm.models.diffusion.ddpm.LatentDiffusion
|
| 4 |
+
params:
|
| 5 |
+
parameterization: "v"
|
| 6 |
+
linear_start: 0.00085
|
| 7 |
+
linear_end: 0.0120
|
| 8 |
+
num_timesteps_cond: 1
|
| 9 |
+
log_every_t: 200
|
| 10 |
+
timesteps: 1000
|
| 11 |
+
first_stage_key: "jpg"
|
| 12 |
+
cond_stage_key: "txt"
|
| 13 |
+
image_size: 64
|
| 14 |
+
channels: 4
|
| 15 |
+
cond_stage_trainable: false # Note: different from the one we trained before
|
| 16 |
+
conditioning_key: crossattn
|
| 17 |
+
monitor: val/loss_simple_ema
|
| 18 |
+
scale_factor: 0.18215
|
| 19 |
+
use_ema: False
|
| 20 |
+
|
| 21 |
+
scheduler_config: # 10000 warmup steps
|
| 22 |
+
target: ldm.lr_scheduler.LambdaLinearScheduler
|
| 23 |
+
params:
|
| 24 |
+
warm_up_steps: [ 10000 ]
|
| 25 |
+
cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
|
| 26 |
+
f_start: [ 1.e-6 ]
|
| 27 |
+
f_max: [ 1. ]
|
| 28 |
+
f_min: [ 1. ]
|
| 29 |
+
|
| 30 |
+
unet_config:
|
| 31 |
+
target: ldm.modules.diffusionmodules.openaimodel.UNetModel
|
| 32 |
+
params:
|
| 33 |
+
use_fp16: True
|
| 34 |
+
image_size: 32 # unused
|
| 35 |
+
in_channels: 4
|
| 36 |
+
out_channels: 4
|
| 37 |
+
model_channels: 320
|
| 38 |
+
attention_resolutions: [ 4, 2, 1 ]
|
| 39 |
+
num_res_blocks: 2
|
| 40 |
+
channel_mult: [ 1, 2, 4, 4 ]
|
| 41 |
+
num_heads: 8
|
| 42 |
+
use_spatial_transformer: True
|
| 43 |
+
transformer_depth: 1
|
| 44 |
+
context_dim: 768
|
| 45 |
+
use_checkpoint: True
|
| 46 |
+
legacy: False
|
| 47 |
+
|
| 48 |
+
first_stage_config:
|
| 49 |
+
target: ldm.models.autoencoder.AutoencoderKL
|
| 50 |
+
params:
|
| 51 |
+
embed_dim: 4
|
| 52 |
+
monitor: val/rec_loss
|
| 53 |
+
ddconfig:
|
| 54 |
+
double_z: true
|
| 55 |
+
z_channels: 4
|
| 56 |
+
resolution: 256
|
| 57 |
+
in_channels: 3
|
| 58 |
+
out_ch: 3
|
| 59 |
+
ch: 128
|
| 60 |
+
ch_mult:
|
| 61 |
+
- 1
|
| 62 |
+
- 2
|
| 63 |
+
- 4
|
| 64 |
+
- 4
|
| 65 |
+
num_res_blocks: 2
|
| 66 |
+
attn_resolutions: []
|
| 67 |
+
dropout: 0.0
|
| 68 |
+
lossconfig:
|
| 69 |
+
target: torch.nn.Identity
|
| 70 |
+
|
| 71 |
+
cond_stage_config:
|
| 72 |
+
target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
|
EasyFluffExperimental.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:792a6336bdf4000fd44e0061ec59ae1304836c1e1b943e663c69094e13b67290
|
| 3 |
+
size 2132626098
|
EasyFluffExperimental.yaml
ADDED
|
@@ -0,0 +1,72 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
model:
|
| 2 |
+
base_learning_rate: 1.0e-04
|
| 3 |
+
target: ldm.models.diffusion.ddpm.LatentDiffusion
|
| 4 |
+
params:
|
| 5 |
+
parameterization: "v"
|
| 6 |
+
linear_start: 0.00085
|
| 7 |
+
linear_end: 0.0120
|
| 8 |
+
num_timesteps_cond: 1
|
| 9 |
+
log_every_t: 200
|
| 10 |
+
timesteps: 1000
|
| 11 |
+
first_stage_key: "jpg"
|
| 12 |
+
cond_stage_key: "txt"
|
| 13 |
+
image_size: 64
|
| 14 |
+
channels: 4
|
| 15 |
+
cond_stage_trainable: false # Note: different from the one we trained before
|
| 16 |
+
conditioning_key: crossattn
|
| 17 |
+
monitor: val/loss_simple_ema
|
| 18 |
+
scale_factor: 0.18215
|
| 19 |
+
use_ema: False
|
| 20 |
+
|
| 21 |
+
scheduler_config: # 10000 warmup steps
|
| 22 |
+
target: ldm.lr_scheduler.LambdaLinearScheduler
|
| 23 |
+
params:
|
| 24 |
+
warm_up_steps: [ 10000 ]
|
| 25 |
+
cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
|
| 26 |
+
f_start: [ 1.e-6 ]
|
| 27 |
+
f_max: [ 1. ]
|
| 28 |
+
f_min: [ 1. ]
|
| 29 |
+
|
| 30 |
+
unet_config:
|
| 31 |
+
target: ldm.modules.diffusionmodules.openaimodel.UNetModel
|
| 32 |
+
params:
|
| 33 |
+
use_fp16: True
|
| 34 |
+
image_size: 32 # unused
|
| 35 |
+
in_channels: 4
|
| 36 |
+
out_channels: 4
|
| 37 |
+
model_channels: 320
|
| 38 |
+
attention_resolutions: [ 4, 2, 1 ]
|
| 39 |
+
num_res_blocks: 2
|
| 40 |
+
channel_mult: [ 1, 2, 4, 4 ]
|
| 41 |
+
num_heads: 8
|
| 42 |
+
use_spatial_transformer: True
|
| 43 |
+
transformer_depth: 1
|
| 44 |
+
context_dim: 768
|
| 45 |
+
use_checkpoint: True
|
| 46 |
+
legacy: False
|
| 47 |
+
|
| 48 |
+
first_stage_config:
|
| 49 |
+
target: ldm.models.autoencoder.AutoencoderKL
|
| 50 |
+
params:
|
| 51 |
+
embed_dim: 4
|
| 52 |
+
monitor: val/rec_loss
|
| 53 |
+
ddconfig:
|
| 54 |
+
double_z: true
|
| 55 |
+
z_channels: 4
|
| 56 |
+
resolution: 256
|
| 57 |
+
in_channels: 3
|
| 58 |
+
out_ch: 3
|
| 59 |
+
ch: 128
|
| 60 |
+
ch_mult:
|
| 61 |
+
- 1
|
| 62 |
+
- 2
|
| 63 |
+
- 4
|
| 64 |
+
- 4
|
| 65 |
+
num_res_blocks: 2
|
| 66 |
+
attn_resolutions: []
|
| 67 |
+
dropout: 0.0
|
| 68 |
+
lossconfig:
|
| 69 |
+
target: torch.nn.Identity
|
| 70 |
+
|
| 71 |
+
cond_stage_config:
|
| 72 |
+
target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
|
EasyFluffV10-FunEdition.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4e7583769c0f75031f689f2d703ab26f090499d4e3075626ed7eaa7ee747c86b
|
| 3 |
+
size 2132646592
|
EasyFluffV10-FunEdition.yaml
ADDED
|
@@ -0,0 +1,72 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
model:
|
| 2 |
+
base_learning_rate: 1.0e-04
|
| 3 |
+
target: ldm.models.diffusion.ddpm.LatentDiffusion
|
| 4 |
+
params:
|
| 5 |
+
parameterization: "v"
|
| 6 |
+
linear_start: 0.00085
|
| 7 |
+
linear_end: 0.0120
|
| 8 |
+
num_timesteps_cond: 1
|
| 9 |
+
log_every_t: 200
|
| 10 |
+
timesteps: 1000
|
| 11 |
+
first_stage_key: "jpg"
|
| 12 |
+
cond_stage_key: "txt"
|
| 13 |
+
image_size: 64
|
| 14 |
+
channels: 4
|
| 15 |
+
cond_stage_trainable: false # Note: different from the one we trained before
|
| 16 |
+
conditioning_key: crossattn
|
| 17 |
+
monitor: val/loss_simple_ema
|
| 18 |
+
scale_factor: 0.18215
|
| 19 |
+
use_ema: False
|
| 20 |
+
|
| 21 |
+
scheduler_config: # 10000 warmup steps
|
| 22 |
+
target: ldm.lr_scheduler.LambdaLinearScheduler
|
| 23 |
+
params:
|
| 24 |
+
warm_up_steps: [ 10000 ]
|
| 25 |
+
cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
|
| 26 |
+
f_start: [ 1.e-6 ]
|
| 27 |
+
f_max: [ 1. ]
|
| 28 |
+
f_min: [ 1. ]
|
| 29 |
+
|
| 30 |
+
unet_config:
|
| 31 |
+
target: ldm.modules.diffusionmodules.openaimodel.UNetModel
|
| 32 |
+
params:
|
| 33 |
+
use_fp16: True
|
| 34 |
+
image_size: 32 # unused
|
| 35 |
+
in_channels: 4
|
| 36 |
+
out_channels: 4
|
| 37 |
+
model_channels: 320
|
| 38 |
+
attention_resolutions: [ 4, 2, 1 ]
|
| 39 |
+
num_res_blocks: 2
|
| 40 |
+
channel_mult: [ 1, 2, 4, 4 ]
|
| 41 |
+
num_heads: 8
|
| 42 |
+
use_spatial_transformer: True
|
| 43 |
+
transformer_depth: 1
|
| 44 |
+
context_dim: 768
|
| 45 |
+
use_checkpoint: True
|
| 46 |
+
legacy: False
|
| 47 |
+
|
| 48 |
+
first_stage_config:
|
| 49 |
+
target: ldm.models.autoencoder.AutoencoderKL
|
| 50 |
+
params:
|
| 51 |
+
embed_dim: 4
|
| 52 |
+
monitor: val/rec_loss
|
| 53 |
+
ddconfig:
|
| 54 |
+
double_z: true
|
| 55 |
+
z_channels: 4
|
| 56 |
+
resolution: 256
|
| 57 |
+
in_channels: 3
|
| 58 |
+
out_ch: 3
|
| 59 |
+
ch: 128
|
| 60 |
+
ch_mult:
|
| 61 |
+
- 1
|
| 62 |
+
- 2
|
| 63 |
+
- 4
|
| 64 |
+
- 4
|
| 65 |
+
num_res_blocks: 2
|
| 66 |
+
attn_resolutions: []
|
| 67 |
+
dropout: 0.0
|
| 68 |
+
lossconfig:
|
| 69 |
+
target: torch.nn.Identity
|
| 70 |
+
|
| 71 |
+
cond_stage_config:
|
| 72 |
+
target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
|
EasyFluffV10-FunEditionAlt.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0dc78c95b0226530e205bd91ebd7048c893ea551b723f5546e3ca53c804c2873
|
| 3 |
+
size 2132648304
|
EasyFluffV10-FunEditionAlt.yaml
ADDED
|
@@ -0,0 +1,72 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
model:
|
| 2 |
+
base_learning_rate: 1.0e-04
|
| 3 |
+
target: ldm.models.diffusion.ddpm.LatentDiffusion
|
| 4 |
+
params:
|
| 5 |
+
parameterization: "v"
|
| 6 |
+
linear_start: 0.00085
|
| 7 |
+
linear_end: 0.0120
|
| 8 |
+
num_timesteps_cond: 1
|
| 9 |
+
log_every_t: 200
|
| 10 |
+
timesteps: 1000
|
| 11 |
+
first_stage_key: "jpg"
|
| 12 |
+
cond_stage_key: "txt"
|
| 13 |
+
image_size: 64
|
| 14 |
+
channels: 4
|
| 15 |
+
cond_stage_trainable: false # Note: different from the one we trained before
|
| 16 |
+
conditioning_key: crossattn
|
| 17 |
+
monitor: val/loss_simple_ema
|
| 18 |
+
scale_factor: 0.18215
|
| 19 |
+
use_ema: False
|
| 20 |
+
|
| 21 |
+
scheduler_config: # 10000 warmup steps
|
| 22 |
+
target: ldm.lr_scheduler.LambdaLinearScheduler
|
| 23 |
+
params:
|
| 24 |
+
warm_up_steps: [ 10000 ]
|
| 25 |
+
cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
|
| 26 |
+
f_start: [ 1.e-6 ]
|
| 27 |
+
f_max: [ 1. ]
|
| 28 |
+
f_min: [ 1. ]
|
| 29 |
+
|
| 30 |
+
unet_config:
|
| 31 |
+
target: ldm.modules.diffusionmodules.openaimodel.UNetModel
|
| 32 |
+
params:
|
| 33 |
+
use_fp16: True
|
| 34 |
+
image_size: 32 # unused
|
| 35 |
+
in_channels: 4
|
| 36 |
+
out_channels: 4
|
| 37 |
+
model_channels: 320
|
| 38 |
+
attention_resolutions: [ 4, 2, 1 ]
|
| 39 |
+
num_res_blocks: 2
|
| 40 |
+
channel_mult: [ 1, 2, 4, 4 ]
|
| 41 |
+
num_heads: 8
|
| 42 |
+
use_spatial_transformer: True
|
| 43 |
+
transformer_depth: 1
|
| 44 |
+
context_dim: 768
|
| 45 |
+
use_checkpoint: True
|
| 46 |
+
legacy: False
|
| 47 |
+
|
| 48 |
+
first_stage_config:
|
| 49 |
+
target: ldm.models.autoencoder.AutoencoderKL
|
| 50 |
+
params:
|
| 51 |
+
embed_dim: 4
|
| 52 |
+
monitor: val/rec_loss
|
| 53 |
+
ddconfig:
|
| 54 |
+
double_z: true
|
| 55 |
+
z_channels: 4
|
| 56 |
+
resolution: 256
|
| 57 |
+
in_channels: 3
|
| 58 |
+
out_ch: 3
|
| 59 |
+
ch: 128
|
| 60 |
+
ch_mult:
|
| 61 |
+
- 1
|
| 62 |
+
- 2
|
| 63 |
+
- 4
|
| 64 |
+
- 4
|
| 65 |
+
num_res_blocks: 2
|
| 66 |
+
attn_resolutions: []
|
| 67 |
+
dropout: 0.0
|
| 68 |
+
lossconfig:
|
| 69 |
+
target: torch.nn.Identity
|
| 70 |
+
|
| 71 |
+
cond_stage_config:
|
| 72 |
+
target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
|
EasyFluffV10-FunnerEdition.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dddf3bc713c462ff32c97d47b9aaee6093b7ce0841a5af1d894d903f588f8019
|
| 3 |
+
size 2132649480
|
EasyFluffV10-FunnerEdition.yaml
ADDED
|
@@ -0,0 +1,72 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
model:
|
| 2 |
+
base_learning_rate: 1.0e-04
|
| 3 |
+
target: ldm.models.diffusion.ddpm.LatentDiffusion
|
| 4 |
+
params:
|
| 5 |
+
parameterization: "v"
|
| 6 |
+
linear_start: 0.00085
|
| 7 |
+
linear_end: 0.0120
|
| 8 |
+
num_timesteps_cond: 1
|
| 9 |
+
log_every_t: 200
|
| 10 |
+
timesteps: 1000
|
| 11 |
+
first_stage_key: "jpg"
|
| 12 |
+
cond_stage_key: "txt"
|
| 13 |
+
image_size: 64
|
| 14 |
+
channels: 4
|
| 15 |
+
cond_stage_trainable: false # Note: different from the one we trained before
|
| 16 |
+
conditioning_key: crossattn
|
| 17 |
+
monitor: val/loss_simple_ema
|
| 18 |
+
scale_factor: 0.18215
|
| 19 |
+
use_ema: False
|
| 20 |
+
|
| 21 |
+
scheduler_config: # 10000 warmup steps
|
| 22 |
+
target: ldm.lr_scheduler.LambdaLinearScheduler
|
| 23 |
+
params:
|
| 24 |
+
warm_up_steps: [ 10000 ]
|
| 25 |
+
cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
|
| 26 |
+
f_start: [ 1.e-6 ]
|
| 27 |
+
f_max: [ 1. ]
|
| 28 |
+
f_min: [ 1. ]
|
| 29 |
+
|
| 30 |
+
unet_config:
|
| 31 |
+
target: ldm.modules.diffusionmodules.openaimodel.UNetModel
|
| 32 |
+
params:
|
| 33 |
+
use_fp16: True
|
| 34 |
+
image_size: 32 # unused
|
| 35 |
+
in_channels: 4
|
| 36 |
+
out_channels: 4
|
| 37 |
+
model_channels: 320
|
| 38 |
+
attention_resolutions: [ 4, 2, 1 ]
|
| 39 |
+
num_res_blocks: 2
|
| 40 |
+
channel_mult: [ 1, 2, 4, 4 ]
|
| 41 |
+
num_heads: 8
|
| 42 |
+
use_spatial_transformer: True
|
| 43 |
+
transformer_depth: 1
|
| 44 |
+
context_dim: 768
|
| 45 |
+
use_checkpoint: True
|
| 46 |
+
legacy: False
|
| 47 |
+
|
| 48 |
+
first_stage_config:
|
| 49 |
+
target: ldm.models.autoencoder.AutoencoderKL
|
| 50 |
+
params:
|
| 51 |
+
embed_dim: 4
|
| 52 |
+
monitor: val/rec_loss
|
| 53 |
+
ddconfig:
|
| 54 |
+
double_z: true
|
| 55 |
+
z_channels: 4
|
| 56 |
+
resolution: 256
|
| 57 |
+
in_channels: 3
|
| 58 |
+
out_ch: 3
|
| 59 |
+
ch: 128
|
| 60 |
+
ch_mult:
|
| 61 |
+
- 1
|
| 62 |
+
- 2
|
| 63 |
+
- 4
|
| 64 |
+
- 4
|
| 65 |
+
num_res_blocks: 2
|
| 66 |
+
attn_resolutions: []
|
| 67 |
+
dropout: 0.0
|
| 68 |
+
lossconfig:
|
| 69 |
+
target: torch.nn.Identity
|
| 70 |
+
|
| 71 |
+
cond_stage_config:
|
| 72 |
+
target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
|
EasyFluffV10-PreRelease.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8101105d3db43b276aeccad54b7854cad487b5d506f9079c47cdf70d622c6e18
|
| 3 |
+
size 2132657048
|
EasyFluffV10-PreRelease.yaml
ADDED
|
@@ -0,0 +1,72 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
model:
|
| 2 |
+
base_learning_rate: 1.0e-04
|
| 3 |
+
target: ldm.models.diffusion.ddpm.LatentDiffusion
|
| 4 |
+
params:
|
| 5 |
+
parameterization: "v"
|
| 6 |
+
linear_start: 0.00085
|
| 7 |
+
linear_end: 0.0120
|
| 8 |
+
num_timesteps_cond: 1
|
| 9 |
+
log_every_t: 200
|
| 10 |
+
timesteps: 1000
|
| 11 |
+
first_stage_key: "jpg"
|
| 12 |
+
cond_stage_key: "txt"
|
| 13 |
+
image_size: 64
|
| 14 |
+
channels: 4
|
| 15 |
+
cond_stage_trainable: false # Note: different from the one we trained before
|
| 16 |
+
conditioning_key: crossattn
|
| 17 |
+
monitor: val/loss_simple_ema
|
| 18 |
+
scale_factor: 0.18215
|
| 19 |
+
use_ema: False
|
| 20 |
+
|
| 21 |
+
scheduler_config: # 10000 warmup steps
|
| 22 |
+
target: ldm.lr_scheduler.LambdaLinearScheduler
|
| 23 |
+
params:
|
| 24 |
+
warm_up_steps: [ 10000 ]
|
| 25 |
+
cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
|
| 26 |
+
f_start: [ 1.e-6 ]
|
| 27 |
+
f_max: [ 1. ]
|
| 28 |
+
f_min: [ 1. ]
|
| 29 |
+
|
| 30 |
+
unet_config:
|
| 31 |
+
target: ldm.modules.diffusionmodules.openaimodel.UNetModel
|
| 32 |
+
params:
|
| 33 |
+
use_fp16: True
|
| 34 |
+
image_size: 32 # unused
|
| 35 |
+
in_channels: 4
|
| 36 |
+
out_channels: 4
|
| 37 |
+
model_channels: 320
|
| 38 |
+
attention_resolutions: [ 4, 2, 1 ]
|
| 39 |
+
num_res_blocks: 2
|
| 40 |
+
channel_mult: [ 1, 2, 4, 4 ]
|
| 41 |
+
num_heads: 8
|
| 42 |
+
use_spatial_transformer: True
|
| 43 |
+
transformer_depth: 1
|
| 44 |
+
context_dim: 768
|
| 45 |
+
use_checkpoint: True
|
| 46 |
+
legacy: False
|
| 47 |
+
|
| 48 |
+
first_stage_config:
|
| 49 |
+
target: ldm.models.autoencoder.AutoencoderKL
|
| 50 |
+
params:
|
| 51 |
+
embed_dim: 4
|
| 52 |
+
monitor: val/rec_loss
|
| 53 |
+
ddconfig:
|
| 54 |
+
double_z: true
|
| 55 |
+
z_channels: 4
|
| 56 |
+
resolution: 256
|
| 57 |
+
in_channels: 3
|
| 58 |
+
out_ch: 3
|
| 59 |
+
ch: 128
|
| 60 |
+
ch_mult:
|
| 61 |
+
- 1
|
| 62 |
+
- 2
|
| 63 |
+
- 4
|
| 64 |
+
- 4
|
| 65 |
+
num_res_blocks: 2
|
| 66 |
+
attn_resolutions: []
|
| 67 |
+
dropout: 0.0
|
| 68 |
+
lossconfig:
|
| 69 |
+
target: torch.nn.Identity
|
| 70 |
+
|
| 71 |
+
cond_stage_config:
|
| 72 |
+
target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
|
EasyFluffV10-PreRelease2.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4685c7fd2f7efe92d346ef6f7b2d6b29282a8c20cffcd1baf8a9733abf16840a
|
| 3 |
+
size 2132662528
|
EasyFluffV10-PreRelease2.yaml
ADDED
|
@@ -0,0 +1,72 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
model:
|
| 2 |
+
base_learning_rate: 1.0e-04
|
| 3 |
+
target: ldm.models.diffusion.ddpm.LatentDiffusion
|
| 4 |
+
params:
|
| 5 |
+
parameterization: "v"
|
| 6 |
+
linear_start: 0.00085
|
| 7 |
+
linear_end: 0.0120
|
| 8 |
+
num_timesteps_cond: 1
|
| 9 |
+
log_every_t: 200
|
| 10 |
+
timesteps: 1000
|
| 11 |
+
first_stage_key: "jpg"
|
| 12 |
+
cond_stage_key: "txt"
|
| 13 |
+
image_size: 64
|
| 14 |
+
channels: 4
|
| 15 |
+
cond_stage_trainable: false # Note: different from the one we trained before
|
| 16 |
+
conditioning_key: crossattn
|
| 17 |
+
monitor: val/loss_simple_ema
|
| 18 |
+
scale_factor: 0.18215
|
| 19 |
+
use_ema: False
|
| 20 |
+
|
| 21 |
+
scheduler_config: # 10000 warmup steps
|
| 22 |
+
target: ldm.lr_scheduler.LambdaLinearScheduler
|
| 23 |
+
params:
|
| 24 |
+
warm_up_steps: [ 10000 ]
|
| 25 |
+
cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
|
| 26 |
+
f_start: [ 1.e-6 ]
|
| 27 |
+
f_max: [ 1. ]
|
| 28 |
+
f_min: [ 1. ]
|
| 29 |
+
|
| 30 |
+
unet_config:
|
| 31 |
+
target: ldm.modules.diffusionmodules.openaimodel.UNetModel
|
| 32 |
+
params:
|
| 33 |
+
use_fp16: True
|
| 34 |
+
image_size: 32 # unused
|
| 35 |
+
in_channels: 4
|
| 36 |
+
out_channels: 4
|
| 37 |
+
model_channels: 320
|
| 38 |
+
attention_resolutions: [ 4, 2, 1 ]
|
| 39 |
+
num_res_blocks: 2
|
| 40 |
+
channel_mult: [ 1, 2, 4, 4 ]
|
| 41 |
+
num_heads: 8
|
| 42 |
+
use_spatial_transformer: True
|
| 43 |
+
transformer_depth: 1
|
| 44 |
+
context_dim: 768
|
| 45 |
+
use_checkpoint: True
|
| 46 |
+
legacy: False
|
| 47 |
+
|
| 48 |
+
first_stage_config:
|
| 49 |
+
target: ldm.models.autoencoder.AutoencoderKL
|
| 50 |
+
params:
|
| 51 |
+
embed_dim: 4
|
| 52 |
+
monitor: val/rec_loss
|
| 53 |
+
ddconfig:
|
| 54 |
+
double_z: true
|
| 55 |
+
z_channels: 4
|
| 56 |
+
resolution: 256
|
| 57 |
+
in_channels: 3
|
| 58 |
+
out_ch: 3
|
| 59 |
+
ch: 128
|
| 60 |
+
ch_mult:
|
| 61 |
+
- 1
|
| 62 |
+
- 2
|
| 63 |
+
- 4
|
| 64 |
+
- 4
|
| 65 |
+
num_res_blocks: 2
|
| 66 |
+
attn_resolutions: []
|
| 67 |
+
dropout: 0.0
|
| 68 |
+
lossconfig:
|
| 69 |
+
target: torch.nn.Identity
|
| 70 |
+
|
| 71 |
+
cond_stage_config:
|
| 72 |
+
target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
|
EasyFluffV10.1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1ac4dcb22c46c59df45f2d6797c3e3ed650ab738f9a2ad45c182471a731f2784
|
| 3 |
+
size 2132637456
|
EasyFluffV10.1.yaml
ADDED
|
@@ -0,0 +1,72 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
model:
|
| 2 |
+
base_learning_rate: 1.0e-04
|
| 3 |
+
target: ldm.models.diffusion.ddpm.LatentDiffusion
|
| 4 |
+
params:
|
| 5 |
+
parameterization: "v"
|
| 6 |
+
linear_start: 0.00085
|
| 7 |
+
linear_end: 0.0120
|
| 8 |
+
num_timesteps_cond: 1
|
| 9 |
+
log_every_t: 200
|
| 10 |
+
timesteps: 1000
|
| 11 |
+
first_stage_key: "jpg"
|
| 12 |
+
cond_stage_key: "txt"
|
| 13 |
+
image_size: 64
|
| 14 |
+
channels: 4
|
| 15 |
+
cond_stage_trainable: false # Note: different from the one we trained before
|
| 16 |
+
conditioning_key: crossattn
|
| 17 |
+
monitor: val/loss_simple_ema
|
| 18 |
+
scale_factor: 0.18215
|
| 19 |
+
use_ema: False
|
| 20 |
+
|
| 21 |
+
scheduler_config: # 10000 warmup steps
|
| 22 |
+
target: ldm.lr_scheduler.LambdaLinearScheduler
|
| 23 |
+
params:
|
| 24 |
+
warm_up_steps: [ 10000 ]
|
| 25 |
+
cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
|
| 26 |
+
f_start: [ 1.e-6 ]
|
| 27 |
+
f_max: [ 1. ]
|
| 28 |
+
f_min: [ 1. ]
|
| 29 |
+
|
| 30 |
+
unet_config:
|
| 31 |
+
target: ldm.modules.diffusionmodules.openaimodel.UNetModel
|
| 32 |
+
params:
|
| 33 |
+
use_fp16: True
|
| 34 |
+
image_size: 32 # unused
|
| 35 |
+
in_channels: 4
|
| 36 |
+
out_channels: 4
|
| 37 |
+
model_channels: 320
|
| 38 |
+
attention_resolutions: [ 4, 2, 1 ]
|
| 39 |
+
num_res_blocks: 2
|
| 40 |
+
channel_mult: [ 1, 2, 4, 4 ]
|
| 41 |
+
num_heads: 8
|
| 42 |
+
use_spatial_transformer: True
|
| 43 |
+
transformer_depth: 1
|
| 44 |
+
context_dim: 768
|
| 45 |
+
use_checkpoint: True
|
| 46 |
+
legacy: False
|
| 47 |
+
|
| 48 |
+
first_stage_config:
|
| 49 |
+
target: ldm.models.autoencoder.AutoencoderKL
|
| 50 |
+
params:
|
| 51 |
+
embed_dim: 4
|
| 52 |
+
monitor: val/rec_loss
|
| 53 |
+
ddconfig:
|
| 54 |
+
double_z: true
|
| 55 |
+
z_channels: 4
|
| 56 |
+
resolution: 256
|
| 57 |
+
in_channels: 3
|
| 58 |
+
out_ch: 3
|
| 59 |
+
ch: 128
|
| 60 |
+
ch_mult:
|
| 61 |
+
- 1
|
| 62 |
+
- 2
|
| 63 |
+
- 4
|
| 64 |
+
- 4
|
| 65 |
+
num_res_blocks: 2
|
| 66 |
+
attn_resolutions: []
|
| 67 |
+
dropout: 0.0
|
| 68 |
+
lossconfig:
|
| 69 |
+
target: torch.nn.Identity
|
| 70 |
+
|
| 71 |
+
cond_stage_config:
|
| 72 |
+
target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
|
EasyFluffV10.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:120af784616881b783cf99933688ed1b1591cedbf8af760fa1fc26e1e9ec0b33
|
| 3 |
+
size 2132637464
|
EasyFluffV10.yaml
ADDED
|
@@ -0,0 +1,72 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
model:
|
| 2 |
+
base_learning_rate: 1.0e-04
|
| 3 |
+
target: ldm.models.diffusion.ddpm.LatentDiffusion
|
| 4 |
+
params:
|
| 5 |
+
parameterization: "v"
|
| 6 |
+
linear_start: 0.00085
|
| 7 |
+
linear_end: 0.0120
|
| 8 |
+
num_timesteps_cond: 1
|
| 9 |
+
log_every_t: 200
|
| 10 |
+
timesteps: 1000
|
| 11 |
+
first_stage_key: "jpg"
|
| 12 |
+
cond_stage_key: "txt"
|
| 13 |
+
image_size: 64
|
| 14 |
+
channels: 4
|
| 15 |
+
cond_stage_trainable: false # Note: different from the one we trained before
|
| 16 |
+
conditioning_key: crossattn
|
| 17 |
+
monitor: val/loss_simple_ema
|
| 18 |
+
scale_factor: 0.18215
|
| 19 |
+
use_ema: False
|
| 20 |
+
|
| 21 |
+
scheduler_config: # 10000 warmup steps
|
| 22 |
+
target: ldm.lr_scheduler.LambdaLinearScheduler
|
| 23 |
+
params:
|
| 24 |
+
warm_up_steps: [ 10000 ]
|
| 25 |
+
cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
|
| 26 |
+
f_start: [ 1.e-6 ]
|
| 27 |
+
f_max: [ 1. ]
|
| 28 |
+
f_min: [ 1. ]
|
| 29 |
+
|
| 30 |
+
unet_config:
|
| 31 |
+
target: ldm.modules.diffusionmodules.openaimodel.UNetModel
|
| 32 |
+
params:
|
| 33 |
+
use_fp16: True
|
| 34 |
+
image_size: 32 # unused
|
| 35 |
+
in_channels: 4
|
| 36 |
+
out_channels: 4
|
| 37 |
+
model_channels: 320
|
| 38 |
+
attention_resolutions: [ 4, 2, 1 ]
|
| 39 |
+
num_res_blocks: 2
|
| 40 |
+
channel_mult: [ 1, 2, 4, 4 ]
|
| 41 |
+
num_heads: 8
|
| 42 |
+
use_spatial_transformer: True
|
| 43 |
+
transformer_depth: 1
|
| 44 |
+
context_dim: 768
|
| 45 |
+
use_checkpoint: True
|
| 46 |
+
legacy: False
|
| 47 |
+
|
| 48 |
+
first_stage_config:
|
| 49 |
+
target: ldm.models.autoencoder.AutoencoderKL
|
| 50 |
+
params:
|
| 51 |
+
embed_dim: 4
|
| 52 |
+
monitor: val/rec_loss
|
| 53 |
+
ddconfig:
|
| 54 |
+
double_z: true
|
| 55 |
+
z_channels: 4
|
| 56 |
+
resolution: 256
|
| 57 |
+
in_channels: 3
|
| 58 |
+
out_ch: 3
|
| 59 |
+
ch: 128
|
| 60 |
+
ch_mult:
|
| 61 |
+
- 1
|
| 62 |
+
- 2
|
| 63 |
+
- 4
|
| 64 |
+
- 4
|
| 65 |
+
num_res_blocks: 2
|
| 66 |
+
attn_resolutions: []
|
| 67 |
+
dropout: 0.0
|
| 68 |
+
lossconfig:
|
| 69 |
+
target: torch.nn.Identity
|
| 70 |
+
|
| 71 |
+
cond_stage_config:
|
| 72 |
+
target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
|
EasyFluffV11.1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d66e4da0d720202f03fb2ac72a0eb491040d3a98350e0bba407ac00f55675d1d
|
| 3 |
+
size 2133823506
|
EasyFluffV11.1.yaml
ADDED
|
@@ -0,0 +1,72 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
model:
|
| 2 |
+
base_learning_rate: 1.0e-04
|
| 3 |
+
target: ldm.models.diffusion.ddpm.LatentDiffusion
|
| 4 |
+
params:
|
| 5 |
+
parameterization: "v"
|
| 6 |
+
linear_start: 0.00085
|
| 7 |
+
linear_end: 0.0120
|
| 8 |
+
num_timesteps_cond: 1
|
| 9 |
+
log_every_t: 200
|
| 10 |
+
timesteps: 1000
|
| 11 |
+
first_stage_key: "jpg"
|
| 12 |
+
cond_stage_key: "txt"
|
| 13 |
+
image_size: 64
|
| 14 |
+
channels: 4
|
| 15 |
+
cond_stage_trainable: false # Note: different from the one we trained before
|
| 16 |
+
conditioning_key: crossattn
|
| 17 |
+
monitor: val/loss_simple_ema
|
| 18 |
+
scale_factor: 0.18215
|
| 19 |
+
use_ema: False
|
| 20 |
+
|
| 21 |
+
scheduler_config: # 10000 warmup steps
|
| 22 |
+
target: ldm.lr_scheduler.LambdaLinearScheduler
|
| 23 |
+
params:
|
| 24 |
+
warm_up_steps: [ 10000 ]
|
| 25 |
+
cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
|
| 26 |
+
f_start: [ 1.e-6 ]
|
| 27 |
+
f_max: [ 1. ]
|
| 28 |
+
f_min: [ 1. ]
|
| 29 |
+
|
| 30 |
+
unet_config:
|
| 31 |
+
target: ldm.modules.diffusionmodules.openaimodel.UNetModel
|
| 32 |
+
params:
|
| 33 |
+
use_fp16: True
|
| 34 |
+
image_size: 32 # unused
|
| 35 |
+
in_channels: 4
|
| 36 |
+
out_channels: 4
|
| 37 |
+
model_channels: 320
|
| 38 |
+
attention_resolutions: [ 4, 2, 1 ]
|
| 39 |
+
num_res_blocks: 2
|
| 40 |
+
channel_mult: [ 1, 2, 4, 4 ]
|
| 41 |
+
num_heads: 8
|
| 42 |
+
use_spatial_transformer: True
|
| 43 |
+
transformer_depth: 1
|
| 44 |
+
context_dim: 768
|
| 45 |
+
use_checkpoint: True
|
| 46 |
+
legacy: False
|
| 47 |
+
|
| 48 |
+
first_stage_config:
|
| 49 |
+
target: ldm.models.autoencoder.AutoencoderKL
|
| 50 |
+
params:
|
| 51 |
+
embed_dim: 4
|
| 52 |
+
monitor: val/rec_loss
|
| 53 |
+
ddconfig:
|
| 54 |
+
double_z: true
|
| 55 |
+
z_channels: 4
|
| 56 |
+
resolution: 256
|
| 57 |
+
in_channels: 3
|
| 58 |
+
out_ch: 3
|
| 59 |
+
ch: 128
|
| 60 |
+
ch_mult:
|
| 61 |
+
- 1
|
| 62 |
+
- 2
|
| 63 |
+
- 4
|
| 64 |
+
- 4
|
| 65 |
+
num_res_blocks: 2
|
| 66 |
+
attn_resolutions: []
|
| 67 |
+
dropout: 0.0
|
| 68 |
+
lossconfig:
|
| 69 |
+
target: torch.nn.Identity
|
| 70 |
+
|
| 71 |
+
cond_stage_config:
|
| 72 |
+
target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
|
EasyFluffV11.2.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:821628644ef37919d541cc824d57b9f8084652e3e388dcc5f5a4dd496d6d556b
|
| 3 |
+
size 2132643672
|
EasyFluffV11.2.yaml
ADDED
|
@@ -0,0 +1,72 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
model:
|
| 2 |
+
base_learning_rate: 1.0e-04
|
| 3 |
+
target: ldm.models.diffusion.ddpm.LatentDiffusion
|
| 4 |
+
params:
|
| 5 |
+
parameterization: "v"
|
| 6 |
+
linear_start: 0.00085
|
| 7 |
+
linear_end: 0.0120
|
| 8 |
+
num_timesteps_cond: 1
|
| 9 |
+
log_every_t: 200
|
| 10 |
+
timesteps: 1000
|
| 11 |
+
first_stage_key: "jpg"
|
| 12 |
+
cond_stage_key: "txt"
|
| 13 |
+
image_size: 64
|
| 14 |
+
channels: 4
|
| 15 |
+
cond_stage_trainable: false # Note: different from the one we trained before
|
| 16 |
+
conditioning_key: crossattn
|
| 17 |
+
monitor: val/loss_simple_ema
|
| 18 |
+
scale_factor: 0.18215
|
| 19 |
+
use_ema: False
|
| 20 |
+
|
| 21 |
+
scheduler_config: # 10000 warmup steps
|
| 22 |
+
target: ldm.lr_scheduler.LambdaLinearScheduler
|
| 23 |
+
params:
|
| 24 |
+
warm_up_steps: [ 10000 ]
|
| 25 |
+
cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
|
| 26 |
+
f_start: [ 1.e-6 ]
|
| 27 |
+
f_max: [ 1. ]
|
| 28 |
+
f_min: [ 1. ]
|
| 29 |
+
|
| 30 |
+
unet_config:
|
| 31 |
+
target: ldm.modules.diffusionmodules.openaimodel.UNetModel
|
| 32 |
+
params:
|
| 33 |
+
use_fp16: True
|
| 34 |
+
image_size: 32 # unused
|
| 35 |
+
in_channels: 4
|
| 36 |
+
out_channels: 4
|
| 37 |
+
model_channels: 320
|
| 38 |
+
attention_resolutions: [ 4, 2, 1 ]
|
| 39 |
+
num_res_blocks: 2
|
| 40 |
+
channel_mult: [ 1, 2, 4, 4 ]
|
| 41 |
+
num_heads: 8
|
| 42 |
+
use_spatial_transformer: True
|
| 43 |
+
transformer_depth: 1
|
| 44 |
+
context_dim: 768
|
| 45 |
+
use_checkpoint: True
|
| 46 |
+
legacy: False
|
| 47 |
+
|
| 48 |
+
first_stage_config:
|
| 49 |
+
target: ldm.models.autoencoder.AutoencoderKL
|
| 50 |
+
params:
|
| 51 |
+
embed_dim: 4
|
| 52 |
+
monitor: val/rec_loss
|
| 53 |
+
ddconfig:
|
| 54 |
+
double_z: true
|
| 55 |
+
z_channels: 4
|
| 56 |
+
resolution: 256
|
| 57 |
+
in_channels: 3
|
| 58 |
+
out_ch: 3
|
| 59 |
+
ch: 128
|
| 60 |
+
ch_mult:
|
| 61 |
+
- 1
|
| 62 |
+
- 2
|
| 63 |
+
- 4
|
| 64 |
+
- 4
|
| 65 |
+
num_res_blocks: 2
|
| 66 |
+
attn_resolutions: []
|
| 67 |
+
dropout: 0.0
|
| 68 |
+
lossconfig:
|
| 69 |
+
target: torch.nn.Identity
|
| 70 |
+
|
| 71 |
+
cond_stage_config:
|
| 72 |
+
target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
|
EasyFluffV11.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c55486da6d56b5fb29f7978bf1ce389c3775ffe3c4d40203571a3cabf5297152
|
| 3 |
+
size 2132641032
|
EasyFluffV11.yaml
ADDED
|
@@ -0,0 +1,72 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
model:
|
| 2 |
+
base_learning_rate: 1.0e-04
|
| 3 |
+
target: ldm.models.diffusion.ddpm.LatentDiffusion
|
| 4 |
+
params:
|
| 5 |
+
parameterization: "v"
|
| 6 |
+
linear_start: 0.00085
|
| 7 |
+
linear_end: 0.0120
|
| 8 |
+
num_timesteps_cond: 1
|
| 9 |
+
log_every_t: 200
|
| 10 |
+
timesteps: 1000
|
| 11 |
+
first_stage_key: "jpg"
|
| 12 |
+
cond_stage_key: "txt"
|
| 13 |
+
image_size: 64
|
| 14 |
+
channels: 4
|
| 15 |
+
cond_stage_trainable: false # Note: different from the one we trained before
|
| 16 |
+
conditioning_key: crossattn
|
| 17 |
+
monitor: val/loss_simple_ema
|
| 18 |
+
scale_factor: 0.18215
|
| 19 |
+
use_ema: False
|
| 20 |
+
|
| 21 |
+
scheduler_config: # 10000 warmup steps
|
| 22 |
+
target: ldm.lr_scheduler.LambdaLinearScheduler
|
| 23 |
+
params:
|
| 24 |
+
warm_up_steps: [ 10000 ]
|
| 25 |
+
cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
|
| 26 |
+
f_start: [ 1.e-6 ]
|
| 27 |
+
f_max: [ 1. ]
|
| 28 |
+
f_min: [ 1. ]
|
| 29 |
+
|
| 30 |
+
unet_config:
|
| 31 |
+
target: ldm.modules.diffusionmodules.openaimodel.UNetModel
|
| 32 |
+
params:
|
| 33 |
+
use_fp16: True
|
| 34 |
+
image_size: 32 # unused
|
| 35 |
+
in_channels: 4
|
| 36 |
+
out_channels: 4
|
| 37 |
+
model_channels: 320
|
| 38 |
+
attention_resolutions: [ 4, 2, 1 ]
|
| 39 |
+
num_res_blocks: 2
|
| 40 |
+
channel_mult: [ 1, 2, 4, 4 ]
|
| 41 |
+
num_heads: 8
|
| 42 |
+
use_spatial_transformer: True
|
| 43 |
+
transformer_depth: 1
|
| 44 |
+
context_dim: 768
|
| 45 |
+
use_checkpoint: True
|
| 46 |
+
legacy: False
|
| 47 |
+
|
| 48 |
+
first_stage_config:
|
| 49 |
+
target: ldm.models.autoencoder.AutoencoderKL
|
| 50 |
+
params:
|
| 51 |
+
embed_dim: 4
|
| 52 |
+
monitor: val/rec_loss
|
| 53 |
+
ddconfig:
|
| 54 |
+
double_z: true
|
| 55 |
+
z_channels: 4
|
| 56 |
+
resolution: 256
|
| 57 |
+
in_channels: 3
|
| 58 |
+
out_ch: 3
|
| 59 |
+
ch: 128
|
| 60 |
+
ch_mult:
|
| 61 |
+
- 1
|
| 62 |
+
- 2
|
| 63 |
+
- 4
|
| 64 |
+
- 4
|
| 65 |
+
num_res_blocks: 2
|
| 66 |
+
attn_resolutions: []
|
| 67 |
+
dropout: 0.0
|
| 68 |
+
lossconfig:
|
| 69 |
+
target: torch.nn.Identity
|
| 70 |
+
|
| 71 |
+
cond_stage_config:
|
| 72 |
+
target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
|
EasyFluffV2.1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:103cd6bd8f2369f295d04c8e15abed85bfbaa7e20a3636f1e59e3755925413a0
|
| 3 |
+
size 2132626098
|
EasyFluffV2.1.yaml
ADDED
|
@@ -0,0 +1,72 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
model:
|
| 2 |
+
base_learning_rate: 1.0e-04
|
| 3 |
+
target: ldm.models.diffusion.ddpm.LatentDiffusion
|
| 4 |
+
params:
|
| 5 |
+
parameterization: "v"
|
| 6 |
+
linear_start: 0.00085
|
| 7 |
+
linear_end: 0.0120
|
| 8 |
+
num_timesteps_cond: 1
|
| 9 |
+
log_every_t: 200
|
| 10 |
+
timesteps: 1000
|
| 11 |
+
first_stage_key: "jpg"
|
| 12 |
+
cond_stage_key: "txt"
|
| 13 |
+
image_size: 64
|
| 14 |
+
channels: 4
|
| 15 |
+
cond_stage_trainable: false # Note: different from the one we trained before
|
| 16 |
+
conditioning_key: crossattn
|
| 17 |
+
monitor: val/loss_simple_ema
|
| 18 |
+
scale_factor: 0.18215
|
| 19 |
+
use_ema: False
|
| 20 |
+
|
| 21 |
+
scheduler_config: # 10000 warmup steps
|
| 22 |
+
target: ldm.lr_scheduler.LambdaLinearScheduler
|
| 23 |
+
params:
|
| 24 |
+
warm_up_steps: [ 10000 ]
|
| 25 |
+
cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
|
| 26 |
+
f_start: [ 1.e-6 ]
|
| 27 |
+
f_max: [ 1. ]
|
| 28 |
+
f_min: [ 1. ]
|
| 29 |
+
|
| 30 |
+
unet_config:
|
| 31 |
+
target: ldm.modules.diffusionmodules.openaimodel.UNetModel
|
| 32 |
+
params:
|
| 33 |
+
use_fp16: True
|
| 34 |
+
image_size: 32 # unused
|
| 35 |
+
in_channels: 4
|
| 36 |
+
out_channels: 4
|
| 37 |
+
model_channels: 320
|
| 38 |
+
attention_resolutions: [ 4, 2, 1 ]
|
| 39 |
+
num_res_blocks: 2
|
| 40 |
+
channel_mult: [ 1, 2, 4, 4 ]
|
| 41 |
+
num_heads: 8
|
| 42 |
+
use_spatial_transformer: True
|
| 43 |
+
transformer_depth: 1
|
| 44 |
+
context_dim: 768
|
| 45 |
+
use_checkpoint: True
|
| 46 |
+
legacy: False
|
| 47 |
+
|
| 48 |
+
first_stage_config:
|
| 49 |
+
target: ldm.models.autoencoder.AutoencoderKL
|
| 50 |
+
params:
|
| 51 |
+
embed_dim: 4
|
| 52 |
+
monitor: val/rec_loss
|
| 53 |
+
ddconfig:
|
| 54 |
+
double_z: true
|
| 55 |
+
z_channels: 4
|
| 56 |
+
resolution: 256
|
| 57 |
+
in_channels: 3
|
| 58 |
+
out_ch: 3
|
| 59 |
+
ch: 128
|
| 60 |
+
ch_mult:
|
| 61 |
+
- 1
|
| 62 |
+
- 2
|
| 63 |
+
- 4
|
| 64 |
+
- 4
|
| 65 |
+
num_res_blocks: 2
|
| 66 |
+
attn_resolutions: []
|
| 67 |
+
dropout: 0.0
|
| 68 |
+
lossconfig:
|
| 69 |
+
target: torch.nn.Identity
|
| 70 |
+
|
| 71 |
+
cond_stage_config:
|
| 72 |
+
target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
|
EasyFluffV2.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bbe1410b37b18d5faacd3f80366ce47b3f50029d4889c091c357747b5ce7631f
|
| 3 |
+
size 2132626098
|
EasyFluffV2.yaml
ADDED
|
@@ -0,0 +1,72 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
model:
|
| 2 |
+
base_learning_rate: 1.0e-04
|
| 3 |
+
target: ldm.models.diffusion.ddpm.LatentDiffusion
|
| 4 |
+
params:
|
| 5 |
+
parameterization: "v"
|
| 6 |
+
linear_start: 0.00085
|
| 7 |
+
linear_end: 0.0120
|
| 8 |
+
num_timesteps_cond: 1
|
| 9 |
+
log_every_t: 200
|
| 10 |
+
timesteps: 1000
|
| 11 |
+
first_stage_key: "jpg"
|
| 12 |
+
cond_stage_key: "txt"
|
| 13 |
+
image_size: 64
|
| 14 |
+
channels: 4
|
| 15 |
+
cond_stage_trainable: false # Note: different from the one we trained before
|
| 16 |
+
conditioning_key: crossattn
|
| 17 |
+
monitor: val/loss_simple_ema
|
| 18 |
+
scale_factor: 0.18215
|
| 19 |
+
use_ema: False
|
| 20 |
+
|
| 21 |
+
scheduler_config: # 10000 warmup steps
|
| 22 |
+
target: ldm.lr_scheduler.LambdaLinearScheduler
|
| 23 |
+
params:
|
| 24 |
+
warm_up_steps: [ 10000 ]
|
| 25 |
+
cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
|
| 26 |
+
f_start: [ 1.e-6 ]
|
| 27 |
+
f_max: [ 1. ]
|
| 28 |
+
f_min: [ 1. ]
|
| 29 |
+
|
| 30 |
+
unet_config:
|
| 31 |
+
target: ldm.modules.diffusionmodules.openaimodel.UNetModel
|
| 32 |
+
params:
|
| 33 |
+
use_fp16: True
|
| 34 |
+
image_size: 32 # unused
|
| 35 |
+
in_channels: 4
|
| 36 |
+
out_channels: 4
|
| 37 |
+
model_channels: 320
|
| 38 |
+
attention_resolutions: [ 4, 2, 1 ]
|
| 39 |
+
num_res_blocks: 2
|
| 40 |
+
channel_mult: [ 1, 2, 4, 4 ]
|
| 41 |
+
num_heads: 8
|
| 42 |
+
use_spatial_transformer: True
|
| 43 |
+
transformer_depth: 1
|
| 44 |
+
context_dim: 768
|
| 45 |
+
use_checkpoint: True
|
| 46 |
+
legacy: False
|
| 47 |
+
|
| 48 |
+
first_stage_config:
|
| 49 |
+
target: ldm.models.autoencoder.AutoencoderKL
|
| 50 |
+
params:
|
| 51 |
+
embed_dim: 4
|
| 52 |
+
monitor: val/rec_loss
|
| 53 |
+
ddconfig:
|
| 54 |
+
double_z: true
|
| 55 |
+
z_channels: 4
|
| 56 |
+
resolution: 256
|
| 57 |
+
in_channels: 3
|
| 58 |
+
out_ch: 3
|
| 59 |
+
ch: 128
|
| 60 |
+
ch_mult:
|
| 61 |
+
- 1
|
| 62 |
+
- 2
|
| 63 |
+
- 4
|
| 64 |
+
- 4
|
| 65 |
+
num_res_blocks: 2
|
| 66 |
+
attn_resolutions: []
|
| 67 |
+
dropout: 0.0
|
| 68 |
+
lossconfig:
|
| 69 |
+
target: torch.nn.Identity
|
| 70 |
+
|
| 71 |
+
cond_stage_config:
|
| 72 |
+
target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
|
EasyFluffV3-LessCreativeMoreAccurate.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:89854ef3d3a55adced737b913b50fd7365b5b8647751fdf18793211bcc547966
|
| 3 |
+
size 2132626098
|
EasyFluffV3-LessCreativeMoreAccurate.yaml
ADDED
|
@@ -0,0 +1,72 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
model:
|
| 2 |
+
base_learning_rate: 1.0e-04
|
| 3 |
+
target: ldm.models.diffusion.ddpm.LatentDiffusion
|
| 4 |
+
params:
|
| 5 |
+
parameterization: "v"
|
| 6 |
+
linear_start: 0.00085
|
| 7 |
+
linear_end: 0.0120
|
| 8 |
+
num_timesteps_cond: 1
|
| 9 |
+
log_every_t: 200
|
| 10 |
+
timesteps: 1000
|
| 11 |
+
first_stage_key: "jpg"
|
| 12 |
+
cond_stage_key: "txt"
|
| 13 |
+
image_size: 64
|
| 14 |
+
channels: 4
|
| 15 |
+
cond_stage_trainable: false # Note: different from the one we trained before
|
| 16 |
+
conditioning_key: crossattn
|
| 17 |
+
monitor: val/loss_simple_ema
|
| 18 |
+
scale_factor: 0.18215
|
| 19 |
+
use_ema: False
|
| 20 |
+
|
| 21 |
+
scheduler_config: # 10000 warmup steps
|
| 22 |
+
target: ldm.lr_scheduler.LambdaLinearScheduler
|
| 23 |
+
params:
|
| 24 |
+
warm_up_steps: [ 10000 ]
|
| 25 |
+
cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
|
| 26 |
+
f_start: [ 1.e-6 ]
|
| 27 |
+
f_max: [ 1. ]
|
| 28 |
+
f_min: [ 1. ]
|
| 29 |
+
|
| 30 |
+
unet_config:
|
| 31 |
+
target: ldm.modules.diffusionmodules.openaimodel.UNetModel
|
| 32 |
+
params:
|
| 33 |
+
use_fp16: True
|
| 34 |
+
image_size: 32 # unused
|
| 35 |
+
in_channels: 4
|
| 36 |
+
out_channels: 4
|
| 37 |
+
model_channels: 320
|
| 38 |
+
attention_resolutions: [ 4, 2, 1 ]
|
| 39 |
+
num_res_blocks: 2
|
| 40 |
+
channel_mult: [ 1, 2, 4, 4 ]
|
| 41 |
+
num_heads: 8
|
| 42 |
+
use_spatial_transformer: True
|
| 43 |
+
transformer_depth: 1
|
| 44 |
+
context_dim: 768
|
| 45 |
+
use_checkpoint: True
|
| 46 |
+
legacy: False
|
| 47 |
+
|
| 48 |
+
first_stage_config:
|
| 49 |
+
target: ldm.models.autoencoder.AutoencoderKL
|
| 50 |
+
params:
|
| 51 |
+
embed_dim: 4
|
| 52 |
+
monitor: val/rec_loss
|
| 53 |
+
ddconfig:
|
| 54 |
+
double_z: true
|
| 55 |
+
z_channels: 4
|
| 56 |
+
resolution: 256
|
| 57 |
+
in_channels: 3
|
| 58 |
+
out_ch: 3
|
| 59 |
+
ch: 128
|
| 60 |
+
ch_mult:
|
| 61 |
+
- 1
|
| 62 |
+
- 2
|
| 63 |
+
- 4
|
| 64 |
+
- 4
|
| 65 |
+
num_res_blocks: 2
|
| 66 |
+
attn_resolutions: []
|
| 67 |
+
dropout: 0.0
|
| 68 |
+
lossconfig:
|
| 69 |
+
target: torch.nn.Identity
|
| 70 |
+
|
| 71 |
+
cond_stage_config:
|
| 72 |
+
target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
|
EasyFluffV3.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:49e8d8c7d92b4495d2add15253cb6c0a312f179696655e3f6e79905caf45eb31
|
| 3 |
+
size 2132626098
|
EasyFluffV3.yaml
ADDED
|
@@ -0,0 +1,72 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
model:
|
| 2 |
+
base_learning_rate: 1.0e-04
|
| 3 |
+
target: ldm.models.diffusion.ddpm.LatentDiffusion
|
| 4 |
+
params:
|
| 5 |
+
parameterization: "v"
|
| 6 |
+
linear_start: 0.00085
|
| 7 |
+
linear_end: 0.0120
|
| 8 |
+
num_timesteps_cond: 1
|
| 9 |
+
log_every_t: 200
|
| 10 |
+
timesteps: 1000
|
| 11 |
+
first_stage_key: "jpg"
|
| 12 |
+
cond_stage_key: "txt"
|
| 13 |
+
image_size: 64
|
| 14 |
+
channels: 4
|
| 15 |
+
cond_stage_trainable: false # Note: different from the one we trained before
|
| 16 |
+
conditioning_key: crossattn
|
| 17 |
+
monitor: val/loss_simple_ema
|
| 18 |
+
scale_factor: 0.18215
|
| 19 |
+
use_ema: False
|
| 20 |
+
|
| 21 |
+
scheduler_config: # 10000 warmup steps
|
| 22 |
+
target: ldm.lr_scheduler.LambdaLinearScheduler
|
| 23 |
+
params:
|
| 24 |
+
warm_up_steps: [ 10000 ]
|
| 25 |
+
cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
|
| 26 |
+
f_start: [ 1.e-6 ]
|
| 27 |
+
f_max: [ 1. ]
|
| 28 |
+
f_min: [ 1. ]
|
| 29 |
+
|
| 30 |
+
unet_config:
|
| 31 |
+
target: ldm.modules.diffusionmodules.openaimodel.UNetModel
|
| 32 |
+
params:
|
| 33 |
+
use_fp16: True
|
| 34 |
+
image_size: 32 # unused
|
| 35 |
+
in_channels: 4
|
| 36 |
+
out_channels: 4
|
| 37 |
+
model_channels: 320
|
| 38 |
+
attention_resolutions: [ 4, 2, 1 ]
|
| 39 |
+
num_res_blocks: 2
|
| 40 |
+
channel_mult: [ 1, 2, 4, 4 ]
|
| 41 |
+
num_heads: 8
|
| 42 |
+
use_spatial_transformer: True
|
| 43 |
+
transformer_depth: 1
|
| 44 |
+
context_dim: 768
|
| 45 |
+
use_checkpoint: True
|
| 46 |
+
legacy: False
|
| 47 |
+
|
| 48 |
+
first_stage_config:
|
| 49 |
+
target: ldm.models.autoencoder.AutoencoderKL
|
| 50 |
+
params:
|
| 51 |
+
embed_dim: 4
|
| 52 |
+
monitor: val/rec_loss
|
| 53 |
+
ddconfig:
|
| 54 |
+
double_z: true
|
| 55 |
+
z_channels: 4
|
| 56 |
+
resolution: 256
|
| 57 |
+
in_channels: 3
|
| 58 |
+
out_ch: 3
|
| 59 |
+
ch: 128
|
| 60 |
+
ch_mult:
|
| 61 |
+
- 1
|
| 62 |
+
- 2
|
| 63 |
+
- 4
|
| 64 |
+
- 4
|
| 65 |
+
num_res_blocks: 2
|
| 66 |
+
attn_resolutions: []
|
| 67 |
+
dropout: 0.0
|
| 68 |
+
lossconfig:
|
| 69 |
+
target: torch.nn.Identity
|
| 70 |
+
|
| 71 |
+
cond_stage_config:
|
| 72 |
+
target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
|
EasyFluffV4-LessCreativeMoreAccurate.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3331ac9574805d977ab06920a1cb9a494713471fd072e968e03e4edd587482dd
|
| 3 |
+
size 2132625644
|
EasyFluffV4-LessCreativeMoreAccurate.yaml
ADDED
|
@@ -0,0 +1,72 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
model:
|
| 2 |
+
base_learning_rate: 1.0e-04
|
| 3 |
+
target: ldm.models.diffusion.ddpm.LatentDiffusion
|
| 4 |
+
params:
|
| 5 |
+
parameterization: "v"
|
| 6 |
+
linear_start: 0.00085
|
| 7 |
+
linear_end: 0.0120
|
| 8 |
+
num_timesteps_cond: 1
|
| 9 |
+
log_every_t: 200
|
| 10 |
+
timesteps: 1000
|
| 11 |
+
first_stage_key: "jpg"
|
| 12 |
+
cond_stage_key: "txt"
|
| 13 |
+
image_size: 64
|
| 14 |
+
channels: 4
|
| 15 |
+
cond_stage_trainable: false # Note: different from the one we trained before
|
| 16 |
+
conditioning_key: crossattn
|
| 17 |
+
monitor: val/loss_simple_ema
|
| 18 |
+
scale_factor: 0.18215
|
| 19 |
+
use_ema: False
|
| 20 |
+
|
| 21 |
+
scheduler_config: # 10000 warmup steps
|
| 22 |
+
target: ldm.lr_scheduler.LambdaLinearScheduler
|
| 23 |
+
params:
|
| 24 |
+
warm_up_steps: [ 10000 ]
|
| 25 |
+
cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
|
| 26 |
+
f_start: [ 1.e-6 ]
|
| 27 |
+
f_max: [ 1. ]
|
| 28 |
+
f_min: [ 1. ]
|
| 29 |
+
|
| 30 |
+
unet_config:
|
| 31 |
+
target: ldm.modules.diffusionmodules.openaimodel.UNetModel
|
| 32 |
+
params:
|
| 33 |
+
use_fp16: True
|
| 34 |
+
image_size: 32 # unused
|
| 35 |
+
in_channels: 4
|
| 36 |
+
out_channels: 4
|
| 37 |
+
model_channels: 320
|
| 38 |
+
attention_resolutions: [ 4, 2, 1 ]
|
| 39 |
+
num_res_blocks: 2
|
| 40 |
+
channel_mult: [ 1, 2, 4, 4 ]
|
| 41 |
+
num_heads: 8
|
| 42 |
+
use_spatial_transformer: True
|
| 43 |
+
transformer_depth: 1
|
| 44 |
+
context_dim: 768
|
| 45 |
+
use_checkpoint: True
|
| 46 |
+
legacy: False
|
| 47 |
+
|
| 48 |
+
first_stage_config:
|
| 49 |
+
target: ldm.models.autoencoder.AutoencoderKL
|
| 50 |
+
params:
|
| 51 |
+
embed_dim: 4
|
| 52 |
+
monitor: val/rec_loss
|
| 53 |
+
ddconfig:
|
| 54 |
+
double_z: true
|
| 55 |
+
z_channels: 4
|
| 56 |
+
resolution: 256
|
| 57 |
+
in_channels: 3
|
| 58 |
+
out_ch: 3
|
| 59 |
+
ch: 128
|
| 60 |
+
ch_mult:
|
| 61 |
+
- 1
|
| 62 |
+
- 2
|
| 63 |
+
- 4
|
| 64 |
+
- 4
|
| 65 |
+
num_res_blocks: 2
|
| 66 |
+
attn_resolutions: []
|
| 67 |
+
dropout: 0.0
|
| 68 |
+
lossconfig:
|
| 69 |
+
target: torch.nn.Identity
|
| 70 |
+
|
| 71 |
+
cond_stage_config:
|
| 72 |
+
target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
|
EasyFluffV4.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a401c94834edb490621bcdf4645688647ef76316485d9646a815191a0526c472
|
| 3 |
+
size 2132625644
|
EasyFluffV4.yaml
ADDED
|
@@ -0,0 +1,72 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
model:
|
| 2 |
+
base_learning_rate: 1.0e-04
|
| 3 |
+
target: ldm.models.diffusion.ddpm.LatentDiffusion
|
| 4 |
+
params:
|
| 5 |
+
parameterization: "v"
|
| 6 |
+
linear_start: 0.00085
|
| 7 |
+
linear_end: 0.0120
|
| 8 |
+
num_timesteps_cond: 1
|
| 9 |
+
log_every_t: 200
|
| 10 |
+
timesteps: 1000
|
| 11 |
+
first_stage_key: "jpg"
|
| 12 |
+
cond_stage_key: "txt"
|
| 13 |
+
image_size: 64
|
| 14 |
+
channels: 4
|
| 15 |
+
cond_stage_trainable: false # Note: different from the one we trained before
|
| 16 |
+
conditioning_key: crossattn
|
| 17 |
+
monitor: val/loss_simple_ema
|
| 18 |
+
scale_factor: 0.18215
|
| 19 |
+
use_ema: False
|
| 20 |
+
|
| 21 |
+
scheduler_config: # 10000 warmup steps
|
| 22 |
+
target: ldm.lr_scheduler.LambdaLinearScheduler
|
| 23 |
+
params:
|
| 24 |
+
warm_up_steps: [ 10000 ]
|
| 25 |
+
cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
|
| 26 |
+
f_start: [ 1.e-6 ]
|
| 27 |
+
f_max: [ 1. ]
|
| 28 |
+
f_min: [ 1. ]
|
| 29 |
+
|
| 30 |
+
unet_config:
|
| 31 |
+
target: ldm.modules.diffusionmodules.openaimodel.UNetModel
|
| 32 |
+
params:
|
| 33 |
+
use_fp16: True
|
| 34 |
+
image_size: 32 # unused
|
| 35 |
+
in_channels: 4
|
| 36 |
+
out_channels: 4
|
| 37 |
+
model_channels: 320
|
| 38 |
+
attention_resolutions: [ 4, 2, 1 ]
|
| 39 |
+
num_res_blocks: 2
|
| 40 |
+
channel_mult: [ 1, 2, 4, 4 ]
|
| 41 |
+
num_heads: 8
|
| 42 |
+
use_spatial_transformer: True
|
| 43 |
+
transformer_depth: 1
|
| 44 |
+
context_dim: 768
|
| 45 |
+
use_checkpoint: True
|
| 46 |
+
legacy: False
|
| 47 |
+
|
| 48 |
+
first_stage_config:
|
| 49 |
+
target: ldm.models.autoencoder.AutoencoderKL
|
| 50 |
+
params:
|
| 51 |
+
embed_dim: 4
|
| 52 |
+
monitor: val/rec_loss
|
| 53 |
+
ddconfig:
|
| 54 |
+
double_z: true
|
| 55 |
+
z_channels: 4
|
| 56 |
+
resolution: 256
|
| 57 |
+
in_channels: 3
|
| 58 |
+
out_ch: 3
|
| 59 |
+
ch: 128
|
| 60 |
+
ch_mult:
|
| 61 |
+
- 1
|
| 62 |
+
- 2
|
| 63 |
+
- 4
|
| 64 |
+
- 4
|
| 65 |
+
num_res_blocks: 2
|
| 66 |
+
attn_resolutions: []
|
| 67 |
+
dropout: 0.0
|
| 68 |
+
lossconfig:
|
| 69 |
+
target: torch.nn.Identity
|
| 70 |
+
|
| 71 |
+
cond_stage_config:
|
| 72 |
+
target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
|
EasyFluffV5.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a9c3908504e2cd6bb9f843ce98deb00d29ec2737d032d54dbe8f66aec4ce6548
|
| 3 |
+
size 2132625644
|
EasyFluffV5.yaml
ADDED
|
@@ -0,0 +1,72 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
model:
|
| 2 |
+
base_learning_rate: 1.0e-04
|
| 3 |
+
target: ldm.models.diffusion.ddpm.LatentDiffusion
|
| 4 |
+
params:
|
| 5 |
+
parameterization: "v"
|
| 6 |
+
linear_start: 0.00085
|
| 7 |
+
linear_end: 0.0120
|
| 8 |
+
num_timesteps_cond: 1
|
| 9 |
+
log_every_t: 200
|
| 10 |
+
timesteps: 1000
|
| 11 |
+
first_stage_key: "jpg"
|
| 12 |
+
cond_stage_key: "txt"
|
| 13 |
+
image_size: 64
|
| 14 |
+
channels: 4
|
| 15 |
+
cond_stage_trainable: false # Note: different from the one we trained before
|
| 16 |
+
conditioning_key: crossattn
|
| 17 |
+
monitor: val/loss_simple_ema
|
| 18 |
+
scale_factor: 0.18215
|
| 19 |
+
use_ema: False
|
| 20 |
+
|
| 21 |
+
scheduler_config: # 10000 warmup steps
|
| 22 |
+
target: ldm.lr_scheduler.LambdaLinearScheduler
|
| 23 |
+
params:
|
| 24 |
+
warm_up_steps: [ 10000 ]
|
| 25 |
+
cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
|
| 26 |
+
f_start: [ 1.e-6 ]
|
| 27 |
+
f_max: [ 1. ]
|
| 28 |
+
f_min: [ 1. ]
|
| 29 |
+
|
| 30 |
+
unet_config:
|
| 31 |
+
target: ldm.modules.diffusionmodules.openaimodel.UNetModel
|
| 32 |
+
params:
|
| 33 |
+
use_fp16: True
|
| 34 |
+
image_size: 32 # unused
|
| 35 |
+
in_channels: 4
|
| 36 |
+
out_channels: 4
|
| 37 |
+
model_channels: 320
|
| 38 |
+
attention_resolutions: [ 4, 2, 1 ]
|
| 39 |
+
num_res_blocks: 2
|
| 40 |
+
channel_mult: [ 1, 2, 4, 4 ]
|
| 41 |
+
num_heads: 8
|
| 42 |
+
use_spatial_transformer: True
|
| 43 |
+
transformer_depth: 1
|
| 44 |
+
context_dim: 768
|
| 45 |
+
use_checkpoint: True
|
| 46 |
+
legacy: False
|
| 47 |
+
|
| 48 |
+
first_stage_config:
|
| 49 |
+
target: ldm.models.autoencoder.AutoencoderKL
|
| 50 |
+
params:
|
| 51 |
+
embed_dim: 4
|
| 52 |
+
monitor: val/rec_loss
|
| 53 |
+
ddconfig:
|
| 54 |
+
double_z: true
|
| 55 |
+
z_channels: 4
|
| 56 |
+
resolution: 256
|
| 57 |
+
in_channels: 3
|
| 58 |
+
out_ch: 3
|
| 59 |
+
ch: 128
|
| 60 |
+
ch_mult:
|
| 61 |
+
- 1
|
| 62 |
+
- 2
|
| 63 |
+
- 4
|
| 64 |
+
- 4
|
| 65 |
+
num_res_blocks: 2
|
| 66 |
+
attn_resolutions: []
|
| 67 |
+
dropout: 0.0
|
| 68 |
+
lossconfig:
|
| 69 |
+
target: torch.nn.Identity
|
| 70 |
+
|
| 71 |
+
cond_stage_config:
|
| 72 |
+
target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
|
EasyFluffV6-OnlyFR-NoPF.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:263175da28d67f76cdb33ec7be1a7edc4a7f849d21b39d4f05cfd2f2c485b96d
|
| 3 |
+
size 2132625644
|
EasyFluffV6-OnlyFR-NoPF.yaml
ADDED
|
@@ -0,0 +1,72 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
model:
|
| 2 |
+
base_learning_rate: 1.0e-04
|
| 3 |
+
target: ldm.models.diffusion.ddpm.LatentDiffusion
|
| 4 |
+
params:
|
| 5 |
+
parameterization: "v"
|
| 6 |
+
linear_start: 0.00085
|
| 7 |
+
linear_end: 0.0120
|
| 8 |
+
num_timesteps_cond: 1
|
| 9 |
+
log_every_t: 200
|
| 10 |
+
timesteps: 1000
|
| 11 |
+
first_stage_key: "jpg"
|
| 12 |
+
cond_stage_key: "txt"
|
| 13 |
+
image_size: 64
|
| 14 |
+
channels: 4
|
| 15 |
+
cond_stage_trainable: false # Note: different from the one we trained before
|
| 16 |
+
conditioning_key: crossattn
|
| 17 |
+
monitor: val/loss_simple_ema
|
| 18 |
+
scale_factor: 0.18215
|
| 19 |
+
use_ema: False
|
| 20 |
+
|
| 21 |
+
scheduler_config: # 10000 warmup steps
|
| 22 |
+
target: ldm.lr_scheduler.LambdaLinearScheduler
|
| 23 |
+
params:
|
| 24 |
+
warm_up_steps: [ 10000 ]
|
| 25 |
+
cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
|
| 26 |
+
f_start: [ 1.e-6 ]
|
| 27 |
+
f_max: [ 1. ]
|
| 28 |
+
f_min: [ 1. ]
|
| 29 |
+
|
| 30 |
+
unet_config:
|
| 31 |
+
target: ldm.modules.diffusionmodules.openaimodel.UNetModel
|
| 32 |
+
params:
|
| 33 |
+
use_fp16: True
|
| 34 |
+
image_size: 32 # unused
|
| 35 |
+
in_channels: 4
|
| 36 |
+
out_channels: 4
|
| 37 |
+
model_channels: 320
|
| 38 |
+
attention_resolutions: [ 4, 2, 1 ]
|
| 39 |
+
num_res_blocks: 2
|
| 40 |
+
channel_mult: [ 1, 2, 4, 4 ]
|
| 41 |
+
num_heads: 8
|
| 42 |
+
use_spatial_transformer: True
|
| 43 |
+
transformer_depth: 1
|
| 44 |
+
context_dim: 768
|
| 45 |
+
use_checkpoint: True
|
| 46 |
+
legacy: False
|
| 47 |
+
|
| 48 |
+
first_stage_config:
|
| 49 |
+
target: ldm.models.autoencoder.AutoencoderKL
|
| 50 |
+
params:
|
| 51 |
+
embed_dim: 4
|
| 52 |
+
monitor: val/rec_loss
|
| 53 |
+
ddconfig:
|
| 54 |
+
double_z: true
|
| 55 |
+
z_channels: 4
|
| 56 |
+
resolution: 256
|
| 57 |
+
in_channels: 3
|
| 58 |
+
out_ch: 3
|
| 59 |
+
ch: 128
|
| 60 |
+
ch_mult:
|
| 61 |
+
- 1
|
| 62 |
+
- 2
|
| 63 |
+
- 4
|
| 64 |
+
- 4
|
| 65 |
+
num_res_blocks: 2
|
| 66 |
+
attn_resolutions: []
|
| 67 |
+
dropout: 0.0
|
| 68 |
+
lossconfig:
|
| 69 |
+
target: torch.nn.Identity
|
| 70 |
+
|
| 71 |
+
cond_stage_config:
|
| 72 |
+
target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
|
EasyFluffV6.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0db21412fcad1d273c3c8c04e1b3fda3f4976145809920af56018b7c3c7f19a5
|
| 3 |
+
size 2132637576
|
EasyFluffV6.yaml
ADDED
|
@@ -0,0 +1,72 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
model:
|
| 2 |
+
base_learning_rate: 1.0e-04
|
| 3 |
+
target: ldm.models.diffusion.ddpm.LatentDiffusion
|
| 4 |
+
params:
|
| 5 |
+
parameterization: "v"
|
| 6 |
+
linear_start: 0.00085
|
| 7 |
+
linear_end: 0.0120
|
| 8 |
+
num_timesteps_cond: 1
|
| 9 |
+
log_every_t: 200
|
| 10 |
+
timesteps: 1000
|
| 11 |
+
first_stage_key: "jpg"
|
| 12 |
+
cond_stage_key: "txt"
|
| 13 |
+
image_size: 64
|
| 14 |
+
channels: 4
|
| 15 |
+
cond_stage_trainable: false # Note: different from the one we trained before
|
| 16 |
+
conditioning_key: crossattn
|
| 17 |
+
monitor: val/loss_simple_ema
|
| 18 |
+
scale_factor: 0.18215
|
| 19 |
+
use_ema: False
|
| 20 |
+
|
| 21 |
+
scheduler_config: # 10000 warmup steps
|
| 22 |
+
target: ldm.lr_scheduler.LambdaLinearScheduler
|
| 23 |
+
params:
|
| 24 |
+
warm_up_steps: [ 10000 ]
|
| 25 |
+
cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
|
| 26 |
+
f_start: [ 1.e-6 ]
|
| 27 |
+
f_max: [ 1. ]
|
| 28 |
+
f_min: [ 1. ]
|
| 29 |
+
|
| 30 |
+
unet_config:
|
| 31 |
+
target: ldm.modules.diffusionmodules.openaimodel.UNetModel
|
| 32 |
+
params:
|
| 33 |
+
use_fp16: True
|
| 34 |
+
image_size: 32 # unused
|
| 35 |
+
in_channels: 4
|
| 36 |
+
out_channels: 4
|
| 37 |
+
model_channels: 320
|
| 38 |
+
attention_resolutions: [ 4, 2, 1 ]
|
| 39 |
+
num_res_blocks: 2
|
| 40 |
+
channel_mult: [ 1, 2, 4, 4 ]
|
| 41 |
+
num_heads: 8
|
| 42 |
+
use_spatial_transformer: True
|
| 43 |
+
transformer_depth: 1
|
| 44 |
+
context_dim: 768
|
| 45 |
+
use_checkpoint: True
|
| 46 |
+
legacy: False
|
| 47 |
+
|
| 48 |
+
first_stage_config:
|
| 49 |
+
target: ldm.models.autoencoder.AutoencoderKL
|
| 50 |
+
params:
|
| 51 |
+
embed_dim: 4
|
| 52 |
+
monitor: val/rec_loss
|
| 53 |
+
ddconfig:
|
| 54 |
+
double_z: true
|
| 55 |
+
z_channels: 4
|
| 56 |
+
resolution: 256
|
| 57 |
+
in_channels: 3
|
| 58 |
+
out_ch: 3
|
| 59 |
+
ch: 128
|
| 60 |
+
ch_mult:
|
| 61 |
+
- 1
|
| 62 |
+
- 2
|
| 63 |
+
- 4
|
| 64 |
+
- 4
|
| 65 |
+
num_res_blocks: 2
|
| 66 |
+
attn_resolutions: []
|
| 67 |
+
dropout: 0.0
|
| 68 |
+
lossconfig:
|
| 69 |
+
target: torch.nn.Identity
|
| 70 |
+
|
| 71 |
+
cond_stage_config:
|
| 72 |
+
target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
|
EasyFluffV7.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3312f299e3c7badc53b0bba562e9f29dce5c80ce07c89f9b1ccd20d8e7ea76e4
|
| 3 |
+
size 2132640440
|
EasyFluffV7.yaml
ADDED
|
@@ -0,0 +1,72 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
model:
|
| 2 |
+
base_learning_rate: 1.0e-04
|
| 3 |
+
target: ldm.models.diffusion.ddpm.LatentDiffusion
|
| 4 |
+
params:
|
| 5 |
+
parameterization: "v"
|
| 6 |
+
linear_start: 0.00085
|
| 7 |
+
linear_end: 0.0120
|
| 8 |
+
num_timesteps_cond: 1
|
| 9 |
+
log_every_t: 200
|
| 10 |
+
timesteps: 1000
|
| 11 |
+
first_stage_key: "jpg"
|
| 12 |
+
cond_stage_key: "txt"
|
| 13 |
+
image_size: 64
|
| 14 |
+
channels: 4
|
| 15 |
+
cond_stage_trainable: false # Note: different from the one we trained before
|
| 16 |
+
conditioning_key: crossattn
|
| 17 |
+
monitor: val/loss_simple_ema
|
| 18 |
+
scale_factor: 0.18215
|
| 19 |
+
use_ema: False
|
| 20 |
+
|
| 21 |
+
scheduler_config: # 10000 warmup steps
|
| 22 |
+
target: ldm.lr_scheduler.LambdaLinearScheduler
|
| 23 |
+
params:
|
| 24 |
+
warm_up_steps: [ 10000 ]
|
| 25 |
+
cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
|
| 26 |
+
f_start: [ 1.e-6 ]
|
| 27 |
+
f_max: [ 1. ]
|
| 28 |
+
f_min: [ 1. ]
|
| 29 |
+
|
| 30 |
+
unet_config:
|
| 31 |
+
target: ldm.modules.diffusionmodules.openaimodel.UNetModel
|
| 32 |
+
params:
|
| 33 |
+
use_fp16: True
|
| 34 |
+
image_size: 32 # unused
|
| 35 |
+
in_channels: 4
|
| 36 |
+
out_channels: 4
|
| 37 |
+
model_channels: 320
|
| 38 |
+
attention_resolutions: [ 4, 2, 1 ]
|
| 39 |
+
num_res_blocks: 2
|
| 40 |
+
channel_mult: [ 1, 2, 4, 4 ]
|
| 41 |
+
num_heads: 8
|
| 42 |
+
use_spatial_transformer: True
|
| 43 |
+
transformer_depth: 1
|
| 44 |
+
context_dim: 768
|
| 45 |
+
use_checkpoint: True
|
| 46 |
+
legacy: False
|
| 47 |
+
|
| 48 |
+
first_stage_config:
|
| 49 |
+
target: ldm.models.autoencoder.AutoencoderKL
|
| 50 |
+
params:
|
| 51 |
+
embed_dim: 4
|
| 52 |
+
monitor: val/rec_loss
|
| 53 |
+
ddconfig:
|
| 54 |
+
double_z: true
|
| 55 |
+
z_channels: 4
|
| 56 |
+
resolution: 256
|
| 57 |
+
in_channels: 3
|
| 58 |
+
out_ch: 3
|
| 59 |
+
ch: 128
|
| 60 |
+
ch_mult:
|
| 61 |
+
- 1
|
| 62 |
+
- 2
|
| 63 |
+
- 4
|
| 64 |
+
- 4
|
| 65 |
+
num_res_blocks: 2
|
| 66 |
+
attn_resolutions: []
|
| 67 |
+
dropout: 0.0
|
| 68 |
+
lossconfig:
|
| 69 |
+
target: torch.nn.Identity
|
| 70 |
+
|
| 71 |
+
cond_stage_config:
|
| 72 |
+
target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
|
EasyFluffV8.1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:664543b577412608b8c39ec7c26731cb23745714c46734f654afa9d7ed875e53
|
| 3 |
+
size 2132634704
|
EasyFluffV8.1.yaml
ADDED
|
@@ -0,0 +1,72 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
model:
|
| 2 |
+
base_learning_rate: 1.0e-04
|
| 3 |
+
target: ldm.models.diffusion.ddpm.LatentDiffusion
|
| 4 |
+
params:
|
| 5 |
+
parameterization: "v"
|
| 6 |
+
linear_start: 0.00085
|
| 7 |
+
linear_end: 0.0120
|
| 8 |
+
num_timesteps_cond: 1
|
| 9 |
+
log_every_t: 200
|
| 10 |
+
timesteps: 1000
|
| 11 |
+
first_stage_key: "jpg"
|
| 12 |
+
cond_stage_key: "txt"
|
| 13 |
+
image_size: 64
|
| 14 |
+
channels: 4
|
| 15 |
+
cond_stage_trainable: false # Note: different from the one we trained before
|
| 16 |
+
conditioning_key: crossattn
|
| 17 |
+
monitor: val/loss_simple_ema
|
| 18 |
+
scale_factor: 0.18215
|
| 19 |
+
use_ema: False
|
| 20 |
+
|
| 21 |
+
scheduler_config: # 10000 warmup steps
|
| 22 |
+
target: ldm.lr_scheduler.LambdaLinearScheduler
|
| 23 |
+
params:
|
| 24 |
+
warm_up_steps: [ 10000 ]
|
| 25 |
+
cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
|
| 26 |
+
f_start: [ 1.e-6 ]
|
| 27 |
+
f_max: [ 1. ]
|
| 28 |
+
f_min: [ 1. ]
|
| 29 |
+
|
| 30 |
+
unet_config:
|
| 31 |
+
target: ldm.modules.diffusionmodules.openaimodel.UNetModel
|
| 32 |
+
params:
|
| 33 |
+
use_fp16: True
|
| 34 |
+
image_size: 32 # unused
|
| 35 |
+
in_channels: 4
|
| 36 |
+
out_channels: 4
|
| 37 |
+
model_channels: 320
|
| 38 |
+
attention_resolutions: [ 4, 2, 1 ]
|
| 39 |
+
num_res_blocks: 2
|
| 40 |
+
channel_mult: [ 1, 2, 4, 4 ]
|
| 41 |
+
num_heads: 8
|
| 42 |
+
use_spatial_transformer: True
|
| 43 |
+
transformer_depth: 1
|
| 44 |
+
context_dim: 768
|
| 45 |
+
use_checkpoint: True
|
| 46 |
+
legacy: False
|
| 47 |
+
|
| 48 |
+
first_stage_config:
|
| 49 |
+
target: ldm.models.autoencoder.AutoencoderKL
|
| 50 |
+
params:
|
| 51 |
+
embed_dim: 4
|
| 52 |
+
monitor: val/rec_loss
|
| 53 |
+
ddconfig:
|
| 54 |
+
double_z: true
|
| 55 |
+
z_channels: 4
|
| 56 |
+
resolution: 256
|
| 57 |
+
in_channels: 3
|
| 58 |
+
out_ch: 3
|
| 59 |
+
ch: 128
|
| 60 |
+
ch_mult:
|
| 61 |
+
- 1
|
| 62 |
+
- 2
|
| 63 |
+
- 4
|
| 64 |
+
- 4
|
| 65 |
+
num_res_blocks: 2
|
| 66 |
+
attn_resolutions: []
|
| 67 |
+
dropout: 0.0
|
| 68 |
+
lossconfig:
|
| 69 |
+
target: torch.nn.Identity
|
| 70 |
+
|
| 71 |
+
cond_stage_config:
|
| 72 |
+
target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
|
EasyFluffV8.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8f76f92e8b19164629183dd6c2c8cd21d75878d31947fc6cf155a858e501b38d
|
| 3 |
+
size 2132634696
|
EasyFluffV8.yaml
ADDED
|
@@ -0,0 +1,72 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
model:
|
| 2 |
+
base_learning_rate: 1.0e-04
|
| 3 |
+
target: ldm.models.diffusion.ddpm.LatentDiffusion
|
| 4 |
+
params:
|
| 5 |
+
parameterization: "v"
|
| 6 |
+
linear_start: 0.00085
|
| 7 |
+
linear_end: 0.0120
|
| 8 |
+
num_timesteps_cond: 1
|
| 9 |
+
log_every_t: 200
|
| 10 |
+
timesteps: 1000
|
| 11 |
+
first_stage_key: "jpg"
|
| 12 |
+
cond_stage_key: "txt"
|
| 13 |
+
image_size: 64
|
| 14 |
+
channels: 4
|
| 15 |
+
cond_stage_trainable: false # Note: different from the one we trained before
|
| 16 |
+
conditioning_key: crossattn
|
| 17 |
+
monitor: val/loss_simple_ema
|
| 18 |
+
scale_factor: 0.18215
|
| 19 |
+
use_ema: False
|
| 20 |
+
|
| 21 |
+
scheduler_config: # 10000 warmup steps
|
| 22 |
+
target: ldm.lr_scheduler.LambdaLinearScheduler
|
| 23 |
+
params:
|
| 24 |
+
warm_up_steps: [ 10000 ]
|
| 25 |
+
cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
|
| 26 |
+
f_start: [ 1.e-6 ]
|
| 27 |
+
f_max: [ 1. ]
|
| 28 |
+
f_min: [ 1. ]
|
| 29 |
+
|
| 30 |
+
unet_config:
|
| 31 |
+
target: ldm.modules.diffusionmodules.openaimodel.UNetModel
|
| 32 |
+
params:
|
| 33 |
+
use_fp16: True
|
| 34 |
+
image_size: 32 # unused
|
| 35 |
+
in_channels: 4
|
| 36 |
+
out_channels: 4
|
| 37 |
+
model_channels: 320
|
| 38 |
+
attention_resolutions: [ 4, 2, 1 ]
|
| 39 |
+
num_res_blocks: 2
|
| 40 |
+
channel_mult: [ 1, 2, 4, 4 ]
|
| 41 |
+
num_heads: 8
|
| 42 |
+
use_spatial_transformer: True
|
| 43 |
+
transformer_depth: 1
|
| 44 |
+
context_dim: 768
|
| 45 |
+
use_checkpoint: True
|
| 46 |
+
legacy: False
|
| 47 |
+
|
| 48 |
+
first_stage_config:
|
| 49 |
+
target: ldm.models.autoencoder.AutoencoderKL
|
| 50 |
+
params:
|
| 51 |
+
embed_dim: 4
|
| 52 |
+
monitor: val/rec_loss
|
| 53 |
+
ddconfig:
|
| 54 |
+
double_z: true
|
| 55 |
+
z_channels: 4
|
| 56 |
+
resolution: 256
|
| 57 |
+
in_channels: 3
|
| 58 |
+
out_ch: 3
|
| 59 |
+
ch: 128
|
| 60 |
+
ch_mult:
|
| 61 |
+
- 1
|
| 62 |
+
- 2
|
| 63 |
+
- 4
|
| 64 |
+
- 4
|
| 65 |
+
num_res_blocks: 2
|
| 66 |
+
attn_resolutions: []
|
| 67 |
+
dropout: 0.0
|
| 68 |
+
lossconfig:
|
| 69 |
+
target: torch.nn.Identity
|
| 70 |
+
|
| 71 |
+
cond_stage_config:
|
| 72 |
+
target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
|
EasyFluffV9.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ab52d21e8bc7a0d78e66e7ef34a92712aedc38a3ebfa4ebb0e3ef720c3284421
|
| 3 |
+
size 2132638784
|