YoonaAI commited on
Commit
c51487a
·
1 Parent(s): 1971a45

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +35 -25
README.md CHANGED
@@ -1,61 +1,71 @@
1
  ---
2
  widget:
3
- - text: "Jens Peter Hansen kommer fra Danmark"
4
-
5
  model:
6
  target: cldm.cldm.ControlLDM
7
  params:
8
  linear_start: 0.00085
9
- linear_end: 0.0120
10
  num_timesteps_cond: 1
11
  log_every_t: 200
12
  timesteps: 1000
13
- first_stage_key: "jpg"
14
- cond_stage_key: "txt"
15
- control_key: "hint"
16
  image_size: 64
17
  channels: 4
18
  cond_stage_trainable: false
19
  conditioning_key: crossattn
20
  monitor: val/loss_simple_ema
21
  scale_factor: 0.18215
22
- use_ema: False
23
- only_mid_control: False
24
-
25
  control_stage_config:
26
  target: cldm.cldm.ControlNet
27
  params:
28
- image_size: 32 # unused
29
  in_channels: 4
30
  hint_channels: 3
31
  model_channels: 320
32
- attention_resolutions: [ 4, 2, 1 ]
 
 
 
33
  num_res_blocks: 2
34
- channel_mult: [ 1, 2, 4, 4 ]
 
 
 
 
35
  num_heads: 8
36
- use_spatial_transformer: True
37
  transformer_depth: 1
38
  context_dim: 768
39
- use_checkpoint: True
40
- legacy: False
41
-
42
  unet_config:
43
  target: cldm.cldm.ControlledUnetModel
44
  params:
45
- image_size: 32 # unused
46
  in_channels: 4
47
  out_channels: 4
48
  model_channels: 320
49
- attention_resolutions: [ 4, 2, 1 ]
 
 
 
50
  num_res_blocks: 2
51
- channel_mult: [ 1, 2, 4, 4 ]
 
 
 
 
52
  num_heads: 8
53
- use_spatial_transformer: True
54
  transformer_depth: 1
55
  context_dim: 768
56
- use_checkpoint: True
57
- legacy: False
58
-
59
  first_stage_config:
60
  target: ldm.models.autoencoder.AutoencoderKL
61
  params:
@@ -75,10 +85,10 @@ model:
75
  - 4
76
  num_res_blocks: 2
77
  attn_resolutions: []
78
- dropout: 0.0
79
  lossconfig:
80
  target: torch.nn.Identity
81
-
82
  cond_stage_config:
83
  target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
 
84
  ---
 
1
  ---
2
  widget:
3
+ - text: Jens Peter Hansen kommer fra Danmark
 
4
  model:
5
  target: cldm.cldm.ControlLDM
6
  params:
7
  linear_start: 0.00085
8
+ linear_end: 0.012
9
  num_timesteps_cond: 1
10
  log_every_t: 200
11
  timesteps: 1000
12
+ first_stage_key: jpg
13
+ cond_stage_key: txt
14
+ control_key: hint
15
  image_size: 64
16
  channels: 4
17
  cond_stage_trainable: false
18
  conditioning_key: crossattn
19
  monitor: val/loss_simple_ema
20
  scale_factor: 0.18215
21
+ use_ema: false
22
+ only_mid_control: false
 
23
  control_stage_config:
24
  target: cldm.cldm.ControlNet
25
  params:
26
+ image_size: 32
27
  in_channels: 4
28
  hint_channels: 3
29
  model_channels: 320
30
+ attention_resolutions:
31
+ - 4
32
+ - 2
33
+ - 1
34
  num_res_blocks: 2
35
+ channel_mult:
36
+ - 1
37
+ - 2
38
+ - 4
39
+ - 4
40
  num_heads: 8
41
+ use_spatial_transformer: true
42
  transformer_depth: 1
43
  context_dim: 768
44
+ use_checkpoint: true
45
+ legacy: false
 
46
  unet_config:
47
  target: cldm.cldm.ControlledUnetModel
48
  params:
49
+ image_size: 32
50
  in_channels: 4
51
  out_channels: 4
52
  model_channels: 320
53
+ attention_resolutions:
54
+ - 4
55
+ - 2
56
+ - 1
57
  num_res_blocks: 2
58
+ channel_mult:
59
+ - 1
60
+ - 2
61
+ - 4
62
+ - 4
63
  num_heads: 8
64
+ use_spatial_transformer: true
65
  transformer_depth: 1
66
  context_dim: 768
67
+ use_checkpoint: true
68
+ legacy: false
 
69
  first_stage_config:
70
  target: ldm.models.autoencoder.AutoencoderKL
71
  params:
 
85
  - 4
86
  num_res_blocks: 2
87
  attn_resolutions: []
88
+ dropout: 0
89
  lossconfig:
90
  target: torch.nn.Identity
 
91
  cond_stage_config:
92
  target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
93
+ pipeline_tag: text-to-image
94
  ---