bimabk commited on
Commit
9abd8d2
·
verified ·
1 Parent(s): ab20bb1

Upload task output 5f0c9bca-9669-41e6-82b0-b772ea64cfc1

Browse files
checkpoints/last/config.yaml CHANGED
@@ -10,8 +10,8 @@ config:
10
  name_or_path: /cache/models/GraydientPlatformAPI--realism-engine2-xl
11
  quantize: true
12
  network:
13
- linear: 16
14
- linear_alpha: 32
15
  type: lora
16
  sample:
17
  guidance_scale: 7.5
@@ -26,24 +26,20 @@ config:
26
  walk_seed: true
27
  width: 1024
28
  train:
29
- batch_size: 1
30
  dtype: bf16
31
  ema_config:
32
  ema_decay: 0.99
33
  use_ema: true
34
  gradient_accumulation_steps: 1
35
  gradient_checkpointing: true
36
- lr: 2.5014644196170985e-06
37
  lr_scheduler: cosine
38
- lr_warmup_steps: 74
39
  noise_scheduler: flowmatch
40
- optimizer: prodigy
41
  optimizer_params:
42
- d_coef: 1.0
43
- decouple: true
44
- safeguard_warmup: true
45
- use_bias_correction: true
46
- weight_decay: 0.001498059175741326
47
  steps: 2000
48
  train_text_encoder: false
49
  train_unet: true
 
10
  name_or_path: /cache/models/GraydientPlatformAPI--realism-engine2-xl
11
  quantize: true
12
  network:
13
+ linear: 64
14
+ linear_alpha: 64
15
  type: lora
16
  sample:
17
  guidance_scale: 7.5
 
26
  walk_seed: true
27
  width: 1024
28
  train:
29
+ batch_size: 2
30
  dtype: bf16
31
  ema_config:
32
  ema_decay: 0.99
33
  use_ema: true
34
  gradient_accumulation_steps: 1
35
  gradient_checkpointing: true
36
+ lr: 2.7699350639476486e-05
37
  lr_scheduler: cosine
38
+ lr_warmup_steps: 497
39
  noise_scheduler: flowmatch
40
+ optimizer: adamw8bit
41
  optimizer_params:
42
+ weight_decay: 0.0006599631153988919
 
 
 
 
43
  steps: 2000
44
  train_text_encoder: false
45
  train_unet: true
checkpoints/last/samples/1768624370213__000000000_0.jpg ADDED