bimabk commited on
Commit
96dd45f
·
verified ·
1 Parent(s): 9abd8d2

Upload task output 5f0c9bca-9669-41e6-82b0-b772ea64cfc1

Browse files
checkpoints/last/config.yaml CHANGED
@@ -1,7 +1,18 @@
1
  config:
2
  name: last
3
  process:
4
- - device: cuda:0
 
 
 
 
 
 
 
 
 
 
 
5
  model:
6
  is_flux: false
7
  is_v2: false
@@ -10,7 +21,7 @@ config:
10
  name_or_path: /cache/models/GraydientPlatformAPI--realism-engine2-xl
11
  quantize: true
12
  network:
13
- linear: 64
14
  linear_alpha: 64
15
  type: lora
16
  sample:
@@ -18,7 +29,7 @@ config:
18
  height: 1024
19
  neg: ''
20
  prompts:
21
- - '[trigger]'
22
  sample_every: 250
23
  sample_steps: 20
24
  sampler: flowmatch
@@ -26,20 +37,20 @@ config:
26
  walk_seed: true
27
  width: 1024
28
  train:
29
- batch_size: 2
30
  dtype: bf16
31
  ema_config:
32
  ema_decay: 0.99
33
  use_ema: true
34
  gradient_accumulation_steps: 1
35
  gradient_checkpointing: true
36
- lr: 2.7699350639476486e-05
37
  lr_scheduler: cosine
38
- lr_warmup_steps: 497
39
  noise_scheduler: flowmatch
40
  optimizer: adamw8bit
41
  optimizer_params:
42
- weight_decay: 0.0006599631153988919
43
  steps: 2000
44
  train_text_encoder: false
45
  train_unet: true
 
1
  config:
2
  name: last
3
  process:
4
+ - datasets:
5
+ - cache_latents_to_disk: true
6
+ caption_dropout_rate: 0.05
7
+ caption_ext: txt
8
+ folder_path: /dataset/images/5f0c9bca-9669-41e6-82b0-b772ea64cfc1/img/5_lora
9
+ style
10
+ resolution:
11
+ - 512
12
+ - 768
13
+ - 1024
14
+ shuffle_tokens: false
15
+ device: cuda:0
16
  model:
17
  is_flux: false
18
  is_v2: false
 
21
  name_or_path: /cache/models/GraydientPlatformAPI--realism-engine2-xl
22
  quantize: true
23
  network:
24
+ linear: 128
25
  linear_alpha: 64
26
  type: lora
27
  sample:
 
29
  height: 1024
30
  neg: ''
31
  prompts:
32
+ - '[trigger] a person'
33
  sample_every: 250
34
  sample_steps: 20
35
  sampler: flowmatch
 
37
  walk_seed: true
38
  width: 1024
39
  train:
40
+ batch_size: 4
41
  dtype: bf16
42
  ema_config:
43
  ema_decay: 0.99
44
  use_ema: true
45
  gradient_accumulation_steps: 1
46
  gradient_checkpointing: true
47
+ lr: 2.277636174853309e-06
48
  lr_scheduler: cosine
49
+ lr_warmup_steps: 426
50
  noise_scheduler: flowmatch
51
  optimizer: adamw8bit
52
  optimizer_params:
53
+ weight_decay: 0.0014429686228454891
54
  steps: 2000
55
  train_text_encoder: false
56
  train_unet: true
checkpoints/last/samples/1768626768296__000000000_0.jpg ADDED