tantanbei commited on
Commit
5c9418c
·
verified ·
1 Parent(s): b49baf9

Create train_wan_1_3b.toml

Browse files
Files changed (1) hide show
  1. train_wan_1_3b.toml +50 -0
train_wan_1_3b.toml ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # This configuration should allow you to train Wan 14b t2v on 512x512x81 sized videos (or varying aspect ratios of the same size), with 24GB VRAM.
2
+
3
+ # change this
4
+ output_dir = '/data/data/model_training/shangche_wan2_1_lora/output'
5
+ # and this
6
+ dataset = '/data/app/diffusion-pipe/examples/dataset.toml'
7
+
8
+ # training settings
9
+ epochs = 1000
10
+ micro_batch_size_per_gpu = 5
11
+ pipeline_stages = 1
12
+ gradient_accumulation_steps = 2
13
+ gradient_clipping = 1
14
+ warmup_steps = 30
15
+
16
+ # eval settings
17
+ eval_every_n_epochs = 1
18
+ eval_before_first_step = true
19
+ eval_micro_batch_size_per_gpu = 5
20
+ eval_gradient_accumulation_steps = 2
21
+
22
+ # misc settings
23
+ save_every_n_epochs = 1
24
+ checkpoint_every_n_minutes = 120
25
+ activation_checkpointing = 'unsloth'
26
+ partition_method = 'parameters'
27
+ save_dtype = 'bfloat16'
28
+ caching_batch_size = 1
29
+ steps_per_print = 1
30
+ video_clip_mode = 'single_beginning'
31
+ # blocks_to_swap = 32
32
+
33
+ [model]
34
+ type = 'wan'
35
+ ckpt_path = '/data/app/diffusion-pipe/submodules/Wan2_1/Wan2.1-T2V-1.3B'
36
+ dtype = 'bfloat16'
37
+ transformer_dtype = 'float8'
38
+ timestep_sample_method = 'logit_normal'
39
+
40
+ [adapter]
41
+ type = 'lora'
42
+ rank = 32
43
+ dtype = 'bfloat16'
44
+
45
+ [optimizer]
46
+ type = 'AdamW8bitKahan'
47
+ lr = 1e-5
48
+ betas = [0.9, 0.99]
49
+ weight_decay = 0.01
50
+ stabilize = false