ThalisAI commited on
Commit
410661f
·
verified ·
1 Parent(s): af95fd5

Upload train_v6.sh with huggingface_hub

Browse files
Files changed (1) hide show
  1. train_v6.sh +47 -0
train_v6.sh ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ DATASET_CONFIG=/mnt/training/config/graveyard-ghoulz/v06-musubi/config.toml
2
+ MODEL_BASE=/mnt/data/models/musubi
3
+
4
+ python src/musubi_tuner/qwen_image_cache_latents.py \
5
+ --dataset_config $DATASET_CONFIG \
6
+ --vae $MODEL_BASE/vae.safetensors \
7
+ --skip_existing
8
+
9
+ python src/musubi_tuner/qwen_image_cache_text_encoder_outputs.py \
10
+ --dataset_config $DATASET_CONFIG \
11
+ --text_encoder $MODEL_BASE/text_encoder.safetensors \
12
+ --batch_size 8 \
13
+ --skip_existing
14
+
15
+ accelerate launch \
16
+ --num_cpu_threads_per_process 1 \
17
+ --mixed_precision bf16 \
18
+ src/musubi_tuner/qwen_image_train.py \
19
+ --dit $MODEL_BASE/unet.safetensors \
20
+ --vae $MODEL_BASE/vae.safetensors \
21
+ --text_encoder $MODEL_BASE/text_encoder.safetensors \
22
+ --dataset_config $DATASET_CONFIG \
23
+ --sdpa \
24
+ --mixed_precision bf16 \
25
+ --full_bf16 \
26
+ --gradient_accumulation_steps 8 \
27
+ --gradient_checkpointing \
28
+ --timestep_sampling shift \
29
+ --weighting_scheme none \
30
+ --discrete_flow_shift 2.2 \
31
+ --optimizer_type adafactor \
32
+ --learning_rate 1e-5 \
33
+ --fused_backward_pass \
34
+ --optimizer_args "relative_step=False" "scale_parameter=False" "warmup_init=False" \
35
+ --max_grad_norm 0 \
36
+ --lr_scheduler cosine \
37
+ --lr_warmup_steps 0.05 \
38
+ --max_data_loader_n_workers 0 \
39
+ --max_train_epochs 5 \
40
+ --save_every_n_epochs 1 \
41
+ --seed 42 \
42
+ --output_dir /mnt/training/output/musubi \
43
+ --output_name graveyard_ghoulz_musubi_fine_v6 \
44
+ --log_with wandb \
45
+ --wandb_run_name graveyard_ghoulz_musubi_fine_v6 \
46
+ --save_state \
47
+ --save_every_n_steps 250