| | export MODEL_NAME="models/Diffusion_Transformer/Wan2.2-I2V-A14B" |
| | export DATASET_NAME="datasets/internal_datasets/" |
| | export DATASET_META_NAME="datasets/internal_datasets/metadata.json" |
| | |
| | |
| | |
| | NCCL_DEBUG=INFO |
| |
|
| | accelerate launch --mixed_precision="bf16" scripts/wan2.2/train_distill_lora.py \ |
| | --config_path="config/wan2.2/wan_civitai_i2v.yaml" \ |
| | --pretrained_model_name_or_path=$MODEL_NAME \ |
| | --train_data_dir=$DATASET_NAME \ |
| | --train_data_meta=$DATASET_META_NAME \ |
| | --image_sample_size=640 \ |
| | --video_sample_size=640 \ |
| | --token_sample_size=640 \ |
| | --video_sample_stride=2 \ |
| | --video_sample_n_frames=81 \ |
| | --train_batch_size=1 \ |
| | --video_repeat=1 \ |
| | --gradient_accumulation_steps=1 \ |
| | --dataloader_num_workers=8 \ |
| | --num_train_epochs=100 \ |
| | --checkpointing_steps=50 \ |
| | --learning_rate=1e-05 \ |
| | --learning_rate_critic=1e-06 \ |
| | --seed=42 \ |
| | --output_dir="output_dir_distill" \ |
| | --gradient_checkpointing \ |
| | --mixed_precision="bf16" \ |
| | --adam_weight_decay=3e-2 \ |
| | --adam_epsilon=1e-10 \ |
| | --vae_mini_batch=1 \ |
| | --max_grad_norm=0.05 \ |
| | --random_hw_adapt \ |
| | --training_with_video_token_length \ |
| | --enable_bucket \ |
| | --uniform_sampling \ |
| | --boundary_type="low" \ |
| | --rank=64 \ |
| | --network_alpha=32 \ |
| | --target_name="q,k,v,ffn.0,ffn.2" \ |
| | --use_peft_lora \ |
| | --train_mode="i2v" \ |
| | --low_vram |