| #!/bin/bash |
|
|
|
|
| |
| |
| MODEL_NAME="Qwen/Qwen2.5-VL-3B-Instruct" |
| |
|
|
| export PYTHONPATH=src:$PYTHONPATH |
|
|
| GLOBAL_BATCH_SIZE=128 |
| BATCH_PER_DEVICE=4 |
| NUM_DEVICES=8 |
| GRAD_ACCUM_STEPS=$((GLOBAL_BATCH_SIZE / (BATCH_PER_DEVICE * NUM_DEVICES))) |
|
|
| |
| |
|
|
| |
|
|
| deepspeed src/train/train_sft.py \ |
| --use_liger_kernel True \ |
| --lora_enable True \ |
| --vision_lora True \ |
| --use_dora False \ |
| --lora_namespan_exclude "['lm_head', 'embed_tokens']" \ |
| --lora_rank 32 \ |
| --lora_alpha 64 \ |
| --lora_dropout 0.05 \ |
| --num_lora_modules -1 \ |
| --deepspeed scripts/zero3.json \ |
| --model_id $MODEL_NAME \ |
| --data_path /path/to/your/training/data.json \ |
| --image_folder /path/to/your/image/folder \ |
| --remove_unused_columns False \ |
| --freeze_vision_tower True \ |
| --freeze_llm True \ |
| --freeze_merger True \ |
| --bf16 True \ |
| --fp16 False \ |
| --disable_flash_attn2 False \ |
| --output_dir output/lora_vision_test \ |
| --num_train_epochs 1 \ |
| --per_device_train_batch_size $BATCH_PER_DEVICE \ |
| --gradient_accumulation_steps $GRAD_ACCUM_STEPS \ |
| --image_min_pixels $((256 * 28 * 28)) \ |
| --image_max_pixels $((1280 * 28 * 28)) \ |
| --learning_rate 2e-4 \ |
| --weight_decay 0.1 \ |
| --warmup_ratio 0.03 \ |
| --lr_scheduler_type "cosine" \ |
| --logging_steps 1 \ |
| --tf32 True \ |
| --gradient_checkpointing False \ |
| --report_to tensorboard \ |
| --lazy_preprocess True \ |
| --save_strategy "steps" \ |
| --save_steps 200 \ |
| --save_total_limit 10 \ |
| --dataloader_num_workers 4 |