finetune-model / scripts /finetune_cls.sh
reiofa's picture
Upload folder using huggingface_hub
89b38b0 verified
#!/bin/bash
# MODEL_NAME="Qwen/Qwen2-VL-7B-Instruct"
# MODEL_NAME="Qwen/Qwen2-VL-2B-Instruct"
MODEL_NAME="Qwen/Qwen2.5-VL-3B-Instruct"
# loss_type should be one of "cross_entropy", "focal_loss", "class_balanced_cross_entropy", or "class_balanced_focal_loss".
export PYTHONPATH=src:$PYTHONPATH
deepspeed src/train/train_cls.py \
--deepspeed scripts/zero3.json \
--model_id $MODEL_NAME \
--data_path /path/to/your/training/data.json \
--image_folder /path/to/your/image/folder \
--eval_path /path/to/your/training/data.json \
--eval_image_folder /path/to/your/image/folder \
--freeze_llm True \
--freeze_vision_tower False \
--freeze_merger False \
--bf16 True \
--fp16 False \
--loss_type "cross_entropy" \
--num_labels 2 \
--disable_flash_attn2 False \
--output_dir output/qwen_cls \
--num_train_epochs 3 \
--per_device_train_batch_size 4 \
--gradient_accumulation_steps 4 \
--learning_rate 3e-5 \
--head_lr 4e-5 \
--vision_lr 6e-6 \
--merger_lr 2e-5 \
--weight_decay 0.02 \
--warmup_ratio 0.05 \
--max_grad_norm 1.0 \
--lr_scheduler_type "cosine" \
--logging_steps 1 \
--tf32 True \
--eval_strategy "epoch" \
--load_best_model_at_end True \
--metric_for_best_model "eval_f1" \
--greater_is_better True \
--gradient_checkpointing True \
--report_to tensorboard \
--lazy_preprocess True \
--save_strategy "epoch" \
--dataloader_num_workers 4