model_id: ./models/final_ft_2_epochs_lr1e-05_llava-phi_full model_family: llava-phi LoRA: r: 0 alpha: 128 dropout: 0.05 loss_type: grad_ascent tune_vision_tower: false tune_mm_projector: true tune_language_model: true data_path: ./dataset/full.json split: full batch_size: 8 gradient_accumulation_steps: 4 max_grad_norm: 1.0 num_epochs: 2 save_dir: models/final_ft_${num_epochs}_epochs_lr${lr}_${model_family}_${split} save_steps: 180 lr: 1.0e-05 weight_decay: 0 seed: 233 workers: 4 lr_scheduler_type: cosine warmup_ratio: 0 max_train_steps: -1 report_to: wandb resume_from_checkpoint: ''