#!/bin/bash # Create output directory OUTPUT_DIR="./solana_model_output" mkdir -p $OUTPUT_DIR # Run training script python finetune_deepseekcoder.py \ --model_name_or_path "deepseek-ai/deepseek-coder-6.7b-instruct" \ --data_path "./solana_1000_yourgpt.jsonl" \ --output_dir $OUTPUT_DIR \ --num_train_epochs 3 \ --per_device_train_batch_size 4 \ --gradient_accumulation_steps 4 \ --learning_rate 2e-5 \ --warmup_ratio 0.03 \ --logging_steps 10 \ --save_steps 100 \ --save_total_limit 3 \ --lr_scheduler_type "cosine" \ --fp16 true \ --report_to "wandb" \ --overwrite_output_dir \ --ddp_find_unused_parameters false \ --gradient_checkpointing true