| # Create output directory | |
| OUTPUT_DIR="./solana_model_output" | |
| mkdir -p $OUTPUT_DIR | |
| # Run training script | |
| python finetune_deepseekcoder.py \ | |
| --model_name_or_path "deepseek-ai/deepseek-coder-6.7b-instruct" \ | |
| --data_path "./solana_1000_yourgpt.jsonl" \ | |
| --output_dir $OUTPUT_DIR \ | |
| --num_train_epochs 3 \ | |
| --per_device_train_batch_size 4 \ | |
| --gradient_accumulation_steps 4 \ | |
| --learning_rate 2e-5 \ | |
| --warmup_ratio 0.03 \ | |
| --logging_steps 10 \ | |
| --save_steps 100 \ | |
| --save_total_limit 3 \ | |
| --lr_scheduler_type "cosine" \ | |
| --fp16 true \ | |
| --report_to "wandb" \ | |
| --overwrite_output_dir \ | |
| --ddp_find_unused_parameters false \ | |
| --gradient_checkpointing true | |