| # BASE_MODEL="facebook/opt-125m" #"meta-llama/Llama-2-7b-hf facebook/opt-125m | |
| BASE_MODEL="meta-llama/Llama-2-7b-hf" | |
| # DATA_PATH="./data/MetaMathQA.json" | |
| DATA_PATH="./data/MetaMathQA-40K.json" | |
| OUTPUT="output/cp3e5" | |
| export WANDB_PROJECT="HRA_MetaMath395" | |
| # python finetune_32.py \ | |
| # --model_name_or_path $BASE_MODEL \ | |
| # --output_dir $OUTPUT \ | |
| # --hrft_r 32 \ | |
| # --init_a 1e-4 \ | |
| # --eps 1e-4 \ | |
| # --add_orth "none" \ | |
| # --lamda 1e-4 \ | |
| # --data_path $DATA_PATH \ | |
| # --dataset_split "train[:100000]"\ | |
| # --dataset_field query response \ | |
| # --num_train_epochs 2 \ | |
| # --per_device_train_batch_size 8 \ | |
| # --gradient_accumulation_steps 4 \ | |
| # --save_strategy "steps" \ | |
| # --save_steps 0 \ | |
| # --save_total_limit 1 \ | |
| # --learning_rate 3e-5 \ | |
| # --weight_decay 0. \ | |
| # --warmup_ratio 0.005 \ | |
| # --lr_scheduler_type "cosine" \ | |
| # --logging_steps 1000 \ | |
| # --bf16 True \ | |
| # --tf32 True \ | |
| # --report_to "none" \ | |
| # wandb sync wandb/latest-run | |
| # OUTPUT="output/cp1e5N" | |
| # python finetune_32.py \ | |
| # --model_name_or_path $BASE_MODEL \ | |
| # --output_dir $OUTPUT \ | |
| # --hrft_r 32 \ | |
| # --init_a 1e-4 \ | |
| # --eps 1e-4 \ | |
| # --add_orth "none" \ | |
| # --lamda 1e-4 \ | |
| # --data_path $DATA_PATH \ | |
| # --dataset_split "train[:100000]"\ | |
| # --dataset_field query response \ | |
| # --num_train_epochs 2 \ | |
| # --per_device_train_batch_size 8 \ | |
| # --gradient_accumulation_steps 4 \ | |
| # --save_strategy "steps" \ | |
| # --save_steps 0 \ | |
| # --save_total_limit 1 \ | |
| # --learning_rate 1e-5 \ | |
| # --weight_decay 0. \ | |
| # --warmup_ratio 0.005 \ | |
| # --lr_scheduler_type "cosine" \ | |
| # --logging_steps 1000 \ | |
| # --bf16 True \ | |
| # --tf32 True \ | |
| # --report_to "wandb" | |
| # wandb sync wandb/latest-run | |
| # OUTPUT="output/cpr1" | |
| # python finetune_32.py \ | |
| # --model_name_or_path $BASE_MODEL \ | |
| # --output_dir $OUTPUT \ | |
| # --hrft_r 1 \ | |
| # --init_a 1e-4 \ | |
| # --eps 1e-4 \ | |
| # --add_orth "none" \ | |
| # --lamda 1e-4 \ | |
| # --data_path $DATA_PATH \ | |
| # --dataset_split "train"\ | |
| # --dataset_field query response \ | |
| # --num_train_epochs 2 \ | |
| # --per_device_train_batch_size 32 \ | |
| # --gradient_accumulation_steps 1 \ | |
| # --save_strategy "steps" \ | |
| # --save_steps 0 \ | |
| # --save_total_limit 1 \ | |
| # --learning_rate 3e-5 \ | |
| # --weight_decay 0. \ | |
| # --warmup_ratio 0.005 \ | |
| # --lr_scheduler_type "cosine" \ | |
| # --logging_steps 1000 \ | |
| # --bf16 True \ | |
| # --tf32 True \ | |
| # --report_to "wandb" | |
| # wandb sync wandb/latest-run | |
| # OUTPUT="output/cpr2" | |
| # python finetune_32.py \ | |
| # --model_name_or_path $BASE_MODEL \ | |
| # --output_dir $OUTPUT \ | |
| # --hrft_r 1 \ | |
| # --init_a 1e-4 \ | |
| # --eps 1e-4 \ | |
| # --add_orth "none" \ | |
| # --lamda 1e-4 \ | |
| # --data_path $DATA_PATH \ | |
| # --dataset_split "train"\ | |
| # --dataset_field query response \ | |
| # --num_train_epochs 3 \ | |
| # --per_device_train_batch_size 32 \ | |
| # --gradient_accumulation_steps 1 \ | |
| # --save_strategy "steps" \ | |
| # --save_steps 0 \ | |
| # --save_total_limit 1 \ | |
| # --learning_rate 3e-5 \ | |
| # --weight_decay 0. \ | |
| # --warmup_ratio 0.005 \ | |
| # --lr_scheduler_type "cosine" \ | |
| # --logging_steps 200 \ | |
| # --bf16 True \ | |
| # --tf32 True \ | |
| # --report_to "wandb" | |
| # wandb sync wandb/latest-run | |
| OUTPUT="output/cms3" | |
| python finetune_32.py \ | |
| --model_name_or_path $BASE_MODEL \ | |
| --output_dir $OUTPUT \ | |
| --hrft_r 32 \ | |
| --init_a 1e-4 \ | |
| --eps 1e-4 \ | |
| --add_orth "none" \ | |
| --lamda 1e-4 \ | |
| --data_path $DATA_PATH \ | |
| --dataset_split "train"\ | |
| --dataset_field query response \ | |
| --num_train_epochs 2 \ | |
| --per_device_train_batch_size 8 \ | |
| --gradient_accumulation_steps 4 \ | |
| --save_strategy "steps" \ | |
| --save_steps 0 \ | |
| --save_total_limit 1 \ | |
| --learning_rate 1e-5 \ | |
| --weight_decay 0. \ | |
| --warmup_ratio 0.005 \ | |
| --lr_scheduler_type "cosine" \ | |
| --logging_steps 200 \ | |
| --bf16 True \ | |
| --tf32 True \ | |
| --report_to "wandb" | |
| date +"%F %T" | |
| # wandb sync wandb/latest-run |