| export WANDB_MODE=disabled | |
| export CUDA_LAUNCH_BLOCKING=1 | |
| DATASET=Games | |
| BASE_MODEL= huggyllama/llama-7b | |
| DATA_PATH=./data | |
| OUTPUT_DIR=./ckpt/$DATASET/ | |
| torchrun --nproc_per_node=8 --master_port=3324 finetune.py \ | |
| --base_model $BASE_MODEL \ | |
| --output_dir $OUTPUT_DIR \ | |
| --dataset $DATASET \ | |
| --data_path $DATA_PATH \ | |
| --per_device_batch_size 8 \ | |
| --gradient_accumulation_steps 2 \ | |
| --learning_rate 5e-5 \ | |
| --epochs 4 \ | |
| --weight_decay 0.01 \ | |
| --save_and_eval_strategy epoch \ | |
| --deepspeed ./config/ds_z3_bf16.json \ | |
| --bf16 \ | |
| --only_train_response \ | |
| --tasks seqrec,item2index,index2item,fusionseqrec,itemsearch,preferenceobtain \ | |
| --train_prompt_sample_num 1,1,1,1,1,1 \ | |
| --train_data_sample_num 0,0,0,100000,0,0 \ | |
| --index_file .index.json | |
| cd convert | |
| nohup ./convert.sh $OUTPUT_DIR >convert.log 2>&1 & | |
| cd .. | |
| DATASET=Arts | |
| BASE_MODEL= huggyllama/llama-7b | |
| DATA_PATH=./data | |
| OUTPUT_DIR=./ckpt/$DATASET/ | |
| torchrun --nproc_per_node=8 --master_port=3324 finetune.py \ | |
| --base_model $BASE_MODEL \ | |
| --output_dir $OUTPUT_DIR \ | |
| --dataset $DATASET \ | |
| --data_path $DATA_PATH \ | |
| --per_device_batch_size 8 \ | |
| --gradient_accumulation_steps 2 \ | |
| --learning_rate 5e-5 \ | |
| --epochs 4 \ | |
| --weight_decay 0.01 \ | |
| --save_and_eval_strategy epoch \ | |
| --deepspeed ./config/ds_z3_bf16.json \ | |
| --bf16 \ | |
| --only_train_response \ | |
| --tasks seqrec,item2index,index2item,fusionseqrec,itemsearch,preferenceobtain \ | |
| --train_prompt_sample_num 1,1,1,1,1,1 \ | |
| --train_data_sample_num 0,0,0,30000,0,0 \ | |
| --index_file .index.json | |
| cd convert | |
| nohup ./convert.sh $OUTPUT_DIR >convert.log 2>&1 & | |
| cd .. | |
| DATASET=Instruments | |
| BASE_MODEL= huggyllama/llama-7b | |
| DATA_PATH=./data | |
| OUTPUT_DIR=./ckpt/$DATASET/ | |
| torchrun --nproc_per_node=8 --master_port=3324 finetune.py \ | |
| --base_model $BASE_MODEL \ | |
| --output_dir $OUTPUT_DIR \ | |
| --dataset $DATASET \ | |
| --data_path $DATA_PATH \ | |
| --per_device_batch_size 8 \ | |
| --gradient_accumulation_steps 2 \ | |
| --learning_rate 5e-5 \ | |
| --epochs 4 \ | |
| --weight_decay 0.01 \ | |
| --save_and_eval_strategy epoch \ | |
| --deepspeed ./config/ds_z3_bf16.json \ | |
| --bf16 \ | |
| --only_train_response \ | |
| --tasks seqrec,item2index,index2item,fusionseqrec,itemsearch,preferenceobtain \ | |
| --train_prompt_sample_num 1,1,1,1,1,1 \ | |
| --train_data_sample_num 0,0,0,20000,0,0 \ | |
| --index_file .index.json | |
| cd convert | |
| nohup ./convert.sh $OUTPUT_DIR >convert.log 2>&1 & | |
| cd .. | |