File size: 1,210 Bytes
8a506a6
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
export WANDB_MODE=disabled
export CUDA_LAUNCH_BLOCKING=1

DATASET=Instruments
BASE_MODEL=/datain/v-yinju/llama-7b
INTER=/datain/v-yinju/LLMBased_Multimodal_RS/Data/Musical_Instruments/Musical_Instruments.inters.numerical.json
FEATURE=/datain/v-yinju/LLMBased_Multimodal_RS/Data/Musical_Instruments/Musical_Instruments.features.numerical.json
INDEX=/datain/v-yinju/LLMBased_Multimodal_RS/Data/Musical_Instruments/Rqvae/Text/indices.json
OUTPUT_DIR=/datain/v-yinju/LLMBased_Multimodal_RS/Model_LLaMA/Ins

mkdir -p $OUTPUT_DIR

torchrun --nproc_per_node=2 lora_finetune.py \
    --base_model $BASE_MODEL \
    --output_dir $OUTPUT_DIR \
    --inter_path $INTER \
    --feature_path $FEATURE \
    --per_device_batch_size 6 \
    --gradient_accumulation_steps 2 \
    --learning_rate 5e-5 \
    --epochs 4 \
    --weight_decay 0.01 \
    --save_and_eval_strategy epoch \
    --fp16 \
    --deepspeed ./config/ds_z2_fp16.json \
    --dataloader_num_workers 4 \
    --only_train_response \
    --tasks seqrec,item2index,index2item,fusionseqrec \
    --train_prompt_sample_num 1,1,1,1 \
    --train_data_sample_num 0,0,0,0 \
    --index_file $INDEX

cd convert
nohup ./convert.sh $OUTPUT_DIR >convert.log 2>&1 &
cd ..