File size: 803 Bytes
0d0ff25 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 |
#!/bin/bash
torchrun --nproc_per_node=8 train.py \
--model_name_or_path tiiuae/falcon-7b \
--bf16 True \
--output_dir ./out_dir/ \
--cache_dir ./hf-cache/ \
--num_train_epochs 1 \
--per_device_train_batch_size 2 \
--per_device_eval_batch_size 2 \
--gradient_accumulation_steps 8 \
--evaluation_strategy "no" \
--save_strategy "steps" \
--save_steps 2000 \
--save_total_limit 2 \
--learning_rate 2e-5 \
--weight_decay 0.1 \
--warmup_ratio 0.03 \
--lr_scheduler_type "cosine" \
--logging_steps 1 \
--tf32 True \
--max_steps 15000 \
--model_max_length 2048 \
--mem_freq 31 \
--fsdp "full_shard auto_wrap" \
--fsdp_transformer_layer_cls_to_wrap 'DecoderLayer'
|